diff --git "a/checkpoint-42659/trainer_state.json" "b/checkpoint-42659/trainer_state.json" new file mode 100644--- /dev/null +++ "b/checkpoint-42659/trainer_state.json" @@ -0,0 +1,298646 @@ +{ + "best_metric": null, + "best_model_checkpoint": null, + "epoch": 0.9999941396062988, + "eval_steps": 500, + "global_step": 42659, + "is_hyper_param_search": false, + "is_local_process_zero": true, + "is_world_process_zero": true, + "log_history": [ + { + "epoch": 2.3441574804995398e-05, + "grad_norm": 0.5949042439460754, + "learning_rate": 2e-05, + "loss": 0.5845, + "step": 1 + }, + { + "epoch": 4.6883149609990796e-05, + "grad_norm": 0.46402573585510254, + "learning_rate": 4e-05, + "loss": 0.2901, + "step": 2 + }, + { + "epoch": 7.03247244149862e-05, + "grad_norm": 2.976228713989258, + "learning_rate": 6e-05, + "loss": 1.1027, + "step": 3 + }, + { + "epoch": 9.376629921998159e-05, + "grad_norm": 1.3883099555969238, + "learning_rate": 8e-05, + "loss": 0.6743, + "step": 4 + }, + { + "epoch": 0.000117207874024977, + "grad_norm": 1.573052167892456, + "learning_rate": 0.0001, + "loss": 0.7602, + "step": 5 + }, + { + "epoch": 0.0001406494488299724, + "grad_norm": 1.6482921838760376, + "learning_rate": 0.00012, + "loss": 0.8481, + "step": 6 + }, + { + "epoch": 0.0001640910236349678, + "grad_norm": 1.5320346355438232, + "learning_rate": 0.00014, + "loss": 0.6771, + "step": 7 + }, + { + "epoch": 0.00018753259843996319, + "grad_norm": 1.107053279876709, + "learning_rate": 0.00016, + "loss": 0.8183, + "step": 8 + }, + { + "epoch": 0.00021097417324495858, + "grad_norm": 1.0839240550994873, + "learning_rate": 0.00018, + "loss": 0.753, + "step": 9 + }, + { + "epoch": 0.000234415748049954, + "grad_norm": 0.2018798589706421, + "learning_rate": 0.0002, + "loss": 0.0796, + "step": 10 + }, + { + "epoch": 0.0002578573228549494, + "grad_norm": 0.6567934155464172, + "learning_rate": 0.0001999999997286986, + "loss": 0.5076, + "step": 11 + }, + { + "epoch": 0.0002812988976599448, + "grad_norm": 0.7991616725921631, + "learning_rate": 0.0001999999989147944, + "loss": 0.4111, + "step": 12 + }, + { + "epoch": 0.0003047404724649402, + "grad_norm": 0.6299276351928711, + "learning_rate": 0.00019999999755828735, + "loss": 0.3175, + "step": 13 + }, + { + "epoch": 0.0003281820472699356, + "grad_norm": 0.2816776633262634, + "learning_rate": 0.00019999999565917755, + "loss": 0.1662, + "step": 14 + }, + { + "epoch": 0.000351623622074931, + "grad_norm": 0.8360978364944458, + "learning_rate": 0.00019999999321746495, + "loss": 0.4858, + "step": 15 + }, + { + "epoch": 0.00037506519687992637, + "grad_norm": 0.9331958293914795, + "learning_rate": 0.00019999999023314954, + "loss": 0.4263, + "step": 16 + }, + { + "epoch": 0.00039850677168492177, + "grad_norm": 0.63945472240448, + "learning_rate": 0.00019999998670623142, + "loss": 0.1949, + "step": 17 + }, + { + "epoch": 0.00042194834648991717, + "grad_norm": 0.9862967133522034, + "learning_rate": 0.00019999998263671055, + "loss": 0.3715, + "step": 18 + }, + { + "epoch": 0.00044538992129491257, + "grad_norm": 0.70915287733078, + "learning_rate": 0.00019999997802458693, + "loss": 0.2862, + "step": 19 + }, + { + "epoch": 0.000468831496099908, + "grad_norm": 1.1273599863052368, + "learning_rate": 0.00019999997286986062, + "loss": 0.4973, + "step": 20 + }, + { + "epoch": 0.0004922730709049034, + "grad_norm": 0.6961119174957275, + "learning_rate": 0.00019999996717253167, + "loss": 0.3369, + "step": 21 + }, + { + "epoch": 0.0005157146457098988, + "grad_norm": 0.8777477741241455, + "learning_rate": 0.0001999999609326001, + "loss": 0.3325, + "step": 22 + }, + { + "epoch": 0.0005391562205148942, + "grad_norm": 0.293773353099823, + "learning_rate": 0.0001999999541500659, + "loss": 0.1387, + "step": 23 + }, + { + "epoch": 0.0005625977953198896, + "grad_norm": 0.963109016418457, + "learning_rate": 0.00019999994682492916, + "loss": 0.3235, + "step": 24 + }, + { + "epoch": 0.000586039370124885, + "grad_norm": 1.1176785230636597, + "learning_rate": 0.00019999993895718988, + "loss": 0.3388, + "step": 25 + }, + { + "epoch": 0.0006094809449298804, + "grad_norm": 0.7970275282859802, + "learning_rate": 0.00019999993054684812, + "loss": 0.3184, + "step": 26 + }, + { + "epoch": 0.0006329225197348758, + "grad_norm": 0.9765394330024719, + "learning_rate": 0.00019999992159390394, + "loss": 0.2695, + "step": 27 + }, + { + "epoch": 0.0006563640945398711, + "grad_norm": 1.0396114587783813, + "learning_rate": 0.00019999991209835734, + "loss": 0.4312, + "step": 28 + }, + { + "epoch": 0.0006798056693448665, + "grad_norm": 0.7726010084152222, + "learning_rate": 0.00019999990206020845, + "loss": 0.4143, + "step": 29 + }, + { + "epoch": 0.000703247244149862, + "grad_norm": 0.8144803047180176, + "learning_rate": 0.00019999989147945722, + "loss": 0.3838, + "step": 30 + }, + { + "epoch": 0.0007266888189548573, + "grad_norm": 0.2968287765979767, + "learning_rate": 0.00019999988035610382, + "loss": 0.1566, + "step": 31 + }, + { + "epoch": 0.0007501303937598527, + "grad_norm": 0.9519981145858765, + "learning_rate": 0.00019999986869014824, + "loss": 0.4405, + "step": 32 + }, + { + "epoch": 0.0007735719685648481, + "grad_norm": 0.6892004609107971, + "learning_rate": 0.00019999985648159056, + "loss": 0.2612, + "step": 33 + }, + { + "epoch": 0.0007970135433698435, + "grad_norm": 0.5350654125213623, + "learning_rate": 0.00019999984373043087, + "loss": 0.2457, + "step": 34 + }, + { + "epoch": 0.0008204551181748389, + "grad_norm": 0.7230672836303711, + "learning_rate": 0.00019999983043666916, + "loss": 0.2782, + "step": 35 + }, + { + "epoch": 0.0008438966929798343, + "grad_norm": 0.6681996583938599, + "learning_rate": 0.0001999998166003056, + "loss": 0.2386, + "step": 36 + }, + { + "epoch": 0.0008673382677848297, + "grad_norm": 0.21338987350463867, + "learning_rate": 0.0001999998022213402, + "loss": 0.0787, + "step": 37 + }, + { + "epoch": 0.0008907798425898251, + "grad_norm": 0.6774462461471558, + "learning_rate": 0.0001999997872997731, + "loss": 0.8007, + "step": 38 + }, + { + "epoch": 0.0009142214173948205, + "grad_norm": 0.6301290392875671, + "learning_rate": 0.00019999977183560432, + "loss": 0.3021, + "step": 39 + }, + { + "epoch": 0.000937662992199816, + "grad_norm": 0.883497416973114, + "learning_rate": 0.00019999975582883398, + "loss": 0.3352, + "step": 40 + }, + { + "epoch": 0.0009611045670048114, + "grad_norm": 0.6740983128547668, + "learning_rate": 0.00019999973927946213, + "loss": 0.4127, + "step": 41 + }, + { + "epoch": 0.0009845461418098067, + "grad_norm": 0.9566420912742615, + "learning_rate": 0.0001999997221874889, + "loss": 0.5157, + "step": 42 + }, + { + "epoch": 0.0010079877166148021, + "grad_norm": 0.5852053761482239, + "learning_rate": 0.00019999970455291438, + "loss": 0.3035, + "step": 43 + }, + { + "epoch": 0.0010314292914197975, + "grad_norm": 0.17220944166183472, + "learning_rate": 0.0001999996863757386, + "loss": 0.0942, + "step": 44 + }, + { + "epoch": 0.001054870866224793, + "grad_norm": 0.1584855169057846, + "learning_rate": 0.00019999966765596176, + "loss": 0.0567, + "step": 45 + }, + { + "epoch": 0.0010783124410297883, + "grad_norm": 0.45523375272750854, + "learning_rate": 0.00019999964839358388, + "loss": 0.8523, + "step": 46 + }, + { + "epoch": 0.0011017540158347837, + "grad_norm": 0.25971096754074097, + "learning_rate": 0.0001999996285886051, + "loss": 0.1482, + "step": 47 + }, + { + "epoch": 0.0011251955906397791, + "grad_norm": 0.1835508644580841, + "learning_rate": 0.00019999960824102555, + "loss": 0.0549, + "step": 48 + }, + { + "epoch": 0.0011486371654447745, + "grad_norm": 0.23798798024654388, + "learning_rate": 0.0001999995873508453, + "loss": 0.4174, + "step": 49 + }, + { + "epoch": 0.00117207874024977, + "grad_norm": 0.5296609997749329, + "learning_rate": 0.00019999956591806447, + "loss": 0.2984, + "step": 50 + }, + { + "epoch": 0.0011955203150547653, + "grad_norm": 0.3088091313838959, + "learning_rate": 0.00019999954394268315, + "loss": 0.1863, + "step": 51 + }, + { + "epoch": 0.0012189618898597607, + "grad_norm": 0.7972097992897034, + "learning_rate": 0.00019999952142470156, + "loss": 0.455, + "step": 52 + }, + { + "epoch": 0.001242403464664756, + "grad_norm": 0.5585108399391174, + "learning_rate": 0.00019999949836411971, + "loss": 0.2941, + "step": 53 + }, + { + "epoch": 0.0012658450394697515, + "grad_norm": 0.5695565342903137, + "learning_rate": 0.00019999947476093778, + "loss": 0.362, + "step": 54 + }, + { + "epoch": 0.001289286614274747, + "grad_norm": 0.7046261429786682, + "learning_rate": 0.00019999945061515591, + "loss": 0.3374, + "step": 55 + }, + { + "epoch": 0.0013127281890797423, + "grad_norm": 0.2330344319343567, + "learning_rate": 0.00019999942592677418, + "loss": 0.1368, + "step": 56 + }, + { + "epoch": 0.0013361697638847377, + "grad_norm": 0.28778040409088135, + "learning_rate": 0.00019999940069579278, + "loss": 0.1702, + "step": 57 + }, + { + "epoch": 0.001359611338689733, + "grad_norm": 0.8627858757972717, + "learning_rate": 0.0001999993749222118, + "loss": 0.4467, + "step": 58 + }, + { + "epoch": 0.0013830529134947285, + "grad_norm": 0.8555877208709717, + "learning_rate": 0.0001999993486060314, + "loss": 0.5048, + "step": 59 + }, + { + "epoch": 0.001406494488299724, + "grad_norm": 0.7191990613937378, + "learning_rate": 0.00019999932174725175, + "loss": 0.2404, + "step": 60 + }, + { + "epoch": 0.0014299360631047193, + "grad_norm": 0.6784783005714417, + "learning_rate": 0.00019999929434587292, + "loss": 0.6007, + "step": 61 + }, + { + "epoch": 0.0014533776379097147, + "grad_norm": 0.4936765432357788, + "learning_rate": 0.0001999992664018951, + "loss": 0.2649, + "step": 62 + }, + { + "epoch": 0.00147681921271471, + "grad_norm": 0.8700072765350342, + "learning_rate": 0.00019999923791531852, + "loss": 0.2812, + "step": 63 + }, + { + "epoch": 0.0015002607875197055, + "grad_norm": 0.7428555488586426, + "learning_rate": 0.00019999920888614324, + "loss": 0.4125, + "step": 64 + }, + { + "epoch": 0.0015237023623247009, + "grad_norm": 0.7245789766311646, + "learning_rate": 0.0001999991793143694, + "loss": 0.2782, + "step": 65 + }, + { + "epoch": 0.0015471439371296963, + "grad_norm": 0.3065873682498932, + "learning_rate": 0.00019999914919999724, + "loss": 0.1644, + "step": 66 + }, + { + "epoch": 0.0015705855119346917, + "grad_norm": 0.7532345056533813, + "learning_rate": 0.00019999911854302687, + "loss": 0.2962, + "step": 67 + }, + { + "epoch": 0.001594027086739687, + "grad_norm": 0.7162583470344543, + "learning_rate": 0.00019999908734345847, + "loss": 0.3131, + "step": 68 + }, + { + "epoch": 0.0016174686615446825, + "grad_norm": 0.8019320368766785, + "learning_rate": 0.00019999905560129223, + "loss": 0.2979, + "step": 69 + }, + { + "epoch": 0.0016409102363496779, + "grad_norm": 0.5570685267448425, + "learning_rate": 0.00019999902331652825, + "loss": 0.2214, + "step": 70 + }, + { + "epoch": 0.0016643518111546733, + "grad_norm": 1.1150927543640137, + "learning_rate": 0.00019999899048916682, + "loss": 0.2471, + "step": 71 + }, + { + "epoch": 0.0016877933859596687, + "grad_norm": 0.6421457529067993, + "learning_rate": 0.00019999895711920803, + "loss": 0.3986, + "step": 72 + }, + { + "epoch": 0.001711234960764664, + "grad_norm": 0.8516373634338379, + "learning_rate": 0.00019999892320665208, + "loss": 0.366, + "step": 73 + }, + { + "epoch": 0.0017346765355696595, + "grad_norm": 0.8823465704917908, + "learning_rate": 0.00019999888875149916, + "loss": 0.3996, + "step": 74 + }, + { + "epoch": 0.0017581181103746549, + "grad_norm": 0.7563062906265259, + "learning_rate": 0.00019999885375374948, + "loss": 0.4637, + "step": 75 + }, + { + "epoch": 0.0017815596851796503, + "grad_norm": 0.8985307216644287, + "learning_rate": 0.00019999881821340318, + "loss": 0.3991, + "step": 76 + }, + { + "epoch": 0.0018050012599846457, + "grad_norm": 0.5766052603721619, + "learning_rate": 0.0001999987821304605, + "loss": 0.2608, + "step": 77 + }, + { + "epoch": 0.001828442834789641, + "grad_norm": 0.677850604057312, + "learning_rate": 0.0001999987455049216, + "loss": 0.3558, + "step": 78 + }, + { + "epoch": 0.0018518844095946367, + "grad_norm": 0.4816989004611969, + "learning_rate": 0.00019999870833678672, + "loss": 0.1489, + "step": 79 + }, + { + "epoch": 0.001875325984399632, + "grad_norm": 0.7620590329170227, + "learning_rate": 0.00019999867062605602, + "loss": 0.3821, + "step": 80 + }, + { + "epoch": 0.0018987675592046275, + "grad_norm": 0.5539602637290955, + "learning_rate": 0.00019999863237272973, + "loss": 0.2767, + "step": 81 + }, + { + "epoch": 0.0019222091340096229, + "grad_norm": 0.6648173332214355, + "learning_rate": 0.00019999859357680805, + "loss": 0.3679, + "step": 82 + }, + { + "epoch": 0.0019456507088146183, + "grad_norm": 0.7620189785957336, + "learning_rate": 0.00019999855423829117, + "loss": 0.3254, + "step": 83 + }, + { + "epoch": 0.0019690922836196134, + "grad_norm": 0.34241870045661926, + "learning_rate": 0.00019999851435717933, + "loss": 0.1729, + "step": 84 + }, + { + "epoch": 0.001992533858424609, + "grad_norm": 0.8453075885772705, + "learning_rate": 0.00019999847393347274, + "loss": 0.2824, + "step": 85 + }, + { + "epoch": 0.0020159754332296042, + "grad_norm": 0.8194279670715332, + "learning_rate": 0.00019999843296717161, + "loss": 0.2742, + "step": 86 + }, + { + "epoch": 0.0020394170080346, + "grad_norm": 1.3069039583206177, + "learning_rate": 0.0001999983914582762, + "loss": 0.3597, + "step": 87 + }, + { + "epoch": 0.002062858582839595, + "grad_norm": 0.3111972510814667, + "learning_rate": 0.0001999983494067867, + "loss": 0.0914, + "step": 88 + }, + { + "epoch": 0.0020863001576445907, + "grad_norm": 0.5526110529899597, + "learning_rate": 0.00019999830681270334, + "loss": 0.6849, + "step": 89 + }, + { + "epoch": 0.002109741732449586, + "grad_norm": 0.5527598261833191, + "learning_rate": 0.00019999826367602633, + "loss": 0.2381, + "step": 90 + }, + { + "epoch": 0.0021331833072545815, + "grad_norm": 0.7751446962356567, + "learning_rate": 0.00019999821999675595, + "loss": 0.3013, + "step": 91 + }, + { + "epoch": 0.0021566248820595766, + "grad_norm": 1.0483684539794922, + "learning_rate": 0.00019999817577489242, + "loss": 0.412, + "step": 92 + }, + { + "epoch": 0.0021800664568645723, + "grad_norm": 0.7636705040931702, + "learning_rate": 0.00019999813101043598, + "loss": 0.3013, + "step": 93 + }, + { + "epoch": 0.0022035080316695674, + "grad_norm": 0.7908076047897339, + "learning_rate": 0.00019999808570338688, + "loss": 0.3503, + "step": 94 + }, + { + "epoch": 0.002226949606474563, + "grad_norm": 0.6745355129241943, + "learning_rate": 0.0001999980398537453, + "loss": 0.2197, + "step": 95 + }, + { + "epoch": 0.0022503911812795582, + "grad_norm": 0.7659450173377991, + "learning_rate": 0.0001999979934615116, + "loss": 0.2852, + "step": 96 + }, + { + "epoch": 0.002273832756084554, + "grad_norm": 0.561555802822113, + "learning_rate": 0.00019999794652668596, + "loss": 0.2563, + "step": 97 + }, + { + "epoch": 0.002297274330889549, + "grad_norm": 0.3912050127983093, + "learning_rate": 0.00019999789904926867, + "loss": 0.1224, + "step": 98 + }, + { + "epoch": 0.0023207159056945446, + "grad_norm": 0.7781040668487549, + "learning_rate": 0.00019999785102925994, + "loss": 0.7994, + "step": 99 + }, + { + "epoch": 0.00234415748049954, + "grad_norm": 0.6666799187660217, + "learning_rate": 0.00019999780246666003, + "loss": 0.346, + "step": 100 + }, + { + "epoch": 0.0023675990553045354, + "grad_norm": 0.1652546375989914, + "learning_rate": 0.00019999775336146928, + "loss": 0.0803, + "step": 101 + }, + { + "epoch": 0.0023910406301095306, + "grad_norm": 0.12952055037021637, + "learning_rate": 0.00019999770371368792, + "loss": 0.0688, + "step": 102 + }, + { + "epoch": 0.0024144822049145262, + "grad_norm": 0.6516351699829102, + "learning_rate": 0.00019999765352331615, + "loss": 0.3544, + "step": 103 + }, + { + "epoch": 0.0024379237797195214, + "grad_norm": 0.4797312915325165, + "learning_rate": 0.00019999760279035435, + "loss": 0.1793, + "step": 104 + }, + { + "epoch": 0.002461365354524517, + "grad_norm": 0.6266962289810181, + "learning_rate": 0.00019999755151480276, + "loss": 0.3958, + "step": 105 + }, + { + "epoch": 0.002484806929329512, + "grad_norm": 0.12748892605304718, + "learning_rate": 0.0001999974996966616, + "loss": 0.0647, + "step": 106 + }, + { + "epoch": 0.002508248504134508, + "grad_norm": 0.7375484704971313, + "learning_rate": 0.0001999974473359312, + "loss": 0.3804, + "step": 107 + }, + { + "epoch": 0.002531690078939503, + "grad_norm": 0.13670992851257324, + "learning_rate": 0.00019999739443261186, + "loss": 0.0543, + "step": 108 + }, + { + "epoch": 0.0025551316537444986, + "grad_norm": 0.8479056358337402, + "learning_rate": 0.00019999734098670381, + "loss": 0.3736, + "step": 109 + }, + { + "epoch": 0.002578573228549494, + "grad_norm": 0.5258725881576538, + "learning_rate": 0.00019999728699820742, + "loss": 0.276, + "step": 110 + }, + { + "epoch": 0.0026020148033544894, + "grad_norm": 0.5404109358787537, + "learning_rate": 0.0001999972324671229, + "loss": 0.3195, + "step": 111 + }, + { + "epoch": 0.0026254563781594846, + "grad_norm": 0.8638169765472412, + "learning_rate": 0.0001999971773934506, + "loss": 0.3335, + "step": 112 + }, + { + "epoch": 0.00264889795296448, + "grad_norm": 0.47846153378486633, + "learning_rate": 0.0001999971217771908, + "loss": 0.2099, + "step": 113 + }, + { + "epoch": 0.0026723395277694754, + "grad_norm": 0.573208749294281, + "learning_rate": 0.00019999706561834384, + "loss": 0.2671, + "step": 114 + }, + { + "epoch": 0.002695781102574471, + "grad_norm": 0.7311113476753235, + "learning_rate": 0.00019999700891690995, + "loss": 0.3654, + "step": 115 + }, + { + "epoch": 0.002719222677379466, + "grad_norm": 0.5757972002029419, + "learning_rate": 0.0001999969516728895, + "loss": 0.2141, + "step": 116 + }, + { + "epoch": 0.002742664252184462, + "grad_norm": 0.20148050785064697, + "learning_rate": 0.00019999689388628276, + "loss": 0.0837, + "step": 117 + }, + { + "epoch": 0.002766105826989457, + "grad_norm": 0.5406265258789062, + "learning_rate": 0.00019999683555709007, + "loss": 0.2404, + "step": 118 + }, + { + "epoch": 0.0027895474017944526, + "grad_norm": 0.9374266862869263, + "learning_rate": 0.00019999677668531174, + "loss": 0.3164, + "step": 119 + }, + { + "epoch": 0.002812988976599448, + "grad_norm": 0.933398425579071, + "learning_rate": 0.0001999967172709481, + "loss": 0.4501, + "step": 120 + }, + { + "epoch": 0.0028364305514044434, + "grad_norm": 0.6541131138801575, + "learning_rate": 0.00019999665731399943, + "loss": 0.335, + "step": 121 + }, + { + "epoch": 0.0028598721262094386, + "grad_norm": 0.5073215961456299, + "learning_rate": 0.00019999659681446612, + "loss": 0.5513, + "step": 122 + }, + { + "epoch": 0.002883313701014434, + "grad_norm": 0.3098054826259613, + "learning_rate": 0.00019999653577234847, + "loss": 0.1315, + "step": 123 + }, + { + "epoch": 0.0029067552758194294, + "grad_norm": 0.5391197204589844, + "learning_rate": 0.00019999647418764678, + "loss": 0.2286, + "step": 124 + }, + { + "epoch": 0.002930196850624425, + "grad_norm": 0.5601248741149902, + "learning_rate": 0.00019999641206036142, + "loss": 0.3223, + "step": 125 + }, + { + "epoch": 0.00295363842542942, + "grad_norm": 0.2623872756958008, + "learning_rate": 0.0001999963493904927, + "loss": 0.0951, + "step": 126 + }, + { + "epoch": 0.002977080000234416, + "grad_norm": 0.6245015263557434, + "learning_rate": 0.00019999628617804098, + "loss": 0.7631, + "step": 127 + }, + { + "epoch": 0.003000521575039411, + "grad_norm": 0.15687838196754456, + "learning_rate": 0.0001999962224230066, + "loss": 0.0567, + "step": 128 + }, + { + "epoch": 0.0030239631498444066, + "grad_norm": 0.598662257194519, + "learning_rate": 0.00019999615812538993, + "loss": 0.3821, + "step": 129 + }, + { + "epoch": 0.0030474047246494018, + "grad_norm": 0.3797200918197632, + "learning_rate": 0.00019999609328519126, + "loss": 0.16, + "step": 130 + }, + { + "epoch": 0.0030708462994543974, + "grad_norm": 0.2790805399417877, + "learning_rate": 0.000199996027902411, + "loss": 0.1671, + "step": 131 + }, + { + "epoch": 0.0030942878742593926, + "grad_norm": 0.4512339234352112, + "learning_rate": 0.00019999596197704946, + "loss": 0.6836, + "step": 132 + }, + { + "epoch": 0.003117729449064388, + "grad_norm": 0.7587285041809082, + "learning_rate": 0.00019999589550910704, + "loss": 0.3187, + "step": 133 + }, + { + "epoch": 0.0031411710238693834, + "grad_norm": 0.7854050397872925, + "learning_rate": 0.00019999582849858407, + "loss": 0.303, + "step": 134 + }, + { + "epoch": 0.003164612598674379, + "grad_norm": 0.8098669052124023, + "learning_rate": 0.00019999576094548093, + "loss": 0.269, + "step": 135 + }, + { + "epoch": 0.003188054173479374, + "grad_norm": 0.535693347454071, + "learning_rate": 0.00019999569284979795, + "loss": 0.251, + "step": 136 + }, + { + "epoch": 0.0032114957482843698, + "grad_norm": 0.6405147314071655, + "learning_rate": 0.00019999562421153558, + "loss": 0.1833, + "step": 137 + }, + { + "epoch": 0.003234937323089365, + "grad_norm": 0.7770189642906189, + "learning_rate": 0.0001999955550306941, + "loss": 0.2643, + "step": 138 + }, + { + "epoch": 0.0032583788978943606, + "grad_norm": 0.8537822961807251, + "learning_rate": 0.00019999548530727395, + "loss": 0.2697, + "step": 139 + }, + { + "epoch": 0.0032818204726993557, + "grad_norm": 0.6494275331497192, + "learning_rate": 0.00019999541504127545, + "loss": 0.2603, + "step": 140 + }, + { + "epoch": 0.0033052620475043514, + "grad_norm": 0.6694074869155884, + "learning_rate": 0.00019999534423269908, + "loss": 0.2221, + "step": 141 + }, + { + "epoch": 0.0033287036223093465, + "grad_norm": 0.8610504865646362, + "learning_rate": 0.00019999527288154512, + "loss": 0.3352, + "step": 142 + }, + { + "epoch": 0.003352145197114342, + "grad_norm": 0.5883650183677673, + "learning_rate": 0.000199995200987814, + "loss": 0.316, + "step": 143 + }, + { + "epoch": 0.0033755867719193373, + "grad_norm": 0.6600243449211121, + "learning_rate": 0.0001999951285515061, + "loss": 0.9615, + "step": 144 + }, + { + "epoch": 0.003399028346724333, + "grad_norm": 0.748359739780426, + "learning_rate": 0.00019999505557262182, + "loss": 0.28, + "step": 145 + }, + { + "epoch": 0.003422469921529328, + "grad_norm": 0.5711257457733154, + "learning_rate": 0.00019999498205116155, + "loss": 0.3411, + "step": 146 + }, + { + "epoch": 0.0034459114963343238, + "grad_norm": 0.7077252864837646, + "learning_rate": 0.00019999490798712572, + "loss": 0.2662, + "step": 147 + }, + { + "epoch": 0.003469353071139319, + "grad_norm": 1.0276670455932617, + "learning_rate": 0.0001999948333805147, + "loss": 0.3337, + "step": 148 + }, + { + "epoch": 0.0034927946459443146, + "grad_norm": 0.24112822115421295, + "learning_rate": 0.0001999947582313289, + "loss": 0.1071, + "step": 149 + }, + { + "epoch": 0.0035162362207493097, + "grad_norm": 0.44446349143981934, + "learning_rate": 0.00019999468253956875, + "loss": 0.1026, + "step": 150 + }, + { + "epoch": 0.0035396777955543053, + "grad_norm": 0.7473758459091187, + "learning_rate": 0.0001999946063052346, + "loss": 0.5218, + "step": 151 + }, + { + "epoch": 0.0035631193703593005, + "grad_norm": 0.41437751054763794, + "learning_rate": 0.00019999452952832695, + "loss": 0.4768, + "step": 152 + }, + { + "epoch": 0.003586560945164296, + "grad_norm": 0.2056046426296234, + "learning_rate": 0.00019999445220884616, + "loss": 0.0881, + "step": 153 + }, + { + "epoch": 0.0036100025199692913, + "grad_norm": 0.7336739897727966, + "learning_rate": 0.00019999437434679266, + "loss": 0.2438, + "step": 154 + }, + { + "epoch": 0.003633444094774287, + "grad_norm": 0.5342663526535034, + "learning_rate": 0.00019999429594216682, + "loss": 0.226, + "step": 155 + }, + { + "epoch": 0.003656885669579282, + "grad_norm": 0.2539614737033844, + "learning_rate": 0.0001999942169949692, + "loss": 0.1021, + "step": 156 + }, + { + "epoch": 0.0036803272443842777, + "grad_norm": 0.9793872237205505, + "learning_rate": 0.0001999941375052001, + "loss": 0.3694, + "step": 157 + }, + { + "epoch": 0.0037037688191892734, + "grad_norm": 0.7477020025253296, + "learning_rate": 0.00019999405747286, + "loss": 0.3563, + "step": 158 + }, + { + "epoch": 0.0037272103939942685, + "grad_norm": 0.24919746816158295, + "learning_rate": 0.00019999397689794934, + "loss": 0.1659, + "step": 159 + }, + { + "epoch": 0.003750651968799264, + "grad_norm": 0.4024282991886139, + "learning_rate": 0.00019999389578046855, + "loss": 0.1535, + "step": 160 + }, + { + "epoch": 0.0037740935436042593, + "grad_norm": 0.7658990025520325, + "learning_rate": 0.00019999381412041806, + "loss": 0.3437, + "step": 161 + }, + { + "epoch": 0.003797535118409255, + "grad_norm": 0.7377162575721741, + "learning_rate": 0.00019999373191779832, + "loss": 0.3145, + "step": 162 + }, + { + "epoch": 0.00382097669321425, + "grad_norm": 0.7219094038009644, + "learning_rate": 0.0001999936491726098, + "loss": 0.2993, + "step": 163 + }, + { + "epoch": 0.0038444182680192457, + "grad_norm": 0.30201277136802673, + "learning_rate": 0.00019999356588485293, + "loss": 0.1176, + "step": 164 + }, + { + "epoch": 0.003867859842824241, + "grad_norm": 0.5947176218032837, + "learning_rate": 0.00019999348205452814, + "loss": 0.2796, + "step": 165 + }, + { + "epoch": 0.0038913014176292365, + "grad_norm": 0.6437216997146606, + "learning_rate": 0.00019999339768163592, + "loss": 0.2012, + "step": 166 + }, + { + "epoch": 0.003914742992434232, + "grad_norm": 0.5345343947410583, + "learning_rate": 0.0001999933127661767, + "loss": 0.1586, + "step": 167 + }, + { + "epoch": 0.003938184567239227, + "grad_norm": 0.5847349166870117, + "learning_rate": 0.000199993227308151, + "loss": 0.2412, + "step": 168 + }, + { + "epoch": 0.003961626142044223, + "grad_norm": 0.712988018989563, + "learning_rate": 0.00019999314130755917, + "loss": 0.2715, + "step": 169 + }, + { + "epoch": 0.003985067716849218, + "grad_norm": 0.6388394832611084, + "learning_rate": 0.0001999930547644018, + "loss": 0.4987, + "step": 170 + }, + { + "epoch": 0.004008509291654213, + "grad_norm": 0.6887701153755188, + "learning_rate": 0.00019999296767867923, + "loss": 0.2207, + "step": 171 + }, + { + "epoch": 0.0040319508664592085, + "grad_norm": 0.6678301692008972, + "learning_rate": 0.00019999288005039206, + "loss": 0.254, + "step": 172 + }, + { + "epoch": 0.0040553924412642045, + "grad_norm": 0.6648051738739014, + "learning_rate": 0.0001999927918795407, + "loss": 0.3993, + "step": 173 + }, + { + "epoch": 0.0040788340160692, + "grad_norm": 0.6412478089332581, + "learning_rate": 0.00019999270316612563, + "loss": 0.3067, + "step": 174 + }, + { + "epoch": 0.004102275590874195, + "grad_norm": 0.2730605900287628, + "learning_rate": 0.00019999261391014737, + "loss": 0.0567, + "step": 175 + }, + { + "epoch": 0.00412571716567919, + "grad_norm": 1.0569084882736206, + "learning_rate": 0.00019999252411160635, + "loss": 0.4632, + "step": 176 + }, + { + "epoch": 0.004149158740484186, + "grad_norm": 0.8004516959190369, + "learning_rate": 0.00019999243377050307, + "loss": 0.448, + "step": 177 + }, + { + "epoch": 0.004172600315289181, + "grad_norm": 0.6045053005218506, + "learning_rate": 0.00019999234288683808, + "loss": 0.4932, + "step": 178 + }, + { + "epoch": 0.0041960418900941765, + "grad_norm": 0.2875787019729614, + "learning_rate": 0.0001999922514606118, + "loss": 0.1086, + "step": 179 + }, + { + "epoch": 0.004219483464899172, + "grad_norm": 0.5860103368759155, + "learning_rate": 0.00019999215949182471, + "loss": 0.2688, + "step": 180 + }, + { + "epoch": 0.004242925039704168, + "grad_norm": 0.27062076330184937, + "learning_rate": 0.00019999206698047744, + "loss": 0.1344, + "step": 181 + }, + { + "epoch": 0.004266366614509163, + "grad_norm": 0.7103718519210815, + "learning_rate": 0.0001999919739265703, + "loss": 0.226, + "step": 182 + }, + { + "epoch": 0.004289808189314158, + "grad_norm": 0.20319703221321106, + "learning_rate": 0.00019999188033010398, + "loss": 0.1087, + "step": 183 + }, + { + "epoch": 0.004313249764119153, + "grad_norm": 0.6077110171318054, + "learning_rate": 0.0001999917861910789, + "loss": 0.3064, + "step": 184 + }, + { + "epoch": 0.004336691338924149, + "grad_norm": 0.5340200066566467, + "learning_rate": 0.00019999169150949553, + "loss": 0.6988, + "step": 185 + }, + { + "epoch": 0.0043601329137291445, + "grad_norm": 0.5805467367172241, + "learning_rate": 0.00019999159628535446, + "loss": 0.2784, + "step": 186 + }, + { + "epoch": 0.00438357448853414, + "grad_norm": 0.5847718119621277, + "learning_rate": 0.00019999150051865617, + "loss": 0.4073, + "step": 187 + }, + { + "epoch": 0.004407016063339135, + "grad_norm": 0.2271495908498764, + "learning_rate": 0.0001999914042094012, + "loss": 0.1461, + "step": 188 + }, + { + "epoch": 0.004430457638144131, + "grad_norm": 0.6872106194496155, + "learning_rate": 0.00019999130735759004, + "loss": 0.4331, + "step": 189 + }, + { + "epoch": 0.004453899212949126, + "grad_norm": 0.5474057197570801, + "learning_rate": 0.00019999120996322324, + "loss": 0.7478, + "step": 190 + }, + { + "epoch": 0.004477340787754121, + "grad_norm": 0.45351818203926086, + "learning_rate": 0.00019999111202630135, + "loss": 0.1787, + "step": 191 + }, + { + "epoch": 0.0045007823625591165, + "grad_norm": 0.6970012784004211, + "learning_rate": 0.00019999101354682483, + "loss": 0.2912, + "step": 192 + }, + { + "epoch": 0.0045242239373641125, + "grad_norm": 0.6965880393981934, + "learning_rate": 0.0001999909145247943, + "loss": 0.2206, + "step": 193 + }, + { + "epoch": 0.004547665512169108, + "grad_norm": 0.5245277881622314, + "learning_rate": 0.00019999081496021022, + "loss": 0.2663, + "step": 194 + }, + { + "epoch": 0.004571107086974103, + "grad_norm": 0.7173568606376648, + "learning_rate": 0.00019999071485307318, + "loss": 0.228, + "step": 195 + }, + { + "epoch": 0.004594548661779098, + "grad_norm": 0.5895550847053528, + "learning_rate": 0.00019999061420338372, + "loss": 0.5861, + "step": 196 + }, + { + "epoch": 0.004617990236584094, + "grad_norm": 0.5454007983207703, + "learning_rate": 0.00019999051301114234, + "loss": 0.2933, + "step": 197 + }, + { + "epoch": 0.004641431811389089, + "grad_norm": 0.42998167872428894, + "learning_rate": 0.0001999904112763497, + "loss": 0.127, + "step": 198 + }, + { + "epoch": 0.0046648733861940845, + "grad_norm": 0.6337180733680725, + "learning_rate": 0.00019999030899900622, + "loss": 0.2621, + "step": 199 + }, + { + "epoch": 0.00468831496099908, + "grad_norm": 0.4967643618583679, + "learning_rate": 0.00019999020617911253, + "loss": 0.2131, + "step": 200 + }, + { + "epoch": 0.004711756535804076, + "grad_norm": 0.33472809195518494, + "learning_rate": 0.00019999010281666915, + "loss": 0.0961, + "step": 201 + }, + { + "epoch": 0.004735198110609071, + "grad_norm": 0.22658176720142365, + "learning_rate": 0.00019998999891167668, + "loss": 0.0679, + "step": 202 + }, + { + "epoch": 0.004758639685414066, + "grad_norm": 0.571151614189148, + "learning_rate": 0.00019998989446413562, + "loss": 0.2945, + "step": 203 + }, + { + "epoch": 0.004782081260219061, + "grad_norm": 1.0746181011199951, + "learning_rate": 0.00019998978947404666, + "loss": 0.4049, + "step": 204 + }, + { + "epoch": 0.004805522835024057, + "grad_norm": 0.5280450582504272, + "learning_rate": 0.00019998968394141023, + "loss": 0.6824, + "step": 205 + }, + { + "epoch": 0.0048289644098290525, + "grad_norm": 0.6076974272727966, + "learning_rate": 0.00019998957786622698, + "loss": 0.2413, + "step": 206 + }, + { + "epoch": 0.004852405984634048, + "grad_norm": 0.17039133608341217, + "learning_rate": 0.00019998947124849747, + "loss": 0.0667, + "step": 207 + }, + { + "epoch": 0.004875847559439043, + "grad_norm": 0.3580293357372284, + "learning_rate": 0.0001999893640882223, + "loss": 0.1136, + "step": 208 + }, + { + "epoch": 0.004899289134244039, + "grad_norm": 0.7673648595809937, + "learning_rate": 0.000199989256385402, + "loss": 0.2669, + "step": 209 + }, + { + "epoch": 0.004922730709049034, + "grad_norm": 0.5456467270851135, + "learning_rate": 0.0001999891481400372, + "loss": 0.1825, + "step": 210 + }, + { + "epoch": 0.004946172283854029, + "grad_norm": 0.22815725207328796, + "learning_rate": 0.00019998903935212846, + "loss": 0.0421, + "step": 211 + }, + { + "epoch": 0.004969613858659024, + "grad_norm": 0.15425407886505127, + "learning_rate": 0.00019998893002167635, + "loss": 0.0383, + "step": 212 + }, + { + "epoch": 0.0049930554334640205, + "grad_norm": 0.617191731929779, + "learning_rate": 0.00019998882014868156, + "loss": 0.2605, + "step": 213 + }, + { + "epoch": 0.005016497008269016, + "grad_norm": 0.8083759546279907, + "learning_rate": 0.00019998870973314458, + "loss": 0.2757, + "step": 214 + }, + { + "epoch": 0.005039938583074011, + "grad_norm": 0.4168655574321747, + "learning_rate": 0.00019998859877506605, + "loss": 0.1314, + "step": 215 + }, + { + "epoch": 0.005063380157879006, + "grad_norm": 0.6286880373954773, + "learning_rate": 0.00019998848727444658, + "loss": 0.1704, + "step": 216 + }, + { + "epoch": 0.005086821732684002, + "grad_norm": 0.6133062243461609, + "learning_rate": 0.00019998837523128678, + "loss": 0.1897, + "step": 217 + }, + { + "epoch": 0.005110263307488997, + "grad_norm": 0.5801655650138855, + "learning_rate": 0.0001999882626455872, + "loss": 0.1935, + "step": 218 + }, + { + "epoch": 0.005133704882293992, + "grad_norm": 0.6554915904998779, + "learning_rate": 0.00019998814951734855, + "loss": 0.2565, + "step": 219 + }, + { + "epoch": 0.005157146457098988, + "grad_norm": 0.2743030786514282, + "learning_rate": 0.00019998803584657138, + "loss": 0.1194, + "step": 220 + }, + { + "epoch": 0.005180588031903984, + "grad_norm": 0.683155357837677, + "learning_rate": 0.00019998792163325628, + "loss": 0.2381, + "step": 221 + }, + { + "epoch": 0.005204029606708979, + "grad_norm": 0.22134950757026672, + "learning_rate": 0.00019998780687740392, + "loss": 0.0875, + "step": 222 + }, + { + "epoch": 0.005227471181513974, + "grad_norm": 0.9469959139823914, + "learning_rate": 0.00019998769157901492, + "loss": 0.4573, + "step": 223 + }, + { + "epoch": 0.005250912756318969, + "grad_norm": 0.6153260469436646, + "learning_rate": 0.00019998757573808987, + "loss": 0.2991, + "step": 224 + }, + { + "epoch": 0.005274354331123965, + "grad_norm": 0.6141527891159058, + "learning_rate": 0.00019998745935462947, + "loss": 0.2398, + "step": 225 + }, + { + "epoch": 0.00529779590592896, + "grad_norm": 0.42826026678085327, + "learning_rate": 0.00019998734242863427, + "loss": 0.1343, + "step": 226 + }, + { + "epoch": 0.005321237480733956, + "grad_norm": 0.6432958841323853, + "learning_rate": 0.00019998722496010496, + "loss": 0.275, + "step": 227 + }, + { + "epoch": 0.005344679055538951, + "grad_norm": 0.11820521950721741, + "learning_rate": 0.00019998710694904212, + "loss": 0.0415, + "step": 228 + }, + { + "epoch": 0.005368120630343947, + "grad_norm": 0.2677289545536041, + "learning_rate": 0.00019998698839544647, + "loss": 0.0791, + "step": 229 + }, + { + "epoch": 0.005391562205148942, + "grad_norm": 1.0672919750213623, + "learning_rate": 0.00019998686929931857, + "loss": 0.5419, + "step": 230 + }, + { + "epoch": 0.005415003779953937, + "grad_norm": 0.668829619884491, + "learning_rate": 0.0001999867496606591, + "loss": 0.2677, + "step": 231 + }, + { + "epoch": 0.005438445354758932, + "grad_norm": 0.5264405608177185, + "learning_rate": 0.00019998662947946877, + "loss": 0.1967, + "step": 232 + }, + { + "epoch": 0.005461886929563928, + "grad_norm": 0.2036389857530594, + "learning_rate": 0.00019998650875574815, + "loss": 0.0925, + "step": 233 + }, + { + "epoch": 0.005485328504368924, + "grad_norm": 0.623810887336731, + "learning_rate": 0.0001999863874894979, + "loss": 0.2569, + "step": 234 + }, + { + "epoch": 0.005508770079173919, + "grad_norm": 0.29454079270362854, + "learning_rate": 0.00019998626568071872, + "loss": 0.0692, + "step": 235 + }, + { + "epoch": 0.005532211653978914, + "grad_norm": 0.16561725735664368, + "learning_rate": 0.00019998614332941123, + "loss": 0.0547, + "step": 236 + }, + { + "epoch": 0.00555565322878391, + "grad_norm": 0.19288848340511322, + "learning_rate": 0.00019998602043557614, + "loss": 0.0857, + "step": 237 + }, + { + "epoch": 0.005579094803588905, + "grad_norm": 0.8369110226631165, + "learning_rate": 0.0001999858969992141, + "loss": 0.3406, + "step": 238 + }, + { + "epoch": 0.0056025363783939, + "grad_norm": 0.8375520706176758, + "learning_rate": 0.00019998577302032575, + "loss": 0.3766, + "step": 239 + }, + { + "epoch": 0.005625977953198896, + "grad_norm": 0.6832995414733887, + "learning_rate": 0.0001999856484989118, + "loss": 0.3119, + "step": 240 + }, + { + "epoch": 0.005649419528003892, + "grad_norm": 0.5383294820785522, + "learning_rate": 0.0001999855234349729, + "loss": 0.25, + "step": 241 + }, + { + "epoch": 0.005672861102808887, + "grad_norm": 0.8301267623901367, + "learning_rate": 0.00019998539782850974, + "loss": 0.4474, + "step": 242 + }, + { + "epoch": 0.005696302677613882, + "grad_norm": 0.4764211177825928, + "learning_rate": 0.000199985271679523, + "loss": 0.2414, + "step": 243 + }, + { + "epoch": 0.005719744252418877, + "grad_norm": 0.6162632703781128, + "learning_rate": 0.00019998514498801334, + "loss": 0.7303, + "step": 244 + }, + { + "epoch": 0.005743185827223873, + "grad_norm": 0.4925750195980072, + "learning_rate": 0.00019998501775398152, + "loss": 0.171, + "step": 245 + }, + { + "epoch": 0.005766627402028868, + "grad_norm": 0.4426475763320923, + "learning_rate": 0.00019998488997742815, + "loss": 0.2037, + "step": 246 + }, + { + "epoch": 0.005790068976833864, + "grad_norm": 0.6759050488471985, + "learning_rate": 0.00019998476165835396, + "loss": 0.2985, + "step": 247 + }, + { + "epoch": 0.005813510551638859, + "grad_norm": 0.7058880925178528, + "learning_rate": 0.00019998463279675966, + "loss": 0.3654, + "step": 248 + }, + { + "epoch": 0.005836952126443855, + "grad_norm": 0.21047967672348022, + "learning_rate": 0.00019998450339264593, + "loss": 0.092, + "step": 249 + }, + { + "epoch": 0.00586039370124885, + "grad_norm": 0.2309301793575287, + "learning_rate": 0.00019998437344601343, + "loss": 0.0699, + "step": 250 + }, + { + "epoch": 0.005883835276053845, + "grad_norm": 0.780332088470459, + "learning_rate": 0.00019998424295686294, + "loss": 0.2965, + "step": 251 + }, + { + "epoch": 0.00590727685085884, + "grad_norm": 0.7872247695922852, + "learning_rate": 0.00019998411192519516, + "loss": 0.3912, + "step": 252 + }, + { + "epoch": 0.005930718425663836, + "grad_norm": 0.5965334177017212, + "learning_rate": 0.00019998398035101072, + "loss": 0.22, + "step": 253 + }, + { + "epoch": 0.005954160000468832, + "grad_norm": 0.4982486665248871, + "learning_rate": 0.00019998384823431041, + "loss": 0.2081, + "step": 254 + }, + { + "epoch": 0.005977601575273827, + "grad_norm": 1.027955174446106, + "learning_rate": 0.00019998371557509496, + "loss": 0.2498, + "step": 255 + }, + { + "epoch": 0.006001043150078822, + "grad_norm": 0.3990105390548706, + "learning_rate": 0.00019998358237336504, + "loss": 0.1723, + "step": 256 + }, + { + "epoch": 0.006024484724883818, + "grad_norm": 0.6770126223564148, + "learning_rate": 0.0001999834486291214, + "loss": 0.2878, + "step": 257 + }, + { + "epoch": 0.006047926299688813, + "grad_norm": 0.5540021061897278, + "learning_rate": 0.00019998331434236476, + "loss": 0.2709, + "step": 258 + }, + { + "epoch": 0.006071367874493808, + "grad_norm": 0.7046794891357422, + "learning_rate": 0.0001999831795130958, + "loss": 0.7881, + "step": 259 + }, + { + "epoch": 0.0060948094492988035, + "grad_norm": 0.8279392719268799, + "learning_rate": 0.00019998304414131534, + "loss": 0.2677, + "step": 260 + }, + { + "epoch": 0.0061182510241038, + "grad_norm": 0.5849525928497314, + "learning_rate": 0.00019998290822702405, + "loss": 0.1714, + "step": 261 + }, + { + "epoch": 0.006141692598908795, + "grad_norm": 0.6247529983520508, + "learning_rate": 0.00019998277177022268, + "loss": 0.3779, + "step": 262 + }, + { + "epoch": 0.00616513417371379, + "grad_norm": 0.20718392729759216, + "learning_rate": 0.00019998263477091205, + "loss": 0.0684, + "step": 263 + }, + { + "epoch": 0.006188575748518785, + "grad_norm": 0.5874770879745483, + "learning_rate": 0.00019998249722909275, + "loss": 0.1717, + "step": 264 + }, + { + "epoch": 0.006212017323323781, + "grad_norm": 0.6086574792861938, + "learning_rate": 0.00019998235914476563, + "loss": 0.3244, + "step": 265 + }, + { + "epoch": 0.006235458898128776, + "grad_norm": 0.8970377445220947, + "learning_rate": 0.00019998222051793144, + "loss": 0.2694, + "step": 266 + }, + { + "epoch": 0.0062589004729337715, + "grad_norm": 0.6833506226539612, + "learning_rate": 0.0001999820813485909, + "loss": 0.2508, + "step": 267 + }, + { + "epoch": 0.006282342047738767, + "grad_norm": 0.16413381695747375, + "learning_rate": 0.00019998194163674477, + "loss": 0.0432, + "step": 268 + }, + { + "epoch": 0.006305783622543763, + "grad_norm": 1.3942104578018188, + "learning_rate": 0.0001999818013823938, + "loss": 0.3097, + "step": 269 + }, + { + "epoch": 0.006329225197348758, + "grad_norm": 0.27007514238357544, + "learning_rate": 0.00019998166058553878, + "loss": 0.1414, + "step": 270 + }, + { + "epoch": 0.006352666772153753, + "grad_norm": 0.6357520818710327, + "learning_rate": 0.00019998151924618046, + "loss": 0.2202, + "step": 271 + }, + { + "epoch": 0.006376108346958748, + "grad_norm": 0.5415841341018677, + "learning_rate": 0.00019998137736431962, + "loss": 0.7762, + "step": 272 + }, + { + "epoch": 0.006399549921763744, + "grad_norm": 0.6225699782371521, + "learning_rate": 0.00019998123493995698, + "loss": 0.2769, + "step": 273 + }, + { + "epoch": 0.0064229914965687395, + "grad_norm": 0.5930852890014648, + "learning_rate": 0.00019998109197309338, + "loss": 0.2888, + "step": 274 + }, + { + "epoch": 0.006446433071373735, + "grad_norm": 0.6408600807189941, + "learning_rate": 0.00019998094846372953, + "loss": 0.2813, + "step": 275 + }, + { + "epoch": 0.00646987464617873, + "grad_norm": 0.27594876289367676, + "learning_rate": 0.0001999808044118663, + "loss": 0.1223, + "step": 276 + }, + { + "epoch": 0.006493316220983726, + "grad_norm": 0.48001447319984436, + "learning_rate": 0.00019998065981750436, + "loss": 0.7353, + "step": 277 + }, + { + "epoch": 0.006516757795788721, + "grad_norm": 0.30030956864356995, + "learning_rate": 0.00019998051468064456, + "loss": 0.1615, + "step": 278 + }, + { + "epoch": 0.006540199370593716, + "grad_norm": 0.19775788486003876, + "learning_rate": 0.0001999803690012877, + "loss": 0.1089, + "step": 279 + }, + { + "epoch": 0.0065636409453987115, + "grad_norm": 0.6999666690826416, + "learning_rate": 0.00019998022277943451, + "loss": 0.3336, + "step": 280 + }, + { + "epoch": 0.0065870825202037075, + "grad_norm": 0.34480974078178406, + "learning_rate": 0.00019998007601508585, + "loss": 0.142, + "step": 281 + }, + { + "epoch": 0.006610524095008703, + "grad_norm": 0.41515403985977173, + "learning_rate": 0.00019997992870824247, + "loss": 0.1908, + "step": 282 + }, + { + "epoch": 0.006633965669813698, + "grad_norm": 0.22067850828170776, + "learning_rate": 0.00019997978085890518, + "loss": 0.0804, + "step": 283 + }, + { + "epoch": 0.006657407244618693, + "grad_norm": 1.0457795858383179, + "learning_rate": 0.0001999796324670748, + "loss": 0.3922, + "step": 284 + }, + { + "epoch": 0.006680848819423689, + "grad_norm": 0.5554715394973755, + "learning_rate": 0.00019997948353275215, + "loss": 0.1965, + "step": 285 + }, + { + "epoch": 0.006704290394228684, + "grad_norm": 0.7129070162773132, + "learning_rate": 0.00019997933405593797, + "loss": 0.3064, + "step": 286 + }, + { + "epoch": 0.0067277319690336795, + "grad_norm": 0.6452338099479675, + "learning_rate": 0.0001999791840366331, + "loss": 0.29, + "step": 287 + }, + { + "epoch": 0.006751173543838675, + "grad_norm": 0.23260188102722168, + "learning_rate": 0.0001999790334748384, + "loss": 0.0672, + "step": 288 + }, + { + "epoch": 0.006774615118643671, + "grad_norm": 0.2568034827709198, + "learning_rate": 0.00019997888237055468, + "loss": 0.1155, + "step": 289 + }, + { + "epoch": 0.006798056693448666, + "grad_norm": 0.40309175848960876, + "learning_rate": 0.00019997873072378269, + "loss": 0.166, + "step": 290 + }, + { + "epoch": 0.006821498268253661, + "grad_norm": 0.3536180853843689, + "learning_rate": 0.00019997857853452328, + "loss": 0.1175, + "step": 291 + }, + { + "epoch": 0.006844939843058656, + "grad_norm": 0.9091312289237976, + "learning_rate": 0.0001999784258027773, + "loss": 0.4353, + "step": 292 + }, + { + "epoch": 0.006868381417863652, + "grad_norm": 0.6268461346626282, + "learning_rate": 0.0001999782725285456, + "loss": 0.8288, + "step": 293 + }, + { + "epoch": 0.0068918229926686475, + "grad_norm": 0.9979032874107361, + "learning_rate": 0.00019997811871182898, + "loss": 0.5519, + "step": 294 + }, + { + "epoch": 0.006915264567473643, + "grad_norm": 0.7042083740234375, + "learning_rate": 0.00019997796435262824, + "loss": 0.1781, + "step": 295 + }, + { + "epoch": 0.006938706142278638, + "grad_norm": 0.47846630215644836, + "learning_rate": 0.00019997780945094431, + "loss": 0.6003, + "step": 296 + }, + { + "epoch": 0.006962147717083634, + "grad_norm": 0.3156568109989166, + "learning_rate": 0.00019997765400677793, + "loss": 0.0638, + "step": 297 + }, + { + "epoch": 0.006985589291888629, + "grad_norm": 0.7269763946533203, + "learning_rate": 0.00019997749802013002, + "loss": 0.2889, + "step": 298 + }, + { + "epoch": 0.007009030866693624, + "grad_norm": 0.36343690752983093, + "learning_rate": 0.00019997734149100136, + "loss": 0.0876, + "step": 299 + }, + { + "epoch": 0.0070324724414986195, + "grad_norm": 0.6919704079627991, + "learning_rate": 0.00019997718441939287, + "loss": 0.1912, + "step": 300 + }, + { + "epoch": 0.0070559140163036155, + "grad_norm": 0.5871474146842957, + "learning_rate": 0.00019997702680530532, + "loss": 0.5779, + "step": 301 + }, + { + "epoch": 0.007079355591108611, + "grad_norm": 0.15146811306476593, + "learning_rate": 0.00019997686864873965, + "loss": 0.0664, + "step": 302 + }, + { + "epoch": 0.007102797165913606, + "grad_norm": 0.5561721324920654, + "learning_rate": 0.00019997670994969667, + "loss": 0.1914, + "step": 303 + }, + { + "epoch": 0.007126238740718601, + "grad_norm": 0.19945819675922394, + "learning_rate": 0.00019997655070817723, + "loss": 0.0828, + "step": 304 + }, + { + "epoch": 0.007149680315523597, + "grad_norm": 0.6353179216384888, + "learning_rate": 0.00019997639092418225, + "loss": 0.2524, + "step": 305 + }, + { + "epoch": 0.007173121890328592, + "grad_norm": 0.6715456247329712, + "learning_rate": 0.00019997623059771254, + "loss": 0.3402, + "step": 306 + }, + { + "epoch": 0.0071965634651335875, + "grad_norm": 0.4307434856891632, + "learning_rate": 0.000199976069728769, + "loss": 0.1869, + "step": 307 + }, + { + "epoch": 0.007220005039938583, + "grad_norm": 0.608390212059021, + "learning_rate": 0.00019997590831735248, + "loss": 0.3036, + "step": 308 + }, + { + "epoch": 0.007243446614743579, + "grad_norm": 0.4761715829372406, + "learning_rate": 0.00019997574636346387, + "loss": 0.2263, + "step": 309 + }, + { + "epoch": 0.007266888189548574, + "grad_norm": 0.6349612474441528, + "learning_rate": 0.00019997558386710407, + "loss": 0.2343, + "step": 310 + }, + { + "epoch": 0.007290329764353569, + "grad_norm": 0.676900327205658, + "learning_rate": 0.00019997542082827394, + "loss": 0.2843, + "step": 311 + }, + { + "epoch": 0.007313771339158564, + "grad_norm": 0.5269830226898193, + "learning_rate": 0.00019997525724697437, + "loss": 0.209, + "step": 312 + }, + { + "epoch": 0.00733721291396356, + "grad_norm": 0.24434977769851685, + "learning_rate": 0.00019997509312320622, + "loss": 0.0655, + "step": 313 + }, + { + "epoch": 0.0073606544887685555, + "grad_norm": 0.6879680752754211, + "learning_rate": 0.0001999749284569704, + "loss": 0.8974, + "step": 314 + }, + { + "epoch": 0.007384096063573551, + "grad_norm": 0.21603448688983917, + "learning_rate": 0.00019997476324826782, + "loss": 0.0715, + "step": 315 + }, + { + "epoch": 0.007407537638378547, + "grad_norm": 0.8295873403549194, + "learning_rate": 0.00019997459749709936, + "loss": 0.1262, + "step": 316 + }, + { + "epoch": 0.007430979213183542, + "grad_norm": 0.9513508677482605, + "learning_rate": 0.00019997443120346593, + "loss": 0.3499, + "step": 317 + }, + { + "epoch": 0.007454420787988537, + "grad_norm": 0.808876633644104, + "learning_rate": 0.00019997426436736844, + "loss": 0.2086, + "step": 318 + }, + { + "epoch": 0.007477862362793532, + "grad_norm": 0.6566169261932373, + "learning_rate": 0.00019997409698880775, + "loss": 0.1757, + "step": 319 + }, + { + "epoch": 0.007501303937598528, + "grad_norm": 0.7490182518959045, + "learning_rate": 0.0001999739290677848, + "loss": 0.2711, + "step": 320 + }, + { + "epoch": 0.0075247455124035235, + "grad_norm": 0.9839372634887695, + "learning_rate": 0.00019997376060430054, + "loss": 0.3519, + "step": 321 + }, + { + "epoch": 0.007548187087208519, + "grad_norm": 0.9783810973167419, + "learning_rate": 0.0001999735915983558, + "loss": 0.3177, + "step": 322 + }, + { + "epoch": 0.007571628662013514, + "grad_norm": 0.7347965240478516, + "learning_rate": 0.00019997342204995156, + "loss": 0.3189, + "step": 323 + }, + { + "epoch": 0.00759507023681851, + "grad_norm": 0.7132028937339783, + "learning_rate": 0.0001999732519590887, + "loss": 0.1736, + "step": 324 + }, + { + "epoch": 0.007618511811623505, + "grad_norm": 0.5780074000358582, + "learning_rate": 0.0001999730813257682, + "loss": 0.1577, + "step": 325 + }, + { + "epoch": 0.0076419533864285, + "grad_norm": 0.7790593504905701, + "learning_rate": 0.00019997291014999094, + "loss": 0.2546, + "step": 326 + }, + { + "epoch": 0.007665394961233495, + "grad_norm": 0.8323804140090942, + "learning_rate": 0.00019997273843175783, + "loss": 0.2378, + "step": 327 + }, + { + "epoch": 0.0076888365360384915, + "grad_norm": 0.9022363424301147, + "learning_rate": 0.00019997256617106985, + "loss": 0.3421, + "step": 328 + }, + { + "epoch": 0.007712278110843487, + "grad_norm": 0.7091406583786011, + "learning_rate": 0.00019997239336792791, + "loss": 0.2942, + "step": 329 + }, + { + "epoch": 0.007735719685648482, + "grad_norm": 0.29742515087127686, + "learning_rate": 0.00019997222002233298, + "loss": 0.1267, + "step": 330 + }, + { + "epoch": 0.007759161260453477, + "grad_norm": 0.8812986612319946, + "learning_rate": 0.00019997204613428596, + "loss": 0.278, + "step": 331 + }, + { + "epoch": 0.007782602835258473, + "grad_norm": 0.8990840911865234, + "learning_rate": 0.00019997187170378778, + "loss": 0.1753, + "step": 332 + }, + { + "epoch": 0.007806044410063468, + "grad_norm": 0.7537962794303894, + "learning_rate": 0.00019997169673083943, + "loss": 0.1863, + "step": 333 + }, + { + "epoch": 0.007829485984868463, + "grad_norm": 0.811258852481842, + "learning_rate": 0.00019997152121544186, + "loss": 0.355, + "step": 334 + }, + { + "epoch": 0.00785292755967346, + "grad_norm": 0.4777142107486725, + "learning_rate": 0.00019997134515759596, + "loss": 0.2285, + "step": 335 + }, + { + "epoch": 0.007876369134478454, + "grad_norm": 0.2254602611064911, + "learning_rate": 0.00019997116855730279, + "loss": 0.0783, + "step": 336 + }, + { + "epoch": 0.00789981070928345, + "grad_norm": 0.5047174096107483, + "learning_rate": 0.00019997099141456318, + "loss": 0.2612, + "step": 337 + }, + { + "epoch": 0.007923252284088446, + "grad_norm": 0.5901991128921509, + "learning_rate": 0.00019997081372937823, + "loss": 0.3295, + "step": 338 + }, + { + "epoch": 0.00794669385889344, + "grad_norm": 0.22092990577220917, + "learning_rate": 0.0001999706355017488, + "loss": 0.1364, + "step": 339 + }, + { + "epoch": 0.007970135433698436, + "grad_norm": 0.6751435399055481, + "learning_rate": 0.0001999704567316759, + "loss": 0.2879, + "step": 340 + }, + { + "epoch": 0.00799357700850343, + "grad_norm": 0.7283354997634888, + "learning_rate": 0.00019997027741916045, + "loss": 0.3392, + "step": 341 + }, + { + "epoch": 0.008017018583308427, + "grad_norm": 0.5604244470596313, + "learning_rate": 0.0001999700975642035, + "loss": 0.1912, + "step": 342 + }, + { + "epoch": 0.008040460158113423, + "grad_norm": 0.6692593693733215, + "learning_rate": 0.000199969917166806, + "loss": 0.2217, + "step": 343 + }, + { + "epoch": 0.008063901732918417, + "grad_norm": 0.7876620292663574, + "learning_rate": 0.00019996973622696888, + "loss": 0.2507, + "step": 344 + }, + { + "epoch": 0.008087343307723413, + "grad_norm": 0.44995588064193726, + "learning_rate": 0.0001999695547446932, + "loss": 0.1702, + "step": 345 + }, + { + "epoch": 0.008110784882528409, + "grad_norm": 0.5608309507369995, + "learning_rate": 0.0001999693727199799, + "loss": 0.2702, + "step": 346 + }, + { + "epoch": 0.008134226457333403, + "grad_norm": 0.6912648677825928, + "learning_rate": 0.00019996919015282996, + "loss": 0.2824, + "step": 347 + }, + { + "epoch": 0.0081576680321384, + "grad_norm": 0.7922164797782898, + "learning_rate": 0.00019996900704324436, + "loss": 0.3248, + "step": 348 + }, + { + "epoch": 0.008181109606943394, + "grad_norm": 0.9546986818313599, + "learning_rate": 0.00019996882339122415, + "loss": 0.3951, + "step": 349 + }, + { + "epoch": 0.00820455118174839, + "grad_norm": 0.2911871671676636, + "learning_rate": 0.0001999686391967703, + "loss": 0.0765, + "step": 350 + }, + { + "epoch": 0.008227992756553386, + "grad_norm": 0.4835963547229767, + "learning_rate": 0.00019996845445988378, + "loss": 0.2021, + "step": 351 + }, + { + "epoch": 0.00825143433135838, + "grad_norm": 0.688918948173523, + "learning_rate": 0.00019996826918056563, + "loss": 0.249, + "step": 352 + }, + { + "epoch": 0.008274875906163376, + "grad_norm": 0.2286163866519928, + "learning_rate": 0.00019996808335881686, + "loss": 0.0857, + "step": 353 + }, + { + "epoch": 0.008298317480968372, + "grad_norm": 0.7363866567611694, + "learning_rate": 0.0001999678969946384, + "loss": 0.1985, + "step": 354 + }, + { + "epoch": 0.008321759055773367, + "grad_norm": 0.6413483619689941, + "learning_rate": 0.00019996771008803138, + "loss": 0.2495, + "step": 355 + }, + { + "epoch": 0.008345200630578363, + "grad_norm": 0.14204461872577667, + "learning_rate": 0.00019996752263899673, + "loss": 0.0777, + "step": 356 + }, + { + "epoch": 0.008368642205383357, + "grad_norm": 0.5384321808815002, + "learning_rate": 0.00019996733464753546, + "loss": 0.2058, + "step": 357 + }, + { + "epoch": 0.008392083780188353, + "grad_norm": 0.6845918893814087, + "learning_rate": 0.00019996714611364866, + "loss": 0.9432, + "step": 358 + }, + { + "epoch": 0.008415525354993349, + "grad_norm": 0.6740832328796387, + "learning_rate": 0.0001999669570373373, + "loss": 0.2829, + "step": 359 + }, + { + "epoch": 0.008438966929798343, + "grad_norm": 0.6737803816795349, + "learning_rate": 0.00019996676741860242, + "loss": 0.3779, + "step": 360 + }, + { + "epoch": 0.00846240850460334, + "grad_norm": 0.5922757983207703, + "learning_rate": 0.00019996657725744504, + "loss": 0.22, + "step": 361 + }, + { + "epoch": 0.008485850079408335, + "grad_norm": 0.2421647161245346, + "learning_rate": 0.00019996638655386624, + "loss": 0.0816, + "step": 362 + }, + { + "epoch": 0.00850929165421333, + "grad_norm": 0.5775250196456909, + "learning_rate": 0.000199966195307867, + "loss": 0.2992, + "step": 363 + }, + { + "epoch": 0.008532733229018326, + "grad_norm": 0.5939027070999146, + "learning_rate": 0.00019996600351944834, + "loss": 0.231, + "step": 364 + }, + { + "epoch": 0.00855617480382332, + "grad_norm": 0.39424362778663635, + "learning_rate": 0.00019996581118861137, + "loss": 0.0777, + "step": 365 + }, + { + "epoch": 0.008579616378628316, + "grad_norm": 0.19475315511226654, + "learning_rate": 0.00019996561831535709, + "loss": 0.0615, + "step": 366 + }, + { + "epoch": 0.008603057953433312, + "grad_norm": 0.13105110824108124, + "learning_rate": 0.00019996542489968656, + "loss": 0.0346, + "step": 367 + }, + { + "epoch": 0.008626499528238307, + "grad_norm": 0.6401360034942627, + "learning_rate": 0.0001999652309416008, + "loss": 0.2085, + "step": 368 + }, + { + "epoch": 0.008649941103043303, + "grad_norm": 0.30108118057250977, + "learning_rate": 0.0001999650364411009, + "loss": 0.1187, + "step": 369 + }, + { + "epoch": 0.008673382677848299, + "grad_norm": 0.7189109325408936, + "learning_rate": 0.00019996484139818792, + "loss": 0.2998, + "step": 370 + }, + { + "epoch": 0.008696824252653293, + "grad_norm": 0.6583682894706726, + "learning_rate": 0.00019996464581286286, + "loss": 0.227, + "step": 371 + }, + { + "epoch": 0.008720265827458289, + "grad_norm": 0.5531545281410217, + "learning_rate": 0.00019996444968512688, + "loss": 0.203, + "step": 372 + }, + { + "epoch": 0.008743707402263283, + "grad_norm": 0.5096223950386047, + "learning_rate": 0.0001999642530149809, + "loss": 0.2079, + "step": 373 + }, + { + "epoch": 0.00876714897706828, + "grad_norm": 0.15910835564136505, + "learning_rate": 0.00019996405580242614, + "loss": 0.0502, + "step": 374 + }, + { + "epoch": 0.008790590551873275, + "grad_norm": 0.28779730200767517, + "learning_rate": 0.0001999638580474636, + "loss": 0.1577, + "step": 375 + }, + { + "epoch": 0.00881403212667827, + "grad_norm": 0.5278774499893188, + "learning_rate": 0.00019996365975009431, + "loss": 0.134, + "step": 376 + }, + { + "epoch": 0.008837473701483266, + "grad_norm": 0.2480696141719818, + "learning_rate": 0.00019996346091031943, + "loss": 0.1631, + "step": 377 + }, + { + "epoch": 0.008860915276288262, + "grad_norm": 0.7121061682701111, + "learning_rate": 0.00019996326152814, + "loss": 0.3263, + "step": 378 + }, + { + "epoch": 0.008884356851093256, + "grad_norm": 0.6154395341873169, + "learning_rate": 0.00019996306160355708, + "loss": 0.2667, + "step": 379 + }, + { + "epoch": 0.008907798425898252, + "grad_norm": 0.6590301394462585, + "learning_rate": 0.00019996286113657178, + "loss": 0.3846, + "step": 380 + }, + { + "epoch": 0.008931240000703247, + "grad_norm": 0.7259660363197327, + "learning_rate": 0.0001999626601271852, + "loss": 0.5183, + "step": 381 + }, + { + "epoch": 0.008954681575508243, + "grad_norm": 0.6820341348648071, + "learning_rate": 0.00019996245857539842, + "loss": 0.2178, + "step": 382 + }, + { + "epoch": 0.008978123150313239, + "grad_norm": 0.259586364030838, + "learning_rate": 0.00019996225648121252, + "loss": 0.0506, + "step": 383 + }, + { + "epoch": 0.009001564725118233, + "grad_norm": 0.6221457123756409, + "learning_rate": 0.0001999620538446286, + "loss": 0.2433, + "step": 384 + }, + { + "epoch": 0.009025006299923229, + "grad_norm": 0.5683369636535645, + "learning_rate": 0.00019996185066564778, + "loss": 0.2262, + "step": 385 + }, + { + "epoch": 0.009048447874728225, + "grad_norm": 0.49663597345352173, + "learning_rate": 0.0001999616469442711, + "loss": 0.2214, + "step": 386 + }, + { + "epoch": 0.00907188944953322, + "grad_norm": 0.20031699538230896, + "learning_rate": 0.00019996144268049978, + "loss": 0.0503, + "step": 387 + }, + { + "epoch": 0.009095331024338215, + "grad_norm": 0.29964399337768555, + "learning_rate": 0.00019996123787433483, + "loss": 0.0979, + "step": 388 + }, + { + "epoch": 0.00911877259914321, + "grad_norm": 0.8354300260543823, + "learning_rate": 0.0001999610325257774, + "loss": 0.2111, + "step": 389 + }, + { + "epoch": 0.009142214173948206, + "grad_norm": 0.5974153280258179, + "learning_rate": 0.00019996082663482857, + "loss": 0.1467, + "step": 390 + }, + { + "epoch": 0.009165655748753202, + "grad_norm": 0.3378881514072418, + "learning_rate": 0.0001999606202014895, + "loss": 0.1357, + "step": 391 + }, + { + "epoch": 0.009189097323558196, + "grad_norm": 0.5921586751937866, + "learning_rate": 0.0001999604132257613, + "loss": 0.1247, + "step": 392 + }, + { + "epoch": 0.009212538898363192, + "grad_norm": 0.6025273203849792, + "learning_rate": 0.00019996020570764512, + "loss": 0.2143, + "step": 393 + }, + { + "epoch": 0.009235980473168188, + "grad_norm": 0.6158013343811035, + "learning_rate": 0.00019995999764714202, + "loss": 0.2431, + "step": 394 + }, + { + "epoch": 0.009259422047973183, + "grad_norm": 1.097886323928833, + "learning_rate": 0.00019995978904425315, + "loss": 0.2904, + "step": 395 + }, + { + "epoch": 0.009282863622778179, + "grad_norm": 0.7492481470108032, + "learning_rate": 0.00019995957989897965, + "loss": 0.3578, + "step": 396 + }, + { + "epoch": 0.009306305197583173, + "grad_norm": 0.886577308177948, + "learning_rate": 0.00019995937021132266, + "loss": 0.3842, + "step": 397 + }, + { + "epoch": 0.009329746772388169, + "grad_norm": 0.8143008351325989, + "learning_rate": 0.00019995915998128334, + "loss": 0.3743, + "step": 398 + }, + { + "epoch": 0.009353188347193165, + "grad_norm": 0.2816913425922394, + "learning_rate": 0.0001999589492088628, + "loss": 0.1559, + "step": 399 + }, + { + "epoch": 0.00937662992199816, + "grad_norm": 0.5959277153015137, + "learning_rate": 0.0001999587378940622, + "loss": 0.2568, + "step": 400 + }, + { + "epoch": 0.009400071496803155, + "grad_norm": 0.6672255992889404, + "learning_rate": 0.00019995852603688267, + "loss": 0.1664, + "step": 401 + }, + { + "epoch": 0.009423513071608151, + "grad_norm": 0.6576876044273376, + "learning_rate": 0.00019995831363732537, + "loss": 0.3663, + "step": 402 + }, + { + "epoch": 0.009446954646413146, + "grad_norm": 0.7870962619781494, + "learning_rate": 0.00019995810069539145, + "loss": 0.2915, + "step": 403 + }, + { + "epoch": 0.009470396221218142, + "grad_norm": 0.2316613793373108, + "learning_rate": 0.00019995788721108206, + "loss": 0.0903, + "step": 404 + }, + { + "epoch": 0.009493837796023136, + "grad_norm": 0.2614646255970001, + "learning_rate": 0.00019995767318439835, + "loss": 0.0913, + "step": 405 + }, + { + "epoch": 0.009517279370828132, + "grad_norm": 0.4343143105506897, + "learning_rate": 0.00019995745861534151, + "loss": 0.2624, + "step": 406 + }, + { + "epoch": 0.009540720945633128, + "grad_norm": 0.7314099669456482, + "learning_rate": 0.00019995724350391268, + "loss": 0.2357, + "step": 407 + }, + { + "epoch": 0.009564162520438122, + "grad_norm": 0.6501837968826294, + "learning_rate": 0.00019995702785011307, + "loss": 0.2823, + "step": 408 + }, + { + "epoch": 0.009587604095243119, + "grad_norm": 0.1931079477071762, + "learning_rate": 0.00019995681165394378, + "loss": 0.0536, + "step": 409 + }, + { + "epoch": 0.009611045670048115, + "grad_norm": 0.5088310837745667, + "learning_rate": 0.00019995659491540602, + "loss": 0.19, + "step": 410 + }, + { + "epoch": 0.009634487244853109, + "grad_norm": 0.30636927485466003, + "learning_rate": 0.00019995637763450098, + "loss": 0.0804, + "step": 411 + }, + { + "epoch": 0.009657928819658105, + "grad_norm": 0.7454654574394226, + "learning_rate": 0.00019995615981122978, + "loss": 0.3115, + "step": 412 + }, + { + "epoch": 0.0096813703944631, + "grad_norm": 0.3022196590900421, + "learning_rate": 0.00019995594144559372, + "loss": 0.0719, + "step": 413 + }, + { + "epoch": 0.009704811969268095, + "grad_norm": 0.7151706218719482, + "learning_rate": 0.00019995572253759386, + "loss": 0.2602, + "step": 414 + }, + { + "epoch": 0.009728253544073091, + "grad_norm": 0.5330536961555481, + "learning_rate": 0.00019995550308723147, + "loss": 0.1581, + "step": 415 + }, + { + "epoch": 0.009751695118878086, + "grad_norm": 0.6118378639221191, + "learning_rate": 0.0001999552830945077, + "loss": 0.2509, + "step": 416 + }, + { + "epoch": 0.009775136693683082, + "grad_norm": 0.44109079241752625, + "learning_rate": 0.0001999550625594237, + "loss": 0.1104, + "step": 417 + }, + { + "epoch": 0.009798578268488078, + "grad_norm": 0.16235236823558807, + "learning_rate": 0.0001999548414819808, + "loss": 0.0212, + "step": 418 + }, + { + "epoch": 0.009822019843293072, + "grad_norm": 0.6356249451637268, + "learning_rate": 0.0001999546198621801, + "loss": 0.2059, + "step": 419 + }, + { + "epoch": 0.009845461418098068, + "grad_norm": 1.1880903244018555, + "learning_rate": 0.0001999543977000228, + "loss": 0.3845, + "step": 420 + }, + { + "epoch": 0.009868902992903062, + "grad_norm": 0.7581347823143005, + "learning_rate": 0.00019995417499551012, + "loss": 0.3444, + "step": 421 + }, + { + "epoch": 0.009892344567708058, + "grad_norm": 0.6067566871643066, + "learning_rate": 0.00019995395174864328, + "loss": 0.1959, + "step": 422 + }, + { + "epoch": 0.009915786142513055, + "grad_norm": 0.29686808586120605, + "learning_rate": 0.0001999537279594235, + "loss": 0.0709, + "step": 423 + }, + { + "epoch": 0.009939227717318049, + "grad_norm": 0.529879629611969, + "learning_rate": 0.00019995350362785202, + "loss": 0.3009, + "step": 424 + }, + { + "epoch": 0.009962669292123045, + "grad_norm": 0.4692983329296112, + "learning_rate": 0.00019995327875392996, + "loss": 0.2523, + "step": 425 + }, + { + "epoch": 0.009986110866928041, + "grad_norm": 0.6187524199485779, + "learning_rate": 0.00019995305333765862, + "loss": 0.2484, + "step": 426 + }, + { + "epoch": 0.010009552441733035, + "grad_norm": 0.5419464707374573, + "learning_rate": 0.0001999528273790392, + "loss": 0.2384, + "step": 427 + }, + { + "epoch": 0.010032994016538031, + "grad_norm": 0.659604012966156, + "learning_rate": 0.00019995260087807295, + "loss": 0.3593, + "step": 428 + }, + { + "epoch": 0.010056435591343026, + "grad_norm": 0.19014348089694977, + "learning_rate": 0.00019995237383476104, + "loss": 0.0763, + "step": 429 + }, + { + "epoch": 0.010079877166148022, + "grad_norm": 0.6400291919708252, + "learning_rate": 0.00019995214624910477, + "loss": 0.2739, + "step": 430 + }, + { + "epoch": 0.010103318740953018, + "grad_norm": 0.16737399995326996, + "learning_rate": 0.00019995191812110532, + "loss": 0.0643, + "step": 431 + }, + { + "epoch": 0.010126760315758012, + "grad_norm": 0.5155094265937805, + "learning_rate": 0.00019995168945076397, + "loss": 0.2395, + "step": 432 + }, + { + "epoch": 0.010150201890563008, + "grad_norm": 0.24244576692581177, + "learning_rate": 0.00019995146023808194, + "loss": 0.0986, + "step": 433 + }, + { + "epoch": 0.010173643465368004, + "grad_norm": 0.6261517405509949, + "learning_rate": 0.00019995123048306045, + "loss": 0.3232, + "step": 434 + }, + { + "epoch": 0.010197085040172998, + "grad_norm": 0.5514913201332092, + "learning_rate": 0.00019995100018570082, + "loss": 0.2747, + "step": 435 + }, + { + "epoch": 0.010220526614977994, + "grad_norm": 0.5182925462722778, + "learning_rate": 0.00019995076934600423, + "loss": 0.2153, + "step": 436 + }, + { + "epoch": 0.010243968189782989, + "grad_norm": 0.6232104301452637, + "learning_rate": 0.00019995053796397193, + "loss": 0.2611, + "step": 437 + }, + { + "epoch": 0.010267409764587985, + "grad_norm": 0.6415002346038818, + "learning_rate": 0.00019995030603960523, + "loss": 0.4166, + "step": 438 + }, + { + "epoch": 0.010290851339392981, + "grad_norm": 0.36174169182777405, + "learning_rate": 0.00019995007357290537, + "loss": 0.1211, + "step": 439 + }, + { + "epoch": 0.010314292914197975, + "grad_norm": 0.6188311576843262, + "learning_rate": 0.00019994984056387357, + "loss": 0.3186, + "step": 440 + }, + { + "epoch": 0.010337734489002971, + "grad_norm": 0.6219453811645508, + "learning_rate": 0.00019994960701251113, + "loss": 0.2703, + "step": 441 + }, + { + "epoch": 0.010361176063807967, + "grad_norm": 0.7587157487869263, + "learning_rate": 0.00019994937291881933, + "loss": 0.3078, + "step": 442 + }, + { + "epoch": 0.010384617638612962, + "grad_norm": 0.5257847309112549, + "learning_rate": 0.00019994913828279943, + "loss": 0.1669, + "step": 443 + }, + { + "epoch": 0.010408059213417958, + "grad_norm": 0.4872528314590454, + "learning_rate": 0.00019994890310445266, + "loss": 0.1373, + "step": 444 + }, + { + "epoch": 0.010431500788222952, + "grad_norm": 0.43843454122543335, + "learning_rate": 0.00019994866738378035, + "loss": 0.2513, + "step": 445 + }, + { + "epoch": 0.010454942363027948, + "grad_norm": 0.6644647121429443, + "learning_rate": 0.00019994843112078377, + "loss": 0.192, + "step": 446 + }, + { + "epoch": 0.010478383937832944, + "grad_norm": 0.12005545943975449, + "learning_rate": 0.00019994819431546414, + "loss": 0.0355, + "step": 447 + }, + { + "epoch": 0.010501825512637938, + "grad_norm": 0.28230252861976624, + "learning_rate": 0.00019994795696782287, + "loss": 0.0419, + "step": 448 + }, + { + "epoch": 0.010525267087442934, + "grad_norm": 0.31346991658210754, + "learning_rate": 0.00019994771907786113, + "loss": 0.1914, + "step": 449 + }, + { + "epoch": 0.01054870866224793, + "grad_norm": 0.14669907093048096, + "learning_rate": 0.0001999474806455803, + "loss": 0.0426, + "step": 450 + }, + { + "epoch": 0.010572150237052925, + "grad_norm": 0.4708583652973175, + "learning_rate": 0.00019994724167098158, + "loss": 0.2017, + "step": 451 + }, + { + "epoch": 0.01059559181185792, + "grad_norm": 0.17829152941703796, + "learning_rate": 0.00019994700215406633, + "loss": 0.0924, + "step": 452 + }, + { + "epoch": 0.010619033386662915, + "grad_norm": 0.8060371279716492, + "learning_rate": 0.00019994676209483583, + "loss": 0.1556, + "step": 453 + }, + { + "epoch": 0.010642474961467911, + "grad_norm": 0.8929669260978699, + "learning_rate": 0.0001999465214932914, + "loss": 0.2771, + "step": 454 + }, + { + "epoch": 0.010665916536272907, + "grad_norm": 1.3240294456481934, + "learning_rate": 0.00019994628034943434, + "loss": 0.2714, + "step": 455 + }, + { + "epoch": 0.010689358111077902, + "grad_norm": 0.7064284682273865, + "learning_rate": 0.00019994603866326595, + "loss": 0.2535, + "step": 456 + }, + { + "epoch": 0.010712799685882898, + "grad_norm": 0.6368045210838318, + "learning_rate": 0.00019994579643478754, + "loss": 0.2058, + "step": 457 + }, + { + "epoch": 0.010736241260687894, + "grad_norm": 0.5467596650123596, + "learning_rate": 0.00019994555366400043, + "loss": 0.2312, + "step": 458 + }, + { + "epoch": 0.010759682835492888, + "grad_norm": 0.5373201966285706, + "learning_rate": 0.00019994531035090596, + "loss": 0.2875, + "step": 459 + }, + { + "epoch": 0.010783124410297884, + "grad_norm": 0.5930894613265991, + "learning_rate": 0.00019994506649550537, + "loss": 0.1827, + "step": 460 + }, + { + "epoch": 0.010806565985102878, + "grad_norm": 0.8144658207893372, + "learning_rate": 0.00019994482209780007, + "loss": 0.2946, + "step": 461 + }, + { + "epoch": 0.010830007559907874, + "grad_norm": 0.7888110876083374, + "learning_rate": 0.0001999445771577914, + "loss": 0.2634, + "step": 462 + }, + { + "epoch": 0.01085344913471287, + "grad_norm": 0.5213248133659363, + "learning_rate": 0.00019994433167548056, + "loss": 0.2002, + "step": 463 + }, + { + "epoch": 0.010876890709517865, + "grad_norm": 0.44147759675979614, + "learning_rate": 0.00019994408565086902, + "loss": 0.1506, + "step": 464 + }, + { + "epoch": 0.01090033228432286, + "grad_norm": 0.44850003719329834, + "learning_rate": 0.00019994383908395803, + "loss": 0.1723, + "step": 465 + }, + { + "epoch": 0.010923773859127857, + "grad_norm": 0.8170858025550842, + "learning_rate": 0.00019994359197474898, + "loss": 0.23, + "step": 466 + }, + { + "epoch": 0.010947215433932851, + "grad_norm": 0.9881705045700073, + "learning_rate": 0.0001999433443232432, + "loss": 0.25, + "step": 467 + }, + { + "epoch": 0.010970657008737847, + "grad_norm": 0.5028934478759766, + "learning_rate": 0.000199943096129442, + "loss": 0.2465, + "step": 468 + }, + { + "epoch": 0.010994098583542842, + "grad_norm": 0.6120350360870361, + "learning_rate": 0.00019994284739334675, + "loss": 0.275, + "step": 469 + }, + { + "epoch": 0.011017540158347838, + "grad_norm": 0.24405772984027863, + "learning_rate": 0.0001999425981149588, + "loss": 0.0793, + "step": 470 + }, + { + "epoch": 0.011040981733152834, + "grad_norm": 0.66557776927948, + "learning_rate": 0.00019994234829427953, + "loss": 0.2135, + "step": 471 + }, + { + "epoch": 0.011064423307957828, + "grad_norm": 0.2423417717218399, + "learning_rate": 0.00019994209793131024, + "loss": 0.158, + "step": 472 + }, + { + "epoch": 0.011087864882762824, + "grad_norm": 0.8622592687606812, + "learning_rate": 0.00019994184702605235, + "loss": 0.3722, + "step": 473 + }, + { + "epoch": 0.01111130645756782, + "grad_norm": 0.6352599859237671, + "learning_rate": 0.00019994159557850717, + "loss": 0.2501, + "step": 474 + }, + { + "epoch": 0.011134748032372814, + "grad_norm": 0.6557714939117432, + "learning_rate": 0.00019994134358867607, + "loss": 0.35, + "step": 475 + }, + { + "epoch": 0.01115818960717781, + "grad_norm": 0.5305690765380859, + "learning_rate": 0.0001999410910565604, + "loss": 0.239, + "step": 476 + }, + { + "epoch": 0.011181631181982806, + "grad_norm": 0.18752551078796387, + "learning_rate": 0.00019994083798216162, + "loss": 0.0635, + "step": 477 + }, + { + "epoch": 0.0112050727567878, + "grad_norm": 0.36575376987457275, + "learning_rate": 0.00019994058436548099, + "loss": 0.1115, + "step": 478 + }, + { + "epoch": 0.011228514331592797, + "grad_norm": 0.6986315846443176, + "learning_rate": 0.00019994033020652, + "loss": 0.2562, + "step": 479 + }, + { + "epoch": 0.011251955906397791, + "grad_norm": 0.81879061460495, + "learning_rate": 0.00019994007550527994, + "loss": 0.3331, + "step": 480 + }, + { + "epoch": 0.011275397481202787, + "grad_norm": 0.5318918228149414, + "learning_rate": 0.0001999398202617622, + "loss": 0.2258, + "step": 481 + }, + { + "epoch": 0.011298839056007783, + "grad_norm": 0.12566931545734406, + "learning_rate": 0.0001999395644759682, + "loss": 0.0335, + "step": 482 + }, + { + "epoch": 0.011322280630812778, + "grad_norm": 0.8205024600028992, + "learning_rate": 0.0001999393081478993, + "loss": 0.2794, + "step": 483 + }, + { + "epoch": 0.011345722205617774, + "grad_norm": 0.5575319528579712, + "learning_rate": 0.00019993905127755693, + "loss": 0.2201, + "step": 484 + }, + { + "epoch": 0.01136916378042277, + "grad_norm": 0.7015597224235535, + "learning_rate": 0.00019993879386494246, + "loss": 0.1862, + "step": 485 + }, + { + "epoch": 0.011392605355227764, + "grad_norm": 0.3827681541442871, + "learning_rate": 0.00019993853591005728, + "loss": 0.1975, + "step": 486 + }, + { + "epoch": 0.01141604693003276, + "grad_norm": 0.2947011888027191, + "learning_rate": 0.0001999382774129028, + "loss": 0.1472, + "step": 487 + }, + { + "epoch": 0.011439488504837754, + "grad_norm": 0.5288210511207581, + "learning_rate": 0.00019993801837348037, + "loss": 0.1514, + "step": 488 + }, + { + "epoch": 0.01146293007964275, + "grad_norm": 0.7433895468711853, + "learning_rate": 0.0001999377587917915, + "loss": 0.2537, + "step": 489 + }, + { + "epoch": 0.011486371654447746, + "grad_norm": 0.5434961318969727, + "learning_rate": 0.00019993749866783749, + "loss": 0.6307, + "step": 490 + }, + { + "epoch": 0.01150981322925274, + "grad_norm": 0.5819007158279419, + "learning_rate": 0.00019993723800161984, + "loss": 0.1698, + "step": 491 + }, + { + "epoch": 0.011533254804057737, + "grad_norm": 0.6132246851921082, + "learning_rate": 0.0001999369767931399, + "loss": 0.2387, + "step": 492 + }, + { + "epoch": 0.011556696378862733, + "grad_norm": 0.6878663897514343, + "learning_rate": 0.00019993671504239913, + "loss": 0.4026, + "step": 493 + }, + { + "epoch": 0.011580137953667727, + "grad_norm": 0.549267590045929, + "learning_rate": 0.00019993645274939893, + "loss": 0.1856, + "step": 494 + }, + { + "epoch": 0.011603579528472723, + "grad_norm": 0.341360867023468, + "learning_rate": 0.0001999361899141407, + "loss": 0.1411, + "step": 495 + }, + { + "epoch": 0.011627021103277718, + "grad_norm": 0.3034430146217346, + "learning_rate": 0.00019993592653662595, + "loss": 0.0994, + "step": 496 + }, + { + "epoch": 0.011650462678082714, + "grad_norm": 0.576898992061615, + "learning_rate": 0.00019993566261685604, + "loss": 0.1793, + "step": 497 + }, + { + "epoch": 0.01167390425288771, + "grad_norm": 0.25170543789863586, + "learning_rate": 0.0001999353981548324, + "loss": 0.0883, + "step": 498 + }, + { + "epoch": 0.011697345827692704, + "grad_norm": 0.47391194105148315, + "learning_rate": 0.0001999351331505565, + "loss": 0.2839, + "step": 499 + }, + { + "epoch": 0.0117207874024977, + "grad_norm": 0.13042376935482025, + "learning_rate": 0.00019993486760402972, + "loss": 0.0305, + "step": 500 + }, + { + "epoch": 0.011744228977302696, + "grad_norm": 0.7089911699295044, + "learning_rate": 0.00019993460151525354, + "loss": 0.2874, + "step": 501 + }, + { + "epoch": 0.01176767055210769, + "grad_norm": 0.6009535193443298, + "learning_rate": 0.00019993433488422942, + "loss": 0.2109, + "step": 502 + }, + { + "epoch": 0.011791112126912686, + "grad_norm": 0.8218595385551453, + "learning_rate": 0.00019993406771095877, + "loss": 0.3646, + "step": 503 + }, + { + "epoch": 0.01181455370171768, + "grad_norm": 0.49470260739326477, + "learning_rate": 0.0001999337999954431, + "loss": 0.2111, + "step": 504 + }, + { + "epoch": 0.011837995276522677, + "grad_norm": 0.7235592603683472, + "learning_rate": 0.0001999335317376838, + "loss": 0.2721, + "step": 505 + }, + { + "epoch": 0.011861436851327673, + "grad_norm": 0.8272290825843811, + "learning_rate": 0.00019993326293768233, + "loss": 0.3595, + "step": 506 + }, + { + "epoch": 0.011884878426132667, + "grad_norm": 0.4907560348510742, + "learning_rate": 0.0001999329935954402, + "loss": 0.2536, + "step": 507 + }, + { + "epoch": 0.011908320000937663, + "grad_norm": 0.8301956653594971, + "learning_rate": 0.0001999327237109588, + "loss": 0.2112, + "step": 508 + }, + { + "epoch": 0.01193176157574266, + "grad_norm": 0.6780012249946594, + "learning_rate": 0.00019993245328423967, + "loss": 0.1926, + "step": 509 + }, + { + "epoch": 0.011955203150547654, + "grad_norm": 0.5232536792755127, + "learning_rate": 0.00019993218231528418, + "loss": 0.95, + "step": 510 + }, + { + "epoch": 0.01197864472535265, + "grad_norm": 0.21924926340579987, + "learning_rate": 0.0001999319108040939, + "loss": 0.0917, + "step": 511 + }, + { + "epoch": 0.012002086300157644, + "grad_norm": 0.2537566125392914, + "learning_rate": 0.00019993163875067026, + "loss": 0.1008, + "step": 512 + }, + { + "epoch": 0.01202552787496264, + "grad_norm": 0.45175445079803467, + "learning_rate": 0.00019993136615501476, + "loss": 0.1565, + "step": 513 + }, + { + "epoch": 0.012048969449767636, + "grad_norm": 0.5711517930030823, + "learning_rate": 0.0001999310930171288, + "loss": 0.9519, + "step": 514 + }, + { + "epoch": 0.01207241102457263, + "grad_norm": 0.2406393438577652, + "learning_rate": 0.00019993081933701397, + "loss": 0.0925, + "step": 515 + }, + { + "epoch": 0.012095852599377626, + "grad_norm": 0.5500733256340027, + "learning_rate": 0.00019993054511467168, + "loss": 0.1487, + "step": 516 + }, + { + "epoch": 0.012119294174182622, + "grad_norm": 0.8804779648780823, + "learning_rate": 0.00019993027035010346, + "loss": 0.2454, + "step": 517 + }, + { + "epoch": 0.012142735748987617, + "grad_norm": 0.42231565713882446, + "learning_rate": 0.00019992999504331077, + "loss": 0.1146, + "step": 518 + }, + { + "epoch": 0.012166177323792613, + "grad_norm": 0.17540986835956573, + "learning_rate": 0.00019992971919429513, + "loss": 0.0567, + "step": 519 + }, + { + "epoch": 0.012189618898597607, + "grad_norm": 0.2802681624889374, + "learning_rate": 0.00019992944280305798, + "loss": 0.0841, + "step": 520 + }, + { + "epoch": 0.012213060473402603, + "grad_norm": 0.5718349814414978, + "learning_rate": 0.00019992916586960093, + "loss": 0.2998, + "step": 521 + }, + { + "epoch": 0.0122365020482076, + "grad_norm": 0.44984936714172363, + "learning_rate": 0.0001999288883939254, + "loss": 0.6132, + "step": 522 + }, + { + "epoch": 0.012259943623012593, + "grad_norm": 0.708790123462677, + "learning_rate": 0.00019992861037603285, + "loss": 0.2313, + "step": 523 + }, + { + "epoch": 0.01228338519781759, + "grad_norm": 0.667547881603241, + "learning_rate": 0.0001999283318159249, + "loss": 0.2559, + "step": 524 + }, + { + "epoch": 0.012306826772622586, + "grad_norm": 0.7393106818199158, + "learning_rate": 0.00019992805271360303, + "loss": 0.2801, + "step": 525 + }, + { + "epoch": 0.01233026834742758, + "grad_norm": 0.7608263492584229, + "learning_rate": 0.0001999277730690687, + "loss": 0.2534, + "step": 526 + }, + { + "epoch": 0.012353709922232576, + "grad_norm": 0.8572748899459839, + "learning_rate": 0.00019992749288232347, + "loss": 0.3037, + "step": 527 + }, + { + "epoch": 0.01237715149703757, + "grad_norm": 0.43121567368507385, + "learning_rate": 0.00019992721215336887, + "loss": 0.1341, + "step": 528 + }, + { + "epoch": 0.012400593071842566, + "grad_norm": 0.4770892262458801, + "learning_rate": 0.0001999269308822064, + "loss": 0.1653, + "step": 529 + }, + { + "epoch": 0.012424034646647562, + "grad_norm": 0.6240641474723816, + "learning_rate": 0.0001999266490688376, + "loss": 0.348, + "step": 530 + }, + { + "epoch": 0.012447476221452557, + "grad_norm": 0.23262853920459747, + "learning_rate": 0.00019992636671326399, + "loss": 0.0817, + "step": 531 + }, + { + "epoch": 0.012470917796257553, + "grad_norm": 0.44457101821899414, + "learning_rate": 0.00019992608381548709, + "loss": 0.1706, + "step": 532 + }, + { + "epoch": 0.012494359371062549, + "grad_norm": 0.4952245354652405, + "learning_rate": 0.00019992580037550846, + "loss": 0.7539, + "step": 533 + }, + { + "epoch": 0.012517800945867543, + "grad_norm": 0.43034395575523376, + "learning_rate": 0.00019992551639332962, + "loss": 0.1776, + "step": 534 + }, + { + "epoch": 0.012541242520672539, + "grad_norm": 0.42816853523254395, + "learning_rate": 0.00019992523186895215, + "loss": 0.4895, + "step": 535 + }, + { + "epoch": 0.012564684095477533, + "grad_norm": 0.24319849908351898, + "learning_rate": 0.00019992494680237753, + "loss": 0.1181, + "step": 536 + }, + { + "epoch": 0.01258812567028253, + "grad_norm": 0.44329744577407837, + "learning_rate": 0.00019992466119360733, + "loss": 0.1522, + "step": 537 + }, + { + "epoch": 0.012611567245087526, + "grad_norm": 0.6837267875671387, + "learning_rate": 0.0001999243750426431, + "loss": 0.2943, + "step": 538 + }, + { + "epoch": 0.01263500881989252, + "grad_norm": 0.5570329427719116, + "learning_rate": 0.00019992408834948645, + "loss": 0.2563, + "step": 539 + }, + { + "epoch": 0.012658450394697516, + "grad_norm": 0.5458884239196777, + "learning_rate": 0.00019992380111413883, + "loss": 0.2184, + "step": 540 + }, + { + "epoch": 0.012681891969502512, + "grad_norm": 0.13460184633731842, + "learning_rate": 0.00019992351333660192, + "loss": 0.0375, + "step": 541 + }, + { + "epoch": 0.012705333544307506, + "grad_norm": 0.7200286984443665, + "learning_rate": 0.00019992322501687716, + "loss": 0.4291, + "step": 542 + }, + { + "epoch": 0.012728775119112502, + "grad_norm": 0.4758561849594116, + "learning_rate": 0.0001999229361549662, + "loss": 0.2149, + "step": 543 + }, + { + "epoch": 0.012752216693917497, + "grad_norm": 0.7225779891014099, + "learning_rate": 0.00019992264675087058, + "loss": 0.2841, + "step": 544 + }, + { + "epoch": 0.012775658268722493, + "grad_norm": 0.6032595634460449, + "learning_rate": 0.00019992235680459187, + "loss": 0.2236, + "step": 545 + }, + { + "epoch": 0.012799099843527489, + "grad_norm": 0.42449676990509033, + "learning_rate": 0.00019992206631613165, + "loss": 0.1407, + "step": 546 + }, + { + "epoch": 0.012822541418332483, + "grad_norm": 0.2103375941514969, + "learning_rate": 0.00019992177528549147, + "loss": 0.0494, + "step": 547 + }, + { + "epoch": 0.012845982993137479, + "grad_norm": 0.6454800963401794, + "learning_rate": 0.00019992148371267293, + "loss": 0.2855, + "step": 548 + }, + { + "epoch": 0.012869424567942475, + "grad_norm": 0.7128476500511169, + "learning_rate": 0.00019992119159767762, + "loss": 0.1585, + "step": 549 + }, + { + "epoch": 0.01289286614274747, + "grad_norm": 0.6816179156303406, + "learning_rate": 0.00019992089894050713, + "loss": 0.2238, + "step": 550 + }, + { + "epoch": 0.012916307717552465, + "grad_norm": 0.23558539152145386, + "learning_rate": 0.00019992060574116298, + "loss": 0.079, + "step": 551 + }, + { + "epoch": 0.01293974929235746, + "grad_norm": 0.46283039450645447, + "learning_rate": 0.00019992031199964685, + "loss": 0.0835, + "step": 552 + }, + { + "epoch": 0.012963190867162456, + "grad_norm": 0.1916387975215912, + "learning_rate": 0.00019992001771596026, + "loss": 0.0592, + "step": 553 + }, + { + "epoch": 0.012986632441967452, + "grad_norm": 0.883942723274231, + "learning_rate": 0.0001999197228901049, + "loss": 0.9411, + "step": 554 + }, + { + "epoch": 0.013010074016772446, + "grad_norm": 0.5870908498764038, + "learning_rate": 0.00019991942752208228, + "loss": 0.1294, + "step": 555 + }, + { + "epoch": 0.013033515591577442, + "grad_norm": 0.17677952349185944, + "learning_rate": 0.00019991913161189405, + "loss": 0.0493, + "step": 556 + }, + { + "epoch": 0.013056957166382438, + "grad_norm": 0.6673420071601868, + "learning_rate": 0.00019991883515954183, + "loss": 0.2706, + "step": 557 + }, + { + "epoch": 0.013080398741187433, + "grad_norm": 0.934374988079071, + "learning_rate": 0.00019991853816502718, + "loss": 0.3044, + "step": 558 + }, + { + "epoch": 0.013103840315992429, + "grad_norm": 0.5648056268692017, + "learning_rate": 0.00019991824062835168, + "loss": 0.2367, + "step": 559 + }, + { + "epoch": 0.013127281890797423, + "grad_norm": 0.6449615955352783, + "learning_rate": 0.00019991794254951706, + "loss": 0.2923, + "step": 560 + }, + { + "epoch": 0.013150723465602419, + "grad_norm": 0.6085073947906494, + "learning_rate": 0.00019991764392852484, + "loss": 0.8168, + "step": 561 + }, + { + "epoch": 0.013174165040407415, + "grad_norm": 0.6915916800498962, + "learning_rate": 0.00019991734476537668, + "loss": 0.3099, + "step": 562 + }, + { + "epoch": 0.01319760661521241, + "grad_norm": 0.5503354072570801, + "learning_rate": 0.00019991704506007422, + "loss": 0.1698, + "step": 563 + }, + { + "epoch": 0.013221048190017405, + "grad_norm": 0.40020129084587097, + "learning_rate": 0.00019991674481261907, + "loss": 0.169, + "step": 564 + }, + { + "epoch": 0.013244489764822402, + "grad_norm": 0.7866767048835754, + "learning_rate": 0.00019991644402301284, + "loss": 0.2964, + "step": 565 + }, + { + "epoch": 0.013267931339627396, + "grad_norm": 0.3078819811344147, + "learning_rate": 0.00019991614269125715, + "loss": 0.1336, + "step": 566 + }, + { + "epoch": 0.013291372914432392, + "grad_norm": 0.49509432911872864, + "learning_rate": 0.00019991584081735368, + "loss": 0.2754, + "step": 567 + }, + { + "epoch": 0.013314814489237386, + "grad_norm": 0.5185694098472595, + "learning_rate": 0.00019991553840130406, + "loss": 0.2063, + "step": 568 + }, + { + "epoch": 0.013338256064042382, + "grad_norm": 0.5430858135223389, + "learning_rate": 0.00019991523544310989, + "loss": 0.2668, + "step": 569 + }, + { + "epoch": 0.013361697638847378, + "grad_norm": 0.4049232602119446, + "learning_rate": 0.00019991493194277285, + "loss": 0.1309, + "step": 570 + }, + { + "epoch": 0.013385139213652373, + "grad_norm": 0.7005218863487244, + "learning_rate": 0.0001999146279002946, + "loss": 0.3248, + "step": 571 + }, + { + "epoch": 0.013408580788457369, + "grad_norm": 0.6463424563407898, + "learning_rate": 0.00019991432331567676, + "loss": 0.1743, + "step": 572 + }, + { + "epoch": 0.013432022363262365, + "grad_norm": 0.5478033423423767, + "learning_rate": 0.000199914018188921, + "loss": 0.091, + "step": 573 + }, + { + "epoch": 0.013455463938067359, + "grad_norm": 0.5143813490867615, + "learning_rate": 0.00019991371252002895, + "loss": 0.167, + "step": 574 + }, + { + "epoch": 0.013478905512872355, + "grad_norm": 0.6184714436531067, + "learning_rate": 0.0001999134063090023, + "loss": 0.1705, + "step": 575 + }, + { + "epoch": 0.01350234708767735, + "grad_norm": 0.6626764535903931, + "learning_rate": 0.0001999130995558427, + "loss": 0.212, + "step": 576 + }, + { + "epoch": 0.013525788662482345, + "grad_norm": 0.8640002012252808, + "learning_rate": 0.0001999127922605518, + "loss": 0.3421, + "step": 577 + }, + { + "epoch": 0.013549230237287341, + "grad_norm": 0.3090156614780426, + "learning_rate": 0.0001999124844231313, + "loss": 0.1276, + "step": 578 + }, + { + "epoch": 0.013572671812092336, + "grad_norm": 0.56193608045578, + "learning_rate": 0.00019991217604358282, + "loss": 0.23, + "step": 579 + }, + { + "epoch": 0.013596113386897332, + "grad_norm": 0.5381534695625305, + "learning_rate": 0.0001999118671219081, + "loss": 0.1702, + "step": 580 + }, + { + "epoch": 0.013619554961702328, + "grad_norm": 0.1784377545118332, + "learning_rate": 0.00019991155765810875, + "loss": 0.0502, + "step": 581 + }, + { + "epoch": 0.013642996536507322, + "grad_norm": 0.581748366355896, + "learning_rate": 0.00019991124765218648, + "loss": 0.215, + "step": 582 + }, + { + "epoch": 0.013666438111312318, + "grad_norm": 0.5499674677848816, + "learning_rate": 0.00019991093710414298, + "loss": 0.2488, + "step": 583 + }, + { + "epoch": 0.013689879686117313, + "grad_norm": 0.7134978175163269, + "learning_rate": 0.00019991062601397994, + "loss": 0.3061, + "step": 584 + }, + { + "epoch": 0.013713321260922309, + "grad_norm": 0.7472466826438904, + "learning_rate": 0.00019991031438169903, + "loss": 0.2748, + "step": 585 + }, + { + "epoch": 0.013736762835727305, + "grad_norm": 0.091724693775177, + "learning_rate": 0.00019991000220730194, + "loss": 0.018, + "step": 586 + }, + { + "epoch": 0.013760204410532299, + "grad_norm": 0.5018664598464966, + "learning_rate": 0.00019990968949079038, + "loss": 0.1332, + "step": 587 + }, + { + "epoch": 0.013783645985337295, + "grad_norm": 0.6025956273078918, + "learning_rate": 0.00019990937623216599, + "loss": 0.1878, + "step": 588 + }, + { + "epoch": 0.013807087560142291, + "grad_norm": 0.3993389308452606, + "learning_rate": 0.00019990906243143056, + "loss": 0.1059, + "step": 589 + }, + { + "epoch": 0.013830529134947285, + "grad_norm": 0.9360159635543823, + "learning_rate": 0.00019990874808858573, + "loss": 0.2004, + "step": 590 + }, + { + "epoch": 0.013853970709752281, + "grad_norm": 0.2304013967514038, + "learning_rate": 0.00019990843320363323, + "loss": 0.0857, + "step": 591 + }, + { + "epoch": 0.013877412284557276, + "grad_norm": 0.6336825489997864, + "learning_rate": 0.00019990811777657473, + "loss": 0.2068, + "step": 592 + }, + { + "epoch": 0.013900853859362272, + "grad_norm": 0.4605937898159027, + "learning_rate": 0.00019990780180741203, + "loss": 0.6579, + "step": 593 + }, + { + "epoch": 0.013924295434167268, + "grad_norm": 0.587249755859375, + "learning_rate": 0.00019990748529614676, + "loss": 0.1847, + "step": 594 + }, + { + "epoch": 0.013947737008972262, + "grad_norm": 0.1613743007183075, + "learning_rate": 0.00019990716824278066, + "loss": 0.0778, + "step": 595 + }, + { + "epoch": 0.013971178583777258, + "grad_norm": 0.7784532308578491, + "learning_rate": 0.00019990685064731547, + "loss": 0.2487, + "step": 596 + }, + { + "epoch": 0.013994620158582254, + "grad_norm": 0.7074769139289856, + "learning_rate": 0.0001999065325097529, + "loss": 0.2522, + "step": 597 + }, + { + "epoch": 0.014018061733387249, + "grad_norm": 0.2792196273803711, + "learning_rate": 0.00019990621383009466, + "loss": 0.1478, + "step": 598 + }, + { + "epoch": 0.014041503308192245, + "grad_norm": 0.6335695385932922, + "learning_rate": 0.00019990589460834252, + "loss": 0.1876, + "step": 599 + }, + { + "epoch": 0.014064944882997239, + "grad_norm": 0.6631762981414795, + "learning_rate": 0.00019990557484449815, + "loss": 0.1944, + "step": 600 + }, + { + "epoch": 0.014088386457802235, + "grad_norm": 0.5583719611167908, + "learning_rate": 0.00019990525453856335, + "loss": 0.1695, + "step": 601 + }, + { + "epoch": 0.014111828032607231, + "grad_norm": 0.6139362454414368, + "learning_rate": 0.00019990493369053983, + "loss": 0.1571, + "step": 602 + }, + { + "epoch": 0.014135269607412225, + "grad_norm": 0.7506747245788574, + "learning_rate": 0.0001999046123004293, + "loss": 0.1671, + "step": 603 + }, + { + "epoch": 0.014158711182217221, + "grad_norm": 0.6077771186828613, + "learning_rate": 0.00019990429036823354, + "loss": 0.1823, + "step": 604 + }, + { + "epoch": 0.014182152757022217, + "grad_norm": 0.5142677426338196, + "learning_rate": 0.0001999039678939543, + "loss": 0.2038, + "step": 605 + }, + { + "epoch": 0.014205594331827212, + "grad_norm": 0.4969141483306885, + "learning_rate": 0.00019990364487759336, + "loss": 0.1493, + "step": 606 + }, + { + "epoch": 0.014229035906632208, + "grad_norm": 0.26726245880126953, + "learning_rate": 0.00019990332131915238, + "loss": 0.0701, + "step": 607 + }, + { + "epoch": 0.014252477481437202, + "grad_norm": 0.6477627754211426, + "learning_rate": 0.0001999029972186332, + "loss": 0.8525, + "step": 608 + }, + { + "epoch": 0.014275919056242198, + "grad_norm": 1.4012728929519653, + "learning_rate": 0.00019990267257603757, + "loss": 0.2497, + "step": 609 + }, + { + "epoch": 0.014299360631047194, + "grad_norm": 0.7440151572227478, + "learning_rate": 0.00019990234739136718, + "loss": 0.3013, + "step": 610 + }, + { + "epoch": 0.014322802205852189, + "grad_norm": 0.635734498500824, + "learning_rate": 0.00019990202166462386, + "loss": 0.2018, + "step": 611 + }, + { + "epoch": 0.014346243780657185, + "grad_norm": 0.2563173472881317, + "learning_rate": 0.00019990169539580937, + "loss": 0.0976, + "step": 612 + }, + { + "epoch": 0.01436968535546218, + "grad_norm": 0.5573287606239319, + "learning_rate": 0.00019990136858492546, + "loss": 0.1273, + "step": 613 + }, + { + "epoch": 0.014393126930267175, + "grad_norm": 0.4459177553653717, + "learning_rate": 0.00019990104123197394, + "loss": 0.1068, + "step": 614 + }, + { + "epoch": 0.014416568505072171, + "grad_norm": 0.16282352805137634, + "learning_rate": 0.00019990071333695654, + "loss": 0.0715, + "step": 615 + }, + { + "epoch": 0.014440010079877165, + "grad_norm": 0.45460405945777893, + "learning_rate": 0.00019990038489987505, + "loss": 0.0924, + "step": 616 + }, + { + "epoch": 0.014463451654682161, + "grad_norm": 0.3604227304458618, + "learning_rate": 0.00019990005592073127, + "loss": 0.0983, + "step": 617 + }, + { + "epoch": 0.014486893229487157, + "grad_norm": 0.7418956756591797, + "learning_rate": 0.000199899726399527, + "loss": 0.1858, + "step": 618 + }, + { + "epoch": 0.014510334804292152, + "grad_norm": 0.5918416976928711, + "learning_rate": 0.000199899396336264, + "loss": 0.2062, + "step": 619 + }, + { + "epoch": 0.014533776379097148, + "grad_norm": 0.4211011528968811, + "learning_rate": 0.00019989906573094404, + "loss": 0.1828, + "step": 620 + }, + { + "epoch": 0.014557217953902144, + "grad_norm": 0.3322004973888397, + "learning_rate": 0.00019989873458356897, + "loss": 0.1685, + "step": 621 + }, + { + "epoch": 0.014580659528707138, + "grad_norm": 0.6451049447059631, + "learning_rate": 0.00019989840289414055, + "loss": 0.1907, + "step": 622 + }, + { + "epoch": 0.014604101103512134, + "grad_norm": 0.28061607480049133, + "learning_rate": 0.00019989807066266057, + "loss": 0.1312, + "step": 623 + }, + { + "epoch": 0.014627542678317128, + "grad_norm": 0.9529527425765991, + "learning_rate": 0.00019989773788913086, + "loss": 0.2773, + "step": 624 + }, + { + "epoch": 0.014650984253122125, + "grad_norm": 0.7876133322715759, + "learning_rate": 0.00019989740457355322, + "loss": 0.2104, + "step": 625 + }, + { + "epoch": 0.01467442582792712, + "grad_norm": 0.18662813305854797, + "learning_rate": 0.00019989707071592944, + "loss": 0.0536, + "step": 626 + }, + { + "epoch": 0.014697867402732115, + "grad_norm": 0.7935323119163513, + "learning_rate": 0.00019989673631626135, + "loss": 0.7655, + "step": 627 + }, + { + "epoch": 0.014721308977537111, + "grad_norm": 0.5465646386146545, + "learning_rate": 0.00019989640137455075, + "loss": 0.7349, + "step": 628 + }, + { + "epoch": 0.014744750552342107, + "grad_norm": 0.8105056285858154, + "learning_rate": 0.0001998960658907995, + "loss": 0.2517, + "step": 629 + }, + { + "epoch": 0.014768192127147101, + "grad_norm": 0.8692085146903992, + "learning_rate": 0.00019989572986500937, + "loss": 0.261, + "step": 630 + }, + { + "epoch": 0.014791633701952097, + "grad_norm": 0.7020800709724426, + "learning_rate": 0.0001998953932971822, + "loss": 0.1853, + "step": 631 + }, + { + "epoch": 0.014815075276757093, + "grad_norm": 0.705134928226471, + "learning_rate": 0.0001998950561873198, + "loss": 0.1761, + "step": 632 + }, + { + "epoch": 0.014838516851562088, + "grad_norm": 0.36112430691719055, + "learning_rate": 0.00019989471853542405, + "loss": 0.0901, + "step": 633 + }, + { + "epoch": 0.014861958426367084, + "grad_norm": 0.9095965027809143, + "learning_rate": 0.00019989438034149674, + "loss": 0.2572, + "step": 634 + }, + { + "epoch": 0.014885400001172078, + "grad_norm": 0.5996527671813965, + "learning_rate": 0.0001998940416055397, + "loss": 0.193, + "step": 635 + }, + { + "epoch": 0.014908841575977074, + "grad_norm": 0.7433411478996277, + "learning_rate": 0.0001998937023275548, + "loss": 0.3962, + "step": 636 + }, + { + "epoch": 0.01493228315078207, + "grad_norm": 0.505038857460022, + "learning_rate": 0.00019989336250754385, + "loss": 0.1641, + "step": 637 + }, + { + "epoch": 0.014955724725587064, + "grad_norm": 0.5721381306648254, + "learning_rate": 0.00019989302214550869, + "loss": 0.1941, + "step": 638 + }, + { + "epoch": 0.01497916630039206, + "grad_norm": 0.4324139654636383, + "learning_rate": 0.0001998926812414512, + "loss": 0.1605, + "step": 639 + }, + { + "epoch": 0.015002607875197057, + "grad_norm": 0.1731496900320053, + "learning_rate": 0.00019989233979537322, + "loss": 0.052, + "step": 640 + }, + { + "epoch": 0.015026049450002051, + "grad_norm": 0.5154364705085754, + "learning_rate": 0.00019989199780727658, + "loss": 0.1206, + "step": 641 + }, + { + "epoch": 0.015049491024807047, + "grad_norm": 0.718360424041748, + "learning_rate": 0.00019989165527716315, + "loss": 0.3095, + "step": 642 + }, + { + "epoch": 0.015072932599612041, + "grad_norm": 0.45002317428588867, + "learning_rate": 0.0001998913122050348, + "loss": 0.232, + "step": 643 + }, + { + "epoch": 0.015096374174417037, + "grad_norm": 0.5010674595832825, + "learning_rate": 0.0001998909685908934, + "loss": 0.3016, + "step": 644 + }, + { + "epoch": 0.015119815749222033, + "grad_norm": 0.21530738472938538, + "learning_rate": 0.00019989062443474077, + "loss": 0.0828, + "step": 645 + }, + { + "epoch": 0.015143257324027028, + "grad_norm": 0.5304890275001526, + "learning_rate": 0.0001998902797365788, + "loss": 0.1741, + "step": 646 + }, + { + "epoch": 0.015166698898832024, + "grad_norm": 0.6440426707267761, + "learning_rate": 0.0001998899344964094, + "loss": 0.2401, + "step": 647 + }, + { + "epoch": 0.01519014047363702, + "grad_norm": 0.7120727300643921, + "learning_rate": 0.00019988958871423437, + "loss": 0.6703, + "step": 648 + }, + { + "epoch": 0.015213582048442014, + "grad_norm": 0.7415287494659424, + "learning_rate": 0.00019988924239005561, + "loss": 0.256, + "step": 649 + }, + { + "epoch": 0.01523702362324701, + "grad_norm": 0.5389621257781982, + "learning_rate": 0.00019988889552387503, + "loss": 0.5537, + "step": 650 + }, + { + "epoch": 0.015260465198052004, + "grad_norm": 0.9104427099227905, + "learning_rate": 0.00019988854811569452, + "loss": 0.3055, + "step": 651 + }, + { + "epoch": 0.015283906772857, + "grad_norm": 0.28037071228027344, + "learning_rate": 0.00019988820016551592, + "loss": 0.0932, + "step": 652 + }, + { + "epoch": 0.015307348347661997, + "grad_norm": 0.5219214558601379, + "learning_rate": 0.00019988785167334114, + "loss": 0.1535, + "step": 653 + }, + { + "epoch": 0.01533078992246699, + "grad_norm": 0.1436864733695984, + "learning_rate": 0.00019988750263917206, + "loss": 0.0397, + "step": 654 + }, + { + "epoch": 0.015354231497271987, + "grad_norm": 0.6337111592292786, + "learning_rate": 0.0001998871530630106, + "loss": 0.7387, + "step": 655 + }, + { + "epoch": 0.015377673072076983, + "grad_norm": 0.629487156867981, + "learning_rate": 0.0001998868029448586, + "loss": 0.1946, + "step": 656 + }, + { + "epoch": 0.015401114646881977, + "grad_norm": 0.5992152094841003, + "learning_rate": 0.00019988645228471806, + "loss": 0.347, + "step": 657 + }, + { + "epoch": 0.015424556221686973, + "grad_norm": 0.5702572464942932, + "learning_rate": 0.00019988610108259076, + "loss": 0.1947, + "step": 658 + }, + { + "epoch": 0.015447997796491968, + "grad_norm": 0.45920059084892273, + "learning_rate": 0.0001998857493384787, + "loss": 0.1202, + "step": 659 + }, + { + "epoch": 0.015471439371296964, + "grad_norm": 0.4662037789821625, + "learning_rate": 0.00019988539705238374, + "loss": 0.1695, + "step": 660 + }, + { + "epoch": 0.01549488094610196, + "grad_norm": 0.6685786247253418, + "learning_rate": 0.00019988504422430782, + "loss": 0.2737, + "step": 661 + }, + { + "epoch": 0.015518322520906954, + "grad_norm": 0.6307722926139832, + "learning_rate": 0.00019988469085425282, + "loss": 0.2513, + "step": 662 + }, + { + "epoch": 0.01554176409571195, + "grad_norm": 0.1641143560409546, + "learning_rate": 0.0001998843369422207, + "loss": 0.0815, + "step": 663 + }, + { + "epoch": 0.015565205670516946, + "grad_norm": 0.5202215313911438, + "learning_rate": 0.00019988398248821335, + "loss": 0.1472, + "step": 664 + }, + { + "epoch": 0.01558864724532194, + "grad_norm": 0.3969402611255646, + "learning_rate": 0.00019988362749223272, + "loss": 0.114, + "step": 665 + }, + { + "epoch": 0.015612088820126937, + "grad_norm": 0.22779987752437592, + "learning_rate": 0.0001998832719542807, + "loss": 0.1054, + "step": 666 + }, + { + "epoch": 0.015635530394931933, + "grad_norm": 0.9606099724769592, + "learning_rate": 0.00019988291587435923, + "loss": 0.2663, + "step": 667 + }, + { + "epoch": 0.015658971969736927, + "grad_norm": 0.5149839520454407, + "learning_rate": 0.00019988255925247028, + "loss": 0.2021, + "step": 668 + }, + { + "epoch": 0.01568241354454192, + "grad_norm": 0.19868052005767822, + "learning_rate": 0.00019988220208861575, + "loss": 0.1188, + "step": 669 + }, + { + "epoch": 0.01570585511934692, + "grad_norm": 0.7900245785713196, + "learning_rate": 0.00019988184438279755, + "loss": 0.1474, + "step": 670 + }, + { + "epoch": 0.015729296694151913, + "grad_norm": 0.659267783164978, + "learning_rate": 0.0001998814861350177, + "loss": 0.2912, + "step": 671 + }, + { + "epoch": 0.015752738268956908, + "grad_norm": 0.2722329795360565, + "learning_rate": 0.00019988112734527805, + "loss": 0.0983, + "step": 672 + }, + { + "epoch": 0.015776179843761905, + "grad_norm": 0.1912352740764618, + "learning_rate": 0.00019988076801358062, + "loss": 0.0754, + "step": 673 + }, + { + "epoch": 0.0157996214185669, + "grad_norm": 0.5369628071784973, + "learning_rate": 0.00019988040813992734, + "loss": 0.1953, + "step": 674 + }, + { + "epoch": 0.015823062993371894, + "grad_norm": 0.8670894503593445, + "learning_rate": 0.00019988004772432015, + "loss": 0.4281, + "step": 675 + }, + { + "epoch": 0.015846504568176892, + "grad_norm": 0.6192448139190674, + "learning_rate": 0.000199879686766761, + "loss": 0.2621, + "step": 676 + }, + { + "epoch": 0.015869946142981886, + "grad_norm": 0.7343265414237976, + "learning_rate": 0.00019987932526725188, + "loss": 0.2331, + "step": 677 + }, + { + "epoch": 0.01589338771778688, + "grad_norm": 0.5789709091186523, + "learning_rate": 0.00019987896322579474, + "loss": 0.21, + "step": 678 + }, + { + "epoch": 0.015916829292591875, + "grad_norm": 0.4419856369495392, + "learning_rate": 0.00019987860064239152, + "loss": 0.1969, + "step": 679 + }, + { + "epoch": 0.015940270867396873, + "grad_norm": 0.4342624545097351, + "learning_rate": 0.00019987823751704419, + "loss": 0.1029, + "step": 680 + }, + { + "epoch": 0.015963712442201867, + "grad_norm": 0.5171717405319214, + "learning_rate": 0.00019987787384975476, + "loss": 0.6722, + "step": 681 + }, + { + "epoch": 0.01598715401700686, + "grad_norm": 1.0158748626708984, + "learning_rate": 0.00019987750964052517, + "loss": 0.1956, + "step": 682 + }, + { + "epoch": 0.01601059559181186, + "grad_norm": 0.1515653133392334, + "learning_rate": 0.0001998771448893574, + "loss": 0.0443, + "step": 683 + }, + { + "epoch": 0.016034037166616853, + "grad_norm": 0.6483715176582336, + "learning_rate": 0.00019987677959625345, + "loss": 0.2853, + "step": 684 + }, + { + "epoch": 0.016057478741421848, + "grad_norm": 0.8484935760498047, + "learning_rate": 0.00019987641376121527, + "loss": 0.266, + "step": 685 + }, + { + "epoch": 0.016080920316226845, + "grad_norm": 0.582958459854126, + "learning_rate": 0.00019987604738424487, + "loss": 0.358, + "step": 686 + }, + { + "epoch": 0.01610436189103184, + "grad_norm": 0.15327893197536469, + "learning_rate": 0.00019987568046534422, + "loss": 0.0502, + "step": 687 + }, + { + "epoch": 0.016127803465836834, + "grad_norm": 0.21786658465862274, + "learning_rate": 0.0001998753130045153, + "loss": 0.0364, + "step": 688 + }, + { + "epoch": 0.016151245040641832, + "grad_norm": 0.6710576415061951, + "learning_rate": 0.00019987494500176017, + "loss": 0.1574, + "step": 689 + }, + { + "epoch": 0.016174686615446826, + "grad_norm": 0.7413833141326904, + "learning_rate": 0.00019987457645708078, + "loss": 0.3334, + "step": 690 + }, + { + "epoch": 0.01619812819025182, + "grad_norm": 0.709855318069458, + "learning_rate": 0.00019987420737047908, + "loss": 0.353, + "step": 691 + }, + { + "epoch": 0.016221569765056818, + "grad_norm": 0.613691508769989, + "learning_rate": 0.00019987383774195715, + "loss": 0.1702, + "step": 692 + }, + { + "epoch": 0.016245011339861812, + "grad_norm": 0.7642368078231812, + "learning_rate": 0.00019987346757151695, + "loss": 0.1297, + "step": 693 + }, + { + "epoch": 0.016268452914666807, + "grad_norm": 0.16106946766376495, + "learning_rate": 0.00019987309685916055, + "loss": 0.0401, + "step": 694 + }, + { + "epoch": 0.0162918944894718, + "grad_norm": 0.20553815364837646, + "learning_rate": 0.00019987272560488987, + "loss": 0.0723, + "step": 695 + }, + { + "epoch": 0.0163153360642768, + "grad_norm": 0.748511552810669, + "learning_rate": 0.00019987235380870702, + "loss": 0.1537, + "step": 696 + }, + { + "epoch": 0.016338777639081793, + "grad_norm": 0.7615611553192139, + "learning_rate": 0.0001998719814706139, + "loss": 0.2414, + "step": 697 + }, + { + "epoch": 0.016362219213886788, + "grad_norm": 0.5420604944229126, + "learning_rate": 0.00019987160859061268, + "loss": 0.1864, + "step": 698 + }, + { + "epoch": 0.016385660788691785, + "grad_norm": 0.620459794998169, + "learning_rate": 0.00019987123516870527, + "loss": 0.2249, + "step": 699 + }, + { + "epoch": 0.01640910236349678, + "grad_norm": 0.34334880113601685, + "learning_rate": 0.00019987086120489372, + "loss": 0.1662, + "step": 700 + }, + { + "epoch": 0.016432543938301774, + "grad_norm": 0.8204813003540039, + "learning_rate": 0.00019987048669918006, + "loss": 0.364, + "step": 701 + }, + { + "epoch": 0.01645598551310677, + "grad_norm": 0.6364107131958008, + "learning_rate": 0.00019987011165156638, + "loss": 0.1408, + "step": 702 + }, + { + "epoch": 0.016479427087911766, + "grad_norm": 0.9615024328231812, + "learning_rate": 0.00019986973606205463, + "loss": 0.1875, + "step": 703 + }, + { + "epoch": 0.01650286866271676, + "grad_norm": 0.7562951445579529, + "learning_rate": 0.00019986935993064687, + "loss": 0.2154, + "step": 704 + }, + { + "epoch": 0.016526310237521758, + "grad_norm": 0.2351832240819931, + "learning_rate": 0.00019986898325734516, + "loss": 0.0454, + "step": 705 + }, + { + "epoch": 0.016549751812326752, + "grad_norm": 0.24975982308387756, + "learning_rate": 0.00019986860604215157, + "loss": 0.0631, + "step": 706 + }, + { + "epoch": 0.016573193387131747, + "grad_norm": 0.6284460425376892, + "learning_rate": 0.0001998682282850681, + "loss": 0.802, + "step": 707 + }, + { + "epoch": 0.016596634961936745, + "grad_norm": 0.7846702337265015, + "learning_rate": 0.0001998678499860968, + "loss": 0.1411, + "step": 708 + }, + { + "epoch": 0.01662007653674174, + "grad_norm": 0.442792683839798, + "learning_rate": 0.0001998674711452398, + "loss": 0.129, + "step": 709 + }, + { + "epoch": 0.016643518111546733, + "grad_norm": 0.6087958812713623, + "learning_rate": 0.000199867091762499, + "loss": 0.236, + "step": 710 + }, + { + "epoch": 0.016666959686351727, + "grad_norm": 0.25554418563842773, + "learning_rate": 0.0001998667118378766, + "loss": 0.0992, + "step": 711 + }, + { + "epoch": 0.016690401261156725, + "grad_norm": 0.9517407417297363, + "learning_rate": 0.00019986633137137464, + "loss": 0.2518, + "step": 712 + }, + { + "epoch": 0.01671384283596172, + "grad_norm": 0.9014322757720947, + "learning_rate": 0.0001998659503629951, + "loss": 0.2596, + "step": 713 + }, + { + "epoch": 0.016737284410766714, + "grad_norm": 0.7990396022796631, + "learning_rate": 0.00019986556881274015, + "loss": 0.2488, + "step": 714 + }, + { + "epoch": 0.01676072598557171, + "grad_norm": 0.49786052107810974, + "learning_rate": 0.0001998651867206118, + "loss": 0.1466, + "step": 715 + }, + { + "epoch": 0.016784167560376706, + "grad_norm": 0.2769588232040405, + "learning_rate": 0.00019986480408661214, + "loss": 0.0801, + "step": 716 + }, + { + "epoch": 0.0168076091351817, + "grad_norm": 0.6695491075515747, + "learning_rate": 0.00019986442091074325, + "loss": 0.2012, + "step": 717 + }, + { + "epoch": 0.016831050709986698, + "grad_norm": 0.7652488350868225, + "learning_rate": 0.00019986403719300717, + "loss": 0.2841, + "step": 718 + }, + { + "epoch": 0.016854492284791692, + "grad_norm": 0.6177108287811279, + "learning_rate": 0.00019986365293340607, + "loss": 0.1498, + "step": 719 + }, + { + "epoch": 0.016877933859596687, + "grad_norm": 0.3895026743412018, + "learning_rate": 0.00019986326813194194, + "loss": 0.1015, + "step": 720 + }, + { + "epoch": 0.016901375434401684, + "grad_norm": 0.3174229860305786, + "learning_rate": 0.00019986288278861692, + "loss": 0.1091, + "step": 721 + }, + { + "epoch": 0.01692481700920668, + "grad_norm": 0.8030871748924255, + "learning_rate": 0.0001998624969034331, + "loss": 0.2233, + "step": 722 + }, + { + "epoch": 0.016948258584011673, + "grad_norm": 0.222248837351799, + "learning_rate": 0.00019986211047639255, + "loss": 0.09, + "step": 723 + }, + { + "epoch": 0.01697170015881667, + "grad_norm": 0.7226714491844177, + "learning_rate": 0.00019986172350749742, + "loss": 0.3252, + "step": 724 + }, + { + "epoch": 0.016995141733621665, + "grad_norm": 0.6229119896888733, + "learning_rate": 0.00019986133599674973, + "loss": 0.186, + "step": 725 + }, + { + "epoch": 0.01701858330842666, + "grad_norm": 0.49077001214027405, + "learning_rate": 0.00019986094794415162, + "loss": 0.1366, + "step": 726 + }, + { + "epoch": 0.017042024883231654, + "grad_norm": 0.45209917426109314, + "learning_rate": 0.0001998605593497052, + "loss": 0.2048, + "step": 727 + }, + { + "epoch": 0.01706546645803665, + "grad_norm": 0.7372874021530151, + "learning_rate": 0.0001998601702134126, + "loss": 0.4372, + "step": 728 + }, + { + "epoch": 0.017088908032841646, + "grad_norm": 0.7262665629386902, + "learning_rate": 0.0001998597805352759, + "loss": 0.2143, + "step": 729 + }, + { + "epoch": 0.01711234960764664, + "grad_norm": 0.4723210632801056, + "learning_rate": 0.0001998593903152972, + "loss": 0.1663, + "step": 730 + }, + { + "epoch": 0.017135791182451638, + "grad_norm": 0.1603878140449524, + "learning_rate": 0.00019985899955347867, + "loss": 0.0344, + "step": 731 + }, + { + "epoch": 0.017159232757256632, + "grad_norm": 0.9798487424850464, + "learning_rate": 0.00019985860824982239, + "loss": 0.5416, + "step": 732 + }, + { + "epoch": 0.017182674332061627, + "grad_norm": 0.5643085241317749, + "learning_rate": 0.00019985821640433051, + "loss": 0.1545, + "step": 733 + }, + { + "epoch": 0.017206115906866624, + "grad_norm": 0.7564507126808167, + "learning_rate": 0.00019985782401700514, + "loss": 0.1824, + "step": 734 + }, + { + "epoch": 0.01722955748167162, + "grad_norm": 0.635838508605957, + "learning_rate": 0.00019985743108784837, + "loss": 0.2408, + "step": 735 + }, + { + "epoch": 0.017252999056476613, + "grad_norm": 0.581157386302948, + "learning_rate": 0.00019985703761686238, + "loss": 0.2423, + "step": 736 + }, + { + "epoch": 0.01727644063128161, + "grad_norm": 0.3222108483314514, + "learning_rate": 0.00019985664360404933, + "loss": 0.1258, + "step": 737 + }, + { + "epoch": 0.017299882206086605, + "grad_norm": 0.6570534706115723, + "learning_rate": 0.00019985624904941132, + "loss": 0.2741, + "step": 738 + }, + { + "epoch": 0.0173233237808916, + "grad_norm": 0.5418664216995239, + "learning_rate": 0.00019985585395295048, + "loss": 0.1992, + "step": 739 + }, + { + "epoch": 0.017346765355696597, + "grad_norm": 0.6183370351791382, + "learning_rate": 0.00019985545831466896, + "loss": 0.1592, + "step": 740 + }, + { + "epoch": 0.01737020693050159, + "grad_norm": 0.5572360754013062, + "learning_rate": 0.00019985506213456892, + "loss": 0.2497, + "step": 741 + }, + { + "epoch": 0.017393648505306586, + "grad_norm": 1.060128092765808, + "learning_rate": 0.0001998546654126525, + "loss": 0.3116, + "step": 742 + }, + { + "epoch": 0.01741709008011158, + "grad_norm": 0.6086611747741699, + "learning_rate": 0.00019985426814892186, + "loss": 0.1867, + "step": 743 + }, + { + "epoch": 0.017440531654916578, + "grad_norm": 0.5583115816116333, + "learning_rate": 0.00019985387034337918, + "loss": 0.304, + "step": 744 + }, + { + "epoch": 0.017463973229721572, + "grad_norm": 0.6076557040214539, + "learning_rate": 0.00019985347199602657, + "loss": 0.1993, + "step": 745 + }, + { + "epoch": 0.017487414804526567, + "grad_norm": 0.16987551748752594, + "learning_rate": 0.00019985307310686624, + "loss": 0.0671, + "step": 746 + }, + { + "epoch": 0.017510856379331564, + "grad_norm": 0.6006963849067688, + "learning_rate": 0.00019985267367590029, + "loss": 0.2809, + "step": 747 + }, + { + "epoch": 0.01753429795413656, + "grad_norm": 0.4998105466365814, + "learning_rate": 0.00019985227370313094, + "loss": 0.1972, + "step": 748 + }, + { + "epoch": 0.017557739528941553, + "grad_norm": 0.7757418155670166, + "learning_rate": 0.00019985187318856032, + "loss": 0.223, + "step": 749 + }, + { + "epoch": 0.01758118110374655, + "grad_norm": 0.45715856552124023, + "learning_rate": 0.00019985147213219068, + "loss": 0.1897, + "step": 750 + }, + { + "epoch": 0.017604622678551545, + "grad_norm": 0.27410733699798584, + "learning_rate": 0.0001998510705340241, + "loss": 0.0866, + "step": 751 + }, + { + "epoch": 0.01762806425335654, + "grad_norm": 0.5733933448791504, + "learning_rate": 0.00019985066839406281, + "loss": 0.1896, + "step": 752 + }, + { + "epoch": 0.017651505828161537, + "grad_norm": 0.6507370471954346, + "learning_rate": 0.00019985026571230901, + "loss": 0.262, + "step": 753 + }, + { + "epoch": 0.01767494740296653, + "grad_norm": 0.6203749775886536, + "learning_rate": 0.00019984986248876486, + "loss": 0.2419, + "step": 754 + }, + { + "epoch": 0.017698388977771526, + "grad_norm": 0.6858208775520325, + "learning_rate": 0.00019984945872343252, + "loss": 0.1332, + "step": 755 + }, + { + "epoch": 0.017721830552576524, + "grad_norm": 0.2775539755821228, + "learning_rate": 0.00019984905441631422, + "loss": 0.108, + "step": 756 + }, + { + "epoch": 0.017745272127381518, + "grad_norm": 0.2330639362335205, + "learning_rate": 0.00019984864956741215, + "loss": 0.0938, + "step": 757 + }, + { + "epoch": 0.017768713702186512, + "grad_norm": 1.14830482006073, + "learning_rate": 0.0001998482441767285, + "loss": 0.5565, + "step": 758 + }, + { + "epoch": 0.017792155276991507, + "grad_norm": 0.28859272599220276, + "learning_rate": 0.0001998478382442655, + "loss": 0.1214, + "step": 759 + }, + { + "epoch": 0.017815596851796504, + "grad_norm": 0.4109710156917572, + "learning_rate": 0.00019984743177002526, + "loss": 0.6518, + "step": 760 + }, + { + "epoch": 0.0178390384266015, + "grad_norm": 0.5149407386779785, + "learning_rate": 0.00019984702475401008, + "loss": 0.2895, + "step": 761 + }, + { + "epoch": 0.017862480001406493, + "grad_norm": 0.3435801565647125, + "learning_rate": 0.00019984661719622216, + "loss": 0.1274, + "step": 762 + }, + { + "epoch": 0.01788592157621149, + "grad_norm": 0.1973167359828949, + "learning_rate": 0.00019984620909666365, + "loss": 0.059, + "step": 763 + }, + { + "epoch": 0.017909363151016485, + "grad_norm": 0.20723921060562134, + "learning_rate": 0.00019984580045533681, + "loss": 0.0844, + "step": 764 + }, + { + "epoch": 0.01793280472582148, + "grad_norm": 0.5382825136184692, + "learning_rate": 0.00019984539127224386, + "loss": 0.6689, + "step": 765 + }, + { + "epoch": 0.017956246300626477, + "grad_norm": 0.6938613653182983, + "learning_rate": 0.000199844981547387, + "loss": 0.1379, + "step": 766 + }, + { + "epoch": 0.01797968787543147, + "grad_norm": 0.16332094371318817, + "learning_rate": 0.00019984457128076848, + "loss": 0.0509, + "step": 767 + }, + { + "epoch": 0.018003129450236466, + "grad_norm": 0.5869006514549255, + "learning_rate": 0.00019984416047239047, + "loss": 0.2165, + "step": 768 + }, + { + "epoch": 0.018026571025041464, + "grad_norm": 0.6051867604255676, + "learning_rate": 0.00019984374912225525, + "loss": 0.2352, + "step": 769 + }, + { + "epoch": 0.018050012599846458, + "grad_norm": 0.6811045408248901, + "learning_rate": 0.00019984333723036506, + "loss": 0.1737, + "step": 770 + }, + { + "epoch": 0.018073454174651452, + "grad_norm": 0.5091527700424194, + "learning_rate": 0.00019984292479672213, + "loss": 0.1898, + "step": 771 + }, + { + "epoch": 0.01809689574945645, + "grad_norm": 0.6732995510101318, + "learning_rate": 0.00019984251182132866, + "loss": 0.2052, + "step": 772 + }, + { + "epoch": 0.018120337324261444, + "grad_norm": 0.30015328526496887, + "learning_rate": 0.0001998420983041869, + "loss": 0.1508, + "step": 773 + }, + { + "epoch": 0.01814377889906644, + "grad_norm": 0.11115707457065582, + "learning_rate": 0.00019984168424529912, + "loss": 0.048, + "step": 774 + }, + { + "epoch": 0.018167220473871433, + "grad_norm": 0.49691563844680786, + "learning_rate": 0.00019984126964466757, + "loss": 0.1568, + "step": 775 + }, + { + "epoch": 0.01819066204867643, + "grad_norm": 0.7290630340576172, + "learning_rate": 0.00019984085450229445, + "loss": 0.6623, + "step": 776 + }, + { + "epoch": 0.018214103623481425, + "grad_norm": 0.7597460150718689, + "learning_rate": 0.00019984043881818208, + "loss": 0.2299, + "step": 777 + }, + { + "epoch": 0.01823754519828642, + "grad_norm": 0.5675699710845947, + "learning_rate": 0.00019984002259233267, + "loss": 0.2928, + "step": 778 + }, + { + "epoch": 0.018260986773091417, + "grad_norm": 0.629774272441864, + "learning_rate": 0.0001998396058247485, + "loss": 0.1544, + "step": 779 + }, + { + "epoch": 0.01828442834789641, + "grad_norm": 0.567751407623291, + "learning_rate": 0.00019983918851543178, + "loss": 0.1237, + "step": 780 + }, + { + "epoch": 0.018307869922701406, + "grad_norm": 0.8562507629394531, + "learning_rate": 0.00019983877066438485, + "loss": 0.2552, + "step": 781 + }, + { + "epoch": 0.018331311497506404, + "grad_norm": 0.8219698071479797, + "learning_rate": 0.00019983835227160995, + "loss": 0.2238, + "step": 782 + }, + { + "epoch": 0.018354753072311398, + "grad_norm": 0.9011059999465942, + "learning_rate": 0.0001998379333371093, + "loss": 0.3118, + "step": 783 + }, + { + "epoch": 0.018378194647116392, + "grad_norm": 0.5809875726699829, + "learning_rate": 0.00019983751386088528, + "loss": 0.7419, + "step": 784 + }, + { + "epoch": 0.01840163622192139, + "grad_norm": 0.7049424052238464, + "learning_rate": 0.00019983709384294006, + "loss": 0.2806, + "step": 785 + }, + { + "epoch": 0.018425077796726384, + "grad_norm": 0.6173737049102783, + "learning_rate": 0.00019983667328327598, + "loss": 0.1656, + "step": 786 + }, + { + "epoch": 0.01844851937153138, + "grad_norm": 0.4937627613544464, + "learning_rate": 0.00019983625218189527, + "loss": 0.1449, + "step": 787 + }, + { + "epoch": 0.018471960946336376, + "grad_norm": 0.5056303143501282, + "learning_rate": 0.0001998358305388003, + "loss": 0.1923, + "step": 788 + }, + { + "epoch": 0.01849540252114137, + "grad_norm": 0.24023230373859406, + "learning_rate": 0.00019983540835399327, + "loss": 0.1033, + "step": 789 + }, + { + "epoch": 0.018518844095946365, + "grad_norm": 0.6816996932029724, + "learning_rate": 0.00019983498562747652, + "loss": 0.7593, + "step": 790 + }, + { + "epoch": 0.018542285670751363, + "grad_norm": 0.464208722114563, + "learning_rate": 0.00019983456235925232, + "loss": 0.1593, + "step": 791 + }, + { + "epoch": 0.018565727245556357, + "grad_norm": 0.4195472002029419, + "learning_rate": 0.00019983413854932297, + "loss": 0.1976, + "step": 792 + }, + { + "epoch": 0.01858916882036135, + "grad_norm": 0.5451998710632324, + "learning_rate": 0.0001998337141976908, + "loss": 0.9307, + "step": 793 + }, + { + "epoch": 0.018612610395166346, + "grad_norm": 0.5875346660614014, + "learning_rate": 0.0001998332893043581, + "loss": 0.2936, + "step": 794 + }, + { + "epoch": 0.018636051969971344, + "grad_norm": 0.7110582590103149, + "learning_rate": 0.00019983286386932714, + "loss": 0.2169, + "step": 795 + }, + { + "epoch": 0.018659493544776338, + "grad_norm": 1.39467191696167, + "learning_rate": 0.00019983243789260024, + "loss": 0.2043, + "step": 796 + }, + { + "epoch": 0.018682935119581332, + "grad_norm": 0.3294582962989807, + "learning_rate": 0.00019983201137417975, + "loss": 0.1655, + "step": 797 + }, + { + "epoch": 0.01870637669438633, + "grad_norm": 0.20827855169773102, + "learning_rate": 0.00019983158431406797, + "loss": 0.0674, + "step": 798 + }, + { + "epoch": 0.018729818269191324, + "grad_norm": 0.5688413381576538, + "learning_rate": 0.00019983115671226716, + "loss": 0.1891, + "step": 799 + }, + { + "epoch": 0.01875325984399632, + "grad_norm": 0.27344411611557007, + "learning_rate": 0.00019983072856877976, + "loss": 0.0945, + "step": 800 + }, + { + "epoch": 0.018776701418801316, + "grad_norm": 0.23606249690055847, + "learning_rate": 0.00019983029988360795, + "loss": 0.0548, + "step": 801 + }, + { + "epoch": 0.01880014299360631, + "grad_norm": 0.15879616141319275, + "learning_rate": 0.00019982987065675416, + "loss": 0.0446, + "step": 802 + }, + { + "epoch": 0.018823584568411305, + "grad_norm": 0.6119068264961243, + "learning_rate": 0.00019982944088822068, + "loss": 0.1879, + "step": 803 + }, + { + "epoch": 0.018847026143216303, + "grad_norm": 0.6196049451828003, + "learning_rate": 0.00019982901057800985, + "loss": 0.3576, + "step": 804 + }, + { + "epoch": 0.018870467718021297, + "grad_norm": 0.4028332233428955, + "learning_rate": 0.000199828579726124, + "loss": 0.519, + "step": 805 + }, + { + "epoch": 0.01889390929282629, + "grad_norm": 0.3934844136238098, + "learning_rate": 0.00019982814833256548, + "loss": 0.1439, + "step": 806 + }, + { + "epoch": 0.01891735086763129, + "grad_norm": 0.7262880206108093, + "learning_rate": 0.0001998277163973366, + "loss": 0.3039, + "step": 807 + }, + { + "epoch": 0.018940792442436283, + "grad_norm": 0.8802475929260254, + "learning_rate": 0.00019982728392043973, + "loss": 0.901, + "step": 808 + }, + { + "epoch": 0.018964234017241278, + "grad_norm": 0.876638650894165, + "learning_rate": 0.0001998268509018772, + "loss": 0.2698, + "step": 809 + }, + { + "epoch": 0.018987675592046272, + "grad_norm": 0.6613205075263977, + "learning_rate": 0.0001998264173416514, + "loss": 0.2348, + "step": 810 + }, + { + "epoch": 0.01901111716685127, + "grad_norm": 0.5422263741493225, + "learning_rate": 0.0001998259832397646, + "loss": 0.1174, + "step": 811 + }, + { + "epoch": 0.019034558741656264, + "grad_norm": 0.38125699758529663, + "learning_rate": 0.00019982554859621928, + "loss": 0.5869, + "step": 812 + }, + { + "epoch": 0.01905800031646126, + "grad_norm": 0.6990520358085632, + "learning_rate": 0.00019982511341101766, + "loss": 0.3153, + "step": 813 + }, + { + "epoch": 0.019081441891266256, + "grad_norm": 0.7107980251312256, + "learning_rate": 0.0001998246776841622, + "loss": 0.2869, + "step": 814 + }, + { + "epoch": 0.01910488346607125, + "grad_norm": 0.8780837655067444, + "learning_rate": 0.00019982424141565524, + "loss": 0.2393, + "step": 815 + }, + { + "epoch": 0.019128325040876245, + "grad_norm": 0.543339192867279, + "learning_rate": 0.00019982380460549913, + "loss": 0.1908, + "step": 816 + }, + { + "epoch": 0.019151766615681243, + "grad_norm": 0.8875446915626526, + "learning_rate": 0.00019982336725369625, + "loss": 0.2228, + "step": 817 + }, + { + "epoch": 0.019175208190486237, + "grad_norm": 0.6630957126617432, + "learning_rate": 0.00019982292936024897, + "loss": 0.1545, + "step": 818 + }, + { + "epoch": 0.01919864976529123, + "grad_norm": 0.7724511623382568, + "learning_rate": 0.00019982249092515965, + "loss": 0.2804, + "step": 819 + }, + { + "epoch": 0.01922209134009623, + "grad_norm": 0.2532182037830353, + "learning_rate": 0.00019982205194843072, + "loss": 0.0909, + "step": 820 + }, + { + "epoch": 0.019245532914901223, + "grad_norm": 0.5681134462356567, + "learning_rate": 0.00019982161243006452, + "loss": 0.199, + "step": 821 + }, + { + "epoch": 0.019268974489706218, + "grad_norm": 0.6192635893821716, + "learning_rate": 0.0001998211723700634, + "loss": 0.2653, + "step": 822 + }, + { + "epoch": 0.019292416064511216, + "grad_norm": 0.13605134189128876, + "learning_rate": 0.00019982073176842986, + "loss": 0.0346, + "step": 823 + }, + { + "epoch": 0.01931585763931621, + "grad_norm": 1.116391658782959, + "learning_rate": 0.00019982029062516618, + "loss": 0.3296, + "step": 824 + }, + { + "epoch": 0.019339299214121204, + "grad_norm": 0.365531861782074, + "learning_rate": 0.0001998198489402748, + "loss": 0.1377, + "step": 825 + }, + { + "epoch": 0.0193627407889262, + "grad_norm": 0.3073025345802307, + "learning_rate": 0.00019981940671375816, + "loss": 0.1295, + "step": 826 + }, + { + "epoch": 0.019386182363731196, + "grad_norm": 0.5730487704277039, + "learning_rate": 0.00019981896394561857, + "loss": 0.1998, + "step": 827 + }, + { + "epoch": 0.01940962393853619, + "grad_norm": 0.6160964965820312, + "learning_rate": 0.0001998185206358585, + "loss": 0.1682, + "step": 828 + }, + { + "epoch": 0.019433065513341185, + "grad_norm": 0.5006136894226074, + "learning_rate": 0.00019981807678448027, + "loss": 0.1556, + "step": 829 + }, + { + "epoch": 0.019456507088146183, + "grad_norm": 0.717441737651825, + "learning_rate": 0.00019981763239148637, + "loss": 0.288, + "step": 830 + }, + { + "epoch": 0.019479948662951177, + "grad_norm": 0.9351259469985962, + "learning_rate": 0.00019981718745687923, + "loss": 0.26, + "step": 831 + }, + { + "epoch": 0.01950339023775617, + "grad_norm": 0.755966067314148, + "learning_rate": 0.00019981674198066122, + "loss": 0.6724, + "step": 832 + }, + { + "epoch": 0.01952683181256117, + "grad_norm": 0.8638993501663208, + "learning_rate": 0.00019981629596283474, + "loss": 0.1957, + "step": 833 + }, + { + "epoch": 0.019550273387366163, + "grad_norm": 0.36098942160606384, + "learning_rate": 0.00019981584940340224, + "loss": 0.155, + "step": 834 + }, + { + "epoch": 0.019573714962171158, + "grad_norm": 0.2532658576965332, + "learning_rate": 0.00019981540230236614, + "loss": 0.1015, + "step": 835 + }, + { + "epoch": 0.019597156536976156, + "grad_norm": 0.6711028814315796, + "learning_rate": 0.00019981495465972885, + "loss": 0.1523, + "step": 836 + }, + { + "epoch": 0.01962059811178115, + "grad_norm": 0.594174861907959, + "learning_rate": 0.0001998145064754928, + "loss": 0.6156, + "step": 837 + }, + { + "epoch": 0.019644039686586144, + "grad_norm": 0.7037493586540222, + "learning_rate": 0.00019981405774966045, + "loss": 0.3085, + "step": 838 + }, + { + "epoch": 0.019667481261391142, + "grad_norm": 0.5806911587715149, + "learning_rate": 0.0001998136084822342, + "loss": 0.246, + "step": 839 + }, + { + "epoch": 0.019690922836196136, + "grad_norm": 0.5944305658340454, + "learning_rate": 0.0001998131586732165, + "loss": 0.2633, + "step": 840 + }, + { + "epoch": 0.01971436441100113, + "grad_norm": 0.7731623649597168, + "learning_rate": 0.00019981270832260983, + "loss": 0.2996, + "step": 841 + }, + { + "epoch": 0.019737805985806125, + "grad_norm": 0.634233832359314, + "learning_rate": 0.00019981225743041656, + "loss": 0.2457, + "step": 842 + }, + { + "epoch": 0.019761247560611123, + "grad_norm": 0.45685169100761414, + "learning_rate": 0.0001998118059966392, + "loss": 0.1562, + "step": 843 + }, + { + "epoch": 0.019784689135416117, + "grad_norm": 0.633160412311554, + "learning_rate": 0.00019981135402128016, + "loss": 0.2613, + "step": 844 + }, + { + "epoch": 0.01980813071022111, + "grad_norm": 0.5905700325965881, + "learning_rate": 0.0001998109015043419, + "loss": 0.2289, + "step": 845 + }, + { + "epoch": 0.01983157228502611, + "grad_norm": 0.5935895442962646, + "learning_rate": 0.0001998104484458269, + "loss": 0.2023, + "step": 846 + }, + { + "epoch": 0.019855013859831103, + "grad_norm": 0.44775840640068054, + "learning_rate": 0.00019980999484573762, + "loss": 0.1177, + "step": 847 + }, + { + "epoch": 0.019878455434636098, + "grad_norm": 0.782088041305542, + "learning_rate": 0.00019980954070407644, + "loss": 0.3015, + "step": 848 + }, + { + "epoch": 0.019901897009441095, + "grad_norm": 0.732153594493866, + "learning_rate": 0.00019980908602084595, + "loss": 0.3097, + "step": 849 + }, + { + "epoch": 0.01992533858424609, + "grad_norm": 0.5082840323448181, + "learning_rate": 0.0001998086307960485, + "loss": 0.2918, + "step": 850 + }, + { + "epoch": 0.019948780159051084, + "grad_norm": 0.5404842495918274, + "learning_rate": 0.00019980817502968665, + "loss": 0.1607, + "step": 851 + }, + { + "epoch": 0.019972221733856082, + "grad_norm": 0.7150543332099915, + "learning_rate": 0.00019980771872176283, + "loss": 0.253, + "step": 852 + }, + { + "epoch": 0.019995663308661076, + "grad_norm": 0.6495437622070312, + "learning_rate": 0.00019980726187227948, + "loss": 0.1222, + "step": 853 + }, + { + "epoch": 0.02001910488346607, + "grad_norm": 0.590838611125946, + "learning_rate": 0.00019980680448123918, + "loss": 0.2657, + "step": 854 + }, + { + "epoch": 0.02004254645827107, + "grad_norm": 0.4214664399623871, + "learning_rate": 0.00019980634654864435, + "loss": 0.1242, + "step": 855 + }, + { + "epoch": 0.020065988033076063, + "grad_norm": 0.5659839510917664, + "learning_rate": 0.00019980588807449743, + "loss": 0.2083, + "step": 856 + }, + { + "epoch": 0.020089429607881057, + "grad_norm": 0.5820354223251343, + "learning_rate": 0.000199805429058801, + "loss": 0.2927, + "step": 857 + }, + { + "epoch": 0.02011287118268605, + "grad_norm": 0.708172619342804, + "learning_rate": 0.00019980496950155747, + "loss": 0.2084, + "step": 858 + }, + { + "epoch": 0.02013631275749105, + "grad_norm": 0.7637394666671753, + "learning_rate": 0.0001998045094027694, + "loss": 0.3589, + "step": 859 + }, + { + "epoch": 0.020159754332296043, + "grad_norm": 0.47605615854263306, + "learning_rate": 0.00019980404876243922, + "loss": 0.2239, + "step": 860 + }, + { + "epoch": 0.020183195907101038, + "grad_norm": 0.18203295767307281, + "learning_rate": 0.00019980358758056948, + "loss": 0.05, + "step": 861 + }, + { + "epoch": 0.020206637481906035, + "grad_norm": 0.5603464841842651, + "learning_rate": 0.0001998031258571627, + "loss": 0.8104, + "step": 862 + }, + { + "epoch": 0.02023007905671103, + "grad_norm": 0.3036572337150574, + "learning_rate": 0.0001998026635922213, + "loss": 0.0732, + "step": 863 + }, + { + "epoch": 0.020253520631516024, + "grad_norm": 0.6009906530380249, + "learning_rate": 0.0001998022007857479, + "loss": 0.2583, + "step": 864 + }, + { + "epoch": 0.020276962206321022, + "grad_norm": 0.3623764216899872, + "learning_rate": 0.0001998017374377449, + "loss": 0.0668, + "step": 865 + }, + { + "epoch": 0.020300403781126016, + "grad_norm": 0.42061322927474976, + "learning_rate": 0.00019980127354821486, + "loss": 0.1546, + "step": 866 + }, + { + "epoch": 0.02032384535593101, + "grad_norm": 0.41968151926994324, + "learning_rate": 0.00019980080911716038, + "loss": 0.1261, + "step": 867 + }, + { + "epoch": 0.020347286930736008, + "grad_norm": 0.28716182708740234, + "learning_rate": 0.00019980034414458385, + "loss": 0.0564, + "step": 868 + }, + { + "epoch": 0.020370728505541003, + "grad_norm": 0.6267808079719543, + "learning_rate": 0.00019979987863048783, + "loss": 0.3441, + "step": 869 + }, + { + "epoch": 0.020394170080345997, + "grad_norm": 0.5394189357757568, + "learning_rate": 0.0001997994125748749, + "loss": 0.1757, + "step": 870 + }, + { + "epoch": 0.020417611655150995, + "grad_norm": 0.4051491618156433, + "learning_rate": 0.00019979894597774752, + "loss": 0.1199, + "step": 871 + }, + { + "epoch": 0.02044105322995599, + "grad_norm": 0.7897968888282776, + "learning_rate": 0.00019979847883910826, + "loss": 0.2223, + "step": 872 + }, + { + "epoch": 0.020464494804760983, + "grad_norm": 0.35679200291633606, + "learning_rate": 0.00019979801115895966, + "loss": 0.122, + "step": 873 + }, + { + "epoch": 0.020487936379565978, + "grad_norm": 0.19271114468574524, + "learning_rate": 0.00019979754293730423, + "loss": 0.1286, + "step": 874 + }, + { + "epoch": 0.020511377954370975, + "grad_norm": 0.6356596946716309, + "learning_rate": 0.00019979707417414454, + "loss": 0.158, + "step": 875 + }, + { + "epoch": 0.02053481952917597, + "grad_norm": 0.542151927947998, + "learning_rate": 0.00019979660486948314, + "loss": 0.2194, + "step": 876 + }, + { + "epoch": 0.020558261103980964, + "grad_norm": 0.9460797309875488, + "learning_rate": 0.00019979613502332252, + "loss": 0.3717, + "step": 877 + }, + { + "epoch": 0.020581702678785962, + "grad_norm": 0.5281556844711304, + "learning_rate": 0.00019979566463566525, + "loss": 0.1298, + "step": 878 + }, + { + "epoch": 0.020605144253590956, + "grad_norm": 0.39157986640930176, + "learning_rate": 0.00019979519370651394, + "loss": 0.1624, + "step": 879 + }, + { + "epoch": 0.02062858582839595, + "grad_norm": 0.3683106601238251, + "learning_rate": 0.0001997947222358711, + "loss": 0.1652, + "step": 880 + }, + { + "epoch": 0.020652027403200948, + "grad_norm": 0.49618688225746155, + "learning_rate": 0.0001997942502237393, + "loss": 0.1483, + "step": 881 + }, + { + "epoch": 0.020675468978005943, + "grad_norm": 0.40317994356155396, + "learning_rate": 0.00019979377767012104, + "loss": 0.1441, + "step": 882 + }, + { + "epoch": 0.020698910552810937, + "grad_norm": 1.184988021850586, + "learning_rate": 0.00019979330457501896, + "loss": 0.2349, + "step": 883 + }, + { + "epoch": 0.020722352127615935, + "grad_norm": 0.6750385165214539, + "learning_rate": 0.0001997928309384356, + "loss": 0.2915, + "step": 884 + }, + { + "epoch": 0.02074579370242093, + "grad_norm": 0.46977338194847107, + "learning_rate": 0.00019979235676037356, + "loss": 0.1447, + "step": 885 + }, + { + "epoch": 0.020769235277225923, + "grad_norm": 0.14100733399391174, + "learning_rate": 0.00019979188204083536, + "loss": 0.0427, + "step": 886 + }, + { + "epoch": 0.02079267685203092, + "grad_norm": 0.5261769890785217, + "learning_rate": 0.0001997914067798236, + "loss": 0.2253, + "step": 887 + }, + { + "epoch": 0.020816118426835915, + "grad_norm": 0.4626210629940033, + "learning_rate": 0.00019979093097734088, + "loss": 0.089, + "step": 888 + }, + { + "epoch": 0.02083956000164091, + "grad_norm": 0.1428050696849823, + "learning_rate": 0.00019979045463338976, + "loss": 0.0393, + "step": 889 + }, + { + "epoch": 0.020863001576445904, + "grad_norm": 0.6458396315574646, + "learning_rate": 0.00019978997774797283, + "loss": 0.1376, + "step": 890 + }, + { + "epoch": 0.020886443151250902, + "grad_norm": 0.43315964937210083, + "learning_rate": 0.00019978950032109268, + "loss": 0.581, + "step": 891 + }, + { + "epoch": 0.020909884726055896, + "grad_norm": 0.3294367492198944, + "learning_rate": 0.00019978902235275189, + "loss": 0.1111, + "step": 892 + }, + { + "epoch": 0.02093332630086089, + "grad_norm": 0.3159511387348175, + "learning_rate": 0.00019978854384295302, + "loss": 0.1133, + "step": 893 + }, + { + "epoch": 0.020956767875665888, + "grad_norm": 0.21729202568531036, + "learning_rate": 0.00019978806479169876, + "loss": 0.0672, + "step": 894 + }, + { + "epoch": 0.020980209450470882, + "grad_norm": 0.5190734267234802, + "learning_rate": 0.00019978758519899163, + "loss": 0.1834, + "step": 895 + }, + { + "epoch": 0.021003651025275877, + "grad_norm": 0.6576384902000427, + "learning_rate": 0.00019978710506483424, + "loss": 0.1686, + "step": 896 + }, + { + "epoch": 0.021027092600080875, + "grad_norm": 0.24186009168624878, + "learning_rate": 0.00019978662438922923, + "loss": 0.0782, + "step": 897 + }, + { + "epoch": 0.02105053417488587, + "grad_norm": 0.3849562406539917, + "learning_rate": 0.00019978614317217922, + "loss": 0.1248, + "step": 898 + }, + { + "epoch": 0.021073975749690863, + "grad_norm": 0.48456382751464844, + "learning_rate": 0.00019978566141368674, + "loss": 0.1546, + "step": 899 + }, + { + "epoch": 0.02109741732449586, + "grad_norm": 0.388495534658432, + "learning_rate": 0.0001997851791137545, + "loss": 0.4188, + "step": 900 + }, + { + "epoch": 0.021120858899300855, + "grad_norm": 0.5419715046882629, + "learning_rate": 0.00019978469627238504, + "loss": 0.1743, + "step": 901 + }, + { + "epoch": 0.02114430047410585, + "grad_norm": 1.2148679494857788, + "learning_rate": 0.00019978421288958101, + "loss": 0.2395, + "step": 902 + }, + { + "epoch": 0.021167742048910847, + "grad_norm": 0.6299781203269958, + "learning_rate": 0.00019978372896534507, + "loss": 0.2129, + "step": 903 + }, + { + "epoch": 0.02119118362371584, + "grad_norm": 0.32777729630470276, + "learning_rate": 0.0001997832444996798, + "loss": 0.1698, + "step": 904 + }, + { + "epoch": 0.021214625198520836, + "grad_norm": 0.5747807025909424, + "learning_rate": 0.0001997827594925878, + "loss": 0.119, + "step": 905 + }, + { + "epoch": 0.02123806677332583, + "grad_norm": 0.8726794719696045, + "learning_rate": 0.0001997822739440718, + "loss": 0.1716, + "step": 906 + }, + { + "epoch": 0.021261508348130828, + "grad_norm": 0.44130048155784607, + "learning_rate": 0.00019978178785413434, + "loss": 0.1184, + "step": 907 + }, + { + "epoch": 0.021284949922935822, + "grad_norm": 0.5872118473052979, + "learning_rate": 0.0001997813012227781, + "loss": 0.1784, + "step": 908 + }, + { + "epoch": 0.021308391497740817, + "grad_norm": 0.7269693613052368, + "learning_rate": 0.0001997808140500057, + "loss": 0.2057, + "step": 909 + }, + { + "epoch": 0.021331833072545815, + "grad_norm": 0.4866543412208557, + "learning_rate": 0.00019978032633581982, + "loss": 0.1865, + "step": 910 + }, + { + "epoch": 0.02135527464735081, + "grad_norm": 0.753338098526001, + "learning_rate": 0.00019977983808022306, + "loss": 0.3272, + "step": 911 + }, + { + "epoch": 0.021378716222155803, + "grad_norm": 0.9057136178016663, + "learning_rate": 0.0001997793492832181, + "loss": 0.2167, + "step": 912 + }, + { + "epoch": 0.0214021577969608, + "grad_norm": 0.30720558762550354, + "learning_rate": 0.0001997788599448076, + "loss": 0.2829, + "step": 913 + }, + { + "epoch": 0.021425599371765795, + "grad_norm": 0.6474365592002869, + "learning_rate": 0.0001997783700649942, + "loss": 0.3219, + "step": 914 + }, + { + "epoch": 0.02144904094657079, + "grad_norm": 0.281015008687973, + "learning_rate": 0.00019977787964378055, + "loss": 0.0825, + "step": 915 + }, + { + "epoch": 0.021472482521375787, + "grad_norm": 0.7394567131996155, + "learning_rate": 0.00019977738868116933, + "loss": 0.1965, + "step": 916 + }, + { + "epoch": 0.02149592409618078, + "grad_norm": 0.6128546595573425, + "learning_rate": 0.0001997768971771632, + "loss": 0.7432, + "step": 917 + }, + { + "epoch": 0.021519365670985776, + "grad_norm": 0.6843597292900085, + "learning_rate": 0.00019977640513176478, + "loss": 0.2356, + "step": 918 + }, + { + "epoch": 0.021542807245790774, + "grad_norm": 1.3871873617172241, + "learning_rate": 0.00019977591254497683, + "loss": 0.3022, + "step": 919 + }, + { + "epoch": 0.021566248820595768, + "grad_norm": 0.6924955248832703, + "learning_rate": 0.00019977541941680193, + "loss": 0.2906, + "step": 920 + }, + { + "epoch": 0.021589690395400762, + "grad_norm": 0.5595911145210266, + "learning_rate": 0.0001997749257472428, + "loss": 0.5533, + "step": 921 + }, + { + "epoch": 0.021613131970205757, + "grad_norm": 0.5421935319900513, + "learning_rate": 0.00019977443153630216, + "loss": 0.8074, + "step": 922 + }, + { + "epoch": 0.021636573545010754, + "grad_norm": 0.9225009679794312, + "learning_rate": 0.00019977393678398265, + "loss": 0.1822, + "step": 923 + }, + { + "epoch": 0.02166001511981575, + "grad_norm": 0.3034336268901825, + "learning_rate": 0.0001997734414902869, + "loss": 0.1026, + "step": 924 + }, + { + "epoch": 0.021683456694620743, + "grad_norm": 0.22882544994354248, + "learning_rate": 0.00019977294565521768, + "loss": 0.0554, + "step": 925 + }, + { + "epoch": 0.02170689826942574, + "grad_norm": 0.5400307774543762, + "learning_rate": 0.00019977244927877764, + "loss": 0.1678, + "step": 926 + }, + { + "epoch": 0.021730339844230735, + "grad_norm": 0.25121667981147766, + "learning_rate": 0.0001997719523609695, + "loss": 0.1653, + "step": 927 + }, + { + "epoch": 0.02175378141903573, + "grad_norm": 0.547863781452179, + "learning_rate": 0.00019977145490179593, + "loss": 0.1773, + "step": 928 + }, + { + "epoch": 0.021777222993840727, + "grad_norm": 0.6480681300163269, + "learning_rate": 0.00019977095690125967, + "loss": 0.181, + "step": 929 + }, + { + "epoch": 0.02180066456864572, + "grad_norm": 0.6013523936271667, + "learning_rate": 0.00019977045835936337, + "loss": 0.6133, + "step": 930 + }, + { + "epoch": 0.021824106143450716, + "grad_norm": 0.3693852722644806, + "learning_rate": 0.00019976995927610975, + "loss": 0.1303, + "step": 931 + }, + { + "epoch": 0.021847547718255714, + "grad_norm": 0.3166419565677643, + "learning_rate": 0.00019976945965150154, + "loss": 0.1188, + "step": 932 + }, + { + "epoch": 0.021870989293060708, + "grad_norm": 0.6044907569885254, + "learning_rate": 0.0001997689594855414, + "loss": 0.3486, + "step": 933 + }, + { + "epoch": 0.021894430867865702, + "grad_norm": 0.557529091835022, + "learning_rate": 0.00019976845877823214, + "loss": 0.1066, + "step": 934 + }, + { + "epoch": 0.0219178724426707, + "grad_norm": 0.15929526090621948, + "learning_rate": 0.00019976795752957638, + "loss": 0.0397, + "step": 935 + }, + { + "epoch": 0.021941314017475694, + "grad_norm": 0.23305262625217438, + "learning_rate": 0.00019976745573957688, + "loss": 0.0599, + "step": 936 + }, + { + "epoch": 0.02196475559228069, + "grad_norm": 0.20420151948928833, + "learning_rate": 0.00019976695340823638, + "loss": 0.1003, + "step": 937 + }, + { + "epoch": 0.021988197167085683, + "grad_norm": 0.238168403506279, + "learning_rate": 0.00019976645053555755, + "loss": 0.0644, + "step": 938 + }, + { + "epoch": 0.02201163874189068, + "grad_norm": 0.7006150484085083, + "learning_rate": 0.00019976594712154316, + "loss": 0.2412, + "step": 939 + }, + { + "epoch": 0.022035080316695675, + "grad_norm": 0.7413314580917358, + "learning_rate": 0.00019976544316619598, + "loss": 0.204, + "step": 940 + }, + { + "epoch": 0.02205852189150067, + "grad_norm": 0.3220592737197876, + "learning_rate": 0.00019976493866951867, + "loss": 0.0834, + "step": 941 + }, + { + "epoch": 0.022081963466305667, + "grad_norm": 0.520555853843689, + "learning_rate": 0.000199764433631514, + "loss": 0.1868, + "step": 942 + }, + { + "epoch": 0.02210540504111066, + "grad_norm": 0.6681939363479614, + "learning_rate": 0.0001997639280521847, + "loss": 0.1863, + "step": 943 + }, + { + "epoch": 0.022128846615915656, + "grad_norm": 0.7549737095832825, + "learning_rate": 0.0001997634219315335, + "loss": 0.272, + "step": 944 + }, + { + "epoch": 0.022152288190720654, + "grad_norm": 0.6871306896209717, + "learning_rate": 0.0001997629152695632, + "loss": 0.2357, + "step": 945 + }, + { + "epoch": 0.022175729765525648, + "grad_norm": 0.7043282389640808, + "learning_rate": 0.00019976240806627652, + "loss": 0.2494, + "step": 946 + }, + { + "epoch": 0.022199171340330642, + "grad_norm": 0.5292699337005615, + "learning_rate": 0.00019976190032167622, + "loss": 0.2262, + "step": 947 + }, + { + "epoch": 0.02222261291513564, + "grad_norm": 0.30569684505462646, + "learning_rate": 0.000199761392035765, + "loss": 0.1236, + "step": 948 + }, + { + "epoch": 0.022246054489940634, + "grad_norm": 0.4223461449146271, + "learning_rate": 0.0001997608832085457, + "loss": 0.6393, + "step": 949 + }, + { + "epoch": 0.02226949606474563, + "grad_norm": 0.4625244736671448, + "learning_rate": 0.00019976037384002103, + "loss": 0.1815, + "step": 950 + }, + { + "epoch": 0.022292937639550627, + "grad_norm": 0.30356115102767944, + "learning_rate": 0.00019975986393019378, + "loss": 0.0972, + "step": 951 + }, + { + "epoch": 0.02231637921435562, + "grad_norm": 0.7602765560150146, + "learning_rate": 0.00019975935347906666, + "loss": 0.1775, + "step": 952 + }, + { + "epoch": 0.022339820789160615, + "grad_norm": 0.4764317572116852, + "learning_rate": 0.0001997588424866425, + "loss": 0.5091, + "step": 953 + }, + { + "epoch": 0.022363262363965613, + "grad_norm": 0.8186021447181702, + "learning_rate": 0.0001997583309529241, + "loss": 0.2713, + "step": 954 + }, + { + "epoch": 0.022386703938770607, + "grad_norm": 0.6189667582511902, + "learning_rate": 0.00019975781887791416, + "loss": 0.2127, + "step": 955 + }, + { + "epoch": 0.0224101455135756, + "grad_norm": 0.6619798541069031, + "learning_rate": 0.0001997573062616155, + "loss": 0.1754, + "step": 956 + }, + { + "epoch": 0.022433587088380596, + "grad_norm": 0.4600893259048462, + "learning_rate": 0.00019975679310403087, + "loss": 0.2576, + "step": 957 + }, + { + "epoch": 0.022457028663185594, + "grad_norm": 0.6538305878639221, + "learning_rate": 0.00019975627940516306, + "loss": 0.1748, + "step": 958 + }, + { + "epoch": 0.022480470237990588, + "grad_norm": 0.4385824501514435, + "learning_rate": 0.0001997557651650149, + "loss": 0.6562, + "step": 959 + }, + { + "epoch": 0.022503911812795582, + "grad_norm": 0.4114258587360382, + "learning_rate": 0.00019975525038358916, + "loss": 0.1176, + "step": 960 + }, + { + "epoch": 0.02252735338760058, + "grad_norm": 0.4903501272201538, + "learning_rate": 0.0001997547350608886, + "loss": 0.2021, + "step": 961 + }, + { + "epoch": 0.022550794962405574, + "grad_norm": 0.8052154779434204, + "learning_rate": 0.00019975421919691608, + "loss": 0.3616, + "step": 962 + }, + { + "epoch": 0.02257423653721057, + "grad_norm": 0.4685530364513397, + "learning_rate": 0.00019975370279167435, + "loss": 0.1272, + "step": 963 + }, + { + "epoch": 0.022597678112015566, + "grad_norm": 0.7080509066581726, + "learning_rate": 0.0001997531858451662, + "loss": 0.1588, + "step": 964 + }, + { + "epoch": 0.02262111968682056, + "grad_norm": 0.45451825857162476, + "learning_rate": 0.00019975266835739447, + "loss": 0.1528, + "step": 965 + }, + { + "epoch": 0.022644561261625555, + "grad_norm": 0.6237311959266663, + "learning_rate": 0.000199752150328362, + "loss": 0.2349, + "step": 966 + }, + { + "epoch": 0.022668002836430553, + "grad_norm": 0.4807731807231903, + "learning_rate": 0.0001997516317580715, + "loss": 0.1384, + "step": 967 + }, + { + "epoch": 0.022691444411235547, + "grad_norm": 0.586226761341095, + "learning_rate": 0.00019975111264652586, + "loss": 0.2084, + "step": 968 + }, + { + "epoch": 0.02271488598604054, + "grad_norm": 0.4857969284057617, + "learning_rate": 0.00019975059299372788, + "loss": 0.1398, + "step": 969 + }, + { + "epoch": 0.02273832756084554, + "grad_norm": 0.834439218044281, + "learning_rate": 0.00019975007279968038, + "loss": 0.2234, + "step": 970 + }, + { + "epoch": 0.022761769135650534, + "grad_norm": 0.2713853418827057, + "learning_rate": 0.00019974955206438616, + "loss": 0.072, + "step": 971 + }, + { + "epoch": 0.022785210710455528, + "grad_norm": 0.7392247319221497, + "learning_rate": 0.0001997490307878481, + "loss": 0.1541, + "step": 972 + }, + { + "epoch": 0.022808652285260522, + "grad_norm": 0.5263757109642029, + "learning_rate": 0.00019974850897006897, + "loss": 0.1915, + "step": 973 + }, + { + "epoch": 0.02283209386006552, + "grad_norm": 0.5948006510734558, + "learning_rate": 0.00019974798661105166, + "loss": 0.2073, + "step": 974 + }, + { + "epoch": 0.022855535434870514, + "grad_norm": 0.5414551496505737, + "learning_rate": 0.0001997474637107989, + "loss": 0.1764, + "step": 975 + }, + { + "epoch": 0.02287897700967551, + "grad_norm": 0.9705998301506042, + "learning_rate": 0.00019974694026931363, + "loss": 0.2796, + "step": 976 + }, + { + "epoch": 0.022902418584480506, + "grad_norm": 0.7276982069015503, + "learning_rate": 0.0001997464162865987, + "loss": 0.3575, + "step": 977 + }, + { + "epoch": 0.0229258601592855, + "grad_norm": 0.6535611152648926, + "learning_rate": 0.0001997458917626569, + "loss": 0.2416, + "step": 978 + }, + { + "epoch": 0.022949301734090495, + "grad_norm": 0.5487162470817566, + "learning_rate": 0.00019974536669749104, + "loss": 0.2356, + "step": 979 + }, + { + "epoch": 0.022972743308895493, + "grad_norm": 0.5740652680397034, + "learning_rate": 0.00019974484109110404, + "loss": 0.1603, + "step": 980 + }, + { + "epoch": 0.022996184883700487, + "grad_norm": 0.6367218494415283, + "learning_rate": 0.00019974431494349873, + "loss": 0.3099, + "step": 981 + }, + { + "epoch": 0.02301962645850548, + "grad_norm": 0.18016217648983002, + "learning_rate": 0.00019974378825467795, + "loss": 0.0689, + "step": 982 + }, + { + "epoch": 0.02304306803331048, + "grad_norm": 0.6329823732376099, + "learning_rate": 0.00019974326102464458, + "loss": 0.2885, + "step": 983 + }, + { + "epoch": 0.023066509608115474, + "grad_norm": 0.8090145587921143, + "learning_rate": 0.00019974273325340148, + "loss": 0.3562, + "step": 984 + }, + { + "epoch": 0.023089951182920468, + "grad_norm": 0.6770017743110657, + "learning_rate": 0.0001997422049409515, + "loss": 0.375, + "step": 985 + }, + { + "epoch": 0.023113392757725466, + "grad_norm": 0.5361506342887878, + "learning_rate": 0.0001997416760872975, + "loss": 0.1819, + "step": 986 + }, + { + "epoch": 0.02313683433253046, + "grad_norm": 0.42407292127609253, + "learning_rate": 0.00019974114669244237, + "loss": 0.2689, + "step": 987 + }, + { + "epoch": 0.023160275907335454, + "grad_norm": 1.2921723127365112, + "learning_rate": 0.00019974061675638897, + "loss": 0.2404, + "step": 988 + }, + { + "epoch": 0.02318371748214045, + "grad_norm": 0.4806138873100281, + "learning_rate": 0.00019974008627914017, + "loss": 0.1615, + "step": 989 + }, + { + "epoch": 0.023207159056945446, + "grad_norm": 0.23234735429286957, + "learning_rate": 0.00019973955526069888, + "loss": 0.085, + "step": 990 + }, + { + "epoch": 0.02323060063175044, + "grad_norm": 0.4859183132648468, + "learning_rate": 0.00019973902370106796, + "loss": 0.1914, + "step": 991 + }, + { + "epoch": 0.023254042206555435, + "grad_norm": 0.24454697966575623, + "learning_rate": 0.00019973849160025026, + "loss": 0.1062, + "step": 992 + }, + { + "epoch": 0.023277483781360433, + "grad_norm": 0.4459204375743866, + "learning_rate": 0.00019973795895824874, + "loss": 0.1852, + "step": 993 + }, + { + "epoch": 0.023300925356165427, + "grad_norm": 0.2503916323184967, + "learning_rate": 0.00019973742577506623, + "loss": 0.1165, + "step": 994 + }, + { + "epoch": 0.02332436693097042, + "grad_norm": 0.7631092667579651, + "learning_rate": 0.00019973689205070566, + "loss": 0.274, + "step": 995 + }, + { + "epoch": 0.02334780850577542, + "grad_norm": 0.5033678412437439, + "learning_rate": 0.00019973635778516987, + "loss": 0.2388, + "step": 996 + }, + { + "epoch": 0.023371250080580414, + "grad_norm": 0.614150881767273, + "learning_rate": 0.00019973582297846184, + "loss": 0.2843, + "step": 997 + }, + { + "epoch": 0.023394691655385408, + "grad_norm": 0.11222164332866669, + "learning_rate": 0.00019973528763058443, + "loss": 0.0365, + "step": 998 + }, + { + "epoch": 0.023418133230190406, + "grad_norm": 0.3843648135662079, + "learning_rate": 0.0001997347517415405, + "loss": 0.5336, + "step": 999 + }, + { + "epoch": 0.0234415748049954, + "grad_norm": 0.6190709471702576, + "learning_rate": 0.00019973421531133302, + "loss": 0.1506, + "step": 1000 + }, + { + "epoch": 0.023465016379800394, + "grad_norm": 0.3511750102043152, + "learning_rate": 0.0001997336783399649, + "loss": 0.148, + "step": 1001 + }, + { + "epoch": 0.023488457954605392, + "grad_norm": 0.17774604260921478, + "learning_rate": 0.00019973314082743905, + "loss": 0.071, + "step": 1002 + }, + { + "epoch": 0.023511899529410386, + "grad_norm": 0.235946387052536, + "learning_rate": 0.00019973260277375836, + "loss": 0.0563, + "step": 1003 + }, + { + "epoch": 0.02353534110421538, + "grad_norm": 0.4997907876968384, + "learning_rate": 0.00019973206417892573, + "loss": 0.1255, + "step": 1004 + }, + { + "epoch": 0.023558782679020375, + "grad_norm": 0.40170514583587646, + "learning_rate": 0.00019973152504294416, + "loss": 0.1709, + "step": 1005 + }, + { + "epoch": 0.023582224253825373, + "grad_norm": 0.5228164196014404, + "learning_rate": 0.00019973098536581652, + "loss": 0.1775, + "step": 1006 + }, + { + "epoch": 0.023605665828630367, + "grad_norm": 0.6925870776176453, + "learning_rate": 0.00019973044514754574, + "loss": 0.2544, + "step": 1007 + }, + { + "epoch": 0.02362910740343536, + "grad_norm": 0.8529369831085205, + "learning_rate": 0.00019972990438813475, + "loss": 0.3061, + "step": 1008 + }, + { + "epoch": 0.02365254897824036, + "grad_norm": 0.20368193089962006, + "learning_rate": 0.00019972936308758652, + "loss": 0.0844, + "step": 1009 + }, + { + "epoch": 0.023675990553045353, + "grad_norm": 0.5020921230316162, + "learning_rate": 0.00019972882124590395, + "loss": 0.15, + "step": 1010 + }, + { + "epoch": 0.023699432127850348, + "grad_norm": 0.517438530921936, + "learning_rate": 0.00019972827886309, + "loss": 0.2299, + "step": 1011 + }, + { + "epoch": 0.023722873702655346, + "grad_norm": 0.2502062916755676, + "learning_rate": 0.0001997277359391476, + "loss": 0.0599, + "step": 1012 + }, + { + "epoch": 0.02374631527746034, + "grad_norm": 0.6453468799591064, + "learning_rate": 0.0001997271924740797, + "loss": 0.2528, + "step": 1013 + }, + { + "epoch": 0.023769756852265334, + "grad_norm": 0.4083835184574127, + "learning_rate": 0.00019972664846788925, + "loss": 0.1122, + "step": 1014 + }, + { + "epoch": 0.023793198427070332, + "grad_norm": 0.3937140703201294, + "learning_rate": 0.00019972610392057918, + "loss": 0.1399, + "step": 1015 + }, + { + "epoch": 0.023816640001875326, + "grad_norm": 0.8884684443473816, + "learning_rate": 0.0001997255588321525, + "loss": 0.2408, + "step": 1016 + }, + { + "epoch": 0.02384008157668032, + "grad_norm": 0.16219137609004974, + "learning_rate": 0.0001997250132026121, + "loss": 0.0514, + "step": 1017 + }, + { + "epoch": 0.02386352315148532, + "grad_norm": 0.4636850357055664, + "learning_rate": 0.00019972446703196098, + "loss": 0.2035, + "step": 1018 + }, + { + "epoch": 0.023886964726290313, + "grad_norm": 0.40922340750694275, + "learning_rate": 0.00019972392032020214, + "loss": 0.128, + "step": 1019 + }, + { + "epoch": 0.023910406301095307, + "grad_norm": 0.9046550393104553, + "learning_rate": 0.00019972337306733845, + "loss": 0.3277, + "step": 1020 + }, + { + "epoch": 0.0239338478759003, + "grad_norm": 0.5399825572967529, + "learning_rate": 0.000199722825273373, + "loss": 0.2141, + "step": 1021 + }, + { + "epoch": 0.0239572894507053, + "grad_norm": 0.473209947347641, + "learning_rate": 0.00019972227693830863, + "loss": 0.1668, + "step": 1022 + }, + { + "epoch": 0.023980731025510293, + "grad_norm": 0.39264723658561707, + "learning_rate": 0.0001997217280621484, + "loss": 0.0797, + "step": 1023 + }, + { + "epoch": 0.024004172600315288, + "grad_norm": 0.41059085726737976, + "learning_rate": 0.00019972117864489527, + "loss": 0.1357, + "step": 1024 + }, + { + "epoch": 0.024027614175120286, + "grad_norm": 0.5010368227958679, + "learning_rate": 0.00019972062868655225, + "loss": 0.2129, + "step": 1025 + }, + { + "epoch": 0.02405105574992528, + "grad_norm": 0.3485681712627411, + "learning_rate": 0.00019972007818712224, + "loss": 0.1457, + "step": 1026 + }, + { + "epoch": 0.024074497324730274, + "grad_norm": 0.6708214282989502, + "learning_rate": 0.00019971952714660832, + "loss": 0.2195, + "step": 1027 + }, + { + "epoch": 0.024097938899535272, + "grad_norm": 0.5401322841644287, + "learning_rate": 0.00019971897556501344, + "loss": 0.1567, + "step": 1028 + }, + { + "epoch": 0.024121380474340266, + "grad_norm": 0.5246657133102417, + "learning_rate": 0.00019971842344234059, + "loss": 0.1703, + "step": 1029 + }, + { + "epoch": 0.02414482204914526, + "grad_norm": 0.38511908054351807, + "learning_rate": 0.00019971787077859276, + "loss": 0.4765, + "step": 1030 + }, + { + "epoch": 0.02416826362395026, + "grad_norm": 0.557888388633728, + "learning_rate": 0.000199717317573773, + "loss": 0.2136, + "step": 1031 + }, + { + "epoch": 0.024191705198755253, + "grad_norm": 0.45465362071990967, + "learning_rate": 0.00019971676382788423, + "loss": 0.2646, + "step": 1032 + }, + { + "epoch": 0.024215146773560247, + "grad_norm": 0.6017799973487854, + "learning_rate": 0.0001997162095409295, + "loss": 0.1971, + "step": 1033 + }, + { + "epoch": 0.024238588348365245, + "grad_norm": 0.2942996323108673, + "learning_rate": 0.0001997156547129118, + "loss": 0.1266, + "step": 1034 + }, + { + "epoch": 0.02426202992317024, + "grad_norm": 0.3480985760688782, + "learning_rate": 0.00019971509934383418, + "loss": 0.1539, + "step": 1035 + }, + { + "epoch": 0.024285471497975233, + "grad_norm": 0.64544278383255, + "learning_rate": 0.00019971454343369963, + "loss": 0.1828, + "step": 1036 + }, + { + "epoch": 0.024308913072780228, + "grad_norm": 0.13123035430908203, + "learning_rate": 0.00019971398698251117, + "loss": 0.0377, + "step": 1037 + }, + { + "epoch": 0.024332354647585226, + "grad_norm": 0.6999850273132324, + "learning_rate": 0.00019971342999027179, + "loss": 0.1771, + "step": 1038 + }, + { + "epoch": 0.02435579622239022, + "grad_norm": 0.2407456487417221, + "learning_rate": 0.00019971287245698453, + "loss": 0.078, + "step": 1039 + }, + { + "epoch": 0.024379237797195214, + "grad_norm": 0.2001028060913086, + "learning_rate": 0.00019971231438265243, + "loss": 0.0743, + "step": 1040 + }, + { + "epoch": 0.024402679372000212, + "grad_norm": 0.5092536807060242, + "learning_rate": 0.00019971175576727853, + "loss": 0.1476, + "step": 1041 + }, + { + "epoch": 0.024426120946805206, + "grad_norm": 0.5253428220748901, + "learning_rate": 0.0001997111966108658, + "loss": 0.2938, + "step": 1042 + }, + { + "epoch": 0.0244495625216102, + "grad_norm": 0.6158435940742493, + "learning_rate": 0.00019971063691341735, + "loss": 0.2108, + "step": 1043 + }, + { + "epoch": 0.0244730040964152, + "grad_norm": 0.679253339767456, + "learning_rate": 0.00019971007667493615, + "loss": 0.1576, + "step": 1044 + }, + { + "epoch": 0.024496445671220193, + "grad_norm": 0.4670245349407196, + "learning_rate": 0.0001997095158954253, + "loss": 0.1648, + "step": 1045 + }, + { + "epoch": 0.024519887246025187, + "grad_norm": 0.38780876994132996, + "learning_rate": 0.0001997089545748878, + "loss": 0.1385, + "step": 1046 + }, + { + "epoch": 0.024543328820830185, + "grad_norm": 0.6800023317337036, + "learning_rate": 0.00019970839271332672, + "loss": 0.1915, + "step": 1047 + }, + { + "epoch": 0.02456677039563518, + "grad_norm": 0.31648173928260803, + "learning_rate": 0.00019970783031074508, + "loss": 0.0915, + "step": 1048 + }, + { + "epoch": 0.024590211970440173, + "grad_norm": 0.5470262765884399, + "learning_rate": 0.00019970726736714592, + "loss": 0.1475, + "step": 1049 + }, + { + "epoch": 0.02461365354524517, + "grad_norm": 0.7768846154212952, + "learning_rate": 0.00019970670388253238, + "loss": 0.2463, + "step": 1050 + }, + { + "epoch": 0.024637095120050165, + "grad_norm": 0.5324990749359131, + "learning_rate": 0.00019970613985690743, + "loss": 0.0674, + "step": 1051 + }, + { + "epoch": 0.02466053669485516, + "grad_norm": 0.5183433294296265, + "learning_rate": 0.00019970557529027419, + "loss": 0.1357, + "step": 1052 + }, + { + "epoch": 0.024683978269660154, + "grad_norm": 0.7040920853614807, + "learning_rate": 0.00019970501018263566, + "loss": 0.1581, + "step": 1053 + }, + { + "epoch": 0.024707419844465152, + "grad_norm": 0.19400742650032043, + "learning_rate": 0.00019970444453399497, + "loss": 0.0421, + "step": 1054 + }, + { + "epoch": 0.024730861419270146, + "grad_norm": 0.6002979874610901, + "learning_rate": 0.00019970387834435514, + "loss": 0.2748, + "step": 1055 + }, + { + "epoch": 0.02475430299407514, + "grad_norm": 0.8603864312171936, + "learning_rate": 0.00019970331161371928, + "loss": 0.3568, + "step": 1056 + }, + { + "epoch": 0.02477774456888014, + "grad_norm": 0.21206480264663696, + "learning_rate": 0.0001997027443420904, + "loss": 0.082, + "step": 1057 + }, + { + "epoch": 0.024801186143685133, + "grad_norm": 0.5642871260643005, + "learning_rate": 0.00019970217652947168, + "loss": 0.1617, + "step": 1058 + }, + { + "epoch": 0.024824627718490127, + "grad_norm": 0.4343988299369812, + "learning_rate": 0.00019970160817586613, + "loss": 0.0906, + "step": 1059 + }, + { + "epoch": 0.024848069293295125, + "grad_norm": 0.1946047693490982, + "learning_rate": 0.00019970103928127687, + "loss": 0.0572, + "step": 1060 + }, + { + "epoch": 0.02487151086810012, + "grad_norm": 0.26811620593070984, + "learning_rate": 0.00019970046984570694, + "loss": 0.0865, + "step": 1061 + }, + { + "epoch": 0.024894952442905113, + "grad_norm": 0.7342778444290161, + "learning_rate": 0.00019969989986915948, + "loss": 0.2078, + "step": 1062 + }, + { + "epoch": 0.02491839401771011, + "grad_norm": 0.7648444771766663, + "learning_rate": 0.00019969932935163753, + "loss": 0.2937, + "step": 1063 + }, + { + "epoch": 0.024941835592515105, + "grad_norm": 0.6230491995811462, + "learning_rate": 0.00019969875829314423, + "loss": 0.2526, + "step": 1064 + }, + { + "epoch": 0.0249652771673201, + "grad_norm": 0.8571129441261292, + "learning_rate": 0.00019969818669368267, + "loss": 0.3134, + "step": 1065 + }, + { + "epoch": 0.024988718742125098, + "grad_norm": 0.41301393508911133, + "learning_rate": 0.00019969761455325595, + "loss": 0.118, + "step": 1066 + }, + { + "epoch": 0.025012160316930092, + "grad_norm": 0.760718584060669, + "learning_rate": 0.00019969704187186715, + "loss": 0.7617, + "step": 1067 + }, + { + "epoch": 0.025035601891735086, + "grad_norm": 0.6238318681716919, + "learning_rate": 0.00019969646864951944, + "loss": 0.2595, + "step": 1068 + }, + { + "epoch": 0.02505904346654008, + "grad_norm": 0.7750898003578186, + "learning_rate": 0.00019969589488621583, + "loss": 0.2157, + "step": 1069 + }, + { + "epoch": 0.025082485041345078, + "grad_norm": 0.1825062781572342, + "learning_rate": 0.00019969532058195953, + "loss": 0.0493, + "step": 1070 + }, + { + "epoch": 0.025105926616150073, + "grad_norm": 0.4364195764064789, + "learning_rate": 0.0001996947457367536, + "loss": 0.1318, + "step": 1071 + }, + { + "epoch": 0.025129368190955067, + "grad_norm": 0.29587799310684204, + "learning_rate": 0.00019969417035060118, + "loss": 0.0748, + "step": 1072 + }, + { + "epoch": 0.025152809765760065, + "grad_norm": 0.4329134523868561, + "learning_rate": 0.0001996935944235054, + "loss": 0.4049, + "step": 1073 + }, + { + "epoch": 0.02517625134056506, + "grad_norm": 0.24999503791332245, + "learning_rate": 0.00019969301795546936, + "loss": 0.0683, + "step": 1074 + }, + { + "epoch": 0.025199692915370053, + "grad_norm": 0.5839853882789612, + "learning_rate": 0.0001996924409464962, + "loss": 0.2183, + "step": 1075 + }, + { + "epoch": 0.02522313449017505, + "grad_norm": 0.5867292284965515, + "learning_rate": 0.00019969186339658906, + "loss": 0.1521, + "step": 1076 + }, + { + "epoch": 0.025246576064980045, + "grad_norm": 0.5998815298080444, + "learning_rate": 0.00019969128530575105, + "loss": 0.2622, + "step": 1077 + }, + { + "epoch": 0.02527001763978504, + "grad_norm": 0.7387522459030151, + "learning_rate": 0.00019969070667398534, + "loss": 0.2962, + "step": 1078 + }, + { + "epoch": 0.025293459214590037, + "grad_norm": 0.7580485939979553, + "learning_rate": 0.00019969012750129507, + "loss": 0.262, + "step": 1079 + }, + { + "epoch": 0.025316900789395032, + "grad_norm": 0.12392041832208633, + "learning_rate": 0.00019968954778768334, + "loss": 0.0307, + "step": 1080 + }, + { + "epoch": 0.025340342364200026, + "grad_norm": 0.33363768458366394, + "learning_rate": 0.00019968896753315332, + "loss": 0.1006, + "step": 1081 + }, + { + "epoch": 0.025363783939005024, + "grad_norm": 0.5312449932098389, + "learning_rate": 0.00019968838673770819, + "loss": 0.1767, + "step": 1082 + }, + { + "epoch": 0.025387225513810018, + "grad_norm": 0.15879371762275696, + "learning_rate": 0.00019968780540135104, + "loss": 0.06, + "step": 1083 + }, + { + "epoch": 0.025410667088615013, + "grad_norm": 0.5332093834877014, + "learning_rate": 0.00019968722352408505, + "loss": 0.1114, + "step": 1084 + }, + { + "epoch": 0.025434108663420007, + "grad_norm": 0.40765124559402466, + "learning_rate": 0.00019968664110591336, + "loss": 0.0694, + "step": 1085 + }, + { + "epoch": 0.025457550238225005, + "grad_norm": 0.09878119826316833, + "learning_rate": 0.0001996860581468392, + "loss": 0.0246, + "step": 1086 + }, + { + "epoch": 0.02548099181303, + "grad_norm": 0.5734668970108032, + "learning_rate": 0.00019968547464686566, + "loss": 0.2063, + "step": 1087 + }, + { + "epoch": 0.025504433387834993, + "grad_norm": 0.6225250959396362, + "learning_rate": 0.00019968489060599596, + "loss": 0.1866, + "step": 1088 + }, + { + "epoch": 0.02552787496263999, + "grad_norm": 0.4748706519603729, + "learning_rate": 0.00019968430602423324, + "loss": 0.5442, + "step": 1089 + }, + { + "epoch": 0.025551316537444985, + "grad_norm": 0.7612010836601257, + "learning_rate": 0.00019968372090158062, + "loss": 0.1794, + "step": 1090 + }, + { + "epoch": 0.02557475811224998, + "grad_norm": 0.7566714882850647, + "learning_rate": 0.00019968313523804137, + "loss": 0.264, + "step": 1091 + }, + { + "epoch": 0.025598199687054977, + "grad_norm": 0.33736711740493774, + "learning_rate": 0.00019968254903361863, + "loss": 0.5861, + "step": 1092 + }, + { + "epoch": 0.025621641261859972, + "grad_norm": 0.5684875249862671, + "learning_rate": 0.00019968196228831557, + "loss": 0.716, + "step": 1093 + }, + { + "epoch": 0.025645082836664966, + "grad_norm": 0.9097557067871094, + "learning_rate": 0.00019968137500213536, + "loss": 0.3552, + "step": 1094 + }, + { + "epoch": 0.025668524411469964, + "grad_norm": 0.8374289870262146, + "learning_rate": 0.00019968078717508123, + "loss": 0.2797, + "step": 1095 + }, + { + "epoch": 0.025691965986274958, + "grad_norm": 0.5399090051651001, + "learning_rate": 0.00019968019880715635, + "loss": 0.63, + "step": 1096 + }, + { + "epoch": 0.025715407561079952, + "grad_norm": 0.4865534007549286, + "learning_rate": 0.00019967960989836388, + "loss": 0.1769, + "step": 1097 + }, + { + "epoch": 0.02573884913588495, + "grad_norm": 0.5646209716796875, + "learning_rate": 0.00019967902044870704, + "loss": 0.2969, + "step": 1098 + }, + { + "epoch": 0.025762290710689945, + "grad_norm": 0.5581073760986328, + "learning_rate": 0.0001996784304581891, + "loss": 0.1978, + "step": 1099 + }, + { + "epoch": 0.02578573228549494, + "grad_norm": 0.5247354507446289, + "learning_rate": 0.0001996778399268131, + "loss": 0.1725, + "step": 1100 + }, + { + "epoch": 0.025809173860299937, + "grad_norm": 0.30659621953964233, + "learning_rate": 0.0001996772488545824, + "loss": 0.1317, + "step": 1101 + }, + { + "epoch": 0.02583261543510493, + "grad_norm": 0.5895461440086365, + "learning_rate": 0.00019967665724150013, + "loss": 0.2239, + "step": 1102 + }, + { + "epoch": 0.025856057009909925, + "grad_norm": 0.7179093956947327, + "learning_rate": 0.0001996760650875695, + "loss": 0.2056, + "step": 1103 + }, + { + "epoch": 0.02587949858471492, + "grad_norm": 0.43903642892837524, + "learning_rate": 0.00019967547239279376, + "loss": 0.1713, + "step": 1104 + }, + { + "epoch": 0.025902940159519917, + "grad_norm": 0.6397926807403564, + "learning_rate": 0.0001996748791571761, + "loss": 0.1659, + "step": 1105 + }, + { + "epoch": 0.02592638173432491, + "grad_norm": 0.1950692981481552, + "learning_rate": 0.00019967428538071973, + "loss": 0.06, + "step": 1106 + }, + { + "epoch": 0.025949823309129906, + "grad_norm": 0.9632493853569031, + "learning_rate": 0.0001996736910634279, + "loss": 0.2439, + "step": 1107 + }, + { + "epoch": 0.025973264883934904, + "grad_norm": 0.2680678963661194, + "learning_rate": 0.0001996730962053038, + "loss": 0.0823, + "step": 1108 + }, + { + "epoch": 0.025996706458739898, + "grad_norm": 0.1447686105966568, + "learning_rate": 0.0001996725008063507, + "loss": 0.0494, + "step": 1109 + }, + { + "epoch": 0.026020148033544892, + "grad_norm": 0.2815142869949341, + "learning_rate": 0.0001996719048665718, + "loss": 0.0833, + "step": 1110 + }, + { + "epoch": 0.02604358960834989, + "grad_norm": 0.25236809253692627, + "learning_rate": 0.00019967130838597036, + "loss": 0.0591, + "step": 1111 + }, + { + "epoch": 0.026067031183154885, + "grad_norm": 0.6185254454612732, + "learning_rate": 0.00019967071136454956, + "loss": 0.1901, + "step": 1112 + }, + { + "epoch": 0.02609047275795988, + "grad_norm": 0.6582789421081543, + "learning_rate": 0.0001996701138023127, + "loss": 0.2049, + "step": 1113 + }, + { + "epoch": 0.026113914332764877, + "grad_norm": 0.3048769533634186, + "learning_rate": 0.000199669515699263, + "loss": 0.0766, + "step": 1114 + }, + { + "epoch": 0.02613735590756987, + "grad_norm": 0.5987125039100647, + "learning_rate": 0.0001996689170554037, + "loss": 0.1619, + "step": 1115 + }, + { + "epoch": 0.026160797482374865, + "grad_norm": 0.4608689248561859, + "learning_rate": 0.00019966831787073805, + "loss": 0.1678, + "step": 1116 + }, + { + "epoch": 0.026184239057179863, + "grad_norm": 0.4778953194618225, + "learning_rate": 0.0001996677181452693, + "loss": 0.108, + "step": 1117 + }, + { + "epoch": 0.026207680631984857, + "grad_norm": 0.6389352083206177, + "learning_rate": 0.00019966711787900072, + "loss": 0.2429, + "step": 1118 + }, + { + "epoch": 0.02623112220678985, + "grad_norm": 0.6152185201644897, + "learning_rate": 0.00019966651707193554, + "loss": 0.1768, + "step": 1119 + }, + { + "epoch": 0.026254563781594846, + "grad_norm": 1.0019820928573608, + "learning_rate": 0.00019966591572407707, + "loss": 0.2965, + "step": 1120 + }, + { + "epoch": 0.026278005356399844, + "grad_norm": 0.6235135793685913, + "learning_rate": 0.00019966531383542853, + "loss": 0.8929, + "step": 1121 + }, + { + "epoch": 0.026301446931204838, + "grad_norm": 0.571590781211853, + "learning_rate": 0.00019966471140599317, + "loss": 0.7203, + "step": 1122 + }, + { + "epoch": 0.026324888506009832, + "grad_norm": 0.6303793787956238, + "learning_rate": 0.0001996641084357743, + "loss": 0.8813, + "step": 1123 + }, + { + "epoch": 0.02634833008081483, + "grad_norm": 0.7525303363800049, + "learning_rate": 0.00019966350492477516, + "loss": 0.4099, + "step": 1124 + }, + { + "epoch": 0.026371771655619825, + "grad_norm": 1.1141490936279297, + "learning_rate": 0.00019966290087299906, + "loss": 0.2338, + "step": 1125 + }, + { + "epoch": 0.02639521323042482, + "grad_norm": 0.6719129085540771, + "learning_rate": 0.00019966229628044926, + "loss": 0.1782, + "step": 1126 + }, + { + "epoch": 0.026418654805229817, + "grad_norm": 0.8386567831039429, + "learning_rate": 0.00019966169114712901, + "loss": 0.2806, + "step": 1127 + }, + { + "epoch": 0.02644209638003481, + "grad_norm": 0.49595221877098083, + "learning_rate": 0.00019966108547304166, + "loss": 0.1624, + "step": 1128 + }, + { + "epoch": 0.026465537954839805, + "grad_norm": 0.5514326691627502, + "learning_rate": 0.00019966047925819042, + "loss": 0.2424, + "step": 1129 + }, + { + "epoch": 0.026488979529644803, + "grad_norm": 0.5910806655883789, + "learning_rate": 0.00019965987250257866, + "loss": 0.8386, + "step": 1130 + }, + { + "epoch": 0.026512421104449797, + "grad_norm": 0.7105093002319336, + "learning_rate": 0.00019965926520620962, + "loss": 0.1007, + "step": 1131 + }, + { + "epoch": 0.02653586267925479, + "grad_norm": 0.2535845935344696, + "learning_rate": 0.00019965865736908657, + "loss": 0.082, + "step": 1132 + }, + { + "epoch": 0.02655930425405979, + "grad_norm": 0.4647004008293152, + "learning_rate": 0.00019965804899121285, + "loss": 0.1905, + "step": 1133 + }, + { + "epoch": 0.026582745828864784, + "grad_norm": 0.7136042714118958, + "learning_rate": 0.00019965744007259176, + "loss": 0.2853, + "step": 1134 + }, + { + "epoch": 0.026606187403669778, + "grad_norm": 0.22646699845790863, + "learning_rate": 0.00019965683061322662, + "loss": 0.0371, + "step": 1135 + }, + { + "epoch": 0.026629628978474772, + "grad_norm": 0.4119760990142822, + "learning_rate": 0.0001996562206131207, + "loss": 0.2316, + "step": 1136 + }, + { + "epoch": 0.02665307055327977, + "grad_norm": 0.8824479579925537, + "learning_rate": 0.00019965561007227734, + "loss": 0.2691, + "step": 1137 + }, + { + "epoch": 0.026676512128084764, + "grad_norm": 0.3271239399909973, + "learning_rate": 0.00019965499899069982, + "loss": 0.1188, + "step": 1138 + }, + { + "epoch": 0.02669995370288976, + "grad_norm": 0.29687076807022095, + "learning_rate": 0.00019965438736839147, + "loss": 0.0837, + "step": 1139 + }, + { + "epoch": 0.026723395277694757, + "grad_norm": 0.5426803827285767, + "learning_rate": 0.00019965377520535563, + "loss": 0.1761, + "step": 1140 + }, + { + "epoch": 0.02674683685249975, + "grad_norm": 0.5315082669258118, + "learning_rate": 0.0001996531625015956, + "loss": 0.8283, + "step": 1141 + }, + { + "epoch": 0.026770278427304745, + "grad_norm": 0.5164610743522644, + "learning_rate": 0.00019965254925711468, + "loss": 0.9326, + "step": 1142 + }, + { + "epoch": 0.026793720002109743, + "grad_norm": 0.5724196434020996, + "learning_rate": 0.00019965193547191626, + "loss": 0.164, + "step": 1143 + }, + { + "epoch": 0.026817161576914737, + "grad_norm": 0.6802283525466919, + "learning_rate": 0.0001996513211460036, + "loss": 0.2161, + "step": 1144 + }, + { + "epoch": 0.02684060315171973, + "grad_norm": 0.22340132296085358, + "learning_rate": 0.00019965070627938012, + "loss": 0.0649, + "step": 1145 + }, + { + "epoch": 0.02686404472652473, + "grad_norm": 0.4663689136505127, + "learning_rate": 0.00019965009087204906, + "loss": 0.1267, + "step": 1146 + }, + { + "epoch": 0.026887486301329724, + "grad_norm": 1.2326210737228394, + "learning_rate": 0.0001996494749240138, + "loss": 0.2506, + "step": 1147 + }, + { + "epoch": 0.026910927876134718, + "grad_norm": 0.5408186316490173, + "learning_rate": 0.00019964885843527772, + "loss": 0.1755, + "step": 1148 + }, + { + "epoch": 0.026934369450939716, + "grad_norm": 0.20826566219329834, + "learning_rate": 0.00019964824140584412, + "loss": 0.0647, + "step": 1149 + }, + { + "epoch": 0.02695781102574471, + "grad_norm": 0.43343210220336914, + "learning_rate": 0.00019964762383571638, + "loss": 0.1224, + "step": 1150 + }, + { + "epoch": 0.026981252600549704, + "grad_norm": 1.0555871725082397, + "learning_rate": 0.0001996470057248978, + "loss": 0.219, + "step": 1151 + }, + { + "epoch": 0.0270046941753547, + "grad_norm": 0.1997964084148407, + "learning_rate": 0.00019964638707339178, + "loss": 0.0581, + "step": 1152 + }, + { + "epoch": 0.027028135750159697, + "grad_norm": 0.6265400648117065, + "learning_rate": 0.0001996457678812017, + "loss": 0.1985, + "step": 1153 + }, + { + "epoch": 0.02705157732496469, + "grad_norm": 0.16196909546852112, + "learning_rate": 0.00019964514814833082, + "loss": 0.0418, + "step": 1154 + }, + { + "epoch": 0.027075018899769685, + "grad_norm": 0.7015427947044373, + "learning_rate": 0.00019964452787478258, + "loss": 0.2001, + "step": 1155 + }, + { + "epoch": 0.027098460474574683, + "grad_norm": 0.35411080718040466, + "learning_rate": 0.00019964390706056034, + "loss": 0.0892, + "step": 1156 + }, + { + "epoch": 0.027121902049379677, + "grad_norm": 0.5563951134681702, + "learning_rate": 0.00019964328570566746, + "loss": 0.1745, + "step": 1157 + }, + { + "epoch": 0.02714534362418467, + "grad_norm": 0.1392686516046524, + "learning_rate": 0.0001996426638101073, + "loss": 0.0537, + "step": 1158 + }, + { + "epoch": 0.02716878519898967, + "grad_norm": 0.4488295316696167, + "learning_rate": 0.00019964204137388325, + "loss": 0.652, + "step": 1159 + }, + { + "epoch": 0.027192226773794664, + "grad_norm": 0.4817724823951721, + "learning_rate": 0.00019964141839699868, + "loss": 0.1273, + "step": 1160 + }, + { + "epoch": 0.027215668348599658, + "grad_norm": 0.6721146106719971, + "learning_rate": 0.00019964079487945697, + "loss": 0.1883, + "step": 1161 + }, + { + "epoch": 0.027239109923404656, + "grad_norm": 0.45874667167663574, + "learning_rate": 0.00019964017082126152, + "loss": 0.2561, + "step": 1162 + }, + { + "epoch": 0.02726255149820965, + "grad_norm": 0.49036905169487, + "learning_rate": 0.0001996395462224157, + "loss": 0.0979, + "step": 1163 + }, + { + "epoch": 0.027285993073014644, + "grad_norm": 0.5567464232444763, + "learning_rate": 0.00019963892108292288, + "loss": 0.2112, + "step": 1164 + }, + { + "epoch": 0.027309434647819642, + "grad_norm": 0.39623817801475525, + "learning_rate": 0.0001996382954027865, + "loss": 0.122, + "step": 1165 + }, + { + "epoch": 0.027332876222624636, + "grad_norm": 0.6837006211280823, + "learning_rate": 0.00019963766918200987, + "loss": 0.2174, + "step": 1166 + }, + { + "epoch": 0.02735631779742963, + "grad_norm": 0.8154470324516296, + "learning_rate": 0.0001996370424205965, + "loss": 0.2975, + "step": 1167 + }, + { + "epoch": 0.027379759372234625, + "grad_norm": 0.18226215243339539, + "learning_rate": 0.00019963641511854975, + "loss": 0.0463, + "step": 1168 + }, + { + "epoch": 0.027403200947039623, + "grad_norm": 0.3827337324619293, + "learning_rate": 0.00019963578727587295, + "loss": 0.1366, + "step": 1169 + }, + { + "epoch": 0.027426642521844617, + "grad_norm": 0.693473219871521, + "learning_rate": 0.0001996351588925696, + "loss": 0.2756, + "step": 1170 + }, + { + "epoch": 0.02745008409664961, + "grad_norm": 0.7621276378631592, + "learning_rate": 0.0001996345299686431, + "loss": 0.221, + "step": 1171 + }, + { + "epoch": 0.02747352567145461, + "grad_norm": 0.739479660987854, + "learning_rate": 0.0001996339005040968, + "loss": 0.2402, + "step": 1172 + }, + { + "epoch": 0.027496967246259604, + "grad_norm": 0.22739636898040771, + "learning_rate": 0.00019963327049893418, + "loss": 0.0682, + "step": 1173 + }, + { + "epoch": 0.027520408821064598, + "grad_norm": 0.8709672689437866, + "learning_rate": 0.00019963263995315862, + "loss": 0.3022, + "step": 1174 + }, + { + "epoch": 0.027543850395869596, + "grad_norm": 0.73809415102005, + "learning_rate": 0.00019963200886677357, + "loss": 0.2955, + "step": 1175 + }, + { + "epoch": 0.02756729197067459, + "grad_norm": 0.464041531085968, + "learning_rate": 0.00019963137723978244, + "loss": 0.163, + "step": 1176 + }, + { + "epoch": 0.027590733545479584, + "grad_norm": 0.5818406939506531, + "learning_rate": 0.0001996307450721886, + "loss": 0.7194, + "step": 1177 + }, + { + "epoch": 0.027614175120284582, + "grad_norm": 0.6443785429000854, + "learning_rate": 0.0001996301123639956, + "loss": 0.2376, + "step": 1178 + }, + { + "epoch": 0.027637616695089576, + "grad_norm": 0.36906594038009644, + "learning_rate": 0.0001996294791152068, + "loss": 0.0837, + "step": 1179 + }, + { + "epoch": 0.02766105826989457, + "grad_norm": 0.5143387913703918, + "learning_rate": 0.0001996288453258256, + "loss": 0.9518, + "step": 1180 + }, + { + "epoch": 0.02768449984469957, + "grad_norm": 0.15207533538341522, + "learning_rate": 0.0001996282109958555, + "loss": 0.0345, + "step": 1181 + }, + { + "epoch": 0.027707941419504563, + "grad_norm": 0.538331151008606, + "learning_rate": 0.00019962757612529996, + "loss": 0.1673, + "step": 1182 + }, + { + "epoch": 0.027731382994309557, + "grad_norm": 0.4748792350292206, + "learning_rate": 0.00019962694071416236, + "loss": 0.1689, + "step": 1183 + }, + { + "epoch": 0.02775482456911455, + "grad_norm": 0.5940277576446533, + "learning_rate": 0.00019962630476244618, + "loss": 0.2174, + "step": 1184 + }, + { + "epoch": 0.02777826614391955, + "grad_norm": 0.6126834750175476, + "learning_rate": 0.00019962566827015488, + "loss": 0.164, + "step": 1185 + }, + { + "epoch": 0.027801707718724544, + "grad_norm": 0.1922556608915329, + "learning_rate": 0.0001996250312372919, + "loss": 0.061, + "step": 1186 + }, + { + "epoch": 0.027825149293529538, + "grad_norm": 0.4600471556186676, + "learning_rate": 0.00019962439366386073, + "loss": 0.1464, + "step": 1187 + }, + { + "epoch": 0.027848590868334536, + "grad_norm": 0.6949896216392517, + "learning_rate": 0.00019962375554986473, + "loss": 0.3857, + "step": 1188 + }, + { + "epoch": 0.02787203244313953, + "grad_norm": 0.541475236415863, + "learning_rate": 0.00019962311689530752, + "loss": 0.1467, + "step": 1189 + }, + { + "epoch": 0.027895474017944524, + "grad_norm": 0.7371475696563721, + "learning_rate": 0.0001996224777001924, + "loss": 0.1794, + "step": 1190 + }, + { + "epoch": 0.027918915592749522, + "grad_norm": 0.19657616317272186, + "learning_rate": 0.00019962183796452294, + "loss": 0.0475, + "step": 1191 + }, + { + "epoch": 0.027942357167554516, + "grad_norm": 0.19369672238826752, + "learning_rate": 0.00019962119768830258, + "loss": 0.053, + "step": 1192 + }, + { + "epoch": 0.02796579874235951, + "grad_norm": 0.4549216032028198, + "learning_rate": 0.00019962055687153482, + "loss": 0.1253, + "step": 1193 + }, + { + "epoch": 0.02798924031716451, + "grad_norm": 1.0522340536117554, + "learning_rate": 0.0001996199155142231, + "loss": 0.3467, + "step": 1194 + }, + { + "epoch": 0.028012681891969503, + "grad_norm": 0.7234682440757751, + "learning_rate": 0.00019961927361637094, + "loss": 0.234, + "step": 1195 + }, + { + "epoch": 0.028036123466774497, + "grad_norm": 0.9738910794258118, + "learning_rate": 0.00019961863117798177, + "loss": 0.2864, + "step": 1196 + }, + { + "epoch": 0.028059565041579495, + "grad_norm": 0.6285727024078369, + "learning_rate": 0.00019961798819905912, + "loss": 0.2205, + "step": 1197 + }, + { + "epoch": 0.02808300661638449, + "grad_norm": 0.6200679540634155, + "learning_rate": 0.00019961734467960647, + "loss": 0.2239, + "step": 1198 + }, + { + "epoch": 0.028106448191189484, + "grad_norm": 0.5757086277008057, + "learning_rate": 0.00019961670061962732, + "loss": 0.2, + "step": 1199 + }, + { + "epoch": 0.028129889765994478, + "grad_norm": 0.22837944328784943, + "learning_rate": 0.0001996160560191251, + "loss": 0.0714, + "step": 1200 + }, + { + "epoch": 0.028153331340799476, + "grad_norm": 0.4323669373989105, + "learning_rate": 0.0001996154108781034, + "loss": 0.1063, + "step": 1201 + }, + { + "epoch": 0.02817677291560447, + "grad_norm": 0.47299060225486755, + "learning_rate": 0.00019961476519656567, + "loss": 0.1622, + "step": 1202 + }, + { + "epoch": 0.028200214490409464, + "grad_norm": 0.820660412311554, + "learning_rate": 0.00019961411897451544, + "loss": 0.1898, + "step": 1203 + }, + { + "epoch": 0.028223656065214462, + "grad_norm": 0.4817592203617096, + "learning_rate": 0.00019961347221195616, + "loss": 0.1436, + "step": 1204 + }, + { + "epoch": 0.028247097640019456, + "grad_norm": 0.2987964451313019, + "learning_rate": 0.0001996128249088914, + "loss": 0.0909, + "step": 1205 + }, + { + "epoch": 0.02827053921482445, + "grad_norm": 0.4126628041267395, + "learning_rate": 0.00019961217706532467, + "loss": 0.1815, + "step": 1206 + }, + { + "epoch": 0.02829398078962945, + "grad_norm": 0.6709031462669373, + "learning_rate": 0.00019961152868125945, + "loss": 0.1379, + "step": 1207 + }, + { + "epoch": 0.028317422364434443, + "grad_norm": 0.5807145833969116, + "learning_rate": 0.0001996108797566993, + "loss": 0.1724, + "step": 1208 + }, + { + "epoch": 0.028340863939239437, + "grad_norm": 0.6477484703063965, + "learning_rate": 0.00019961023029164767, + "loss": 0.2286, + "step": 1209 + }, + { + "epoch": 0.028364305514044435, + "grad_norm": 0.5267817974090576, + "learning_rate": 0.00019960958028610818, + "loss": 0.1438, + "step": 1210 + }, + { + "epoch": 0.02838774708884943, + "grad_norm": 0.6550439596176147, + "learning_rate": 0.00019960892974008425, + "loss": 0.8812, + "step": 1211 + }, + { + "epoch": 0.028411188663654423, + "grad_norm": 0.6972312927246094, + "learning_rate": 0.0001996082786535795, + "loss": 0.7748, + "step": 1212 + }, + { + "epoch": 0.02843463023845942, + "grad_norm": 0.5339691042900085, + "learning_rate": 0.00019960762702659742, + "loss": 0.1762, + "step": 1213 + }, + { + "epoch": 0.028458071813264416, + "grad_norm": 0.5307449102401733, + "learning_rate": 0.00019960697485914156, + "loss": 0.5175, + "step": 1214 + }, + { + "epoch": 0.02848151338806941, + "grad_norm": 0.6192423701286316, + "learning_rate": 0.00019960632215121544, + "loss": 0.1623, + "step": 1215 + }, + { + "epoch": 0.028504954962874404, + "grad_norm": 0.6464714407920837, + "learning_rate": 0.00019960566890282264, + "loss": 0.7831, + "step": 1216 + }, + { + "epoch": 0.028528396537679402, + "grad_norm": 0.682843804359436, + "learning_rate": 0.00019960501511396666, + "loss": 0.1524, + "step": 1217 + }, + { + "epoch": 0.028551838112484396, + "grad_norm": 0.2247849702835083, + "learning_rate": 0.00019960436078465108, + "loss": 0.1063, + "step": 1218 + }, + { + "epoch": 0.02857527968728939, + "grad_norm": 0.4936310648918152, + "learning_rate": 0.00019960370591487941, + "loss": 0.8186, + "step": 1219 + }, + { + "epoch": 0.02859872126209439, + "grad_norm": 0.4363958239555359, + "learning_rate": 0.00019960305050465527, + "loss": 0.1777, + "step": 1220 + }, + { + "epoch": 0.028622162836899383, + "grad_norm": 0.25307610630989075, + "learning_rate": 0.00019960239455398216, + "loss": 0.0534, + "step": 1221 + }, + { + "epoch": 0.028645604411704377, + "grad_norm": 0.40488383173942566, + "learning_rate": 0.00019960173806286364, + "loss": 0.12, + "step": 1222 + }, + { + "epoch": 0.028669045986509375, + "grad_norm": 0.5167030096054077, + "learning_rate": 0.00019960108103130332, + "loss": 0.1876, + "step": 1223 + }, + { + "epoch": 0.02869248756131437, + "grad_norm": 0.4426465332508087, + "learning_rate": 0.00019960042345930473, + "loss": 0.0615, + "step": 1224 + }, + { + "epoch": 0.028715929136119363, + "grad_norm": 0.6080380082130432, + "learning_rate": 0.00019959976534687145, + "loss": 0.1508, + "step": 1225 + }, + { + "epoch": 0.02873937071092436, + "grad_norm": 0.6739064455032349, + "learning_rate": 0.00019959910669400702, + "loss": 0.1436, + "step": 1226 + }, + { + "epoch": 0.028762812285729356, + "grad_norm": 0.4195094108581543, + "learning_rate": 0.000199598447500715, + "loss": 0.1206, + "step": 1227 + }, + { + "epoch": 0.02878625386053435, + "grad_norm": 0.7400779128074646, + "learning_rate": 0.00019959778776699905, + "loss": 0.1706, + "step": 1228 + }, + { + "epoch": 0.028809695435339348, + "grad_norm": 0.7760661840438843, + "learning_rate": 0.0001995971274928627, + "loss": 0.2204, + "step": 1229 + }, + { + "epoch": 0.028833137010144342, + "grad_norm": 0.7142547965049744, + "learning_rate": 0.00019959646667830954, + "loss": 0.1271, + "step": 1230 + }, + { + "epoch": 0.028856578584949336, + "grad_norm": 0.17523755133152008, + "learning_rate": 0.00019959580532334314, + "loss": 0.036, + "step": 1231 + }, + { + "epoch": 0.02888002015975433, + "grad_norm": 0.8468749523162842, + "learning_rate": 0.00019959514342796712, + "loss": 0.1273, + "step": 1232 + }, + { + "epoch": 0.02890346173455933, + "grad_norm": 0.31802281737327576, + "learning_rate": 0.00019959448099218503, + "loss": 0.0551, + "step": 1233 + }, + { + "epoch": 0.028926903309364323, + "grad_norm": 0.49737030267715454, + "learning_rate": 0.0001995938180160005, + "loss": 0.1424, + "step": 1234 + }, + { + "epoch": 0.028950344884169317, + "grad_norm": 0.3948785662651062, + "learning_rate": 0.00019959315449941708, + "loss": 0.0961, + "step": 1235 + }, + { + "epoch": 0.028973786458974315, + "grad_norm": 0.8143609762191772, + "learning_rate": 0.00019959249044243843, + "loss": 0.2018, + "step": 1236 + }, + { + "epoch": 0.02899722803377931, + "grad_norm": 0.5830464363098145, + "learning_rate": 0.00019959182584506816, + "loss": 0.8536, + "step": 1237 + }, + { + "epoch": 0.029020669608584303, + "grad_norm": 0.334064781665802, + "learning_rate": 0.00019959116070730976, + "loss": 0.1106, + "step": 1238 + }, + { + "epoch": 0.0290441111833893, + "grad_norm": 0.488286554813385, + "learning_rate": 0.00019959049502916697, + "loss": 0.1555, + "step": 1239 + }, + { + "epoch": 0.029067552758194296, + "grad_norm": 0.8234584331512451, + "learning_rate": 0.0001995898288106434, + "loss": 0.2519, + "step": 1240 + }, + { + "epoch": 0.02909099433299929, + "grad_norm": 0.5184747576713562, + "learning_rate": 0.00019958916205174254, + "loss": 0.1617, + "step": 1241 + }, + { + "epoch": 0.029114435907804288, + "grad_norm": 0.2879592180252075, + "learning_rate": 0.00019958849475246813, + "loss": 0.4278, + "step": 1242 + }, + { + "epoch": 0.029137877482609282, + "grad_norm": 0.7179453372955322, + "learning_rate": 0.00019958782691282374, + "loss": 0.7281, + "step": 1243 + }, + { + "epoch": 0.029161319057414276, + "grad_norm": 0.4979240596294403, + "learning_rate": 0.00019958715853281297, + "loss": 0.5244, + "step": 1244 + }, + { + "epoch": 0.029184760632219274, + "grad_norm": 0.5438482165336609, + "learning_rate": 0.0001995864896124395, + "loss": 0.1439, + "step": 1245 + }, + { + "epoch": 0.02920820220702427, + "grad_norm": 0.48561322689056396, + "learning_rate": 0.00019958582015170695, + "loss": 0.1372, + "step": 1246 + }, + { + "epoch": 0.029231643781829263, + "grad_norm": 0.6108757257461548, + "learning_rate": 0.00019958515015061894, + "loss": 0.2013, + "step": 1247 + }, + { + "epoch": 0.029255085356634257, + "grad_norm": 0.517255961894989, + "learning_rate": 0.00019958447960917908, + "loss": 0.1569, + "step": 1248 + }, + { + "epoch": 0.029278526931439255, + "grad_norm": 0.8751998543739319, + "learning_rate": 0.00019958380852739106, + "loss": 0.2792, + "step": 1249 + }, + { + "epoch": 0.02930196850624425, + "grad_norm": 0.6474177837371826, + "learning_rate": 0.00019958313690525848, + "loss": 0.1237, + "step": 1250 + }, + { + "epoch": 0.029325410081049243, + "grad_norm": 0.21159625053405762, + "learning_rate": 0.00019958246474278501, + "loss": 0.0605, + "step": 1251 + }, + { + "epoch": 0.02934885165585424, + "grad_norm": 0.5283594131469727, + "learning_rate": 0.0001995817920399743, + "loss": 0.5716, + "step": 1252 + }, + { + "epoch": 0.029372293230659235, + "grad_norm": 0.6649013161659241, + "learning_rate": 0.00019958111879682996, + "loss": 0.187, + "step": 1253 + }, + { + "epoch": 0.02939573480546423, + "grad_norm": 0.5090125203132629, + "learning_rate": 0.00019958044501335566, + "loss": 0.1434, + "step": 1254 + }, + { + "epoch": 0.029419176380269228, + "grad_norm": 0.5175171494483948, + "learning_rate": 0.00019957977068955506, + "loss": 0.3012, + "step": 1255 + }, + { + "epoch": 0.029442617955074222, + "grad_norm": 0.1934339702129364, + "learning_rate": 0.00019957909582543187, + "loss": 0.0417, + "step": 1256 + }, + { + "epoch": 0.029466059529879216, + "grad_norm": 0.7397926449775696, + "learning_rate": 0.00019957842042098967, + "loss": 0.1664, + "step": 1257 + }, + { + "epoch": 0.029489501104684214, + "grad_norm": 0.8641571998596191, + "learning_rate": 0.00019957774447623217, + "loss": 0.2392, + "step": 1258 + }, + { + "epoch": 0.02951294267948921, + "grad_norm": 0.49381229281425476, + "learning_rate": 0.00019957706799116304, + "loss": 0.128, + "step": 1259 + }, + { + "epoch": 0.029536384254294203, + "grad_norm": 0.19269219040870667, + "learning_rate": 0.00019957639096578592, + "loss": 0.0632, + "step": 1260 + }, + { + "epoch": 0.0295598258290992, + "grad_norm": 0.4609513282775879, + "learning_rate": 0.00019957571340010452, + "loss": 0.1947, + "step": 1261 + }, + { + "epoch": 0.029583267403904195, + "grad_norm": 0.3575972616672516, + "learning_rate": 0.00019957503529412248, + "loss": 0.1401, + "step": 1262 + }, + { + "epoch": 0.02960670897870919, + "grad_norm": 1.099858283996582, + "learning_rate": 0.00019957435664784355, + "loss": 0.3258, + "step": 1263 + }, + { + "epoch": 0.029630150553514187, + "grad_norm": 0.4878140985965729, + "learning_rate": 0.0001995736774612713, + "loss": 0.1728, + "step": 1264 + }, + { + "epoch": 0.02965359212831918, + "grad_norm": 0.9486998319625854, + "learning_rate": 0.0001995729977344095, + "loss": 0.3255, + "step": 1265 + }, + { + "epoch": 0.029677033703124175, + "grad_norm": 0.7537282109260559, + "learning_rate": 0.0001995723174672618, + "loss": 0.146, + "step": 1266 + }, + { + "epoch": 0.02970047527792917, + "grad_norm": 0.5823652744293213, + "learning_rate": 0.00019957163665983193, + "loss": 0.2056, + "step": 1267 + }, + { + "epoch": 0.029723916852734168, + "grad_norm": 0.6414916515350342, + "learning_rate": 0.00019957095531212356, + "loss": 0.2047, + "step": 1268 + }, + { + "epoch": 0.029747358427539162, + "grad_norm": 0.35955068469047546, + "learning_rate": 0.00019957027342414036, + "loss": 0.4937, + "step": 1269 + }, + { + "epoch": 0.029770800002344156, + "grad_norm": 0.18163366615772247, + "learning_rate": 0.00019956959099588607, + "loss": 0.0783, + "step": 1270 + }, + { + "epoch": 0.029794241577149154, + "grad_norm": 0.3634130656719208, + "learning_rate": 0.00019956890802736438, + "loss": 0.1101, + "step": 1271 + }, + { + "epoch": 0.029817683151954148, + "grad_norm": 0.5205760598182678, + "learning_rate": 0.000199568224518579, + "loss": 0.1788, + "step": 1272 + }, + { + "epoch": 0.029841124726759143, + "grad_norm": 0.48325830698013306, + "learning_rate": 0.00019956754046953362, + "loss": 0.6197, + "step": 1273 + }, + { + "epoch": 0.02986456630156414, + "grad_norm": 0.1597793996334076, + "learning_rate": 0.00019956685588023198, + "loss": 0.0647, + "step": 1274 + }, + { + "epoch": 0.029888007876369135, + "grad_norm": 0.6650166511535645, + "learning_rate": 0.0001995661707506778, + "loss": 0.4733, + "step": 1275 + }, + { + "epoch": 0.02991144945117413, + "grad_norm": 0.6105600595474243, + "learning_rate": 0.00019956548508087476, + "loss": 0.165, + "step": 1276 + }, + { + "epoch": 0.029934891025979127, + "grad_norm": 0.7650709748268127, + "learning_rate": 0.00019956479887082656, + "loss": 0.2443, + "step": 1277 + }, + { + "epoch": 0.02995833260078412, + "grad_norm": 0.40413057804107666, + "learning_rate": 0.00019956411212053703, + "loss": 0.165, + "step": 1278 + }, + { + "epoch": 0.029981774175589115, + "grad_norm": 1.131652593612671, + "learning_rate": 0.00019956342483000978, + "loss": 0.2546, + "step": 1279 + }, + { + "epoch": 0.030005215750394113, + "grad_norm": 0.15005774796009064, + "learning_rate": 0.0001995627369992486, + "loss": 0.0536, + "step": 1280 + }, + { + "epoch": 0.030028657325199107, + "grad_norm": 0.5942365527153015, + "learning_rate": 0.0001995620486282572, + "loss": 0.1784, + "step": 1281 + }, + { + "epoch": 0.030052098900004102, + "grad_norm": 0.4907914102077484, + "learning_rate": 0.00019956135971703934, + "loss": 0.893, + "step": 1282 + }, + { + "epoch": 0.030075540474809096, + "grad_norm": 0.5523700714111328, + "learning_rate": 0.0001995606702655987, + "loss": 0.2096, + "step": 1283 + }, + { + "epoch": 0.030098982049614094, + "grad_norm": 0.1255091279745102, + "learning_rate": 0.00019955998027393915, + "loss": 0.0528, + "step": 1284 + }, + { + "epoch": 0.030122423624419088, + "grad_norm": 0.52999347448349, + "learning_rate": 0.0001995592897420643, + "loss": 0.1227, + "step": 1285 + }, + { + "epoch": 0.030145865199224083, + "grad_norm": 0.3798665404319763, + "learning_rate": 0.00019955859866997795, + "loss": 0.4505, + "step": 1286 + }, + { + "epoch": 0.03016930677402908, + "grad_norm": 0.5448628067970276, + "learning_rate": 0.00019955790705768382, + "loss": 0.1385, + "step": 1287 + }, + { + "epoch": 0.030192748348834075, + "grad_norm": 0.7611163258552551, + "learning_rate": 0.00019955721490518572, + "loss": 0.2529, + "step": 1288 + }, + { + "epoch": 0.03021618992363907, + "grad_norm": 0.6013275980949402, + "learning_rate": 0.00019955652221248735, + "loss": 0.2737, + "step": 1289 + }, + { + "epoch": 0.030239631498444067, + "grad_norm": 0.3704405128955841, + "learning_rate": 0.00019955582897959251, + "loss": 0.128, + "step": 1290 + }, + { + "epoch": 0.03026307307324906, + "grad_norm": 0.4946976602077484, + "learning_rate": 0.0001995551352065049, + "loss": 0.6119, + "step": 1291 + }, + { + "epoch": 0.030286514648054055, + "grad_norm": 0.5734766721725464, + "learning_rate": 0.0001995544408932284, + "loss": 0.1888, + "step": 1292 + }, + { + "epoch": 0.030309956222859053, + "grad_norm": 0.5023686289787292, + "learning_rate": 0.00019955374603976663, + "loss": 0.1468, + "step": 1293 + }, + { + "epoch": 0.030333397797664047, + "grad_norm": 0.5786041021347046, + "learning_rate": 0.0001995530506461235, + "loss": 0.1403, + "step": 1294 + }, + { + "epoch": 0.030356839372469042, + "grad_norm": 1.0352083444595337, + "learning_rate": 0.00019955235471230266, + "loss": 0.2647, + "step": 1295 + }, + { + "epoch": 0.03038028094727404, + "grad_norm": 0.4946807622909546, + "learning_rate": 0.00019955165823830798, + "loss": 0.178, + "step": 1296 + }, + { + "epoch": 0.030403722522079034, + "grad_norm": 0.6379843354225159, + "learning_rate": 0.0001995509612241432, + "loss": 0.2379, + "step": 1297 + }, + { + "epoch": 0.030427164096884028, + "grad_norm": 0.46945175528526306, + "learning_rate": 0.00019955026366981208, + "loss": 0.1254, + "step": 1298 + }, + { + "epoch": 0.030450605671689022, + "grad_norm": 0.640883207321167, + "learning_rate": 0.00019954956557531844, + "loss": 0.2403, + "step": 1299 + }, + { + "epoch": 0.03047404724649402, + "grad_norm": 0.5586436986923218, + "learning_rate": 0.0001995488669406661, + "loss": 0.1782, + "step": 1300 + }, + { + "epoch": 0.030497488821299015, + "grad_norm": 0.44823551177978516, + "learning_rate": 0.00019954816776585876, + "loss": 0.1656, + "step": 1301 + }, + { + "epoch": 0.03052093039610401, + "grad_norm": 0.5214378833770752, + "learning_rate": 0.0001995474680509003, + "loss": 0.1996, + "step": 1302 + }, + { + "epoch": 0.030544371970909007, + "grad_norm": 0.9339608550071716, + "learning_rate": 0.0001995467677957944, + "loss": 0.2459, + "step": 1303 + }, + { + "epoch": 0.030567813545714, + "grad_norm": 0.5359817743301392, + "learning_rate": 0.00019954606700054502, + "loss": 0.1276, + "step": 1304 + }, + { + "epoch": 0.030591255120518995, + "grad_norm": 0.48439162969589233, + "learning_rate": 0.00019954536566515587, + "loss": 0.1368, + "step": 1305 + }, + { + "epoch": 0.030614696695323993, + "grad_norm": 0.7634085416793823, + "learning_rate": 0.00019954466378963078, + "loss": 0.1381, + "step": 1306 + }, + { + "epoch": 0.030638138270128987, + "grad_norm": 0.4741237163543701, + "learning_rate": 0.00019954396137397348, + "loss": 0.2214, + "step": 1307 + }, + { + "epoch": 0.03066157984493398, + "grad_norm": 0.7365843653678894, + "learning_rate": 0.00019954325841818791, + "loss": 0.1866, + "step": 1308 + }, + { + "epoch": 0.03068502141973898, + "grad_norm": 0.6131802797317505, + "learning_rate": 0.00019954255492227781, + "loss": 0.225, + "step": 1309 + }, + { + "epoch": 0.030708462994543974, + "grad_norm": 0.40357154607772827, + "learning_rate": 0.000199541850886247, + "loss": 0.5238, + "step": 1310 + }, + { + "epoch": 0.030731904569348968, + "grad_norm": 0.5441321730613708, + "learning_rate": 0.00019954114631009928, + "loss": 0.1629, + "step": 1311 + }, + { + "epoch": 0.030755346144153966, + "grad_norm": 0.2254931628704071, + "learning_rate": 0.00019954044119383855, + "loss": 0.0838, + "step": 1312 + }, + { + "epoch": 0.03077878771895896, + "grad_norm": 0.7056419253349304, + "learning_rate": 0.00019953973553746854, + "loss": 0.3754, + "step": 1313 + }, + { + "epoch": 0.030802229293763955, + "grad_norm": 0.5924163460731506, + "learning_rate": 0.00019953902934099316, + "loss": 0.2384, + "step": 1314 + }, + { + "epoch": 0.03082567086856895, + "grad_norm": 0.33284640312194824, + "learning_rate": 0.0001995383226044162, + "loss": 0.1064, + "step": 1315 + }, + { + "epoch": 0.030849112443373947, + "grad_norm": 0.4977656900882721, + "learning_rate": 0.0001995376153277415, + "loss": 0.1366, + "step": 1316 + }, + { + "epoch": 0.03087255401817894, + "grad_norm": 0.10834868252277374, + "learning_rate": 0.00019953690751097289, + "loss": 0.0252, + "step": 1317 + }, + { + "epoch": 0.030895995592983935, + "grad_norm": 0.5983797907829285, + "learning_rate": 0.00019953619915411423, + "loss": 0.1482, + "step": 1318 + }, + { + "epoch": 0.030919437167788933, + "grad_norm": 0.23064617812633514, + "learning_rate": 0.00019953549025716937, + "loss": 0.0732, + "step": 1319 + }, + { + "epoch": 0.030942878742593927, + "grad_norm": 0.6093079447746277, + "learning_rate": 0.00019953478082014212, + "loss": 0.2772, + "step": 1320 + }, + { + "epoch": 0.03096632031739892, + "grad_norm": 0.45635536313056946, + "learning_rate": 0.00019953407084303634, + "loss": 0.2566, + "step": 1321 + }, + { + "epoch": 0.03098976189220392, + "grad_norm": 0.22575706243515015, + "learning_rate": 0.00019953336032585593, + "loss": 0.0618, + "step": 1322 + }, + { + "epoch": 0.031013203467008914, + "grad_norm": 0.2830718755722046, + "learning_rate": 0.0001995326492686047, + "loss": 0.1008, + "step": 1323 + }, + { + "epoch": 0.031036645041813908, + "grad_norm": 0.1035655215382576, + "learning_rate": 0.00019953193767128648, + "loss": 0.0287, + "step": 1324 + }, + { + "epoch": 0.031060086616618906, + "grad_norm": 0.4913709759712219, + "learning_rate": 0.0001995312255339052, + "loss": 0.6734, + "step": 1325 + }, + { + "epoch": 0.0310835281914239, + "grad_norm": 0.5051928758621216, + "learning_rate": 0.0001995305128564647, + "loss": 0.2025, + "step": 1326 + }, + { + "epoch": 0.031106969766228895, + "grad_norm": 0.2192583829164505, + "learning_rate": 0.0001995297996389688, + "loss": 0.0686, + "step": 1327 + }, + { + "epoch": 0.031130411341033892, + "grad_norm": 0.4489094018936157, + "learning_rate": 0.00019952908588142143, + "loss": 0.1611, + "step": 1328 + }, + { + "epoch": 0.031153852915838887, + "grad_norm": 0.3734390139579773, + "learning_rate": 0.00019952837158382646, + "loss": 0.1276, + "step": 1329 + }, + { + "epoch": 0.03117729449064388, + "grad_norm": 0.36243754625320435, + "learning_rate": 0.00019952765674618772, + "loss": 0.0526, + "step": 1330 + }, + { + "epoch": 0.031200736065448875, + "grad_norm": 0.5671990513801575, + "learning_rate": 0.00019952694136850913, + "loss": 0.1414, + "step": 1331 + }, + { + "epoch": 0.031224177640253873, + "grad_norm": 0.5083608031272888, + "learning_rate": 0.00019952622545079456, + "loss": 0.1661, + "step": 1332 + }, + { + "epoch": 0.031247619215058867, + "grad_norm": 0.5848331451416016, + "learning_rate": 0.0001995255089930479, + "loss": 0.128, + "step": 1333 + }, + { + "epoch": 0.031271060789863865, + "grad_norm": 0.366942822933197, + "learning_rate": 0.00019952479199527302, + "loss": 0.1253, + "step": 1334 + }, + { + "epoch": 0.03129450236466886, + "grad_norm": 0.6393371820449829, + "learning_rate": 0.0001995240744574738, + "loss": 0.1366, + "step": 1335 + }, + { + "epoch": 0.031317943939473854, + "grad_norm": 0.4556933641433716, + "learning_rate": 0.00019952335637965417, + "loss": 0.6138, + "step": 1336 + }, + { + "epoch": 0.03134138551427885, + "grad_norm": 0.5450738072395325, + "learning_rate": 0.000199522637761818, + "loss": 0.0748, + "step": 1337 + }, + { + "epoch": 0.03136482708908384, + "grad_norm": 0.559502363204956, + "learning_rate": 0.00019952191860396923, + "loss": 0.8341, + "step": 1338 + }, + { + "epoch": 0.03138826866388884, + "grad_norm": 0.29984813928604126, + "learning_rate": 0.0001995211989061117, + "loss": 0.1103, + "step": 1339 + }, + { + "epoch": 0.03141171023869384, + "grad_norm": 0.7225408554077148, + "learning_rate": 0.00019952047866824937, + "loss": 0.218, + "step": 1340 + }, + { + "epoch": 0.03143515181349883, + "grad_norm": 0.37625327706336975, + "learning_rate": 0.0001995197578903861, + "loss": 0.6726, + "step": 1341 + }, + { + "epoch": 0.03145859338830383, + "grad_norm": 0.6080980896949768, + "learning_rate": 0.00019951903657252585, + "loss": 0.1465, + "step": 1342 + }, + { + "epoch": 0.03148203496310882, + "grad_norm": 0.7998687624931335, + "learning_rate": 0.0001995183147146725, + "loss": 0.3562, + "step": 1343 + }, + { + "epoch": 0.031505476537913815, + "grad_norm": 0.2847980856895447, + "learning_rate": 0.00019951759231682996, + "loss": 0.0669, + "step": 1344 + }, + { + "epoch": 0.03152891811271881, + "grad_norm": 0.5033867359161377, + "learning_rate": 0.00019951686937900217, + "loss": 0.7949, + "step": 1345 + }, + { + "epoch": 0.03155235968752381, + "grad_norm": 0.31807830929756165, + "learning_rate": 0.00019951614590119307, + "loss": 0.1145, + "step": 1346 + }, + { + "epoch": 0.031575801262328805, + "grad_norm": 0.3002892732620239, + "learning_rate": 0.00019951542188340656, + "loss": 0.1343, + "step": 1347 + }, + { + "epoch": 0.0315992428371338, + "grad_norm": 0.19227789342403412, + "learning_rate": 0.00019951469732564653, + "loss": 0.052, + "step": 1348 + }, + { + "epoch": 0.031622684411938794, + "grad_norm": 0.6773321628570557, + "learning_rate": 0.000199513972227917, + "loss": 0.2398, + "step": 1349 + }, + { + "epoch": 0.03164612598674379, + "grad_norm": 0.1713387370109558, + "learning_rate": 0.00019951324659022186, + "loss": 0.0537, + "step": 1350 + }, + { + "epoch": 0.03166956756154878, + "grad_norm": 0.4832536280155182, + "learning_rate": 0.00019951252041256503, + "loss": 0.1369, + "step": 1351 + }, + { + "epoch": 0.031693009136353784, + "grad_norm": 0.7787949442863464, + "learning_rate": 0.00019951179369495048, + "loss": 0.3028, + "step": 1352 + }, + { + "epoch": 0.03171645071115878, + "grad_norm": 0.4839560389518738, + "learning_rate": 0.00019951106643738213, + "loss": 0.719, + "step": 1353 + }, + { + "epoch": 0.03173989228596377, + "grad_norm": 0.6670080423355103, + "learning_rate": 0.00019951033863986392, + "loss": 0.1804, + "step": 1354 + }, + { + "epoch": 0.031763333860768767, + "grad_norm": 0.5687954425811768, + "learning_rate": 0.00019950961030239982, + "loss": 0.2772, + "step": 1355 + }, + { + "epoch": 0.03178677543557376, + "grad_norm": 0.5086057782173157, + "learning_rate": 0.0001995088814249938, + "loss": 0.1881, + "step": 1356 + }, + { + "epoch": 0.031810217010378755, + "grad_norm": 0.7868693470954895, + "learning_rate": 0.00019950815200764978, + "loss": 0.1242, + "step": 1357 + }, + { + "epoch": 0.03183365858518375, + "grad_norm": 0.7812328934669495, + "learning_rate": 0.00019950742205037173, + "loss": 0.2005, + "step": 1358 + }, + { + "epoch": 0.03185710015998875, + "grad_norm": 0.5139791965484619, + "learning_rate": 0.0001995066915531636, + "loss": 0.1595, + "step": 1359 + }, + { + "epoch": 0.031880541734793745, + "grad_norm": 0.7708167433738708, + "learning_rate": 0.00019950596051602937, + "loss": 0.158, + "step": 1360 + }, + { + "epoch": 0.03190398330959874, + "grad_norm": 0.2919766306877136, + "learning_rate": 0.000199505228938973, + "loss": 0.1135, + "step": 1361 + }, + { + "epoch": 0.031927424884403734, + "grad_norm": 0.5775530338287354, + "learning_rate": 0.00019950449682199845, + "loss": 0.114, + "step": 1362 + }, + { + "epoch": 0.03195086645920873, + "grad_norm": 0.7231960296630859, + "learning_rate": 0.0001995037641651097, + "loss": 0.2896, + "step": 1363 + }, + { + "epoch": 0.03197430803401372, + "grad_norm": 0.2148783802986145, + "learning_rate": 0.00019950303096831075, + "loss": 0.0468, + "step": 1364 + }, + { + "epoch": 0.031997749608818724, + "grad_norm": 0.6225679516792297, + "learning_rate": 0.00019950229723160557, + "loss": 0.2144, + "step": 1365 + }, + { + "epoch": 0.03202119118362372, + "grad_norm": 0.5414308309555054, + "learning_rate": 0.0001995015629549981, + "loss": 0.1691, + "step": 1366 + }, + { + "epoch": 0.03204463275842871, + "grad_norm": 0.25999462604522705, + "learning_rate": 0.00019950082813849232, + "loss": 0.0844, + "step": 1367 + }, + { + "epoch": 0.032068074333233706, + "grad_norm": 0.15623898804187775, + "learning_rate": 0.0001995000927820923, + "loss": 0.0616, + "step": 1368 + }, + { + "epoch": 0.0320915159080387, + "grad_norm": 0.7916480302810669, + "learning_rate": 0.00019949935688580198, + "loss": 0.2997, + "step": 1369 + }, + { + "epoch": 0.032114957482843695, + "grad_norm": 0.7750642895698547, + "learning_rate": 0.00019949862044962532, + "loss": 0.3771, + "step": 1370 + }, + { + "epoch": 0.03213839905764869, + "grad_norm": 0.7214958071708679, + "learning_rate": 0.00019949788347356636, + "loss": 0.2498, + "step": 1371 + }, + { + "epoch": 0.03216184063245369, + "grad_norm": 0.31311067938804626, + "learning_rate": 0.0001994971459576291, + "loss": 0.0732, + "step": 1372 + }, + { + "epoch": 0.032185282207258685, + "grad_norm": 0.6309042572975159, + "learning_rate": 0.0001994964079018175, + "loss": 0.2019, + "step": 1373 + }, + { + "epoch": 0.03220872378206368, + "grad_norm": 0.7834305167198181, + "learning_rate": 0.00019949566930613563, + "loss": 0.2482, + "step": 1374 + }, + { + "epoch": 0.032232165356868674, + "grad_norm": 0.8589005470275879, + "learning_rate": 0.00019949493017058746, + "loss": 0.2042, + "step": 1375 + }, + { + "epoch": 0.03225560693167367, + "grad_norm": 0.4949837327003479, + "learning_rate": 0.00019949419049517696, + "loss": 0.1599, + "step": 1376 + }, + { + "epoch": 0.03227904850647866, + "grad_norm": 0.539564847946167, + "learning_rate": 0.00019949345027990822, + "loss": 0.144, + "step": 1377 + }, + { + "epoch": 0.032302490081283664, + "grad_norm": 0.4626608192920685, + "learning_rate": 0.0001994927095247852, + "loss": 0.1471, + "step": 1378 + }, + { + "epoch": 0.03232593165608866, + "grad_norm": 0.11935852468013763, + "learning_rate": 0.00019949196822981196, + "loss": 0.0437, + "step": 1379 + }, + { + "epoch": 0.03234937323089365, + "grad_norm": 0.1032068207859993, + "learning_rate": 0.00019949122639499248, + "loss": 0.0361, + "step": 1380 + }, + { + "epoch": 0.032372814805698646, + "grad_norm": 0.2297845333814621, + "learning_rate": 0.00019949048402033084, + "loss": 0.0781, + "step": 1381 + }, + { + "epoch": 0.03239625638050364, + "grad_norm": 0.2640823423862457, + "learning_rate": 0.00019948974110583101, + "loss": 0.1203, + "step": 1382 + }, + { + "epoch": 0.032419697955308635, + "grad_norm": 0.5162749886512756, + "learning_rate": 0.00019948899765149706, + "loss": 0.1206, + "step": 1383 + }, + { + "epoch": 0.032443139530113636, + "grad_norm": 0.310268759727478, + "learning_rate": 0.00019948825365733298, + "loss": 0.1306, + "step": 1384 + }, + { + "epoch": 0.03246658110491863, + "grad_norm": 0.5007949471473694, + "learning_rate": 0.00019948750912334289, + "loss": 0.1496, + "step": 1385 + }, + { + "epoch": 0.032490022679723625, + "grad_norm": 0.9015179872512817, + "learning_rate": 0.00019948676404953075, + "loss": 0.2487, + "step": 1386 + }, + { + "epoch": 0.03251346425452862, + "grad_norm": 0.6375516057014465, + "learning_rate": 0.00019948601843590064, + "loss": 0.27, + "step": 1387 + }, + { + "epoch": 0.032536905829333614, + "grad_norm": 0.17853212356567383, + "learning_rate": 0.0001994852722824566, + "loss": 0.0484, + "step": 1388 + }, + { + "epoch": 0.03256034740413861, + "grad_norm": 0.1275033950805664, + "learning_rate": 0.00019948452558920266, + "loss": 0.0447, + "step": 1389 + }, + { + "epoch": 0.0325837889789436, + "grad_norm": 0.5999669432640076, + "learning_rate": 0.00019948377835614286, + "loss": 0.3504, + "step": 1390 + }, + { + "epoch": 0.0326072305537486, + "grad_norm": 0.5204381942749023, + "learning_rate": 0.00019948303058328132, + "loss": 0.5569, + "step": 1391 + }, + { + "epoch": 0.0326306721285536, + "grad_norm": 0.27047091722488403, + "learning_rate": 0.00019948228227062206, + "loss": 0.0906, + "step": 1392 + }, + { + "epoch": 0.03265411370335859, + "grad_norm": 0.6264225244522095, + "learning_rate": 0.00019948153341816914, + "loss": 0.2164, + "step": 1393 + }, + { + "epoch": 0.032677555278163586, + "grad_norm": 0.4897654950618744, + "learning_rate": 0.0001994807840259266, + "loss": 0.0991, + "step": 1394 + }, + { + "epoch": 0.03270099685296858, + "grad_norm": 0.5795860290527344, + "learning_rate": 0.00019948003409389858, + "loss": 0.1925, + "step": 1395 + }, + { + "epoch": 0.032724438427773575, + "grad_norm": 0.3977848291397095, + "learning_rate": 0.00019947928362208902, + "loss": 0.1951, + "step": 1396 + }, + { + "epoch": 0.032747880002578576, + "grad_norm": 0.5540663003921509, + "learning_rate": 0.00019947853261050215, + "loss": 0.1862, + "step": 1397 + }, + { + "epoch": 0.03277132157738357, + "grad_norm": 0.4354701042175293, + "learning_rate": 0.0001994777810591419, + "loss": 0.078, + "step": 1398 + }, + { + "epoch": 0.032794763152188565, + "grad_norm": 0.2681378722190857, + "learning_rate": 0.00019947702896801244, + "loss": 0.0975, + "step": 1399 + }, + { + "epoch": 0.03281820472699356, + "grad_norm": 0.672390878200531, + "learning_rate": 0.00019947627633711782, + "loss": 0.2667, + "step": 1400 + }, + { + "epoch": 0.032841646301798554, + "grad_norm": 0.3623005449771881, + "learning_rate": 0.00019947552316646215, + "loss": 0.0894, + "step": 1401 + }, + { + "epoch": 0.03286508787660355, + "grad_norm": 0.5057967305183411, + "learning_rate": 0.00019947476945604946, + "loss": 0.1368, + "step": 1402 + }, + { + "epoch": 0.03288852945140854, + "grad_norm": 0.68674236536026, + "learning_rate": 0.00019947401520588388, + "loss": 0.2207, + "step": 1403 + }, + { + "epoch": 0.03291197102621354, + "grad_norm": 0.4129534065723419, + "learning_rate": 0.00019947326041596953, + "loss": 0.127, + "step": 1404 + }, + { + "epoch": 0.03293541260101854, + "grad_norm": 0.6966304779052734, + "learning_rate": 0.00019947250508631044, + "loss": 0.2635, + "step": 1405 + }, + { + "epoch": 0.03295885417582353, + "grad_norm": 0.7108837962150574, + "learning_rate": 0.00019947174921691074, + "loss": 0.2016, + "step": 1406 + }, + { + "epoch": 0.032982295750628526, + "grad_norm": 0.937193751335144, + "learning_rate": 0.00019947099280777454, + "loss": 0.222, + "step": 1407 + }, + { + "epoch": 0.03300573732543352, + "grad_norm": 0.5149775743484497, + "learning_rate": 0.0001994702358589059, + "loss": 0.1163, + "step": 1408 + }, + { + "epoch": 0.033029178900238515, + "grad_norm": 0.8320923447608948, + "learning_rate": 0.00019946947837030905, + "loss": 0.2652, + "step": 1409 + }, + { + "epoch": 0.033052620475043516, + "grad_norm": 0.5638698935508728, + "learning_rate": 0.00019946872034198794, + "loss": 0.1551, + "step": 1410 + }, + { + "epoch": 0.03307606204984851, + "grad_norm": 0.7935977578163147, + "learning_rate": 0.00019946796177394677, + "loss": 0.2142, + "step": 1411 + }, + { + "epoch": 0.033099503624653505, + "grad_norm": 0.68088299036026, + "learning_rate": 0.00019946720266618968, + "loss": 0.1244, + "step": 1412 + }, + { + "epoch": 0.0331229451994585, + "grad_norm": 0.34091776609420776, + "learning_rate": 0.0001994664430187207, + "loss": 0.3526, + "step": 1413 + }, + { + "epoch": 0.033146386774263494, + "grad_norm": 0.719480574131012, + "learning_rate": 0.00019946568283154404, + "loss": 0.2267, + "step": 1414 + }, + { + "epoch": 0.03316982834906849, + "grad_norm": 0.5436756014823914, + "learning_rate": 0.00019946492210466378, + "loss": 0.1489, + "step": 1415 + }, + { + "epoch": 0.03319326992387349, + "grad_norm": 0.5614482164382935, + "learning_rate": 0.00019946416083808406, + "loss": 0.2202, + "step": 1416 + }, + { + "epoch": 0.03321671149867848, + "grad_norm": 0.5843610167503357, + "learning_rate": 0.00019946339903180896, + "loss": 0.1623, + "step": 1417 + }, + { + "epoch": 0.03324015307348348, + "grad_norm": 0.21810437738895416, + "learning_rate": 0.00019946263668584271, + "loss": 0.0418, + "step": 1418 + }, + { + "epoch": 0.03326359464828847, + "grad_norm": 0.3877335488796234, + "learning_rate": 0.00019946187380018938, + "loss": 0.1083, + "step": 1419 + }, + { + "epoch": 0.033287036223093466, + "grad_norm": 0.5312466621398926, + "learning_rate": 0.00019946111037485314, + "loss": 0.7331, + "step": 1420 + }, + { + "epoch": 0.03331047779789846, + "grad_norm": 0.7420642375946045, + "learning_rate": 0.00019946034640983812, + "loss": 0.2463, + "step": 1421 + }, + { + "epoch": 0.033333919372703455, + "grad_norm": 0.4980773627758026, + "learning_rate": 0.00019945958190514845, + "loss": 0.8346, + "step": 1422 + }, + { + "epoch": 0.033357360947508456, + "grad_norm": 0.6287073493003845, + "learning_rate": 0.00019945881686078828, + "loss": 0.2006, + "step": 1423 + }, + { + "epoch": 0.03338080252231345, + "grad_norm": 0.52288419008255, + "learning_rate": 0.0001994580512767618, + "loss": 0.2039, + "step": 1424 + }, + { + "epoch": 0.033404244097118445, + "grad_norm": 0.32891911268234253, + "learning_rate": 0.0001994572851530731, + "loss": 0.1417, + "step": 1425 + }, + { + "epoch": 0.03342768567192344, + "grad_norm": 0.8882091641426086, + "learning_rate": 0.00019945651848972642, + "loss": 0.2588, + "step": 1426 + }, + { + "epoch": 0.03345112724672843, + "grad_norm": 0.7421273589134216, + "learning_rate": 0.00019945575128672586, + "loss": 0.1719, + "step": 1427 + }, + { + "epoch": 0.03347456882153343, + "grad_norm": 0.6317639350891113, + "learning_rate": 0.0001994549835440756, + "loss": 0.1735, + "step": 1428 + }, + { + "epoch": 0.03349801039633843, + "grad_norm": 0.6305322647094727, + "learning_rate": 0.00019945421526177976, + "loss": 0.2535, + "step": 1429 + }, + { + "epoch": 0.03352145197114342, + "grad_norm": 0.3368450403213501, + "learning_rate": 0.0001994534464398426, + "loss": 0.1225, + "step": 1430 + }, + { + "epoch": 0.03354489354594842, + "grad_norm": 0.8812502026557922, + "learning_rate": 0.00019945267707826824, + "loss": 0.2582, + "step": 1431 + }, + { + "epoch": 0.03356833512075341, + "grad_norm": 0.537307620048523, + "learning_rate": 0.00019945190717706083, + "loss": 0.2671, + "step": 1432 + }, + { + "epoch": 0.033591776695558406, + "grad_norm": 0.7708643674850464, + "learning_rate": 0.0001994511367362246, + "loss": 0.2129, + "step": 1433 + }, + { + "epoch": 0.0336152182703634, + "grad_norm": 0.4790114760398865, + "learning_rate": 0.0001994503657557637, + "loss": 0.1944, + "step": 1434 + }, + { + "epoch": 0.0336386598451684, + "grad_norm": 0.4827374815940857, + "learning_rate": 0.00019944959423568232, + "loss": 0.1582, + "step": 1435 + }, + { + "epoch": 0.033662101419973396, + "grad_norm": 0.383157879114151, + "learning_rate": 0.00019944882217598465, + "loss": 0.0522, + "step": 1436 + }, + { + "epoch": 0.03368554299477839, + "grad_norm": 0.5356507301330566, + "learning_rate": 0.00019944804957667484, + "loss": 0.1824, + "step": 1437 + }, + { + "epoch": 0.033708984569583385, + "grad_norm": 0.9170346260070801, + "learning_rate": 0.00019944727643775716, + "loss": 0.2675, + "step": 1438 + }, + { + "epoch": 0.03373242614438838, + "grad_norm": 0.5859565734863281, + "learning_rate": 0.00019944650275923573, + "loss": 0.21, + "step": 1439 + }, + { + "epoch": 0.03375586771919337, + "grad_norm": 0.4387784004211426, + "learning_rate": 0.0001994457285411148, + "loss": 0.1354, + "step": 1440 + }, + { + "epoch": 0.03377930929399837, + "grad_norm": 0.7216665744781494, + "learning_rate": 0.00019944495378339854, + "loss": 0.2345, + "step": 1441 + }, + { + "epoch": 0.03380275086880337, + "grad_norm": 0.7390990257263184, + "learning_rate": 0.00019944417848609118, + "loss": 0.1822, + "step": 1442 + }, + { + "epoch": 0.03382619244360836, + "grad_norm": 0.6986528635025024, + "learning_rate": 0.0001994434026491969, + "loss": 0.2346, + "step": 1443 + }, + { + "epoch": 0.03384963401841336, + "grad_norm": 0.5563347935676575, + "learning_rate": 0.00019944262627271991, + "loss": 0.1188, + "step": 1444 + }, + { + "epoch": 0.03387307559321835, + "grad_norm": 0.4763752520084381, + "learning_rate": 0.00019944184935666444, + "loss": 0.0802, + "step": 1445 + }, + { + "epoch": 0.033896517168023346, + "grad_norm": 0.7656883001327515, + "learning_rate": 0.0001994410719010347, + "loss": 0.1864, + "step": 1446 + }, + { + "epoch": 0.03391995874282834, + "grad_norm": 0.20474442839622498, + "learning_rate": 0.00019944029390583495, + "loss": 0.0559, + "step": 1447 + }, + { + "epoch": 0.03394340031763334, + "grad_norm": 0.6368665099143982, + "learning_rate": 0.00019943951537106932, + "loss": 0.1763, + "step": 1448 + }, + { + "epoch": 0.033966841892438336, + "grad_norm": 0.5481422543525696, + "learning_rate": 0.0001994387362967421, + "loss": 0.2062, + "step": 1449 + }, + { + "epoch": 0.03399028346724333, + "grad_norm": 0.4000488519668579, + "learning_rate": 0.0001994379566828575, + "loss": 0.1175, + "step": 1450 + }, + { + "epoch": 0.034013725042048325, + "grad_norm": 0.6314788460731506, + "learning_rate": 0.00019943717652941975, + "loss": 0.7582, + "step": 1451 + }, + { + "epoch": 0.03403716661685332, + "grad_norm": 0.2895454168319702, + "learning_rate": 0.0001994363958364331, + "loss": 0.1499, + "step": 1452 + }, + { + "epoch": 0.03406060819165831, + "grad_norm": 0.7281842827796936, + "learning_rate": 0.0001994356146039017, + "loss": 0.2448, + "step": 1453 + }, + { + "epoch": 0.03408404976646331, + "grad_norm": 0.7268853783607483, + "learning_rate": 0.00019943483283182994, + "loss": 0.1484, + "step": 1454 + }, + { + "epoch": 0.03410749134126831, + "grad_norm": 0.8516520261764526, + "learning_rate": 0.00019943405052022195, + "loss": 0.2366, + "step": 1455 + }, + { + "epoch": 0.0341309329160733, + "grad_norm": 0.12079279869794846, + "learning_rate": 0.00019943326766908205, + "loss": 0.024, + "step": 1456 + }, + { + "epoch": 0.0341543744908783, + "grad_norm": 0.22875623404979706, + "learning_rate": 0.00019943248427841436, + "loss": 0.048, + "step": 1457 + }, + { + "epoch": 0.03417781606568329, + "grad_norm": 0.44669094681739807, + "learning_rate": 0.00019943170034822328, + "loss": 0.1363, + "step": 1458 + }, + { + "epoch": 0.034201257640488286, + "grad_norm": 0.3731606602668762, + "learning_rate": 0.00019943091587851296, + "loss": 0.1038, + "step": 1459 + }, + { + "epoch": 0.03422469921529328, + "grad_norm": 0.5608958601951599, + "learning_rate": 0.00019943013086928772, + "loss": 0.1722, + "step": 1460 + }, + { + "epoch": 0.03424814079009828, + "grad_norm": 0.515383243560791, + "learning_rate": 0.0001994293453205518, + "loss": 0.2433, + "step": 1461 + }, + { + "epoch": 0.034271582364903276, + "grad_norm": 0.4520937502384186, + "learning_rate": 0.00019942855923230944, + "loss": 0.1419, + "step": 1462 + }, + { + "epoch": 0.03429502393970827, + "grad_norm": 0.6170895099639893, + "learning_rate": 0.00019942777260456492, + "loss": 0.2156, + "step": 1463 + }, + { + "epoch": 0.034318465514513265, + "grad_norm": 0.2649894058704376, + "learning_rate": 0.0001994269854373225, + "loss": 0.1036, + "step": 1464 + }, + { + "epoch": 0.03434190708931826, + "grad_norm": 0.1776680201292038, + "learning_rate": 0.00019942619773058646, + "loss": 0.0578, + "step": 1465 + }, + { + "epoch": 0.03436534866412325, + "grad_norm": 0.36533549427986145, + "learning_rate": 0.00019942540948436111, + "loss": 0.0906, + "step": 1466 + }, + { + "epoch": 0.034388790238928255, + "grad_norm": 0.41770902276039124, + "learning_rate": 0.00019942462069865067, + "loss": 0.1442, + "step": 1467 + }, + { + "epoch": 0.03441223181373325, + "grad_norm": 0.43451231718063354, + "learning_rate": 0.00019942383137345945, + "loss": 0.122, + "step": 1468 + }, + { + "epoch": 0.03443567338853824, + "grad_norm": 0.6064940094947815, + "learning_rate": 0.00019942304150879172, + "loss": 0.1683, + "step": 1469 + }, + { + "epoch": 0.03445911496334324, + "grad_norm": 0.7216470837593079, + "learning_rate": 0.0001994222511046518, + "loss": 0.2605, + "step": 1470 + }, + { + "epoch": 0.03448255653814823, + "grad_norm": 0.21833743155002594, + "learning_rate": 0.00019942146016104387, + "loss": 0.0777, + "step": 1471 + }, + { + "epoch": 0.034505998112953226, + "grad_norm": 0.4405701160430908, + "learning_rate": 0.00019942066867797238, + "loss": 0.1469, + "step": 1472 + }, + { + "epoch": 0.03452943968775822, + "grad_norm": 0.19066959619522095, + "learning_rate": 0.00019941987665544151, + "loss": 0.0704, + "step": 1473 + }, + { + "epoch": 0.03455288126256322, + "grad_norm": 1.110748291015625, + "learning_rate": 0.0001994190840934556, + "loss": 0.2576, + "step": 1474 + }, + { + "epoch": 0.034576322837368216, + "grad_norm": 0.4706651270389557, + "learning_rate": 0.00019941829099201895, + "loss": 0.1624, + "step": 1475 + }, + { + "epoch": 0.03459976441217321, + "grad_norm": 0.3899070918560028, + "learning_rate": 0.00019941749735113585, + "loss": 0.1589, + "step": 1476 + }, + { + "epoch": 0.034623205986978205, + "grad_norm": 0.4899284243583679, + "learning_rate": 0.00019941670317081064, + "loss": 0.2335, + "step": 1477 + }, + { + "epoch": 0.0346466475617832, + "grad_norm": 0.6976783871650696, + "learning_rate": 0.00019941590845104758, + "loss": 0.2168, + "step": 1478 + }, + { + "epoch": 0.03467008913658819, + "grad_norm": 1.6537820100784302, + "learning_rate": 0.000199415113191851, + "loss": 0.1542, + "step": 1479 + }, + { + "epoch": 0.034693530711393195, + "grad_norm": 0.46323543787002563, + "learning_rate": 0.00019941431739322525, + "loss": 0.1213, + "step": 1480 + }, + { + "epoch": 0.03471697228619819, + "grad_norm": 0.5881422162055969, + "learning_rate": 0.0001994135210551746, + "loss": 0.2419, + "step": 1481 + }, + { + "epoch": 0.03474041386100318, + "grad_norm": 0.7006497383117676, + "learning_rate": 0.0001994127241777034, + "loss": 0.2656, + "step": 1482 + }, + { + "epoch": 0.03476385543580818, + "grad_norm": 0.37066030502319336, + "learning_rate": 0.00019941192676081594, + "loss": 0.1117, + "step": 1483 + }, + { + "epoch": 0.03478729701061317, + "grad_norm": 0.6707988381385803, + "learning_rate": 0.0001994111288045166, + "loss": 0.1639, + "step": 1484 + }, + { + "epoch": 0.034810738585418166, + "grad_norm": 0.9128818511962891, + "learning_rate": 0.00019941033030880962, + "loss": 0.1959, + "step": 1485 + }, + { + "epoch": 0.03483418016022316, + "grad_norm": 0.7623459696769714, + "learning_rate": 0.00019940953127369943, + "loss": 0.1965, + "step": 1486 + }, + { + "epoch": 0.03485762173502816, + "grad_norm": 0.3549376130104065, + "learning_rate": 0.00019940873169919034, + "loss": 0.1755, + "step": 1487 + }, + { + "epoch": 0.034881063309833156, + "grad_norm": 0.6285916566848755, + "learning_rate": 0.00019940793158528666, + "loss": 0.2136, + "step": 1488 + }, + { + "epoch": 0.03490450488463815, + "grad_norm": 0.4287380874156952, + "learning_rate": 0.00019940713093199276, + "loss": 0.1647, + "step": 1489 + }, + { + "epoch": 0.034927946459443145, + "grad_norm": 0.47663530707359314, + "learning_rate": 0.00019940632973931296, + "loss": 0.7745, + "step": 1490 + }, + { + "epoch": 0.03495138803424814, + "grad_norm": 1.0636008977890015, + "learning_rate": 0.00019940552800725163, + "loss": 0.3448, + "step": 1491 + }, + { + "epoch": 0.03497482960905313, + "grad_norm": 0.06852046400308609, + "learning_rate": 0.0001994047257358131, + "loss": 0.0131, + "step": 1492 + }, + { + "epoch": 0.034998271183858135, + "grad_norm": 0.9414334297180176, + "learning_rate": 0.00019940392292500176, + "loss": 0.1308, + "step": 1493 + }, + { + "epoch": 0.03502171275866313, + "grad_norm": 0.2715790569782257, + "learning_rate": 0.00019940311957482193, + "loss": 0.0479, + "step": 1494 + }, + { + "epoch": 0.03504515433346812, + "grad_norm": 0.583857536315918, + "learning_rate": 0.00019940231568527797, + "loss": 0.9888, + "step": 1495 + }, + { + "epoch": 0.03506859590827312, + "grad_norm": 1.005668044090271, + "learning_rate": 0.00019940151125637424, + "loss": 0.3238, + "step": 1496 + }, + { + "epoch": 0.03509203748307811, + "grad_norm": 0.6315475106239319, + "learning_rate": 0.00019940070628811515, + "loss": 0.2173, + "step": 1497 + }, + { + "epoch": 0.035115479057883106, + "grad_norm": 0.6719298958778381, + "learning_rate": 0.00019939990078050496, + "loss": 0.2196, + "step": 1498 + }, + { + "epoch": 0.03513892063268811, + "grad_norm": 0.8462068438529968, + "learning_rate": 0.0001993990947335482, + "loss": 0.21, + "step": 1499 + }, + { + "epoch": 0.0351623622074931, + "grad_norm": 1.2403252124786377, + "learning_rate": 0.0001993982881472491, + "loss": 0.3272, + "step": 1500 + }, + { + "epoch": 0.035185803782298096, + "grad_norm": 0.5896390676498413, + "learning_rate": 0.0001993974810216121, + "loss": 0.7075, + "step": 1501 + }, + { + "epoch": 0.03520924535710309, + "grad_norm": 0.30274176597595215, + "learning_rate": 0.00019939667335664162, + "loss": 0.1012, + "step": 1502 + }, + { + "epoch": 0.035232686931908085, + "grad_norm": 0.478086918592453, + "learning_rate": 0.00019939586515234194, + "loss": 0.2418, + "step": 1503 + }, + { + "epoch": 0.03525612850671308, + "grad_norm": 0.5664384961128235, + "learning_rate": 0.0001993950564087175, + "loss": 0.2205, + "step": 1504 + }, + { + "epoch": 0.03527957008151807, + "grad_norm": 0.6076618432998657, + "learning_rate": 0.00019939424712577274, + "loss": 0.1828, + "step": 1505 + }, + { + "epoch": 0.035303011656323074, + "grad_norm": 0.5271397233009338, + "learning_rate": 0.00019939343730351197, + "loss": 0.199, + "step": 1506 + }, + { + "epoch": 0.03532645323112807, + "grad_norm": 0.5535656213760376, + "learning_rate": 0.00019939262694193958, + "loss": 0.1652, + "step": 1507 + }, + { + "epoch": 0.03534989480593306, + "grad_norm": 0.48601067066192627, + "learning_rate": 0.00019939181604106005, + "loss": 0.1593, + "step": 1508 + }, + { + "epoch": 0.03537333638073806, + "grad_norm": 0.15992216765880585, + "learning_rate": 0.00019939100460087772, + "loss": 0.063, + "step": 1509 + }, + { + "epoch": 0.03539677795554305, + "grad_norm": 0.4650883674621582, + "learning_rate": 0.00019939019262139702, + "loss": 0.8437, + "step": 1510 + }, + { + "epoch": 0.035420219530348046, + "grad_norm": 0.5139392614364624, + "learning_rate": 0.00019938938010262232, + "loss": 0.1602, + "step": 1511 + }, + { + "epoch": 0.03544366110515305, + "grad_norm": 0.09309602528810501, + "learning_rate": 0.00019938856704455803, + "loss": 0.032, + "step": 1512 + }, + { + "epoch": 0.03546710267995804, + "grad_norm": 0.4388226568698883, + "learning_rate": 0.0001993877534472086, + "loss": 0.1034, + "step": 1513 + }, + { + "epoch": 0.035490544254763036, + "grad_norm": 0.4483087658882141, + "learning_rate": 0.00019938693931057845, + "loss": 0.1444, + "step": 1514 + }, + { + "epoch": 0.03551398582956803, + "grad_norm": 0.40976935625076294, + "learning_rate": 0.00019938612463467193, + "loss": 0.119, + "step": 1515 + }, + { + "epoch": 0.035537427404373025, + "grad_norm": 0.5402634739875793, + "learning_rate": 0.00019938530941949353, + "loss": 0.1934, + "step": 1516 + }, + { + "epoch": 0.03556086897917802, + "grad_norm": 0.426124632358551, + "learning_rate": 0.00019938449366504766, + "loss": 0.1437, + "step": 1517 + }, + { + "epoch": 0.03558431055398301, + "grad_norm": 0.19560696184635162, + "learning_rate": 0.0001993836773713387, + "loss": 0.084, + "step": 1518 + }, + { + "epoch": 0.035607752128788014, + "grad_norm": 0.7218527793884277, + "learning_rate": 0.00019938286053837112, + "loss": 0.9045, + "step": 1519 + }, + { + "epoch": 0.03563119370359301, + "grad_norm": 0.19447897374629974, + "learning_rate": 0.00019938204316614938, + "loss": 0.0499, + "step": 1520 + }, + { + "epoch": 0.035654635278398, + "grad_norm": 0.16653206944465637, + "learning_rate": 0.0001993812252546778, + "loss": 0.0704, + "step": 1521 + }, + { + "epoch": 0.035678076853203, + "grad_norm": 0.7457926273345947, + "learning_rate": 0.00019938040680396094, + "loss": 0.2182, + "step": 1522 + }, + { + "epoch": 0.03570151842800799, + "grad_norm": 0.1753389835357666, + "learning_rate": 0.00019937958781400325, + "loss": 0.0601, + "step": 1523 + }, + { + "epoch": 0.035724960002812986, + "grad_norm": 0.38723838329315186, + "learning_rate": 0.00019937876828480906, + "loss": 0.1433, + "step": 1524 + }, + { + "epoch": 0.03574840157761799, + "grad_norm": 0.733288049697876, + "learning_rate": 0.0001993779482163829, + "loss": 0.2855, + "step": 1525 + }, + { + "epoch": 0.03577184315242298, + "grad_norm": 0.8172062039375305, + "learning_rate": 0.0001993771276087292, + "loss": 0.1313, + "step": 1526 + }, + { + "epoch": 0.035795284727227976, + "grad_norm": 0.5077277421951294, + "learning_rate": 0.0001993763064618524, + "loss": 0.2266, + "step": 1527 + }, + { + "epoch": 0.03581872630203297, + "grad_norm": 0.5727379322052002, + "learning_rate": 0.00019937548477575695, + "loss": 0.1448, + "step": 1528 + }, + { + "epoch": 0.035842167876837965, + "grad_norm": 0.552718460559845, + "learning_rate": 0.00019937466255044736, + "loss": 0.1898, + "step": 1529 + }, + { + "epoch": 0.03586560945164296, + "grad_norm": 0.5990809202194214, + "learning_rate": 0.00019937383978592806, + "loss": 0.1505, + "step": 1530 + }, + { + "epoch": 0.03588905102644796, + "grad_norm": 0.34167054295539856, + "learning_rate": 0.0001993730164822035, + "loss": 0.0966, + "step": 1531 + }, + { + "epoch": 0.035912492601252954, + "grad_norm": 0.6499657034873962, + "learning_rate": 0.00019937219263927816, + "loss": 0.6786, + "step": 1532 + }, + { + "epoch": 0.03593593417605795, + "grad_norm": 0.5614050626754761, + "learning_rate": 0.00019937136825715648, + "loss": 0.1479, + "step": 1533 + }, + { + "epoch": 0.03595937575086294, + "grad_norm": 0.8230414390563965, + "learning_rate": 0.000199370543335843, + "loss": 0.3259, + "step": 1534 + }, + { + "epoch": 0.03598281732566794, + "grad_norm": 0.4668335020542145, + "learning_rate": 0.00019936971787534212, + "loss": 0.1132, + "step": 1535 + }, + { + "epoch": 0.03600625890047293, + "grad_norm": 0.24762609601020813, + "learning_rate": 0.0001993688918756584, + "loss": 0.0635, + "step": 1536 + }, + { + "epoch": 0.036029700475277926, + "grad_norm": 0.42154037952423096, + "learning_rate": 0.00019936806533679623, + "loss": 0.1, + "step": 1537 + }, + { + "epoch": 0.03605314205008293, + "grad_norm": 0.7844969630241394, + "learning_rate": 0.00019936723825876018, + "loss": 0.1882, + "step": 1538 + }, + { + "epoch": 0.03607658362488792, + "grad_norm": 0.6626172661781311, + "learning_rate": 0.0001993664106415547, + "loss": 0.208, + "step": 1539 + }, + { + "epoch": 0.036100025199692916, + "grad_norm": 0.278112530708313, + "learning_rate": 0.00019936558248518429, + "loss": 0.0658, + "step": 1540 + }, + { + "epoch": 0.03612346677449791, + "grad_norm": 0.73876953125, + "learning_rate": 0.00019936475378965341, + "loss": 0.3082, + "step": 1541 + }, + { + "epoch": 0.036146908349302904, + "grad_norm": 0.8610652089118958, + "learning_rate": 0.00019936392455496658, + "loss": 0.2188, + "step": 1542 + }, + { + "epoch": 0.0361703499241079, + "grad_norm": 0.7485872507095337, + "learning_rate": 0.00019936309478112832, + "loss": 0.2343, + "step": 1543 + }, + { + "epoch": 0.0361937914989129, + "grad_norm": 0.7803905010223389, + "learning_rate": 0.00019936226446814312, + "loss": 0.3395, + "step": 1544 + }, + { + "epoch": 0.036217233073717894, + "grad_norm": 0.9817821979522705, + "learning_rate": 0.00019936143361601547, + "loss": 0.1931, + "step": 1545 + }, + { + "epoch": 0.03624067464852289, + "grad_norm": 0.3995992839336395, + "learning_rate": 0.00019936060222474987, + "loss": 0.084, + "step": 1546 + }, + { + "epoch": 0.03626411622332788, + "grad_norm": 0.45571795105934143, + "learning_rate": 0.00019935977029435087, + "loss": 0.1817, + "step": 1547 + }, + { + "epoch": 0.03628755779813288, + "grad_norm": 0.35392269492149353, + "learning_rate": 0.00019935893782482297, + "loss": 0.0785, + "step": 1548 + }, + { + "epoch": 0.03631099937293787, + "grad_norm": 0.797389566898346, + "learning_rate": 0.00019935810481617065, + "loss": 0.2219, + "step": 1549 + }, + { + "epoch": 0.036334440947742866, + "grad_norm": 0.5008671283721924, + "learning_rate": 0.00019935727126839847, + "loss": 0.2153, + "step": 1550 + }, + { + "epoch": 0.03635788252254787, + "grad_norm": 0.9872144460678101, + "learning_rate": 0.00019935643718151097, + "loss": 0.2187, + "step": 1551 + }, + { + "epoch": 0.03638132409735286, + "grad_norm": 0.6145446300506592, + "learning_rate": 0.0001993556025555126, + "loss": 0.1515, + "step": 1552 + }, + { + "epoch": 0.036404765672157856, + "grad_norm": 0.5923949480056763, + "learning_rate": 0.00019935476739040798, + "loss": 0.1705, + "step": 1553 + }, + { + "epoch": 0.03642820724696285, + "grad_norm": 0.19945025444030762, + "learning_rate": 0.0001993539316862016, + "loss": 0.035, + "step": 1554 + }, + { + "epoch": 0.036451648821767844, + "grad_norm": 0.8144315481185913, + "learning_rate": 0.00019935309544289796, + "loss": 0.1879, + "step": 1555 + }, + { + "epoch": 0.03647509039657284, + "grad_norm": 0.9629526138305664, + "learning_rate": 0.00019935225866050163, + "loss": 0.2221, + "step": 1556 + }, + { + "epoch": 0.03649853197137784, + "grad_norm": 0.44918006658554077, + "learning_rate": 0.00019935142133901715, + "loss": 0.1556, + "step": 1557 + }, + { + "epoch": 0.036521973546182834, + "grad_norm": 0.35598745942115784, + "learning_rate": 0.00019935058347844912, + "loss": 0.072, + "step": 1558 + }, + { + "epoch": 0.03654541512098783, + "grad_norm": 0.6084961295127869, + "learning_rate": 0.00019934974507880197, + "loss": 0.2602, + "step": 1559 + }, + { + "epoch": 0.03656885669579282, + "grad_norm": 0.8145936131477356, + "learning_rate": 0.00019934890614008036, + "loss": 0.2431, + "step": 1560 + }, + { + "epoch": 0.03659229827059782, + "grad_norm": 0.12920810282230377, + "learning_rate": 0.00019934806666228874, + "loss": 0.0229, + "step": 1561 + }, + { + "epoch": 0.03661573984540281, + "grad_norm": 0.36000287532806396, + "learning_rate": 0.00019934722664543175, + "loss": 0.1299, + "step": 1562 + }, + { + "epoch": 0.03663918142020781, + "grad_norm": 0.7396690249443054, + "learning_rate": 0.00019934638608951391, + "loss": 0.1784, + "step": 1563 + }, + { + "epoch": 0.03666262299501281, + "grad_norm": 0.46396055817604065, + "learning_rate": 0.00019934554499453976, + "loss": 0.0882, + "step": 1564 + }, + { + "epoch": 0.0366860645698178, + "grad_norm": 1.7557995319366455, + "learning_rate": 0.00019934470336051388, + "loss": 0.2184, + "step": 1565 + }, + { + "epoch": 0.036709506144622796, + "grad_norm": 0.3661404848098755, + "learning_rate": 0.0001993438611874409, + "loss": 0.106, + "step": 1566 + }, + { + "epoch": 0.03673294771942779, + "grad_norm": 0.3686346113681793, + "learning_rate": 0.00019934301847532532, + "loss": 0.124, + "step": 1567 + }, + { + "epoch": 0.036756389294232784, + "grad_norm": 0.5505778789520264, + "learning_rate": 0.0001993421752241717, + "loss": 0.1959, + "step": 1568 + }, + { + "epoch": 0.03677983086903778, + "grad_norm": 0.6041340231895447, + "learning_rate": 0.00019934133143398468, + "loss": 0.2953, + "step": 1569 + }, + { + "epoch": 0.03680327244384278, + "grad_norm": 0.25849035382270813, + "learning_rate": 0.0001993404871047688, + "loss": 0.0853, + "step": 1570 + }, + { + "epoch": 0.036826714018647774, + "grad_norm": 0.5496562719345093, + "learning_rate": 0.00019933964223652862, + "loss": 0.1572, + "step": 1571 + }, + { + "epoch": 0.03685015559345277, + "grad_norm": 0.15931156277656555, + "learning_rate": 0.00019933879682926874, + "loss": 0.0459, + "step": 1572 + }, + { + "epoch": 0.03687359716825776, + "grad_norm": 0.6565028429031372, + "learning_rate": 0.0001993379508829938, + "loss": 0.3902, + "step": 1573 + }, + { + "epoch": 0.03689703874306276, + "grad_norm": 0.4672311246395111, + "learning_rate": 0.0001993371043977083, + "loss": 0.1411, + "step": 1574 + }, + { + "epoch": 0.03692048031786775, + "grad_norm": 0.794900119304657, + "learning_rate": 0.0001993362573734169, + "loss": 0.4258, + "step": 1575 + }, + { + "epoch": 0.03694392189267275, + "grad_norm": 0.9526443481445312, + "learning_rate": 0.00019933540981012418, + "loss": 0.2233, + "step": 1576 + }, + { + "epoch": 0.03696736346747775, + "grad_norm": 0.5868130326271057, + "learning_rate": 0.00019933456170783472, + "loss": 0.205, + "step": 1577 + }, + { + "epoch": 0.03699080504228274, + "grad_norm": 0.6385966539382935, + "learning_rate": 0.00019933371306655312, + "loss": 0.1597, + "step": 1578 + }, + { + "epoch": 0.037014246617087736, + "grad_norm": 0.6843447089195251, + "learning_rate": 0.00019933286388628405, + "loss": 0.7977, + "step": 1579 + }, + { + "epoch": 0.03703768819189273, + "grad_norm": 0.3484320640563965, + "learning_rate": 0.00019933201416703203, + "loss": 0.1592, + "step": 1580 + }, + { + "epoch": 0.037061129766697724, + "grad_norm": 0.47097527980804443, + "learning_rate": 0.0001993311639088017, + "loss": 0.1418, + "step": 1581 + }, + { + "epoch": 0.037084571341502726, + "grad_norm": 0.5379921197891235, + "learning_rate": 0.0001993303131115977, + "loss": 0.2071, + "step": 1582 + }, + { + "epoch": 0.03710801291630772, + "grad_norm": 0.7015400528907776, + "learning_rate": 0.0001993294617754246, + "loss": 0.1849, + "step": 1583 + }, + { + "epoch": 0.037131454491112714, + "grad_norm": 0.45825645327568054, + "learning_rate": 0.0001993286099002871, + "loss": 0.1892, + "step": 1584 + }, + { + "epoch": 0.03715489606591771, + "grad_norm": 0.16867394745349884, + "learning_rate": 0.00019932775748618975, + "loss": 0.044, + "step": 1585 + }, + { + "epoch": 0.0371783376407227, + "grad_norm": 0.49758344888687134, + "learning_rate": 0.0001993269045331372, + "loss": 0.1982, + "step": 1586 + }, + { + "epoch": 0.0372017792155277, + "grad_norm": 0.4135840833187103, + "learning_rate": 0.00019932605104113404, + "loss": 0.1731, + "step": 1587 + }, + { + "epoch": 0.03722522079033269, + "grad_norm": 0.5072572827339172, + "learning_rate": 0.00019932519701018495, + "loss": 0.1704, + "step": 1588 + }, + { + "epoch": 0.03724866236513769, + "grad_norm": 0.24640750885009766, + "learning_rate": 0.00019932434244029456, + "loss": 0.0393, + "step": 1589 + }, + { + "epoch": 0.03727210393994269, + "grad_norm": 0.28754159808158875, + "learning_rate": 0.00019932348733146747, + "loss": 0.0553, + "step": 1590 + }, + { + "epoch": 0.03729554551474768, + "grad_norm": 0.7846813797950745, + "learning_rate": 0.00019932263168370837, + "loss": 0.2229, + "step": 1591 + }, + { + "epoch": 0.037318987089552676, + "grad_norm": 0.46751701831817627, + "learning_rate": 0.0001993217754970219, + "loss": 0.1706, + "step": 1592 + }, + { + "epoch": 0.03734242866435767, + "grad_norm": 0.17591482400894165, + "learning_rate": 0.00019932091877141264, + "loss": 0.0443, + "step": 1593 + }, + { + "epoch": 0.037365870239162664, + "grad_norm": 0.475734680891037, + "learning_rate": 0.0001993200615068853, + "loss": 0.1087, + "step": 1594 + }, + { + "epoch": 0.037389311813967666, + "grad_norm": 0.43333572149276733, + "learning_rate": 0.00019931920370344451, + "loss": 0.148, + "step": 1595 + }, + { + "epoch": 0.03741275338877266, + "grad_norm": 0.6129709482192993, + "learning_rate": 0.00019931834536109492, + "loss": 0.7946, + "step": 1596 + }, + { + "epoch": 0.037436194963577654, + "grad_norm": 0.6196077466011047, + "learning_rate": 0.0001993174864798412, + "loss": 0.1957, + "step": 1597 + }, + { + "epoch": 0.03745963653838265, + "grad_norm": 0.6119499206542969, + "learning_rate": 0.000199316627059688, + "loss": 0.8694, + "step": 1598 + }, + { + "epoch": 0.03748307811318764, + "grad_norm": 0.5263822078704834, + "learning_rate": 0.00019931576710064002, + "loss": 0.1069, + "step": 1599 + }, + { + "epoch": 0.03750651968799264, + "grad_norm": 0.6613817811012268, + "learning_rate": 0.00019931490660270187, + "loss": 0.1713, + "step": 1600 + }, + { + "epoch": 0.03752996126279763, + "grad_norm": 0.8510591387748718, + "learning_rate": 0.00019931404556587824, + "loss": 0.2411, + "step": 1601 + }, + { + "epoch": 0.03755340283760263, + "grad_norm": 0.514175534248352, + "learning_rate": 0.00019931318399017383, + "loss": 0.6791, + "step": 1602 + }, + { + "epoch": 0.03757684441240763, + "grad_norm": 0.44856610894203186, + "learning_rate": 0.00019931232187559328, + "loss": 0.1542, + "step": 1603 + }, + { + "epoch": 0.03760028598721262, + "grad_norm": 0.44843825697898865, + "learning_rate": 0.00019931145922214127, + "loss": 0.0498, + "step": 1604 + }, + { + "epoch": 0.037623727562017616, + "grad_norm": 0.22615213692188263, + "learning_rate": 0.0001993105960298225, + "loss": 0.0781, + "step": 1605 + }, + { + "epoch": 0.03764716913682261, + "grad_norm": 0.6839895844459534, + "learning_rate": 0.00019930973229864162, + "loss": 0.2025, + "step": 1606 + }, + { + "epoch": 0.037670610711627604, + "grad_norm": 0.2365608662366867, + "learning_rate": 0.00019930886802860335, + "loss": 0.0545, + "step": 1607 + }, + { + "epoch": 0.037694052286432606, + "grad_norm": 0.2703697681427002, + "learning_rate": 0.0001993080032197124, + "loss": 0.0868, + "step": 1608 + }, + { + "epoch": 0.0377174938612376, + "grad_norm": 0.6457381248474121, + "learning_rate": 0.00019930713787197338, + "loss": 0.2317, + "step": 1609 + }, + { + "epoch": 0.037740935436042594, + "grad_norm": 0.36271512508392334, + "learning_rate": 0.00019930627198539108, + "loss": 0.1164, + "step": 1610 + }, + { + "epoch": 0.03776437701084759, + "grad_norm": 0.8240512609481812, + "learning_rate": 0.00019930540555997017, + "loss": 0.212, + "step": 1611 + }, + { + "epoch": 0.03778781858565258, + "grad_norm": 0.6675083041191101, + "learning_rate": 0.00019930453859571528, + "loss": 0.198, + "step": 1612 + }, + { + "epoch": 0.03781126016045758, + "grad_norm": 0.7259770035743713, + "learning_rate": 0.0001993036710926312, + "loss": 0.2975, + "step": 1613 + }, + { + "epoch": 0.03783470173526258, + "grad_norm": 0.6814106106758118, + "learning_rate": 0.00019930280305072265, + "loss": 0.2541, + "step": 1614 + }, + { + "epoch": 0.03785814331006757, + "grad_norm": 0.28688302636146545, + "learning_rate": 0.00019930193446999423, + "loss": 0.4265, + "step": 1615 + }, + { + "epoch": 0.03788158488487257, + "grad_norm": 0.36137011647224426, + "learning_rate": 0.00019930106535045075, + "loss": 0.1233, + "step": 1616 + }, + { + "epoch": 0.03790502645967756, + "grad_norm": 0.4859457314014435, + "learning_rate": 0.00019930019569209693, + "loss": 0.1639, + "step": 1617 + }, + { + "epoch": 0.037928468034482556, + "grad_norm": 0.4534970819950104, + "learning_rate": 0.0001992993254949374, + "loss": 0.6694, + "step": 1618 + }, + { + "epoch": 0.03795190960928755, + "grad_norm": 0.10828510671854019, + "learning_rate": 0.00019929845475897696, + "loss": 0.031, + "step": 1619 + }, + { + "epoch": 0.037975351184092544, + "grad_norm": 0.5528334379196167, + "learning_rate": 0.00019929758348422033, + "loss": 0.1884, + "step": 1620 + }, + { + "epoch": 0.037998792758897545, + "grad_norm": 0.2594635486602783, + "learning_rate": 0.0001992967116706722, + "loss": 0.125, + "step": 1621 + }, + { + "epoch": 0.03802223433370254, + "grad_norm": 0.19368258118629456, + "learning_rate": 0.00019929583931833732, + "loss": 0.0809, + "step": 1622 + }, + { + "epoch": 0.038045675908507534, + "grad_norm": 0.7335733771324158, + "learning_rate": 0.00019929496642722043, + "loss": 0.2802, + "step": 1623 + }, + { + "epoch": 0.03806911748331253, + "grad_norm": 0.2118058055639267, + "learning_rate": 0.00019929409299732626, + "loss": 0.1009, + "step": 1624 + }, + { + "epoch": 0.03809255905811752, + "grad_norm": 0.24377822875976562, + "learning_rate": 0.00019929321902865957, + "loss": 0.0864, + "step": 1625 + }, + { + "epoch": 0.03811600063292252, + "grad_norm": 0.20871761441230774, + "learning_rate": 0.00019929234452122505, + "loss": 0.1163, + "step": 1626 + }, + { + "epoch": 0.03813944220772752, + "grad_norm": 0.4525724947452545, + "learning_rate": 0.00019929146947502747, + "loss": 0.2446, + "step": 1627 + }, + { + "epoch": 0.03816288378253251, + "grad_norm": 0.1893165111541748, + "learning_rate": 0.0001992905938900716, + "loss": 0.0877, + "step": 1628 + }, + { + "epoch": 0.03818632535733751, + "grad_norm": 0.45095178484916687, + "learning_rate": 0.00019928971776636217, + "loss": 0.0666, + "step": 1629 + }, + { + "epoch": 0.0382097669321425, + "grad_norm": 0.729732871055603, + "learning_rate": 0.00019928884110390397, + "loss": 0.2206, + "step": 1630 + }, + { + "epoch": 0.038233208506947496, + "grad_norm": 0.32675325870513916, + "learning_rate": 0.0001992879639027017, + "loss": 0.0668, + "step": 1631 + }, + { + "epoch": 0.03825665008175249, + "grad_norm": 0.3338560163974762, + "learning_rate": 0.00019928708616276015, + "loss": 0.0825, + "step": 1632 + }, + { + "epoch": 0.038280091656557484, + "grad_norm": 0.40583157539367676, + "learning_rate": 0.00019928620788408407, + "loss": 0.1558, + "step": 1633 + }, + { + "epoch": 0.038303533231362485, + "grad_norm": 0.6264913082122803, + "learning_rate": 0.00019928532906667826, + "loss": 0.7745, + "step": 1634 + }, + { + "epoch": 0.03832697480616748, + "grad_norm": 0.5317113399505615, + "learning_rate": 0.00019928444971054746, + "loss": 0.1688, + "step": 1635 + }, + { + "epoch": 0.038350416380972474, + "grad_norm": 0.5262113213539124, + "learning_rate": 0.00019928356981569641, + "loss": 0.1039, + "step": 1636 + }, + { + "epoch": 0.03837385795577747, + "grad_norm": 0.25725647807121277, + "learning_rate": 0.00019928268938212992, + "loss": 0.0993, + "step": 1637 + }, + { + "epoch": 0.03839729953058246, + "grad_norm": 0.6527292132377625, + "learning_rate": 0.0001992818084098528, + "loss": 0.2919, + "step": 1638 + }, + { + "epoch": 0.03842074110538746, + "grad_norm": 0.49602261185646057, + "learning_rate": 0.00019928092689886978, + "loss": 0.1449, + "step": 1639 + }, + { + "epoch": 0.03844418268019246, + "grad_norm": 0.6044578552246094, + "learning_rate": 0.00019928004484918565, + "loss": 0.2653, + "step": 1640 + }, + { + "epoch": 0.03846762425499745, + "grad_norm": 0.6067953109741211, + "learning_rate": 0.00019927916226080523, + "loss": 0.2601, + "step": 1641 + }, + { + "epoch": 0.03849106582980245, + "grad_norm": 0.6558887958526611, + "learning_rate": 0.00019927827913373327, + "loss": 0.2257, + "step": 1642 + }, + { + "epoch": 0.03851450740460744, + "grad_norm": 0.4780043661594391, + "learning_rate": 0.00019927739546797457, + "loss": 0.1943, + "step": 1643 + }, + { + "epoch": 0.038537948979412436, + "grad_norm": 0.18573758006095886, + "learning_rate": 0.00019927651126353392, + "loss": 0.0344, + "step": 1644 + }, + { + "epoch": 0.03856139055421743, + "grad_norm": 0.41209444403648376, + "learning_rate": 0.00019927562652041615, + "loss": 0.1009, + "step": 1645 + }, + { + "epoch": 0.03858483212902243, + "grad_norm": 0.4514060914516449, + "learning_rate": 0.00019927474123862605, + "loss": 0.1116, + "step": 1646 + }, + { + "epoch": 0.038608273703827425, + "grad_norm": 0.5895359516143799, + "learning_rate": 0.00019927385541816838, + "loss": 0.2794, + "step": 1647 + }, + { + "epoch": 0.03863171527863242, + "grad_norm": 0.1901797652244568, + "learning_rate": 0.00019927296905904803, + "loss": 0.0716, + "step": 1648 + }, + { + "epoch": 0.038655156853437414, + "grad_norm": 0.6226240396499634, + "learning_rate": 0.0001992720821612697, + "loss": 0.191, + "step": 1649 + }, + { + "epoch": 0.03867859842824241, + "grad_norm": 0.6236171126365662, + "learning_rate": 0.00019927119472483827, + "loss": 0.2724, + "step": 1650 + }, + { + "epoch": 0.0387020400030474, + "grad_norm": 0.13706813752651215, + "learning_rate": 0.00019927030674975858, + "loss": 0.0703, + "step": 1651 + }, + { + "epoch": 0.0387254815778524, + "grad_norm": 0.4480455815792084, + "learning_rate": 0.00019926941823603537, + "loss": 0.1112, + "step": 1652 + }, + { + "epoch": 0.0387489231526574, + "grad_norm": 0.32004448771476746, + "learning_rate": 0.00019926852918367356, + "loss": 0.0979, + "step": 1653 + }, + { + "epoch": 0.03877236472746239, + "grad_norm": 0.33769190311431885, + "learning_rate": 0.00019926763959267788, + "loss": 0.088, + "step": 1654 + }, + { + "epoch": 0.03879580630226739, + "grad_norm": 0.381352037191391, + "learning_rate": 0.00019926674946305322, + "loss": 0.6034, + "step": 1655 + }, + { + "epoch": 0.03881924787707238, + "grad_norm": 0.7063488960266113, + "learning_rate": 0.00019926585879480438, + "loss": 0.1709, + "step": 1656 + }, + { + "epoch": 0.038842689451877375, + "grad_norm": 0.399135023355484, + "learning_rate": 0.00019926496758793615, + "loss": 0.1452, + "step": 1657 + }, + { + "epoch": 0.03886613102668237, + "grad_norm": 0.5358358025550842, + "learning_rate": 0.0001992640758424535, + "loss": 0.1187, + "step": 1658 + }, + { + "epoch": 0.03888957260148737, + "grad_norm": 0.6456049680709839, + "learning_rate": 0.00019926318355836112, + "loss": 0.1522, + "step": 1659 + }, + { + "epoch": 0.038913014176292365, + "grad_norm": 0.7441787719726562, + "learning_rate": 0.00019926229073566394, + "loss": 0.2277, + "step": 1660 + }, + { + "epoch": 0.03893645575109736, + "grad_norm": 0.4388434588909149, + "learning_rate": 0.00019926139737436676, + "loss": 0.0759, + "step": 1661 + }, + { + "epoch": 0.038959897325902354, + "grad_norm": 0.6938778162002563, + "learning_rate": 0.00019926050347447447, + "loss": 0.2175, + "step": 1662 + }, + { + "epoch": 0.03898333890070735, + "grad_norm": 0.16479900479316711, + "learning_rate": 0.0001992596090359919, + "loss": 0.0315, + "step": 1663 + }, + { + "epoch": 0.03900678047551234, + "grad_norm": 0.33091527223587036, + "learning_rate": 0.0001992587140589239, + "loss": 0.0959, + "step": 1664 + }, + { + "epoch": 0.03903022205031734, + "grad_norm": 0.5634067058563232, + "learning_rate": 0.00019925781854327528, + "loss": 0.1853, + "step": 1665 + }, + { + "epoch": 0.03905366362512234, + "grad_norm": 0.6287477612495422, + "learning_rate": 0.000199256922489051, + "loss": 0.2071, + "step": 1666 + }, + { + "epoch": 0.03907710519992733, + "grad_norm": 0.29043883085250854, + "learning_rate": 0.00019925602589625582, + "loss": 0.0883, + "step": 1667 + }, + { + "epoch": 0.03910054677473233, + "grad_norm": 0.6233084797859192, + "learning_rate": 0.0001992551287648947, + "loss": 0.193, + "step": 1668 + }, + { + "epoch": 0.03912398834953732, + "grad_norm": 0.20771604776382446, + "learning_rate": 0.00019925423109497245, + "loss": 0.0585, + "step": 1669 + }, + { + "epoch": 0.039147429924342315, + "grad_norm": 0.6694942712783813, + "learning_rate": 0.00019925333288649393, + "loss": 0.136, + "step": 1670 + }, + { + "epoch": 0.03917087149914731, + "grad_norm": 0.770063042640686, + "learning_rate": 0.00019925243413946407, + "loss": 0.2608, + "step": 1671 + }, + { + "epoch": 0.03919431307395231, + "grad_norm": 0.5700552463531494, + "learning_rate": 0.00019925153485388767, + "loss": 0.1797, + "step": 1672 + }, + { + "epoch": 0.039217754648757305, + "grad_norm": 0.2982397973537445, + "learning_rate": 0.00019925063502976967, + "loss": 0.0814, + "step": 1673 + }, + { + "epoch": 0.0392411962235623, + "grad_norm": 0.49938786029815674, + "learning_rate": 0.00019924973466711493, + "loss": 0.1434, + "step": 1674 + }, + { + "epoch": 0.039264637798367294, + "grad_norm": 0.8002720475196838, + "learning_rate": 0.00019924883376592836, + "loss": 0.2314, + "step": 1675 + }, + { + "epoch": 0.03928807937317229, + "grad_norm": 0.5844964981079102, + "learning_rate": 0.00019924793232621482, + "loss": 0.2043, + "step": 1676 + }, + { + "epoch": 0.03931152094797728, + "grad_norm": 0.6919118165969849, + "learning_rate": 0.0001992470303479792, + "loss": 0.1378, + "step": 1677 + }, + { + "epoch": 0.039334962522782284, + "grad_norm": 0.49743103981018066, + "learning_rate": 0.0001992461278312264, + "loss": 0.1794, + "step": 1678 + }, + { + "epoch": 0.03935840409758728, + "grad_norm": 0.5044227242469788, + "learning_rate": 0.0001992452247759613, + "loss": 0.1781, + "step": 1679 + }, + { + "epoch": 0.03938184567239227, + "grad_norm": 0.5732187032699585, + "learning_rate": 0.00019924432118218884, + "loss": 0.2197, + "step": 1680 + }, + { + "epoch": 0.03940528724719727, + "grad_norm": 0.451901376247406, + "learning_rate": 0.0001992434170499139, + "loss": 0.1528, + "step": 1681 + }, + { + "epoch": 0.03942872882200226, + "grad_norm": 0.5129866003990173, + "learning_rate": 0.0001992425123791414, + "loss": 0.8098, + "step": 1682 + }, + { + "epoch": 0.039452170396807255, + "grad_norm": 0.6314334869384766, + "learning_rate": 0.00019924160716987627, + "loss": 0.7687, + "step": 1683 + }, + { + "epoch": 0.03947561197161225, + "grad_norm": 0.5465731620788574, + "learning_rate": 0.00019924070142212332, + "loss": 0.2139, + "step": 1684 + }, + { + "epoch": 0.03949905354641725, + "grad_norm": 0.4384244978427887, + "learning_rate": 0.0001992397951358876, + "loss": 0.0931, + "step": 1685 + }, + { + "epoch": 0.039522495121222245, + "grad_norm": 0.710204541683197, + "learning_rate": 0.0001992388883111739, + "loss": 0.2225, + "step": 1686 + }, + { + "epoch": 0.03954593669602724, + "grad_norm": 0.48375579714775085, + "learning_rate": 0.00019923798094798723, + "loss": 0.1746, + "step": 1687 + }, + { + "epoch": 0.039569378270832234, + "grad_norm": 0.7485866546630859, + "learning_rate": 0.00019923707304633247, + "loss": 0.2278, + "step": 1688 + }, + { + "epoch": 0.03959281984563723, + "grad_norm": 0.6946219801902771, + "learning_rate": 0.00019923616460621458, + "loss": 0.7949, + "step": 1689 + }, + { + "epoch": 0.03961626142044222, + "grad_norm": 0.16691404581069946, + "learning_rate": 0.00019923525562763844, + "loss": 0.0399, + "step": 1690 + }, + { + "epoch": 0.039639702995247224, + "grad_norm": 0.5346314311027527, + "learning_rate": 0.00019923434611060907, + "loss": 0.6364, + "step": 1691 + }, + { + "epoch": 0.03966314457005222, + "grad_norm": 0.18761007487773895, + "learning_rate": 0.0001992334360551313, + "loss": 0.0512, + "step": 1692 + }, + { + "epoch": 0.03968658614485721, + "grad_norm": 0.5926586389541626, + "learning_rate": 0.00019923252546121014, + "loss": 0.2367, + "step": 1693 + }, + { + "epoch": 0.03971002771966221, + "grad_norm": 0.49482643604278564, + "learning_rate": 0.0001992316143288505, + "loss": 0.1672, + "step": 1694 + }, + { + "epoch": 0.0397334692944672, + "grad_norm": 0.7799727320671082, + "learning_rate": 0.0001992307026580573, + "loss": 0.1834, + "step": 1695 + }, + { + "epoch": 0.039756910869272195, + "grad_norm": 0.3025036156177521, + "learning_rate": 0.00019922979044883555, + "loss": 0.0891, + "step": 1696 + }, + { + "epoch": 0.03978035244407719, + "grad_norm": 0.3143922984600067, + "learning_rate": 0.00019922887770119015, + "loss": 0.1243, + "step": 1697 + }, + { + "epoch": 0.03980379401888219, + "grad_norm": 0.5592930912971497, + "learning_rate": 0.0001992279644151261, + "loss": 0.2082, + "step": 1698 + }, + { + "epoch": 0.039827235593687185, + "grad_norm": 0.5269222855567932, + "learning_rate": 0.0001992270505906483, + "loss": 0.2894, + "step": 1699 + }, + { + "epoch": 0.03985067716849218, + "grad_norm": 0.39612480998039246, + "learning_rate": 0.00019922613622776173, + "loss": 0.201, + "step": 1700 + }, + { + "epoch": 0.039874118743297174, + "grad_norm": 0.5695582032203674, + "learning_rate": 0.00019922522132647137, + "loss": 0.851, + "step": 1701 + }, + { + "epoch": 0.03989756031810217, + "grad_norm": 0.49594125151634216, + "learning_rate": 0.00019922430588678215, + "loss": 0.1685, + "step": 1702 + }, + { + "epoch": 0.03992100189290716, + "grad_norm": 0.13203534483909607, + "learning_rate": 0.00019922338990869908, + "loss": 0.047, + "step": 1703 + }, + { + "epoch": 0.039944443467712164, + "grad_norm": 0.31274324655532837, + "learning_rate": 0.0001992224733922271, + "loss": 0.0727, + "step": 1704 + }, + { + "epoch": 0.03996788504251716, + "grad_norm": 0.6376890540122986, + "learning_rate": 0.00019922155633737121, + "loss": 0.6533, + "step": 1705 + }, + { + "epoch": 0.03999132661732215, + "grad_norm": 0.7118820548057556, + "learning_rate": 0.0001992206387441363, + "loss": 0.1456, + "step": 1706 + }, + { + "epoch": 0.04001476819212715, + "grad_norm": 0.12975150346755981, + "learning_rate": 0.00019921972061252746, + "loss": 0.0392, + "step": 1707 + }, + { + "epoch": 0.04003820976693214, + "grad_norm": 0.5486299395561218, + "learning_rate": 0.00019921880194254964, + "loss": 0.1366, + "step": 1708 + }, + { + "epoch": 0.040061651341737135, + "grad_norm": 0.7750927805900574, + "learning_rate": 0.0001992178827342078, + "loss": 0.1913, + "step": 1709 + }, + { + "epoch": 0.04008509291654214, + "grad_norm": 0.38043850660324097, + "learning_rate": 0.00019921696298750693, + "loss": 0.175, + "step": 1710 + }, + { + "epoch": 0.04010853449134713, + "grad_norm": 0.5204871892929077, + "learning_rate": 0.00019921604270245203, + "loss": 0.1816, + "step": 1711 + }, + { + "epoch": 0.040131976066152125, + "grad_norm": 0.6216399073600769, + "learning_rate": 0.00019921512187904807, + "loss": 0.1872, + "step": 1712 + }, + { + "epoch": 0.04015541764095712, + "grad_norm": 0.824190080165863, + "learning_rate": 0.0001992142005173001, + "loss": 0.2576, + "step": 1713 + }, + { + "epoch": 0.040178859215762114, + "grad_norm": 0.46807053685188293, + "learning_rate": 0.00019921327861721307, + "loss": 0.1069, + "step": 1714 + }, + { + "epoch": 0.04020230079056711, + "grad_norm": 0.47530075907707214, + "learning_rate": 0.00019921235617879202, + "loss": 0.1032, + "step": 1715 + }, + { + "epoch": 0.0402257423653721, + "grad_norm": 0.5137298107147217, + "learning_rate": 0.00019921143320204192, + "loss": 0.1317, + "step": 1716 + }, + { + "epoch": 0.040249183940177104, + "grad_norm": 0.3378320038318634, + "learning_rate": 0.00019921050968696783, + "loss": 0.0854, + "step": 1717 + }, + { + "epoch": 0.0402726255149821, + "grad_norm": 0.5448839068412781, + "learning_rate": 0.00019920958563357467, + "loss": 0.1952, + "step": 1718 + }, + { + "epoch": 0.04029606708978709, + "grad_norm": 0.5422284007072449, + "learning_rate": 0.00019920866104186752, + "loss": 0.1536, + "step": 1719 + }, + { + "epoch": 0.04031950866459209, + "grad_norm": 0.6034518480300903, + "learning_rate": 0.00019920773591185141, + "loss": 0.2058, + "step": 1720 + }, + { + "epoch": 0.04034295023939708, + "grad_norm": 0.7414153218269348, + "learning_rate": 0.00019920681024353134, + "loss": 0.2116, + "step": 1721 + }, + { + "epoch": 0.040366391814202075, + "grad_norm": 0.6217705607414246, + "learning_rate": 0.00019920588403691232, + "loss": 0.1772, + "step": 1722 + }, + { + "epoch": 0.04038983338900708, + "grad_norm": 0.32548192143440247, + "learning_rate": 0.00019920495729199939, + "loss": 0.0829, + "step": 1723 + }, + { + "epoch": 0.04041327496381207, + "grad_norm": 0.6723328232765198, + "learning_rate": 0.00019920403000879753, + "loss": 0.1885, + "step": 1724 + }, + { + "epoch": 0.040436716538617065, + "grad_norm": 0.4960329830646515, + "learning_rate": 0.00019920310218731185, + "loss": 0.1007, + "step": 1725 + }, + { + "epoch": 0.04046015811342206, + "grad_norm": 1.0251127481460571, + "learning_rate": 0.00019920217382754735, + "loss": 0.1968, + "step": 1726 + }, + { + "epoch": 0.040483599688227054, + "grad_norm": 0.9479556083679199, + "learning_rate": 0.00019920124492950905, + "loss": 0.5095, + "step": 1727 + }, + { + "epoch": 0.04050704126303205, + "grad_norm": 0.4221051335334778, + "learning_rate": 0.00019920031549320203, + "loss": 0.1178, + "step": 1728 + }, + { + "epoch": 0.04053048283783705, + "grad_norm": 0.9238879680633545, + "learning_rate": 0.0001991993855186313, + "loss": 0.1845, + "step": 1729 + }, + { + "epoch": 0.040553924412642044, + "grad_norm": 0.8250933885574341, + "learning_rate": 0.00019919845500580187, + "loss": 0.2244, + "step": 1730 + }, + { + "epoch": 0.04057736598744704, + "grad_norm": 0.604939341545105, + "learning_rate": 0.00019919752395471888, + "loss": 0.1388, + "step": 1731 + }, + { + "epoch": 0.04060080756225203, + "grad_norm": 0.5638744831085205, + "learning_rate": 0.00019919659236538733, + "loss": 0.2219, + "step": 1732 + }, + { + "epoch": 0.04062424913705703, + "grad_norm": 0.5629063844680786, + "learning_rate": 0.00019919566023781228, + "loss": 0.1319, + "step": 1733 + }, + { + "epoch": 0.04064769071186202, + "grad_norm": 0.4719081223011017, + "learning_rate": 0.00019919472757199877, + "loss": 0.1262, + "step": 1734 + }, + { + "epoch": 0.040671132286667015, + "grad_norm": 0.6441231369972229, + "learning_rate": 0.0001991937943679519, + "loss": 0.1583, + "step": 1735 + }, + { + "epoch": 0.040694573861472016, + "grad_norm": 0.48827409744262695, + "learning_rate": 0.0001991928606256767, + "loss": 0.2341, + "step": 1736 + }, + { + "epoch": 0.04071801543627701, + "grad_norm": 0.6976703405380249, + "learning_rate": 0.00019919192634517826, + "loss": 0.1925, + "step": 1737 + }, + { + "epoch": 0.040741457011082005, + "grad_norm": 0.3677330017089844, + "learning_rate": 0.00019919099152646163, + "loss": 0.1616, + "step": 1738 + }, + { + "epoch": 0.040764898585887, + "grad_norm": 0.20517286658287048, + "learning_rate": 0.0001991900561695319, + "loss": 0.061, + "step": 1739 + }, + { + "epoch": 0.040788340160691994, + "grad_norm": 0.28854313492774963, + "learning_rate": 0.00019918912027439413, + "loss": 0.0664, + "step": 1740 + }, + { + "epoch": 0.04081178173549699, + "grad_norm": 0.6391483545303345, + "learning_rate": 0.00019918818384105343, + "loss": 0.2486, + "step": 1741 + }, + { + "epoch": 0.04083522331030199, + "grad_norm": 0.13423945009708405, + "learning_rate": 0.0001991872468695148, + "loss": 0.0645, + "step": 1742 + }, + { + "epoch": 0.040858664885106984, + "grad_norm": 0.5481492877006531, + "learning_rate": 0.00019918630935978344, + "loss": 0.1895, + "step": 1743 + }, + { + "epoch": 0.04088210645991198, + "grad_norm": 0.5315002799034119, + "learning_rate": 0.0001991853713118643, + "loss": 0.1189, + "step": 1744 + }, + { + "epoch": 0.04090554803471697, + "grad_norm": 0.3325328528881073, + "learning_rate": 0.00019918443272576265, + "loss": 0.0633, + "step": 1745 + }, + { + "epoch": 0.04092898960952197, + "grad_norm": 0.5043154358863831, + "learning_rate": 0.0001991834936014834, + "loss": 0.1418, + "step": 1746 + }, + { + "epoch": 0.04095243118432696, + "grad_norm": 0.4649564027786255, + "learning_rate": 0.00019918255393903175, + "loss": 0.242, + "step": 1747 + }, + { + "epoch": 0.040975872759131955, + "grad_norm": 0.6542818546295166, + "learning_rate": 0.00019918161373841278, + "loss": 0.2345, + "step": 1748 + }, + { + "epoch": 0.040999314333936956, + "grad_norm": 0.45570993423461914, + "learning_rate": 0.00019918067299963156, + "loss": 0.1728, + "step": 1749 + }, + { + "epoch": 0.04102275590874195, + "grad_norm": 0.24479039013385773, + "learning_rate": 0.00019917973172269325, + "loss": 0.0466, + "step": 1750 + }, + { + "epoch": 0.041046197483546945, + "grad_norm": 0.31829947233200073, + "learning_rate": 0.00019917878990760292, + "loss": 0.1007, + "step": 1751 + }, + { + "epoch": 0.04106963905835194, + "grad_norm": 0.557985782623291, + "learning_rate": 0.00019917784755436565, + "loss": 0.0663, + "step": 1752 + }, + { + "epoch": 0.041093080633156934, + "grad_norm": 0.608532190322876, + "learning_rate": 0.00019917690466298663, + "loss": 0.1793, + "step": 1753 + }, + { + "epoch": 0.04111652220796193, + "grad_norm": 0.8442514538764954, + "learning_rate": 0.00019917596123347094, + "loss": 0.2233, + "step": 1754 + }, + { + "epoch": 0.04113996378276693, + "grad_norm": 0.43332818150520325, + "learning_rate": 0.00019917501726582368, + "loss": 0.0837, + "step": 1755 + }, + { + "epoch": 0.041163405357571924, + "grad_norm": 0.7713133692741394, + "learning_rate": 0.00019917407276005, + "loss": 0.2654, + "step": 1756 + }, + { + "epoch": 0.04118684693237692, + "grad_norm": 0.8558149933815002, + "learning_rate": 0.000199173127716155, + "loss": 0.2393, + "step": 1757 + }, + { + "epoch": 0.04121028850718191, + "grad_norm": 0.9868559241294861, + "learning_rate": 0.00019917218213414382, + "loss": 0.2026, + "step": 1758 + }, + { + "epoch": 0.04123373008198691, + "grad_norm": 0.7058619260787964, + "learning_rate": 0.0001991712360140216, + "loss": 0.3267, + "step": 1759 + }, + { + "epoch": 0.0412571716567919, + "grad_norm": 0.6148595213890076, + "learning_rate": 0.00019917028935579345, + "loss": 0.1669, + "step": 1760 + }, + { + "epoch": 0.0412806132315969, + "grad_norm": 0.2843749523162842, + "learning_rate": 0.00019916934215946454, + "loss": 0.0799, + "step": 1761 + }, + { + "epoch": 0.041304054806401896, + "grad_norm": 0.404199481010437, + "learning_rate": 0.00019916839442504, + "loss": 0.1019, + "step": 1762 + }, + { + "epoch": 0.04132749638120689, + "grad_norm": 0.42404699325561523, + "learning_rate": 0.00019916744615252493, + "loss": 0.1111, + "step": 1763 + }, + { + "epoch": 0.041350937956011885, + "grad_norm": 0.506147563457489, + "learning_rate": 0.00019916649734192453, + "loss": 0.1249, + "step": 1764 + }, + { + "epoch": 0.04137437953081688, + "grad_norm": 0.7014893889427185, + "learning_rate": 0.00019916554799324394, + "loss": 0.212, + "step": 1765 + }, + { + "epoch": 0.041397821105621874, + "grad_norm": 0.49428993463516235, + "learning_rate": 0.00019916459810648827, + "loss": 0.1467, + "step": 1766 + }, + { + "epoch": 0.04142126268042687, + "grad_norm": 0.28411629796028137, + "learning_rate": 0.0001991636476816627, + "loss": 0.1166, + "step": 1767 + }, + { + "epoch": 0.04144470425523187, + "grad_norm": 0.2608960270881653, + "learning_rate": 0.00019916269671877243, + "loss": 0.1023, + "step": 1768 + }, + { + "epoch": 0.041468145830036864, + "grad_norm": 0.67303067445755, + "learning_rate": 0.00019916174521782254, + "loss": 0.1981, + "step": 1769 + }, + { + "epoch": 0.04149158740484186, + "grad_norm": 0.6723349094390869, + "learning_rate": 0.00019916079317881824, + "loss": 0.1973, + "step": 1770 + }, + { + "epoch": 0.04151502897964685, + "grad_norm": 0.10023821145296097, + "learning_rate": 0.0001991598406017647, + "loss": 0.0316, + "step": 1771 + }, + { + "epoch": 0.041538470554451846, + "grad_norm": 0.13553085923194885, + "learning_rate": 0.00019915888748666706, + "loss": 0.0274, + "step": 1772 + }, + { + "epoch": 0.04156191212925684, + "grad_norm": 0.2229439914226532, + "learning_rate": 0.0001991579338335305, + "loss": 0.0919, + "step": 1773 + }, + { + "epoch": 0.04158535370406184, + "grad_norm": 0.6054831743240356, + "learning_rate": 0.00019915697964236024, + "loss": 0.2576, + "step": 1774 + }, + { + "epoch": 0.041608795278866836, + "grad_norm": 0.6266111135482788, + "learning_rate": 0.0001991560249131614, + "loss": 0.1731, + "step": 1775 + }, + { + "epoch": 0.04163223685367183, + "grad_norm": 0.7065443396568298, + "learning_rate": 0.00019915506964593918, + "loss": 0.1791, + "step": 1776 + }, + { + "epoch": 0.041655678428476825, + "grad_norm": 0.7078781723976135, + "learning_rate": 0.00019915411384069877, + "loss": 0.1739, + "step": 1777 + }, + { + "epoch": 0.04167912000328182, + "grad_norm": 0.19526322185993195, + "learning_rate": 0.00019915315749744535, + "loss": 0.0478, + "step": 1778 + }, + { + "epoch": 0.041702561578086814, + "grad_norm": 0.2569836676120758, + "learning_rate": 0.0001991522006161841, + "loss": 0.062, + "step": 1779 + }, + { + "epoch": 0.04172600315289181, + "grad_norm": 0.46231094002723694, + "learning_rate": 0.00019915124319692023, + "loss": 0.2142, + "step": 1780 + }, + { + "epoch": 0.04174944472769681, + "grad_norm": 0.57157963514328, + "learning_rate": 0.00019915028523965892, + "loss": 0.7391, + "step": 1781 + }, + { + "epoch": 0.041772886302501804, + "grad_norm": 0.7306718230247498, + "learning_rate": 0.00019914932674440538, + "loss": 0.2653, + "step": 1782 + }, + { + "epoch": 0.0417963278773068, + "grad_norm": 0.3274949789047241, + "learning_rate": 0.00019914836771116483, + "loss": 0.0988, + "step": 1783 + }, + { + "epoch": 0.04181976945211179, + "grad_norm": 0.21942639350891113, + "learning_rate": 0.00019914740813994238, + "loss": 0.0519, + "step": 1784 + }, + { + "epoch": 0.041843211026916786, + "grad_norm": 0.5905762910842896, + "learning_rate": 0.00019914644803074337, + "loss": 0.2079, + "step": 1785 + }, + { + "epoch": 0.04186665260172178, + "grad_norm": 0.8238871693611145, + "learning_rate": 0.00019914548738357292, + "loss": 0.2473, + "step": 1786 + }, + { + "epoch": 0.04189009417652678, + "grad_norm": 0.36299535632133484, + "learning_rate": 0.00019914452619843628, + "loss": 0.1125, + "step": 1787 + }, + { + "epoch": 0.041913535751331776, + "grad_norm": 0.36244997382164, + "learning_rate": 0.00019914356447533864, + "loss": 0.097, + "step": 1788 + }, + { + "epoch": 0.04193697732613677, + "grad_norm": 0.6816521883010864, + "learning_rate": 0.00019914260221428525, + "loss": 0.1932, + "step": 1789 + }, + { + "epoch": 0.041960418900941765, + "grad_norm": 0.7758482694625854, + "learning_rate": 0.00019914163941528128, + "loss": 0.1243, + "step": 1790 + }, + { + "epoch": 0.04198386047574676, + "grad_norm": 0.6706167459487915, + "learning_rate": 0.000199140676078332, + "loss": 0.7889, + "step": 1791 + }, + { + "epoch": 0.042007302050551754, + "grad_norm": 0.8989217877388, + "learning_rate": 0.00019913971220344266, + "loss": 0.1543, + "step": 1792 + }, + { + "epoch": 0.042030743625356755, + "grad_norm": 0.38498640060424805, + "learning_rate": 0.0001991387477906184, + "loss": 0.1084, + "step": 1793 + }, + { + "epoch": 0.04205418520016175, + "grad_norm": 0.5135555863380432, + "learning_rate": 0.00019913778283986452, + "loss": 0.2233, + "step": 1794 + }, + { + "epoch": 0.042077626774966743, + "grad_norm": 0.5050081014633179, + "learning_rate": 0.00019913681735118628, + "loss": 0.1502, + "step": 1795 + }, + { + "epoch": 0.04210106834977174, + "grad_norm": 0.5672056078910828, + "learning_rate": 0.00019913585132458886, + "loss": 0.6951, + "step": 1796 + }, + { + "epoch": 0.04212450992457673, + "grad_norm": 0.5607157349586487, + "learning_rate": 0.00019913488476007752, + "loss": 0.2054, + "step": 1797 + }, + { + "epoch": 0.042147951499381726, + "grad_norm": 0.20557183027267456, + "learning_rate": 0.0001991339176576575, + "loss": 0.0595, + "step": 1798 + }, + { + "epoch": 0.04217139307418672, + "grad_norm": 0.8417203426361084, + "learning_rate": 0.00019913295001733406, + "loss": 0.2211, + "step": 1799 + }, + { + "epoch": 0.04219483464899172, + "grad_norm": 0.5100123882293701, + "learning_rate": 0.00019913198183911247, + "loss": 0.2158, + "step": 1800 + }, + { + "epoch": 0.042218276223796716, + "grad_norm": 0.5250687599182129, + "learning_rate": 0.0001991310131229979, + "loss": 0.122, + "step": 1801 + }, + { + "epoch": 0.04224171779860171, + "grad_norm": 0.5690917372703552, + "learning_rate": 0.00019913004386899573, + "loss": 0.8431, + "step": 1802 + }, + { + "epoch": 0.042265159373406705, + "grad_norm": 0.18540935218334198, + "learning_rate": 0.0001991290740771111, + "loss": 0.0669, + "step": 1803 + }, + { + "epoch": 0.0422886009482117, + "grad_norm": 0.5862986445426941, + "learning_rate": 0.00019912810374734936, + "loss": 0.2046, + "step": 1804 + }, + { + "epoch": 0.042312042523016694, + "grad_norm": 0.5275234580039978, + "learning_rate": 0.00019912713287971573, + "loss": 0.1625, + "step": 1805 + }, + { + "epoch": 0.042335484097821695, + "grad_norm": 0.7181383371353149, + "learning_rate": 0.00019912616147421553, + "loss": 0.6641, + "step": 1806 + }, + { + "epoch": 0.04235892567262669, + "grad_norm": 0.21302586793899536, + "learning_rate": 0.00019912518953085393, + "loss": 0.0771, + "step": 1807 + }, + { + "epoch": 0.04238236724743168, + "grad_norm": 0.8797650933265686, + "learning_rate": 0.0001991242170496363, + "loss": 0.1649, + "step": 1808 + }, + { + "epoch": 0.04240580882223668, + "grad_norm": 0.34292933344841003, + "learning_rate": 0.00019912324403056789, + "loss": 0.095, + "step": 1809 + }, + { + "epoch": 0.04242925039704167, + "grad_norm": 0.8133600354194641, + "learning_rate": 0.00019912227047365393, + "loss": 0.86, + "step": 1810 + }, + { + "epoch": 0.042452691971846666, + "grad_norm": 1.178648591041565, + "learning_rate": 0.00019912129637889977, + "loss": 0.3137, + "step": 1811 + }, + { + "epoch": 0.04247613354665166, + "grad_norm": 0.147113636136055, + "learning_rate": 0.00019912032174631067, + "loss": 0.0437, + "step": 1812 + }, + { + "epoch": 0.04249957512145666, + "grad_norm": 0.6738823056221008, + "learning_rate": 0.00019911934657589192, + "loss": 0.2349, + "step": 1813 + }, + { + "epoch": 0.042523016696261656, + "grad_norm": 0.5004909634590149, + "learning_rate": 0.00019911837086764877, + "loss": 0.1477, + "step": 1814 + }, + { + "epoch": 0.04254645827106665, + "grad_norm": 0.6581906676292419, + "learning_rate": 0.0001991173946215866, + "loss": 0.1869, + "step": 1815 + }, + { + "epoch": 0.042569899845871645, + "grad_norm": 0.7514281868934631, + "learning_rate": 0.00019911641783771063, + "loss": 0.1816, + "step": 1816 + }, + { + "epoch": 0.04259334142067664, + "grad_norm": 0.785004734992981, + "learning_rate": 0.00019911544051602622, + "loss": 0.1656, + "step": 1817 + }, + { + "epoch": 0.042616782995481634, + "grad_norm": 0.7024988532066345, + "learning_rate": 0.00019911446265653861, + "loss": 0.2562, + "step": 1818 + }, + { + "epoch": 0.042640224570286635, + "grad_norm": 0.5778740048408508, + "learning_rate": 0.00019911348425925317, + "loss": 0.1536, + "step": 1819 + }, + { + "epoch": 0.04266366614509163, + "grad_norm": 1.1179652214050293, + "learning_rate": 0.00019911250532417512, + "loss": 0.1745, + "step": 1820 + }, + { + "epoch": 0.04268710771989662, + "grad_norm": 0.44877511262893677, + "learning_rate": 0.0001991115258513099, + "loss": 0.1464, + "step": 1821 + }, + { + "epoch": 0.04271054929470162, + "grad_norm": 0.5141311287879944, + "learning_rate": 0.00019911054584066274, + "loss": 0.0869, + "step": 1822 + }, + { + "epoch": 0.04273399086950661, + "grad_norm": 0.6576289534568787, + "learning_rate": 0.00019910956529223893, + "loss": 0.2733, + "step": 1823 + }, + { + "epoch": 0.042757432444311606, + "grad_norm": 0.7121341228485107, + "learning_rate": 0.00019910858420604387, + "loss": 0.1783, + "step": 1824 + }, + { + "epoch": 0.04278087401911661, + "grad_norm": 0.28353652358055115, + "learning_rate": 0.00019910760258208283, + "loss": 0.0833, + "step": 1825 + }, + { + "epoch": 0.0428043155939216, + "grad_norm": 0.17091436684131622, + "learning_rate": 0.00019910662042036116, + "loss": 0.0469, + "step": 1826 + }, + { + "epoch": 0.042827757168726596, + "grad_norm": 0.4646387994289398, + "learning_rate": 0.0001991056377208842, + "loss": 0.1254, + "step": 1827 + }, + { + "epoch": 0.04285119874353159, + "grad_norm": 0.212777242064476, + "learning_rate": 0.00019910465448365723, + "loss": 0.0852, + "step": 1828 + }, + { + "epoch": 0.042874640318336585, + "grad_norm": 0.5436779856681824, + "learning_rate": 0.00019910367070868562, + "loss": 0.1542, + "step": 1829 + }, + { + "epoch": 0.04289808189314158, + "grad_norm": 0.6845353245735168, + "learning_rate": 0.00019910268639597472, + "loss": 0.1747, + "step": 1830 + }, + { + "epoch": 0.042921523467946573, + "grad_norm": 0.592204749584198, + "learning_rate": 0.00019910170154552984, + "loss": 0.187, + "step": 1831 + }, + { + "epoch": 0.042944965042751575, + "grad_norm": 0.5602278709411621, + "learning_rate": 0.00019910071615735636, + "loss": 0.2875, + "step": 1832 + }, + { + "epoch": 0.04296840661755657, + "grad_norm": 0.6878783106803894, + "learning_rate": 0.0001990997302314596, + "loss": 0.1947, + "step": 1833 + }, + { + "epoch": 0.04299184819236156, + "grad_norm": 0.8545064926147461, + "learning_rate": 0.00019909874376784493, + "loss": 0.2301, + "step": 1834 + }, + { + "epoch": 0.04301528976716656, + "grad_norm": 0.9090492725372314, + "learning_rate": 0.00019909775676651766, + "loss": 0.1404, + "step": 1835 + }, + { + "epoch": 0.04303873134197155, + "grad_norm": 0.8256489634513855, + "learning_rate": 0.0001990967692274832, + "loss": 0.1964, + "step": 1836 + }, + { + "epoch": 0.043062172916776546, + "grad_norm": 0.49588435888290405, + "learning_rate": 0.0001990957811507469, + "loss": 0.5413, + "step": 1837 + }, + { + "epoch": 0.04308561449158155, + "grad_norm": 0.6365183591842651, + "learning_rate": 0.00019909479253631406, + "loss": 0.147, + "step": 1838 + }, + { + "epoch": 0.04310905606638654, + "grad_norm": 0.5482841730117798, + "learning_rate": 0.0001990938033841901, + "loss": 0.1695, + "step": 1839 + }, + { + "epoch": 0.043132497641191536, + "grad_norm": 0.6995058655738831, + "learning_rate": 0.0001990928136943804, + "loss": 0.2117, + "step": 1840 + }, + { + "epoch": 0.04315593921599653, + "grad_norm": 0.8740023970603943, + "learning_rate": 0.00019909182346689028, + "loss": 0.2637, + "step": 1841 + }, + { + "epoch": 0.043179380790801525, + "grad_norm": 0.9642997980117798, + "learning_rate": 0.00019909083270172517, + "loss": 0.1541, + "step": 1842 + }, + { + "epoch": 0.04320282236560652, + "grad_norm": 0.23267291486263275, + "learning_rate": 0.00019908984139889037, + "loss": 0.0649, + "step": 1843 + }, + { + "epoch": 0.04322626394041151, + "grad_norm": 0.6412885189056396, + "learning_rate": 0.00019908884955839134, + "loss": 0.1763, + "step": 1844 + }, + { + "epoch": 0.043249705515216515, + "grad_norm": 0.19580042362213135, + "learning_rate": 0.00019908785718023345, + "loss": 0.0973, + "step": 1845 + }, + { + "epoch": 0.04327314709002151, + "grad_norm": 0.6535825729370117, + "learning_rate": 0.000199086864264422, + "loss": 0.2497, + "step": 1846 + }, + { + "epoch": 0.0432965886648265, + "grad_norm": 0.9230482578277588, + "learning_rate": 0.00019908587081096248, + "loss": 0.2908, + "step": 1847 + }, + { + "epoch": 0.0433200302396315, + "grad_norm": 0.783065915107727, + "learning_rate": 0.00019908487681986024, + "loss": 0.256, + "step": 1848 + }, + { + "epoch": 0.04334347181443649, + "grad_norm": 0.4739764332771301, + "learning_rate": 0.00019908388229112066, + "loss": 0.1935, + "step": 1849 + }, + { + "epoch": 0.043366913389241486, + "grad_norm": 0.5858936905860901, + "learning_rate": 0.00019908288722474916, + "loss": 0.2522, + "step": 1850 + }, + { + "epoch": 0.04339035496404649, + "grad_norm": 0.7401807904243469, + "learning_rate": 0.0001990818916207511, + "loss": 0.3023, + "step": 1851 + }, + { + "epoch": 0.04341379653885148, + "grad_norm": 0.6032687425613403, + "learning_rate": 0.00019908089547913192, + "loss": 0.2434, + "step": 1852 + }, + { + "epoch": 0.043437238113656476, + "grad_norm": 0.6473916172981262, + "learning_rate": 0.00019907989879989704, + "loss": 0.1607, + "step": 1853 + }, + { + "epoch": 0.04346067968846147, + "grad_norm": 0.2577202022075653, + "learning_rate": 0.00019907890158305183, + "loss": 0.0892, + "step": 1854 + }, + { + "epoch": 0.043484121263266465, + "grad_norm": 0.14895012974739075, + "learning_rate": 0.0001990779038286017, + "loss": 0.0353, + "step": 1855 + }, + { + "epoch": 0.04350756283807146, + "grad_norm": 0.6274799704551697, + "learning_rate": 0.0001990769055365521, + "loss": 0.2456, + "step": 1856 + }, + { + "epoch": 0.04353100441287646, + "grad_norm": 0.29036185145378113, + "learning_rate": 0.0001990759067069084, + "loss": 0.1415, + "step": 1857 + }, + { + "epoch": 0.043554445987681455, + "grad_norm": 0.5381824374198914, + "learning_rate": 0.00019907490733967603, + "loss": 0.1481, + "step": 1858 + }, + { + "epoch": 0.04357788756248645, + "grad_norm": 0.7319858074188232, + "learning_rate": 0.00019907390743486049, + "loss": 0.4301, + "step": 1859 + }, + { + "epoch": 0.04360132913729144, + "grad_norm": 0.5010014772415161, + "learning_rate": 0.00019907290699246711, + "loss": 0.1343, + "step": 1860 + }, + { + "epoch": 0.04362477071209644, + "grad_norm": 0.48137685656547546, + "learning_rate": 0.00019907190601250134, + "loss": 0.2175, + "step": 1861 + }, + { + "epoch": 0.04364821228690143, + "grad_norm": 1.1243845224380493, + "learning_rate": 0.00019907090449496862, + "loss": 0.4688, + "step": 1862 + }, + { + "epoch": 0.043671653861706426, + "grad_norm": 0.3779436945915222, + "learning_rate": 0.0001990699024398744, + "loss": 0.1286, + "step": 1863 + }, + { + "epoch": 0.04369509543651143, + "grad_norm": 0.5491756796836853, + "learning_rate": 0.0001990688998472241, + "loss": 0.1308, + "step": 1864 + }, + { + "epoch": 0.04371853701131642, + "grad_norm": 0.23047488927841187, + "learning_rate": 0.00019906789671702316, + "loss": 0.0451, + "step": 1865 + }, + { + "epoch": 0.043741978586121416, + "grad_norm": 0.584831178188324, + "learning_rate": 0.00019906689304927705, + "loss": 0.1412, + "step": 1866 + }, + { + "epoch": 0.04376542016092641, + "grad_norm": 0.1996651142835617, + "learning_rate": 0.00019906588884399115, + "loss": 0.0699, + "step": 1867 + }, + { + "epoch": 0.043788861735731405, + "grad_norm": 0.28243109583854675, + "learning_rate": 0.00019906488410117098, + "loss": 0.0729, + "step": 1868 + }, + { + "epoch": 0.0438123033105364, + "grad_norm": 0.4525586664676666, + "learning_rate": 0.00019906387882082197, + "loss": 0.3524, + "step": 1869 + }, + { + "epoch": 0.0438357448853414, + "grad_norm": 0.5289713144302368, + "learning_rate": 0.00019906287300294957, + "loss": 0.1737, + "step": 1870 + }, + { + "epoch": 0.043859186460146395, + "grad_norm": 0.20201215147972107, + "learning_rate": 0.00019906186664755922, + "loss": 0.0502, + "step": 1871 + }, + { + "epoch": 0.04388262803495139, + "grad_norm": 0.7126545310020447, + "learning_rate": 0.0001990608597546564, + "loss": 0.2255, + "step": 1872 + }, + { + "epoch": 0.04390606960975638, + "grad_norm": 0.8620218634605408, + "learning_rate": 0.00019905985232424655, + "loss": 0.2245, + "step": 1873 + }, + { + "epoch": 0.04392951118456138, + "grad_norm": 0.31868216395378113, + "learning_rate": 0.00019905884435633518, + "loss": 0.0553, + "step": 1874 + }, + { + "epoch": 0.04395295275936637, + "grad_norm": 1.242000699043274, + "learning_rate": 0.0001990578358509277, + "loss": 0.2585, + "step": 1875 + }, + { + "epoch": 0.043976394334171366, + "grad_norm": 0.19323663413524628, + "learning_rate": 0.00019905682680802967, + "loss": 0.0619, + "step": 1876 + }, + { + "epoch": 0.04399983590897637, + "grad_norm": 0.646359920501709, + "learning_rate": 0.0001990558172276465, + "loss": 0.1719, + "step": 1877 + }, + { + "epoch": 0.04402327748378136, + "grad_norm": 0.3687511384487152, + "learning_rate": 0.00019905480710978365, + "loss": 0.1011, + "step": 1878 + }, + { + "epoch": 0.044046719058586356, + "grad_norm": 0.815467357635498, + "learning_rate": 0.00019905379645444667, + "loss": 0.1942, + "step": 1879 + }, + { + "epoch": 0.04407016063339135, + "grad_norm": 0.4083411395549774, + "learning_rate": 0.00019905278526164097, + "loss": 0.4241, + "step": 1880 + }, + { + "epoch": 0.044093602208196345, + "grad_norm": 0.5493329763412476, + "learning_rate": 0.0001990517735313721, + "loss": 0.1792, + "step": 1881 + }, + { + "epoch": 0.04411704378300134, + "grad_norm": 0.473382830619812, + "learning_rate": 0.0001990507612636455, + "loss": 0.1377, + "step": 1882 + }, + { + "epoch": 0.04414048535780634, + "grad_norm": 1.5251052379608154, + "learning_rate": 0.00019904974845846672, + "loss": 0.2045, + "step": 1883 + }, + { + "epoch": 0.044163926932611335, + "grad_norm": 0.2789760231971741, + "learning_rate": 0.00019904873511584117, + "loss": 0.0928, + "step": 1884 + }, + { + "epoch": 0.04418736850741633, + "grad_norm": 0.8948970437049866, + "learning_rate": 0.0001990477212357744, + "loss": 0.1618, + "step": 1885 + }, + { + "epoch": 0.04421081008222132, + "grad_norm": 0.5438262224197388, + "learning_rate": 0.00019904670681827197, + "loss": 0.126, + "step": 1886 + }, + { + "epoch": 0.04423425165702632, + "grad_norm": 0.6533169150352478, + "learning_rate": 0.0001990456918633393, + "loss": 0.1452, + "step": 1887 + }, + { + "epoch": 0.04425769323183131, + "grad_norm": 0.18695902824401855, + "learning_rate": 0.0001990446763709819, + "loss": 0.0687, + "step": 1888 + }, + { + "epoch": 0.04428113480663631, + "grad_norm": 0.6636790633201599, + "learning_rate": 0.0001990436603412053, + "loss": 0.1782, + "step": 1889 + }, + { + "epoch": 0.04430457638144131, + "grad_norm": 0.7937007546424866, + "learning_rate": 0.00019904264377401502, + "loss": 0.3131, + "step": 1890 + }, + { + "epoch": 0.0443280179562463, + "grad_norm": 0.5718887448310852, + "learning_rate": 0.00019904162666941656, + "loss": 0.2512, + "step": 1891 + }, + { + "epoch": 0.044351459531051296, + "grad_norm": 0.4530855417251587, + "learning_rate": 0.00019904060902741549, + "loss": 0.1343, + "step": 1892 + }, + { + "epoch": 0.04437490110585629, + "grad_norm": 0.8870995044708252, + "learning_rate": 0.00019903959084801724, + "loss": 0.2201, + "step": 1893 + }, + { + "epoch": 0.044398342680661285, + "grad_norm": 0.41944620013237, + "learning_rate": 0.00019903857213122742, + "loss": 0.0977, + "step": 1894 + }, + { + "epoch": 0.04442178425546628, + "grad_norm": 0.6139050126075745, + "learning_rate": 0.0001990375528770515, + "loss": 0.1919, + "step": 1895 + }, + { + "epoch": 0.04444522583027128, + "grad_norm": 0.7133302688598633, + "learning_rate": 0.00019903653308549505, + "loss": 0.7262, + "step": 1896 + }, + { + "epoch": 0.044468667405076275, + "grad_norm": 0.7944154143333435, + "learning_rate": 0.0001990355127565636, + "loss": 0.2003, + "step": 1897 + }, + { + "epoch": 0.04449210897988127, + "grad_norm": 0.5575345158576965, + "learning_rate": 0.00019903449189026263, + "loss": 0.1969, + "step": 1898 + }, + { + "epoch": 0.04451555055468626, + "grad_norm": 0.4743904173374176, + "learning_rate": 0.00019903347048659775, + "loss": 0.1372, + "step": 1899 + }, + { + "epoch": 0.04453899212949126, + "grad_norm": 0.6575098037719727, + "learning_rate": 0.0001990324485455745, + "loss": 0.1991, + "step": 1900 + }, + { + "epoch": 0.04456243370429625, + "grad_norm": 0.49272388219833374, + "learning_rate": 0.00019903142606719833, + "loss": 0.1352, + "step": 1901 + }, + { + "epoch": 0.04458587527910125, + "grad_norm": 0.23346592485904694, + "learning_rate": 0.00019903040305147492, + "loss": 0.0475, + "step": 1902 + }, + { + "epoch": 0.04460931685390625, + "grad_norm": 0.559798002243042, + "learning_rate": 0.00019902937949840973, + "loss": 0.1981, + "step": 1903 + }, + { + "epoch": 0.04463275842871124, + "grad_norm": 0.5438728332519531, + "learning_rate": 0.00019902835540800835, + "loss": 0.1298, + "step": 1904 + }, + { + "epoch": 0.044656200003516236, + "grad_norm": 0.7353297472000122, + "learning_rate": 0.00019902733078027633, + "loss": 0.2279, + "step": 1905 + }, + { + "epoch": 0.04467964157832123, + "grad_norm": 0.42874976992607117, + "learning_rate": 0.00019902630561521924, + "loss": 0.1204, + "step": 1906 + }, + { + "epoch": 0.044703083153126225, + "grad_norm": 0.4591049253940582, + "learning_rate": 0.0001990252799128426, + "loss": 0.1335, + "step": 1907 + }, + { + "epoch": 0.044726524727931226, + "grad_norm": 0.618068516254425, + "learning_rate": 0.00019902425367315202, + "loss": 0.2506, + "step": 1908 + }, + { + "epoch": 0.04474996630273622, + "grad_norm": 0.6666290163993835, + "learning_rate": 0.00019902322689615307, + "loss": 0.1611, + "step": 1909 + }, + { + "epoch": 0.044773407877541214, + "grad_norm": 0.15207991003990173, + "learning_rate": 0.0001990221995818513, + "loss": 0.0244, + "step": 1910 + }, + { + "epoch": 0.04479684945234621, + "grad_norm": 0.24474023282527924, + "learning_rate": 0.00019902117173025227, + "loss": 0.0463, + "step": 1911 + }, + { + "epoch": 0.0448202910271512, + "grad_norm": 0.5203870534896851, + "learning_rate": 0.0001990201433413616, + "loss": 0.1149, + "step": 1912 + }, + { + "epoch": 0.0448437326019562, + "grad_norm": 0.2642010748386383, + "learning_rate": 0.00019901911441518486, + "loss": 0.0695, + "step": 1913 + }, + { + "epoch": 0.04486717417676119, + "grad_norm": 0.3563833236694336, + "learning_rate": 0.0001990180849517276, + "loss": 0.1267, + "step": 1914 + }, + { + "epoch": 0.04489061575156619, + "grad_norm": 0.2433224469423294, + "learning_rate": 0.00019901705495099545, + "loss": 0.0656, + "step": 1915 + }, + { + "epoch": 0.04491405732637119, + "grad_norm": 0.4913172423839569, + "learning_rate": 0.00019901602441299394, + "loss": 0.143, + "step": 1916 + }, + { + "epoch": 0.04493749890117618, + "grad_norm": 0.4540841579437256, + "learning_rate": 0.00019901499333772874, + "loss": 0.1657, + "step": 1917 + }, + { + "epoch": 0.044960940475981176, + "grad_norm": 0.41943037509918213, + "learning_rate": 0.00019901396172520536, + "loss": 0.161, + "step": 1918 + }, + { + "epoch": 0.04498438205078617, + "grad_norm": 0.5120501518249512, + "learning_rate": 0.00019901292957542945, + "loss": 0.791, + "step": 1919 + }, + { + "epoch": 0.045007823625591165, + "grad_norm": 0.19925403594970703, + "learning_rate": 0.0001990118968884066, + "loss": 0.0538, + "step": 1920 + }, + { + "epoch": 0.045031265200396166, + "grad_norm": 1.0554518699645996, + "learning_rate": 0.00019901086366414243, + "loss": 0.2361, + "step": 1921 + }, + { + "epoch": 0.04505470677520116, + "grad_norm": 0.7474130392074585, + "learning_rate": 0.00019900982990264253, + "loss": 0.1909, + "step": 1922 + }, + { + "epoch": 0.045078148350006154, + "grad_norm": 0.9372864961624146, + "learning_rate": 0.00019900879560391248, + "loss": 0.3268, + "step": 1923 + }, + { + "epoch": 0.04510158992481115, + "grad_norm": 0.7117167115211487, + "learning_rate": 0.00019900776076795795, + "loss": 0.1329, + "step": 1924 + }, + { + "epoch": 0.04512503149961614, + "grad_norm": 0.30228880047798157, + "learning_rate": 0.00019900672539478452, + "loss": 0.034, + "step": 1925 + }, + { + "epoch": 0.04514847307442114, + "grad_norm": 0.714878261089325, + "learning_rate": 0.0001990056894843978, + "loss": 0.4051, + "step": 1926 + }, + { + "epoch": 0.04517191464922613, + "grad_norm": 0.16953498125076294, + "learning_rate": 0.00019900465303680345, + "loss": 0.0286, + "step": 1927 + }, + { + "epoch": 0.04519535622403113, + "grad_norm": 0.6700881123542786, + "learning_rate": 0.00019900361605200705, + "loss": 0.2924, + "step": 1928 + }, + { + "epoch": 0.04521879779883613, + "grad_norm": 0.5703471302986145, + "learning_rate": 0.00019900257853001425, + "loss": 0.2445, + "step": 1929 + }, + { + "epoch": 0.04524223937364112, + "grad_norm": 0.21867702901363373, + "learning_rate": 0.00019900154047083066, + "loss": 0.0923, + "step": 1930 + }, + { + "epoch": 0.045265680948446116, + "grad_norm": 0.483244925737381, + "learning_rate": 0.00019900050187446198, + "loss": 0.1435, + "step": 1931 + }, + { + "epoch": 0.04528912252325111, + "grad_norm": 0.42652618885040283, + "learning_rate": 0.00019899946274091375, + "loss": 0.119, + "step": 1932 + }, + { + "epoch": 0.045312564098056105, + "grad_norm": 0.2793349325656891, + "learning_rate": 0.00019899842307019165, + "loss": 0.0697, + "step": 1933 + }, + { + "epoch": 0.045336005672861106, + "grad_norm": 0.6768976449966431, + "learning_rate": 0.00019899738286230134, + "loss": 0.1529, + "step": 1934 + }, + { + "epoch": 0.0453594472476661, + "grad_norm": 0.6536403894424438, + "learning_rate": 0.00019899634211724843, + "loss": 0.1357, + "step": 1935 + }, + { + "epoch": 0.045382888822471094, + "grad_norm": 0.11086787283420563, + "learning_rate": 0.0001989953008350386, + "loss": 0.0238, + "step": 1936 + }, + { + "epoch": 0.04540633039727609, + "grad_norm": 0.2656776010990143, + "learning_rate": 0.00019899425901567748, + "loss": 0.08, + "step": 1937 + }, + { + "epoch": 0.04542977197208108, + "grad_norm": 0.3277013599872589, + "learning_rate": 0.0001989932166591707, + "loss": 0.0727, + "step": 1938 + }, + { + "epoch": 0.04545321354688608, + "grad_norm": 0.8907303214073181, + "learning_rate": 0.00019899217376552396, + "loss": 0.3209, + "step": 1939 + }, + { + "epoch": 0.04547665512169108, + "grad_norm": 0.38226133584976196, + "learning_rate": 0.00019899113033474292, + "loss": 0.0801, + "step": 1940 + }, + { + "epoch": 0.04550009669649607, + "grad_norm": 0.6086671352386475, + "learning_rate": 0.00019899008636683322, + "loss": 0.1372, + "step": 1941 + }, + { + "epoch": 0.04552353827130107, + "grad_norm": 0.28260084986686707, + "learning_rate": 0.0001989890418618005, + "loss": 0.1251, + "step": 1942 + }, + { + "epoch": 0.04554697984610606, + "grad_norm": 0.6015377044677734, + "learning_rate": 0.00019898799681965047, + "loss": 0.8325, + "step": 1943 + }, + { + "epoch": 0.045570421420911056, + "grad_norm": 0.22205543518066406, + "learning_rate": 0.0001989869512403888, + "loss": 0.0575, + "step": 1944 + }, + { + "epoch": 0.04559386299571605, + "grad_norm": 0.5084449648857117, + "learning_rate": 0.00019898590512402114, + "loss": 0.1739, + "step": 1945 + }, + { + "epoch": 0.045617304570521044, + "grad_norm": 0.5243961811065674, + "learning_rate": 0.0001989848584705532, + "loss": 0.1164, + "step": 1946 + }, + { + "epoch": 0.045640746145326046, + "grad_norm": 0.28687217831611633, + "learning_rate": 0.0001989838112799906, + "loss": 0.0959, + "step": 1947 + }, + { + "epoch": 0.04566418772013104, + "grad_norm": 0.29733237624168396, + "learning_rate": 0.00019898276355233906, + "loss": 0.0826, + "step": 1948 + }, + { + "epoch": 0.045687629294936034, + "grad_norm": 0.8049404621124268, + "learning_rate": 0.00019898171528760425, + "loss": 0.2496, + "step": 1949 + }, + { + "epoch": 0.04571107086974103, + "grad_norm": 0.7480884790420532, + "learning_rate": 0.00019898066648579193, + "loss": 0.7455, + "step": 1950 + }, + { + "epoch": 0.04573451244454602, + "grad_norm": 0.5457778573036194, + "learning_rate": 0.00019897961714690767, + "loss": 0.0875, + "step": 1951 + }, + { + "epoch": 0.04575795401935102, + "grad_norm": 0.49032095074653625, + "learning_rate": 0.00019897856727095725, + "loss": 0.6239, + "step": 1952 + }, + { + "epoch": 0.04578139559415602, + "grad_norm": 0.17519041895866394, + "learning_rate": 0.0001989775168579463, + "loss": 0.0446, + "step": 1953 + }, + { + "epoch": 0.04580483716896101, + "grad_norm": 0.478712260723114, + "learning_rate": 0.00019897646590788062, + "loss": 0.1482, + "step": 1954 + }, + { + "epoch": 0.04582827874376601, + "grad_norm": 1.0633354187011719, + "learning_rate": 0.00019897541442076585, + "loss": 0.1945, + "step": 1955 + }, + { + "epoch": 0.045851720318571, + "grad_norm": 0.7383019328117371, + "learning_rate": 0.00019897436239660766, + "loss": 0.1549, + "step": 1956 + }, + { + "epoch": 0.045875161893375996, + "grad_norm": 0.5157790780067444, + "learning_rate": 0.0001989733098354118, + "loss": 0.1261, + "step": 1957 + }, + { + "epoch": 0.04589860346818099, + "grad_norm": 0.7624489665031433, + "learning_rate": 0.000198972256737184, + "loss": 0.204, + "step": 1958 + }, + { + "epoch": 0.045922045042985984, + "grad_norm": 0.43477579951286316, + "learning_rate": 0.00019897120310192996, + "loss": 0.1728, + "step": 1959 + }, + { + "epoch": 0.045945486617790986, + "grad_norm": 0.2620481252670288, + "learning_rate": 0.00019897014892965537, + "loss": 0.0537, + "step": 1960 + }, + { + "epoch": 0.04596892819259598, + "grad_norm": 0.17499636113643646, + "learning_rate": 0.00019896909422036595, + "loss": 0.054, + "step": 1961 + }, + { + "epoch": 0.045992369767400974, + "grad_norm": 0.5280943512916565, + "learning_rate": 0.0001989680389740675, + "loss": 0.1287, + "step": 1962 + }, + { + "epoch": 0.04601581134220597, + "grad_norm": 0.7755686640739441, + "learning_rate": 0.00019896698319076564, + "loss": 0.2524, + "step": 1963 + }, + { + "epoch": 0.04603925291701096, + "grad_norm": 0.7438785433769226, + "learning_rate": 0.00019896592687046612, + "loss": 0.2017, + "step": 1964 + }, + { + "epoch": 0.04606269449181596, + "grad_norm": 0.2553136646747589, + "learning_rate": 0.00019896487001317477, + "loss": 0.0853, + "step": 1965 + }, + { + "epoch": 0.04608613606662096, + "grad_norm": 0.8106321096420288, + "learning_rate": 0.0001989638126188972, + "loss": 0.2795, + "step": 1966 + }, + { + "epoch": 0.04610957764142595, + "grad_norm": 0.7611567974090576, + "learning_rate": 0.00019896275468763924, + "loss": 0.3324, + "step": 1967 + }, + { + "epoch": 0.04613301921623095, + "grad_norm": 0.5558561086654663, + "learning_rate": 0.00019896169621940656, + "loss": 0.0989, + "step": 1968 + }, + { + "epoch": 0.04615646079103594, + "grad_norm": 0.8820204138755798, + "learning_rate": 0.00019896063721420494, + "loss": 0.3352, + "step": 1969 + }, + { + "epoch": 0.046179902365840936, + "grad_norm": 0.6518877744674683, + "learning_rate": 0.00019895957767204013, + "loss": 0.1382, + "step": 1970 + }, + { + "epoch": 0.04620334394064593, + "grad_norm": 0.5028252601623535, + "learning_rate": 0.00019895851759291783, + "loss": 0.1315, + "step": 1971 + }, + { + "epoch": 0.04622678551545093, + "grad_norm": 0.24052710831165314, + "learning_rate": 0.00019895745697684388, + "loss": 0.0819, + "step": 1972 + }, + { + "epoch": 0.046250227090255926, + "grad_norm": 0.4823521077632904, + "learning_rate": 0.00019895639582382396, + "loss": 0.1072, + "step": 1973 + }, + { + "epoch": 0.04627366866506092, + "grad_norm": 0.21663276851177216, + "learning_rate": 0.00019895533413386387, + "loss": 0.0498, + "step": 1974 + }, + { + "epoch": 0.046297110239865914, + "grad_norm": 0.1785694807767868, + "learning_rate": 0.00019895427190696935, + "loss": 0.0679, + "step": 1975 + }, + { + "epoch": 0.04632055181467091, + "grad_norm": 0.3989429175853729, + "learning_rate": 0.00019895320914314617, + "loss": 0.1461, + "step": 1976 + }, + { + "epoch": 0.0463439933894759, + "grad_norm": 0.23429200053215027, + "learning_rate": 0.0001989521458424001, + "loss": 0.0928, + "step": 1977 + }, + { + "epoch": 0.0463674349642809, + "grad_norm": 0.5178514122962952, + "learning_rate": 0.00019895108200473688, + "loss": 0.2416, + "step": 1978 + }, + { + "epoch": 0.0463908765390859, + "grad_norm": 0.39489197731018066, + "learning_rate": 0.00019895001763016232, + "loss": 0.1669, + "step": 1979 + }, + { + "epoch": 0.04641431811389089, + "grad_norm": 0.47763949632644653, + "learning_rate": 0.0001989489527186822, + "loss": 0.1547, + "step": 1980 + }, + { + "epoch": 0.04643775968869589, + "grad_norm": 0.16933897137641907, + "learning_rate": 0.00019894788727030225, + "loss": 0.0453, + "step": 1981 + }, + { + "epoch": 0.04646120126350088, + "grad_norm": 0.1492980718612671, + "learning_rate": 0.0001989468212850283, + "loss": 0.0354, + "step": 1982 + }, + { + "epoch": 0.046484642838305876, + "grad_norm": 0.5901780128479004, + "learning_rate": 0.00019894575476286613, + "loss": 0.19, + "step": 1983 + }, + { + "epoch": 0.04650808441311087, + "grad_norm": 0.3083515763282776, + "learning_rate": 0.00019894468770382146, + "loss": 0.0756, + "step": 1984 + }, + { + "epoch": 0.04653152598791587, + "grad_norm": 0.38928306102752686, + "learning_rate": 0.0001989436201079002, + "loss": 0.1463, + "step": 1985 + }, + { + "epoch": 0.046554967562720866, + "grad_norm": 0.749664306640625, + "learning_rate": 0.00019894255197510803, + "loss": 0.1966, + "step": 1986 + }, + { + "epoch": 0.04657840913752586, + "grad_norm": 0.3064092695713043, + "learning_rate": 0.00019894148330545081, + "loss": 0.1056, + "step": 1987 + }, + { + "epoch": 0.046601850712330854, + "grad_norm": 0.579096257686615, + "learning_rate": 0.0001989404140989343, + "loss": 0.1599, + "step": 1988 + }, + { + "epoch": 0.04662529228713585, + "grad_norm": 0.2989484965801239, + "learning_rate": 0.00019893934435556433, + "loss": 0.1098, + "step": 1989 + }, + { + "epoch": 0.04664873386194084, + "grad_norm": 0.4789062440395355, + "learning_rate": 0.0001989382740753467, + "loss": 0.1638, + "step": 1990 + }, + { + "epoch": 0.04667217543674584, + "grad_norm": 0.6478590369224548, + "learning_rate": 0.00019893720325828723, + "loss": 0.1698, + "step": 1991 + }, + { + "epoch": 0.04669561701155084, + "grad_norm": 0.44414210319519043, + "learning_rate": 0.0001989361319043917, + "loss": 0.4867, + "step": 1992 + }, + { + "epoch": 0.04671905858635583, + "grad_norm": 0.6348602175712585, + "learning_rate": 0.0001989350600136659, + "loss": 0.2174, + "step": 1993 + }, + { + "epoch": 0.04674250016116083, + "grad_norm": 1.1521695852279663, + "learning_rate": 0.0001989339875861157, + "loss": 0.219, + "step": 1994 + }, + { + "epoch": 0.04676594173596582, + "grad_norm": 0.5909392237663269, + "learning_rate": 0.00019893291462174694, + "loss": 0.1324, + "step": 1995 + }, + { + "epoch": 0.046789383310770816, + "grad_norm": 0.42345064878463745, + "learning_rate": 0.0001989318411205654, + "loss": 0.1359, + "step": 1996 + }, + { + "epoch": 0.04681282488557581, + "grad_norm": 0.6724593639373779, + "learning_rate": 0.00019893076708257687, + "loss": 0.7762, + "step": 1997 + }, + { + "epoch": 0.04683626646038081, + "grad_norm": 0.5398660898208618, + "learning_rate": 0.00019892969250778723, + "loss": 0.8098, + "step": 1998 + }, + { + "epoch": 0.046859708035185806, + "grad_norm": 0.34341955184936523, + "learning_rate": 0.0001989286173962023, + "loss": 0.0778, + "step": 1999 + }, + { + "epoch": 0.0468831496099908, + "grad_norm": 0.5907673835754395, + "learning_rate": 0.00019892754174782791, + "loss": 0.7452, + "step": 2000 + }, + { + "epoch": 0.046906591184795794, + "grad_norm": 0.46365392208099365, + "learning_rate": 0.00019892646556266992, + "loss": 0.1283, + "step": 2001 + }, + { + "epoch": 0.04693003275960079, + "grad_norm": 0.45642462372779846, + "learning_rate": 0.0001989253888407341, + "loss": 0.1338, + "step": 2002 + }, + { + "epoch": 0.04695347433440578, + "grad_norm": 0.4205925166606903, + "learning_rate": 0.00019892431158202638, + "loss": 0.1142, + "step": 2003 + }, + { + "epoch": 0.046976915909210784, + "grad_norm": 0.24989046156406403, + "learning_rate": 0.00019892323378655253, + "loss": 0.1049, + "step": 2004 + }, + { + "epoch": 0.04700035748401578, + "grad_norm": 0.6635686755180359, + "learning_rate": 0.00019892215545431845, + "loss": 0.2691, + "step": 2005 + }, + { + "epoch": 0.04702379905882077, + "grad_norm": 0.35612669587135315, + "learning_rate": 0.00019892107658533, + "loss": 0.0855, + "step": 2006 + }, + { + "epoch": 0.04704724063362577, + "grad_norm": 0.37297961115837097, + "learning_rate": 0.00019891999717959295, + "loss": 0.1102, + "step": 2007 + }, + { + "epoch": 0.04707068220843076, + "grad_norm": 0.6410475969314575, + "learning_rate": 0.00019891891723711327, + "loss": 0.1878, + "step": 2008 + }, + { + "epoch": 0.047094123783235756, + "grad_norm": 1.3443571329116821, + "learning_rate": 0.00019891783675789673, + "loss": 0.1979, + "step": 2009 + }, + { + "epoch": 0.04711756535804075, + "grad_norm": 0.17031948268413544, + "learning_rate": 0.0001989167557419492, + "loss": 0.0821, + "step": 2010 + }, + { + "epoch": 0.04714100693284575, + "grad_norm": 0.7464780807495117, + "learning_rate": 0.00019891567418927665, + "loss": 0.1994, + "step": 2011 + }, + { + "epoch": 0.047164448507650746, + "grad_norm": 0.35685276985168457, + "learning_rate": 0.00019891459209988481, + "loss": 0.1126, + "step": 2012 + }, + { + "epoch": 0.04718789008245574, + "grad_norm": 0.21158932149410248, + "learning_rate": 0.0001989135094737796, + "loss": 0.0586, + "step": 2013 + }, + { + "epoch": 0.047211331657260734, + "grad_norm": 1.1954760551452637, + "learning_rate": 0.00019891242631096695, + "loss": 0.1828, + "step": 2014 + }, + { + "epoch": 0.04723477323206573, + "grad_norm": 0.8926751613616943, + "learning_rate": 0.00019891134261145268, + "loss": 0.2213, + "step": 2015 + }, + { + "epoch": 0.04725821480687072, + "grad_norm": 0.413129985332489, + "learning_rate": 0.00019891025837524267, + "loss": 0.1394, + "step": 2016 + }, + { + "epoch": 0.047281656381675724, + "grad_norm": 0.7268370389938354, + "learning_rate": 0.00019890917360234283, + "loss": 0.2221, + "step": 2017 + }, + { + "epoch": 0.04730509795648072, + "grad_norm": 0.552194356918335, + "learning_rate": 0.00019890808829275904, + "loss": 0.159, + "step": 2018 + }, + { + "epoch": 0.04732853953128571, + "grad_norm": 0.6325605511665344, + "learning_rate": 0.00019890700244649717, + "loss": 0.1493, + "step": 2019 + }, + { + "epoch": 0.04735198110609071, + "grad_norm": 0.5944485068321228, + "learning_rate": 0.00019890591606356312, + "loss": 0.1711, + "step": 2020 + }, + { + "epoch": 0.0473754226808957, + "grad_norm": 0.22156988084316254, + "learning_rate": 0.0001989048291439628, + "loss": 0.0566, + "step": 2021 + }, + { + "epoch": 0.047398864255700696, + "grad_norm": 0.31351006031036377, + "learning_rate": 0.0001989037416877021, + "loss": 0.0672, + "step": 2022 + }, + { + "epoch": 0.04742230583050569, + "grad_norm": 0.4472307860851288, + "learning_rate": 0.0001989026536947869, + "loss": 0.1361, + "step": 2023 + }, + { + "epoch": 0.04744574740531069, + "grad_norm": 0.8522812128067017, + "learning_rate": 0.00019890156516522313, + "loss": 0.1947, + "step": 2024 + }, + { + "epoch": 0.047469188980115685, + "grad_norm": 0.8829949498176575, + "learning_rate": 0.0001989004760990167, + "loss": 0.2172, + "step": 2025 + }, + { + "epoch": 0.04749263055492068, + "grad_norm": 0.19738318026065826, + "learning_rate": 0.00019889938649617347, + "loss": 0.0316, + "step": 2026 + }, + { + "epoch": 0.047516072129725674, + "grad_norm": 0.870404839515686, + "learning_rate": 0.00019889829635669944, + "loss": 0.2608, + "step": 2027 + }, + { + "epoch": 0.04753951370453067, + "grad_norm": 0.985802412033081, + "learning_rate": 0.00019889720568060042, + "loss": 0.3028, + "step": 2028 + }, + { + "epoch": 0.04756295527933566, + "grad_norm": 1.0302300453186035, + "learning_rate": 0.00019889611446788239, + "loss": 0.211, + "step": 2029 + }, + { + "epoch": 0.047586396854140664, + "grad_norm": 0.9034472703933716, + "learning_rate": 0.00019889502271855128, + "loss": 0.142, + "step": 2030 + }, + { + "epoch": 0.04760983842894566, + "grad_norm": 0.6777682900428772, + "learning_rate": 0.000198893930432613, + "loss": 0.2223, + "step": 2031 + }, + { + "epoch": 0.04763328000375065, + "grad_norm": 0.6903116106987, + "learning_rate": 0.00019889283761007347, + "loss": 0.2503, + "step": 2032 + }, + { + "epoch": 0.04765672157855565, + "grad_norm": 0.4699556827545166, + "learning_rate": 0.0001988917442509386, + "loss": 0.0756, + "step": 2033 + }, + { + "epoch": 0.04768016315336064, + "grad_norm": 0.5100730061531067, + "learning_rate": 0.00019889065035521436, + "loss": 0.1632, + "step": 2034 + }, + { + "epoch": 0.047703604728165636, + "grad_norm": 0.6534309983253479, + "learning_rate": 0.0001988895559229067, + "loss": 0.197, + "step": 2035 + }, + { + "epoch": 0.04772704630297064, + "grad_norm": 0.6867474317550659, + "learning_rate": 0.0001988884609540215, + "loss": 0.2698, + "step": 2036 + }, + { + "epoch": 0.04775048787777563, + "grad_norm": 0.1962049901485443, + "learning_rate": 0.00019888736544856472, + "loss": 0.063, + "step": 2037 + }, + { + "epoch": 0.047773929452580625, + "grad_norm": 0.3664695918560028, + "learning_rate": 0.0001988862694065423, + "loss": 0.1566, + "step": 2038 + }, + { + "epoch": 0.04779737102738562, + "grad_norm": 0.14117223024368286, + "learning_rate": 0.00019888517282796024, + "loss": 0.0387, + "step": 2039 + }, + { + "epoch": 0.047820812602190614, + "grad_norm": 0.23648154735565186, + "learning_rate": 0.00019888407571282447, + "loss": 0.0839, + "step": 2040 + }, + { + "epoch": 0.04784425417699561, + "grad_norm": 0.5273813605308533, + "learning_rate": 0.0001988829780611409, + "loss": 0.1253, + "step": 2041 + }, + { + "epoch": 0.0478676957518006, + "grad_norm": 0.48707541823387146, + "learning_rate": 0.00019888187987291548, + "loss": 0.1541, + "step": 2042 + }, + { + "epoch": 0.047891137326605604, + "grad_norm": 0.5017040371894836, + "learning_rate": 0.00019888078114815424, + "loss": 0.1875, + "step": 2043 + }, + { + "epoch": 0.0479145789014106, + "grad_norm": 0.4936288595199585, + "learning_rate": 0.0001988796818868631, + "loss": 0.1768, + "step": 2044 + }, + { + "epoch": 0.04793802047621559, + "grad_norm": 0.5252894163131714, + "learning_rate": 0.00019887858208904796, + "loss": 0.2112, + "step": 2045 + }, + { + "epoch": 0.04796146205102059, + "grad_norm": 0.48786792159080505, + "learning_rate": 0.00019887748175471495, + "loss": 0.7257, + "step": 2046 + }, + { + "epoch": 0.04798490362582558, + "grad_norm": 0.4279000461101532, + "learning_rate": 0.0001988763808838699, + "loss": 0.207, + "step": 2047 + }, + { + "epoch": 0.048008345200630576, + "grad_norm": 0.3888835906982422, + "learning_rate": 0.0001988752794765188, + "loss": 0.0761, + "step": 2048 + }, + { + "epoch": 0.04803178677543558, + "grad_norm": 0.5260328054428101, + "learning_rate": 0.0001988741775326677, + "loss": 0.2121, + "step": 2049 + }, + { + "epoch": 0.04805522835024057, + "grad_norm": 0.49794629216194153, + "learning_rate": 0.00019887307505232248, + "loss": 0.2197, + "step": 2050 + }, + { + "epoch": 0.048078669925045565, + "grad_norm": 0.7458280324935913, + "learning_rate": 0.00019887197203548925, + "loss": 0.6565, + "step": 2051 + }, + { + "epoch": 0.04810211149985056, + "grad_norm": 0.6641652584075928, + "learning_rate": 0.00019887086848217384, + "loss": 0.788, + "step": 2052 + }, + { + "epoch": 0.048125553074655554, + "grad_norm": 0.7224602699279785, + "learning_rate": 0.00019886976439238235, + "loss": 0.1682, + "step": 2053 + }, + { + "epoch": 0.04814899464946055, + "grad_norm": 0.547616720199585, + "learning_rate": 0.00019886865976612077, + "loss": 0.1545, + "step": 2054 + }, + { + "epoch": 0.04817243622426555, + "grad_norm": 0.9120610356330872, + "learning_rate": 0.00019886755460339503, + "loss": 0.1607, + "step": 2055 + }, + { + "epoch": 0.048195877799070544, + "grad_norm": 0.2350284457206726, + "learning_rate": 0.00019886644890421116, + "loss": 0.0467, + "step": 2056 + }, + { + "epoch": 0.04821931937387554, + "grad_norm": 0.7163190841674805, + "learning_rate": 0.00019886534266857516, + "loss": 0.2398, + "step": 2057 + }, + { + "epoch": 0.04824276094868053, + "grad_norm": 0.4435460567474365, + "learning_rate": 0.00019886423589649303, + "loss": 0.4949, + "step": 2058 + }, + { + "epoch": 0.04826620252348553, + "grad_norm": 0.5414316058158875, + "learning_rate": 0.00019886312858797078, + "loss": 0.137, + "step": 2059 + }, + { + "epoch": 0.04828964409829052, + "grad_norm": 0.7417379021644592, + "learning_rate": 0.00019886202074301444, + "loss": 0.8285, + "step": 2060 + }, + { + "epoch": 0.048313085673095515, + "grad_norm": 0.30057528614997864, + "learning_rate": 0.00019886091236162993, + "loss": 0.048, + "step": 2061 + }, + { + "epoch": 0.04833652724790052, + "grad_norm": 0.189932718873024, + "learning_rate": 0.00019885980344382337, + "loss": 0.0386, + "step": 2062 + }, + { + "epoch": 0.04835996882270551, + "grad_norm": 0.5621135234832764, + "learning_rate": 0.00019885869398960076, + "loss": 0.3428, + "step": 2063 + }, + { + "epoch": 0.048383410397510505, + "grad_norm": 0.8518016934394836, + "learning_rate": 0.00019885758399896806, + "loss": 0.2846, + "step": 2064 + }, + { + "epoch": 0.0484068519723155, + "grad_norm": 0.11999205499887466, + "learning_rate": 0.00019885647347193136, + "loss": 0.0363, + "step": 2065 + }, + { + "epoch": 0.048430293547120494, + "grad_norm": 0.7900104522705078, + "learning_rate": 0.00019885536240849664, + "loss": 0.1698, + "step": 2066 + }, + { + "epoch": 0.04845373512192549, + "grad_norm": 0.610983669757843, + "learning_rate": 0.00019885425080866994, + "loss": 0.1735, + "step": 2067 + }, + { + "epoch": 0.04847717669673049, + "grad_norm": 0.5809565186500549, + "learning_rate": 0.0001988531386724573, + "loss": 0.7462, + "step": 2068 + }, + { + "epoch": 0.048500618271535484, + "grad_norm": 0.20025190711021423, + "learning_rate": 0.00019885202599986474, + "loss": 0.0444, + "step": 2069 + }, + { + "epoch": 0.04852405984634048, + "grad_norm": 0.6328809261322021, + "learning_rate": 0.00019885091279089832, + "loss": 0.5816, + "step": 2070 + }, + { + "epoch": 0.04854750142114547, + "grad_norm": 0.7870927453041077, + "learning_rate": 0.00019884979904556406, + "loss": 0.3238, + "step": 2071 + }, + { + "epoch": 0.04857094299595047, + "grad_norm": 0.7154505848884583, + "learning_rate": 0.000198848684763868, + "loss": 0.1628, + "step": 2072 + }, + { + "epoch": 0.04859438457075546, + "grad_norm": 0.1456042230129242, + "learning_rate": 0.00019884756994581623, + "loss": 0.0416, + "step": 2073 + }, + { + "epoch": 0.048617826145560455, + "grad_norm": 0.7658641338348389, + "learning_rate": 0.00019884645459141472, + "loss": 0.1636, + "step": 2074 + }, + { + "epoch": 0.04864126772036546, + "grad_norm": 0.46978166699409485, + "learning_rate": 0.0001988453387006696, + "loss": 0.1597, + "step": 2075 + }, + { + "epoch": 0.04866470929517045, + "grad_norm": 0.2945898473262787, + "learning_rate": 0.0001988442222735869, + "loss": 0.0684, + "step": 2076 + }, + { + "epoch": 0.048688150869975445, + "grad_norm": 0.29798591136932373, + "learning_rate": 0.00019884310531017264, + "loss": 0.0943, + "step": 2077 + }, + { + "epoch": 0.04871159244478044, + "grad_norm": 0.46228650212287903, + "learning_rate": 0.00019884198781043293, + "loss": 0.1065, + "step": 2078 + }, + { + "epoch": 0.048735034019585434, + "grad_norm": 0.2476292848587036, + "learning_rate": 0.00019884086977437382, + "loss": 0.0685, + "step": 2079 + }, + { + "epoch": 0.04875847559439043, + "grad_norm": 0.6454391479492188, + "learning_rate": 0.00019883975120200137, + "loss": 0.2007, + "step": 2080 + }, + { + "epoch": 0.04878191716919543, + "grad_norm": 0.309614360332489, + "learning_rate": 0.00019883863209332165, + "loss": 0.073, + "step": 2081 + }, + { + "epoch": 0.048805358744000424, + "grad_norm": 0.6307048201560974, + "learning_rate": 0.00019883751244834073, + "loss": 0.1719, + "step": 2082 + }, + { + "epoch": 0.04882880031880542, + "grad_norm": 0.7963147759437561, + "learning_rate": 0.00019883639226706467, + "loss": 0.7475, + "step": 2083 + }, + { + "epoch": 0.04885224189361041, + "grad_norm": 0.5858368277549744, + "learning_rate": 0.00019883527154949962, + "loss": 0.2004, + "step": 2084 + }, + { + "epoch": 0.04887568346841541, + "grad_norm": 0.6350700855255127, + "learning_rate": 0.00019883415029565155, + "loss": 0.1444, + "step": 2085 + }, + { + "epoch": 0.0488991250432204, + "grad_norm": 0.6136678457260132, + "learning_rate": 0.00019883302850552662, + "loss": 0.2362, + "step": 2086 + }, + { + "epoch": 0.0489225666180254, + "grad_norm": 0.7214853167533875, + "learning_rate": 0.00019883190617913093, + "loss": 0.2217, + "step": 2087 + }, + { + "epoch": 0.0489460081928304, + "grad_norm": 0.7862026691436768, + "learning_rate": 0.0001988307833164705, + "loss": 0.1587, + "step": 2088 + }, + { + "epoch": 0.04896944976763539, + "grad_norm": 0.35666748881340027, + "learning_rate": 0.00019882965991755146, + "loss": 0.1269, + "step": 2089 + }, + { + "epoch": 0.048992891342440385, + "grad_norm": 0.4204428791999817, + "learning_rate": 0.0001988285359823799, + "loss": 0.4302, + "step": 2090 + }, + { + "epoch": 0.04901633291724538, + "grad_norm": 0.7530508041381836, + "learning_rate": 0.00019882741151096197, + "loss": 0.1953, + "step": 2091 + }, + { + "epoch": 0.049039774492050374, + "grad_norm": 0.4112527072429657, + "learning_rate": 0.0001988262865033037, + "loss": 0.1408, + "step": 2092 + }, + { + "epoch": 0.04906321606685537, + "grad_norm": 0.3111885190010071, + "learning_rate": 0.0001988251609594112, + "loss": 0.1037, + "step": 2093 + }, + { + "epoch": 0.04908665764166037, + "grad_norm": 0.4151128828525543, + "learning_rate": 0.00019882403487929065, + "loss": 0.1148, + "step": 2094 + }, + { + "epoch": 0.049110099216465364, + "grad_norm": 0.4900517463684082, + "learning_rate": 0.00019882290826294806, + "loss": 0.1819, + "step": 2095 + }, + { + "epoch": 0.04913354079127036, + "grad_norm": 1.2883208990097046, + "learning_rate": 0.0001988217811103896, + "loss": 0.2113, + "step": 2096 + }, + { + "epoch": 0.04915698236607535, + "grad_norm": 0.5141022205352783, + "learning_rate": 0.0001988206534216214, + "loss": 0.133, + "step": 2097 + }, + { + "epoch": 0.04918042394088035, + "grad_norm": 0.608485996723175, + "learning_rate": 0.00019881952519664955, + "loss": 0.253, + "step": 2098 + }, + { + "epoch": 0.04920386551568534, + "grad_norm": 0.4267202317714691, + "learning_rate": 0.00019881839643548014, + "loss": 0.1036, + "step": 2099 + }, + { + "epoch": 0.04922730709049034, + "grad_norm": 0.6238290071487427, + "learning_rate": 0.00019881726713811936, + "loss": 0.2248, + "step": 2100 + }, + { + "epoch": 0.04925074866529534, + "grad_norm": 0.23460575938224792, + "learning_rate": 0.00019881613730457333, + "loss": 0.0595, + "step": 2101 + }, + { + "epoch": 0.04927419024010033, + "grad_norm": 0.6301714777946472, + "learning_rate": 0.00019881500693484815, + "loss": 0.1902, + "step": 2102 + }, + { + "epoch": 0.049297631814905325, + "grad_norm": 0.6128697991371155, + "learning_rate": 0.00019881387602894995, + "loss": 0.1924, + "step": 2103 + }, + { + "epoch": 0.04932107338971032, + "grad_norm": 0.23099221289157867, + "learning_rate": 0.00019881274458688486, + "loss": 0.0856, + "step": 2104 + }, + { + "epoch": 0.049344514964515314, + "grad_norm": 0.5754351019859314, + "learning_rate": 0.00019881161260865908, + "loss": 0.1678, + "step": 2105 + }, + { + "epoch": 0.04936795653932031, + "grad_norm": 0.5725845098495483, + "learning_rate": 0.0001988104800942787, + "loss": 0.0854, + "step": 2106 + }, + { + "epoch": 0.04939139811412531, + "grad_norm": 0.6691819429397583, + "learning_rate": 0.00019880934704374984, + "loss": 0.1677, + "step": 2107 + }, + { + "epoch": 0.049414839688930304, + "grad_norm": 0.41007986664772034, + "learning_rate": 0.00019880821345707875, + "loss": 0.1616, + "step": 2108 + }, + { + "epoch": 0.0494382812637353, + "grad_norm": 0.6026476621627808, + "learning_rate": 0.00019880707933427147, + "loss": 0.1919, + "step": 2109 + }, + { + "epoch": 0.04946172283854029, + "grad_norm": 0.6863310933113098, + "learning_rate": 0.00019880594467533422, + "loss": 0.1933, + "step": 2110 + }, + { + "epoch": 0.04948516441334529, + "grad_norm": 0.42256981134414673, + "learning_rate": 0.0001988048094802731, + "loss": 0.5538, + "step": 2111 + }, + { + "epoch": 0.04950860598815028, + "grad_norm": 0.45790547132492065, + "learning_rate": 0.00019880367374909436, + "loss": 0.1822, + "step": 2112 + }, + { + "epoch": 0.04953204756295528, + "grad_norm": 0.39906248450279236, + "learning_rate": 0.00019880253748180408, + "loss": 0.0986, + "step": 2113 + }, + { + "epoch": 0.04955548913776028, + "grad_norm": 1.0061113834381104, + "learning_rate": 0.00019880140067840847, + "loss": 0.3622, + "step": 2114 + }, + { + "epoch": 0.04957893071256527, + "grad_norm": 0.6678358912467957, + "learning_rate": 0.00019880026333891367, + "loss": 0.2444, + "step": 2115 + }, + { + "epoch": 0.049602372287370265, + "grad_norm": 0.3792705535888672, + "learning_rate": 0.00019879912546332586, + "loss": 0.1502, + "step": 2116 + }, + { + "epoch": 0.04962581386217526, + "grad_norm": 0.4002247452735901, + "learning_rate": 0.0001987979870516512, + "loss": 0.1693, + "step": 2117 + }, + { + "epoch": 0.049649255436980254, + "grad_norm": 0.6611244678497314, + "learning_rate": 0.0001987968481038959, + "loss": 0.2294, + "step": 2118 + }, + { + "epoch": 0.049672697011785255, + "grad_norm": 0.8062430024147034, + "learning_rate": 0.00019879570862006615, + "loss": 0.2493, + "step": 2119 + }, + { + "epoch": 0.04969613858659025, + "grad_norm": 0.5445297360420227, + "learning_rate": 0.00019879456860016808, + "loss": 1.0332, + "step": 2120 + }, + { + "epoch": 0.049719580161395244, + "grad_norm": 0.5417014956474304, + "learning_rate": 0.0001987934280442079, + "loss": 0.1948, + "step": 2121 + }, + { + "epoch": 0.04974302173620024, + "grad_norm": 0.23567089438438416, + "learning_rate": 0.00019879228695219182, + "loss": 0.0527, + "step": 2122 + }, + { + "epoch": 0.04976646331100523, + "grad_norm": 0.1453011929988861, + "learning_rate": 0.00019879114532412597, + "loss": 0.0399, + "step": 2123 + }, + { + "epoch": 0.04978990488581023, + "grad_norm": 0.2076706439256668, + "learning_rate": 0.00019879000316001664, + "loss": 0.0236, + "step": 2124 + }, + { + "epoch": 0.04981334646061522, + "grad_norm": 0.5446987152099609, + "learning_rate": 0.00019878886045986995, + "loss": 0.2205, + "step": 2125 + }, + { + "epoch": 0.04983678803542022, + "grad_norm": 0.11442125588655472, + "learning_rate": 0.00019878771722369213, + "loss": 0.0463, + "step": 2126 + }, + { + "epoch": 0.04986022961022522, + "grad_norm": 0.8007189631462097, + "learning_rate": 0.00019878657345148937, + "loss": 0.324, + "step": 2127 + }, + { + "epoch": 0.04988367118503021, + "grad_norm": 0.8348619937896729, + "learning_rate": 0.0001987854291432679, + "loss": 0.259, + "step": 2128 + }, + { + "epoch": 0.049907112759835205, + "grad_norm": 0.4529968500137329, + "learning_rate": 0.0001987842842990339, + "loss": 0.1188, + "step": 2129 + }, + { + "epoch": 0.0499305543346402, + "grad_norm": 0.733301043510437, + "learning_rate": 0.0001987831389187936, + "loss": 0.2, + "step": 2130 + }, + { + "epoch": 0.049953995909445194, + "grad_norm": 0.5701751708984375, + "learning_rate": 0.0001987819930025532, + "loss": 0.1722, + "step": 2131 + }, + { + "epoch": 0.049977437484250195, + "grad_norm": 0.1551630198955536, + "learning_rate": 0.00019878084655031896, + "loss": 0.0282, + "step": 2132 + }, + { + "epoch": 0.05000087905905519, + "grad_norm": 0.5466103553771973, + "learning_rate": 0.00019877969956209705, + "loss": 0.1657, + "step": 2133 + }, + { + "epoch": 0.050024320633860184, + "grad_norm": 0.7800057530403137, + "learning_rate": 0.0001987785520378937, + "loss": 0.2307, + "step": 2134 + }, + { + "epoch": 0.05004776220866518, + "grad_norm": 0.45169809460639954, + "learning_rate": 0.00019877740397771518, + "loss": 0.1379, + "step": 2135 + }, + { + "epoch": 0.05007120378347017, + "grad_norm": 0.17302943766117096, + "learning_rate": 0.00019877625538156768, + "loss": 0.0331, + "step": 2136 + }, + { + "epoch": 0.05009464535827517, + "grad_norm": 0.49467283487319946, + "learning_rate": 0.00019877510624945744, + "loss": 0.1532, + "step": 2137 + }, + { + "epoch": 0.05011808693308016, + "grad_norm": 0.2742874324321747, + "learning_rate": 0.0001987739565813907, + "loss": 0.1034, + "step": 2138 + }, + { + "epoch": 0.05014152850788516, + "grad_norm": 0.5844963192939758, + "learning_rate": 0.00019877280637737365, + "loss": 0.135, + "step": 2139 + }, + { + "epoch": 0.050164970082690157, + "grad_norm": 0.662998616695404, + "learning_rate": 0.00019877165563741262, + "loss": 0.7057, + "step": 2140 + }, + { + "epoch": 0.05018841165749515, + "grad_norm": 2.8857638835906982, + "learning_rate": 0.00019877050436151379, + "loss": 0.1166, + "step": 2141 + }, + { + "epoch": 0.050211853232300145, + "grad_norm": 0.781008780002594, + "learning_rate": 0.00019876935254968344, + "loss": 0.1113, + "step": 2142 + }, + { + "epoch": 0.05023529480710514, + "grad_norm": 0.5011653900146484, + "learning_rate": 0.0001987682002019278, + "loss": 0.1143, + "step": 2143 + }, + { + "epoch": 0.050258736381910134, + "grad_norm": 0.39529088139533997, + "learning_rate": 0.00019876704731825313, + "loss": 0.1191, + "step": 2144 + }, + { + "epoch": 0.050282177956715135, + "grad_norm": 0.29195451736450195, + "learning_rate": 0.00019876589389866567, + "loss": 0.1082, + "step": 2145 + }, + { + "epoch": 0.05030561953152013, + "grad_norm": 0.5188444256782532, + "learning_rate": 0.0001987647399431717, + "loss": 0.1487, + "step": 2146 + }, + { + "epoch": 0.050329061106325124, + "grad_norm": 0.6525402069091797, + "learning_rate": 0.0001987635854517775, + "loss": 0.7267, + "step": 2147 + }, + { + "epoch": 0.05035250268113012, + "grad_norm": 0.6707928776741028, + "learning_rate": 0.00019876243042448927, + "loss": 0.2092, + "step": 2148 + }, + { + "epoch": 0.05037594425593511, + "grad_norm": 1.516955852508545, + "learning_rate": 0.00019876127486131336, + "loss": 0.2499, + "step": 2149 + }, + { + "epoch": 0.05039938583074011, + "grad_norm": 0.601209282875061, + "learning_rate": 0.00019876011876225594, + "loss": 0.2593, + "step": 2150 + }, + { + "epoch": 0.05042282740554511, + "grad_norm": 0.5392762422561646, + "learning_rate": 0.0001987589621273234, + "loss": 0.2041, + "step": 2151 + }, + { + "epoch": 0.0504462689803501, + "grad_norm": 0.6597854495048523, + "learning_rate": 0.0001987578049565219, + "loss": 0.1707, + "step": 2152 + }, + { + "epoch": 0.050469710555155096, + "grad_norm": 0.18389223515987396, + "learning_rate": 0.0001987566472498578, + "loss": 0.0455, + "step": 2153 + }, + { + "epoch": 0.05049315212996009, + "grad_norm": 0.24131426215171814, + "learning_rate": 0.00019875548900733737, + "loss": 0.0511, + "step": 2154 + }, + { + "epoch": 0.050516593704765085, + "grad_norm": 0.8738335371017456, + "learning_rate": 0.00019875433022896685, + "loss": 0.2453, + "step": 2155 + }, + { + "epoch": 0.05054003527957008, + "grad_norm": 0.7199626564979553, + "learning_rate": 0.00019875317091475256, + "loss": 0.2226, + "step": 2156 + }, + { + "epoch": 0.050563476854375074, + "grad_norm": 0.7222450375556946, + "learning_rate": 0.00019875201106470082, + "loss": 0.1983, + "step": 2157 + }, + { + "epoch": 0.050586918429180075, + "grad_norm": 0.8086413145065308, + "learning_rate": 0.00019875085067881784, + "loss": 0.2997, + "step": 2158 + }, + { + "epoch": 0.05061036000398507, + "grad_norm": 0.6644537448883057, + "learning_rate": 0.00019874968975710998, + "loss": 0.206, + "step": 2159 + }, + { + "epoch": 0.050633801578790064, + "grad_norm": 0.815019965171814, + "learning_rate": 0.00019874852829958358, + "loss": 0.1939, + "step": 2160 + }, + { + "epoch": 0.05065724315359506, + "grad_norm": 0.6278868913650513, + "learning_rate": 0.00019874736630624482, + "loss": 0.1805, + "step": 2161 + }, + { + "epoch": 0.05068068472840005, + "grad_norm": 1.093216896057129, + "learning_rate": 0.00019874620377710012, + "loss": 0.2046, + "step": 2162 + }, + { + "epoch": 0.05070412630320505, + "grad_norm": 0.8395475149154663, + "learning_rate": 0.00019874504071215574, + "loss": 0.3573, + "step": 2163 + }, + { + "epoch": 0.05072756787801005, + "grad_norm": 0.688896119594574, + "learning_rate": 0.00019874387711141798, + "loss": 0.1785, + "step": 2164 + }, + { + "epoch": 0.05075100945281504, + "grad_norm": 0.18212299048900604, + "learning_rate": 0.00019874271297489314, + "loss": 0.0545, + "step": 2165 + }, + { + "epoch": 0.050774451027620036, + "grad_norm": 0.8580446243286133, + "learning_rate": 0.0001987415483025876, + "loss": 0.9196, + "step": 2166 + }, + { + "epoch": 0.05079789260242503, + "grad_norm": 0.7333237528800964, + "learning_rate": 0.00019874038309450765, + "loss": 0.2229, + "step": 2167 + }, + { + "epoch": 0.050821334177230025, + "grad_norm": 0.7817240953445435, + "learning_rate": 0.00019873921735065957, + "loss": 0.2416, + "step": 2168 + }, + { + "epoch": 0.05084477575203502, + "grad_norm": 0.5492011904716492, + "learning_rate": 0.00019873805107104974, + "loss": 0.1341, + "step": 2169 + }, + { + "epoch": 0.050868217326840014, + "grad_norm": 0.49294552206993103, + "learning_rate": 0.00019873688425568447, + "loss": 0.2117, + "step": 2170 + }, + { + "epoch": 0.050891658901645015, + "grad_norm": 0.49254781007766724, + "learning_rate": 0.00019873571690457006, + "loss": 0.1438, + "step": 2171 + }, + { + "epoch": 0.05091510047645001, + "grad_norm": 0.4338245987892151, + "learning_rate": 0.0001987345490177129, + "loss": 0.9601, + "step": 2172 + }, + { + "epoch": 0.050938542051255004, + "grad_norm": 0.7286648154258728, + "learning_rate": 0.0001987333805951193, + "loss": 0.1766, + "step": 2173 + }, + { + "epoch": 0.05096198362606, + "grad_norm": 0.718829333782196, + "learning_rate": 0.0001987322116367956, + "loss": 0.2424, + "step": 2174 + }, + { + "epoch": 0.05098542520086499, + "grad_norm": 0.3685149550437927, + "learning_rate": 0.00019873104214274815, + "loss": 0.4075, + "step": 2175 + }, + { + "epoch": 0.051008866775669987, + "grad_norm": 0.5036562085151672, + "learning_rate": 0.00019872987211298328, + "loss": 0.1386, + "step": 2176 + }, + { + "epoch": 0.05103230835047499, + "grad_norm": 0.38913625478744507, + "learning_rate": 0.00019872870154750735, + "loss": 0.1305, + "step": 2177 + }, + { + "epoch": 0.05105574992527998, + "grad_norm": 0.2680993378162384, + "learning_rate": 0.0001987275304463267, + "loss": 0.0684, + "step": 2178 + }, + { + "epoch": 0.051079191500084976, + "grad_norm": 0.6105185747146606, + "learning_rate": 0.0001987263588094477, + "loss": 0.161, + "step": 2179 + }, + { + "epoch": 0.05110263307488997, + "grad_norm": 0.27071303129196167, + "learning_rate": 0.0001987251866368767, + "loss": 0.0909, + "step": 2180 + }, + { + "epoch": 0.051126074649694965, + "grad_norm": 0.2814388573169708, + "learning_rate": 0.0001987240139286201, + "loss": 0.0737, + "step": 2181 + }, + { + "epoch": 0.05114951622449996, + "grad_norm": 0.9963940978050232, + "learning_rate": 0.0001987228406846842, + "loss": 0.23, + "step": 2182 + }, + { + "epoch": 0.05117295779930496, + "grad_norm": 0.5999456644058228, + "learning_rate": 0.00019872166690507536, + "loss": 0.1835, + "step": 2183 + }, + { + "epoch": 0.051196399374109955, + "grad_norm": 0.3077094554901123, + "learning_rate": 0.0001987204925898, + "loss": 0.1196, + "step": 2184 + }, + { + "epoch": 0.05121984094891495, + "grad_norm": 0.5308026671409607, + "learning_rate": 0.00019871931773886447, + "loss": 0.1504, + "step": 2185 + }, + { + "epoch": 0.051243282523719944, + "grad_norm": 0.5905784964561462, + "learning_rate": 0.00019871814235227517, + "loss": 0.1545, + "step": 2186 + }, + { + "epoch": 0.05126672409852494, + "grad_norm": 0.6928746700286865, + "learning_rate": 0.00019871696643003844, + "loss": 0.173, + "step": 2187 + }, + { + "epoch": 0.05129016567332993, + "grad_norm": 0.6738356947898865, + "learning_rate": 0.00019871578997216065, + "loss": 0.2138, + "step": 2188 + }, + { + "epoch": 0.051313607248134926, + "grad_norm": 0.6120879650115967, + "learning_rate": 0.0001987146129786482, + "loss": 0.2147, + "step": 2189 + }, + { + "epoch": 0.05133704882293993, + "grad_norm": 0.7318762540817261, + "learning_rate": 0.00019871343544950753, + "loss": 0.2601, + "step": 2190 + }, + { + "epoch": 0.05136049039774492, + "grad_norm": 0.7671405673027039, + "learning_rate": 0.00019871225738474498, + "loss": 0.7851, + "step": 2191 + }, + { + "epoch": 0.051383931972549916, + "grad_norm": 0.6975974440574646, + "learning_rate": 0.00019871107878436693, + "loss": 0.17, + "step": 2192 + }, + { + "epoch": 0.05140737354735491, + "grad_norm": 0.6786423921585083, + "learning_rate": 0.00019870989964837978, + "loss": 0.231, + "step": 2193 + }, + { + "epoch": 0.051430815122159905, + "grad_norm": 0.18676602840423584, + "learning_rate": 0.00019870871997678996, + "loss": 0.0576, + "step": 2194 + }, + { + "epoch": 0.0514542566969649, + "grad_norm": 0.550095796585083, + "learning_rate": 0.00019870753976960383, + "loss": 0.1917, + "step": 2195 + }, + { + "epoch": 0.0514776982717699, + "grad_norm": 0.4834291636943817, + "learning_rate": 0.00019870635902682784, + "loss": 0.1092, + "step": 2196 + }, + { + "epoch": 0.051501139846574895, + "grad_norm": 0.7989000678062439, + "learning_rate": 0.00019870517774846834, + "loss": 0.2739, + "step": 2197 + }, + { + "epoch": 0.05152458142137989, + "grad_norm": 0.6990967988967896, + "learning_rate": 0.00019870399593453178, + "loss": 0.5917, + "step": 2198 + }, + { + "epoch": 0.051548022996184883, + "grad_norm": 0.413695752620697, + "learning_rate": 0.00019870281358502458, + "loss": 0.1509, + "step": 2199 + }, + { + "epoch": 0.05157146457098988, + "grad_norm": 0.5161677598953247, + "learning_rate": 0.0001987016306999531, + "loss": 0.1373, + "step": 2200 + }, + { + "epoch": 0.05159490614579487, + "grad_norm": 0.19090932607650757, + "learning_rate": 0.00019870044727932383, + "loss": 0.0609, + "step": 2201 + }, + { + "epoch": 0.05161834772059987, + "grad_norm": 0.28352805972099304, + "learning_rate": 0.00019869926332314317, + "loss": 0.0961, + "step": 2202 + }, + { + "epoch": 0.05164178929540487, + "grad_norm": 0.36746907234191895, + "learning_rate": 0.0001986980788314175, + "loss": 0.0806, + "step": 2203 + }, + { + "epoch": 0.05166523087020986, + "grad_norm": 0.10974553227424622, + "learning_rate": 0.00019869689380415327, + "loss": 0.027, + "step": 2204 + }, + { + "epoch": 0.051688672445014856, + "grad_norm": 0.5859951972961426, + "learning_rate": 0.00019869570824135698, + "loss": 0.1276, + "step": 2205 + }, + { + "epoch": 0.05171211401981985, + "grad_norm": 0.6472352147102356, + "learning_rate": 0.0001986945221430349, + "loss": 0.1648, + "step": 2206 + }, + { + "epoch": 0.051735555594624845, + "grad_norm": 0.37484678626060486, + "learning_rate": 0.00019869333550919367, + "loss": 0.1584, + "step": 2207 + }, + { + "epoch": 0.05175899716942984, + "grad_norm": 0.5457265973091125, + "learning_rate": 0.00019869214833983959, + "loss": 0.2, + "step": 2208 + }, + { + "epoch": 0.05178243874423484, + "grad_norm": 0.7368552088737488, + "learning_rate": 0.00019869096063497913, + "loss": 0.1289, + "step": 2209 + }, + { + "epoch": 0.051805880319039835, + "grad_norm": 0.5050601363182068, + "learning_rate": 0.00019868977239461874, + "loss": 0.1973, + "step": 2210 + }, + { + "epoch": 0.05182932189384483, + "grad_norm": 0.7067375183105469, + "learning_rate": 0.00019868858361876487, + "loss": 0.1625, + "step": 2211 + }, + { + "epoch": 0.05185276346864982, + "grad_norm": 0.3710794448852539, + "learning_rate": 0.00019868739430742397, + "loss": 0.1127, + "step": 2212 + }, + { + "epoch": 0.05187620504345482, + "grad_norm": 0.6895796656608582, + "learning_rate": 0.0001986862044606025, + "loss": 0.124, + "step": 2213 + }, + { + "epoch": 0.05189964661825981, + "grad_norm": 0.36536723375320435, + "learning_rate": 0.00019868501407830692, + "loss": 0.0989, + "step": 2214 + }, + { + "epoch": 0.05192308819306481, + "grad_norm": 0.4762769341468811, + "learning_rate": 0.00019868382316054367, + "loss": 0.5323, + "step": 2215 + }, + { + "epoch": 0.05194652976786981, + "grad_norm": 0.5482944846153259, + "learning_rate": 0.00019868263170731923, + "loss": 0.1245, + "step": 2216 + }, + { + "epoch": 0.0519699713426748, + "grad_norm": 0.37369439005851746, + "learning_rate": 0.00019868143971864006, + "loss": 0.1157, + "step": 2217 + }, + { + "epoch": 0.051993412917479796, + "grad_norm": 0.5988585948944092, + "learning_rate": 0.0001986802471945126, + "loss": 0.7167, + "step": 2218 + }, + { + "epoch": 0.05201685449228479, + "grad_norm": 0.5382434129714966, + "learning_rate": 0.00019867905413494335, + "loss": 0.7087, + "step": 2219 + }, + { + "epoch": 0.052040296067089785, + "grad_norm": 0.51520174741745, + "learning_rate": 0.00019867786053993878, + "loss": 0.178, + "step": 2220 + }, + { + "epoch": 0.05206373764189478, + "grad_norm": 0.31004101037979126, + "learning_rate": 0.00019867666640950536, + "loss": 0.087, + "step": 2221 + }, + { + "epoch": 0.05208717921669978, + "grad_norm": 0.5697863698005676, + "learning_rate": 0.0001986754717436496, + "loss": 0.0954, + "step": 2222 + }, + { + "epoch": 0.052110620791504775, + "grad_norm": 0.1721567064523697, + "learning_rate": 0.00019867427654237792, + "loss": 0.0581, + "step": 2223 + }, + { + "epoch": 0.05213406236630977, + "grad_norm": 0.3260755240917206, + "learning_rate": 0.00019867308080569686, + "loss": 0.0828, + "step": 2224 + }, + { + "epoch": 0.05215750394111476, + "grad_norm": 0.4178527891635895, + "learning_rate": 0.00019867188453361287, + "loss": 0.1679, + "step": 2225 + }, + { + "epoch": 0.05218094551591976, + "grad_norm": 0.9203223586082458, + "learning_rate": 0.0001986706877261325, + "loss": 0.2743, + "step": 2226 + }, + { + "epoch": 0.05220438709072475, + "grad_norm": 0.34570714831352234, + "learning_rate": 0.00019866949038326216, + "loss": 0.0696, + "step": 2227 + }, + { + "epoch": 0.05222782866552975, + "grad_norm": 0.481893926858902, + "learning_rate": 0.00019866829250500843, + "loss": 0.2162, + "step": 2228 + }, + { + "epoch": 0.05225127024033475, + "grad_norm": 0.7239258885383606, + "learning_rate": 0.00019866709409137773, + "loss": 0.2336, + "step": 2229 + }, + { + "epoch": 0.05227471181513974, + "grad_norm": 0.4879460036754608, + "learning_rate": 0.00019866589514237664, + "loss": 0.1719, + "step": 2230 + }, + { + "epoch": 0.052298153389944736, + "grad_norm": 0.5051473379135132, + "learning_rate": 0.0001986646956580116, + "loss": 0.1531, + "step": 2231 + }, + { + "epoch": 0.05232159496474973, + "grad_norm": 0.9794089198112488, + "learning_rate": 0.00019866349563828916, + "loss": 0.2514, + "step": 2232 + }, + { + "epoch": 0.052345036539554725, + "grad_norm": 0.7525996565818787, + "learning_rate": 0.0001986622950832158, + "loss": 0.1835, + "step": 2233 + }, + { + "epoch": 0.052368478114359726, + "grad_norm": 0.21176040172576904, + "learning_rate": 0.0001986610939927981, + "loss": 0.0472, + "step": 2234 + }, + { + "epoch": 0.05239191968916472, + "grad_norm": 0.5788783431053162, + "learning_rate": 0.0001986598923670425, + "loss": 0.1564, + "step": 2235 + }, + { + "epoch": 0.052415361263969715, + "grad_norm": 0.3714841604232788, + "learning_rate": 0.00019865869020595553, + "loss": 0.1132, + "step": 2236 + }, + { + "epoch": 0.05243880283877471, + "grad_norm": 0.30418071150779724, + "learning_rate": 0.00019865748750954375, + "loss": 0.0855, + "step": 2237 + }, + { + "epoch": 0.0524622444135797, + "grad_norm": 0.9492006897926331, + "learning_rate": 0.00019865628427781368, + "loss": 0.244, + "step": 2238 + }, + { + "epoch": 0.0524856859883847, + "grad_norm": 0.19368426501750946, + "learning_rate": 0.00019865508051077182, + "loss": 0.0641, + "step": 2239 + }, + { + "epoch": 0.05250912756318969, + "grad_norm": 0.6158435940742493, + "learning_rate": 0.00019865387620842471, + "loss": 0.1847, + "step": 2240 + }, + { + "epoch": 0.05253256913799469, + "grad_norm": 0.4737758934497833, + "learning_rate": 0.00019865267137077894, + "loss": 0.1167, + "step": 2241 + }, + { + "epoch": 0.05255601071279969, + "grad_norm": 0.3220529556274414, + "learning_rate": 0.00019865146599784097, + "loss": 0.0472, + "step": 2242 + }, + { + "epoch": 0.05257945228760468, + "grad_norm": 0.6433259844779968, + "learning_rate": 0.00019865026008961738, + "loss": 0.1731, + "step": 2243 + }, + { + "epoch": 0.052602893862409676, + "grad_norm": 0.5422577857971191, + "learning_rate": 0.0001986490536461147, + "loss": 0.2258, + "step": 2244 + }, + { + "epoch": 0.05262633543721467, + "grad_norm": 1.0530593395233154, + "learning_rate": 0.00019864784666733947, + "loss": 0.21, + "step": 2245 + }, + { + "epoch": 0.052649777012019665, + "grad_norm": 0.7974436283111572, + "learning_rate": 0.00019864663915329826, + "loss": 0.1778, + "step": 2246 + }, + { + "epoch": 0.052673218586824666, + "grad_norm": 0.8156370520591736, + "learning_rate": 0.0001986454311039976, + "loss": 0.2311, + "step": 2247 + }, + { + "epoch": 0.05269666016162966, + "grad_norm": 0.6229825615882874, + "learning_rate": 0.0001986442225194441, + "loss": 0.1883, + "step": 2248 + }, + { + "epoch": 0.052720101736434655, + "grad_norm": 0.6397926807403564, + "learning_rate": 0.00019864301339964424, + "loss": 0.758, + "step": 2249 + }, + { + "epoch": 0.05274354331123965, + "grad_norm": 0.3806785047054291, + "learning_rate": 0.00019864180374460464, + "loss": 0.1174, + "step": 2250 + }, + { + "epoch": 0.05276698488604464, + "grad_norm": 0.40822911262512207, + "learning_rate": 0.0001986405935543318, + "loss": 0.1184, + "step": 2251 + }, + { + "epoch": 0.05279042646084964, + "grad_norm": 0.6263301968574524, + "learning_rate": 0.00019863938282883235, + "loss": 0.2513, + "step": 2252 + }, + { + "epoch": 0.05281386803565463, + "grad_norm": 0.7235010266304016, + "learning_rate": 0.00019863817156811283, + "loss": 0.1883, + "step": 2253 + }, + { + "epoch": 0.05283730961045963, + "grad_norm": 0.35637253522872925, + "learning_rate": 0.0001986369597721798, + "loss": 0.1213, + "step": 2254 + }, + { + "epoch": 0.05286075118526463, + "grad_norm": 0.5202324986457825, + "learning_rate": 0.0001986357474410399, + "loss": 0.1648, + "step": 2255 + }, + { + "epoch": 0.05288419276006962, + "grad_norm": 0.2967534363269806, + "learning_rate": 0.00019863453457469966, + "loss": 0.0759, + "step": 2256 + }, + { + "epoch": 0.052907634334874616, + "grad_norm": 0.7514863610267639, + "learning_rate": 0.0001986333211731656, + "loss": 0.1715, + "step": 2257 + }, + { + "epoch": 0.05293107590967961, + "grad_norm": 0.3684980571269989, + "learning_rate": 0.0001986321072364444, + "loss": 0.2842, + "step": 2258 + }, + { + "epoch": 0.052954517484484605, + "grad_norm": 0.24615886807441711, + "learning_rate": 0.00019863089276454262, + "loss": 0.0279, + "step": 2259 + }, + { + "epoch": 0.052977959059289606, + "grad_norm": 0.3624086380004883, + "learning_rate": 0.00019862967775746685, + "loss": 0.167, + "step": 2260 + }, + { + "epoch": 0.0530014006340946, + "grad_norm": 0.6718865036964417, + "learning_rate": 0.00019862846221522365, + "loss": 0.1888, + "step": 2261 + }, + { + "epoch": 0.053024842208899595, + "grad_norm": 0.6568745374679565, + "learning_rate": 0.00019862724613781967, + "loss": 0.8207, + "step": 2262 + }, + { + "epoch": 0.05304828378370459, + "grad_norm": 0.5137897729873657, + "learning_rate": 0.00019862602952526145, + "loss": 0.1444, + "step": 2263 + }, + { + "epoch": 0.05307172535850958, + "grad_norm": 0.29411306977272034, + "learning_rate": 0.00019862481237755563, + "loss": 0.1193, + "step": 2264 + }, + { + "epoch": 0.05309516693331458, + "grad_norm": 0.5452072024345398, + "learning_rate": 0.00019862359469470883, + "loss": 0.1574, + "step": 2265 + }, + { + "epoch": 0.05311860850811958, + "grad_norm": 0.269125759601593, + "learning_rate": 0.00019862237647672758, + "loss": 0.1225, + "step": 2266 + }, + { + "epoch": 0.05314205008292457, + "grad_norm": 0.9626508355140686, + "learning_rate": 0.0001986211577236186, + "loss": 0.2143, + "step": 2267 + }, + { + "epoch": 0.05316549165772957, + "grad_norm": 0.141776904463768, + "learning_rate": 0.0001986199384353884, + "loss": 0.058, + "step": 2268 + }, + { + "epoch": 0.05318893323253456, + "grad_norm": 0.5136770009994507, + "learning_rate": 0.00019861871861204367, + "loss": 0.0979, + "step": 2269 + }, + { + "epoch": 0.053212374807339556, + "grad_norm": 0.6499060392379761, + "learning_rate": 0.00019861749825359097, + "loss": 0.2524, + "step": 2270 + }, + { + "epoch": 0.05323581638214455, + "grad_norm": 0.16377583146095276, + "learning_rate": 0.000198616277360037, + "loss": 0.0708, + "step": 2271 + }, + { + "epoch": 0.053259257956949545, + "grad_norm": 0.6038290858268738, + "learning_rate": 0.00019861505593138826, + "loss": 0.1798, + "step": 2272 + }, + { + "epoch": 0.053282699531754546, + "grad_norm": 0.167611226439476, + "learning_rate": 0.00019861383396765155, + "loss": 0.048, + "step": 2273 + }, + { + "epoch": 0.05330614110655954, + "grad_norm": 0.41910091042518616, + "learning_rate": 0.00019861261146883333, + "loss": 0.1003, + "step": 2274 + }, + { + "epoch": 0.053329582681364535, + "grad_norm": 0.4460001587867737, + "learning_rate": 0.00019861138843494034, + "loss": 0.1702, + "step": 2275 + }, + { + "epoch": 0.05335302425616953, + "grad_norm": 0.5871982574462891, + "learning_rate": 0.00019861016486597918, + "loss": 0.2203, + "step": 2276 + }, + { + "epoch": 0.05337646583097452, + "grad_norm": 0.6683955788612366, + "learning_rate": 0.00019860894076195647, + "loss": 0.3143, + "step": 2277 + }, + { + "epoch": 0.05339990740577952, + "grad_norm": 0.266801118850708, + "learning_rate": 0.00019860771612287888, + "loss": 0.0975, + "step": 2278 + }, + { + "epoch": 0.05342334898058452, + "grad_norm": 0.15415488183498383, + "learning_rate": 0.0001986064909487531, + "loss": 0.052, + "step": 2279 + }, + { + "epoch": 0.05344679055538951, + "grad_norm": 0.4620078504085541, + "learning_rate": 0.0001986052652395857, + "loss": 0.17, + "step": 2280 + }, + { + "epoch": 0.05347023213019451, + "grad_norm": 0.5433036684989929, + "learning_rate": 0.00019860403899538332, + "loss": 0.1797, + "step": 2281 + }, + { + "epoch": 0.0534936737049995, + "grad_norm": 0.5093512535095215, + "learning_rate": 0.00019860281221615269, + "loss": 0.1708, + "step": 2282 + }, + { + "epoch": 0.053517115279804496, + "grad_norm": 0.8322006464004517, + "learning_rate": 0.00019860158490190042, + "loss": 0.3001, + "step": 2283 + }, + { + "epoch": 0.05354055685460949, + "grad_norm": 0.5627433657646179, + "learning_rate": 0.00019860035705263318, + "loss": 0.1656, + "step": 2284 + }, + { + "epoch": 0.053563998429414485, + "grad_norm": 0.8625827431678772, + "learning_rate": 0.0001985991286683576, + "loss": 0.1257, + "step": 2285 + }, + { + "epoch": 0.053587440004219486, + "grad_norm": 0.2179257571697235, + "learning_rate": 0.0001985978997490804, + "loss": 0.0634, + "step": 2286 + }, + { + "epoch": 0.05361088157902448, + "grad_norm": 0.6341155767440796, + "learning_rate": 0.00019859667029480823, + "loss": 0.18, + "step": 2287 + }, + { + "epoch": 0.053634323153829475, + "grad_norm": 0.34979182481765747, + "learning_rate": 0.00019859544030554775, + "loss": 0.0711, + "step": 2288 + }, + { + "epoch": 0.05365776472863447, + "grad_norm": 0.4277438819408417, + "learning_rate": 0.00019859420978130565, + "loss": 0.1368, + "step": 2289 + }, + { + "epoch": 0.05368120630343946, + "grad_norm": 0.7987305521965027, + "learning_rate": 0.0001985929787220886, + "loss": 0.2347, + "step": 2290 + }, + { + "epoch": 0.05370464787824446, + "grad_norm": 0.5988703966140747, + "learning_rate": 0.00019859174712790324, + "loss": 0.1509, + "step": 2291 + }, + { + "epoch": 0.05372808945304946, + "grad_norm": 0.7971715927124023, + "learning_rate": 0.00019859051499875632, + "loss": 0.2084, + "step": 2292 + }, + { + "epoch": 0.05375153102785445, + "grad_norm": 0.5874369740486145, + "learning_rate": 0.00019858928233465448, + "loss": 0.2614, + "step": 2293 + }, + { + "epoch": 0.05377497260265945, + "grad_norm": 0.5413160920143127, + "learning_rate": 0.00019858804913560442, + "loss": 0.139, + "step": 2294 + }, + { + "epoch": 0.05379841417746444, + "grad_norm": 0.5715190768241882, + "learning_rate": 0.00019858681540161285, + "loss": 0.1631, + "step": 2295 + }, + { + "epoch": 0.053821855752269436, + "grad_norm": 0.6513562798500061, + "learning_rate": 0.00019858558113268644, + "loss": 0.5033, + "step": 2296 + }, + { + "epoch": 0.05384529732707443, + "grad_norm": 1.145174503326416, + "learning_rate": 0.00019858434632883187, + "loss": 0.3368, + "step": 2297 + }, + { + "epoch": 0.05386873890187943, + "grad_norm": 0.9468703269958496, + "learning_rate": 0.00019858311099005588, + "loss": 0.2102, + "step": 2298 + }, + { + "epoch": 0.053892180476684426, + "grad_norm": 1.0505231618881226, + "learning_rate": 0.00019858187511636517, + "loss": 0.2817, + "step": 2299 + }, + { + "epoch": 0.05391562205148942, + "grad_norm": 0.5671220421791077, + "learning_rate": 0.00019858063870776643, + "loss": 0.1757, + "step": 2300 + }, + { + "epoch": 0.053939063626294415, + "grad_norm": 0.4889351427555084, + "learning_rate": 0.00019857940176426638, + "loss": 0.0907, + "step": 2301 + }, + { + "epoch": 0.05396250520109941, + "grad_norm": 0.7746083736419678, + "learning_rate": 0.0001985781642858717, + "loss": 0.1906, + "step": 2302 + }, + { + "epoch": 0.0539859467759044, + "grad_norm": 0.18350230157375336, + "learning_rate": 0.00019857692627258912, + "loss": 0.0476, + "step": 2303 + }, + { + "epoch": 0.0540093883507094, + "grad_norm": 0.9205737113952637, + "learning_rate": 0.0001985756877244254, + "loss": 0.6328, + "step": 2304 + }, + { + "epoch": 0.0540328299255144, + "grad_norm": 0.9389607906341553, + "learning_rate": 0.00019857444864138722, + "loss": 0.3429, + "step": 2305 + }, + { + "epoch": 0.05405627150031939, + "grad_norm": 0.13248978555202484, + "learning_rate": 0.00019857320902348131, + "loss": 0.0267, + "step": 2306 + }, + { + "epoch": 0.05407971307512439, + "grad_norm": 0.5779540538787842, + "learning_rate": 0.00019857196887071438, + "loss": 0.1441, + "step": 2307 + }, + { + "epoch": 0.05410315464992938, + "grad_norm": 0.6001352071762085, + "learning_rate": 0.0001985707281830932, + "loss": 0.171, + "step": 2308 + }, + { + "epoch": 0.054126596224734376, + "grad_norm": 0.32884681224823, + "learning_rate": 0.00019856948696062444, + "loss": 0.0752, + "step": 2309 + }, + { + "epoch": 0.05415003779953937, + "grad_norm": 0.4972202181816101, + "learning_rate": 0.00019856824520331491, + "loss": 0.1764, + "step": 2310 + }, + { + "epoch": 0.05417347937434437, + "grad_norm": 0.5011967420578003, + "learning_rate": 0.0001985670029111713, + "loss": 0.1481, + "step": 2311 + }, + { + "epoch": 0.054196920949149366, + "grad_norm": 0.5646238327026367, + "learning_rate": 0.00019856576008420035, + "loss": 0.1504, + "step": 2312 + }, + { + "epoch": 0.05422036252395436, + "grad_norm": 0.6572834849357605, + "learning_rate": 0.00019856451672240884, + "loss": 0.1272, + "step": 2313 + }, + { + "epoch": 0.054243804098759354, + "grad_norm": 0.15241190791130066, + "learning_rate": 0.00019856327282580347, + "loss": 0.0283, + "step": 2314 + }, + { + "epoch": 0.05426724567356435, + "grad_norm": 0.4788760244846344, + "learning_rate": 0.000198562028394391, + "loss": 0.072, + "step": 2315 + }, + { + "epoch": 0.05429068724836934, + "grad_norm": 0.7805551290512085, + "learning_rate": 0.00019856078342817818, + "loss": 0.6098, + "step": 2316 + }, + { + "epoch": 0.05431412882317434, + "grad_norm": 0.7757556438446045, + "learning_rate": 0.0001985595379271718, + "loss": 0.3006, + "step": 2317 + }, + { + "epoch": 0.05433757039797934, + "grad_norm": 0.7176744341850281, + "learning_rate": 0.00019855829189137863, + "loss": 0.143, + "step": 2318 + }, + { + "epoch": 0.05436101197278433, + "grad_norm": 0.5866415500640869, + "learning_rate": 0.00019855704532080535, + "loss": 0.6473, + "step": 2319 + }, + { + "epoch": 0.05438445354758933, + "grad_norm": 0.9018004536628723, + "learning_rate": 0.0001985557982154588, + "loss": 0.1627, + "step": 2320 + }, + { + "epoch": 0.05440789512239432, + "grad_norm": 0.7108887434005737, + "learning_rate": 0.00019855455057534571, + "loss": 0.6954, + "step": 2321 + }, + { + "epoch": 0.054431336697199316, + "grad_norm": 0.4861343502998352, + "learning_rate": 0.00019855330240047282, + "loss": 0.1932, + "step": 2322 + }, + { + "epoch": 0.05445477827200431, + "grad_norm": 0.29019609093666077, + "learning_rate": 0.000198552053690847, + "loss": 0.0774, + "step": 2323 + }, + { + "epoch": 0.05447821984680931, + "grad_norm": 0.4545411169528961, + "learning_rate": 0.0001985508044464749, + "loss": 0.1163, + "step": 2324 + }, + { + "epoch": 0.054501661421614306, + "grad_norm": 0.776145339012146, + "learning_rate": 0.00019854955466736343, + "loss": 0.1053, + "step": 2325 + }, + { + "epoch": 0.0545251029964193, + "grad_norm": 0.6379009485244751, + "learning_rate": 0.00019854830435351927, + "loss": 0.193, + "step": 2326 + }, + { + "epoch": 0.054548544571224294, + "grad_norm": 0.6198801398277283, + "learning_rate": 0.00019854705350494926, + "loss": 0.1761, + "step": 2327 + }, + { + "epoch": 0.05457198614602929, + "grad_norm": 0.4884800314903259, + "learning_rate": 0.00019854580212166014, + "loss": 0.141, + "step": 2328 + }, + { + "epoch": 0.05459542772083428, + "grad_norm": 0.6641183495521545, + "learning_rate": 0.00019854455020365874, + "loss": 0.2023, + "step": 2329 + }, + { + "epoch": 0.054618869295639284, + "grad_norm": 0.635323166847229, + "learning_rate": 0.00019854329775095185, + "loss": 0.161, + "step": 2330 + }, + { + "epoch": 0.05464231087044428, + "grad_norm": 0.6058541536331177, + "learning_rate": 0.00019854204476354626, + "loss": 0.2256, + "step": 2331 + }, + { + "epoch": 0.05466575244524927, + "grad_norm": 0.5898990631103516, + "learning_rate": 0.00019854079124144872, + "loss": 0.2244, + "step": 2332 + }, + { + "epoch": 0.05468919402005427, + "grad_norm": 0.5205602645874023, + "learning_rate": 0.00019853953718466613, + "loss": 0.2042, + "step": 2333 + }, + { + "epoch": 0.05471263559485926, + "grad_norm": 0.15347446501255035, + "learning_rate": 0.0001985382825932052, + "loss": 0.0489, + "step": 2334 + }, + { + "epoch": 0.054736077169664256, + "grad_norm": 0.7468143105506897, + "learning_rate": 0.00019853702746707282, + "loss": 0.165, + "step": 2335 + }, + { + "epoch": 0.05475951874446925, + "grad_norm": 0.4155959188938141, + "learning_rate": 0.0001985357718062757, + "loss": 0.1051, + "step": 2336 + }, + { + "epoch": 0.05478296031927425, + "grad_norm": 0.29167869687080383, + "learning_rate": 0.00019853451561082074, + "loss": 0.0737, + "step": 2337 + }, + { + "epoch": 0.054806401894079246, + "grad_norm": 0.5965293049812317, + "learning_rate": 0.00019853325888071475, + "loss": 0.1639, + "step": 2338 + }, + { + "epoch": 0.05482984346888424, + "grad_norm": 0.5956575274467468, + "learning_rate": 0.0001985320016159645, + "loss": 0.2366, + "step": 2339 + }, + { + "epoch": 0.054853285043689234, + "grad_norm": 0.5370386242866516, + "learning_rate": 0.00019853074381657684, + "loss": 0.7048, + "step": 2340 + }, + { + "epoch": 0.05487672661849423, + "grad_norm": 0.5763437151908875, + "learning_rate": 0.0001985294854825586, + "loss": 0.1767, + "step": 2341 + }, + { + "epoch": 0.05490016819329922, + "grad_norm": 0.5927015542984009, + "learning_rate": 0.0001985282266139166, + "loss": 0.3857, + "step": 2342 + }, + { + "epoch": 0.054923609768104224, + "grad_norm": 0.4068201184272766, + "learning_rate": 0.00019852696721065768, + "loss": 0.1697, + "step": 2343 + }, + { + "epoch": 0.05494705134290922, + "grad_norm": 0.17072567343711853, + "learning_rate": 0.00019852570727278866, + "loss": 0.05, + "step": 2344 + }, + { + "epoch": 0.05497049291771421, + "grad_norm": 0.15568053722381592, + "learning_rate": 0.0001985244468003164, + "loss": 0.0477, + "step": 2345 + }, + { + "epoch": 0.05499393449251921, + "grad_norm": 0.3294607102870941, + "learning_rate": 0.00019852318579324769, + "loss": 0.0898, + "step": 2346 + }, + { + "epoch": 0.0550173760673242, + "grad_norm": 0.5457280278205872, + "learning_rate": 0.0001985219242515894, + "loss": 0.1578, + "step": 2347 + }, + { + "epoch": 0.055040817642129196, + "grad_norm": 0.563982367515564, + "learning_rate": 0.00019852066217534842, + "loss": 0.2804, + "step": 2348 + }, + { + "epoch": 0.05506425921693419, + "grad_norm": 0.35728177428245544, + "learning_rate": 0.00019851939956453153, + "loss": 0.1648, + "step": 2349 + }, + { + "epoch": 0.05508770079173919, + "grad_norm": 0.2787947654724121, + "learning_rate": 0.0001985181364191456, + "loss": 0.0617, + "step": 2350 + }, + { + "epoch": 0.055111142366544186, + "grad_norm": 0.2315514236688614, + "learning_rate": 0.00019851687273919751, + "loss": 0.0676, + "step": 2351 + }, + { + "epoch": 0.05513458394134918, + "grad_norm": 0.34581369161605835, + "learning_rate": 0.00019851560852469411, + "loss": 0.0808, + "step": 2352 + }, + { + "epoch": 0.055158025516154174, + "grad_norm": 0.11497363448143005, + "learning_rate": 0.0001985143437756422, + "loss": 0.0275, + "step": 2353 + }, + { + "epoch": 0.05518146709095917, + "grad_norm": 0.42557135224342346, + "learning_rate": 0.00019851307849204873, + "loss": 0.1581, + "step": 2354 + }, + { + "epoch": 0.05520490866576416, + "grad_norm": 0.6940592527389526, + "learning_rate": 0.00019851181267392052, + "loss": 0.2215, + "step": 2355 + }, + { + "epoch": 0.055228350240569164, + "grad_norm": 0.5116882920265198, + "learning_rate": 0.00019851054632126443, + "loss": 0.1318, + "step": 2356 + }, + { + "epoch": 0.05525179181537416, + "grad_norm": 1.0079264640808105, + "learning_rate": 0.00019850927943408734, + "loss": 0.2162, + "step": 2357 + }, + { + "epoch": 0.05527523339017915, + "grad_norm": 0.26263925433158875, + "learning_rate": 0.00019850801201239617, + "loss": 0.05, + "step": 2358 + }, + { + "epoch": 0.05529867496498415, + "grad_norm": 0.30153658986091614, + "learning_rate": 0.00019850674405619774, + "loss": 0.1072, + "step": 2359 + }, + { + "epoch": 0.05532211653978914, + "grad_norm": 0.23645247519016266, + "learning_rate": 0.00019850547556549893, + "loss": 0.081, + "step": 2360 + }, + { + "epoch": 0.055345558114594136, + "grad_norm": 0.2712523639202118, + "learning_rate": 0.00019850420654030665, + "loss": 0.0724, + "step": 2361 + }, + { + "epoch": 0.05536899968939914, + "grad_norm": 0.43684980273246765, + "learning_rate": 0.00019850293698062777, + "loss": 0.1162, + "step": 2362 + }, + { + "epoch": 0.05539244126420413, + "grad_norm": 0.7960320115089417, + "learning_rate": 0.0001985016668864692, + "loss": 0.2023, + "step": 2363 + }, + { + "epoch": 0.055415882839009126, + "grad_norm": 0.4832562804222107, + "learning_rate": 0.0001985003962578378, + "loss": 0.6019, + "step": 2364 + }, + { + "epoch": 0.05543932441381412, + "grad_norm": 0.6829421520233154, + "learning_rate": 0.00019849912509474049, + "loss": 0.2076, + "step": 2365 + }, + { + "epoch": 0.055462765988619114, + "grad_norm": 0.3097965121269226, + "learning_rate": 0.00019849785339718414, + "loss": 0.1174, + "step": 2366 + }, + { + "epoch": 0.05548620756342411, + "grad_norm": 0.5722771883010864, + "learning_rate": 0.00019849658116517568, + "loss": 0.2739, + "step": 2367 + }, + { + "epoch": 0.0555096491382291, + "grad_norm": 0.3667926788330078, + "learning_rate": 0.000198495308398722, + "loss": 0.0951, + "step": 2368 + }, + { + "epoch": 0.055533090713034104, + "grad_norm": 0.5967243909835815, + "learning_rate": 0.00019849403509783, + "loss": 0.1067, + "step": 2369 + }, + { + "epoch": 0.0555565322878391, + "grad_norm": 0.35737675428390503, + "learning_rate": 0.00019849276126250663, + "loss": 0.1263, + "step": 2370 + }, + { + "epoch": 0.05557997386264409, + "grad_norm": 0.4242089092731476, + "learning_rate": 0.00019849148689275877, + "loss": 0.0955, + "step": 2371 + }, + { + "epoch": 0.05560341543744909, + "grad_norm": 0.7584924101829529, + "learning_rate": 0.00019849021198859327, + "loss": 0.2427, + "step": 2372 + }, + { + "epoch": 0.05562685701225408, + "grad_norm": 0.8136000633239746, + "learning_rate": 0.00019848893655001715, + "loss": 1.0788, + "step": 2373 + }, + { + "epoch": 0.055650298587059076, + "grad_norm": 0.5435866713523865, + "learning_rate": 0.00019848766057703728, + "loss": 0.8115, + "step": 2374 + }, + { + "epoch": 0.05567374016186408, + "grad_norm": 0.2494216412305832, + "learning_rate": 0.00019848638406966062, + "loss": 0.0848, + "step": 2375 + }, + { + "epoch": 0.05569718173666907, + "grad_norm": 0.6073143482208252, + "learning_rate": 0.00019848510702789404, + "loss": 0.1488, + "step": 2376 + }, + { + "epoch": 0.055720623311474066, + "grad_norm": 0.39744946360588074, + "learning_rate": 0.00019848382945174454, + "loss": 0.5014, + "step": 2377 + }, + { + "epoch": 0.05574406488627906, + "grad_norm": 0.36557677388191223, + "learning_rate": 0.00019848255134121895, + "loss": 0.4348, + "step": 2378 + }, + { + "epoch": 0.055767506461084054, + "grad_norm": 0.5983520150184631, + "learning_rate": 0.00019848127269632432, + "loss": 0.1878, + "step": 2379 + }, + { + "epoch": 0.05579094803588905, + "grad_norm": 0.5777316093444824, + "learning_rate": 0.00019847999351706755, + "loss": 0.1564, + "step": 2380 + }, + { + "epoch": 0.05581438961069405, + "grad_norm": 0.34505802392959595, + "learning_rate": 0.0001984787138034555, + "loss": 0.0887, + "step": 2381 + }, + { + "epoch": 0.055837831185499044, + "grad_norm": 0.5256333947181702, + "learning_rate": 0.0001984774335554952, + "loss": 0.1195, + "step": 2382 + }, + { + "epoch": 0.05586127276030404, + "grad_norm": 0.14470724761486053, + "learning_rate": 0.00019847615277319362, + "loss": 0.0222, + "step": 2383 + }, + { + "epoch": 0.05588471433510903, + "grad_norm": 0.7868782877922058, + "learning_rate": 0.00019847487145655763, + "loss": 0.1987, + "step": 2384 + }, + { + "epoch": 0.05590815590991403, + "grad_norm": 0.6733342409133911, + "learning_rate": 0.00019847358960559423, + "loss": 0.3215, + "step": 2385 + }, + { + "epoch": 0.05593159748471902, + "grad_norm": 0.6138713955879211, + "learning_rate": 0.00019847230722031036, + "loss": 0.1196, + "step": 2386 + }, + { + "epoch": 0.055955039059524016, + "grad_norm": 0.4183715581893921, + "learning_rate": 0.00019847102430071298, + "loss": 0.1935, + "step": 2387 + }, + { + "epoch": 0.05597848063432902, + "grad_norm": 0.5684915781021118, + "learning_rate": 0.00019846974084680908, + "loss": 0.3132, + "step": 2388 + }, + { + "epoch": 0.05600192220913401, + "grad_norm": 0.4549114406108856, + "learning_rate": 0.00019846845685860556, + "loss": 0.5788, + "step": 2389 + }, + { + "epoch": 0.056025363783939006, + "grad_norm": 0.6836646795272827, + "learning_rate": 0.00019846717233610944, + "loss": 0.2335, + "step": 2390 + }, + { + "epoch": 0.056048805358744, + "grad_norm": 0.5451881289482117, + "learning_rate": 0.00019846588727932764, + "loss": 0.2338, + "step": 2391 + }, + { + "epoch": 0.056072246933548994, + "grad_norm": 0.611516535282135, + "learning_rate": 0.00019846460168826722, + "loss": 0.1903, + "step": 2392 + }, + { + "epoch": 0.05609568850835399, + "grad_norm": 0.46074265241622925, + "learning_rate": 0.0001984633155629351, + "loss": 0.1467, + "step": 2393 + }, + { + "epoch": 0.05611913008315899, + "grad_norm": 0.6711487174034119, + "learning_rate": 0.00019846202890333822, + "loss": 0.2059, + "step": 2394 + }, + { + "epoch": 0.056142571657963984, + "grad_norm": 0.32262900471687317, + "learning_rate": 0.00019846074170948362, + "loss": 0.1138, + "step": 2395 + }, + { + "epoch": 0.05616601323276898, + "grad_norm": 0.3450776934623718, + "learning_rate": 0.0001984594539813783, + "loss": 0.1395, + "step": 2396 + }, + { + "epoch": 0.05618945480757397, + "grad_norm": 0.6597931385040283, + "learning_rate": 0.00019845816571902915, + "loss": 0.1629, + "step": 2397 + }, + { + "epoch": 0.05621289638237897, + "grad_norm": 0.2994259297847748, + "learning_rate": 0.00019845687692244328, + "loss": 0.0895, + "step": 2398 + }, + { + "epoch": 0.05623633795718396, + "grad_norm": 0.1553717404603958, + "learning_rate": 0.00019845558759162759, + "loss": 0.0329, + "step": 2399 + }, + { + "epoch": 0.056259779531988956, + "grad_norm": 0.5419518947601318, + "learning_rate": 0.00019845429772658913, + "loss": 0.1737, + "step": 2400 + }, + { + "epoch": 0.05628322110679396, + "grad_norm": 0.46419334411621094, + "learning_rate": 0.0001984530073273349, + "loss": 0.1295, + "step": 2401 + }, + { + "epoch": 0.05630666268159895, + "grad_norm": 0.4945458471775055, + "learning_rate": 0.00019845171639387187, + "loss": 0.1714, + "step": 2402 + }, + { + "epoch": 0.056330104256403946, + "grad_norm": 0.42228278517723083, + "learning_rate": 0.00019845042492620704, + "loss": 0.1534, + "step": 2403 + }, + { + "epoch": 0.05635354583120894, + "grad_norm": 0.5672913193702698, + "learning_rate": 0.00019844913292434746, + "loss": 0.1907, + "step": 2404 + }, + { + "epoch": 0.056376987406013934, + "grad_norm": 0.2033247947692871, + "learning_rate": 0.0001984478403883001, + "loss": 0.0704, + "step": 2405 + }, + { + "epoch": 0.05640042898081893, + "grad_norm": 0.41466259956359863, + "learning_rate": 0.000198446547318072, + "loss": 0.2483, + "step": 2406 + }, + { + "epoch": 0.05642387055562393, + "grad_norm": 0.32958686351776123, + "learning_rate": 0.0001984452537136702, + "loss": 0.0841, + "step": 2407 + }, + { + "epoch": 0.056447312130428924, + "grad_norm": 0.4820334315299988, + "learning_rate": 0.0001984439595751016, + "loss": 0.1461, + "step": 2408 + }, + { + "epoch": 0.05647075370523392, + "grad_norm": 0.45516112446784973, + "learning_rate": 0.0001984426649023734, + "loss": 0.1487, + "step": 2409 + }, + { + "epoch": 0.05649419528003891, + "grad_norm": 0.3823481798171997, + "learning_rate": 0.0001984413696954925, + "loss": 0.1483, + "step": 2410 + }, + { + "epoch": 0.05651763685484391, + "grad_norm": 0.43259161710739136, + "learning_rate": 0.00019844007395446593, + "loss": 0.1561, + "step": 2411 + }, + { + "epoch": 0.0565410784296489, + "grad_norm": 0.19127610325813293, + "learning_rate": 0.00019843877767930076, + "loss": 0.0706, + "step": 2412 + }, + { + "epoch": 0.0565645200044539, + "grad_norm": 0.14368073642253876, + "learning_rate": 0.00019843748087000404, + "loss": 0.0429, + "step": 2413 + }, + { + "epoch": 0.0565879615792589, + "grad_norm": 0.5794817805290222, + "learning_rate": 0.00019843618352658276, + "loss": 0.6139, + "step": 2414 + }, + { + "epoch": 0.05661140315406389, + "grad_norm": 0.4142094552516937, + "learning_rate": 0.000198434885649044, + "loss": 0.1041, + "step": 2415 + }, + { + "epoch": 0.056634844728868886, + "grad_norm": 0.5067589282989502, + "learning_rate": 0.00019843358723739475, + "loss": 0.1478, + "step": 2416 + }, + { + "epoch": 0.05665828630367388, + "grad_norm": 0.44935914874076843, + "learning_rate": 0.00019843228829164212, + "loss": 0.0884, + "step": 2417 + }, + { + "epoch": 0.056681727878478874, + "grad_norm": 0.3567822575569153, + "learning_rate": 0.00019843098881179312, + "loss": 0.0774, + "step": 2418 + }, + { + "epoch": 0.05670516945328387, + "grad_norm": 0.27320459485054016, + "learning_rate": 0.0001984296887978548, + "loss": 0.0503, + "step": 2419 + }, + { + "epoch": 0.05672861102808887, + "grad_norm": 0.9425749182701111, + "learning_rate": 0.0001984283882498342, + "loss": 0.8292, + "step": 2420 + }, + { + "epoch": 0.056752052602893864, + "grad_norm": 0.9763092994689941, + "learning_rate": 0.00019842708716773843, + "loss": 0.2644, + "step": 2421 + }, + { + "epoch": 0.05677549417769886, + "grad_norm": 0.7208408713340759, + "learning_rate": 0.0001984257855515745, + "loss": 0.2298, + "step": 2422 + }, + { + "epoch": 0.05679893575250385, + "grad_norm": 0.14823009073734283, + "learning_rate": 0.00019842448340134952, + "loss": 0.0726, + "step": 2423 + }, + { + "epoch": 0.05682237732730885, + "grad_norm": 0.2723791003227234, + "learning_rate": 0.0001984231807170705, + "loss": 0.378, + "step": 2424 + }, + { + "epoch": 0.05684581890211384, + "grad_norm": 0.49811533093452454, + "learning_rate": 0.00019842187749874451, + "loss": 0.1565, + "step": 2425 + }, + { + "epoch": 0.05686926047691884, + "grad_norm": 0.8069016933441162, + "learning_rate": 0.0001984205737463787, + "loss": 0.1963, + "step": 2426 + }, + { + "epoch": 0.05689270205172384, + "grad_norm": 0.7035778760910034, + "learning_rate": 0.00019841926945998007, + "loss": 0.2208, + "step": 2427 + }, + { + "epoch": 0.05691614362652883, + "grad_norm": 0.7637605667114258, + "learning_rate": 0.00019841796463955567, + "loss": 0.2314, + "step": 2428 + }, + { + "epoch": 0.056939585201333826, + "grad_norm": 0.6151102185249329, + "learning_rate": 0.0001984166592851127, + "loss": 0.0676, + "step": 2429 + }, + { + "epoch": 0.05696302677613882, + "grad_norm": 0.28432777523994446, + "learning_rate": 0.00019841535339665815, + "loss": 0.0729, + "step": 2430 + }, + { + "epoch": 0.056986468350943814, + "grad_norm": 0.35146477818489075, + "learning_rate": 0.00019841404697419908, + "loss": 0.0846, + "step": 2431 + }, + { + "epoch": 0.05700990992574881, + "grad_norm": 0.4170227348804474, + "learning_rate": 0.00019841274001774266, + "loss": 0.1097, + "step": 2432 + }, + { + "epoch": 0.05703335150055381, + "grad_norm": 0.2175893634557724, + "learning_rate": 0.00019841143252729592, + "loss": 0.0629, + "step": 2433 + }, + { + "epoch": 0.057056793075358804, + "grad_norm": 0.5943698883056641, + "learning_rate": 0.00019841012450286602, + "loss": 0.1829, + "step": 2434 + }, + { + "epoch": 0.0570802346501638, + "grad_norm": 0.2298794537782669, + "learning_rate": 0.00019840881594446, + "loss": 0.065, + "step": 2435 + }, + { + "epoch": 0.05710367622496879, + "grad_norm": 0.12216923385858536, + "learning_rate": 0.000198407506852085, + "loss": 0.045, + "step": 2436 + }, + { + "epoch": 0.05712711779977379, + "grad_norm": 0.20255014300346375, + "learning_rate": 0.00019840619722574806, + "loss": 0.0739, + "step": 2437 + }, + { + "epoch": 0.05715055937457878, + "grad_norm": 0.4931810200214386, + "learning_rate": 0.00019840488706545633, + "loss": 0.2131, + "step": 2438 + }, + { + "epoch": 0.05717400094938378, + "grad_norm": 0.475273072719574, + "learning_rate": 0.0001984035763712169, + "loss": 0.5852, + "step": 2439 + }, + { + "epoch": 0.05719744252418878, + "grad_norm": 0.7135085463523865, + "learning_rate": 0.00019840226514303695, + "loss": 0.268, + "step": 2440 + }, + { + "epoch": 0.05722088409899377, + "grad_norm": 0.7812580466270447, + "learning_rate": 0.0001984009533809235, + "loss": 0.2157, + "step": 2441 + }, + { + "epoch": 0.057244325673798765, + "grad_norm": 0.5030418038368225, + "learning_rate": 0.00019839964108488376, + "loss": 0.11, + "step": 2442 + }, + { + "epoch": 0.05726776724860376, + "grad_norm": 0.48561012744903564, + "learning_rate": 0.00019839832825492475, + "loss": 0.1868, + "step": 2443 + }, + { + "epoch": 0.057291208823408754, + "grad_norm": 0.38335779309272766, + "learning_rate": 0.00019839701489105367, + "loss": 0.1078, + "step": 2444 + }, + { + "epoch": 0.057314650398213755, + "grad_norm": 0.2193153202533722, + "learning_rate": 0.0001983957009932776, + "loss": 0.0557, + "step": 2445 + }, + { + "epoch": 0.05733809197301875, + "grad_norm": 0.659632682800293, + "learning_rate": 0.00019839438656160373, + "loss": 0.1259, + "step": 2446 + }, + { + "epoch": 0.057361533547823744, + "grad_norm": 0.11739759147167206, + "learning_rate": 0.0001983930715960391, + "loss": 0.0306, + "step": 2447 + }, + { + "epoch": 0.05738497512262874, + "grad_norm": 0.5848631262779236, + "learning_rate": 0.00019839175609659093, + "loss": 0.1386, + "step": 2448 + }, + { + "epoch": 0.05740841669743373, + "grad_norm": 0.7234845161437988, + "learning_rate": 0.00019839044006326632, + "loss": 0.2378, + "step": 2449 + }, + { + "epoch": 0.05743185827223873, + "grad_norm": 0.9248813390731812, + "learning_rate": 0.0001983891234960724, + "loss": 0.2213, + "step": 2450 + }, + { + "epoch": 0.05745529984704372, + "grad_norm": 0.45674794912338257, + "learning_rate": 0.00019838780639501636, + "loss": 0.1401, + "step": 2451 + }, + { + "epoch": 0.05747874142184872, + "grad_norm": 0.4697823226451874, + "learning_rate": 0.0001983864887601053, + "loss": 0.1703, + "step": 2452 + }, + { + "epoch": 0.05750218299665372, + "grad_norm": 0.30139291286468506, + "learning_rate": 0.0001983851705913464, + "loss": 0.1126, + "step": 2453 + }, + { + "epoch": 0.05752562457145871, + "grad_norm": 0.5537699460983276, + "learning_rate": 0.00019838385188874677, + "loss": 0.2072, + "step": 2454 + }, + { + "epoch": 0.057549066146263705, + "grad_norm": 0.5483168959617615, + "learning_rate": 0.00019838253265231363, + "loss": 0.2337, + "step": 2455 + }, + { + "epoch": 0.0575725077210687, + "grad_norm": 0.4835502505302429, + "learning_rate": 0.00019838121288205405, + "loss": 0.1369, + "step": 2456 + }, + { + "epoch": 0.057595949295873694, + "grad_norm": 0.5936371088027954, + "learning_rate": 0.00019837989257797526, + "loss": 0.1453, + "step": 2457 + }, + { + "epoch": 0.057619390870678695, + "grad_norm": 0.5478885769844055, + "learning_rate": 0.00019837857174008443, + "loss": 0.1792, + "step": 2458 + }, + { + "epoch": 0.05764283244548369, + "grad_norm": 0.6472013592720032, + "learning_rate": 0.0001983772503683887, + "loss": 0.1156, + "step": 2459 + }, + { + "epoch": 0.057666274020288684, + "grad_norm": 0.2826383411884308, + "learning_rate": 0.0001983759284628952, + "loss": 0.0806, + "step": 2460 + }, + { + "epoch": 0.05768971559509368, + "grad_norm": 0.1980988085269928, + "learning_rate": 0.0001983746060236112, + "loss": 0.0494, + "step": 2461 + }, + { + "epoch": 0.05771315716989867, + "grad_norm": 0.1685679405927658, + "learning_rate": 0.0001983732830505438, + "loss": 0.0314, + "step": 2462 + }, + { + "epoch": 0.05773659874470367, + "grad_norm": 0.6603812575340271, + "learning_rate": 0.0001983719595437002, + "loss": 0.2135, + "step": 2463 + }, + { + "epoch": 0.05776004031950866, + "grad_norm": 0.6933236718177795, + "learning_rate": 0.00019837063550308759, + "loss": 0.17, + "step": 2464 + }, + { + "epoch": 0.05778348189431366, + "grad_norm": 0.37596988677978516, + "learning_rate": 0.00019836931092871314, + "loss": 0.0818, + "step": 2465 + }, + { + "epoch": 0.05780692346911866, + "grad_norm": 0.8507265448570251, + "learning_rate": 0.000198367985820584, + "loss": 0.2484, + "step": 2466 + }, + { + "epoch": 0.05783036504392365, + "grad_norm": 0.26318010687828064, + "learning_rate": 0.0001983666601787075, + "loss": 0.0729, + "step": 2467 + }, + { + "epoch": 0.057853806618728645, + "grad_norm": 0.5309605002403259, + "learning_rate": 0.00019836533400309066, + "loss": 0.1559, + "step": 2468 + }, + { + "epoch": 0.05787724819353364, + "grad_norm": 0.3243257999420166, + "learning_rate": 0.00019836400729374075, + "loss": 0.095, + "step": 2469 + }, + { + "epoch": 0.057900689768338634, + "grad_norm": 0.6953548789024353, + "learning_rate": 0.000198362680050665, + "loss": 0.2766, + "step": 2470 + }, + { + "epoch": 0.057924131343143635, + "grad_norm": 0.5745687484741211, + "learning_rate": 0.0001983613522738706, + "loss": 0.2124, + "step": 2471 + }, + { + "epoch": 0.05794757291794863, + "grad_norm": 0.5631153583526611, + "learning_rate": 0.0001983600239633647, + "loss": 0.1606, + "step": 2472 + }, + { + "epoch": 0.057971014492753624, + "grad_norm": 0.6472281813621521, + "learning_rate": 0.00019835869511915455, + "loss": 0.1448, + "step": 2473 + }, + { + "epoch": 0.05799445606755862, + "grad_norm": 0.7035263180732727, + "learning_rate": 0.00019835736574124737, + "loss": 0.1986, + "step": 2474 + }, + { + "epoch": 0.05801789764236361, + "grad_norm": 0.8421187996864319, + "learning_rate": 0.00019835603582965035, + "loss": 0.2634, + "step": 2475 + }, + { + "epoch": 0.05804133921716861, + "grad_norm": 0.8194336891174316, + "learning_rate": 0.0001983547053843707, + "loss": 0.247, + "step": 2476 + }, + { + "epoch": 0.05806478079197361, + "grad_norm": 0.3971225917339325, + "learning_rate": 0.00019835337440541568, + "loss": 0.1105, + "step": 2477 + }, + { + "epoch": 0.0580882223667786, + "grad_norm": 0.4019365906715393, + "learning_rate": 0.00019835204289279248, + "loss": 0.1711, + "step": 2478 + }, + { + "epoch": 0.0581116639415836, + "grad_norm": 0.6231613159179688, + "learning_rate": 0.0001983507108465083, + "loss": 0.2248, + "step": 2479 + }, + { + "epoch": 0.05813510551638859, + "grad_norm": 0.46208271384239197, + "learning_rate": 0.00019834937826657045, + "loss": 0.1001, + "step": 2480 + }, + { + "epoch": 0.058158547091193585, + "grad_norm": 0.16937454044818878, + "learning_rate": 0.00019834804515298605, + "loss": 0.051, + "step": 2481 + }, + { + "epoch": 0.05818198866599858, + "grad_norm": 0.5051255822181702, + "learning_rate": 0.00019834671150576242, + "loss": 0.1359, + "step": 2482 + }, + { + "epoch": 0.058205430240803574, + "grad_norm": 0.32102248072624207, + "learning_rate": 0.00019834537732490678, + "loss": 0.0686, + "step": 2483 + }, + { + "epoch": 0.058228871815608575, + "grad_norm": 0.6496878862380981, + "learning_rate": 0.00019834404261042635, + "loss": 0.2109, + "step": 2484 + }, + { + "epoch": 0.05825231339041357, + "grad_norm": 0.9084927439689636, + "learning_rate": 0.00019834270736232839, + "loss": 0.2393, + "step": 2485 + }, + { + "epoch": 0.058275754965218564, + "grad_norm": 0.5371273756027222, + "learning_rate": 0.0001983413715806201, + "loss": 0.1531, + "step": 2486 + }, + { + "epoch": 0.05829919654002356, + "grad_norm": 0.5560205578804016, + "learning_rate": 0.00019834003526530877, + "loss": 0.1508, + "step": 2487 + }, + { + "epoch": 0.05832263811482855, + "grad_norm": 0.8812174201011658, + "learning_rate": 0.00019833869841640167, + "loss": 0.2357, + "step": 2488 + }, + { + "epoch": 0.05834607968963355, + "grad_norm": 0.13447394967079163, + "learning_rate": 0.000198337361033906, + "loss": 0.0289, + "step": 2489 + }, + { + "epoch": 0.05836952126443855, + "grad_norm": 0.5412095189094543, + "learning_rate": 0.00019833602311782907, + "loss": 0.1626, + "step": 2490 + }, + { + "epoch": 0.05839296283924354, + "grad_norm": 0.4034751355648041, + "learning_rate": 0.00019833468466817807, + "loss": 0.1338, + "step": 2491 + }, + { + "epoch": 0.05841640441404854, + "grad_norm": 0.4459722936153412, + "learning_rate": 0.00019833334568496037, + "loss": 0.0957, + "step": 2492 + }, + { + "epoch": 0.05843984598885353, + "grad_norm": 0.6097776889801025, + "learning_rate": 0.00019833200616818314, + "loss": 0.227, + "step": 2493 + }, + { + "epoch": 0.058463287563658525, + "grad_norm": 0.8450313806533813, + "learning_rate": 0.00019833066611785367, + "loss": 0.2738, + "step": 2494 + }, + { + "epoch": 0.05848672913846352, + "grad_norm": 0.6441365480422974, + "learning_rate": 0.00019832932553397924, + "loss": 0.2205, + "step": 2495 + }, + { + "epoch": 0.058510170713268514, + "grad_norm": 0.6242318153381348, + "learning_rate": 0.00019832798441656713, + "loss": 0.1679, + "step": 2496 + }, + { + "epoch": 0.058533612288073515, + "grad_norm": 0.8799024820327759, + "learning_rate": 0.00019832664276562463, + "loss": 0.4137, + "step": 2497 + }, + { + "epoch": 0.05855705386287851, + "grad_norm": 0.3524859845638275, + "learning_rate": 0.00019832530058115899, + "loss": 0.0887, + "step": 2498 + }, + { + "epoch": 0.058580495437683504, + "grad_norm": 0.1334221363067627, + "learning_rate": 0.00019832395786317748, + "loss": 0.0394, + "step": 2499 + }, + { + "epoch": 0.0586039370124885, + "grad_norm": 0.6738414168357849, + "learning_rate": 0.00019832261461168743, + "loss": 0.1749, + "step": 2500 + }, + { + "epoch": 0.05862737858729349, + "grad_norm": 0.9191380739212036, + "learning_rate": 0.0001983212708266961, + "loss": 0.2895, + "step": 2501 + }, + { + "epoch": 0.05865082016209849, + "grad_norm": 0.44715645909309387, + "learning_rate": 0.00019831992650821083, + "loss": 0.1276, + "step": 2502 + }, + { + "epoch": 0.05867426173690349, + "grad_norm": 0.4354608356952667, + "learning_rate": 0.00019831858165623883, + "loss": 0.1718, + "step": 2503 + }, + { + "epoch": 0.05869770331170848, + "grad_norm": 0.35200420022010803, + "learning_rate": 0.00019831723627078746, + "loss": 0.1623, + "step": 2504 + }, + { + "epoch": 0.05872114488651348, + "grad_norm": 0.4481732249259949, + "learning_rate": 0.000198315890351864, + "loss": 0.1384, + "step": 2505 + }, + { + "epoch": 0.05874458646131847, + "grad_norm": 0.5656375288963318, + "learning_rate": 0.00019831454389947575, + "loss": 0.2504, + "step": 2506 + }, + { + "epoch": 0.058768028036123465, + "grad_norm": 0.5405871272087097, + "learning_rate": 0.00019831319691363002, + "loss": 0.1655, + "step": 2507 + }, + { + "epoch": 0.05879146961092846, + "grad_norm": 0.8019562363624573, + "learning_rate": 0.00019831184939433412, + "loss": 0.3049, + "step": 2508 + }, + { + "epoch": 0.05881491118573346, + "grad_norm": 0.5327983498573303, + "learning_rate": 0.00019831050134159534, + "loss": 0.1003, + "step": 2509 + }, + { + "epoch": 0.058838352760538455, + "grad_norm": 0.5261302590370178, + "learning_rate": 0.00019830915275542105, + "loss": 0.1502, + "step": 2510 + }, + { + "epoch": 0.05886179433534345, + "grad_norm": 0.2701726257801056, + "learning_rate": 0.00019830780363581853, + "loss": 0.067, + "step": 2511 + }, + { + "epoch": 0.058885235910148444, + "grad_norm": 0.18894420564174652, + "learning_rate": 0.0001983064539827951, + "loss": 0.0317, + "step": 2512 + }, + { + "epoch": 0.05890867748495344, + "grad_norm": 0.32092970609664917, + "learning_rate": 0.00019830510379635808, + "loss": 0.0701, + "step": 2513 + }, + { + "epoch": 0.05893211905975843, + "grad_norm": 0.5563562512397766, + "learning_rate": 0.0001983037530765148, + "loss": 0.1769, + "step": 2514 + }, + { + "epoch": 0.05895556063456343, + "grad_norm": 0.7568148374557495, + "learning_rate": 0.0001983024018232726, + "loss": 0.7383, + "step": 2515 + }, + { + "epoch": 0.05897900220936843, + "grad_norm": 0.6047705411911011, + "learning_rate": 0.00019830105003663877, + "loss": 0.2512, + "step": 2516 + }, + { + "epoch": 0.05900244378417342, + "grad_norm": 0.7137753963470459, + "learning_rate": 0.0001982996977166207, + "loss": 0.2811, + "step": 2517 + }, + { + "epoch": 0.05902588535897842, + "grad_norm": 0.6393406987190247, + "learning_rate": 0.00019829834486322575, + "loss": 0.1556, + "step": 2518 + }, + { + "epoch": 0.05904932693378341, + "grad_norm": 0.4579344391822815, + "learning_rate": 0.00019829699147646117, + "loss": 0.1159, + "step": 2519 + }, + { + "epoch": 0.059072768508588405, + "grad_norm": 0.23998263478279114, + "learning_rate": 0.00019829563755633436, + "loss": 0.0351, + "step": 2520 + }, + { + "epoch": 0.0590962100833934, + "grad_norm": 0.39282867312431335, + "learning_rate": 0.00019829428310285264, + "loss": 0.1547, + "step": 2521 + }, + { + "epoch": 0.0591196516581984, + "grad_norm": 0.7492336630821228, + "learning_rate": 0.0001982929281160234, + "loss": 0.2139, + "step": 2522 + }, + { + "epoch": 0.059143093233003395, + "grad_norm": 1.0883280038833618, + "learning_rate": 0.00019829157259585397, + "loss": 0.2152, + "step": 2523 + }, + { + "epoch": 0.05916653480780839, + "grad_norm": 0.1319207400083542, + "learning_rate": 0.0001982902165423517, + "loss": 0.0304, + "step": 2524 + }, + { + "epoch": 0.059189976382613384, + "grad_norm": 0.61649090051651, + "learning_rate": 0.00019828885995552394, + "loss": 0.5978, + "step": 2525 + }, + { + "epoch": 0.05921341795741838, + "grad_norm": 0.5041993856430054, + "learning_rate": 0.00019828750283537808, + "loss": 0.1861, + "step": 2526 + }, + { + "epoch": 0.05923685953222337, + "grad_norm": 0.35397857427597046, + "learning_rate": 0.00019828614518192144, + "loss": 0.1472, + "step": 2527 + }, + { + "epoch": 0.059260301107028374, + "grad_norm": 0.5100482106208801, + "learning_rate": 0.00019828478699516144, + "loss": 0.1099, + "step": 2528 + }, + { + "epoch": 0.05928374268183337, + "grad_norm": 0.12211043387651443, + "learning_rate": 0.0001982834282751054, + "loss": 0.0477, + "step": 2529 + }, + { + "epoch": 0.05930718425663836, + "grad_norm": 0.4411783814430237, + "learning_rate": 0.00019828206902176073, + "loss": 0.134, + "step": 2530 + }, + { + "epoch": 0.05933062583144336, + "grad_norm": 0.2756865918636322, + "learning_rate": 0.00019828070923513477, + "loss": 0.0695, + "step": 2531 + }, + { + "epoch": 0.05935406740624835, + "grad_norm": 0.6280815601348877, + "learning_rate": 0.00019827934891523494, + "loss": 0.2387, + "step": 2532 + }, + { + "epoch": 0.059377508981053345, + "grad_norm": 0.6154405474662781, + "learning_rate": 0.0001982779880620686, + "loss": 0.1923, + "step": 2533 + }, + { + "epoch": 0.05940095055585834, + "grad_norm": 0.40427765250205994, + "learning_rate": 0.0001982766266756431, + "loss": 0.147, + "step": 2534 + }, + { + "epoch": 0.05942439213066334, + "grad_norm": 0.43436023592948914, + "learning_rate": 0.00019827526475596588, + "loss": 0.1872, + "step": 2535 + }, + { + "epoch": 0.059447833705468335, + "grad_norm": 0.6057941317558289, + "learning_rate": 0.00019827390230304435, + "loss": 0.1986, + "step": 2536 + }, + { + "epoch": 0.05947127528027333, + "grad_norm": 0.7966145873069763, + "learning_rate": 0.00019827253931688582, + "loss": 0.2556, + "step": 2537 + }, + { + "epoch": 0.059494716855078324, + "grad_norm": 0.5452943444252014, + "learning_rate": 0.00019827117579749773, + "loss": 0.1448, + "step": 2538 + }, + { + "epoch": 0.05951815842988332, + "grad_norm": 0.5584341287612915, + "learning_rate": 0.00019826981174488748, + "loss": 0.7523, + "step": 2539 + }, + { + "epoch": 0.05954160000468831, + "grad_norm": 0.2799418568611145, + "learning_rate": 0.00019826844715906247, + "loss": 0.06, + "step": 2540 + }, + { + "epoch": 0.059565041579493314, + "grad_norm": 0.6520624160766602, + "learning_rate": 0.00019826708204003011, + "loss": 0.1879, + "step": 2541 + }, + { + "epoch": 0.05958848315429831, + "grad_norm": 0.14394016563892365, + "learning_rate": 0.0001982657163877978, + "loss": 0.0409, + "step": 2542 + }, + { + "epoch": 0.0596119247291033, + "grad_norm": 0.20464393496513367, + "learning_rate": 0.00019826435020237293, + "loss": 0.0317, + "step": 2543 + }, + { + "epoch": 0.059635366303908297, + "grad_norm": 0.6972588300704956, + "learning_rate": 0.00019826298348376295, + "loss": 0.2293, + "step": 2544 + }, + { + "epoch": 0.05965880787871329, + "grad_norm": 0.738837480545044, + "learning_rate": 0.00019826161623197526, + "loss": 0.214, + "step": 2545 + }, + { + "epoch": 0.059682249453518285, + "grad_norm": 0.7829197645187378, + "learning_rate": 0.0001982602484470173, + "loss": 0.1608, + "step": 2546 + }, + { + "epoch": 0.05970569102832328, + "grad_norm": 0.7539836764335632, + "learning_rate": 0.00019825888012889642, + "loss": 0.188, + "step": 2547 + }, + { + "epoch": 0.05972913260312828, + "grad_norm": 0.6430259346961975, + "learning_rate": 0.00019825751127762012, + "loss": 0.2626, + "step": 2548 + }, + { + "epoch": 0.059752574177933275, + "grad_norm": 0.6180445551872253, + "learning_rate": 0.0001982561418931958, + "loss": 0.1423, + "step": 2549 + }, + { + "epoch": 0.05977601575273827, + "grad_norm": 0.21523836255073547, + "learning_rate": 0.00019825477197563088, + "loss": 0.0437, + "step": 2550 + }, + { + "epoch": 0.059799457327543264, + "grad_norm": 0.8996677398681641, + "learning_rate": 0.0001982534015249328, + "loss": 0.2844, + "step": 2551 + }, + { + "epoch": 0.05982289890234826, + "grad_norm": 0.5617738962173462, + "learning_rate": 0.00019825203054110903, + "loss": 0.204, + "step": 2552 + }, + { + "epoch": 0.05984634047715325, + "grad_norm": 0.826661229133606, + "learning_rate": 0.00019825065902416696, + "loss": 0.2045, + "step": 2553 + }, + { + "epoch": 0.059869782051958254, + "grad_norm": 0.8053309917449951, + "learning_rate": 0.00019824928697411407, + "loss": 0.3321, + "step": 2554 + }, + { + "epoch": 0.05989322362676325, + "grad_norm": 0.2919444143772125, + "learning_rate": 0.00019824791439095778, + "loss": 0.0905, + "step": 2555 + }, + { + "epoch": 0.05991666520156824, + "grad_norm": 0.38292351365089417, + "learning_rate": 0.00019824654127470552, + "loss": 0.1306, + "step": 2556 + }, + { + "epoch": 0.059940106776373236, + "grad_norm": 0.2207963615655899, + "learning_rate": 0.00019824516762536478, + "loss": 0.0771, + "step": 2557 + }, + { + "epoch": 0.05996354835117823, + "grad_norm": 0.6700311303138733, + "learning_rate": 0.00019824379344294299, + "loss": 0.1813, + "step": 2558 + }, + { + "epoch": 0.059986989925983225, + "grad_norm": 0.6495943665504456, + "learning_rate": 0.00019824241872744763, + "loss": 0.2246, + "step": 2559 + }, + { + "epoch": 0.060010431500788226, + "grad_norm": 0.4217361807823181, + "learning_rate": 0.00019824104347888612, + "loss": 0.1581, + "step": 2560 + }, + { + "epoch": 0.06003387307559322, + "grad_norm": 0.1607251763343811, + "learning_rate": 0.00019823966769726598, + "loss": 0.0595, + "step": 2561 + }, + { + "epoch": 0.060057314650398215, + "grad_norm": 0.750238835811615, + "learning_rate": 0.00019823829138259462, + "loss": 0.2054, + "step": 2562 + }, + { + "epoch": 0.06008075622520321, + "grad_norm": 0.1920137256383896, + "learning_rate": 0.00019823691453487954, + "loss": 0.0457, + "step": 2563 + }, + { + "epoch": 0.060104197800008204, + "grad_norm": 0.33823361992836, + "learning_rate": 0.0001982355371541282, + "loss": 0.1088, + "step": 2564 + }, + { + "epoch": 0.0601276393748132, + "grad_norm": 0.1980198174715042, + "learning_rate": 0.00019823415924034805, + "loss": 0.0839, + "step": 2565 + }, + { + "epoch": 0.06015108094961819, + "grad_norm": 0.40981751680374146, + "learning_rate": 0.00019823278079354658, + "loss": 0.1087, + "step": 2566 + }, + { + "epoch": 0.060174522524423193, + "grad_norm": 0.3936941921710968, + "learning_rate": 0.00019823140181373132, + "loss": 0.0831, + "step": 2567 + }, + { + "epoch": 0.06019796409922819, + "grad_norm": 0.4938303828239441, + "learning_rate": 0.0001982300223009097, + "loss": 0.1164, + "step": 2568 + }, + { + "epoch": 0.06022140567403318, + "grad_norm": 0.5949918031692505, + "learning_rate": 0.00019822864225508923, + "loss": 0.2593, + "step": 2569 + }, + { + "epoch": 0.060244847248838176, + "grad_norm": 0.6587717533111572, + "learning_rate": 0.00019822726167627734, + "loss": 0.1929, + "step": 2570 + }, + { + "epoch": 0.06026828882364317, + "grad_norm": 0.8452132940292358, + "learning_rate": 0.0001982258805644816, + "loss": 0.2311, + "step": 2571 + }, + { + "epoch": 0.060291730398448165, + "grad_norm": 0.7384014129638672, + "learning_rate": 0.00019822449891970947, + "loss": 0.1166, + "step": 2572 + }, + { + "epoch": 0.060315171973253166, + "grad_norm": 0.7323882579803467, + "learning_rate": 0.00019822311674196844, + "loss": 0.2361, + "step": 2573 + }, + { + "epoch": 0.06033861354805816, + "grad_norm": 0.15373171865940094, + "learning_rate": 0.00019822173403126603, + "loss": 0.042, + "step": 2574 + }, + { + "epoch": 0.060362055122863155, + "grad_norm": 0.9196078181266785, + "learning_rate": 0.0001982203507876097, + "loss": 0.1627, + "step": 2575 + }, + { + "epoch": 0.06038549669766815, + "grad_norm": 0.48333442211151123, + "learning_rate": 0.00019821896701100703, + "loss": 0.164, + "step": 2576 + }, + { + "epoch": 0.060408938272473144, + "grad_norm": 0.5219484567642212, + "learning_rate": 0.00019821758270146544, + "loss": 0.152, + "step": 2577 + }, + { + "epoch": 0.06043237984727814, + "grad_norm": 0.5905430912971497, + "learning_rate": 0.0001982161978589925, + "loss": 0.1492, + "step": 2578 + }, + { + "epoch": 0.06045582142208313, + "grad_norm": 0.6413229703903198, + "learning_rate": 0.0001982148124835957, + "loss": 0.1963, + "step": 2579 + }, + { + "epoch": 0.06047926299688813, + "grad_norm": 0.38900548219680786, + "learning_rate": 0.0001982134265752826, + "loss": 0.1191, + "step": 2580 + }, + { + "epoch": 0.06050270457169313, + "grad_norm": 0.48206818103790283, + "learning_rate": 0.00019821204013406062, + "loss": 0.1655, + "step": 2581 + }, + { + "epoch": 0.06052614614649812, + "grad_norm": 0.6120655536651611, + "learning_rate": 0.00019821065315993742, + "loss": 0.1106, + "step": 2582 + }, + { + "epoch": 0.060549587721303116, + "grad_norm": 0.4174582064151764, + "learning_rate": 0.00019820926565292042, + "loss": 0.5996, + "step": 2583 + }, + { + "epoch": 0.06057302929610811, + "grad_norm": 0.12165321409702301, + "learning_rate": 0.0001982078776130172, + "loss": 0.0383, + "step": 2584 + }, + { + "epoch": 0.060596470870913105, + "grad_norm": 0.1047314703464508, + "learning_rate": 0.00019820648904023523, + "loss": 0.0375, + "step": 2585 + }, + { + "epoch": 0.060619912445718106, + "grad_norm": 0.48196133971214294, + "learning_rate": 0.00019820509993458214, + "loss": 0.1609, + "step": 2586 + }, + { + "epoch": 0.0606433540205231, + "grad_norm": 0.1202399730682373, + "learning_rate": 0.00019820371029606539, + "loss": 0.0208, + "step": 2587 + }, + { + "epoch": 0.060666795595328095, + "grad_norm": 0.9508033394813538, + "learning_rate": 0.00019820232012469257, + "loss": 0.3285, + "step": 2588 + }, + { + "epoch": 0.06069023717013309, + "grad_norm": 0.5451314449310303, + "learning_rate": 0.00019820092942047116, + "loss": 0.21, + "step": 2589 + }, + { + "epoch": 0.060713678744938084, + "grad_norm": 0.5210623145103455, + "learning_rate": 0.00019819953818340879, + "loss": 0.1243, + "step": 2590 + }, + { + "epoch": 0.06073712031974308, + "grad_norm": 0.5952955484390259, + "learning_rate": 0.00019819814641351292, + "loss": 0.2538, + "step": 2591 + }, + { + "epoch": 0.06076056189454808, + "grad_norm": 0.5883393287658691, + "learning_rate": 0.00019819675411079118, + "loss": 0.2191, + "step": 2592 + }, + { + "epoch": 0.06078400346935307, + "grad_norm": 0.7623121738433838, + "learning_rate": 0.00019819536127525108, + "loss": 0.2598, + "step": 2593 + }, + { + "epoch": 0.06080744504415807, + "grad_norm": 0.4845884442329407, + "learning_rate": 0.0001981939679069002, + "loss": 0.1209, + "step": 2594 + }, + { + "epoch": 0.06083088661896306, + "grad_norm": 0.5188612937927246, + "learning_rate": 0.00019819257400574607, + "loss": 0.1321, + "step": 2595 + }, + { + "epoch": 0.060854328193768056, + "grad_norm": 0.73914635181427, + "learning_rate": 0.0001981911795717963, + "loss": 0.1486, + "step": 2596 + }, + { + "epoch": 0.06087776976857305, + "grad_norm": 0.7988701462745667, + "learning_rate": 0.0001981897846050584, + "loss": 0.1679, + "step": 2597 + }, + { + "epoch": 0.060901211343378045, + "grad_norm": 0.8035656213760376, + "learning_rate": 0.00019818838910554, + "loss": 0.6424, + "step": 2598 + }, + { + "epoch": 0.060924652918183046, + "grad_norm": 0.7214480638504028, + "learning_rate": 0.0001981869930732486, + "loss": 0.1567, + "step": 2599 + }, + { + "epoch": 0.06094809449298804, + "grad_norm": 0.625278115272522, + "learning_rate": 0.00019818559650819183, + "loss": 0.0699, + "step": 2600 + }, + { + "epoch": 0.060971536067793035, + "grad_norm": 0.8982163071632385, + "learning_rate": 0.00019818419941037727, + "loss": 0.211, + "step": 2601 + }, + { + "epoch": 0.06099497764259803, + "grad_norm": 1.0377644300460815, + "learning_rate": 0.00019818280177981245, + "loss": 0.2606, + "step": 2602 + }, + { + "epoch": 0.061018419217403023, + "grad_norm": 0.36188262701034546, + "learning_rate": 0.000198181403616505, + "loss": 0.1583, + "step": 2603 + }, + { + "epoch": 0.06104186079220802, + "grad_norm": 0.4987989664077759, + "learning_rate": 0.0001981800049204625, + "loss": 0.1471, + "step": 2604 + }, + { + "epoch": 0.06106530236701302, + "grad_norm": 0.49581360816955566, + "learning_rate": 0.00019817860569169255, + "loss": 0.1628, + "step": 2605 + }, + { + "epoch": 0.06108874394181801, + "grad_norm": 0.6287886500358582, + "learning_rate": 0.00019817720593020272, + "loss": 0.1416, + "step": 2606 + }, + { + "epoch": 0.06111218551662301, + "grad_norm": 0.7053158283233643, + "learning_rate": 0.00019817580563600057, + "loss": 0.2273, + "step": 2607 + }, + { + "epoch": 0.061135627091428, + "grad_norm": 0.7091062664985657, + "learning_rate": 0.0001981744048090938, + "loss": 0.157, + "step": 2608 + }, + { + "epoch": 0.061159068666232996, + "grad_norm": 0.6268508434295654, + "learning_rate": 0.0001981730034494899, + "loss": 0.2464, + "step": 2609 + }, + { + "epoch": 0.06118251024103799, + "grad_norm": 0.5842845439910889, + "learning_rate": 0.00019817160155719655, + "loss": 0.1199, + "step": 2610 + }, + { + "epoch": 0.061205951815842985, + "grad_norm": 0.22487173974514008, + "learning_rate": 0.0001981701991322213, + "loss": 0.072, + "step": 2611 + }, + { + "epoch": 0.061229393390647986, + "grad_norm": 0.9067367315292358, + "learning_rate": 0.00019816879617457183, + "loss": 0.2187, + "step": 2612 + }, + { + "epoch": 0.06125283496545298, + "grad_norm": 0.07693948596715927, + "learning_rate": 0.00019816739268425568, + "loss": 0.0165, + "step": 2613 + }, + { + "epoch": 0.061276276540257975, + "grad_norm": 0.7738196849822998, + "learning_rate": 0.0001981659886612805, + "loss": 0.349, + "step": 2614 + }, + { + "epoch": 0.06129971811506297, + "grad_norm": 0.617399275302887, + "learning_rate": 0.0001981645841056539, + "loss": 0.192, + "step": 2615 + }, + { + "epoch": 0.06132315968986796, + "grad_norm": 0.3379840552806854, + "learning_rate": 0.00019816317901738352, + "loss": 0.0565, + "step": 2616 + }, + { + "epoch": 0.06134660126467296, + "grad_norm": 0.5405941009521484, + "learning_rate": 0.000198161773396477, + "loss": 0.1876, + "step": 2617 + }, + { + "epoch": 0.06137004283947796, + "grad_norm": 0.3603545129299164, + "learning_rate": 0.00019816036724294188, + "loss": 0.074, + "step": 2618 + }, + { + "epoch": 0.06139348441428295, + "grad_norm": 0.6750705242156982, + "learning_rate": 0.00019815896055678587, + "loss": 0.2428, + "step": 2619 + }, + { + "epoch": 0.06141692598908795, + "grad_norm": 0.6277392506599426, + "learning_rate": 0.00019815755333801657, + "loss": 0.3122, + "step": 2620 + }, + { + "epoch": 0.06144036756389294, + "grad_norm": 0.6673246026039124, + "learning_rate": 0.00019815614558664165, + "loss": 0.2311, + "step": 2621 + }, + { + "epoch": 0.061463809138697936, + "grad_norm": 0.5996283292770386, + "learning_rate": 0.00019815473730266873, + "loss": 0.2446, + "step": 2622 + }, + { + "epoch": 0.06148725071350293, + "grad_norm": 0.4810495674610138, + "learning_rate": 0.0001981533284861054, + "loss": 0.1809, + "step": 2623 + }, + { + "epoch": 0.06151069228830793, + "grad_norm": 0.5437614321708679, + "learning_rate": 0.0001981519191369594, + "loss": 0.1387, + "step": 2624 + }, + { + "epoch": 0.061534133863112926, + "grad_norm": 0.40053874254226685, + "learning_rate": 0.0001981505092552383, + "loss": 0.1542, + "step": 2625 + }, + { + "epoch": 0.06155757543791792, + "grad_norm": 0.42541420459747314, + "learning_rate": 0.00019814909884094978, + "loss": 0.081, + "step": 2626 + }, + { + "epoch": 0.061581017012722915, + "grad_norm": 0.8702730536460876, + "learning_rate": 0.0001981476878941015, + "loss": 0.1746, + "step": 2627 + }, + { + "epoch": 0.06160445858752791, + "grad_norm": 0.6392454504966736, + "learning_rate": 0.0001981462764147011, + "loss": 0.2443, + "step": 2628 + }, + { + "epoch": 0.0616279001623329, + "grad_norm": 0.6542054414749146, + "learning_rate": 0.00019814486440275625, + "loss": 0.1817, + "step": 2629 + }, + { + "epoch": 0.0616513417371379, + "grad_norm": 0.6098841428756714, + "learning_rate": 0.00019814345185827458, + "loss": 0.2092, + "step": 2630 + }, + { + "epoch": 0.0616747833119429, + "grad_norm": 0.9168913960456848, + "learning_rate": 0.00019814203878126382, + "loss": 0.2708, + "step": 2631 + }, + { + "epoch": 0.06169822488674789, + "grad_norm": 0.3708637058734894, + "learning_rate": 0.00019814062517173157, + "loss": 0.1065, + "step": 2632 + }, + { + "epoch": 0.06172166646155289, + "grad_norm": 0.5121538639068604, + "learning_rate": 0.00019813921102968554, + "loss": 0.1468, + "step": 2633 + }, + { + "epoch": 0.06174510803635788, + "grad_norm": 0.34327301383018494, + "learning_rate": 0.00019813779635513337, + "loss": 0.1751, + "step": 2634 + }, + { + "epoch": 0.061768549611162876, + "grad_norm": 0.7261533141136169, + "learning_rate": 0.00019813638114808278, + "loss": 0.1991, + "step": 2635 + }, + { + "epoch": 0.06179199118596787, + "grad_norm": 0.5273656249046326, + "learning_rate": 0.00019813496540854142, + "loss": 0.3203, + "step": 2636 + }, + { + "epoch": 0.06181543276077287, + "grad_norm": 0.6393256187438965, + "learning_rate": 0.000198133549136517, + "loss": 0.1461, + "step": 2637 + }, + { + "epoch": 0.061838874335577866, + "grad_norm": 0.6272023916244507, + "learning_rate": 0.00019813213233201715, + "loss": 0.1763, + "step": 2638 + }, + { + "epoch": 0.06186231591038286, + "grad_norm": 0.5753744840621948, + "learning_rate": 0.00019813071499504958, + "loss": 0.2736, + "step": 2639 + }, + { + "epoch": 0.061885757485187855, + "grad_norm": 0.554560124874115, + "learning_rate": 0.00019812929712562203, + "loss": 0.1359, + "step": 2640 + }, + { + "epoch": 0.06190919905999285, + "grad_norm": 0.6524834036827087, + "learning_rate": 0.00019812787872374214, + "loss": 0.2137, + "step": 2641 + }, + { + "epoch": 0.06193264063479784, + "grad_norm": 0.8529788255691528, + "learning_rate": 0.0001981264597894176, + "loss": 0.3368, + "step": 2642 + }, + { + "epoch": 0.06195608220960284, + "grad_norm": 0.888479471206665, + "learning_rate": 0.00019812504032265616, + "loss": 0.3774, + "step": 2643 + }, + { + "epoch": 0.06197952378440784, + "grad_norm": 0.7107552289962769, + "learning_rate": 0.0001981236203234655, + "loss": 0.2399, + "step": 2644 + }, + { + "epoch": 0.06200296535921283, + "grad_norm": 0.19115066528320312, + "learning_rate": 0.0001981221997918533, + "loss": 0.0383, + "step": 2645 + }, + { + "epoch": 0.06202640693401783, + "grad_norm": 0.6347227096557617, + "learning_rate": 0.0001981207787278273, + "loss": 0.3357, + "step": 2646 + }, + { + "epoch": 0.06204984850882282, + "grad_norm": 0.7831333875656128, + "learning_rate": 0.00019811935713139517, + "loss": 0.1433, + "step": 2647 + }, + { + "epoch": 0.062073290083627816, + "grad_norm": 0.6326016187667847, + "learning_rate": 0.00019811793500256464, + "loss": 0.2342, + "step": 2648 + }, + { + "epoch": 0.06209673165843281, + "grad_norm": 0.4360705614089966, + "learning_rate": 0.00019811651234134345, + "loss": 0.1779, + "step": 2649 + }, + { + "epoch": 0.06212017323323781, + "grad_norm": 0.7302815914154053, + "learning_rate": 0.00019811508914773933, + "loss": 0.2803, + "step": 2650 + }, + { + "epoch": 0.062143614808042806, + "grad_norm": 0.5367488861083984, + "learning_rate": 0.00019811366542175996, + "loss": 0.1962, + "step": 2651 + }, + { + "epoch": 0.0621670563828478, + "grad_norm": 0.6751317977905273, + "learning_rate": 0.00019811224116341307, + "loss": 0.2708, + "step": 2652 + }, + { + "epoch": 0.062190497957652795, + "grad_norm": 0.6518232822418213, + "learning_rate": 0.00019811081637270643, + "loss": 0.2251, + "step": 2653 + }, + { + "epoch": 0.06221393953245779, + "grad_norm": 0.536491334438324, + "learning_rate": 0.00019810939104964768, + "loss": 0.091, + "step": 2654 + }, + { + "epoch": 0.06223738110726278, + "grad_norm": 0.6721421480178833, + "learning_rate": 0.0001981079651942447, + "loss": 0.1593, + "step": 2655 + }, + { + "epoch": 0.062260822682067785, + "grad_norm": 0.3902250826358795, + "learning_rate": 0.00019810653880650509, + "loss": 0.401, + "step": 2656 + }, + { + "epoch": 0.06228426425687278, + "grad_norm": 0.1910688877105713, + "learning_rate": 0.00019810511188643665, + "loss": 0.0396, + "step": 2657 + }, + { + "epoch": 0.06230770583167777, + "grad_norm": 0.3482339084148407, + "learning_rate": 0.00019810368443404713, + "loss": 0.1144, + "step": 2658 + }, + { + "epoch": 0.06233114740648277, + "grad_norm": 0.22409319877624512, + "learning_rate": 0.00019810225644934423, + "loss": 0.0776, + "step": 2659 + }, + { + "epoch": 0.06235458898128776, + "grad_norm": 0.2237289845943451, + "learning_rate": 0.00019810082793233574, + "loss": 0.0618, + "step": 2660 + }, + { + "epoch": 0.062378030556092756, + "grad_norm": 0.4980141818523407, + "learning_rate": 0.00019809939888302942, + "loss": 0.2111, + "step": 2661 + }, + { + "epoch": 0.06240147213089775, + "grad_norm": 0.7451149225234985, + "learning_rate": 0.00019809796930143298, + "loss": 0.1728, + "step": 2662 + }, + { + "epoch": 0.06242491370570275, + "grad_norm": 0.1178559735417366, + "learning_rate": 0.00019809653918755423, + "loss": 0.0284, + "step": 2663 + }, + { + "epoch": 0.062448355280507746, + "grad_norm": 0.6291905641555786, + "learning_rate": 0.00019809510854140085, + "loss": 0.8258, + "step": 2664 + }, + { + "epoch": 0.06247179685531274, + "grad_norm": 0.6358388662338257, + "learning_rate": 0.0001980936773629807, + "loss": 0.2901, + "step": 2665 + }, + { + "epoch": 0.062495238430117735, + "grad_norm": 0.6450452208518982, + "learning_rate": 0.00019809224565230148, + "loss": 0.1642, + "step": 2666 + }, + { + "epoch": 0.06251868000492274, + "grad_norm": 0.4998084604740143, + "learning_rate": 0.000198090813409371, + "loss": 0.1013, + "step": 2667 + }, + { + "epoch": 0.06254212157972773, + "grad_norm": 0.46892645955085754, + "learning_rate": 0.00019808938063419697, + "loss": 0.0638, + "step": 2668 + }, + { + "epoch": 0.06256556315453272, + "grad_norm": 0.13505814969539642, + "learning_rate": 0.0001980879473267872, + "loss": 0.0406, + "step": 2669 + }, + { + "epoch": 0.06258900472933772, + "grad_norm": 0.5627804398536682, + "learning_rate": 0.00019808651348714952, + "loss": 0.1573, + "step": 2670 + }, + { + "epoch": 0.06261244630414271, + "grad_norm": 0.22840337455272675, + "learning_rate": 0.0001980850791152916, + "loss": 0.2331, + "step": 2671 + }, + { + "epoch": 0.06263588787894771, + "grad_norm": 0.0918140709400177, + "learning_rate": 0.0001980836442112213, + "loss": 0.0289, + "step": 2672 + }, + { + "epoch": 0.0626593294537527, + "grad_norm": 0.22925221920013428, + "learning_rate": 0.00019808220877494642, + "loss": 0.0372, + "step": 2673 + }, + { + "epoch": 0.0626827710285577, + "grad_norm": 0.5407826900482178, + "learning_rate": 0.00019808077280647468, + "loss": 0.6722, + "step": 2674 + }, + { + "epoch": 0.06270621260336269, + "grad_norm": 0.6957663297653198, + "learning_rate": 0.00019807933630581394, + "loss": 0.7966, + "step": 2675 + }, + { + "epoch": 0.06272965417816768, + "grad_norm": 0.9432571530342102, + "learning_rate": 0.00019807789927297193, + "loss": 0.1863, + "step": 2676 + }, + { + "epoch": 0.06275309575297268, + "grad_norm": 0.510155200958252, + "learning_rate": 0.0001980764617079565, + "loss": 0.0687, + "step": 2677 + }, + { + "epoch": 0.06277653732777767, + "grad_norm": 0.5645017027854919, + "learning_rate": 0.00019807502361077545, + "loss": 0.1147, + "step": 2678 + }, + { + "epoch": 0.06279997890258268, + "grad_norm": 0.6116880774497986, + "learning_rate": 0.00019807358498143653, + "loss": 0.2675, + "step": 2679 + }, + { + "epoch": 0.06282342047738768, + "grad_norm": 0.7444850206375122, + "learning_rate": 0.00019807214581994763, + "loss": 0.255, + "step": 2680 + }, + { + "epoch": 0.06284686205219267, + "grad_norm": 1.180886149406433, + "learning_rate": 0.00019807070612631647, + "loss": 0.2559, + "step": 2681 + }, + { + "epoch": 0.06287030362699766, + "grad_norm": 0.8681117296218872, + "learning_rate": 0.0001980692659005509, + "loss": 0.2938, + "step": 2682 + }, + { + "epoch": 0.06289374520180266, + "grad_norm": 0.4048895537853241, + "learning_rate": 0.00019806782514265875, + "loss": 0.4542, + "step": 2683 + }, + { + "epoch": 0.06291718677660765, + "grad_norm": 0.24937906861305237, + "learning_rate": 0.0001980663838526478, + "loss": 0.0549, + "step": 2684 + }, + { + "epoch": 0.06294062835141265, + "grad_norm": 0.6761670112609863, + "learning_rate": 0.0001980649420305259, + "loss": 0.3168, + "step": 2685 + }, + { + "epoch": 0.06296406992621764, + "grad_norm": 0.4401816129684448, + "learning_rate": 0.00019806349967630088, + "loss": 0.1497, + "step": 2686 + }, + { + "epoch": 0.06298751150102264, + "grad_norm": 0.6104252934455872, + "learning_rate": 0.00019806205678998056, + "loss": 0.2166, + "step": 2687 + }, + { + "epoch": 0.06301095307582763, + "grad_norm": 0.2531585395336151, + "learning_rate": 0.00019806061337157273, + "loss": 0.0824, + "step": 2688 + }, + { + "epoch": 0.06303439465063262, + "grad_norm": 0.4636375606060028, + "learning_rate": 0.00019805916942108528, + "loss": 0.1091, + "step": 2689 + }, + { + "epoch": 0.06305783622543762, + "grad_norm": 0.2758193910121918, + "learning_rate": 0.00019805772493852603, + "loss": 0.0609, + "step": 2690 + }, + { + "epoch": 0.06308127780024263, + "grad_norm": 0.896906852722168, + "learning_rate": 0.0001980562799239028, + "loss": 0.1781, + "step": 2691 + }, + { + "epoch": 0.06310471937504762, + "grad_norm": 0.6297450661659241, + "learning_rate": 0.0001980548343772234, + "loss": 0.2286, + "step": 2692 + }, + { + "epoch": 0.06312816094985262, + "grad_norm": 0.29831477999687195, + "learning_rate": 0.00019805338829849577, + "loss": 0.0494, + "step": 2693 + }, + { + "epoch": 0.06315160252465761, + "grad_norm": 0.7110785245895386, + "learning_rate": 0.00019805194168772764, + "loss": 0.2367, + "step": 2694 + }, + { + "epoch": 0.0631750440994626, + "grad_norm": 0.813372790813446, + "learning_rate": 0.00019805049454492697, + "loss": 0.2417, + "step": 2695 + }, + { + "epoch": 0.0631984856742676, + "grad_norm": 0.3253214955329895, + "learning_rate": 0.00019804904687010155, + "loss": 0.0927, + "step": 2696 + }, + { + "epoch": 0.0632219272490726, + "grad_norm": 0.5511049032211304, + "learning_rate": 0.00019804759866325923, + "loss": 0.2053, + "step": 2697 + }, + { + "epoch": 0.06324536882387759, + "grad_norm": 0.5421412587165833, + "learning_rate": 0.00019804614992440786, + "loss": 0.1877, + "step": 2698 + }, + { + "epoch": 0.06326881039868258, + "grad_norm": 0.5515120625495911, + "learning_rate": 0.0001980447006535554, + "loss": 0.0817, + "step": 2699 + }, + { + "epoch": 0.06329225197348758, + "grad_norm": 0.7370232343673706, + "learning_rate": 0.00019804325085070955, + "loss": 0.0897, + "step": 2700 + }, + { + "epoch": 0.06331569354829257, + "grad_norm": 0.24264952540397644, + "learning_rate": 0.0001980418005158783, + "loss": 0.0917, + "step": 2701 + }, + { + "epoch": 0.06333913512309756, + "grad_norm": 0.5973063707351685, + "learning_rate": 0.00019804034964906948, + "loss": 0.2015, + "step": 2702 + }, + { + "epoch": 0.06336257669790256, + "grad_norm": 0.6143434643745422, + "learning_rate": 0.00019803889825029098, + "loss": 0.1509, + "step": 2703 + }, + { + "epoch": 0.06338601827270757, + "grad_norm": 0.44271618127822876, + "learning_rate": 0.00019803744631955066, + "loss": 0.1237, + "step": 2704 + }, + { + "epoch": 0.06340945984751256, + "grad_norm": 0.4739818572998047, + "learning_rate": 0.00019803599385685637, + "loss": 0.0899, + "step": 2705 + }, + { + "epoch": 0.06343290142231756, + "grad_norm": 0.3455488681793213, + "learning_rate": 0.00019803454086221608, + "loss": 0.157, + "step": 2706 + }, + { + "epoch": 0.06345634299712255, + "grad_norm": 0.19690151512622833, + "learning_rate": 0.00019803308733563758, + "loss": 0.0299, + "step": 2707 + }, + { + "epoch": 0.06347978457192754, + "grad_norm": 0.7890163064002991, + "learning_rate": 0.00019803163327712878, + "loss": 0.1849, + "step": 2708 + }, + { + "epoch": 0.06350322614673254, + "grad_norm": 0.3171910345554352, + "learning_rate": 0.00019803017868669757, + "loss": 0.1087, + "step": 2709 + }, + { + "epoch": 0.06352666772153753, + "grad_norm": 0.5063906908035278, + "learning_rate": 0.00019802872356435188, + "loss": 0.1885, + "step": 2710 + }, + { + "epoch": 0.06355010929634253, + "grad_norm": 0.5835399627685547, + "learning_rate": 0.00019802726791009958, + "loss": 0.3935, + "step": 2711 + }, + { + "epoch": 0.06357355087114752, + "grad_norm": 0.44595274329185486, + "learning_rate": 0.00019802581172394856, + "loss": 0.14, + "step": 2712 + }, + { + "epoch": 0.06359699244595252, + "grad_norm": 0.6006901860237122, + "learning_rate": 0.00019802435500590674, + "loss": 0.2169, + "step": 2713 + }, + { + "epoch": 0.06362043402075751, + "grad_norm": 0.5527702569961548, + "learning_rate": 0.000198022897755982, + "loss": 0.1766, + "step": 2714 + }, + { + "epoch": 0.0636438755955625, + "grad_norm": 0.620920717716217, + "learning_rate": 0.00019802143997418224, + "loss": 0.2563, + "step": 2715 + }, + { + "epoch": 0.0636673171703675, + "grad_norm": 0.8799117207527161, + "learning_rate": 0.00019801998166051545, + "loss": 0.2316, + "step": 2716 + }, + { + "epoch": 0.06369075874517251, + "grad_norm": 0.5660324692726135, + "learning_rate": 0.00019801852281498946, + "loss": 0.1741, + "step": 2717 + }, + { + "epoch": 0.0637142003199775, + "grad_norm": 0.33556705713272095, + "learning_rate": 0.00019801706343761216, + "loss": 0.111, + "step": 2718 + }, + { + "epoch": 0.0637376418947825, + "grad_norm": 0.6013329029083252, + "learning_rate": 0.0001980156035283916, + "loss": 0.1332, + "step": 2719 + }, + { + "epoch": 0.06376108346958749, + "grad_norm": 0.7427898049354553, + "learning_rate": 0.00019801414308733554, + "loss": 0.8786, + "step": 2720 + }, + { + "epoch": 0.06378452504439248, + "grad_norm": 0.6373158693313599, + "learning_rate": 0.00019801268211445204, + "loss": 0.1821, + "step": 2721 + }, + { + "epoch": 0.06380796661919748, + "grad_norm": 0.16652800142765045, + "learning_rate": 0.00019801122060974895, + "loss": 0.0376, + "step": 2722 + }, + { + "epoch": 0.06383140819400247, + "grad_norm": 0.49045804142951965, + "learning_rate": 0.00019800975857323422, + "loss": 0.1563, + "step": 2723 + }, + { + "epoch": 0.06385484976880747, + "grad_norm": 0.3478959798812866, + "learning_rate": 0.0001980082960049158, + "loss": 0.1043, + "step": 2724 + }, + { + "epoch": 0.06387829134361246, + "grad_norm": 0.39094945788383484, + "learning_rate": 0.00019800683290480158, + "loss": 0.1369, + "step": 2725 + }, + { + "epoch": 0.06390173291841746, + "grad_norm": 0.8593403100967407, + "learning_rate": 0.00019800536927289954, + "loss": 0.2985, + "step": 2726 + }, + { + "epoch": 0.06392517449322245, + "grad_norm": 0.4675174057483673, + "learning_rate": 0.00019800390510921762, + "loss": 0.0779, + "step": 2727 + }, + { + "epoch": 0.06394861606802744, + "grad_norm": 0.46070143580436707, + "learning_rate": 0.00019800244041376375, + "loss": 0.2024, + "step": 2728 + }, + { + "epoch": 0.06397205764283244, + "grad_norm": 0.24619634449481964, + "learning_rate": 0.00019800097518654587, + "loss": 0.0588, + "step": 2729 + }, + { + "epoch": 0.06399549921763745, + "grad_norm": 0.5610895156860352, + "learning_rate": 0.00019799950942757198, + "loss": 0.1377, + "step": 2730 + }, + { + "epoch": 0.06401894079244244, + "grad_norm": 0.5859628915786743, + "learning_rate": 0.00019799804313684998, + "loss": 0.1759, + "step": 2731 + }, + { + "epoch": 0.06404238236724744, + "grad_norm": 0.6518580913543701, + "learning_rate": 0.00019799657631438785, + "loss": 0.1774, + "step": 2732 + }, + { + "epoch": 0.06406582394205243, + "grad_norm": 0.236823171377182, + "learning_rate": 0.0001979951089601935, + "loss": 0.0527, + "step": 2733 + }, + { + "epoch": 0.06408926551685742, + "grad_norm": 0.8657407164573669, + "learning_rate": 0.00019799364107427495, + "loss": 0.1494, + "step": 2734 + }, + { + "epoch": 0.06411270709166242, + "grad_norm": 0.611821174621582, + "learning_rate": 0.00019799217265664018, + "loss": 0.2129, + "step": 2735 + }, + { + "epoch": 0.06413614866646741, + "grad_norm": 0.4930969774723053, + "learning_rate": 0.0001979907037072971, + "loss": 0.1504, + "step": 2736 + }, + { + "epoch": 0.06415959024127241, + "grad_norm": 0.4380182921886444, + "learning_rate": 0.0001979892342262537, + "loss": 0.1199, + "step": 2737 + }, + { + "epoch": 0.0641830318160774, + "grad_norm": 0.4784340560436249, + "learning_rate": 0.000197987764213518, + "loss": 0.1838, + "step": 2738 + }, + { + "epoch": 0.0642064733908824, + "grad_norm": 0.5343209505081177, + "learning_rate": 0.0001979862936690979, + "loss": 0.1976, + "step": 2739 + }, + { + "epoch": 0.06422991496568739, + "grad_norm": 0.42767876386642456, + "learning_rate": 0.00019798482259300143, + "loss": 0.1489, + "step": 2740 + }, + { + "epoch": 0.06425335654049238, + "grad_norm": 0.5708357691764832, + "learning_rate": 0.00019798335098523655, + "loss": 0.1508, + "step": 2741 + }, + { + "epoch": 0.06427679811529738, + "grad_norm": 0.2812480032444, + "learning_rate": 0.00019798187884581124, + "loss": 0.0773, + "step": 2742 + }, + { + "epoch": 0.06430023969010239, + "grad_norm": 0.4780259132385254, + "learning_rate": 0.0001979804061747335, + "loss": 0.1081, + "step": 2743 + }, + { + "epoch": 0.06432368126490738, + "grad_norm": 0.8084869980812073, + "learning_rate": 0.00019797893297201133, + "loss": 0.2393, + "step": 2744 + }, + { + "epoch": 0.06434712283971238, + "grad_norm": 0.48696011304855347, + "learning_rate": 0.00019797745923765274, + "loss": 0.1745, + "step": 2745 + }, + { + "epoch": 0.06437056441451737, + "grad_norm": 0.8136163353919983, + "learning_rate": 0.00019797598497166567, + "loss": 0.1899, + "step": 2746 + }, + { + "epoch": 0.06439400598932236, + "grad_norm": 0.2915462255477905, + "learning_rate": 0.00019797451017405818, + "loss": 0.0911, + "step": 2747 + }, + { + "epoch": 0.06441744756412736, + "grad_norm": 0.8766478300094604, + "learning_rate": 0.00019797303484483822, + "loss": 0.6812, + "step": 2748 + }, + { + "epoch": 0.06444088913893235, + "grad_norm": 0.7226137518882751, + "learning_rate": 0.00019797155898401383, + "loss": 0.2165, + "step": 2749 + }, + { + "epoch": 0.06446433071373735, + "grad_norm": 0.5628411173820496, + "learning_rate": 0.00019797008259159297, + "loss": 0.1243, + "step": 2750 + }, + { + "epoch": 0.06448777228854234, + "grad_norm": 0.6896789073944092, + "learning_rate": 0.00019796860566758374, + "loss": 0.175, + "step": 2751 + }, + { + "epoch": 0.06451121386334734, + "grad_norm": 0.7248033285140991, + "learning_rate": 0.00019796712821199406, + "loss": 0.131, + "step": 2752 + }, + { + "epoch": 0.06453465543815233, + "grad_norm": 0.22570449113845825, + "learning_rate": 0.00019796565022483202, + "loss": 0.0799, + "step": 2753 + }, + { + "epoch": 0.06455809701295732, + "grad_norm": 0.5545620918273926, + "learning_rate": 0.0001979641717061056, + "loss": 0.1307, + "step": 2754 + }, + { + "epoch": 0.06458153858776233, + "grad_norm": 0.33159059286117554, + "learning_rate": 0.00019796269265582282, + "loss": 0.1333, + "step": 2755 + }, + { + "epoch": 0.06460498016256733, + "grad_norm": 0.4734344184398651, + "learning_rate": 0.00019796121307399172, + "loss": 0.2036, + "step": 2756 + }, + { + "epoch": 0.06462842173737232, + "grad_norm": 0.6109330058097839, + "learning_rate": 0.0001979597329606203, + "loss": 0.2038, + "step": 2757 + }, + { + "epoch": 0.06465186331217732, + "grad_norm": 0.690758228302002, + "learning_rate": 0.00019795825231571664, + "loss": 0.7339, + "step": 2758 + }, + { + "epoch": 0.06467530488698231, + "grad_norm": 0.5693872570991516, + "learning_rate": 0.00019795677113928874, + "loss": 0.7108, + "step": 2759 + }, + { + "epoch": 0.0646987464617873, + "grad_norm": 0.9171745181083679, + "learning_rate": 0.00019795528943134465, + "loss": 0.132, + "step": 2760 + }, + { + "epoch": 0.0647221880365923, + "grad_norm": 0.5812225341796875, + "learning_rate": 0.0001979538071918924, + "loss": 0.1658, + "step": 2761 + }, + { + "epoch": 0.06474562961139729, + "grad_norm": 0.19362536072731018, + "learning_rate": 0.00019795232442094005, + "loss": 0.0471, + "step": 2762 + }, + { + "epoch": 0.06476907118620229, + "grad_norm": 0.9023741483688354, + "learning_rate": 0.00019795084111849561, + "loss": 0.245, + "step": 2763 + }, + { + "epoch": 0.06479251276100728, + "grad_norm": 0.8002048134803772, + "learning_rate": 0.0001979493572845672, + "loss": 0.1167, + "step": 2764 + }, + { + "epoch": 0.06481595433581228, + "grad_norm": 0.562515914440155, + "learning_rate": 0.00019794787291916278, + "loss": 0.1173, + "step": 2765 + }, + { + "epoch": 0.06483939591061727, + "grad_norm": 0.7785167098045349, + "learning_rate": 0.00019794638802229044, + "loss": 0.2283, + "step": 2766 + }, + { + "epoch": 0.06486283748542226, + "grad_norm": 0.32261478900909424, + "learning_rate": 0.0001979449025939583, + "loss": 0.0708, + "step": 2767 + }, + { + "epoch": 0.06488627906022727, + "grad_norm": 0.27449050545692444, + "learning_rate": 0.0001979434166341743, + "loss": 0.298, + "step": 2768 + }, + { + "epoch": 0.06490972063503227, + "grad_norm": 0.9218088388442993, + "learning_rate": 0.00019794193014294659, + "loss": 0.1659, + "step": 2769 + }, + { + "epoch": 0.06493316220983726, + "grad_norm": 0.7394668459892273, + "learning_rate": 0.00019794044312028324, + "loss": 0.1432, + "step": 2770 + }, + { + "epoch": 0.06495660378464226, + "grad_norm": 0.5088191628456116, + "learning_rate": 0.00019793895556619227, + "loss": 0.1569, + "step": 2771 + }, + { + "epoch": 0.06498004535944725, + "grad_norm": 0.2709334194660187, + "learning_rate": 0.00019793746748068177, + "loss": 0.0855, + "step": 2772 + }, + { + "epoch": 0.06500348693425224, + "grad_norm": 0.5712910294532776, + "learning_rate": 0.00019793597886375982, + "loss": 0.0611, + "step": 2773 + }, + { + "epoch": 0.06502692850905724, + "grad_norm": 0.5834510326385498, + "learning_rate": 0.0001979344897154345, + "loss": 0.091, + "step": 2774 + }, + { + "epoch": 0.06505037008386223, + "grad_norm": 0.6464034914970398, + "learning_rate": 0.00019793300003571385, + "loss": 0.1983, + "step": 2775 + }, + { + "epoch": 0.06507381165866723, + "grad_norm": 0.5647456049919128, + "learning_rate": 0.00019793150982460605, + "loss": 0.1316, + "step": 2776 + }, + { + "epoch": 0.06509725323347222, + "grad_norm": 0.49010705947875977, + "learning_rate": 0.00019793001908211907, + "loss": 0.1745, + "step": 2777 + }, + { + "epoch": 0.06512069480827722, + "grad_norm": 0.2390076369047165, + "learning_rate": 0.00019792852780826107, + "loss": 0.0614, + "step": 2778 + }, + { + "epoch": 0.06514413638308221, + "grad_norm": 0.5871204137802124, + "learning_rate": 0.00019792703600304015, + "loss": 0.159, + "step": 2779 + }, + { + "epoch": 0.0651675779578872, + "grad_norm": 0.5794686079025269, + "learning_rate": 0.00019792554366646436, + "loss": 0.1867, + "step": 2780 + }, + { + "epoch": 0.06519101953269221, + "grad_norm": 0.3912273049354553, + "learning_rate": 0.00019792405079854182, + "loss": 0.062, + "step": 2781 + }, + { + "epoch": 0.0652144611074972, + "grad_norm": 0.663577675819397, + "learning_rate": 0.00019792255739928063, + "loss": 0.1757, + "step": 2782 + }, + { + "epoch": 0.0652379026823022, + "grad_norm": 0.9375927448272705, + "learning_rate": 0.00019792106346868887, + "loss": 0.2001, + "step": 2783 + }, + { + "epoch": 0.0652613442571072, + "grad_norm": 0.4907112717628479, + "learning_rate": 0.00019791956900677468, + "loss": 0.1084, + "step": 2784 + }, + { + "epoch": 0.06528478583191219, + "grad_norm": 0.7999123334884644, + "learning_rate": 0.00019791807401354616, + "loss": 0.5935, + "step": 2785 + }, + { + "epoch": 0.06530822740671718, + "grad_norm": 0.47881627082824707, + "learning_rate": 0.0001979165784890114, + "loss": 0.1064, + "step": 2786 + }, + { + "epoch": 0.06533166898152218, + "grad_norm": 0.5495904088020325, + "learning_rate": 0.00019791508243317856, + "loss": 0.1392, + "step": 2787 + }, + { + "epoch": 0.06535511055632717, + "grad_norm": 0.9611420631408691, + "learning_rate": 0.00019791358584605572, + "loss": 0.1776, + "step": 2788 + }, + { + "epoch": 0.06537855213113217, + "grad_norm": 0.5917728543281555, + "learning_rate": 0.00019791208872765103, + "loss": 0.192, + "step": 2789 + }, + { + "epoch": 0.06540199370593716, + "grad_norm": 0.5299933552742004, + "learning_rate": 0.00019791059107797254, + "loss": 0.1312, + "step": 2790 + }, + { + "epoch": 0.06542543528074216, + "grad_norm": 0.5182585120201111, + "learning_rate": 0.00019790909289702847, + "loss": 0.1633, + "step": 2791 + }, + { + "epoch": 0.06544887685554715, + "grad_norm": 0.26131293177604675, + "learning_rate": 0.00019790759418482687, + "loss": 0.0784, + "step": 2792 + }, + { + "epoch": 0.06547231843035214, + "grad_norm": 0.583564043045044, + "learning_rate": 0.00019790609494137594, + "loss": 0.1585, + "step": 2793 + }, + { + "epoch": 0.06549576000515715, + "grad_norm": 0.5572071075439453, + "learning_rate": 0.0001979045951666838, + "loss": 0.1568, + "step": 2794 + }, + { + "epoch": 0.06551920157996215, + "grad_norm": 0.965817391872406, + "learning_rate": 0.00019790309486075855, + "loss": 0.3209, + "step": 2795 + }, + { + "epoch": 0.06554264315476714, + "grad_norm": 0.5359334945678711, + "learning_rate": 0.00019790159402360838, + "loss": 0.1633, + "step": 2796 + }, + { + "epoch": 0.06556608472957214, + "grad_norm": 0.9533064365386963, + "learning_rate": 0.00019790009265524138, + "loss": 0.1318, + "step": 2797 + }, + { + "epoch": 0.06558952630437713, + "grad_norm": 0.3699539601802826, + "learning_rate": 0.00019789859075566572, + "loss": 0.1193, + "step": 2798 + }, + { + "epoch": 0.06561296787918212, + "grad_norm": 0.49512404203414917, + "learning_rate": 0.00019789708832488958, + "loss": 0.1053, + "step": 2799 + }, + { + "epoch": 0.06563640945398712, + "grad_norm": 0.3488523066043854, + "learning_rate": 0.00019789558536292105, + "loss": 0.0998, + "step": 2800 + }, + { + "epoch": 0.06565985102879211, + "grad_norm": 0.9262507557868958, + "learning_rate": 0.00019789408186976837, + "loss": 0.3956, + "step": 2801 + }, + { + "epoch": 0.06568329260359711, + "grad_norm": 0.6043416261672974, + "learning_rate": 0.00019789257784543962, + "loss": 0.6699, + "step": 2802 + }, + { + "epoch": 0.0657067341784021, + "grad_norm": 0.46750739216804504, + "learning_rate": 0.000197891073289943, + "loss": 0.5848, + "step": 2803 + }, + { + "epoch": 0.0657301757532071, + "grad_norm": 0.5544444918632507, + "learning_rate": 0.00019788956820328667, + "loss": 0.0804, + "step": 2804 + }, + { + "epoch": 0.06575361732801209, + "grad_norm": 0.501983106136322, + "learning_rate": 0.00019788806258547875, + "loss": 0.164, + "step": 2805 + }, + { + "epoch": 0.06577705890281708, + "grad_norm": 0.9354178309440613, + "learning_rate": 0.0001978865564365275, + "loss": 0.2335, + "step": 2806 + }, + { + "epoch": 0.06580050047762209, + "grad_norm": 0.2625947892665863, + "learning_rate": 0.000197885049756441, + "loss": 0.0724, + "step": 2807 + }, + { + "epoch": 0.06582394205242709, + "grad_norm": 0.37445083260536194, + "learning_rate": 0.00019788354254522748, + "loss": 0.0912, + "step": 2808 + }, + { + "epoch": 0.06584738362723208, + "grad_norm": 0.6297499537467957, + "learning_rate": 0.00019788203480289513, + "loss": 0.1969, + "step": 2809 + }, + { + "epoch": 0.06587082520203708, + "grad_norm": 0.7785255908966064, + "learning_rate": 0.00019788052652945211, + "loss": 0.2009, + "step": 2810 + }, + { + "epoch": 0.06589426677684207, + "grad_norm": 0.38819193840026855, + "learning_rate": 0.0001978790177249066, + "loss": 0.1102, + "step": 2811 + }, + { + "epoch": 0.06591770835164706, + "grad_norm": 0.5835219621658325, + "learning_rate": 0.00019787750838926675, + "loss": 0.1917, + "step": 2812 + }, + { + "epoch": 0.06594114992645206, + "grad_norm": 0.8843295574188232, + "learning_rate": 0.00019787599852254083, + "loss": 0.1578, + "step": 2813 + }, + { + "epoch": 0.06596459150125705, + "grad_norm": 0.6216434240341187, + "learning_rate": 0.00019787448812473696, + "loss": 0.1613, + "step": 2814 + }, + { + "epoch": 0.06598803307606205, + "grad_norm": 0.5103811621665955, + "learning_rate": 0.00019787297719586335, + "loss": 0.1234, + "step": 2815 + }, + { + "epoch": 0.06601147465086704, + "grad_norm": 0.5459553599357605, + "learning_rate": 0.00019787146573592825, + "loss": 0.0918, + "step": 2816 + }, + { + "epoch": 0.06603491622567204, + "grad_norm": 0.6766862273216248, + "learning_rate": 0.0001978699537449398, + "loss": 0.181, + "step": 2817 + }, + { + "epoch": 0.06605835780047703, + "grad_norm": 0.3845318555831909, + "learning_rate": 0.00019786844122290624, + "loss": 0.0882, + "step": 2818 + }, + { + "epoch": 0.06608179937528204, + "grad_norm": 0.796278178691864, + "learning_rate": 0.0001978669281698358, + "loss": 0.1674, + "step": 2819 + }, + { + "epoch": 0.06610524095008703, + "grad_norm": 0.45387589931488037, + "learning_rate": 0.0001978654145857366, + "loss": 0.4305, + "step": 2820 + }, + { + "epoch": 0.06612868252489203, + "grad_norm": 1.4948612451553345, + "learning_rate": 0.00019786390047061693, + "loss": 0.2095, + "step": 2821 + }, + { + "epoch": 0.06615212409969702, + "grad_norm": 0.4365345537662506, + "learning_rate": 0.000197862385824485, + "loss": 0.1007, + "step": 2822 + }, + { + "epoch": 0.06617556567450202, + "grad_norm": 0.17364849150180817, + "learning_rate": 0.00019786087064734898, + "loss": 0.0521, + "step": 2823 + }, + { + "epoch": 0.06619900724930701, + "grad_norm": 0.19794614613056183, + "learning_rate": 0.00019785935493921717, + "loss": 0.0438, + "step": 2824 + }, + { + "epoch": 0.066222448824112, + "grad_norm": 0.2390986829996109, + "learning_rate": 0.00019785783870009772, + "loss": 0.0622, + "step": 2825 + }, + { + "epoch": 0.066245890398917, + "grad_norm": 1.0970613956451416, + "learning_rate": 0.00019785632192999888, + "loss": 0.351, + "step": 2826 + }, + { + "epoch": 0.06626933197372199, + "grad_norm": 0.25908252596855164, + "learning_rate": 0.00019785480462892888, + "loss": 0.08, + "step": 2827 + }, + { + "epoch": 0.06629277354852699, + "grad_norm": 0.7911470532417297, + "learning_rate": 0.000197853286796896, + "loss": 0.3294, + "step": 2828 + }, + { + "epoch": 0.06631621512333198, + "grad_norm": 0.6282100081443787, + "learning_rate": 0.0001978517684339084, + "loss": 0.1883, + "step": 2829 + }, + { + "epoch": 0.06633965669813698, + "grad_norm": 0.4679584801197052, + "learning_rate": 0.00019785024953997436, + "loss": 0.1276, + "step": 2830 + }, + { + "epoch": 0.06636309827294197, + "grad_norm": 0.759888768196106, + "learning_rate": 0.0001978487301151021, + "loss": 0.232, + "step": 2831 + }, + { + "epoch": 0.06638653984774698, + "grad_norm": 0.1798451840877533, + "learning_rate": 0.0001978472101592999, + "loss": 0.0314, + "step": 2832 + }, + { + "epoch": 0.06640998142255197, + "grad_norm": 0.49935048818588257, + "learning_rate": 0.000197845689672576, + "loss": 0.1057, + "step": 2833 + }, + { + "epoch": 0.06643342299735697, + "grad_norm": 0.26330631971359253, + "learning_rate": 0.0001978441686549386, + "loss": 0.0486, + "step": 2834 + }, + { + "epoch": 0.06645686457216196, + "grad_norm": 0.663611888885498, + "learning_rate": 0.00019784264710639602, + "loss": 0.1676, + "step": 2835 + }, + { + "epoch": 0.06648030614696696, + "grad_norm": 0.5152636170387268, + "learning_rate": 0.00019784112502695646, + "loss": 0.1777, + "step": 2836 + }, + { + "epoch": 0.06650374772177195, + "grad_norm": 0.543063759803772, + "learning_rate": 0.00019783960241662822, + "loss": 0.1464, + "step": 2837 + }, + { + "epoch": 0.06652718929657694, + "grad_norm": 0.6395561099052429, + "learning_rate": 0.00019783807927541955, + "loss": 0.1581, + "step": 2838 + }, + { + "epoch": 0.06655063087138194, + "grad_norm": 0.5409878492355347, + "learning_rate": 0.0001978365556033387, + "loss": 0.7806, + "step": 2839 + }, + { + "epoch": 0.06657407244618693, + "grad_norm": 0.34830814599990845, + "learning_rate": 0.00019783503140039397, + "loss": 0.122, + "step": 2840 + }, + { + "epoch": 0.06659751402099193, + "grad_norm": 0.2707661986351013, + "learning_rate": 0.0001978335066665936, + "loss": 0.1148, + "step": 2841 + }, + { + "epoch": 0.06662095559579692, + "grad_norm": 0.8662911057472229, + "learning_rate": 0.00019783198140194586, + "loss": 0.1286, + "step": 2842 + }, + { + "epoch": 0.06664439717060192, + "grad_norm": 0.522050678730011, + "learning_rate": 0.00019783045560645902, + "loss": 0.1361, + "step": 2843 + }, + { + "epoch": 0.06666783874540691, + "grad_norm": 0.5023254752159119, + "learning_rate": 0.0001978289292801414, + "loss": 0.1051, + "step": 2844 + }, + { + "epoch": 0.06669128032021192, + "grad_norm": 0.5527602434158325, + "learning_rate": 0.00019782740242300126, + "loss": 0.131, + "step": 2845 + }, + { + "epoch": 0.06671472189501691, + "grad_norm": 0.5452955961227417, + "learning_rate": 0.0001978258750350469, + "loss": 0.1466, + "step": 2846 + }, + { + "epoch": 0.0667381634698219, + "grad_norm": 0.1808152198791504, + "learning_rate": 0.00019782434711628657, + "loss": 0.0388, + "step": 2847 + }, + { + "epoch": 0.0667616050446269, + "grad_norm": 0.6203113198280334, + "learning_rate": 0.00019782281866672857, + "loss": 0.2041, + "step": 2848 + }, + { + "epoch": 0.0667850466194319, + "grad_norm": 0.6355486512184143, + "learning_rate": 0.00019782128968638124, + "loss": 0.1525, + "step": 2849 + }, + { + "epoch": 0.06680848819423689, + "grad_norm": 0.8963161706924438, + "learning_rate": 0.0001978197601752528, + "loss": 0.2463, + "step": 2850 + }, + { + "epoch": 0.06683192976904188, + "grad_norm": 0.2266157865524292, + "learning_rate": 0.00019781823013335163, + "loss": 0.0509, + "step": 2851 + }, + { + "epoch": 0.06685537134384688, + "grad_norm": 0.558499276638031, + "learning_rate": 0.00019781669956068595, + "loss": 0.1484, + "step": 2852 + }, + { + "epoch": 0.06687881291865187, + "grad_norm": 0.6636942028999329, + "learning_rate": 0.00019781516845726415, + "loss": 0.1891, + "step": 2853 + }, + { + "epoch": 0.06690225449345687, + "grad_norm": 0.41642627120018005, + "learning_rate": 0.00019781363682309447, + "loss": 0.1092, + "step": 2854 + }, + { + "epoch": 0.06692569606826186, + "grad_norm": 0.7207500338554382, + "learning_rate": 0.00019781210465818526, + "loss": 0.1794, + "step": 2855 + }, + { + "epoch": 0.06694913764306686, + "grad_norm": 0.6803163886070251, + "learning_rate": 0.00019781057196254482, + "loss": 0.1954, + "step": 2856 + }, + { + "epoch": 0.06697257921787185, + "grad_norm": 0.27261510491371155, + "learning_rate": 0.00019780903873618146, + "loss": 0.0701, + "step": 2857 + }, + { + "epoch": 0.06699602079267686, + "grad_norm": 0.692280650138855, + "learning_rate": 0.0001978075049791035, + "loss": 0.1474, + "step": 2858 + }, + { + "epoch": 0.06701946236748185, + "grad_norm": 0.5898426175117493, + "learning_rate": 0.0001978059706913193, + "loss": 0.2138, + "step": 2859 + }, + { + "epoch": 0.06704290394228685, + "grad_norm": 0.766025185585022, + "learning_rate": 0.00019780443587283714, + "loss": 0.2807, + "step": 2860 + }, + { + "epoch": 0.06706634551709184, + "grad_norm": 0.6999529600143433, + "learning_rate": 0.00019780290052366534, + "loss": 0.2249, + "step": 2861 + }, + { + "epoch": 0.06708978709189684, + "grad_norm": 0.4122408926486969, + "learning_rate": 0.00019780136464381224, + "loss": 0.1819, + "step": 2862 + }, + { + "epoch": 0.06711322866670183, + "grad_norm": 0.7094532251358032, + "learning_rate": 0.0001977998282332862, + "loss": 0.782, + "step": 2863 + }, + { + "epoch": 0.06713667024150682, + "grad_norm": 0.4587853252887726, + "learning_rate": 0.00019779829129209557, + "loss": 0.4419, + "step": 2864 + }, + { + "epoch": 0.06716011181631182, + "grad_norm": 0.5567826628684998, + "learning_rate": 0.00019779675382024863, + "loss": 0.2212, + "step": 2865 + }, + { + "epoch": 0.06718355339111681, + "grad_norm": 0.6397860050201416, + "learning_rate": 0.00019779521581775376, + "loss": 0.1658, + "step": 2866 + }, + { + "epoch": 0.0672069949659218, + "grad_norm": 0.631488025188446, + "learning_rate": 0.0001977936772846193, + "loss": 0.2293, + "step": 2867 + }, + { + "epoch": 0.0672304365407268, + "grad_norm": 0.5686105489730835, + "learning_rate": 0.00019779213822085362, + "loss": 0.1455, + "step": 2868 + }, + { + "epoch": 0.0672538781155318, + "grad_norm": 0.4499674141407013, + "learning_rate": 0.00019779059862646504, + "loss": 0.0699, + "step": 2869 + }, + { + "epoch": 0.0672773196903368, + "grad_norm": 1.0766396522521973, + "learning_rate": 0.0001977890585014619, + "loss": 0.1902, + "step": 2870 + }, + { + "epoch": 0.0673007612651418, + "grad_norm": 0.6293129920959473, + "learning_rate": 0.0001977875178458526, + "loss": 0.2306, + "step": 2871 + }, + { + "epoch": 0.06732420283994679, + "grad_norm": 0.34327948093414307, + "learning_rate": 0.00019778597665964546, + "loss": 0.0717, + "step": 2872 + }, + { + "epoch": 0.06734764441475179, + "grad_norm": 0.6436792016029358, + "learning_rate": 0.00019778443494284886, + "loss": 0.1669, + "step": 2873 + }, + { + "epoch": 0.06737108598955678, + "grad_norm": 0.5578792691230774, + "learning_rate": 0.0001977828926954712, + "loss": 0.2176, + "step": 2874 + }, + { + "epoch": 0.06739452756436178, + "grad_norm": 0.2680445909500122, + "learning_rate": 0.00019778134991752078, + "loss": 0.0831, + "step": 2875 + }, + { + "epoch": 0.06741796913916677, + "grad_norm": 0.23430374264717102, + "learning_rate": 0.00019777980660900602, + "loss": 0.0415, + "step": 2876 + }, + { + "epoch": 0.06744141071397176, + "grad_norm": 0.5628007650375366, + "learning_rate": 0.00019777826276993525, + "loss": 0.2064, + "step": 2877 + }, + { + "epoch": 0.06746485228877676, + "grad_norm": 0.48105040192604065, + "learning_rate": 0.00019777671840031692, + "loss": 0.1459, + "step": 2878 + }, + { + "epoch": 0.06748829386358175, + "grad_norm": 0.5826212167739868, + "learning_rate": 0.00019777517350015936, + "loss": 0.2524, + "step": 2879 + }, + { + "epoch": 0.06751173543838675, + "grad_norm": 0.6156630516052246, + "learning_rate": 0.00019777362806947096, + "loss": 0.1458, + "step": 2880 + }, + { + "epoch": 0.06753517701319174, + "grad_norm": 0.5170184373855591, + "learning_rate": 0.00019777208210826008, + "loss": 0.1321, + "step": 2881 + }, + { + "epoch": 0.06755861858799674, + "grad_norm": 0.4024691581726074, + "learning_rate": 0.00019777053561653515, + "loss": 0.1282, + "step": 2882 + }, + { + "epoch": 0.06758206016280174, + "grad_norm": 0.6924253702163696, + "learning_rate": 0.00019776898859430456, + "loss": 0.19, + "step": 2883 + }, + { + "epoch": 0.06760550173760674, + "grad_norm": 0.4415542781352997, + "learning_rate": 0.00019776744104157664, + "loss": 0.0734, + "step": 2884 + }, + { + "epoch": 0.06762894331241173, + "grad_norm": 0.685192346572876, + "learning_rate": 0.0001977658929583599, + "loss": 0.1436, + "step": 2885 + }, + { + "epoch": 0.06765238488721673, + "grad_norm": 0.7438700199127197, + "learning_rate": 0.00019776434434466263, + "loss": 0.2161, + "step": 2886 + }, + { + "epoch": 0.06767582646202172, + "grad_norm": 0.5897936820983887, + "learning_rate": 0.00019776279520049327, + "loss": 0.1717, + "step": 2887 + }, + { + "epoch": 0.06769926803682672, + "grad_norm": 0.34064847230911255, + "learning_rate": 0.0001977612455258603, + "loss": 0.0519, + "step": 2888 + }, + { + "epoch": 0.06772270961163171, + "grad_norm": 0.40103834867477417, + "learning_rate": 0.000197759695320772, + "loss": 0.0813, + "step": 2889 + }, + { + "epoch": 0.0677461511864367, + "grad_norm": 0.2213754802942276, + "learning_rate": 0.00019775814458523688, + "loss": 0.0598, + "step": 2890 + }, + { + "epoch": 0.0677695927612417, + "grad_norm": 0.5298768281936646, + "learning_rate": 0.0001977565933192633, + "loss": 0.1155, + "step": 2891 + }, + { + "epoch": 0.06779303433604669, + "grad_norm": 0.6603960394859314, + "learning_rate": 0.00019775504152285967, + "loss": 0.1938, + "step": 2892 + }, + { + "epoch": 0.06781647591085169, + "grad_norm": 0.9571162462234497, + "learning_rate": 0.00019775348919603446, + "loss": 0.903, + "step": 2893 + }, + { + "epoch": 0.06783991748565668, + "grad_norm": 0.40735653042793274, + "learning_rate": 0.0001977519363387961, + "loss": 0.1322, + "step": 2894 + }, + { + "epoch": 0.06786335906046168, + "grad_norm": 0.4962051808834076, + "learning_rate": 0.00019775038295115295, + "loss": 0.1118, + "step": 2895 + }, + { + "epoch": 0.06788680063526668, + "grad_norm": 0.712087094783783, + "learning_rate": 0.00019774882903311348, + "loss": 0.1601, + "step": 2896 + }, + { + "epoch": 0.06791024221007168, + "grad_norm": 0.32640185952186584, + "learning_rate": 0.00019774727458468614, + "loss": 0.3542, + "step": 2897 + }, + { + "epoch": 0.06793368378487667, + "grad_norm": 0.3169027268886566, + "learning_rate": 0.00019774571960587932, + "loss": 0.0537, + "step": 2898 + }, + { + "epoch": 0.06795712535968167, + "grad_norm": 0.4911491870880127, + "learning_rate": 0.00019774416409670147, + "loss": 0.2283, + "step": 2899 + }, + { + "epoch": 0.06798056693448666, + "grad_norm": 0.1473390907049179, + "learning_rate": 0.00019774260805716104, + "loss": 0.0257, + "step": 2900 + }, + { + "epoch": 0.06800400850929166, + "grad_norm": 0.6564428806304932, + "learning_rate": 0.00019774105148726647, + "loss": 0.1681, + "step": 2901 + }, + { + "epoch": 0.06802745008409665, + "grad_norm": 0.7858428955078125, + "learning_rate": 0.00019773949438702622, + "loss": 0.8433, + "step": 2902 + }, + { + "epoch": 0.06805089165890164, + "grad_norm": 0.6717924475669861, + "learning_rate": 0.0001977379367564487, + "loss": 0.2096, + "step": 2903 + }, + { + "epoch": 0.06807433323370664, + "grad_norm": 0.2939845323562622, + "learning_rate": 0.0001977363785955424, + "loss": 0.0524, + "step": 2904 + }, + { + "epoch": 0.06809777480851163, + "grad_norm": 0.11165302246809006, + "learning_rate": 0.00019773481990431577, + "loss": 0.0271, + "step": 2905 + }, + { + "epoch": 0.06812121638331663, + "grad_norm": 0.7513301372528076, + "learning_rate": 0.00019773326068277727, + "loss": 0.2007, + "step": 2906 + }, + { + "epoch": 0.06814465795812162, + "grad_norm": 0.4032916724681854, + "learning_rate": 0.00019773170093093532, + "loss": 0.3684, + "step": 2907 + }, + { + "epoch": 0.06816809953292662, + "grad_norm": 0.5654211640357971, + "learning_rate": 0.00019773014064879843, + "loss": 0.1798, + "step": 2908 + }, + { + "epoch": 0.06819154110773162, + "grad_norm": 0.4155657887458801, + "learning_rate": 0.00019772857983637503, + "loss": 0.1871, + "step": 2909 + }, + { + "epoch": 0.06821498268253662, + "grad_norm": 0.2306881994009018, + "learning_rate": 0.0001977270184936736, + "loss": 0.0863, + "step": 2910 + }, + { + "epoch": 0.06823842425734161, + "grad_norm": 0.5956360101699829, + "learning_rate": 0.00019772545662070264, + "loss": 0.6673, + "step": 2911 + }, + { + "epoch": 0.0682618658321466, + "grad_norm": 0.34437844157218933, + "learning_rate": 0.00019772389421747062, + "loss": 0.4094, + "step": 2912 + }, + { + "epoch": 0.0682853074069516, + "grad_norm": 0.45788902044296265, + "learning_rate": 0.00019772233128398598, + "loss": 0.204, + "step": 2913 + }, + { + "epoch": 0.0683087489817566, + "grad_norm": 0.6984483599662781, + "learning_rate": 0.00019772076782025724, + "loss": 0.1788, + "step": 2914 + }, + { + "epoch": 0.06833219055656159, + "grad_norm": 0.24061790108680725, + "learning_rate": 0.00019771920382629285, + "loss": 0.0232, + "step": 2915 + }, + { + "epoch": 0.06835563213136658, + "grad_norm": 0.587103545665741, + "learning_rate": 0.00019771763930210133, + "loss": 0.1956, + "step": 2916 + }, + { + "epoch": 0.06837907370617158, + "grad_norm": 0.7869679927825928, + "learning_rate": 0.00019771607424769115, + "loss": 0.1667, + "step": 2917 + }, + { + "epoch": 0.06840251528097657, + "grad_norm": 0.6018959879875183, + "learning_rate": 0.0001977145086630708, + "loss": 0.5304, + "step": 2918 + }, + { + "epoch": 0.06842595685578157, + "grad_norm": 0.5860605835914612, + "learning_rate": 0.00019771294254824877, + "loss": 0.1822, + "step": 2919 + }, + { + "epoch": 0.06844939843058656, + "grad_norm": 0.663175642490387, + "learning_rate": 0.00019771137590323355, + "loss": 0.6904, + "step": 2920 + }, + { + "epoch": 0.06847284000539156, + "grad_norm": 0.46324652433395386, + "learning_rate": 0.0001977098087280337, + "loss": 0.1253, + "step": 2921 + }, + { + "epoch": 0.06849628158019656, + "grad_norm": 0.5957575440406799, + "learning_rate": 0.00019770824102265767, + "loss": 0.867, + "step": 2922 + }, + { + "epoch": 0.06851972315500156, + "grad_norm": 0.5177921652793884, + "learning_rate": 0.00019770667278711394, + "loss": 0.1459, + "step": 2923 + }, + { + "epoch": 0.06854316472980655, + "grad_norm": 0.8009453415870667, + "learning_rate": 0.00019770510402141106, + "loss": 0.2941, + "step": 2924 + }, + { + "epoch": 0.06856660630461155, + "grad_norm": 0.4674127697944641, + "learning_rate": 0.0001977035347255576, + "loss": 0.0924, + "step": 2925 + }, + { + "epoch": 0.06859004787941654, + "grad_norm": 0.18763142824172974, + "learning_rate": 0.00019770196489956194, + "loss": 0.0611, + "step": 2926 + }, + { + "epoch": 0.06861348945422154, + "grad_norm": 0.4743148982524872, + "learning_rate": 0.0001977003945434327, + "loss": 0.2037, + "step": 2927 + }, + { + "epoch": 0.06863693102902653, + "grad_norm": 0.4970839023590088, + "learning_rate": 0.0001976988236571784, + "loss": 0.6025, + "step": 2928 + }, + { + "epoch": 0.06866037260383152, + "grad_norm": 0.5509539246559143, + "learning_rate": 0.0001976972522408075, + "loss": 0.1703, + "step": 2929 + }, + { + "epoch": 0.06868381417863652, + "grad_norm": 0.5159246921539307, + "learning_rate": 0.00019769568029432856, + "loss": 0.1107, + "step": 2930 + }, + { + "epoch": 0.06870725575344151, + "grad_norm": 0.3698968291282654, + "learning_rate": 0.00019769410781775014, + "loss": 0.0884, + "step": 2931 + }, + { + "epoch": 0.0687306973282465, + "grad_norm": 0.7422848343849182, + "learning_rate": 0.00019769253481108075, + "loss": 0.2017, + "step": 2932 + }, + { + "epoch": 0.0687541389030515, + "grad_norm": 0.4528050422668457, + "learning_rate": 0.00019769096127432887, + "loss": 0.1355, + "step": 2933 + }, + { + "epoch": 0.06877758047785651, + "grad_norm": 0.5355364084243774, + "learning_rate": 0.00019768938720750312, + "loss": 0.1339, + "step": 2934 + }, + { + "epoch": 0.0688010220526615, + "grad_norm": 0.6774614453315735, + "learning_rate": 0.000197687812610612, + "loss": 0.1346, + "step": 2935 + }, + { + "epoch": 0.0688244636274665, + "grad_norm": 0.4042290151119232, + "learning_rate": 0.00019768623748366406, + "loss": 0.3698, + "step": 2936 + }, + { + "epoch": 0.06884790520227149, + "grad_norm": 0.49371930956840515, + "learning_rate": 0.00019768466182666785, + "loss": 0.1658, + "step": 2937 + }, + { + "epoch": 0.06887134677707649, + "grad_norm": 0.8079572916030884, + "learning_rate": 0.00019768308563963193, + "loss": 0.2961, + "step": 2938 + }, + { + "epoch": 0.06889478835188148, + "grad_norm": 0.6443371772766113, + "learning_rate": 0.00019768150892256485, + "loss": 0.1842, + "step": 2939 + }, + { + "epoch": 0.06891822992668648, + "grad_norm": 0.694534957408905, + "learning_rate": 0.0001976799316754751, + "loss": 0.1108, + "step": 2940 + }, + { + "epoch": 0.06894167150149147, + "grad_norm": 0.5042530298233032, + "learning_rate": 0.00019767835389837134, + "loss": 0.0729, + "step": 2941 + }, + { + "epoch": 0.06896511307629646, + "grad_norm": 0.9814221262931824, + "learning_rate": 0.00019767677559126212, + "loss": 0.1549, + "step": 2942 + }, + { + "epoch": 0.06898855465110146, + "grad_norm": 0.40109676122665405, + "learning_rate": 0.0001976751967541559, + "loss": 0.096, + "step": 2943 + }, + { + "epoch": 0.06901199622590645, + "grad_norm": 0.7379030585289001, + "learning_rate": 0.00019767361738706136, + "loss": 0.1695, + "step": 2944 + }, + { + "epoch": 0.06903543780071145, + "grad_norm": 0.5002499222755432, + "learning_rate": 0.00019767203748998703, + "loss": 0.6908, + "step": 2945 + }, + { + "epoch": 0.06905887937551644, + "grad_norm": 0.3648214638233185, + "learning_rate": 0.00019767045706294142, + "loss": 0.1786, + "step": 2946 + }, + { + "epoch": 0.06908232095032145, + "grad_norm": 0.6223915219306946, + "learning_rate": 0.00019766887610593324, + "loss": 0.1104, + "step": 2947 + }, + { + "epoch": 0.06910576252512644, + "grad_norm": 0.7540935277938843, + "learning_rate": 0.00019766729461897093, + "loss": 0.1834, + "step": 2948 + }, + { + "epoch": 0.06912920409993144, + "grad_norm": 0.6722345352172852, + "learning_rate": 0.00019766571260206316, + "loss": 0.7679, + "step": 2949 + }, + { + "epoch": 0.06915264567473643, + "grad_norm": 0.6967960000038147, + "learning_rate": 0.00019766413005521848, + "loss": 0.1547, + "step": 2950 + }, + { + "epoch": 0.06917608724954143, + "grad_norm": 0.3503660261631012, + "learning_rate": 0.00019766254697844548, + "loss": 0.0939, + "step": 2951 + }, + { + "epoch": 0.06919952882434642, + "grad_norm": 0.19204233586788177, + "learning_rate": 0.00019766096337175276, + "loss": 0.0364, + "step": 2952 + }, + { + "epoch": 0.06922297039915142, + "grad_norm": 0.8548502922058105, + "learning_rate": 0.0001976593792351489, + "loss": 0.2581, + "step": 2953 + }, + { + "epoch": 0.06924641197395641, + "grad_norm": 0.8131521344184875, + "learning_rate": 0.00019765779456864255, + "loss": 0.1495, + "step": 2954 + }, + { + "epoch": 0.0692698535487614, + "grad_norm": 0.5627081990242004, + "learning_rate": 0.00019765620937224222, + "loss": 0.1562, + "step": 2955 + }, + { + "epoch": 0.0692932951235664, + "grad_norm": 0.5045002102851868, + "learning_rate": 0.00019765462364595656, + "loss": 0.1387, + "step": 2956 + }, + { + "epoch": 0.06931673669837139, + "grad_norm": 0.4574079215526581, + "learning_rate": 0.00019765303738979414, + "loss": 0.1121, + "step": 2957 + }, + { + "epoch": 0.06934017827317639, + "grad_norm": 1.1939356327056885, + "learning_rate": 0.00019765145060376363, + "loss": 0.3986, + "step": 2958 + }, + { + "epoch": 0.06936361984798138, + "grad_norm": 0.3857404887676239, + "learning_rate": 0.00019764986328787358, + "loss": 0.5129, + "step": 2959 + }, + { + "epoch": 0.06938706142278639, + "grad_norm": 0.6314467787742615, + "learning_rate": 0.00019764827544213262, + "loss": 0.1079, + "step": 2960 + }, + { + "epoch": 0.06941050299759138, + "grad_norm": 0.21587346494197845, + "learning_rate": 0.00019764668706654938, + "loss": 0.0844, + "step": 2961 + }, + { + "epoch": 0.06943394457239638, + "grad_norm": 0.24210430681705475, + "learning_rate": 0.0001976450981611325, + "loss": 0.0651, + "step": 2962 + }, + { + "epoch": 0.06945738614720137, + "grad_norm": 0.7584033012390137, + "learning_rate": 0.00019764350872589055, + "loss": 0.1975, + "step": 2963 + }, + { + "epoch": 0.06948082772200637, + "grad_norm": 0.6116912961006165, + "learning_rate": 0.00019764191876083217, + "loss": 0.1251, + "step": 2964 + }, + { + "epoch": 0.06950426929681136, + "grad_norm": 0.507893979549408, + "learning_rate": 0.000197640328265966, + "loss": 0.1085, + "step": 2965 + }, + { + "epoch": 0.06952771087161635, + "grad_norm": 0.9910212159156799, + "learning_rate": 0.00019763873724130068, + "loss": 0.2496, + "step": 2966 + }, + { + "epoch": 0.06955115244642135, + "grad_norm": 0.6406585574150085, + "learning_rate": 0.0001976371456868448, + "loss": 0.2312, + "step": 2967 + }, + { + "epoch": 0.06957459402122634, + "grad_norm": 0.42499300837516785, + "learning_rate": 0.00019763555360260702, + "loss": 0.104, + "step": 2968 + }, + { + "epoch": 0.06959803559603134, + "grad_norm": 0.4936681091785431, + "learning_rate": 0.000197633960988596, + "loss": 0.1858, + "step": 2969 + }, + { + "epoch": 0.06962147717083633, + "grad_norm": 0.5983418822288513, + "learning_rate": 0.00019763236784482033, + "loss": 0.2051, + "step": 2970 + }, + { + "epoch": 0.06964491874564133, + "grad_norm": 0.8753874897956848, + "learning_rate": 0.0001976307741712887, + "loss": 0.208, + "step": 2971 + }, + { + "epoch": 0.06966836032044632, + "grad_norm": 0.5375290513038635, + "learning_rate": 0.00019762917996800976, + "loss": 0.1897, + "step": 2972 + }, + { + "epoch": 0.06969180189525133, + "grad_norm": 0.3082475960254669, + "learning_rate": 0.00019762758523499216, + "loss": 0.065, + "step": 2973 + }, + { + "epoch": 0.06971524347005632, + "grad_norm": 0.6081576943397522, + "learning_rate": 0.0001976259899722445, + "loss": 0.1728, + "step": 2974 + }, + { + "epoch": 0.06973868504486132, + "grad_norm": 0.17558078467845917, + "learning_rate": 0.0001976243941797755, + "loss": 0.0362, + "step": 2975 + }, + { + "epoch": 0.06976212661966631, + "grad_norm": 0.5744110941886902, + "learning_rate": 0.0001976227978575938, + "loss": 0.1601, + "step": 2976 + }, + { + "epoch": 0.0697855681944713, + "grad_norm": 0.5946105122566223, + "learning_rate": 0.000197621201005708, + "loss": 0.2193, + "step": 2977 + }, + { + "epoch": 0.0698090097692763, + "grad_norm": 0.5541887283325195, + "learning_rate": 0.00019761960362412687, + "loss": 0.2433, + "step": 2978 + }, + { + "epoch": 0.0698324513440813, + "grad_norm": 0.6773604154586792, + "learning_rate": 0.000197618005712859, + "loss": 0.1654, + "step": 2979 + }, + { + "epoch": 0.06985589291888629, + "grad_norm": 0.4061916470527649, + "learning_rate": 0.0001976164072719131, + "loss": 0.1232, + "step": 2980 + }, + { + "epoch": 0.06987933449369128, + "grad_norm": 0.6316904425621033, + "learning_rate": 0.0001976148083012978, + "loss": 0.2054, + "step": 2981 + }, + { + "epoch": 0.06990277606849628, + "grad_norm": 0.5959118008613586, + "learning_rate": 0.00019761320880102185, + "loss": 0.1583, + "step": 2982 + }, + { + "epoch": 0.06992621764330127, + "grad_norm": 0.8021090030670166, + "learning_rate": 0.00019761160877109388, + "loss": 0.2777, + "step": 2983 + }, + { + "epoch": 0.06994965921810627, + "grad_norm": 0.730717658996582, + "learning_rate": 0.00019761000821152255, + "loss": 0.2312, + "step": 2984 + }, + { + "epoch": 0.06997310079291127, + "grad_norm": 0.7209658622741699, + "learning_rate": 0.0001976084071223166, + "loss": 0.1353, + "step": 2985 + }, + { + "epoch": 0.06999654236771627, + "grad_norm": 0.6175776124000549, + "learning_rate": 0.00019760680550348468, + "loss": 0.1608, + "step": 2986 + }, + { + "epoch": 0.07001998394252126, + "grad_norm": 0.5193164348602295, + "learning_rate": 0.00019760520335503548, + "loss": 0.0717, + "step": 2987 + }, + { + "epoch": 0.07004342551732626, + "grad_norm": 0.505129873752594, + "learning_rate": 0.00019760360067697767, + "loss": 0.6235, + "step": 2988 + }, + { + "epoch": 0.07006686709213125, + "grad_norm": 0.4627044200897217, + "learning_rate": 0.00019760199746932003, + "loss": 0.1119, + "step": 2989 + }, + { + "epoch": 0.07009030866693625, + "grad_norm": 0.6622577905654907, + "learning_rate": 0.00019760039373207118, + "loss": 0.1045, + "step": 2990 + }, + { + "epoch": 0.07011375024174124, + "grad_norm": 0.1620742678642273, + "learning_rate": 0.00019759878946523985, + "loss": 0.0358, + "step": 2991 + }, + { + "epoch": 0.07013719181654623, + "grad_norm": 0.8061487674713135, + "learning_rate": 0.00019759718466883473, + "loss": 0.1952, + "step": 2992 + }, + { + "epoch": 0.07016063339135123, + "grad_norm": 0.13489937782287598, + "learning_rate": 0.0001975955793428646, + "loss": 0.0331, + "step": 2993 + }, + { + "epoch": 0.07018407496615622, + "grad_norm": 1.092830777168274, + "learning_rate": 0.00019759397348733808, + "loss": 0.0795, + "step": 2994 + }, + { + "epoch": 0.07020751654096122, + "grad_norm": 0.1628320813179016, + "learning_rate": 0.00019759236710226389, + "loss": 0.0416, + "step": 2995 + }, + { + "epoch": 0.07023095811576621, + "grad_norm": 0.7330593466758728, + "learning_rate": 0.00019759076018765082, + "loss": 0.3359, + "step": 2996 + }, + { + "epoch": 0.0702543996905712, + "grad_norm": 0.19073647260665894, + "learning_rate": 0.00019758915274350748, + "loss": 0.0834, + "step": 2997 + }, + { + "epoch": 0.07027784126537621, + "grad_norm": 0.1880292147397995, + "learning_rate": 0.00019758754476984273, + "loss": 0.057, + "step": 2998 + }, + { + "epoch": 0.07030128284018121, + "grad_norm": 0.7399777770042419, + "learning_rate": 0.00019758593626666515, + "loss": 0.2292, + "step": 2999 + }, + { + "epoch": 0.0703247244149862, + "grad_norm": 0.6589105129241943, + "learning_rate": 0.00019758432723398353, + "loss": 0.1402, + "step": 3000 + }, + { + "epoch": 0.0703481659897912, + "grad_norm": 0.5588381886482239, + "learning_rate": 0.00019758271767180666, + "loss": 0.342, + "step": 3001 + }, + { + "epoch": 0.07037160756459619, + "grad_norm": 0.7927196621894836, + "learning_rate": 0.0001975811075801432, + "loss": 0.2068, + "step": 3002 + }, + { + "epoch": 0.07039504913940119, + "grad_norm": 0.6380198001861572, + "learning_rate": 0.00019757949695900192, + "loss": 0.1666, + "step": 3003 + }, + { + "epoch": 0.07041849071420618, + "grad_norm": 0.6712199449539185, + "learning_rate": 0.0001975778858083915, + "loss": 0.1954, + "step": 3004 + }, + { + "epoch": 0.07044193228901117, + "grad_norm": 0.5118662714958191, + "learning_rate": 0.00019757627412832077, + "loss": 0.2321, + "step": 3005 + }, + { + "epoch": 0.07046537386381617, + "grad_norm": 0.8980170488357544, + "learning_rate": 0.00019757466191879842, + "loss": 0.2072, + "step": 3006 + }, + { + "epoch": 0.07048881543862116, + "grad_norm": 1.0479631423950195, + "learning_rate": 0.00019757304917983323, + "loss": 0.2302, + "step": 3007 + }, + { + "epoch": 0.07051225701342616, + "grad_norm": 0.42021214962005615, + "learning_rate": 0.0001975714359114339, + "loss": 0.1343, + "step": 3008 + }, + { + "epoch": 0.07053569858823115, + "grad_norm": 0.5695638060569763, + "learning_rate": 0.00019756982211360924, + "loss": 0.1405, + "step": 3009 + }, + { + "epoch": 0.07055914016303615, + "grad_norm": 0.15578952431678772, + "learning_rate": 0.00019756820778636796, + "loss": 0.0574, + "step": 3010 + }, + { + "epoch": 0.07058258173784115, + "grad_norm": 0.22329343855381012, + "learning_rate": 0.00019756659292971887, + "loss": 0.0432, + "step": 3011 + }, + { + "epoch": 0.07060602331264615, + "grad_norm": 0.4478393793106079, + "learning_rate": 0.00019756497754367067, + "loss": 0.0248, + "step": 3012 + }, + { + "epoch": 0.07062946488745114, + "grad_norm": 0.40987908840179443, + "learning_rate": 0.00019756336162823218, + "loss": 0.1087, + "step": 3013 + }, + { + "epoch": 0.07065290646225614, + "grad_norm": 0.2393629103899002, + "learning_rate": 0.00019756174518341214, + "loss": 0.0759, + "step": 3014 + }, + { + "epoch": 0.07067634803706113, + "grad_norm": 0.49267566204071045, + "learning_rate": 0.00019756012820921934, + "loss": 0.7268, + "step": 3015 + }, + { + "epoch": 0.07069978961186613, + "grad_norm": 0.38486894965171814, + "learning_rate": 0.00019755851070566251, + "loss": 0.0758, + "step": 3016 + }, + { + "epoch": 0.07072323118667112, + "grad_norm": 0.31363445520401, + "learning_rate": 0.0001975568926727505, + "loss": 0.0987, + "step": 3017 + }, + { + "epoch": 0.07074667276147611, + "grad_norm": 0.821697473526001, + "learning_rate": 0.000197555274110492, + "loss": 0.7801, + "step": 3018 + }, + { + "epoch": 0.07077011433628111, + "grad_norm": 0.30461421608924866, + "learning_rate": 0.0001975536550188959, + "loss": 0.0778, + "step": 3019 + }, + { + "epoch": 0.0707935559110861, + "grad_norm": 0.4205400049686432, + "learning_rate": 0.00019755203539797086, + "loss": 0.0694, + "step": 3020 + }, + { + "epoch": 0.0708169974858911, + "grad_norm": 0.4571243226528168, + "learning_rate": 0.00019755041524772575, + "loss": 0.1749, + "step": 3021 + }, + { + "epoch": 0.07084043906069609, + "grad_norm": 0.6060929298400879, + "learning_rate": 0.00019754879456816937, + "loss": 0.2261, + "step": 3022 + }, + { + "epoch": 0.07086388063550109, + "grad_norm": 0.7765220403671265, + "learning_rate": 0.00019754717335931048, + "loss": 0.1747, + "step": 3023 + }, + { + "epoch": 0.0708873222103061, + "grad_norm": 0.9781676530838013, + "learning_rate": 0.00019754555162115786, + "loss": 0.2147, + "step": 3024 + }, + { + "epoch": 0.07091076378511109, + "grad_norm": 0.6231436729431152, + "learning_rate": 0.00019754392935372037, + "loss": 0.1348, + "step": 3025 + }, + { + "epoch": 0.07093420535991608, + "grad_norm": 0.5863490104675293, + "learning_rate": 0.00019754230655700676, + "loss": 0.9092, + "step": 3026 + }, + { + "epoch": 0.07095764693472108, + "grad_norm": 0.6382483839988708, + "learning_rate": 0.00019754068323102586, + "loss": 0.1634, + "step": 3027 + }, + { + "epoch": 0.07098108850952607, + "grad_norm": 0.6687260866165161, + "learning_rate": 0.00019753905937578645, + "loss": 0.2088, + "step": 3028 + }, + { + "epoch": 0.07100453008433107, + "grad_norm": 0.9934439063072205, + "learning_rate": 0.00019753743499129737, + "loss": 0.1383, + "step": 3029 + }, + { + "epoch": 0.07102797165913606, + "grad_norm": 0.7598995566368103, + "learning_rate": 0.00019753581007756743, + "loss": 0.1018, + "step": 3030 + }, + { + "epoch": 0.07105141323394105, + "grad_norm": 0.47262826561927795, + "learning_rate": 0.00019753418463460545, + "loss": 0.6334, + "step": 3031 + }, + { + "epoch": 0.07107485480874605, + "grad_norm": 0.3431902229785919, + "learning_rate": 0.00019753255866242022, + "loss": 0.0916, + "step": 3032 + }, + { + "epoch": 0.07109829638355104, + "grad_norm": 0.6518743634223938, + "learning_rate": 0.0001975309321610206, + "loss": 0.1143, + "step": 3033 + }, + { + "epoch": 0.07112173795835604, + "grad_norm": 0.6473389863967896, + "learning_rate": 0.0001975293051304154, + "loss": 0.2135, + "step": 3034 + }, + { + "epoch": 0.07114517953316103, + "grad_norm": 0.17880530655384064, + "learning_rate": 0.00019752767757061346, + "loss": 0.0542, + "step": 3035 + }, + { + "epoch": 0.07116862110796603, + "grad_norm": 0.582767903804779, + "learning_rate": 0.00019752604948162356, + "loss": 0.1466, + "step": 3036 + }, + { + "epoch": 0.07119206268277103, + "grad_norm": 0.4044455587863922, + "learning_rate": 0.00019752442086345463, + "loss": 0.1049, + "step": 3037 + }, + { + "epoch": 0.07121550425757603, + "grad_norm": 0.5283302068710327, + "learning_rate": 0.00019752279171611538, + "loss": 0.164, + "step": 3038 + }, + { + "epoch": 0.07123894583238102, + "grad_norm": 0.2310810387134552, + "learning_rate": 0.00019752116203961478, + "loss": 0.0526, + "step": 3039 + }, + { + "epoch": 0.07126238740718602, + "grad_norm": 0.2347443699836731, + "learning_rate": 0.00019751953183396159, + "loss": 0.0295, + "step": 3040 + }, + { + "epoch": 0.07128582898199101, + "grad_norm": 0.28831467032432556, + "learning_rate": 0.00019751790109916466, + "loss": 0.1073, + "step": 3041 + }, + { + "epoch": 0.071309270556796, + "grad_norm": 0.7304638624191284, + "learning_rate": 0.0001975162698352329, + "loss": 0.1641, + "step": 3042 + }, + { + "epoch": 0.071332712131601, + "grad_norm": 1.1474775075912476, + "learning_rate": 0.00019751463804217506, + "loss": 0.3296, + "step": 3043 + }, + { + "epoch": 0.071356153706406, + "grad_norm": 0.39663228392601013, + "learning_rate": 0.0001975130057200001, + "loss": 0.0627, + "step": 3044 + }, + { + "epoch": 0.07137959528121099, + "grad_norm": 0.7680560350418091, + "learning_rate": 0.0001975113728687168, + "loss": 0.5155, + "step": 3045 + }, + { + "epoch": 0.07140303685601598, + "grad_norm": 0.36431851983070374, + "learning_rate": 0.00019750973948833407, + "loss": 0.1051, + "step": 3046 + }, + { + "epoch": 0.07142647843082098, + "grad_norm": 0.7237342596054077, + "learning_rate": 0.00019750810557886075, + "loss": 0.1307, + "step": 3047 + }, + { + "epoch": 0.07144992000562597, + "grad_norm": 0.6581172943115234, + "learning_rate": 0.0001975064711403057, + "loss": 0.2109, + "step": 3048 + }, + { + "epoch": 0.07147336158043098, + "grad_norm": 1.1607940196990967, + "learning_rate": 0.00019750483617267782, + "loss": 0.4684, + "step": 3049 + }, + { + "epoch": 0.07149680315523597, + "grad_norm": 0.6742085814476013, + "learning_rate": 0.00019750320067598592, + "loss": 0.7483, + "step": 3050 + }, + { + "epoch": 0.07152024473004097, + "grad_norm": 0.7564075589179993, + "learning_rate": 0.0001975015646502389, + "loss": 0.24, + "step": 3051 + }, + { + "epoch": 0.07154368630484596, + "grad_norm": 0.11819720268249512, + "learning_rate": 0.0001974999280954457, + "loss": 0.0304, + "step": 3052 + }, + { + "epoch": 0.07156712787965096, + "grad_norm": 0.6576065421104431, + "learning_rate": 0.0001974982910116151, + "loss": 0.2473, + "step": 3053 + }, + { + "epoch": 0.07159056945445595, + "grad_norm": 0.6757521629333496, + "learning_rate": 0.00019749665339875607, + "loss": 0.2247, + "step": 3054 + }, + { + "epoch": 0.07161401102926095, + "grad_norm": 0.6797861456871033, + "learning_rate": 0.00019749501525687745, + "loss": 0.1424, + "step": 3055 + }, + { + "epoch": 0.07163745260406594, + "grad_norm": 0.5489090085029602, + "learning_rate": 0.00019749337658598814, + "loss": 0.2103, + "step": 3056 + }, + { + "epoch": 0.07166089417887093, + "grad_norm": 0.5486713647842407, + "learning_rate": 0.000197491737386097, + "loss": 0.1099, + "step": 3057 + }, + { + "epoch": 0.07168433575367593, + "grad_norm": 0.7214705348014832, + "learning_rate": 0.00019749009765721297, + "loss": 0.1654, + "step": 3058 + }, + { + "epoch": 0.07170777732848092, + "grad_norm": 0.7033182978630066, + "learning_rate": 0.0001974884573993449, + "loss": 0.2765, + "step": 3059 + }, + { + "epoch": 0.07173121890328592, + "grad_norm": 0.683692216873169, + "learning_rate": 0.00019748681661250177, + "loss": 0.1827, + "step": 3060 + }, + { + "epoch": 0.07175466047809091, + "grad_norm": 0.5752551555633545, + "learning_rate": 0.00019748517529669243, + "loss": 0.118, + "step": 3061 + }, + { + "epoch": 0.07177810205289592, + "grad_norm": 0.26842573285102844, + "learning_rate": 0.00019748353345192573, + "loss": 0.0774, + "step": 3062 + }, + { + "epoch": 0.07180154362770091, + "grad_norm": 0.813735842704773, + "learning_rate": 0.00019748189107821068, + "loss": 0.3135, + "step": 3063 + }, + { + "epoch": 0.07182498520250591, + "grad_norm": 0.2506135106086731, + "learning_rate": 0.00019748024817555618, + "loss": 0.0641, + "step": 3064 + }, + { + "epoch": 0.0718484267773109, + "grad_norm": 0.2050292044878006, + "learning_rate": 0.00019747860474397103, + "loss": 0.0633, + "step": 3065 + }, + { + "epoch": 0.0718718683521159, + "grad_norm": 0.15822413563728333, + "learning_rate": 0.0001974769607834643, + "loss": 0.0443, + "step": 3066 + }, + { + "epoch": 0.07189530992692089, + "grad_norm": 0.6794048547744751, + "learning_rate": 0.00019747531629404482, + "loss": 0.1601, + "step": 3067 + }, + { + "epoch": 0.07191875150172589, + "grad_norm": 0.6546835899353027, + "learning_rate": 0.00019747367127572152, + "loss": 0.191, + "step": 3068 + }, + { + "epoch": 0.07194219307653088, + "grad_norm": 0.42777219414711, + "learning_rate": 0.00019747202572850334, + "loss": 0.0829, + "step": 3069 + }, + { + "epoch": 0.07196563465133587, + "grad_norm": 0.7703954577445984, + "learning_rate": 0.00019747037965239922, + "loss": 0.1987, + "step": 3070 + }, + { + "epoch": 0.07198907622614087, + "grad_norm": 0.8675500750541687, + "learning_rate": 0.00019746873304741806, + "loss": 0.2901, + "step": 3071 + }, + { + "epoch": 0.07201251780094586, + "grad_norm": 0.4944847524166107, + "learning_rate": 0.00019746708591356882, + "loss": 0.131, + "step": 3072 + }, + { + "epoch": 0.07203595937575086, + "grad_norm": 0.45164427161216736, + "learning_rate": 0.00019746543825086045, + "loss": 0.0781, + "step": 3073 + }, + { + "epoch": 0.07205940095055585, + "grad_norm": 0.23765729367733002, + "learning_rate": 0.00019746379005930185, + "loss": 0.052, + "step": 3074 + }, + { + "epoch": 0.07208284252536086, + "grad_norm": 0.16807900369167328, + "learning_rate": 0.000197462141338902, + "loss": 0.0481, + "step": 3075 + }, + { + "epoch": 0.07210628410016585, + "grad_norm": 0.677668571472168, + "learning_rate": 0.00019746049208966982, + "loss": 0.6304, + "step": 3076 + }, + { + "epoch": 0.07212972567497085, + "grad_norm": 0.399628221988678, + "learning_rate": 0.00019745884231161428, + "loss": 0.097, + "step": 3077 + }, + { + "epoch": 0.07215316724977584, + "grad_norm": 0.27469268441200256, + "learning_rate": 0.0001974571920047443, + "loss": 0.0684, + "step": 3078 + }, + { + "epoch": 0.07217660882458084, + "grad_norm": 0.19631905853748322, + "learning_rate": 0.0001974555411690689, + "loss": 0.0402, + "step": 3079 + }, + { + "epoch": 0.07220005039938583, + "grad_norm": 1.0630875825881958, + "learning_rate": 0.00019745388980459694, + "loss": 0.3654, + "step": 3080 + }, + { + "epoch": 0.07222349197419083, + "grad_norm": 0.5556409955024719, + "learning_rate": 0.00019745223791133745, + "loss": 0.1481, + "step": 3081 + }, + { + "epoch": 0.07224693354899582, + "grad_norm": 0.5698478817939758, + "learning_rate": 0.0001974505854892994, + "loss": 0.11, + "step": 3082 + }, + { + "epoch": 0.07227037512380081, + "grad_norm": 0.324849396944046, + "learning_rate": 0.0001974489325384917, + "loss": 0.061, + "step": 3083 + }, + { + "epoch": 0.07229381669860581, + "grad_norm": 0.2234215885400772, + "learning_rate": 0.00019744727905892338, + "loss": 0.0768, + "step": 3084 + }, + { + "epoch": 0.0723172582734108, + "grad_norm": 0.5088491439819336, + "learning_rate": 0.00019744562505060338, + "loss": 0.123, + "step": 3085 + }, + { + "epoch": 0.0723406998482158, + "grad_norm": 0.6305775046348572, + "learning_rate": 0.00019744397051354065, + "loss": 0.1156, + "step": 3086 + }, + { + "epoch": 0.07236414142302079, + "grad_norm": 0.6764863133430481, + "learning_rate": 0.00019744231544774422, + "loss": 0.7931, + "step": 3087 + }, + { + "epoch": 0.0723875829978258, + "grad_norm": 0.5178343057632446, + "learning_rate": 0.00019744065985322307, + "loss": 0.1376, + "step": 3088 + }, + { + "epoch": 0.0724110245726308, + "grad_norm": 0.15926413238048553, + "learning_rate": 0.00019743900372998612, + "loss": 0.0285, + "step": 3089 + }, + { + "epoch": 0.07243446614743579, + "grad_norm": 0.8300385475158691, + "learning_rate": 0.00019743734707804243, + "loss": 0.2406, + "step": 3090 + }, + { + "epoch": 0.07245790772224078, + "grad_norm": 0.2712288200855255, + "learning_rate": 0.00019743568989740094, + "loss": 0.0697, + "step": 3091 + }, + { + "epoch": 0.07248134929704578, + "grad_norm": 0.6900508999824524, + "learning_rate": 0.00019743403218807064, + "loss": 0.1472, + "step": 3092 + }, + { + "epoch": 0.07250479087185077, + "grad_norm": 0.6035946011543274, + "learning_rate": 0.00019743237395006055, + "loss": 0.1135, + "step": 3093 + }, + { + "epoch": 0.07252823244665577, + "grad_norm": 0.6090536713600159, + "learning_rate": 0.00019743071518337968, + "loss": 0.8231, + "step": 3094 + }, + { + "epoch": 0.07255167402146076, + "grad_norm": 0.3540598750114441, + "learning_rate": 0.000197429055888037, + "loss": 0.1396, + "step": 3095 + }, + { + "epoch": 0.07257511559626575, + "grad_norm": 0.7455745935440063, + "learning_rate": 0.0001974273960640415, + "loss": 0.1802, + "step": 3096 + }, + { + "epoch": 0.07259855717107075, + "grad_norm": 0.1775999516248703, + "learning_rate": 0.00019742573571140222, + "loss": 0.0265, + "step": 3097 + }, + { + "epoch": 0.07262199874587574, + "grad_norm": 0.66279536485672, + "learning_rate": 0.00019742407483012818, + "loss": 0.1869, + "step": 3098 + }, + { + "epoch": 0.07264544032068074, + "grad_norm": 0.36322563886642456, + "learning_rate": 0.00019742241342022836, + "loss": 0.1244, + "step": 3099 + }, + { + "epoch": 0.07266888189548573, + "grad_norm": 0.47039794921875, + "learning_rate": 0.00019742075148171178, + "loss": 0.0818, + "step": 3100 + }, + { + "epoch": 0.07269232347029074, + "grad_norm": 0.19588926434516907, + "learning_rate": 0.00019741908901458745, + "loss": 0.0636, + "step": 3101 + }, + { + "epoch": 0.07271576504509573, + "grad_norm": 0.48259595036506653, + "learning_rate": 0.00019741742601886442, + "loss": 0.087, + "step": 3102 + }, + { + "epoch": 0.07273920661990073, + "grad_norm": 0.5523844957351685, + "learning_rate": 0.0001974157624945517, + "loss": 0.1622, + "step": 3103 + }, + { + "epoch": 0.07276264819470572, + "grad_norm": 0.2015560269355774, + "learning_rate": 0.0001974140984416583, + "loss": 0.0475, + "step": 3104 + }, + { + "epoch": 0.07278608976951072, + "grad_norm": 0.1939472109079361, + "learning_rate": 0.00019741243386019327, + "loss": 0.049, + "step": 3105 + }, + { + "epoch": 0.07280953134431571, + "grad_norm": 0.19091713428497314, + "learning_rate": 0.00019741076875016565, + "loss": 0.0507, + "step": 3106 + }, + { + "epoch": 0.0728329729191207, + "grad_norm": 0.45684030652046204, + "learning_rate": 0.00019740910311158444, + "loss": 0.1129, + "step": 3107 + }, + { + "epoch": 0.0728564144939257, + "grad_norm": 0.29042795300483704, + "learning_rate": 0.0001974074369444587, + "loss": 0.0479, + "step": 3108 + }, + { + "epoch": 0.0728798560687307, + "grad_norm": 0.793425977230072, + "learning_rate": 0.00019740577024879747, + "loss": 0.1617, + "step": 3109 + }, + { + "epoch": 0.07290329764353569, + "grad_norm": 0.6870112419128418, + "learning_rate": 0.00019740410302460982, + "loss": 0.1706, + "step": 3110 + }, + { + "epoch": 0.07292673921834068, + "grad_norm": 1.177152156829834, + "learning_rate": 0.00019740243527190473, + "loss": 0.2644, + "step": 3111 + }, + { + "epoch": 0.07295018079314568, + "grad_norm": 0.7179206609725952, + "learning_rate": 0.0001974007669906913, + "loss": 0.2319, + "step": 3112 + }, + { + "epoch": 0.07297362236795069, + "grad_norm": 0.5777270197868347, + "learning_rate": 0.00019739909818097858, + "loss": 0.1377, + "step": 3113 + }, + { + "epoch": 0.07299706394275568, + "grad_norm": 0.44075897336006165, + "learning_rate": 0.00019739742884277562, + "loss": 0.1684, + "step": 3114 + }, + { + "epoch": 0.07302050551756067, + "grad_norm": 1.1906245946884155, + "learning_rate": 0.00019739575897609145, + "loss": 0.2202, + "step": 3115 + }, + { + "epoch": 0.07304394709236567, + "grad_norm": 0.8072792291641235, + "learning_rate": 0.00019739408858093516, + "loss": 0.1571, + "step": 3116 + }, + { + "epoch": 0.07306738866717066, + "grad_norm": 0.6312717199325562, + "learning_rate": 0.0001973924176573158, + "loss": 0.1578, + "step": 3117 + }, + { + "epoch": 0.07309083024197566, + "grad_norm": 0.46754154562950134, + "learning_rate": 0.00019739074620524246, + "loss": 0.4808, + "step": 3118 + }, + { + "epoch": 0.07311427181678065, + "grad_norm": 0.12462183833122253, + "learning_rate": 0.00019738907422472418, + "loss": 0.0293, + "step": 3119 + }, + { + "epoch": 0.07313771339158565, + "grad_norm": 0.27480924129486084, + "learning_rate": 0.00019738740171577004, + "loss": 0.0561, + "step": 3120 + }, + { + "epoch": 0.07316115496639064, + "grad_norm": 0.6890031695365906, + "learning_rate": 0.00019738572867838912, + "loss": 0.2682, + "step": 3121 + }, + { + "epoch": 0.07318459654119563, + "grad_norm": 0.937913715839386, + "learning_rate": 0.0001973840551125905, + "loss": 0.2493, + "step": 3122 + }, + { + "epoch": 0.07320803811600063, + "grad_norm": 0.5166339874267578, + "learning_rate": 0.00019738238101838328, + "loss": 0.2489, + "step": 3123 + }, + { + "epoch": 0.07323147969080562, + "grad_norm": 0.6622734069824219, + "learning_rate": 0.0001973807063957765, + "loss": 0.2011, + "step": 3124 + }, + { + "epoch": 0.07325492126561062, + "grad_norm": 0.7259514927864075, + "learning_rate": 0.00019737903124477927, + "loss": 0.1355, + "step": 3125 + }, + { + "epoch": 0.07327836284041563, + "grad_norm": 0.6751295328140259, + "learning_rate": 0.00019737735556540068, + "loss": 0.1692, + "step": 3126 + }, + { + "epoch": 0.07330180441522062, + "grad_norm": 0.23125210404396057, + "learning_rate": 0.00019737567935764985, + "loss": 0.0458, + "step": 3127 + }, + { + "epoch": 0.07332524599002561, + "grad_norm": 0.6284736394882202, + "learning_rate": 0.00019737400262153583, + "loss": 0.1556, + "step": 3128 + }, + { + "epoch": 0.07334868756483061, + "grad_norm": 0.3690071105957031, + "learning_rate": 0.00019737232535706772, + "loss": 0.6655, + "step": 3129 + }, + { + "epoch": 0.0733721291396356, + "grad_norm": 0.22172878682613373, + "learning_rate": 0.00019737064756425462, + "loss": 0.0391, + "step": 3130 + }, + { + "epoch": 0.0733955707144406, + "grad_norm": 0.7061000466346741, + "learning_rate": 0.00019736896924310568, + "loss": 0.2248, + "step": 3131 + }, + { + "epoch": 0.07341901228924559, + "grad_norm": 0.39302006363868713, + "learning_rate": 0.00019736729039362998, + "loss": 0.1382, + "step": 3132 + }, + { + "epoch": 0.07344245386405059, + "grad_norm": 0.765044629573822, + "learning_rate": 0.0001973656110158366, + "loss": 0.1463, + "step": 3133 + }, + { + "epoch": 0.07346589543885558, + "grad_norm": 0.4787522256374359, + "learning_rate": 0.00019736393110973468, + "loss": 0.0832, + "step": 3134 + }, + { + "epoch": 0.07348933701366057, + "grad_norm": 0.6342676877975464, + "learning_rate": 0.00019736225067533333, + "loss": 0.1392, + "step": 3135 + }, + { + "epoch": 0.07351277858846557, + "grad_norm": 0.6664438247680664, + "learning_rate": 0.00019736056971264167, + "loss": 0.1953, + "step": 3136 + }, + { + "epoch": 0.07353622016327056, + "grad_norm": 0.5551084876060486, + "learning_rate": 0.00019735888822166887, + "loss": 0.1063, + "step": 3137 + }, + { + "epoch": 0.07355966173807556, + "grad_norm": 0.7548152208328247, + "learning_rate": 0.00019735720620242395, + "loss": 0.1186, + "step": 3138 + }, + { + "epoch": 0.07358310331288057, + "grad_norm": 0.6441261172294617, + "learning_rate": 0.00019735552365491608, + "loss": 0.2711, + "step": 3139 + }, + { + "epoch": 0.07360654488768556, + "grad_norm": 0.429417222738266, + "learning_rate": 0.00019735384057915445, + "loss": 0.1191, + "step": 3140 + }, + { + "epoch": 0.07362998646249055, + "grad_norm": 0.7868626117706299, + "learning_rate": 0.0001973521569751481, + "loss": 0.1767, + "step": 3141 + }, + { + "epoch": 0.07365342803729555, + "grad_norm": 1.0374584197998047, + "learning_rate": 0.00019735047284290624, + "loss": 0.2618, + "step": 3142 + }, + { + "epoch": 0.07367686961210054, + "grad_norm": 0.6177417039871216, + "learning_rate": 0.00019734878818243799, + "loss": 0.2475, + "step": 3143 + }, + { + "epoch": 0.07370031118690554, + "grad_norm": 0.33270353078842163, + "learning_rate": 0.00019734710299375242, + "loss": 0.0633, + "step": 3144 + }, + { + "epoch": 0.07372375276171053, + "grad_norm": 0.6571106314659119, + "learning_rate": 0.00019734541727685877, + "loss": 0.8632, + "step": 3145 + }, + { + "epoch": 0.07374719433651553, + "grad_norm": 0.41483956575393677, + "learning_rate": 0.00019734373103176615, + "loss": 0.088, + "step": 3146 + }, + { + "epoch": 0.07377063591132052, + "grad_norm": 0.5709042549133301, + "learning_rate": 0.00019734204425848371, + "loss": 0.1832, + "step": 3147 + }, + { + "epoch": 0.07379407748612551, + "grad_norm": 0.47740674018859863, + "learning_rate": 0.0001973403569570206, + "loss": 0.0817, + "step": 3148 + }, + { + "epoch": 0.07381751906093051, + "grad_norm": 0.6653018593788147, + "learning_rate": 0.00019733866912738596, + "loss": 0.1151, + "step": 3149 + }, + { + "epoch": 0.0738409606357355, + "grad_norm": 0.20368455350399017, + "learning_rate": 0.00019733698076958897, + "loss": 0.0906, + "step": 3150 + }, + { + "epoch": 0.0738644022105405, + "grad_norm": 0.4542419910430908, + "learning_rate": 0.0001973352918836388, + "loss": 0.1021, + "step": 3151 + }, + { + "epoch": 0.0738878437853455, + "grad_norm": 0.35212790966033936, + "learning_rate": 0.0001973336024695446, + "loss": 0.0678, + "step": 3152 + }, + { + "epoch": 0.0739112853601505, + "grad_norm": 0.3801686465740204, + "learning_rate": 0.0001973319125273155, + "loss": 0.1559, + "step": 3153 + }, + { + "epoch": 0.0739347269349555, + "grad_norm": 0.1748773604631424, + "learning_rate": 0.00019733022205696075, + "loss": 0.0809, + "step": 3154 + }, + { + "epoch": 0.07395816850976049, + "grad_norm": 0.5437548756599426, + "learning_rate": 0.00019732853105848945, + "loss": 0.1052, + "step": 3155 + }, + { + "epoch": 0.07398161008456548, + "grad_norm": 0.2554425895214081, + "learning_rate": 0.00019732683953191082, + "loss": 0.0741, + "step": 3156 + }, + { + "epoch": 0.07400505165937048, + "grad_norm": 0.5250517129898071, + "learning_rate": 0.000197325147477234, + "loss": 0.1748, + "step": 3157 + }, + { + "epoch": 0.07402849323417547, + "grad_norm": 0.7426334023475647, + "learning_rate": 0.00019732345489446824, + "loss": 0.1799, + "step": 3158 + }, + { + "epoch": 0.07405193480898047, + "grad_norm": 0.6833359599113464, + "learning_rate": 0.00019732176178362262, + "loss": 0.2019, + "step": 3159 + }, + { + "epoch": 0.07407537638378546, + "grad_norm": 0.1502295285463333, + "learning_rate": 0.00019732006814470642, + "loss": 0.0577, + "step": 3160 + }, + { + "epoch": 0.07409881795859045, + "grad_norm": 0.8678262829780579, + "learning_rate": 0.00019731837397772881, + "loss": 0.5988, + "step": 3161 + }, + { + "epoch": 0.07412225953339545, + "grad_norm": 0.5297439694404602, + "learning_rate": 0.00019731667928269892, + "loss": 0.1331, + "step": 3162 + }, + { + "epoch": 0.07414570110820044, + "grad_norm": 0.44021767377853394, + "learning_rate": 0.000197314984059626, + "loss": 0.1108, + "step": 3163 + }, + { + "epoch": 0.07416914268300545, + "grad_norm": 0.5866394639015198, + "learning_rate": 0.00019731328830851927, + "loss": 0.6666, + "step": 3164 + }, + { + "epoch": 0.07419258425781045, + "grad_norm": 0.48086780309677124, + "learning_rate": 0.0001973115920293879, + "loss": 0.1448, + "step": 3165 + }, + { + "epoch": 0.07421602583261544, + "grad_norm": 0.5798571705818176, + "learning_rate": 0.00019730989522224107, + "loss": 0.1626, + "step": 3166 + }, + { + "epoch": 0.07423946740742043, + "grad_norm": 0.3254583775997162, + "learning_rate": 0.000197308197887088, + "loss": 0.0687, + "step": 3167 + }, + { + "epoch": 0.07426290898222543, + "grad_norm": 0.16586749255657196, + "learning_rate": 0.00019730650002393796, + "loss": 0.0401, + "step": 3168 + }, + { + "epoch": 0.07428635055703042, + "grad_norm": 0.5056729316711426, + "learning_rate": 0.00019730480163280008, + "loss": 0.1281, + "step": 3169 + }, + { + "epoch": 0.07430979213183542, + "grad_norm": 0.23724666237831116, + "learning_rate": 0.0001973031027136836, + "loss": 0.083, + "step": 3170 + }, + { + "epoch": 0.07433323370664041, + "grad_norm": 0.6201127171516418, + "learning_rate": 0.0001973014032665978, + "loss": 0.236, + "step": 3171 + }, + { + "epoch": 0.0743566752814454, + "grad_norm": 0.3784458041191101, + "learning_rate": 0.00019729970329155182, + "loss": 0.1015, + "step": 3172 + }, + { + "epoch": 0.0743801168562504, + "grad_norm": 0.5908719301223755, + "learning_rate": 0.00019729800278855492, + "loss": 0.881, + "step": 3173 + }, + { + "epoch": 0.0744035584310554, + "grad_norm": 0.4960664212703705, + "learning_rate": 0.0001972963017576163, + "loss": 0.17, + "step": 3174 + }, + { + "epoch": 0.07442700000586039, + "grad_norm": 0.7892377972602844, + "learning_rate": 0.00019729460019874524, + "loss": 0.2017, + "step": 3175 + }, + { + "epoch": 0.07445044158066538, + "grad_norm": 0.2139170914888382, + "learning_rate": 0.00019729289811195092, + "loss": 0.0475, + "step": 3176 + }, + { + "epoch": 0.07447388315547039, + "grad_norm": 0.43913963437080383, + "learning_rate": 0.0001972911954972426, + "loss": 0.1144, + "step": 3177 + }, + { + "epoch": 0.07449732473027539, + "grad_norm": 0.533353865146637, + "learning_rate": 0.00019728949235462952, + "loss": 0.1287, + "step": 3178 + }, + { + "epoch": 0.07452076630508038, + "grad_norm": 0.7270887494087219, + "learning_rate": 0.00019728778868412095, + "loss": 0.1913, + "step": 3179 + }, + { + "epoch": 0.07454420787988537, + "grad_norm": 0.3738802671432495, + "learning_rate": 0.00019728608448572605, + "loss": 0.0747, + "step": 3180 + }, + { + "epoch": 0.07456764945469037, + "grad_norm": 0.49824026226997375, + "learning_rate": 0.00019728437975945418, + "loss": 0.2095, + "step": 3181 + }, + { + "epoch": 0.07459109102949536, + "grad_norm": 0.7779057025909424, + "learning_rate": 0.0001972826745053145, + "loss": 0.2993, + "step": 3182 + }, + { + "epoch": 0.07461453260430036, + "grad_norm": 0.38560783863067627, + "learning_rate": 0.00019728096872331628, + "loss": 0.0777, + "step": 3183 + }, + { + "epoch": 0.07463797417910535, + "grad_norm": 0.5115217566490173, + "learning_rate": 0.0001972792624134688, + "loss": 0.1534, + "step": 3184 + }, + { + "epoch": 0.07466141575391035, + "grad_norm": 0.3155441880226135, + "learning_rate": 0.0001972775555757813, + "loss": 0.0861, + "step": 3185 + }, + { + "epoch": 0.07468485732871534, + "grad_norm": 0.5373808741569519, + "learning_rate": 0.00019727584821026305, + "loss": 0.1891, + "step": 3186 + }, + { + "epoch": 0.07470829890352033, + "grad_norm": 0.5356204509735107, + "learning_rate": 0.00019727414031692333, + "loss": 0.1607, + "step": 3187 + }, + { + "epoch": 0.07473174047832533, + "grad_norm": 0.36898213624954224, + "learning_rate": 0.00019727243189577136, + "loss": 0.1054, + "step": 3188 + }, + { + "epoch": 0.07475518205313032, + "grad_norm": 0.6019520163536072, + "learning_rate": 0.00019727072294681647, + "loss": 0.2364, + "step": 3189 + }, + { + "epoch": 0.07477862362793533, + "grad_norm": 0.43046626448631287, + "learning_rate": 0.0001972690134700679, + "loss": 0.13, + "step": 3190 + }, + { + "epoch": 0.07480206520274033, + "grad_norm": 0.16783104836940765, + "learning_rate": 0.0001972673034655349, + "loss": 0.0494, + "step": 3191 + }, + { + "epoch": 0.07482550677754532, + "grad_norm": 0.37699344754219055, + "learning_rate": 0.00019726559293322678, + "loss": 0.0581, + "step": 3192 + }, + { + "epoch": 0.07484894835235031, + "grad_norm": 0.6425116062164307, + "learning_rate": 0.00019726388187315283, + "loss": 0.2746, + "step": 3193 + }, + { + "epoch": 0.07487238992715531, + "grad_norm": 0.6611656546592712, + "learning_rate": 0.00019726217028532235, + "loss": 0.1881, + "step": 3194 + }, + { + "epoch": 0.0748958315019603, + "grad_norm": 0.49243783950805664, + "learning_rate": 0.00019726045816974454, + "loss": 0.1414, + "step": 3195 + }, + { + "epoch": 0.0749192730767653, + "grad_norm": 0.5551742911338806, + "learning_rate": 0.0001972587455264288, + "loss": 0.7398, + "step": 3196 + }, + { + "epoch": 0.07494271465157029, + "grad_norm": 0.2164343297481537, + "learning_rate": 0.00019725703235538435, + "loss": 0.0486, + "step": 3197 + }, + { + "epoch": 0.07496615622637529, + "grad_norm": 0.9302752017974854, + "learning_rate": 0.00019725531865662053, + "loss": 0.2127, + "step": 3198 + }, + { + "epoch": 0.07498959780118028, + "grad_norm": 0.5271553993225098, + "learning_rate": 0.00019725360443014658, + "loss": 0.1741, + "step": 3199 + }, + { + "epoch": 0.07501303937598527, + "grad_norm": 0.5474498867988586, + "learning_rate": 0.00019725188967597185, + "loss": 0.1291, + "step": 3200 + }, + { + "epoch": 0.07503648095079027, + "grad_norm": 0.32210588455200195, + "learning_rate": 0.00019725017439410565, + "loss": 0.084, + "step": 3201 + }, + { + "epoch": 0.07505992252559526, + "grad_norm": 0.4331153631210327, + "learning_rate": 0.00019724845858455726, + "loss": 0.1055, + "step": 3202 + }, + { + "epoch": 0.07508336410040027, + "grad_norm": 0.29366937279701233, + "learning_rate": 0.00019724674224733598, + "loss": 0.0856, + "step": 3203 + }, + { + "epoch": 0.07510680567520527, + "grad_norm": 0.4389234483242035, + "learning_rate": 0.00019724502538245116, + "loss": 0.0573, + "step": 3204 + }, + { + "epoch": 0.07513024725001026, + "grad_norm": 0.45061251521110535, + "learning_rate": 0.00019724330798991208, + "loss": 0.1657, + "step": 3205 + }, + { + "epoch": 0.07515368882481525, + "grad_norm": 0.7059164047241211, + "learning_rate": 0.00019724159006972812, + "loss": 0.1865, + "step": 3206 + }, + { + "epoch": 0.07517713039962025, + "grad_norm": 0.8623042106628418, + "learning_rate": 0.0001972398716219085, + "loss": 0.3502, + "step": 3207 + }, + { + "epoch": 0.07520057197442524, + "grad_norm": 0.5168110132217407, + "learning_rate": 0.00019723815264646266, + "loss": 0.6508, + "step": 3208 + }, + { + "epoch": 0.07522401354923024, + "grad_norm": 0.25933629274368286, + "learning_rate": 0.00019723643314339983, + "loss": 0.0622, + "step": 3209 + }, + { + "epoch": 0.07524745512403523, + "grad_norm": 0.40620267391204834, + "learning_rate": 0.00019723471311272942, + "loss": 0.1036, + "step": 3210 + }, + { + "epoch": 0.07527089669884023, + "grad_norm": 0.6758156418800354, + "learning_rate": 0.00019723299255446067, + "loss": 0.1071, + "step": 3211 + }, + { + "epoch": 0.07529433827364522, + "grad_norm": 0.42248862981796265, + "learning_rate": 0.00019723127146860302, + "loss": 0.0575, + "step": 3212 + }, + { + "epoch": 0.07531777984845021, + "grad_norm": 0.6010007858276367, + "learning_rate": 0.00019722954985516573, + "loss": 0.107, + "step": 3213 + }, + { + "epoch": 0.07534122142325521, + "grad_norm": 0.4091894030570984, + "learning_rate": 0.00019722782771415816, + "loss": 0.3576, + "step": 3214 + }, + { + "epoch": 0.0753646629980602, + "grad_norm": 0.31607702374458313, + "learning_rate": 0.00019722610504558968, + "loss": 0.043, + "step": 3215 + }, + { + "epoch": 0.07538810457286521, + "grad_norm": 0.6917475461959839, + "learning_rate": 0.00019722438184946957, + "loss": 0.2674, + "step": 3216 + }, + { + "epoch": 0.0754115461476702, + "grad_norm": 0.6316292881965637, + "learning_rate": 0.00019722265812580729, + "loss": 0.1481, + "step": 3217 + }, + { + "epoch": 0.0754349877224752, + "grad_norm": 0.6428734064102173, + "learning_rate": 0.00019722093387461215, + "loss": 0.7324, + "step": 3218 + }, + { + "epoch": 0.0754584292972802, + "grad_norm": 0.7015413045883179, + "learning_rate": 0.00019721920909589346, + "loss": 0.3017, + "step": 3219 + }, + { + "epoch": 0.07548187087208519, + "grad_norm": 0.20436231791973114, + "learning_rate": 0.0001972174837896606, + "loss": 0.068, + "step": 3220 + }, + { + "epoch": 0.07550531244689018, + "grad_norm": 0.45898351073265076, + "learning_rate": 0.00019721575795592292, + "loss": 0.2739, + "step": 3221 + }, + { + "epoch": 0.07552875402169518, + "grad_norm": 0.16632576286792755, + "learning_rate": 0.00019721403159468984, + "loss": 0.0357, + "step": 3222 + }, + { + "epoch": 0.07555219559650017, + "grad_norm": 0.4683643579483032, + "learning_rate": 0.00019721230470597068, + "loss": 0.1208, + "step": 3223 + }, + { + "epoch": 0.07557563717130517, + "grad_norm": 0.5204114317893982, + "learning_rate": 0.0001972105772897748, + "loss": 0.0992, + "step": 3224 + }, + { + "epoch": 0.07559907874611016, + "grad_norm": 0.3589335083961487, + "learning_rate": 0.00019720884934611163, + "loss": 0.085, + "step": 3225 + }, + { + "epoch": 0.07562252032091515, + "grad_norm": 0.3143129348754883, + "learning_rate": 0.00019720712087499046, + "loss": 0.0843, + "step": 3226 + }, + { + "epoch": 0.07564596189572015, + "grad_norm": 0.37709474563598633, + "learning_rate": 0.00019720539187642077, + "loss": 0.3305, + "step": 3227 + }, + { + "epoch": 0.07566940347052516, + "grad_norm": 0.7249090671539307, + "learning_rate": 0.00019720366235041187, + "loss": 0.1634, + "step": 3228 + }, + { + "epoch": 0.07569284504533015, + "grad_norm": 0.7184211015701294, + "learning_rate": 0.0001972019322969732, + "loss": 0.2544, + "step": 3229 + }, + { + "epoch": 0.07571628662013515, + "grad_norm": 0.6350836753845215, + "learning_rate": 0.00019720020171611408, + "loss": 0.1718, + "step": 3230 + }, + { + "epoch": 0.07573972819494014, + "grad_norm": 0.5678260326385498, + "learning_rate": 0.00019719847060784392, + "loss": 0.1132, + "step": 3231 + }, + { + "epoch": 0.07576316976974513, + "grad_norm": 0.8223496079444885, + "learning_rate": 0.00019719673897217215, + "loss": 0.1646, + "step": 3232 + }, + { + "epoch": 0.07578661134455013, + "grad_norm": 0.6253158450126648, + "learning_rate": 0.0001971950068091081, + "loss": 1.0543, + "step": 3233 + }, + { + "epoch": 0.07581005291935512, + "grad_norm": 0.17873017489910126, + "learning_rate": 0.00019719327411866125, + "loss": 0.0325, + "step": 3234 + }, + { + "epoch": 0.07583349449416012, + "grad_norm": 0.9663310050964355, + "learning_rate": 0.00019719154090084095, + "loss": 0.2087, + "step": 3235 + }, + { + "epoch": 0.07585693606896511, + "grad_norm": 0.6938252449035645, + "learning_rate": 0.00019718980715565662, + "loss": 0.5014, + "step": 3236 + }, + { + "epoch": 0.0758803776437701, + "grad_norm": 0.5855813026428223, + "learning_rate": 0.00019718807288311767, + "loss": 0.2345, + "step": 3237 + }, + { + "epoch": 0.0759038192185751, + "grad_norm": 0.5711536407470703, + "learning_rate": 0.0001971863380832335, + "loss": 0.6371, + "step": 3238 + }, + { + "epoch": 0.0759272607933801, + "grad_norm": 0.6892685890197754, + "learning_rate": 0.00019718460275601352, + "loss": 0.7921, + "step": 3239 + }, + { + "epoch": 0.07595070236818509, + "grad_norm": 0.43929365277290344, + "learning_rate": 0.00019718286690146715, + "loss": 0.1054, + "step": 3240 + }, + { + "epoch": 0.0759741439429901, + "grad_norm": 0.608704686164856, + "learning_rate": 0.0001971811305196038, + "loss": 0.1776, + "step": 3241 + }, + { + "epoch": 0.07599758551779509, + "grad_norm": 0.22255797684192657, + "learning_rate": 0.00019717939361043295, + "loss": 0.0703, + "step": 3242 + }, + { + "epoch": 0.07602102709260009, + "grad_norm": 0.15565234422683716, + "learning_rate": 0.00019717765617396392, + "loss": 0.0513, + "step": 3243 + }, + { + "epoch": 0.07604446866740508, + "grad_norm": 0.733096182346344, + "learning_rate": 0.00019717591821020626, + "loss": 0.1592, + "step": 3244 + }, + { + "epoch": 0.07606791024221007, + "grad_norm": 0.2271246612071991, + "learning_rate": 0.00019717417971916928, + "loss": 0.0437, + "step": 3245 + }, + { + "epoch": 0.07609135181701507, + "grad_norm": 0.42609456181526184, + "learning_rate": 0.00019717244070086248, + "loss": 0.661, + "step": 3246 + }, + { + "epoch": 0.07611479339182006, + "grad_norm": 0.9580833315849304, + "learning_rate": 0.0001971707011552953, + "loss": 0.2317, + "step": 3247 + }, + { + "epoch": 0.07613823496662506, + "grad_norm": 0.2614049017429352, + "learning_rate": 0.00019716896108247715, + "loss": 0.0583, + "step": 3248 + }, + { + "epoch": 0.07616167654143005, + "grad_norm": 0.9050033688545227, + "learning_rate": 0.0001971672204824175, + "loss": 0.2374, + "step": 3249 + }, + { + "epoch": 0.07618511811623505, + "grad_norm": 0.6106353402137756, + "learning_rate": 0.00019716547935512578, + "loss": 0.2011, + "step": 3250 + }, + { + "epoch": 0.07620855969104004, + "grad_norm": 0.9754160046577454, + "learning_rate": 0.0001971637377006114, + "loss": 0.1713, + "step": 3251 + }, + { + "epoch": 0.07623200126584503, + "grad_norm": 0.17407061159610748, + "learning_rate": 0.00019716199551888389, + "loss": 0.0444, + "step": 3252 + }, + { + "epoch": 0.07625544284065003, + "grad_norm": 0.5506469011306763, + "learning_rate": 0.00019716025280995262, + "loss": 0.838, + "step": 3253 + }, + { + "epoch": 0.07627888441545504, + "grad_norm": 0.5688194036483765, + "learning_rate": 0.00019715850957382713, + "loss": 0.1651, + "step": 3254 + }, + { + "epoch": 0.07630232599026003, + "grad_norm": 0.6302314400672913, + "learning_rate": 0.00019715676581051682, + "loss": 0.1545, + "step": 3255 + }, + { + "epoch": 0.07632576756506503, + "grad_norm": 0.47799181938171387, + "learning_rate": 0.00019715502152003116, + "loss": 0.1093, + "step": 3256 + }, + { + "epoch": 0.07634920913987002, + "grad_norm": 0.37884578108787537, + "learning_rate": 0.0001971532767023796, + "loss": 0.1148, + "step": 3257 + }, + { + "epoch": 0.07637265071467501, + "grad_norm": 0.36793434619903564, + "learning_rate": 0.00019715153135757167, + "loss": 0.116, + "step": 3258 + }, + { + "epoch": 0.07639609228948001, + "grad_norm": 0.7635891437530518, + "learning_rate": 0.0001971497854856168, + "loss": 0.2504, + "step": 3259 + }, + { + "epoch": 0.076419533864285, + "grad_norm": 0.6724544167518616, + "learning_rate": 0.00019714803908652443, + "loss": 0.699, + "step": 3260 + }, + { + "epoch": 0.07644297543909, + "grad_norm": 0.5477925539016724, + "learning_rate": 0.0001971462921603041, + "loss": 0.2025, + "step": 3261 + }, + { + "epoch": 0.07646641701389499, + "grad_norm": 0.20272797346115112, + "learning_rate": 0.00019714454470696522, + "loss": 0.0789, + "step": 3262 + }, + { + "epoch": 0.07648985858869999, + "grad_norm": 0.5393373370170593, + "learning_rate": 0.0001971427967265173, + "loss": 0.122, + "step": 3263 + }, + { + "epoch": 0.07651330016350498, + "grad_norm": 0.19724978506565094, + "learning_rate": 0.0001971410482189699, + "loss": 0.0556, + "step": 3264 + }, + { + "epoch": 0.07653674173830997, + "grad_norm": 0.3239450752735138, + "learning_rate": 0.0001971392991843324, + "loss": 0.0651, + "step": 3265 + }, + { + "epoch": 0.07656018331311497, + "grad_norm": 0.3613870143890381, + "learning_rate": 0.00019713754962261435, + "loss": 0.1282, + "step": 3266 + }, + { + "epoch": 0.07658362488791998, + "grad_norm": 0.22994475066661835, + "learning_rate": 0.0001971357995338252, + "loss": 0.0885, + "step": 3267 + }, + { + "epoch": 0.07660706646272497, + "grad_norm": 0.53934645652771, + "learning_rate": 0.0001971340489179745, + "loss": 0.1705, + "step": 3268 + }, + { + "epoch": 0.07663050803752997, + "grad_norm": 0.5203676819801331, + "learning_rate": 0.00019713229777507167, + "loss": 0.0857, + "step": 3269 + }, + { + "epoch": 0.07665394961233496, + "grad_norm": 0.07367340475320816, + "learning_rate": 0.00019713054610512632, + "loss": 0.0276, + "step": 3270 + }, + { + "epoch": 0.07667739118713995, + "grad_norm": 0.7478379011154175, + "learning_rate": 0.00019712879390814785, + "loss": 0.2666, + "step": 3271 + }, + { + "epoch": 0.07670083276194495, + "grad_norm": 0.23750583827495575, + "learning_rate": 0.00019712704118414581, + "loss": 0.0794, + "step": 3272 + }, + { + "epoch": 0.07672427433674994, + "grad_norm": 0.6280679702758789, + "learning_rate": 0.00019712528793312974, + "loss": 0.2179, + "step": 3273 + }, + { + "epoch": 0.07674771591155494, + "grad_norm": 0.3712387681007385, + "learning_rate": 0.00019712353415510912, + "loss": 0.0611, + "step": 3274 + }, + { + "epoch": 0.07677115748635993, + "grad_norm": 0.18518033623695374, + "learning_rate": 0.0001971217798500935, + "loss": 0.0435, + "step": 3275 + }, + { + "epoch": 0.07679459906116493, + "grad_norm": 0.8565661907196045, + "learning_rate": 0.00019712002501809232, + "loss": 0.1639, + "step": 3276 + }, + { + "epoch": 0.07681804063596992, + "grad_norm": 0.651785671710968, + "learning_rate": 0.00019711826965911517, + "loss": 0.2532, + "step": 3277 + }, + { + "epoch": 0.07684148221077491, + "grad_norm": 0.6226869225502014, + "learning_rate": 0.0001971165137731716, + "loss": 0.1092, + "step": 3278 + }, + { + "epoch": 0.07686492378557991, + "grad_norm": 0.5982236266136169, + "learning_rate": 0.00019711475736027104, + "loss": 0.2383, + "step": 3279 + }, + { + "epoch": 0.07688836536038492, + "grad_norm": 0.1675049364566803, + "learning_rate": 0.0001971130004204231, + "loss": 0.0315, + "step": 3280 + }, + { + "epoch": 0.07691180693518991, + "grad_norm": 0.542087733745575, + "learning_rate": 0.0001971112429536373, + "loss": 0.1407, + "step": 3281 + }, + { + "epoch": 0.0769352485099949, + "grad_norm": 0.26511090993881226, + "learning_rate": 0.00019710948495992313, + "loss": 0.0414, + "step": 3282 + }, + { + "epoch": 0.0769586900847999, + "grad_norm": 0.2627033293247223, + "learning_rate": 0.0001971077264392902, + "loss": 0.0641, + "step": 3283 + }, + { + "epoch": 0.0769821316596049, + "grad_norm": 0.5352437496185303, + "learning_rate": 0.000197105967391748, + "loss": 0.1615, + "step": 3284 + }, + { + "epoch": 0.07700557323440989, + "grad_norm": 0.5524579882621765, + "learning_rate": 0.00019710420781730606, + "loss": 0.1009, + "step": 3285 + }, + { + "epoch": 0.07702901480921488, + "grad_norm": 0.3163284659385681, + "learning_rate": 0.00019710244771597402, + "loss": 0.0289, + "step": 3286 + }, + { + "epoch": 0.07705245638401988, + "grad_norm": 0.5992369055747986, + "learning_rate": 0.00019710068708776133, + "loss": 0.2189, + "step": 3287 + }, + { + "epoch": 0.07707589795882487, + "grad_norm": 0.6674913763999939, + "learning_rate": 0.0001970989259326776, + "loss": 0.1644, + "step": 3288 + }, + { + "epoch": 0.07709933953362987, + "grad_norm": 0.7229846715927124, + "learning_rate": 0.00019709716425073235, + "loss": 0.2261, + "step": 3289 + }, + { + "epoch": 0.07712278110843486, + "grad_norm": 0.684005081653595, + "learning_rate": 0.00019709540204193517, + "loss": 0.3265, + "step": 3290 + }, + { + "epoch": 0.07714622268323985, + "grad_norm": 0.5245772004127502, + "learning_rate": 0.0001970936393062956, + "loss": 0.1137, + "step": 3291 + }, + { + "epoch": 0.07716966425804486, + "grad_norm": 0.5687353014945984, + "learning_rate": 0.00019709187604382322, + "loss": 0.1733, + "step": 3292 + }, + { + "epoch": 0.07719310583284986, + "grad_norm": 0.6001117825508118, + "learning_rate": 0.00019709011225452758, + "loss": 0.2217, + "step": 3293 + }, + { + "epoch": 0.07721654740765485, + "grad_norm": 0.19574099779129028, + "learning_rate": 0.00019708834793841828, + "loss": 0.0643, + "step": 3294 + }, + { + "epoch": 0.07723998898245985, + "grad_norm": 0.6759383678436279, + "learning_rate": 0.00019708658309550487, + "loss": 0.1589, + "step": 3295 + }, + { + "epoch": 0.07726343055726484, + "grad_norm": 0.21110899746418, + "learning_rate": 0.0001970848177257969, + "loss": 0.0534, + "step": 3296 + }, + { + "epoch": 0.07728687213206983, + "grad_norm": 0.7893348336219788, + "learning_rate": 0.00019708305182930403, + "loss": 0.212, + "step": 3297 + }, + { + "epoch": 0.07731031370687483, + "grad_norm": 0.8165237307548523, + "learning_rate": 0.00019708128540603577, + "loss": 0.173, + "step": 3298 + }, + { + "epoch": 0.07733375528167982, + "grad_norm": 0.6541932821273804, + "learning_rate": 0.0001970795184560017, + "loss": 0.1404, + "step": 3299 + }, + { + "epoch": 0.07735719685648482, + "grad_norm": 0.8600411415100098, + "learning_rate": 0.00019707775097921144, + "loss": 0.3027, + "step": 3300 + }, + { + "epoch": 0.07738063843128981, + "grad_norm": 0.34670543670654297, + "learning_rate": 0.0001970759829756746, + "loss": 0.1056, + "step": 3301 + }, + { + "epoch": 0.0774040800060948, + "grad_norm": 0.4590158462524414, + "learning_rate": 0.00019707421444540072, + "loss": 0.6018, + "step": 3302 + }, + { + "epoch": 0.0774275215808998, + "grad_norm": 0.22700604796409607, + "learning_rate": 0.00019707244538839943, + "loss": 0.0624, + "step": 3303 + }, + { + "epoch": 0.0774509631557048, + "grad_norm": 0.19863735139369965, + "learning_rate": 0.00019707067580468035, + "loss": 0.0429, + "step": 3304 + }, + { + "epoch": 0.0774744047305098, + "grad_norm": 0.28841453790664673, + "learning_rate": 0.00019706890569425303, + "loss": 0.0928, + "step": 3305 + }, + { + "epoch": 0.0774978463053148, + "grad_norm": 0.5058981776237488, + "learning_rate": 0.00019706713505712712, + "loss": 0.182, + "step": 3306 + }, + { + "epoch": 0.07752128788011979, + "grad_norm": 0.19116389751434326, + "learning_rate": 0.00019706536389331217, + "loss": 0.0455, + "step": 3307 + }, + { + "epoch": 0.07754472945492479, + "grad_norm": 0.7532224655151367, + "learning_rate": 0.00019706359220281782, + "loss": 0.2621, + "step": 3308 + }, + { + "epoch": 0.07756817102972978, + "grad_norm": 0.5861573219299316, + "learning_rate": 0.0001970618199856537, + "loss": 0.1813, + "step": 3309 + }, + { + "epoch": 0.07759161260453477, + "grad_norm": 0.7029127478599548, + "learning_rate": 0.00019706004724182944, + "loss": 0.205, + "step": 3310 + }, + { + "epoch": 0.07761505417933977, + "grad_norm": 0.6041339635848999, + "learning_rate": 0.00019705827397135465, + "loss": 0.1724, + "step": 3311 + }, + { + "epoch": 0.07763849575414476, + "grad_norm": 0.3675556480884552, + "learning_rate": 0.00019705650017423886, + "loss": 0.6833, + "step": 3312 + }, + { + "epoch": 0.07766193732894976, + "grad_norm": 0.39072418212890625, + "learning_rate": 0.00019705472585049183, + "loss": 0.0579, + "step": 3313 + }, + { + "epoch": 0.07768537890375475, + "grad_norm": 0.33774226903915405, + "learning_rate": 0.00019705295100012313, + "loss": 0.1003, + "step": 3314 + }, + { + "epoch": 0.07770882047855975, + "grad_norm": 0.45493513345718384, + "learning_rate": 0.0001970511756231424, + "loss": 0.211, + "step": 3315 + }, + { + "epoch": 0.07773226205336474, + "grad_norm": 0.525829017162323, + "learning_rate": 0.0001970493997195592, + "loss": 0.1603, + "step": 3316 + }, + { + "epoch": 0.07775570362816973, + "grad_norm": 0.7530962228775024, + "learning_rate": 0.0001970476232893833, + "loss": 0.3088, + "step": 3317 + }, + { + "epoch": 0.07777914520297474, + "grad_norm": 0.6996729969978333, + "learning_rate": 0.00019704584633262423, + "loss": 0.1551, + "step": 3318 + }, + { + "epoch": 0.07780258677777974, + "grad_norm": 0.219340980052948, + "learning_rate": 0.00019704406884929164, + "loss": 0.0605, + "step": 3319 + }, + { + "epoch": 0.07782602835258473, + "grad_norm": 0.3001663386821747, + "learning_rate": 0.00019704229083939527, + "loss": 0.0772, + "step": 3320 + }, + { + "epoch": 0.07784946992738973, + "grad_norm": 0.45327767729759216, + "learning_rate": 0.00019704051230294467, + "loss": 0.0916, + "step": 3321 + }, + { + "epoch": 0.07787291150219472, + "grad_norm": 0.38910743594169617, + "learning_rate": 0.0001970387332399495, + "loss": 0.116, + "step": 3322 + }, + { + "epoch": 0.07789635307699971, + "grad_norm": 0.8471893072128296, + "learning_rate": 0.00019703695365041945, + "loss": 0.2296, + "step": 3323 + }, + { + "epoch": 0.07791979465180471, + "grad_norm": 0.3062788248062134, + "learning_rate": 0.00019703517353436415, + "loss": 0.0738, + "step": 3324 + }, + { + "epoch": 0.0779432362266097, + "grad_norm": 0.8267127275466919, + "learning_rate": 0.00019703339289179331, + "loss": 0.1634, + "step": 3325 + }, + { + "epoch": 0.0779666778014147, + "grad_norm": 0.5300244688987732, + "learning_rate": 0.00019703161172271653, + "loss": 0.1262, + "step": 3326 + }, + { + "epoch": 0.07799011937621969, + "grad_norm": 0.24654646217823029, + "learning_rate": 0.0001970298300271435, + "loss": 0.0684, + "step": 3327 + }, + { + "epoch": 0.07801356095102469, + "grad_norm": 0.5833593010902405, + "learning_rate": 0.0001970280478050839, + "loss": 0.1523, + "step": 3328 + }, + { + "epoch": 0.07803700252582968, + "grad_norm": 0.6933130025863647, + "learning_rate": 0.00019702626505654734, + "loss": 0.1485, + "step": 3329 + }, + { + "epoch": 0.07806044410063467, + "grad_norm": 0.1896199882030487, + "learning_rate": 0.00019702448178154355, + "loss": 0.0433, + "step": 3330 + }, + { + "epoch": 0.07808388567543968, + "grad_norm": 0.48619672656059265, + "learning_rate": 0.00019702269798008222, + "loss": 0.1098, + "step": 3331 + }, + { + "epoch": 0.07810732725024468, + "grad_norm": 0.4655081331729889, + "learning_rate": 0.00019702091365217297, + "loss": 0.0725, + "step": 3332 + }, + { + "epoch": 0.07813076882504967, + "grad_norm": 0.7860180139541626, + "learning_rate": 0.00019701912879782553, + "loss": 0.2913, + "step": 3333 + }, + { + "epoch": 0.07815421039985467, + "grad_norm": 0.7365726828575134, + "learning_rate": 0.00019701734341704959, + "loss": 0.217, + "step": 3334 + }, + { + "epoch": 0.07817765197465966, + "grad_norm": 0.12334848940372467, + "learning_rate": 0.00019701555750985478, + "loss": 0.0295, + "step": 3335 + }, + { + "epoch": 0.07820109354946465, + "grad_norm": 0.8464415073394775, + "learning_rate": 0.00019701377107625084, + "loss": 0.7115, + "step": 3336 + }, + { + "epoch": 0.07822453512426965, + "grad_norm": 0.8929542899131775, + "learning_rate": 0.00019701198411624745, + "loss": 0.1697, + "step": 3337 + }, + { + "epoch": 0.07824797669907464, + "grad_norm": 0.6009935140609741, + "learning_rate": 0.0001970101966298543, + "loss": 0.1303, + "step": 3338 + }, + { + "epoch": 0.07827141827387964, + "grad_norm": 0.2503918409347534, + "learning_rate": 0.00019700840861708115, + "loss": 0.0616, + "step": 3339 + }, + { + "epoch": 0.07829485984868463, + "grad_norm": 0.6553563475608826, + "learning_rate": 0.0001970066200779376, + "loss": 0.269, + "step": 3340 + }, + { + "epoch": 0.07831830142348963, + "grad_norm": 0.24984265863895416, + "learning_rate": 0.0001970048310124334, + "loss": 0.0738, + "step": 3341 + }, + { + "epoch": 0.07834174299829462, + "grad_norm": 0.9425820708274841, + "learning_rate": 0.00019700304142057827, + "loss": 0.2109, + "step": 3342 + }, + { + "epoch": 0.07836518457309963, + "grad_norm": 0.29333630204200745, + "learning_rate": 0.0001970012513023819, + "loss": 0.1012, + "step": 3343 + }, + { + "epoch": 0.07838862614790462, + "grad_norm": 0.35886773467063904, + "learning_rate": 0.00019699946065785401, + "loss": 0.0524, + "step": 3344 + }, + { + "epoch": 0.07841206772270962, + "grad_norm": 0.38375699520111084, + "learning_rate": 0.00019699766948700433, + "loss": 0.1274, + "step": 3345 + }, + { + "epoch": 0.07843550929751461, + "grad_norm": 0.4624890089035034, + "learning_rate": 0.00019699587778984257, + "loss": 0.2529, + "step": 3346 + }, + { + "epoch": 0.0784589508723196, + "grad_norm": 0.6124911904335022, + "learning_rate": 0.00019699408556637845, + "loss": 0.2254, + "step": 3347 + }, + { + "epoch": 0.0784823924471246, + "grad_norm": 0.7630993723869324, + "learning_rate": 0.00019699229281662167, + "loss": 0.7449, + "step": 3348 + }, + { + "epoch": 0.0785058340219296, + "grad_norm": 0.4782824218273163, + "learning_rate": 0.00019699049954058202, + "loss": 0.1815, + "step": 3349 + }, + { + "epoch": 0.07852927559673459, + "grad_norm": 0.7765792012214661, + "learning_rate": 0.00019698870573826914, + "loss": 0.1942, + "step": 3350 + }, + { + "epoch": 0.07855271717153958, + "grad_norm": 0.6478104591369629, + "learning_rate": 0.00019698691140969288, + "loss": 0.1563, + "step": 3351 + }, + { + "epoch": 0.07857615874634458, + "grad_norm": 0.5362963080406189, + "learning_rate": 0.00019698511655486287, + "loss": 0.8754, + "step": 3352 + }, + { + "epoch": 0.07859960032114957, + "grad_norm": 0.44379884004592896, + "learning_rate": 0.0001969833211737889, + "loss": 0.1683, + "step": 3353 + }, + { + "epoch": 0.07862304189595457, + "grad_norm": 0.45300740003585815, + "learning_rate": 0.0001969815252664807, + "loss": 0.17, + "step": 3354 + }, + { + "epoch": 0.07864648347075956, + "grad_norm": 0.2637466490268707, + "learning_rate": 0.00019697972883294803, + "loss": 0.0691, + "step": 3355 + }, + { + "epoch": 0.07866992504556457, + "grad_norm": 0.44470474123954773, + "learning_rate": 0.00019697793187320063, + "loss": 0.0792, + "step": 3356 + }, + { + "epoch": 0.07869336662036956, + "grad_norm": 0.4707695245742798, + "learning_rate": 0.00019697613438724823, + "loss": 0.1689, + "step": 3357 + }, + { + "epoch": 0.07871680819517456, + "grad_norm": 0.3365781903266907, + "learning_rate": 0.00019697433637510059, + "loss": 0.1421, + "step": 3358 + }, + { + "epoch": 0.07874024976997955, + "grad_norm": 0.46322473883628845, + "learning_rate": 0.00019697253783676748, + "loss": 0.1202, + "step": 3359 + }, + { + "epoch": 0.07876369134478454, + "grad_norm": 0.39445027709007263, + "learning_rate": 0.00019697073877225868, + "loss": 0.5015, + "step": 3360 + }, + { + "epoch": 0.07878713291958954, + "grad_norm": 0.5675919055938721, + "learning_rate": 0.00019696893918158388, + "loss": 0.1547, + "step": 3361 + }, + { + "epoch": 0.07881057449439453, + "grad_norm": 0.7720356583595276, + "learning_rate": 0.0001969671390647529, + "loss": 0.1805, + "step": 3362 + }, + { + "epoch": 0.07883401606919953, + "grad_norm": 0.402239054441452, + "learning_rate": 0.0001969653384217755, + "loss": 0.168, + "step": 3363 + }, + { + "epoch": 0.07885745764400452, + "grad_norm": 0.6927630305290222, + "learning_rate": 0.00019696353725266147, + "loss": 0.11, + "step": 3364 + }, + { + "epoch": 0.07888089921880952, + "grad_norm": 0.39841845631599426, + "learning_rate": 0.00019696173555742053, + "loss": 0.2041, + "step": 3365 + }, + { + "epoch": 0.07890434079361451, + "grad_norm": 0.503152072429657, + "learning_rate": 0.0001969599333360625, + "loss": 0.692, + "step": 3366 + }, + { + "epoch": 0.0789277823684195, + "grad_norm": 0.17600803077220917, + "learning_rate": 0.00019695813058859715, + "loss": 0.0431, + "step": 3367 + }, + { + "epoch": 0.0789512239432245, + "grad_norm": 0.22685539722442627, + "learning_rate": 0.00019695632731503426, + "loss": 0.0574, + "step": 3368 + }, + { + "epoch": 0.07897466551802951, + "grad_norm": 0.5681820511817932, + "learning_rate": 0.0001969545235153836, + "loss": 0.1652, + "step": 3369 + }, + { + "epoch": 0.0789981070928345, + "grad_norm": 0.3631137013435364, + "learning_rate": 0.00019695271918965495, + "loss": 0.058, + "step": 3370 + }, + { + "epoch": 0.0790215486676395, + "grad_norm": 0.2378411740064621, + "learning_rate": 0.00019695091433785818, + "loss": 0.0472, + "step": 3371 + }, + { + "epoch": 0.07904499024244449, + "grad_norm": 0.20580261945724487, + "learning_rate": 0.000196949108960003, + "loss": 0.05, + "step": 3372 + }, + { + "epoch": 0.07906843181724948, + "grad_norm": 0.4839774966239929, + "learning_rate": 0.00019694730305609917, + "loss": 0.126, + "step": 3373 + }, + { + "epoch": 0.07909187339205448, + "grad_norm": 0.5923928022384644, + "learning_rate": 0.0001969454966261566, + "loss": 0.6722, + "step": 3374 + }, + { + "epoch": 0.07911531496685947, + "grad_norm": 0.67839515209198, + "learning_rate": 0.00019694368967018507, + "loss": 0.1108, + "step": 3375 + }, + { + "epoch": 0.07913875654166447, + "grad_norm": 0.4238930344581604, + "learning_rate": 0.0001969418821881943, + "loss": 0.0847, + "step": 3376 + }, + { + "epoch": 0.07916219811646946, + "grad_norm": 0.765417754650116, + "learning_rate": 0.00019694007418019415, + "loss": 0.1401, + "step": 3377 + }, + { + "epoch": 0.07918563969127446, + "grad_norm": 0.18471483886241913, + "learning_rate": 0.00019693826564619449, + "loss": 0.0276, + "step": 3378 + }, + { + "epoch": 0.07920908126607945, + "grad_norm": 0.7070007920265198, + "learning_rate": 0.00019693645658620503, + "loss": 0.2196, + "step": 3379 + }, + { + "epoch": 0.07923252284088445, + "grad_norm": 0.7856169939041138, + "learning_rate": 0.00019693464700023562, + "loss": 0.1547, + "step": 3380 + }, + { + "epoch": 0.07925596441568944, + "grad_norm": 0.8233810067176819, + "learning_rate": 0.0001969328368882961, + "loss": 0.2074, + "step": 3381 + }, + { + "epoch": 0.07927940599049445, + "grad_norm": 0.5059824585914612, + "learning_rate": 0.0001969310262503963, + "loss": 0.7262, + "step": 3382 + }, + { + "epoch": 0.07930284756529944, + "grad_norm": 0.6324707269668579, + "learning_rate": 0.000196929215086546, + "loss": 0.2338, + "step": 3383 + }, + { + "epoch": 0.07932628914010444, + "grad_norm": 0.492562860250473, + "learning_rate": 0.00019692740339675508, + "loss": 0.0824, + "step": 3384 + }, + { + "epoch": 0.07934973071490943, + "grad_norm": 0.6762421727180481, + "learning_rate": 0.00019692559118103333, + "loss": 0.2306, + "step": 3385 + }, + { + "epoch": 0.07937317228971442, + "grad_norm": 1.0182840824127197, + "learning_rate": 0.0001969237784393906, + "loss": 0.1734, + "step": 3386 + }, + { + "epoch": 0.07939661386451942, + "grad_norm": 0.851369321346283, + "learning_rate": 0.0001969219651718367, + "loss": 0.2074, + "step": 3387 + }, + { + "epoch": 0.07942005543932441, + "grad_norm": 0.6136981844902039, + "learning_rate": 0.00019692015137838154, + "loss": 0.5625, + "step": 3388 + }, + { + "epoch": 0.07944349701412941, + "grad_norm": 0.5399134159088135, + "learning_rate": 0.00019691833705903487, + "loss": 0.7914, + "step": 3389 + }, + { + "epoch": 0.0794669385889344, + "grad_norm": 0.5097196698188782, + "learning_rate": 0.00019691652221380662, + "loss": 0.121, + "step": 3390 + }, + { + "epoch": 0.0794903801637394, + "grad_norm": 0.4309489130973816, + "learning_rate": 0.00019691470684270654, + "loss": 0.1282, + "step": 3391 + }, + { + "epoch": 0.07951382173854439, + "grad_norm": 0.27175426483154297, + "learning_rate": 0.00019691289094574458, + "loss": 0.0365, + "step": 3392 + }, + { + "epoch": 0.07953726331334939, + "grad_norm": 0.4616539776325226, + "learning_rate": 0.00019691107452293055, + "loss": 0.0936, + "step": 3393 + }, + { + "epoch": 0.07956070488815438, + "grad_norm": 0.5676476359367371, + "learning_rate": 0.00019690925757427427, + "loss": 0.1912, + "step": 3394 + }, + { + "epoch": 0.07958414646295939, + "grad_norm": 0.4127635359764099, + "learning_rate": 0.00019690744009978565, + "loss": 0.2237, + "step": 3395 + }, + { + "epoch": 0.07960758803776438, + "grad_norm": 0.24629522860050201, + "learning_rate": 0.00019690562209947454, + "loss": 0.0602, + "step": 3396 + }, + { + "epoch": 0.07963102961256938, + "grad_norm": 0.6488888263702393, + "learning_rate": 0.0001969038035733508, + "loss": 0.2731, + "step": 3397 + }, + { + "epoch": 0.07965447118737437, + "grad_norm": 0.3242536187171936, + "learning_rate": 0.0001969019845214243, + "loss": 0.4619, + "step": 3398 + }, + { + "epoch": 0.07967791276217936, + "grad_norm": 0.6491689085960388, + "learning_rate": 0.00019690016494370492, + "loss": 0.1746, + "step": 3399 + }, + { + "epoch": 0.07970135433698436, + "grad_norm": 0.404083788394928, + "learning_rate": 0.0001968983448402025, + "loss": 0.0671, + "step": 3400 + }, + { + "epoch": 0.07972479591178935, + "grad_norm": 0.6536796689033508, + "learning_rate": 0.00019689652421092692, + "loss": 0.1583, + "step": 3401 + }, + { + "epoch": 0.07974823748659435, + "grad_norm": 0.42821142077445984, + "learning_rate": 0.00019689470305588809, + "loss": 0.1531, + "step": 3402 + }, + { + "epoch": 0.07977167906139934, + "grad_norm": 0.27636149525642395, + "learning_rate": 0.00019689288137509587, + "loss": 0.0404, + "step": 3403 + }, + { + "epoch": 0.07979512063620434, + "grad_norm": 0.14426742494106293, + "learning_rate": 0.00019689105916856017, + "loss": 0.0334, + "step": 3404 + }, + { + "epoch": 0.07981856221100933, + "grad_norm": 0.5653432607650757, + "learning_rate": 0.00019688923643629088, + "loss": 0.2171, + "step": 3405 + }, + { + "epoch": 0.07984200378581433, + "grad_norm": 0.46304482221603394, + "learning_rate": 0.0001968874131782978, + "loss": 0.1334, + "step": 3406 + }, + { + "epoch": 0.07986544536061933, + "grad_norm": 0.6779239773750305, + "learning_rate": 0.00019688558939459094, + "loss": 0.2351, + "step": 3407 + }, + { + "epoch": 0.07988888693542433, + "grad_norm": 0.33286383748054504, + "learning_rate": 0.00019688376508518015, + "loss": 0.0495, + "step": 3408 + }, + { + "epoch": 0.07991232851022932, + "grad_norm": 0.5647602081298828, + "learning_rate": 0.00019688194025007532, + "loss": 0.9027, + "step": 3409 + }, + { + "epoch": 0.07993577008503432, + "grad_norm": 0.5380653142929077, + "learning_rate": 0.00019688011488928634, + "loss": 0.1432, + "step": 3410 + }, + { + "epoch": 0.07995921165983931, + "grad_norm": 0.209493950009346, + "learning_rate": 0.00019687828900282316, + "loss": 0.0556, + "step": 3411 + }, + { + "epoch": 0.0799826532346443, + "grad_norm": 0.6179551482200623, + "learning_rate": 0.00019687646259069564, + "loss": 0.8217, + "step": 3412 + }, + { + "epoch": 0.0800060948094493, + "grad_norm": 0.1775025725364685, + "learning_rate": 0.00019687463565291373, + "loss": 0.0413, + "step": 3413 + }, + { + "epoch": 0.0800295363842543, + "grad_norm": 0.5947052836418152, + "learning_rate": 0.0001968728081894873, + "loss": 0.1999, + "step": 3414 + }, + { + "epoch": 0.08005297795905929, + "grad_norm": 0.6143971085548401, + "learning_rate": 0.00019687098020042628, + "loss": 0.2561, + "step": 3415 + }, + { + "epoch": 0.08007641953386428, + "grad_norm": 0.4250313639640808, + "learning_rate": 0.00019686915168574062, + "loss": 0.1812, + "step": 3416 + }, + { + "epoch": 0.08009986110866928, + "grad_norm": 0.6003297567367554, + "learning_rate": 0.00019686732264544024, + "loss": 0.208, + "step": 3417 + }, + { + "epoch": 0.08012330268347427, + "grad_norm": 0.1185346469283104, + "learning_rate": 0.000196865493079535, + "loss": 0.0216, + "step": 3418 + }, + { + "epoch": 0.08014674425827926, + "grad_norm": 0.391910582780838, + "learning_rate": 0.0001968636629880349, + "loss": 0.1315, + "step": 3419 + }, + { + "epoch": 0.08017018583308427, + "grad_norm": 0.5781988501548767, + "learning_rate": 0.0001968618323709498, + "loss": 0.152, + "step": 3420 + }, + { + "epoch": 0.08019362740788927, + "grad_norm": 0.6430306434631348, + "learning_rate": 0.00019686000122828975, + "loss": 0.1422, + "step": 3421 + }, + { + "epoch": 0.08021706898269426, + "grad_norm": 0.6622399687767029, + "learning_rate": 0.00019685816956006456, + "loss": 0.1918, + "step": 3422 + }, + { + "epoch": 0.08024051055749926, + "grad_norm": 0.3584354519844055, + "learning_rate": 0.00019685633736628421, + "loss": 0.0805, + "step": 3423 + }, + { + "epoch": 0.08026395213230425, + "grad_norm": 0.49216434359550476, + "learning_rate": 0.00019685450464695866, + "loss": 0.7446, + "step": 3424 + }, + { + "epoch": 0.08028739370710924, + "grad_norm": 0.45553937554359436, + "learning_rate": 0.00019685267140209786, + "loss": 0.0792, + "step": 3425 + }, + { + "epoch": 0.08031083528191424, + "grad_norm": 0.1979358196258545, + "learning_rate": 0.0001968508376317117, + "loss": 0.0639, + "step": 3426 + }, + { + "epoch": 0.08033427685671923, + "grad_norm": 0.20126570761203766, + "learning_rate": 0.00019684900333581022, + "loss": 0.0413, + "step": 3427 + }, + { + "epoch": 0.08035771843152423, + "grad_norm": 0.8626415133476257, + "learning_rate": 0.00019684716851440333, + "loss": 0.1902, + "step": 3428 + }, + { + "epoch": 0.08038116000632922, + "grad_norm": 0.757691502571106, + "learning_rate": 0.00019684533316750093, + "loss": 0.207, + "step": 3429 + }, + { + "epoch": 0.08040460158113422, + "grad_norm": 0.8669591546058655, + "learning_rate": 0.00019684349729511306, + "loss": 0.3598, + "step": 3430 + }, + { + "epoch": 0.08042804315593921, + "grad_norm": 0.5605417490005493, + "learning_rate": 0.00019684166089724964, + "loss": 0.1009, + "step": 3431 + }, + { + "epoch": 0.0804514847307442, + "grad_norm": 0.25720322132110596, + "learning_rate": 0.00019683982397392065, + "loss": 0.1683, + "step": 3432 + }, + { + "epoch": 0.08047492630554921, + "grad_norm": 0.8706725835800171, + "learning_rate": 0.00019683798652513605, + "loss": 0.2105, + "step": 3433 + }, + { + "epoch": 0.08049836788035421, + "grad_norm": 0.57289057970047, + "learning_rate": 0.00019683614855090583, + "loss": 0.2505, + "step": 3434 + }, + { + "epoch": 0.0805218094551592, + "grad_norm": 0.5324033498764038, + "learning_rate": 0.00019683431005123992, + "loss": 0.1412, + "step": 3435 + }, + { + "epoch": 0.0805452510299642, + "grad_norm": 1.165209174156189, + "learning_rate": 0.00019683247102614834, + "loss": 0.1948, + "step": 3436 + }, + { + "epoch": 0.08056869260476919, + "grad_norm": 0.6398192048072815, + "learning_rate": 0.00019683063147564104, + "loss": 0.2407, + "step": 3437 + }, + { + "epoch": 0.08059213417957418, + "grad_norm": 0.5613313913345337, + "learning_rate": 0.00019682879139972802, + "loss": 0.1239, + "step": 3438 + }, + { + "epoch": 0.08061557575437918, + "grad_norm": 0.40165477991104126, + "learning_rate": 0.00019682695079841923, + "loss": 0.1408, + "step": 3439 + }, + { + "epoch": 0.08063901732918417, + "grad_norm": 0.18315941095352173, + "learning_rate": 0.00019682510967172472, + "loss": 0.0409, + "step": 3440 + }, + { + "epoch": 0.08066245890398917, + "grad_norm": 0.44724225997924805, + "learning_rate": 0.0001968232680196544, + "loss": 0.1229, + "step": 3441 + }, + { + "epoch": 0.08068590047879416, + "grad_norm": 0.760633111000061, + "learning_rate": 0.00019682142584221833, + "loss": 0.2426, + "step": 3442 + }, + { + "epoch": 0.08070934205359916, + "grad_norm": 0.39492183923721313, + "learning_rate": 0.0001968195831394265, + "loss": 0.1095, + "step": 3443 + }, + { + "epoch": 0.08073278362840415, + "grad_norm": 0.7603346705436707, + "learning_rate": 0.00019681773991128888, + "loss": 0.2884, + "step": 3444 + }, + { + "epoch": 0.08075622520320914, + "grad_norm": 0.5072410106658936, + "learning_rate": 0.00019681589615781544, + "loss": 0.0865, + "step": 3445 + }, + { + "epoch": 0.08077966677801415, + "grad_norm": 0.5827265977859497, + "learning_rate": 0.00019681405187901626, + "loss": 0.6213, + "step": 3446 + }, + { + "epoch": 0.08080310835281915, + "grad_norm": 0.6795193552970886, + "learning_rate": 0.00019681220707490132, + "loss": 0.201, + "step": 3447 + }, + { + "epoch": 0.08082654992762414, + "grad_norm": 0.1698797047138214, + "learning_rate": 0.0001968103617454806, + "loss": 0.0488, + "step": 3448 + }, + { + "epoch": 0.08084999150242914, + "grad_norm": 0.26648449897766113, + "learning_rate": 0.00019680851589076414, + "loss": 0.0706, + "step": 3449 + }, + { + "epoch": 0.08087343307723413, + "grad_norm": 0.22123970091342926, + "learning_rate": 0.00019680666951076196, + "loss": 0.0414, + "step": 3450 + }, + { + "epoch": 0.08089687465203912, + "grad_norm": 0.24542462825775146, + "learning_rate": 0.00019680482260548408, + "loss": 0.0429, + "step": 3451 + }, + { + "epoch": 0.08092031622684412, + "grad_norm": 0.3195931017398834, + "learning_rate": 0.0001968029751749405, + "loss": 0.076, + "step": 3452 + }, + { + "epoch": 0.08094375780164911, + "grad_norm": 0.5867433547973633, + "learning_rate": 0.00019680112721914123, + "loss": 0.7147, + "step": 3453 + }, + { + "epoch": 0.08096719937645411, + "grad_norm": 0.407906174659729, + "learning_rate": 0.00019679927873809635, + "loss": 0.0729, + "step": 3454 + }, + { + "epoch": 0.0809906409512591, + "grad_norm": 0.25335589051246643, + "learning_rate": 0.00019679742973181585, + "loss": 0.0661, + "step": 3455 + }, + { + "epoch": 0.0810140825260641, + "grad_norm": 0.4289347231388092, + "learning_rate": 0.00019679558020030977, + "loss": 0.1607, + "step": 3456 + }, + { + "epoch": 0.08103752410086909, + "grad_norm": 0.7811866998672485, + "learning_rate": 0.00019679373014358818, + "loss": 0.0936, + "step": 3457 + }, + { + "epoch": 0.0810609656756741, + "grad_norm": 0.541293740272522, + "learning_rate": 0.00019679187956166106, + "loss": 0.2076, + "step": 3458 + }, + { + "epoch": 0.08108440725047909, + "grad_norm": 0.2514490783214569, + "learning_rate": 0.0001967900284545385, + "loss": 0.0626, + "step": 3459 + }, + { + "epoch": 0.08110784882528409, + "grad_norm": 0.17491544783115387, + "learning_rate": 0.0001967881768222305, + "loss": 0.0413, + "step": 3460 + }, + { + "epoch": 0.08113129040008908, + "grad_norm": 0.31301236152648926, + "learning_rate": 0.00019678632466474714, + "loss": 0.2725, + "step": 3461 + }, + { + "epoch": 0.08115473197489408, + "grad_norm": 0.6070270538330078, + "learning_rate": 0.00019678447198209846, + "loss": 0.1388, + "step": 3462 + }, + { + "epoch": 0.08117817354969907, + "grad_norm": 0.7408750653266907, + "learning_rate": 0.00019678261877429453, + "loss": 0.2262, + "step": 3463 + }, + { + "epoch": 0.08120161512450406, + "grad_norm": 0.8615448474884033, + "learning_rate": 0.00019678076504134535, + "loss": 0.0924, + "step": 3464 + }, + { + "epoch": 0.08122505669930906, + "grad_norm": 0.6869056224822998, + "learning_rate": 0.00019677891078326108, + "loss": 0.2574, + "step": 3465 + }, + { + "epoch": 0.08124849827411405, + "grad_norm": 0.2041648030281067, + "learning_rate": 0.00019677705600005166, + "loss": 0.0468, + "step": 3466 + }, + { + "epoch": 0.08127193984891905, + "grad_norm": 0.36616840958595276, + "learning_rate": 0.00019677520069172723, + "loss": 0.1009, + "step": 3467 + }, + { + "epoch": 0.08129538142372404, + "grad_norm": 0.23598717153072357, + "learning_rate": 0.00019677334485829784, + "loss": 0.0772, + "step": 3468 + }, + { + "epoch": 0.08131882299852904, + "grad_norm": 0.8107052445411682, + "learning_rate": 0.00019677148849977355, + "loss": 0.217, + "step": 3469 + }, + { + "epoch": 0.08134226457333403, + "grad_norm": 0.6137067079544067, + "learning_rate": 0.0001967696316161645, + "loss": 0.8422, + "step": 3470 + }, + { + "epoch": 0.08136570614813904, + "grad_norm": 0.4745425879955292, + "learning_rate": 0.00019676777420748066, + "loss": 0.1423, + "step": 3471 + }, + { + "epoch": 0.08138914772294403, + "grad_norm": 0.38012614846229553, + "learning_rate": 0.00019676591627373216, + "loss": 0.1158, + "step": 3472 + }, + { + "epoch": 0.08141258929774903, + "grad_norm": 0.6560646891593933, + "learning_rate": 0.00019676405781492905, + "loss": 0.1532, + "step": 3473 + }, + { + "epoch": 0.08143603087255402, + "grad_norm": 0.338238924741745, + "learning_rate": 0.0001967621988310815, + "loss": 0.0816, + "step": 3474 + }, + { + "epoch": 0.08145947244735902, + "grad_norm": 0.39229127764701843, + "learning_rate": 0.0001967603393221995, + "loss": 0.1067, + "step": 3475 + }, + { + "epoch": 0.08148291402216401, + "grad_norm": 0.47316649556159973, + "learning_rate": 0.0001967584792882932, + "loss": 0.0837, + "step": 3476 + }, + { + "epoch": 0.081506355596969, + "grad_norm": 0.3630748987197876, + "learning_rate": 0.00019675661872937267, + "loss": 0.0965, + "step": 3477 + }, + { + "epoch": 0.081529797171774, + "grad_norm": 0.7894173860549927, + "learning_rate": 0.00019675475764544797, + "loss": 0.2181, + "step": 3478 + }, + { + "epoch": 0.081553238746579, + "grad_norm": 0.2812057435512543, + "learning_rate": 0.00019675289603652928, + "loss": 0.0268, + "step": 3479 + }, + { + "epoch": 0.08157668032138399, + "grad_norm": 0.5025647878646851, + "learning_rate": 0.0001967510339026266, + "loss": 0.1955, + "step": 3480 + }, + { + "epoch": 0.08160012189618898, + "grad_norm": 0.7690475583076477, + "learning_rate": 0.00019674917124375013, + "loss": 0.0939, + "step": 3481 + }, + { + "epoch": 0.08162356347099398, + "grad_norm": 0.45086589455604553, + "learning_rate": 0.00019674730805990996, + "loss": 0.1312, + "step": 3482 + }, + { + "epoch": 0.08164700504579897, + "grad_norm": 0.2599353790283203, + "learning_rate": 0.00019674544435111614, + "loss": 0.0583, + "step": 3483 + }, + { + "epoch": 0.08167044662060398, + "grad_norm": 0.8666149973869324, + "learning_rate": 0.00019674358011737881, + "loss": 0.2292, + "step": 3484 + }, + { + "epoch": 0.08169388819540897, + "grad_norm": 0.37658363580703735, + "learning_rate": 0.00019674171535870812, + "loss": 0.1029, + "step": 3485 + }, + { + "epoch": 0.08171732977021397, + "grad_norm": 0.5934263467788696, + "learning_rate": 0.00019673985007511414, + "loss": 0.108, + "step": 3486 + }, + { + "epoch": 0.08174077134501896, + "grad_norm": 0.7546018362045288, + "learning_rate": 0.00019673798426660704, + "loss": 0.1336, + "step": 3487 + }, + { + "epoch": 0.08176421291982396, + "grad_norm": 0.21827904880046844, + "learning_rate": 0.00019673611793319688, + "loss": 0.0516, + "step": 3488 + }, + { + "epoch": 0.08178765449462895, + "grad_norm": 0.9096226692199707, + "learning_rate": 0.00019673425107489387, + "loss": 0.2294, + "step": 3489 + }, + { + "epoch": 0.08181109606943394, + "grad_norm": 0.16302600502967834, + "learning_rate": 0.00019673238369170804, + "loss": 0.0533, + "step": 3490 + }, + { + "epoch": 0.08183453764423894, + "grad_norm": 0.4903559386730194, + "learning_rate": 0.00019673051578364965, + "loss": 0.1116, + "step": 3491 + }, + { + "epoch": 0.08185797921904393, + "grad_norm": 0.40027254819869995, + "learning_rate": 0.00019672864735072868, + "loss": 0.6812, + "step": 3492 + }, + { + "epoch": 0.08188142079384893, + "grad_norm": 0.6077809929847717, + "learning_rate": 0.0001967267783929554, + "loss": 0.1813, + "step": 3493 + }, + { + "epoch": 0.08190486236865392, + "grad_norm": 0.21274502575397491, + "learning_rate": 0.0001967249089103399, + "loss": 0.0336, + "step": 3494 + }, + { + "epoch": 0.08192830394345892, + "grad_norm": 0.6645164489746094, + "learning_rate": 0.00019672303890289232, + "loss": 0.1735, + "step": 3495 + }, + { + "epoch": 0.08195174551826391, + "grad_norm": 0.17424069344997406, + "learning_rate": 0.0001967211683706228, + "loss": 0.063, + "step": 3496 + }, + { + "epoch": 0.08197518709306892, + "grad_norm": 0.23965485394001007, + "learning_rate": 0.0001967192973135415, + "loss": 0.0725, + "step": 3497 + }, + { + "epoch": 0.08199862866787391, + "grad_norm": 0.937967836856842, + "learning_rate": 0.0001967174257316586, + "loss": 0.211, + "step": 3498 + }, + { + "epoch": 0.08202207024267891, + "grad_norm": 0.4304336905479431, + "learning_rate": 0.00019671555362498418, + "loss": 0.1499, + "step": 3499 + }, + { + "epoch": 0.0820455118174839, + "grad_norm": 0.5649681091308594, + "learning_rate": 0.0001967136809935285, + "loss": 0.2256, + "step": 3500 + }, + { + "epoch": 0.0820689533922889, + "grad_norm": 0.49253928661346436, + "learning_rate": 0.00019671180783730163, + "loss": 0.078, + "step": 3501 + }, + { + "epoch": 0.08209239496709389, + "grad_norm": 0.6492027044296265, + "learning_rate": 0.0001967099341563138, + "loss": 0.1333, + "step": 3502 + }, + { + "epoch": 0.08211583654189888, + "grad_norm": 0.7957311868667603, + "learning_rate": 0.0001967080599505751, + "loss": 0.2404, + "step": 3503 + }, + { + "epoch": 0.08213927811670388, + "grad_norm": 0.46836820244789124, + "learning_rate": 0.0001967061852200958, + "loss": 0.1854, + "step": 3504 + }, + { + "epoch": 0.08216271969150887, + "grad_norm": 0.5039299130439758, + "learning_rate": 0.00019670430996488597, + "loss": 0.1884, + "step": 3505 + }, + { + "epoch": 0.08218616126631387, + "grad_norm": 0.5670678019523621, + "learning_rate": 0.00019670243418495588, + "loss": 0.1866, + "step": 3506 + }, + { + "epoch": 0.08220960284111886, + "grad_norm": 0.6804062724113464, + "learning_rate": 0.00019670055788031566, + "loss": 0.6893, + "step": 3507 + }, + { + "epoch": 0.08223304441592386, + "grad_norm": 0.5273105502128601, + "learning_rate": 0.0001966986810509755, + "loss": 0.1649, + "step": 3508 + }, + { + "epoch": 0.08225648599072885, + "grad_norm": 0.1340843290090561, + "learning_rate": 0.00019669680369694553, + "loss": 0.0413, + "step": 3509 + }, + { + "epoch": 0.08227992756553386, + "grad_norm": 0.07686217129230499, + "learning_rate": 0.000196694925818236, + "loss": 0.0181, + "step": 3510 + }, + { + "epoch": 0.08230336914033885, + "grad_norm": 0.42379114031791687, + "learning_rate": 0.00019669304741485712, + "loss": 0.6237, + "step": 3511 + }, + { + "epoch": 0.08232681071514385, + "grad_norm": 0.3370814621448517, + "learning_rate": 0.000196691168486819, + "loss": 0.1398, + "step": 3512 + }, + { + "epoch": 0.08235025228994884, + "grad_norm": 0.4305917024612427, + "learning_rate": 0.00019668928903413191, + "loss": 0.173, + "step": 3513 + }, + { + "epoch": 0.08237369386475384, + "grad_norm": 0.4941258430480957, + "learning_rate": 0.000196687409056806, + "loss": 0.6494, + "step": 3514 + }, + { + "epoch": 0.08239713543955883, + "grad_norm": 0.5196722745895386, + "learning_rate": 0.00019668552855485154, + "loss": 0.1598, + "step": 3515 + }, + { + "epoch": 0.08242057701436382, + "grad_norm": 0.39713016152381897, + "learning_rate": 0.00019668364752827862, + "loss": 0.1604, + "step": 3516 + }, + { + "epoch": 0.08244401858916882, + "grad_norm": 0.6442750096321106, + "learning_rate": 0.00019668176597709754, + "loss": 0.1988, + "step": 3517 + }, + { + "epoch": 0.08246746016397381, + "grad_norm": 0.6172959208488464, + "learning_rate": 0.0001966798839013185, + "loss": 0.1745, + "step": 3518 + }, + { + "epoch": 0.08249090173877881, + "grad_norm": 0.6106070876121521, + "learning_rate": 0.00019667800130095163, + "loss": 0.1909, + "step": 3519 + }, + { + "epoch": 0.0825143433135838, + "grad_norm": 0.3249143362045288, + "learning_rate": 0.00019667611817600727, + "loss": 0.0312, + "step": 3520 + }, + { + "epoch": 0.0825377848883888, + "grad_norm": 0.6753798127174377, + "learning_rate": 0.00019667423452649554, + "loss": 0.2084, + "step": 3521 + }, + { + "epoch": 0.0825612264631938, + "grad_norm": 0.6426877975463867, + "learning_rate": 0.0001966723503524267, + "loss": 0.1921, + "step": 3522 + }, + { + "epoch": 0.0825846680379988, + "grad_norm": 0.7159570455551147, + "learning_rate": 0.00019667046565381097, + "loss": 0.1921, + "step": 3523 + }, + { + "epoch": 0.08260810961280379, + "grad_norm": 1.05588698387146, + "learning_rate": 0.00019666858043065858, + "loss": 0.2909, + "step": 3524 + }, + { + "epoch": 0.08263155118760879, + "grad_norm": 0.19869595766067505, + "learning_rate": 0.00019666669468297975, + "loss": 0.0388, + "step": 3525 + }, + { + "epoch": 0.08265499276241378, + "grad_norm": 0.40780898928642273, + "learning_rate": 0.00019666480841078472, + "loss": 0.2061, + "step": 3526 + }, + { + "epoch": 0.08267843433721878, + "grad_norm": 0.5423533916473389, + "learning_rate": 0.0001966629216140837, + "loss": 0.1608, + "step": 3527 + }, + { + "epoch": 0.08270187591202377, + "grad_norm": 0.685696542263031, + "learning_rate": 0.000196661034292887, + "loss": 0.6902, + "step": 3528 + }, + { + "epoch": 0.08272531748682876, + "grad_norm": 0.5680442452430725, + "learning_rate": 0.00019665914644720477, + "loss": 0.2093, + "step": 3529 + }, + { + "epoch": 0.08274875906163376, + "grad_norm": 0.3876146972179413, + "learning_rate": 0.0001966572580770473, + "loss": 0.1082, + "step": 3530 + }, + { + "epoch": 0.08277220063643875, + "grad_norm": 1.4398223161697388, + "learning_rate": 0.00019665536918242484, + "loss": 0.2864, + "step": 3531 + }, + { + "epoch": 0.08279564221124375, + "grad_norm": 0.861741840839386, + "learning_rate": 0.0001966534797633476, + "loss": 0.3685, + "step": 3532 + }, + { + "epoch": 0.08281908378604874, + "grad_norm": 0.7002941966056824, + "learning_rate": 0.0001966515898198259, + "loss": 0.2133, + "step": 3533 + }, + { + "epoch": 0.08284252536085374, + "grad_norm": 0.608247697353363, + "learning_rate": 0.00019664969935186995, + "loss": 0.1917, + "step": 3534 + }, + { + "epoch": 0.08286596693565874, + "grad_norm": 0.31718677282333374, + "learning_rate": 0.00019664780835948998, + "loss": 0.076, + "step": 3535 + }, + { + "epoch": 0.08288940851046374, + "grad_norm": 1.095902919769287, + "learning_rate": 0.0001966459168426963, + "loss": 0.1153, + "step": 3536 + }, + { + "epoch": 0.08291285008526873, + "grad_norm": 0.5310402512550354, + "learning_rate": 0.00019664402480149918, + "loss": 0.1754, + "step": 3537 + }, + { + "epoch": 0.08293629166007373, + "grad_norm": 0.5677681565284729, + "learning_rate": 0.00019664213223590885, + "loss": 0.1909, + "step": 3538 + }, + { + "epoch": 0.08295973323487872, + "grad_norm": 0.5517361164093018, + "learning_rate": 0.0001966402391459356, + "loss": 0.1855, + "step": 3539 + }, + { + "epoch": 0.08298317480968372, + "grad_norm": 0.6141646504402161, + "learning_rate": 0.00019663834553158968, + "loss": 0.7698, + "step": 3540 + }, + { + "epoch": 0.08300661638448871, + "grad_norm": 0.4132867157459259, + "learning_rate": 0.00019663645139288139, + "loss": 0.1554, + "step": 3541 + }, + { + "epoch": 0.0830300579592937, + "grad_norm": 0.6450992822647095, + "learning_rate": 0.000196634556729821, + "loss": 0.189, + "step": 3542 + }, + { + "epoch": 0.0830534995340987, + "grad_norm": 0.30153775215148926, + "learning_rate": 0.00019663266154241877, + "loss": 0.1229, + "step": 3543 + }, + { + "epoch": 0.08307694110890369, + "grad_norm": 0.272026389837265, + "learning_rate": 0.00019663076583068502, + "loss": 0.0735, + "step": 3544 + }, + { + "epoch": 0.08310038268370869, + "grad_norm": 0.4961797893047333, + "learning_rate": 0.00019662886959463002, + "loss": 0.2177, + "step": 3545 + }, + { + "epoch": 0.08312382425851368, + "grad_norm": 0.4575781524181366, + "learning_rate": 0.00019662697283426406, + "loss": 0.1187, + "step": 3546 + }, + { + "epoch": 0.08314726583331868, + "grad_norm": 0.09081536531448364, + "learning_rate": 0.0001966250755495974, + "loss": 0.0223, + "step": 3547 + }, + { + "epoch": 0.08317070740812368, + "grad_norm": 0.7216640114784241, + "learning_rate": 0.0001966231777406404, + "loss": 0.2025, + "step": 3548 + }, + { + "epoch": 0.08319414898292868, + "grad_norm": 0.6050297021865845, + "learning_rate": 0.0001966212794074033, + "loss": 0.2101, + "step": 3549 + }, + { + "epoch": 0.08321759055773367, + "grad_norm": 0.19247505068778992, + "learning_rate": 0.00019661938054989643, + "loss": 0.046, + "step": 3550 + }, + { + "epoch": 0.08324103213253867, + "grad_norm": 0.7009627819061279, + "learning_rate": 0.00019661748116813006, + "loss": 0.2274, + "step": 3551 + }, + { + "epoch": 0.08326447370734366, + "grad_norm": 0.7981820702552795, + "learning_rate": 0.00019661558126211453, + "loss": 0.2014, + "step": 3552 + }, + { + "epoch": 0.08328791528214866, + "grad_norm": 0.7736815810203552, + "learning_rate": 0.00019661368083186012, + "loss": 0.3486, + "step": 3553 + }, + { + "epoch": 0.08331135685695365, + "grad_norm": 0.45632922649383545, + "learning_rate": 0.00019661177987737721, + "loss": 0.1673, + "step": 3554 + }, + { + "epoch": 0.08333479843175864, + "grad_norm": 0.6731275916099548, + "learning_rate": 0.00019660987839867605, + "loss": 0.2252, + "step": 3555 + }, + { + "epoch": 0.08335824000656364, + "grad_norm": 0.470397412776947, + "learning_rate": 0.00019660797639576696, + "loss": 0.1728, + "step": 3556 + }, + { + "epoch": 0.08338168158136863, + "grad_norm": 0.26503366231918335, + "learning_rate": 0.0001966060738686603, + "loss": 0.0814, + "step": 3557 + }, + { + "epoch": 0.08340512315617363, + "grad_norm": 0.630397617816925, + "learning_rate": 0.00019660417081736632, + "loss": 0.226, + "step": 3558 + }, + { + "epoch": 0.08342856473097862, + "grad_norm": 0.4645870327949524, + "learning_rate": 0.00019660226724189543, + "loss": 0.5185, + "step": 3559 + }, + { + "epoch": 0.08345200630578362, + "grad_norm": 0.3771645724773407, + "learning_rate": 0.00019660036314225792, + "loss": 0.0769, + "step": 3560 + }, + { + "epoch": 0.08347544788058862, + "grad_norm": 0.3072541058063507, + "learning_rate": 0.0001965984585184641, + "loss": 0.0751, + "step": 3561 + }, + { + "epoch": 0.08349888945539362, + "grad_norm": 0.44199079275131226, + "learning_rate": 0.00019659655337052434, + "loss": 0.0808, + "step": 3562 + }, + { + "epoch": 0.08352233103019861, + "grad_norm": 0.24939262866973877, + "learning_rate": 0.000196594647698449, + "loss": 0.0699, + "step": 3563 + }, + { + "epoch": 0.08354577260500361, + "grad_norm": 0.441205233335495, + "learning_rate": 0.00019659274150224835, + "loss": 0.4208, + "step": 3564 + }, + { + "epoch": 0.0835692141798086, + "grad_norm": 0.6497467160224915, + "learning_rate": 0.00019659083478193275, + "loss": 0.1892, + "step": 3565 + }, + { + "epoch": 0.0835926557546136, + "grad_norm": 0.5880810618400574, + "learning_rate": 0.0001965889275375126, + "loss": 0.1587, + "step": 3566 + }, + { + "epoch": 0.08361609732941859, + "grad_norm": 0.5987945795059204, + "learning_rate": 0.0001965870197689982, + "loss": 0.2073, + "step": 3567 + }, + { + "epoch": 0.08363953890422358, + "grad_norm": 0.6534876227378845, + "learning_rate": 0.0001965851114763999, + "loss": 0.1202, + "step": 3568 + }, + { + "epoch": 0.08366298047902858, + "grad_norm": 0.3571818768978119, + "learning_rate": 0.0001965832026597281, + "loss": 0.0943, + "step": 3569 + }, + { + "epoch": 0.08368642205383357, + "grad_norm": 0.25961700081825256, + "learning_rate": 0.0001965812933189931, + "loss": 0.0722, + "step": 3570 + }, + { + "epoch": 0.08370986362863857, + "grad_norm": 0.7540719509124756, + "learning_rate": 0.0001965793834542053, + "loss": 0.2558, + "step": 3571 + }, + { + "epoch": 0.08373330520344356, + "grad_norm": 0.4919765889644623, + "learning_rate": 0.00019657747306537507, + "loss": 0.0823, + "step": 3572 + }, + { + "epoch": 0.08375674677824856, + "grad_norm": 0.6257020235061646, + "learning_rate": 0.0001965755621525127, + "loss": 0.0859, + "step": 3573 + }, + { + "epoch": 0.08378018835305356, + "grad_norm": 0.4458496868610382, + "learning_rate": 0.00019657365071562867, + "loss": 0.0813, + "step": 3574 + }, + { + "epoch": 0.08380362992785856, + "grad_norm": 0.45646533370018005, + "learning_rate": 0.00019657173875473326, + "loss": 0.5494, + "step": 3575 + }, + { + "epoch": 0.08382707150266355, + "grad_norm": 0.5075370073318481, + "learning_rate": 0.0001965698262698369, + "loss": 0.1242, + "step": 3576 + }, + { + "epoch": 0.08385051307746855, + "grad_norm": 0.5289875864982605, + "learning_rate": 0.00019656791326094997, + "loss": 0.1226, + "step": 3577 + }, + { + "epoch": 0.08387395465227354, + "grad_norm": 0.5985813736915588, + "learning_rate": 0.0001965659997280828, + "loss": 0.1772, + "step": 3578 + }, + { + "epoch": 0.08389739622707854, + "grad_norm": 0.4535789489746094, + "learning_rate": 0.00019656408567124577, + "loss": 0.1415, + "step": 3579 + }, + { + "epoch": 0.08392083780188353, + "grad_norm": 0.5983160734176636, + "learning_rate": 0.00019656217109044932, + "loss": 0.2362, + "step": 3580 + }, + { + "epoch": 0.08394427937668852, + "grad_norm": 0.6650323271751404, + "learning_rate": 0.00019656025598570383, + "loss": 0.2178, + "step": 3581 + }, + { + "epoch": 0.08396772095149352, + "grad_norm": 0.4328135550022125, + "learning_rate": 0.00019655834035701967, + "loss": 0.1163, + "step": 3582 + }, + { + "epoch": 0.08399116252629851, + "grad_norm": 0.1614588350057602, + "learning_rate": 0.00019655642420440724, + "loss": 0.0316, + "step": 3583 + }, + { + "epoch": 0.08401460410110351, + "grad_norm": 0.6280043721199036, + "learning_rate": 0.00019655450752787692, + "loss": 0.2, + "step": 3584 + }, + { + "epoch": 0.0840380456759085, + "grad_norm": 0.21352629363536835, + "learning_rate": 0.00019655259032743915, + "loss": 0.0404, + "step": 3585 + }, + { + "epoch": 0.08406148725071351, + "grad_norm": 0.3249623477458954, + "learning_rate": 0.0001965506726031043, + "loss": 0.0634, + "step": 3586 + }, + { + "epoch": 0.0840849288255185, + "grad_norm": 0.44545039534568787, + "learning_rate": 0.0001965487543548828, + "loss": 0.0923, + "step": 3587 + }, + { + "epoch": 0.0841083704003235, + "grad_norm": 0.575790286064148, + "learning_rate": 0.00019654683558278504, + "loss": 0.1168, + "step": 3588 + }, + { + "epoch": 0.08413181197512849, + "grad_norm": 0.8074888586997986, + "learning_rate": 0.00019654491628682142, + "loss": 0.5909, + "step": 3589 + }, + { + "epoch": 0.08415525354993349, + "grad_norm": 0.5595147013664246, + "learning_rate": 0.00019654299646700238, + "loss": 0.1911, + "step": 3590 + }, + { + "epoch": 0.08417869512473848, + "grad_norm": 0.68758225440979, + "learning_rate": 0.0001965410761233383, + "loss": 0.0976, + "step": 3591 + }, + { + "epoch": 0.08420213669954348, + "grad_norm": 0.5876624584197998, + "learning_rate": 0.00019653915525583965, + "loss": 0.2394, + "step": 3592 + }, + { + "epoch": 0.08422557827434847, + "grad_norm": 0.6713118553161621, + "learning_rate": 0.00019653723386451683, + "loss": 0.1723, + "step": 3593 + }, + { + "epoch": 0.08424901984915346, + "grad_norm": 0.528383731842041, + "learning_rate": 0.0001965353119493803, + "loss": 0.1153, + "step": 3594 + }, + { + "epoch": 0.08427246142395846, + "grad_norm": 0.7718040347099304, + "learning_rate": 0.00019653338951044038, + "loss": 0.2321, + "step": 3595 + }, + { + "epoch": 0.08429590299876345, + "grad_norm": 0.3366459012031555, + "learning_rate": 0.00019653146654770762, + "loss": 0.0794, + "step": 3596 + }, + { + "epoch": 0.08431934457356845, + "grad_norm": 0.607346773147583, + "learning_rate": 0.0001965295430611924, + "loss": 0.742, + "step": 3597 + }, + { + "epoch": 0.08434278614837344, + "grad_norm": 0.5420230031013489, + "learning_rate": 0.00019652761905090515, + "loss": 0.0871, + "step": 3598 + }, + { + "epoch": 0.08436622772317845, + "grad_norm": 0.4621739089488983, + "learning_rate": 0.00019652569451685637, + "loss": 0.1332, + "step": 3599 + }, + { + "epoch": 0.08438966929798344, + "grad_norm": 0.5955418944358826, + "learning_rate": 0.00019652376945905638, + "loss": 0.2746, + "step": 3600 + }, + { + "epoch": 0.08441311087278844, + "grad_norm": 0.6094779372215271, + "learning_rate": 0.00019652184387751578, + "loss": 0.1255, + "step": 3601 + }, + { + "epoch": 0.08443655244759343, + "grad_norm": 0.6725063920021057, + "learning_rate": 0.0001965199177722449, + "loss": 0.7339, + "step": 3602 + }, + { + "epoch": 0.08445999402239843, + "grad_norm": 0.6175028085708618, + "learning_rate": 0.00019651799114325425, + "loss": 0.1584, + "step": 3603 + }, + { + "epoch": 0.08448343559720342, + "grad_norm": 0.19221358001232147, + "learning_rate": 0.00019651606399055424, + "loss": 0.0543, + "step": 3604 + }, + { + "epoch": 0.08450687717200842, + "grad_norm": 0.9894387125968933, + "learning_rate": 0.0001965141363141554, + "loss": 0.2084, + "step": 3605 + }, + { + "epoch": 0.08453031874681341, + "grad_norm": 0.3512328863143921, + "learning_rate": 0.00019651220811406808, + "loss": 0.1212, + "step": 3606 + }, + { + "epoch": 0.0845537603216184, + "grad_norm": 0.4059489369392395, + "learning_rate": 0.00019651027939030286, + "loss": 0.5148, + "step": 3607 + }, + { + "epoch": 0.0845772018964234, + "grad_norm": 0.46366557478904724, + "learning_rate": 0.00019650835014287013, + "loss": 0.175, + "step": 3608 + }, + { + "epoch": 0.08460064347122839, + "grad_norm": 0.30651646852493286, + "learning_rate": 0.00019650642037178037, + "loss": 0.0912, + "step": 3609 + }, + { + "epoch": 0.08462408504603339, + "grad_norm": 0.3868820369243622, + "learning_rate": 0.0001965044900770441, + "loss": 0.4533, + "step": 3610 + }, + { + "epoch": 0.08464752662083838, + "grad_norm": 0.5182563662528992, + "learning_rate": 0.0001965025592586717, + "loss": 0.2038, + "step": 3611 + }, + { + "epoch": 0.08467096819564339, + "grad_norm": 0.5170130133628845, + "learning_rate": 0.0001965006279166737, + "loss": 0.1406, + "step": 3612 + }, + { + "epoch": 0.08469440977044838, + "grad_norm": 0.4310877025127411, + "learning_rate": 0.00019649869605106065, + "loss": 0.0626, + "step": 3613 + }, + { + "epoch": 0.08471785134525338, + "grad_norm": 0.574131965637207, + "learning_rate": 0.00019649676366184288, + "loss": 0.281, + "step": 3614 + }, + { + "epoch": 0.08474129292005837, + "grad_norm": 0.5104807615280151, + "learning_rate": 0.000196494830749031, + "loss": 0.1022, + "step": 3615 + }, + { + "epoch": 0.08476473449486337, + "grad_norm": 0.7585474252700806, + "learning_rate": 0.00019649289731263542, + "loss": 0.1255, + "step": 3616 + }, + { + "epoch": 0.08478817606966836, + "grad_norm": 0.2554749846458435, + "learning_rate": 0.00019649096335266672, + "loss": 0.0724, + "step": 3617 + }, + { + "epoch": 0.08481161764447336, + "grad_norm": 0.4699517786502838, + "learning_rate": 0.00019648902886913532, + "loss": 0.0941, + "step": 3618 + }, + { + "epoch": 0.08483505921927835, + "grad_norm": 0.6983384490013123, + "learning_rate": 0.00019648709386205172, + "loss": 0.2062, + "step": 3619 + }, + { + "epoch": 0.08485850079408334, + "grad_norm": 0.5241436958312988, + "learning_rate": 0.00019648515833142643, + "loss": 0.1125, + "step": 3620 + }, + { + "epoch": 0.08488194236888834, + "grad_norm": 0.19182226061820984, + "learning_rate": 0.00019648322227726998, + "loss": 0.0342, + "step": 3621 + }, + { + "epoch": 0.08490538394369333, + "grad_norm": 0.6182647943496704, + "learning_rate": 0.00019648128569959284, + "loss": 0.1547, + "step": 3622 + }, + { + "epoch": 0.08492882551849833, + "grad_norm": 0.3386211395263672, + "learning_rate": 0.00019647934859840554, + "loss": 0.1379, + "step": 3623 + }, + { + "epoch": 0.08495226709330332, + "grad_norm": 0.36476361751556396, + "learning_rate": 0.00019647741097371857, + "loss": 0.0788, + "step": 3624 + }, + { + "epoch": 0.08497570866810833, + "grad_norm": 1.117967963218689, + "learning_rate": 0.00019647547282554246, + "loss": 0.2767, + "step": 3625 + }, + { + "epoch": 0.08499915024291332, + "grad_norm": 0.6383253931999207, + "learning_rate": 0.00019647353415388775, + "loss": 0.243, + "step": 3626 + }, + { + "epoch": 0.08502259181771832, + "grad_norm": 0.526532769203186, + "learning_rate": 0.00019647159495876488, + "loss": 0.0764, + "step": 3627 + }, + { + "epoch": 0.08504603339252331, + "grad_norm": 0.6400803923606873, + "learning_rate": 0.00019646965524018444, + "loss": 0.1281, + "step": 3628 + }, + { + "epoch": 0.0850694749673283, + "grad_norm": 0.6436559557914734, + "learning_rate": 0.00019646771499815695, + "loss": 0.1887, + "step": 3629 + }, + { + "epoch": 0.0850929165421333, + "grad_norm": 0.44219738245010376, + "learning_rate": 0.00019646577423269293, + "loss": 0.1142, + "step": 3630 + }, + { + "epoch": 0.0851163581169383, + "grad_norm": 1.132869839668274, + "learning_rate": 0.00019646383294380287, + "loss": 0.2553, + "step": 3631 + }, + { + "epoch": 0.08513979969174329, + "grad_norm": 0.2674187123775482, + "learning_rate": 0.0001964618911314974, + "loss": 0.0661, + "step": 3632 + }, + { + "epoch": 0.08516324126654828, + "grad_norm": 0.15321479737758636, + "learning_rate": 0.00019645994879578696, + "loss": 0.0335, + "step": 3633 + }, + { + "epoch": 0.08518668284135328, + "grad_norm": 0.9375066161155701, + "learning_rate": 0.00019645800593668213, + "loss": 0.2262, + "step": 3634 + }, + { + "epoch": 0.08521012441615827, + "grad_norm": 0.9779962301254272, + "learning_rate": 0.00019645606255419345, + "loss": 0.1593, + "step": 3635 + }, + { + "epoch": 0.08523356599096327, + "grad_norm": 0.8032666444778442, + "learning_rate": 0.0001964541186483315, + "loss": 0.2162, + "step": 3636 + }, + { + "epoch": 0.08525700756576828, + "grad_norm": 0.7826479077339172, + "learning_rate": 0.0001964521742191067, + "loss": 0.2043, + "step": 3637 + }, + { + "epoch": 0.08528044914057327, + "grad_norm": 0.5933001041412354, + "learning_rate": 0.00019645022926652973, + "loss": 0.1919, + "step": 3638 + }, + { + "epoch": 0.08530389071537826, + "grad_norm": 0.2845117151737213, + "learning_rate": 0.00019644828379061116, + "loss": 0.0886, + "step": 3639 + }, + { + "epoch": 0.08532733229018326, + "grad_norm": 0.6054075956344604, + "learning_rate": 0.00019644633779136142, + "loss": 0.1991, + "step": 3640 + }, + { + "epoch": 0.08535077386498825, + "grad_norm": 0.44253358244895935, + "learning_rate": 0.00019644439126879118, + "loss": 0.1614, + "step": 3641 + }, + { + "epoch": 0.08537421543979325, + "grad_norm": 0.3644881546497345, + "learning_rate": 0.00019644244422291094, + "loss": 0.0772, + "step": 3642 + }, + { + "epoch": 0.08539765701459824, + "grad_norm": 0.2735404670238495, + "learning_rate": 0.0001964404966537313, + "loss": 0.0985, + "step": 3643 + }, + { + "epoch": 0.08542109858940324, + "grad_norm": 0.40403687953948975, + "learning_rate": 0.0001964385485612628, + "loss": 0.1369, + "step": 3644 + }, + { + "epoch": 0.08544454016420823, + "grad_norm": 0.16527056694030762, + "learning_rate": 0.00019643659994551604, + "loss": 0.0444, + "step": 3645 + }, + { + "epoch": 0.08546798173901322, + "grad_norm": 0.5183283090591431, + "learning_rate": 0.00019643465080650154, + "loss": 0.1374, + "step": 3646 + }, + { + "epoch": 0.08549142331381822, + "grad_norm": 0.6067399978637695, + "learning_rate": 0.00019643270114422996, + "loss": 0.8177, + "step": 3647 + }, + { + "epoch": 0.08551486488862321, + "grad_norm": 0.9288958311080933, + "learning_rate": 0.00019643075095871182, + "loss": 0.2701, + "step": 3648 + }, + { + "epoch": 0.08553830646342821, + "grad_norm": 0.25887709856033325, + "learning_rate": 0.0001964288002499577, + "loss": 0.0775, + "step": 3649 + }, + { + "epoch": 0.08556174803823322, + "grad_norm": 0.8164992928504944, + "learning_rate": 0.0001964268490179782, + "loss": 0.2207, + "step": 3650 + }, + { + "epoch": 0.08558518961303821, + "grad_norm": 0.46519458293914795, + "learning_rate": 0.00019642489726278393, + "loss": 0.1778, + "step": 3651 + }, + { + "epoch": 0.0856086311878432, + "grad_norm": 0.7260604500770569, + "learning_rate": 0.00019642294498438542, + "loss": 0.5727, + "step": 3652 + }, + { + "epoch": 0.0856320727626482, + "grad_norm": 0.5469567179679871, + "learning_rate": 0.0001964209921827933, + "loss": 0.1354, + "step": 3653 + }, + { + "epoch": 0.08565551433745319, + "grad_norm": 0.10462471842765808, + "learning_rate": 0.0001964190388580182, + "loss": 0.0294, + "step": 3654 + }, + { + "epoch": 0.08567895591225819, + "grad_norm": 0.4233132600784302, + "learning_rate": 0.00019641708501007065, + "loss": 0.121, + "step": 3655 + }, + { + "epoch": 0.08570239748706318, + "grad_norm": 0.6424537301063538, + "learning_rate": 0.00019641513063896131, + "loss": 0.9127, + "step": 3656 + }, + { + "epoch": 0.08572583906186818, + "grad_norm": 0.27915674448013306, + "learning_rate": 0.00019641317574470074, + "loss": 0.1413, + "step": 3657 + }, + { + "epoch": 0.08574928063667317, + "grad_norm": 0.3270205855369568, + "learning_rate": 0.0001964112203272996, + "loss": 0.1039, + "step": 3658 + }, + { + "epoch": 0.08577272221147816, + "grad_norm": 0.365887850522995, + "learning_rate": 0.00019640926438676846, + "loss": 0.0801, + "step": 3659 + }, + { + "epoch": 0.08579616378628316, + "grad_norm": 0.2698613405227661, + "learning_rate": 0.0001964073079231179, + "loss": 0.0882, + "step": 3660 + }, + { + "epoch": 0.08581960536108815, + "grad_norm": 0.802311897277832, + "learning_rate": 0.00019640535093635864, + "loss": 0.1445, + "step": 3661 + }, + { + "epoch": 0.08584304693589315, + "grad_norm": 0.4196336567401886, + "learning_rate": 0.00019640339342650117, + "loss": 0.1334, + "step": 3662 + }, + { + "epoch": 0.08586648851069816, + "grad_norm": 1.0428255796432495, + "learning_rate": 0.0001964014353935562, + "loss": 0.2444, + "step": 3663 + }, + { + "epoch": 0.08588993008550315, + "grad_norm": 0.3549928665161133, + "learning_rate": 0.00019639947683753434, + "loss": 0.0775, + "step": 3664 + }, + { + "epoch": 0.08591337166030814, + "grad_norm": 0.2204071581363678, + "learning_rate": 0.0001963975177584462, + "loss": 0.0425, + "step": 3665 + }, + { + "epoch": 0.08593681323511314, + "grad_norm": 0.4537513554096222, + "learning_rate": 0.00019639555815630243, + "loss": 0.1155, + "step": 3666 + }, + { + "epoch": 0.08596025480991813, + "grad_norm": 0.5085839629173279, + "learning_rate": 0.00019639359803111365, + "loss": 0.1173, + "step": 3667 + }, + { + "epoch": 0.08598369638472313, + "grad_norm": 0.6055182218551636, + "learning_rate": 0.0001963916373828905, + "loss": 0.1577, + "step": 3668 + }, + { + "epoch": 0.08600713795952812, + "grad_norm": 0.559695303440094, + "learning_rate": 0.00019638967621164358, + "loss": 0.1688, + "step": 3669 + }, + { + "epoch": 0.08603057953433312, + "grad_norm": 0.6908040642738342, + "learning_rate": 0.00019638771451738362, + "loss": 0.1379, + "step": 3670 + }, + { + "epoch": 0.08605402110913811, + "grad_norm": 0.4274834394454956, + "learning_rate": 0.00019638575230012116, + "loss": 0.1195, + "step": 3671 + }, + { + "epoch": 0.0860774626839431, + "grad_norm": 0.8888225555419922, + "learning_rate": 0.00019638378955986693, + "loss": 0.3256, + "step": 3672 + }, + { + "epoch": 0.0861009042587481, + "grad_norm": 0.5337581634521484, + "learning_rate": 0.00019638182629663154, + "loss": 0.1388, + "step": 3673 + }, + { + "epoch": 0.08612434583355309, + "grad_norm": 0.7338758707046509, + "learning_rate": 0.00019637986251042565, + "loss": 0.1581, + "step": 3674 + }, + { + "epoch": 0.08614778740835809, + "grad_norm": 0.8152349591255188, + "learning_rate": 0.00019637789820125994, + "loss": 0.3031, + "step": 3675 + }, + { + "epoch": 0.0861712289831631, + "grad_norm": 0.6984524130821228, + "learning_rate": 0.000196375933369145, + "loss": 0.1304, + "step": 3676 + }, + { + "epoch": 0.08619467055796809, + "grad_norm": 0.5661036372184753, + "learning_rate": 0.00019637396801409156, + "loss": 0.1316, + "step": 3677 + }, + { + "epoch": 0.08621811213277308, + "grad_norm": 0.598730742931366, + "learning_rate": 0.00019637200213611027, + "loss": 0.1009, + "step": 3678 + }, + { + "epoch": 0.08624155370757808, + "grad_norm": 0.8464269638061523, + "learning_rate": 0.00019637003573521177, + "loss": 0.222, + "step": 3679 + }, + { + "epoch": 0.08626499528238307, + "grad_norm": 0.6316831707954407, + "learning_rate": 0.00019636806881140676, + "loss": 0.1356, + "step": 3680 + }, + { + "epoch": 0.08628843685718807, + "grad_norm": 0.43410977721214294, + "learning_rate": 0.0001963661013647059, + "loss": 0.1133, + "step": 3681 + }, + { + "epoch": 0.08631187843199306, + "grad_norm": 0.5316421985626221, + "learning_rate": 0.00019636413339511983, + "loss": 0.1495, + "step": 3682 + }, + { + "epoch": 0.08633532000679806, + "grad_norm": 0.6172183156013489, + "learning_rate": 0.0001963621649026593, + "loss": 0.2189, + "step": 3683 + }, + { + "epoch": 0.08635876158160305, + "grad_norm": 0.11752957105636597, + "learning_rate": 0.00019636019588733491, + "loss": 0.0353, + "step": 3684 + }, + { + "epoch": 0.08638220315640804, + "grad_norm": 0.5068826079368591, + "learning_rate": 0.00019635822634915743, + "loss": 0.1778, + "step": 3685 + }, + { + "epoch": 0.08640564473121304, + "grad_norm": 0.7001904845237732, + "learning_rate": 0.00019635625628813748, + "loss": 0.1637, + "step": 3686 + }, + { + "epoch": 0.08642908630601803, + "grad_norm": 0.5955492258071899, + "learning_rate": 0.0001963542857042858, + "loss": 0.8129, + "step": 3687 + }, + { + "epoch": 0.08645252788082303, + "grad_norm": 0.5354343056678772, + "learning_rate": 0.00019635231459761302, + "loss": 0.1056, + "step": 3688 + }, + { + "epoch": 0.08647596945562804, + "grad_norm": 0.18083281815052032, + "learning_rate": 0.00019635034296812992, + "loss": 0.0442, + "step": 3689 + }, + { + "epoch": 0.08649941103043303, + "grad_norm": 0.588659405708313, + "learning_rate": 0.00019634837081584713, + "loss": 0.1023, + "step": 3690 + }, + { + "epoch": 0.08652285260523802, + "grad_norm": 0.7591397762298584, + "learning_rate": 0.00019634639814077533, + "loss": 0.188, + "step": 3691 + }, + { + "epoch": 0.08654629418004302, + "grad_norm": 0.3163873553276062, + "learning_rate": 0.00019634442494292532, + "loss": 0.0661, + "step": 3692 + }, + { + "epoch": 0.08656973575484801, + "grad_norm": 0.356754332780838, + "learning_rate": 0.0001963424512223077, + "loss": 0.086, + "step": 3693 + }, + { + "epoch": 0.086593177329653, + "grad_norm": 0.39890429377555847, + "learning_rate": 0.0001963404769789333, + "loss": 0.0773, + "step": 3694 + }, + { + "epoch": 0.086616618904458, + "grad_norm": 0.7570562362670898, + "learning_rate": 0.0001963385022128127, + "loss": 0.2409, + "step": 3695 + }, + { + "epoch": 0.086640060479263, + "grad_norm": 1.1111217737197876, + "learning_rate": 0.00019633652692395672, + "loss": 0.2142, + "step": 3696 + }, + { + "epoch": 0.08666350205406799, + "grad_norm": 0.12299273163080215, + "learning_rate": 0.00019633455111237603, + "loss": 0.0233, + "step": 3697 + }, + { + "epoch": 0.08668694362887298, + "grad_norm": 0.2513671815395355, + "learning_rate": 0.00019633257477808132, + "loss": 0.0536, + "step": 3698 + }, + { + "epoch": 0.08671038520367798, + "grad_norm": 0.5504773855209351, + "learning_rate": 0.0001963305979210834, + "loss": 0.1556, + "step": 3699 + }, + { + "epoch": 0.08673382677848297, + "grad_norm": 0.13693545758724213, + "learning_rate": 0.0001963286205413929, + "loss": 0.0422, + "step": 3700 + }, + { + "epoch": 0.08675726835328798, + "grad_norm": 0.925342321395874, + "learning_rate": 0.00019632664263902064, + "loss": 0.2528, + "step": 3701 + }, + { + "epoch": 0.08678070992809298, + "grad_norm": 0.6835268139839172, + "learning_rate": 0.00019632466421397728, + "loss": 0.6512, + "step": 3702 + }, + { + "epoch": 0.08680415150289797, + "grad_norm": 0.3060624301433563, + "learning_rate": 0.00019632268526627357, + "loss": 0.0691, + "step": 3703 + }, + { + "epoch": 0.08682759307770296, + "grad_norm": 0.7371663451194763, + "learning_rate": 0.0001963207057959203, + "loss": 0.8796, + "step": 3704 + }, + { + "epoch": 0.08685103465250796, + "grad_norm": 0.6567893624305725, + "learning_rate": 0.00019631872580292816, + "loss": 0.1336, + "step": 3705 + }, + { + "epoch": 0.08687447622731295, + "grad_norm": 0.6672714948654175, + "learning_rate": 0.0001963167452873079, + "loss": 0.1831, + "step": 3706 + }, + { + "epoch": 0.08689791780211795, + "grad_norm": 0.5877408385276794, + "learning_rate": 0.00019631476424907028, + "loss": 0.1175, + "step": 3707 + }, + { + "epoch": 0.08692135937692294, + "grad_norm": 0.45312851667404175, + "learning_rate": 0.00019631278268822603, + "loss": 0.0976, + "step": 3708 + }, + { + "epoch": 0.08694480095172794, + "grad_norm": 0.4838156998157501, + "learning_rate": 0.0001963108006047859, + "loss": 0.1268, + "step": 3709 + }, + { + "epoch": 0.08696824252653293, + "grad_norm": 0.2152417153120041, + "learning_rate": 0.00019630881799876068, + "loss": 0.0778, + "step": 3710 + }, + { + "epoch": 0.08699168410133792, + "grad_norm": 0.6791559457778931, + "learning_rate": 0.0001963068348701611, + "loss": 0.1936, + "step": 3711 + }, + { + "epoch": 0.08701512567614292, + "grad_norm": 0.4110121726989746, + "learning_rate": 0.00019630485121899794, + "loss": 0.0971, + "step": 3712 + }, + { + "epoch": 0.08703856725094791, + "grad_norm": 0.6953498721122742, + "learning_rate": 0.0001963028670452819, + "loss": 0.2601, + "step": 3713 + }, + { + "epoch": 0.08706200882575292, + "grad_norm": 0.540804922580719, + "learning_rate": 0.00019630088234902387, + "loss": 0.1197, + "step": 3714 + }, + { + "epoch": 0.08708545040055791, + "grad_norm": 0.6310924887657166, + "learning_rate": 0.00019629889713023452, + "loss": 0.175, + "step": 3715 + }, + { + "epoch": 0.08710889197536291, + "grad_norm": 0.5881400108337402, + "learning_rate": 0.0001962969113889246, + "loss": 0.1753, + "step": 3716 + }, + { + "epoch": 0.0871323335501679, + "grad_norm": 0.36434054374694824, + "learning_rate": 0.00019629492512510496, + "loss": 0.3943, + "step": 3717 + }, + { + "epoch": 0.0871557751249729, + "grad_norm": 0.6783537268638611, + "learning_rate": 0.00019629293833878637, + "loss": 0.1562, + "step": 3718 + }, + { + "epoch": 0.08717921669977789, + "grad_norm": 0.49052682518959045, + "learning_rate": 0.00019629095102997957, + "loss": 0.1257, + "step": 3719 + }, + { + "epoch": 0.08720265827458289, + "grad_norm": 0.17171543836593628, + "learning_rate": 0.00019628896319869537, + "loss": 0.0549, + "step": 3720 + }, + { + "epoch": 0.08722609984938788, + "grad_norm": 0.2892797887325287, + "learning_rate": 0.0001962869748449445, + "loss": 0.06, + "step": 3721 + }, + { + "epoch": 0.08724954142419288, + "grad_norm": 0.24773336946964264, + "learning_rate": 0.00019628498596873785, + "loss": 0.0628, + "step": 3722 + }, + { + "epoch": 0.08727298299899787, + "grad_norm": 0.8446877598762512, + "learning_rate": 0.00019628299657008618, + "loss": 0.728, + "step": 3723 + }, + { + "epoch": 0.08729642457380286, + "grad_norm": 0.4816908836364746, + "learning_rate": 0.00019628100664900023, + "loss": 0.1595, + "step": 3724 + }, + { + "epoch": 0.08731986614860786, + "grad_norm": 0.268914133310318, + "learning_rate": 0.00019627901620549083, + "loss": 0.0572, + "step": 3725 + }, + { + "epoch": 0.08734330772341285, + "grad_norm": 0.5174204707145691, + "learning_rate": 0.00019627702523956878, + "loss": 0.7965, + "step": 3726 + }, + { + "epoch": 0.08736674929821786, + "grad_norm": 0.7388238906860352, + "learning_rate": 0.00019627503375124487, + "loss": 0.1483, + "step": 3727 + }, + { + "epoch": 0.08739019087302285, + "grad_norm": 0.9473311305046082, + "learning_rate": 0.00019627304174052993, + "loss": 0.2138, + "step": 3728 + }, + { + "epoch": 0.08741363244782785, + "grad_norm": 0.5183023810386658, + "learning_rate": 0.0001962710492074348, + "loss": 0.1286, + "step": 3729 + }, + { + "epoch": 0.08743707402263284, + "grad_norm": 0.5190600752830505, + "learning_rate": 0.0001962690561519702, + "loss": 0.2572, + "step": 3730 + }, + { + "epoch": 0.08746051559743784, + "grad_norm": 0.6309943795204163, + "learning_rate": 0.000196267062574147, + "loss": 0.6739, + "step": 3731 + }, + { + "epoch": 0.08748395717224283, + "grad_norm": 0.6624492406845093, + "learning_rate": 0.00019626506847397602, + "loss": 0.112, + "step": 3732 + }, + { + "epoch": 0.08750739874704783, + "grad_norm": 0.8715534210205078, + "learning_rate": 0.0001962630738514681, + "loss": 0.2049, + "step": 3733 + }, + { + "epoch": 0.08753084032185282, + "grad_norm": 0.49710679054260254, + "learning_rate": 0.000196261078706634, + "loss": 0.2309, + "step": 3734 + }, + { + "epoch": 0.08755428189665782, + "grad_norm": 0.7883541584014893, + "learning_rate": 0.00019625908303948458, + "loss": 0.1996, + "step": 3735 + }, + { + "epoch": 0.08757772347146281, + "grad_norm": 0.17439058423042297, + "learning_rate": 0.00019625708685003067, + "loss": 0.0482, + "step": 3736 + }, + { + "epoch": 0.0876011650462678, + "grad_norm": 0.7934548258781433, + "learning_rate": 0.00019625509013828312, + "loss": 0.1928, + "step": 3737 + }, + { + "epoch": 0.0876246066210728, + "grad_norm": 0.6327301263809204, + "learning_rate": 0.00019625309290425272, + "loss": 0.2552, + "step": 3738 + }, + { + "epoch": 0.08764804819587779, + "grad_norm": 0.39915505051612854, + "learning_rate": 0.00019625109514795035, + "loss": 0.106, + "step": 3739 + }, + { + "epoch": 0.0876714897706828, + "grad_norm": 0.5784162282943726, + "learning_rate": 0.00019624909686938682, + "loss": 0.1077, + "step": 3740 + }, + { + "epoch": 0.0876949313454878, + "grad_norm": 0.6843301057815552, + "learning_rate": 0.00019624709806857296, + "loss": 0.3007, + "step": 3741 + }, + { + "epoch": 0.08771837292029279, + "grad_norm": 0.3016583025455475, + "learning_rate": 0.00019624509874551968, + "loss": 0.0748, + "step": 3742 + }, + { + "epoch": 0.08774181449509778, + "grad_norm": 0.14349883794784546, + "learning_rate": 0.0001962430989002378, + "loss": 0.0334, + "step": 3743 + }, + { + "epoch": 0.08776525606990278, + "grad_norm": 0.4883895218372345, + "learning_rate": 0.00019624109853273814, + "loss": 0.1649, + "step": 3744 + }, + { + "epoch": 0.08778869764470777, + "grad_norm": 0.41845136880874634, + "learning_rate": 0.00019623909764303156, + "loss": 0.1519, + "step": 3745 + }, + { + "epoch": 0.08781213921951277, + "grad_norm": 0.555961549282074, + "learning_rate": 0.00019623709623112893, + "loss": 0.1711, + "step": 3746 + }, + { + "epoch": 0.08783558079431776, + "grad_norm": 0.41253435611724854, + "learning_rate": 0.00019623509429704116, + "loss": 0.1225, + "step": 3747 + }, + { + "epoch": 0.08785902236912276, + "grad_norm": 0.21144035458564758, + "learning_rate": 0.00019623309184077902, + "loss": 0.0526, + "step": 3748 + }, + { + "epoch": 0.08788246394392775, + "grad_norm": 0.17034028470516205, + "learning_rate": 0.0001962310888623534, + "loss": 0.0539, + "step": 3749 + }, + { + "epoch": 0.08790590551873274, + "grad_norm": 0.5020045042037964, + "learning_rate": 0.0001962290853617752, + "loss": 0.177, + "step": 3750 + }, + { + "epoch": 0.08792934709353774, + "grad_norm": 0.5404404997825623, + "learning_rate": 0.00019622708133905527, + "loss": 0.1473, + "step": 3751 + }, + { + "epoch": 0.08795278866834273, + "grad_norm": 0.4638057351112366, + "learning_rate": 0.00019622507679420453, + "loss": 0.1239, + "step": 3752 + }, + { + "epoch": 0.08797623024314774, + "grad_norm": 0.570706307888031, + "learning_rate": 0.00019622307172723378, + "loss": 0.1028, + "step": 3753 + }, + { + "epoch": 0.08799967181795273, + "grad_norm": 0.5857737064361572, + "learning_rate": 0.00019622106613815393, + "loss": 0.1325, + "step": 3754 + }, + { + "epoch": 0.08802311339275773, + "grad_norm": 0.6873630881309509, + "learning_rate": 0.0001962190600269759, + "loss": 0.1907, + "step": 3755 + }, + { + "epoch": 0.08804655496756272, + "grad_norm": 0.7261998057365417, + "learning_rate": 0.00019621705339371054, + "loss": 0.2005, + "step": 3756 + }, + { + "epoch": 0.08806999654236772, + "grad_norm": 0.3530808389186859, + "learning_rate": 0.00019621504623836873, + "loss": 0.0857, + "step": 3757 + }, + { + "epoch": 0.08809343811717271, + "grad_norm": 0.3936951160430908, + "learning_rate": 0.00019621303856096138, + "loss": 0.1304, + "step": 3758 + }, + { + "epoch": 0.0881168796919777, + "grad_norm": 0.2699289917945862, + "learning_rate": 0.00019621103036149938, + "loss": 0.0369, + "step": 3759 + }, + { + "epoch": 0.0881403212667827, + "grad_norm": 0.680974006652832, + "learning_rate": 0.00019620902163999362, + "loss": 0.7272, + "step": 3760 + }, + { + "epoch": 0.0881637628415877, + "grad_norm": 0.7278835773468018, + "learning_rate": 0.000196207012396455, + "loss": 0.189, + "step": 3761 + }, + { + "epoch": 0.08818720441639269, + "grad_norm": 0.3095150291919708, + "learning_rate": 0.00019620500263089444, + "loss": 0.0853, + "step": 3762 + }, + { + "epoch": 0.08821064599119768, + "grad_norm": 0.5108271837234497, + "learning_rate": 0.0001962029923433228, + "loss": 0.17, + "step": 3763 + }, + { + "epoch": 0.08823408756600268, + "grad_norm": 0.5242434740066528, + "learning_rate": 0.00019620098153375106, + "loss": 0.1787, + "step": 3764 + }, + { + "epoch": 0.08825752914080769, + "grad_norm": 0.5796137452125549, + "learning_rate": 0.00019619897020219007, + "loss": 0.18, + "step": 3765 + }, + { + "epoch": 0.08828097071561268, + "grad_norm": 0.7114071846008301, + "learning_rate": 0.00019619695834865075, + "loss": 0.1389, + "step": 3766 + }, + { + "epoch": 0.08830441229041767, + "grad_norm": 0.6126572489738464, + "learning_rate": 0.00019619494597314404, + "loss": 0.1132, + "step": 3767 + }, + { + "epoch": 0.08832785386522267, + "grad_norm": 0.7096579074859619, + "learning_rate": 0.00019619293307568086, + "loss": 0.6977, + "step": 3768 + }, + { + "epoch": 0.08835129544002766, + "grad_norm": 0.7444016337394714, + "learning_rate": 0.0001961909196562721, + "loss": 0.1816, + "step": 3769 + }, + { + "epoch": 0.08837473701483266, + "grad_norm": 0.2983238101005554, + "learning_rate": 0.00019618890571492873, + "loss": 0.0347, + "step": 3770 + }, + { + "epoch": 0.08839817858963765, + "grad_norm": 0.4385242760181427, + "learning_rate": 0.00019618689125166163, + "loss": 0.1037, + "step": 3771 + }, + { + "epoch": 0.08842162016444265, + "grad_norm": 0.6797798871994019, + "learning_rate": 0.00019618487626648176, + "loss": 0.2068, + "step": 3772 + }, + { + "epoch": 0.08844506173924764, + "grad_norm": 0.2852810323238373, + "learning_rate": 0.00019618286075940005, + "loss": 0.1267, + "step": 3773 + }, + { + "epoch": 0.08846850331405264, + "grad_norm": 0.8879359364509583, + "learning_rate": 0.0001961808447304274, + "loss": 0.2637, + "step": 3774 + }, + { + "epoch": 0.08849194488885763, + "grad_norm": 0.3014128804206848, + "learning_rate": 0.00019617882817957484, + "loss": 0.0835, + "step": 3775 + }, + { + "epoch": 0.08851538646366262, + "grad_norm": 0.21020640432834625, + "learning_rate": 0.00019617681110685323, + "loss": 0.0472, + "step": 3776 + }, + { + "epoch": 0.08853882803846762, + "grad_norm": 0.518513023853302, + "learning_rate": 0.00019617479351227352, + "loss": 0.0988, + "step": 3777 + }, + { + "epoch": 0.08856226961327263, + "grad_norm": 0.41369524598121643, + "learning_rate": 0.0001961727753958467, + "loss": 0.1814, + "step": 3778 + }, + { + "epoch": 0.08858571118807762, + "grad_norm": 0.31183457374572754, + "learning_rate": 0.0001961707567575837, + "loss": 0.1074, + "step": 3779 + }, + { + "epoch": 0.08860915276288261, + "grad_norm": 0.5478755831718445, + "learning_rate": 0.00019616873759749546, + "loss": 0.1078, + "step": 3780 + }, + { + "epoch": 0.08863259433768761, + "grad_norm": 0.4501276910305023, + "learning_rate": 0.00019616671791559295, + "loss": 0.0872, + "step": 3781 + }, + { + "epoch": 0.0886560359124926, + "grad_norm": 0.23655542731285095, + "learning_rate": 0.00019616469771188713, + "loss": 0.0415, + "step": 3782 + }, + { + "epoch": 0.0886794774872976, + "grad_norm": 0.5594651103019714, + "learning_rate": 0.00019616267698638893, + "loss": 0.1473, + "step": 3783 + }, + { + "epoch": 0.08870291906210259, + "grad_norm": 0.16683600842952728, + "learning_rate": 0.00019616065573910935, + "loss": 0.033, + "step": 3784 + }, + { + "epoch": 0.08872636063690759, + "grad_norm": 0.40766724944114685, + "learning_rate": 0.00019615863397005935, + "loss": 0.0851, + "step": 3785 + }, + { + "epoch": 0.08874980221171258, + "grad_norm": 0.19906799495220184, + "learning_rate": 0.0001961566116792499, + "loss": 0.0466, + "step": 3786 + }, + { + "epoch": 0.08877324378651757, + "grad_norm": 0.6150079965591431, + "learning_rate": 0.000196154588866692, + "loss": 0.172, + "step": 3787 + }, + { + "epoch": 0.08879668536132257, + "grad_norm": 0.8610079884529114, + "learning_rate": 0.00019615256553239655, + "loss": 0.2468, + "step": 3788 + }, + { + "epoch": 0.08882012693612756, + "grad_norm": 0.21751046180725098, + "learning_rate": 0.00019615054167637462, + "loss": 0.0592, + "step": 3789 + }, + { + "epoch": 0.08884356851093256, + "grad_norm": 0.2298552691936493, + "learning_rate": 0.0001961485172986371, + "loss": 0.0415, + "step": 3790 + }, + { + "epoch": 0.08886701008573757, + "grad_norm": 0.3569745123386383, + "learning_rate": 0.00019614649239919506, + "loss": 0.1293, + "step": 3791 + }, + { + "epoch": 0.08889045166054256, + "grad_norm": 0.5843790769577026, + "learning_rate": 0.00019614446697805945, + "loss": 0.2202, + "step": 3792 + }, + { + "epoch": 0.08891389323534755, + "grad_norm": 0.1328418105840683, + "learning_rate": 0.0001961424410352412, + "loss": 0.051, + "step": 3793 + }, + { + "epoch": 0.08893733481015255, + "grad_norm": 0.5946836471557617, + "learning_rate": 0.00019614041457075143, + "loss": 0.8358, + "step": 3794 + }, + { + "epoch": 0.08896077638495754, + "grad_norm": 0.3903377950191498, + "learning_rate": 0.00019613838758460102, + "loss": 0.127, + "step": 3795 + }, + { + "epoch": 0.08898421795976254, + "grad_norm": 0.2474740594625473, + "learning_rate": 0.00019613636007680105, + "loss": 0.0367, + "step": 3796 + }, + { + "epoch": 0.08900765953456753, + "grad_norm": 0.41711848974227905, + "learning_rate": 0.0001961343320473625, + "loss": 0.1095, + "step": 3797 + }, + { + "epoch": 0.08903110110937253, + "grad_norm": 0.4983327090740204, + "learning_rate": 0.00019613230349629633, + "loss": 0.2408, + "step": 3798 + }, + { + "epoch": 0.08905454268417752, + "grad_norm": 0.7692375779151917, + "learning_rate": 0.0001961302744236136, + "loss": 0.2146, + "step": 3799 + }, + { + "epoch": 0.08907798425898251, + "grad_norm": 0.3181765675544739, + "learning_rate": 0.00019612824482932525, + "loss": 0.0801, + "step": 3800 + }, + { + "epoch": 0.08910142583378751, + "grad_norm": 0.6340170502662659, + "learning_rate": 0.0001961262147134424, + "loss": 0.1657, + "step": 3801 + }, + { + "epoch": 0.0891248674085925, + "grad_norm": 0.5742592215538025, + "learning_rate": 0.00019612418407597595, + "loss": 0.3091, + "step": 3802 + }, + { + "epoch": 0.0891483089833975, + "grad_norm": 0.2315850406885147, + "learning_rate": 0.00019612215291693702, + "loss": 0.0286, + "step": 3803 + }, + { + "epoch": 0.0891717505582025, + "grad_norm": 0.7993977665901184, + "learning_rate": 0.00019612012123633658, + "loss": 0.1511, + "step": 3804 + }, + { + "epoch": 0.0891951921330075, + "grad_norm": 0.3913721442222595, + "learning_rate": 0.00019611808903418566, + "loss": 0.2277, + "step": 3805 + }, + { + "epoch": 0.0892186337078125, + "grad_norm": 0.517299473285675, + "learning_rate": 0.00019611605631049525, + "loss": 0.1354, + "step": 3806 + }, + { + "epoch": 0.08924207528261749, + "grad_norm": 0.4866601526737213, + "learning_rate": 0.00019611402306527644, + "loss": 0.1764, + "step": 3807 + }, + { + "epoch": 0.08926551685742248, + "grad_norm": 0.6546609401702881, + "learning_rate": 0.00019611198929854026, + "loss": 0.1643, + "step": 3808 + }, + { + "epoch": 0.08928895843222748, + "grad_norm": 0.8106259107589722, + "learning_rate": 0.0001961099550102977, + "loss": 0.1705, + "step": 3809 + }, + { + "epoch": 0.08931240000703247, + "grad_norm": 0.6417884230613708, + "learning_rate": 0.00019610792020055982, + "loss": 0.1317, + "step": 3810 + }, + { + "epoch": 0.08933584158183747, + "grad_norm": 0.6451416611671448, + "learning_rate": 0.00019610588486933765, + "loss": 0.1935, + "step": 3811 + }, + { + "epoch": 0.08935928315664246, + "grad_norm": 0.5214853882789612, + "learning_rate": 0.00019610384901664227, + "loss": 0.1363, + "step": 3812 + }, + { + "epoch": 0.08938272473144745, + "grad_norm": 0.47301262617111206, + "learning_rate": 0.0001961018126424847, + "loss": 0.1325, + "step": 3813 + }, + { + "epoch": 0.08940616630625245, + "grad_norm": 0.4533989429473877, + "learning_rate": 0.00019609977574687598, + "loss": 0.0913, + "step": 3814 + }, + { + "epoch": 0.08942960788105744, + "grad_norm": 0.3510311245918274, + "learning_rate": 0.0001960977383298272, + "loss": 0.1138, + "step": 3815 + }, + { + "epoch": 0.08945304945586245, + "grad_norm": 0.1367504745721817, + "learning_rate": 0.00019609570039134938, + "loss": 0.0306, + "step": 3816 + }, + { + "epoch": 0.08947649103066745, + "grad_norm": 0.630663275718689, + "learning_rate": 0.00019609366193145357, + "loss": 0.1704, + "step": 3817 + }, + { + "epoch": 0.08949993260547244, + "grad_norm": 0.8309840559959412, + "learning_rate": 0.00019609162295015087, + "loss": 0.1651, + "step": 3818 + }, + { + "epoch": 0.08952337418027743, + "grad_norm": 0.5509322285652161, + "learning_rate": 0.0001960895834474523, + "loss": 0.1539, + "step": 3819 + }, + { + "epoch": 0.08954681575508243, + "grad_norm": 0.44131726026535034, + "learning_rate": 0.00019608754342336897, + "loss": 0.1275, + "step": 3820 + }, + { + "epoch": 0.08957025732988742, + "grad_norm": 0.4198302626609802, + "learning_rate": 0.00019608550287791192, + "loss": 0.3825, + "step": 3821 + }, + { + "epoch": 0.08959369890469242, + "grad_norm": 0.29124337434768677, + "learning_rate": 0.00019608346181109222, + "loss": 0.0431, + "step": 3822 + }, + { + "epoch": 0.08961714047949741, + "grad_norm": 0.7446104288101196, + "learning_rate": 0.00019608142022292096, + "loss": 0.2395, + "step": 3823 + }, + { + "epoch": 0.0896405820543024, + "grad_norm": 0.4068812429904938, + "learning_rate": 0.00019607937811340925, + "loss": 0.115, + "step": 3824 + }, + { + "epoch": 0.0896640236291074, + "grad_norm": 0.6942898035049438, + "learning_rate": 0.00019607733548256806, + "loss": 0.7901, + "step": 3825 + }, + { + "epoch": 0.0896874652039124, + "grad_norm": 0.9090409278869629, + "learning_rate": 0.0001960752923304086, + "loss": 0.2055, + "step": 3826 + }, + { + "epoch": 0.08971090677871739, + "grad_norm": 0.7031241655349731, + "learning_rate": 0.00019607324865694188, + "loss": 0.4453, + "step": 3827 + }, + { + "epoch": 0.08973434835352238, + "grad_norm": 0.518531858921051, + "learning_rate": 0.00019607120446217904, + "loss": 0.0989, + "step": 3828 + }, + { + "epoch": 0.08975778992832739, + "grad_norm": 0.44335323572158813, + "learning_rate": 0.0001960691597461311, + "loss": 0.1041, + "step": 3829 + }, + { + "epoch": 0.08978123150313239, + "grad_norm": 0.6499403715133667, + "learning_rate": 0.00019606711450880925, + "loss": 0.9012, + "step": 3830 + }, + { + "epoch": 0.08980467307793738, + "grad_norm": 0.31891462206840515, + "learning_rate": 0.00019606506875022447, + "loss": 0.1199, + "step": 3831 + }, + { + "epoch": 0.08982811465274237, + "grad_norm": 0.21222738921642303, + "learning_rate": 0.00019606302247038797, + "loss": 0.052, + "step": 3832 + }, + { + "epoch": 0.08985155622754737, + "grad_norm": 0.46305131912231445, + "learning_rate": 0.00019606097566931082, + "loss": 0.1147, + "step": 3833 + }, + { + "epoch": 0.08987499780235236, + "grad_norm": 0.4844874143600464, + "learning_rate": 0.00019605892834700407, + "loss": 0.1682, + "step": 3834 + }, + { + "epoch": 0.08989843937715736, + "grad_norm": 0.11179858446121216, + "learning_rate": 0.0001960568805034789, + "loss": 0.0271, + "step": 3835 + }, + { + "epoch": 0.08992188095196235, + "grad_norm": 0.3664592206478119, + "learning_rate": 0.0001960548321387464, + "loss": 0.0825, + "step": 3836 + }, + { + "epoch": 0.08994532252676735, + "grad_norm": 0.7129150629043579, + "learning_rate": 0.00019605278325281767, + "loss": 0.3193, + "step": 3837 + }, + { + "epoch": 0.08996876410157234, + "grad_norm": 0.6911553740501404, + "learning_rate": 0.00019605073384570384, + "loss": 0.1542, + "step": 3838 + }, + { + "epoch": 0.08999220567637733, + "grad_norm": 0.5514938831329346, + "learning_rate": 0.00019604868391741604, + "loss": 0.0922, + "step": 3839 + }, + { + "epoch": 0.09001564725118233, + "grad_norm": 0.4432576596736908, + "learning_rate": 0.00019604663346796538, + "loss": 0.1203, + "step": 3840 + }, + { + "epoch": 0.09003908882598732, + "grad_norm": 0.43996894359588623, + "learning_rate": 0.00019604458249736296, + "loss": 0.1458, + "step": 3841 + }, + { + "epoch": 0.09006253040079233, + "grad_norm": 0.4052284061908722, + "learning_rate": 0.00019604253100561993, + "loss": 0.1028, + "step": 3842 + }, + { + "epoch": 0.09008597197559733, + "grad_norm": 0.49523499608039856, + "learning_rate": 0.00019604047899274746, + "loss": 0.0783, + "step": 3843 + }, + { + "epoch": 0.09010941355040232, + "grad_norm": 0.4449259340763092, + "learning_rate": 0.00019603842645875665, + "loss": 0.1135, + "step": 3844 + }, + { + "epoch": 0.09013285512520731, + "grad_norm": 0.45335328578948975, + "learning_rate": 0.00019603637340365862, + "loss": 0.1172, + "step": 3845 + }, + { + "epoch": 0.09015629670001231, + "grad_norm": 0.8119585514068604, + "learning_rate": 0.00019603431982746452, + "loss": 0.2124, + "step": 3846 + }, + { + "epoch": 0.0901797382748173, + "grad_norm": 0.24540261924266815, + "learning_rate": 0.00019603226573018553, + "loss": 0.0535, + "step": 3847 + }, + { + "epoch": 0.0902031798496223, + "grad_norm": 0.6102820634841919, + "learning_rate": 0.00019603021111183275, + "loss": 0.513, + "step": 3848 + }, + { + "epoch": 0.09022662142442729, + "grad_norm": 1.0708682537078857, + "learning_rate": 0.00019602815597241736, + "loss": 0.2117, + "step": 3849 + }, + { + "epoch": 0.09025006299923229, + "grad_norm": 0.6165540814399719, + "learning_rate": 0.00019602610031195045, + "loss": 0.1314, + "step": 3850 + }, + { + "epoch": 0.09027350457403728, + "grad_norm": 0.5768490433692932, + "learning_rate": 0.00019602404413044324, + "loss": 0.1546, + "step": 3851 + }, + { + "epoch": 0.09029694614884227, + "grad_norm": 0.4902581572532654, + "learning_rate": 0.00019602198742790688, + "loss": 0.7046, + "step": 3852 + }, + { + "epoch": 0.09032038772364727, + "grad_norm": 0.6675931811332703, + "learning_rate": 0.00019601993020435253, + "loss": 0.1751, + "step": 3853 + }, + { + "epoch": 0.09034382929845226, + "grad_norm": 0.49076035618782043, + "learning_rate": 0.0001960178724597913, + "loss": 0.8409, + "step": 3854 + }, + { + "epoch": 0.09036727087325727, + "grad_norm": 0.4949086308479309, + "learning_rate": 0.00019601581419423443, + "loss": 0.0944, + "step": 3855 + }, + { + "epoch": 0.09039071244806227, + "grad_norm": 0.36974039673805237, + "learning_rate": 0.00019601375540769303, + "loss": 0.6625, + "step": 3856 + }, + { + "epoch": 0.09041415402286726, + "grad_norm": 0.49378564953804016, + "learning_rate": 0.00019601169610017832, + "loss": 0.1179, + "step": 3857 + }, + { + "epoch": 0.09043759559767225, + "grad_norm": 0.5481164455413818, + "learning_rate": 0.00019600963627170142, + "loss": 0.1335, + "step": 3858 + }, + { + "epoch": 0.09046103717247725, + "grad_norm": 0.29450032114982605, + "learning_rate": 0.00019600757592227352, + "loss": 0.1102, + "step": 3859 + }, + { + "epoch": 0.09048447874728224, + "grad_norm": 0.3704769015312195, + "learning_rate": 0.00019600551505190585, + "loss": 0.0697, + "step": 3860 + }, + { + "epoch": 0.09050792032208724, + "grad_norm": 1.094264030456543, + "learning_rate": 0.00019600345366060955, + "loss": 0.1611, + "step": 3861 + }, + { + "epoch": 0.09053136189689223, + "grad_norm": 0.6676830053329468, + "learning_rate": 0.00019600139174839578, + "loss": 0.1928, + "step": 3862 + }, + { + "epoch": 0.09055480347169723, + "grad_norm": 0.6044160723686218, + "learning_rate": 0.0001959993293152758, + "loss": 0.1212, + "step": 3863 + }, + { + "epoch": 0.09057824504650222, + "grad_norm": 0.6554685831069946, + "learning_rate": 0.0001959972663612607, + "loss": 0.1567, + "step": 3864 + }, + { + "epoch": 0.09060168662130721, + "grad_norm": 0.6042359471321106, + "learning_rate": 0.00019599520288636181, + "loss": 0.2305, + "step": 3865 + }, + { + "epoch": 0.09062512819611221, + "grad_norm": 0.5317223072052002, + "learning_rate": 0.0001959931388905902, + "loss": 0.1237, + "step": 3866 + }, + { + "epoch": 0.0906485697709172, + "grad_norm": 0.6945502161979675, + "learning_rate": 0.00019599107437395715, + "loss": 0.2088, + "step": 3867 + }, + { + "epoch": 0.09067201134572221, + "grad_norm": 0.3788738548755646, + "learning_rate": 0.00019598900933647382, + "loss": 0.0636, + "step": 3868 + }, + { + "epoch": 0.0906954529205272, + "grad_norm": 0.642842173576355, + "learning_rate": 0.0001959869437781514, + "loss": 0.1174, + "step": 3869 + }, + { + "epoch": 0.0907188944953322, + "grad_norm": 0.5135043859481812, + "learning_rate": 0.00019598487769900116, + "loss": 0.1459, + "step": 3870 + }, + { + "epoch": 0.0907423360701372, + "grad_norm": 0.7926623225212097, + "learning_rate": 0.00019598281109903427, + "loss": 0.2611, + "step": 3871 + }, + { + "epoch": 0.09076577764494219, + "grad_norm": 0.23114651441574097, + "learning_rate": 0.00019598074397826195, + "loss": 0.0349, + "step": 3872 + }, + { + "epoch": 0.09078921921974718, + "grad_norm": 0.44854435324668884, + "learning_rate": 0.0001959786763366954, + "loss": 0.1576, + "step": 3873 + }, + { + "epoch": 0.09081266079455218, + "grad_norm": 0.6956064105033875, + "learning_rate": 0.00019597660817434584, + "loss": 0.1771, + "step": 3874 + }, + { + "epoch": 0.09083610236935717, + "grad_norm": 0.22594857215881348, + "learning_rate": 0.00019597453949122453, + "loss": 0.0706, + "step": 3875 + }, + { + "epoch": 0.09085954394416217, + "grad_norm": 0.5359764695167542, + "learning_rate": 0.00019597247028734267, + "loss": 0.1161, + "step": 3876 + }, + { + "epoch": 0.09088298551896716, + "grad_norm": 0.3717862069606781, + "learning_rate": 0.00019597040056271148, + "loss": 0.1175, + "step": 3877 + }, + { + "epoch": 0.09090642709377215, + "grad_norm": 0.5734536647796631, + "learning_rate": 0.0001959683303173422, + "loss": 0.1479, + "step": 3878 + }, + { + "epoch": 0.09092986866857715, + "grad_norm": 0.41846758127212524, + "learning_rate": 0.00019596625955124606, + "loss": 0.1132, + "step": 3879 + }, + { + "epoch": 0.09095331024338216, + "grad_norm": 0.5692993402481079, + "learning_rate": 0.0001959641882644343, + "loss": 0.177, + "step": 3880 + }, + { + "epoch": 0.09097675181818715, + "grad_norm": 0.5728318095207214, + "learning_rate": 0.00019596211645691813, + "loss": 0.8352, + "step": 3881 + }, + { + "epoch": 0.09100019339299215, + "grad_norm": 0.7019538283348083, + "learning_rate": 0.00019596004412870884, + "loss": 0.1399, + "step": 3882 + }, + { + "epoch": 0.09102363496779714, + "grad_norm": 0.6170162558555603, + "learning_rate": 0.00019595797127981766, + "loss": 0.1528, + "step": 3883 + }, + { + "epoch": 0.09104707654260213, + "grad_norm": 0.8072185516357422, + "learning_rate": 0.00019595589791025582, + "loss": 0.9693, + "step": 3884 + }, + { + "epoch": 0.09107051811740713, + "grad_norm": 0.49075984954833984, + "learning_rate": 0.0001959538240200346, + "loss": 0.1387, + "step": 3885 + }, + { + "epoch": 0.09109395969221212, + "grad_norm": 0.37204012274742126, + "learning_rate": 0.0001959517496091652, + "loss": 0.078, + "step": 3886 + }, + { + "epoch": 0.09111740126701712, + "grad_norm": 0.3289319574832916, + "learning_rate": 0.00019594967467765893, + "loss": 0.0361, + "step": 3887 + }, + { + "epoch": 0.09114084284182211, + "grad_norm": 0.6002510190010071, + "learning_rate": 0.00019594759922552703, + "loss": 0.2092, + "step": 3888 + }, + { + "epoch": 0.0911642844166271, + "grad_norm": 0.5495257377624512, + "learning_rate": 0.00019594552325278075, + "loss": 0.7076, + "step": 3889 + }, + { + "epoch": 0.0911877259914321, + "grad_norm": 0.43886101245880127, + "learning_rate": 0.00019594344675943136, + "loss": 0.0899, + "step": 3890 + }, + { + "epoch": 0.0912111675662371, + "grad_norm": 0.17938552796840668, + "learning_rate": 0.00019594136974549013, + "loss": 0.0611, + "step": 3891 + }, + { + "epoch": 0.09123460914104209, + "grad_norm": 0.7910386919975281, + "learning_rate": 0.00019593929221096835, + "loss": 0.2126, + "step": 3892 + }, + { + "epoch": 0.0912580507158471, + "grad_norm": 0.549855649471283, + "learning_rate": 0.00019593721415587725, + "loss": 0.8257, + "step": 3893 + }, + { + "epoch": 0.09128149229065209, + "grad_norm": 0.34635618329048157, + "learning_rate": 0.00019593513558022814, + "loss": 0.08, + "step": 3894 + }, + { + "epoch": 0.09130493386545709, + "grad_norm": 0.7779479622840881, + "learning_rate": 0.00019593305648403226, + "loss": 0.1439, + "step": 3895 + }, + { + "epoch": 0.09132837544026208, + "grad_norm": 0.5721011757850647, + "learning_rate": 0.00019593097686730095, + "loss": 0.1549, + "step": 3896 + }, + { + "epoch": 0.09135181701506707, + "grad_norm": 0.395631343126297, + "learning_rate": 0.00019592889673004544, + "loss": 0.0873, + "step": 3897 + }, + { + "epoch": 0.09137525858987207, + "grad_norm": 0.4175715446472168, + "learning_rate": 0.00019592681607227705, + "loss": 0.0961, + "step": 3898 + }, + { + "epoch": 0.09139870016467706, + "grad_norm": 0.7899255156517029, + "learning_rate": 0.00019592473489400707, + "loss": 0.2241, + "step": 3899 + }, + { + "epoch": 0.09142214173948206, + "grad_norm": 0.5056964755058289, + "learning_rate": 0.0001959226531952468, + "loss": 0.1284, + "step": 3900 + }, + { + "epoch": 0.09144558331428705, + "grad_norm": 0.5494536757469177, + "learning_rate": 0.00019592057097600748, + "loss": 0.1584, + "step": 3901 + }, + { + "epoch": 0.09146902488909205, + "grad_norm": 0.29968708753585815, + "learning_rate": 0.00019591848823630046, + "loss": 0.0814, + "step": 3902 + }, + { + "epoch": 0.09149246646389704, + "grad_norm": 0.26997724175453186, + "learning_rate": 0.00019591640497613702, + "loss": 0.0573, + "step": 3903 + }, + { + "epoch": 0.09151590803870203, + "grad_norm": 0.4599801301956177, + "learning_rate": 0.00019591432119552845, + "loss": 0.1322, + "step": 3904 + }, + { + "epoch": 0.09153934961350703, + "grad_norm": 0.5586200952529907, + "learning_rate": 0.00019591223689448611, + "loss": 0.1666, + "step": 3905 + }, + { + "epoch": 0.09156279118831204, + "grad_norm": 0.7172501087188721, + "learning_rate": 0.00019591015207302124, + "loss": 0.265, + "step": 3906 + }, + { + "epoch": 0.09158623276311703, + "grad_norm": 0.33797985315322876, + "learning_rate": 0.00019590806673114522, + "loss": 0.062, + "step": 3907 + }, + { + "epoch": 0.09160967433792203, + "grad_norm": 0.6642261147499084, + "learning_rate": 0.00019590598086886932, + "loss": 0.2031, + "step": 3908 + }, + { + "epoch": 0.09163311591272702, + "grad_norm": 0.782748818397522, + "learning_rate": 0.00019590389448620487, + "loss": 0.1203, + "step": 3909 + }, + { + "epoch": 0.09165655748753201, + "grad_norm": 0.5609946250915527, + "learning_rate": 0.00019590180758316322, + "loss": 0.0635, + "step": 3910 + }, + { + "epoch": 0.09167999906233701, + "grad_norm": 0.4188430905342102, + "learning_rate": 0.00019589972015975562, + "loss": 0.1016, + "step": 3911 + }, + { + "epoch": 0.091703440637142, + "grad_norm": 0.8869742751121521, + "learning_rate": 0.00019589763221599348, + "loss": 0.1207, + "step": 3912 + }, + { + "epoch": 0.091726882211947, + "grad_norm": 0.7618728876113892, + "learning_rate": 0.00019589554375188805, + "loss": 0.9208, + "step": 3913 + }, + { + "epoch": 0.09175032378675199, + "grad_norm": 0.730579674243927, + "learning_rate": 0.00019589345476745074, + "loss": 0.1493, + "step": 3914 + }, + { + "epoch": 0.09177376536155699, + "grad_norm": 0.7308956980705261, + "learning_rate": 0.00019589136526269283, + "loss": 0.129, + "step": 3915 + }, + { + "epoch": 0.09179720693636198, + "grad_norm": 0.5493001937866211, + "learning_rate": 0.00019588927523762567, + "loss": 0.6586, + "step": 3916 + }, + { + "epoch": 0.09182064851116697, + "grad_norm": 0.6719971895217896, + "learning_rate": 0.00019588718469226062, + "loss": 0.1222, + "step": 3917 + }, + { + "epoch": 0.09184409008597197, + "grad_norm": 0.6142128109931946, + "learning_rate": 0.00019588509362660901, + "loss": 0.1303, + "step": 3918 + }, + { + "epoch": 0.09186753166077698, + "grad_norm": 1.8856569528579712, + "learning_rate": 0.0001958830020406822, + "loss": 0.2763, + "step": 3919 + }, + { + "epoch": 0.09189097323558197, + "grad_norm": 0.3429468274116516, + "learning_rate": 0.0001958809099344915, + "loss": 0.0559, + "step": 3920 + }, + { + "epoch": 0.09191441481038697, + "grad_norm": 0.5499963760375977, + "learning_rate": 0.00019587881730804828, + "loss": 0.1504, + "step": 3921 + }, + { + "epoch": 0.09193785638519196, + "grad_norm": 0.2826288342475891, + "learning_rate": 0.00019587672416136392, + "loss": 0.0725, + "step": 3922 + }, + { + "epoch": 0.09196129795999695, + "grad_norm": 0.5171840190887451, + "learning_rate": 0.00019587463049444977, + "loss": 0.1651, + "step": 3923 + }, + { + "epoch": 0.09198473953480195, + "grad_norm": 0.21427172422409058, + "learning_rate": 0.00019587253630731717, + "loss": 0.0601, + "step": 3924 + }, + { + "epoch": 0.09200818110960694, + "grad_norm": 0.7432731986045837, + "learning_rate": 0.00019587044159997749, + "loss": 0.2207, + "step": 3925 + }, + { + "epoch": 0.09203162268441194, + "grad_norm": 0.7341137528419495, + "learning_rate": 0.0001958683463724421, + "loss": 0.1408, + "step": 3926 + }, + { + "epoch": 0.09205506425921693, + "grad_norm": 0.17879356443881989, + "learning_rate": 0.00019586625062472234, + "loss": 0.04, + "step": 3927 + }, + { + "epoch": 0.09207850583402193, + "grad_norm": 0.5033879280090332, + "learning_rate": 0.00019586415435682964, + "loss": 0.7431, + "step": 3928 + }, + { + "epoch": 0.09210194740882692, + "grad_norm": 0.4781463146209717, + "learning_rate": 0.00019586205756877534, + "loss": 0.1485, + "step": 3929 + }, + { + "epoch": 0.09212538898363191, + "grad_norm": 0.7594573497772217, + "learning_rate": 0.00019585996026057082, + "loss": 0.2616, + "step": 3930 + }, + { + "epoch": 0.09214883055843692, + "grad_norm": 0.6347326040267944, + "learning_rate": 0.00019585786243222746, + "loss": 0.2123, + "step": 3931 + }, + { + "epoch": 0.09217227213324192, + "grad_norm": 0.5306419134140015, + "learning_rate": 0.00019585576408375662, + "loss": 0.0892, + "step": 3932 + }, + { + "epoch": 0.09219571370804691, + "grad_norm": 0.7928760051727295, + "learning_rate": 0.00019585366521516972, + "loss": 0.207, + "step": 3933 + }, + { + "epoch": 0.0922191552828519, + "grad_norm": 0.4717167913913727, + "learning_rate": 0.00019585156582647815, + "loss": 0.1137, + "step": 3934 + }, + { + "epoch": 0.0922425968576569, + "grad_norm": 0.24632969498634338, + "learning_rate": 0.00019584946591769326, + "loss": 0.0561, + "step": 3935 + }, + { + "epoch": 0.0922660384324619, + "grad_norm": 0.0833762064576149, + "learning_rate": 0.0001958473654888265, + "loss": 0.019, + "step": 3936 + }, + { + "epoch": 0.09228948000726689, + "grad_norm": 0.606147825717926, + "learning_rate": 0.00019584526453988925, + "loss": 0.1394, + "step": 3937 + }, + { + "epoch": 0.09231292158207188, + "grad_norm": 0.4922872483730316, + "learning_rate": 0.00019584316307089288, + "loss": 0.1143, + "step": 3938 + }, + { + "epoch": 0.09233636315687688, + "grad_norm": 0.42705950140953064, + "learning_rate": 0.00019584106108184882, + "loss": 0.1201, + "step": 3939 + }, + { + "epoch": 0.09235980473168187, + "grad_norm": 0.43251535296440125, + "learning_rate": 0.00019583895857276846, + "loss": 0.1672, + "step": 3940 + }, + { + "epoch": 0.09238324630648687, + "grad_norm": 0.48558899760246277, + "learning_rate": 0.00019583685554366324, + "loss": 0.0939, + "step": 3941 + }, + { + "epoch": 0.09240668788129186, + "grad_norm": 0.6345486640930176, + "learning_rate": 0.00019583475199454452, + "loss": 0.1385, + "step": 3942 + }, + { + "epoch": 0.09243012945609685, + "grad_norm": 0.3352912962436676, + "learning_rate": 0.00019583264792542374, + "loss": 0.0899, + "step": 3943 + }, + { + "epoch": 0.09245357103090186, + "grad_norm": 0.8341488242149353, + "learning_rate": 0.00019583054333631236, + "loss": 0.2244, + "step": 3944 + }, + { + "epoch": 0.09247701260570686, + "grad_norm": 0.4388292133808136, + "learning_rate": 0.0001958284382272217, + "loss": 0.1269, + "step": 3945 + }, + { + "epoch": 0.09250045418051185, + "grad_norm": 0.4710147976875305, + "learning_rate": 0.00019582633259816325, + "loss": 0.1062, + "step": 3946 + }, + { + "epoch": 0.09252389575531685, + "grad_norm": 0.6147160530090332, + "learning_rate": 0.00019582422644914846, + "loss": 0.0922, + "step": 3947 + }, + { + "epoch": 0.09254733733012184, + "grad_norm": 0.4279003143310547, + "learning_rate": 0.0001958221197801887, + "loss": 0.0577, + "step": 3948 + }, + { + "epoch": 0.09257077890492683, + "grad_norm": 0.28770074248313904, + "learning_rate": 0.00019582001259129542, + "loss": 0.088, + "step": 3949 + }, + { + "epoch": 0.09259422047973183, + "grad_norm": 0.31845828890800476, + "learning_rate": 0.00019581790488248008, + "loss": 0.0497, + "step": 3950 + }, + { + "epoch": 0.09261766205453682, + "grad_norm": 0.6901636123657227, + "learning_rate": 0.00019581579665375407, + "loss": 0.1684, + "step": 3951 + }, + { + "epoch": 0.09264110362934182, + "grad_norm": 0.4699903726577759, + "learning_rate": 0.00019581368790512883, + "loss": 0.1213, + "step": 3952 + }, + { + "epoch": 0.09266454520414681, + "grad_norm": 0.575778067111969, + "learning_rate": 0.0001958115786366159, + "loss": 0.6998, + "step": 3953 + }, + { + "epoch": 0.0926879867789518, + "grad_norm": 0.7182656526565552, + "learning_rate": 0.0001958094688482266, + "loss": 0.1591, + "step": 3954 + }, + { + "epoch": 0.0927114283537568, + "grad_norm": 0.6344795823097229, + "learning_rate": 0.00019580735853997241, + "loss": 0.1245, + "step": 3955 + }, + { + "epoch": 0.0927348699285618, + "grad_norm": 0.6463645100593567, + "learning_rate": 0.00019580524771186481, + "loss": 0.2741, + "step": 3956 + }, + { + "epoch": 0.0927583115033668, + "grad_norm": 0.612134575843811, + "learning_rate": 0.00019580313636391529, + "loss": 0.171, + "step": 3957 + }, + { + "epoch": 0.0927817530781718, + "grad_norm": 0.3867059648036957, + "learning_rate": 0.0001958010244961352, + "loss": 0.594, + "step": 3958 + }, + { + "epoch": 0.09280519465297679, + "grad_norm": 0.4701751470565796, + "learning_rate": 0.0001957989121085361, + "loss": 0.1021, + "step": 3959 + }, + { + "epoch": 0.09282863622778179, + "grad_norm": 0.4497973918914795, + "learning_rate": 0.00019579679920112936, + "loss": 0.1307, + "step": 3960 + }, + { + "epoch": 0.09285207780258678, + "grad_norm": 0.20544540882110596, + "learning_rate": 0.00019579468577392655, + "loss": 0.0844, + "step": 3961 + }, + { + "epoch": 0.09287551937739177, + "grad_norm": 0.9132511019706726, + "learning_rate": 0.00019579257182693904, + "loss": 0.0749, + "step": 3962 + }, + { + "epoch": 0.09289896095219677, + "grad_norm": 0.5555505156517029, + "learning_rate": 0.00019579045736017837, + "loss": 0.1063, + "step": 3963 + }, + { + "epoch": 0.09292240252700176, + "grad_norm": 0.9112286567687988, + "learning_rate": 0.00019578834237365595, + "loss": 0.1736, + "step": 3964 + }, + { + "epoch": 0.09294584410180676, + "grad_norm": 0.41089943051338196, + "learning_rate": 0.00019578622686738332, + "loss": 0.1298, + "step": 3965 + }, + { + "epoch": 0.09296928567661175, + "grad_norm": 0.7871052622795105, + "learning_rate": 0.00019578411084137195, + "loss": 0.9062, + "step": 3966 + }, + { + "epoch": 0.09299272725141675, + "grad_norm": 0.6654585003852844, + "learning_rate": 0.00019578199429563323, + "loss": 0.1563, + "step": 3967 + }, + { + "epoch": 0.09301616882622174, + "grad_norm": 0.18815091252326965, + "learning_rate": 0.00019577987723017878, + "loss": 0.041, + "step": 3968 + }, + { + "epoch": 0.09303961040102673, + "grad_norm": 0.6500517725944519, + "learning_rate": 0.00019577775964502002, + "loss": 0.138, + "step": 3969 + }, + { + "epoch": 0.09306305197583174, + "grad_norm": 0.5699960589408875, + "learning_rate": 0.0001957756415401684, + "loss": 0.1202, + "step": 3970 + }, + { + "epoch": 0.09308649355063674, + "grad_norm": 0.7108030915260315, + "learning_rate": 0.00019577352291563549, + "loss": 0.2238, + "step": 3971 + }, + { + "epoch": 0.09310993512544173, + "grad_norm": 0.7038835287094116, + "learning_rate": 0.00019577140377143276, + "loss": 0.5596, + "step": 3972 + }, + { + "epoch": 0.09313337670024673, + "grad_norm": 0.38926592469215393, + "learning_rate": 0.00019576928410757166, + "loss": 0.1133, + "step": 3973 + }, + { + "epoch": 0.09315681827505172, + "grad_norm": 0.4975616931915283, + "learning_rate": 0.0001957671639240638, + "loss": 0.2376, + "step": 3974 + }, + { + "epoch": 0.09318025984985671, + "grad_norm": 0.527027428150177, + "learning_rate": 0.00019576504322092056, + "loss": 0.1878, + "step": 3975 + }, + { + "epoch": 0.09320370142466171, + "grad_norm": 0.5105933547019958, + "learning_rate": 0.00019576292199815348, + "loss": 0.1172, + "step": 3976 + }, + { + "epoch": 0.0932271429994667, + "grad_norm": 0.5733494162559509, + "learning_rate": 0.00019576080025577415, + "loss": 0.1586, + "step": 3977 + }, + { + "epoch": 0.0932505845742717, + "grad_norm": 0.4728325605392456, + "learning_rate": 0.000195758677993794, + "loss": 0.1337, + "step": 3978 + }, + { + "epoch": 0.09327402614907669, + "grad_norm": 0.7253130078315735, + "learning_rate": 0.00019575655521222456, + "loss": 0.6002, + "step": 3979 + }, + { + "epoch": 0.09329746772388169, + "grad_norm": 0.2538681626319885, + "learning_rate": 0.00019575443191107737, + "loss": 0.0699, + "step": 3980 + }, + { + "epoch": 0.09332090929868668, + "grad_norm": 0.7477884292602539, + "learning_rate": 0.00019575230809036395, + "loss": 0.0911, + "step": 3981 + }, + { + "epoch": 0.09334435087349167, + "grad_norm": 0.5600191354751587, + "learning_rate": 0.00019575018375009583, + "loss": 0.1174, + "step": 3982 + }, + { + "epoch": 0.09336779244829668, + "grad_norm": 0.5725374817848206, + "learning_rate": 0.00019574805889028448, + "loss": 0.1787, + "step": 3983 + }, + { + "epoch": 0.09339123402310168, + "grad_norm": 0.38931187987327576, + "learning_rate": 0.00019574593351094147, + "loss": 0.4622, + "step": 3984 + }, + { + "epoch": 0.09341467559790667, + "grad_norm": 0.5187137722969055, + "learning_rate": 0.00019574380761207838, + "loss": 0.1697, + "step": 3985 + }, + { + "epoch": 0.09343811717271167, + "grad_norm": 1.1314181089401245, + "learning_rate": 0.00019574168119370665, + "loss": 0.0995, + "step": 3986 + }, + { + "epoch": 0.09346155874751666, + "grad_norm": 0.844215989112854, + "learning_rate": 0.00019573955425583787, + "loss": 0.1393, + "step": 3987 + }, + { + "epoch": 0.09348500032232165, + "grad_norm": 0.11887677758932114, + "learning_rate": 0.0001957374267984836, + "loss": 0.0187, + "step": 3988 + }, + { + "epoch": 0.09350844189712665, + "grad_norm": 0.6458276510238647, + "learning_rate": 0.00019573529882165535, + "loss": 0.1821, + "step": 3989 + }, + { + "epoch": 0.09353188347193164, + "grad_norm": 0.6448591351509094, + "learning_rate": 0.0001957331703253647, + "loss": 0.1608, + "step": 3990 + }, + { + "epoch": 0.09355532504673664, + "grad_norm": 0.7714619040489197, + "learning_rate": 0.0001957310413096231, + "loss": 0.2162, + "step": 3991 + }, + { + "epoch": 0.09357876662154163, + "grad_norm": 0.6591419577598572, + "learning_rate": 0.00019572891177444226, + "loss": 0.1602, + "step": 3992 + }, + { + "epoch": 0.09360220819634663, + "grad_norm": 0.4858422577381134, + "learning_rate": 0.00019572678171983363, + "loss": 0.1475, + "step": 3993 + }, + { + "epoch": 0.09362564977115162, + "grad_norm": 0.46028196811676025, + "learning_rate": 0.0001957246511458088, + "loss": 0.1142, + "step": 3994 + }, + { + "epoch": 0.09364909134595663, + "grad_norm": 0.7768867611885071, + "learning_rate": 0.0001957225200523793, + "loss": 0.1703, + "step": 3995 + }, + { + "epoch": 0.09367253292076162, + "grad_norm": 0.2557813227176666, + "learning_rate": 0.0001957203884395567, + "loss": 0.0586, + "step": 3996 + }, + { + "epoch": 0.09369597449556662, + "grad_norm": 0.20795318484306335, + "learning_rate": 0.00019571825630735262, + "loss": 0.0616, + "step": 3997 + }, + { + "epoch": 0.09371941607037161, + "grad_norm": 0.7042951583862305, + "learning_rate": 0.00019571612365577856, + "loss": 0.1404, + "step": 3998 + }, + { + "epoch": 0.0937428576451766, + "grad_norm": 0.5619989037513733, + "learning_rate": 0.00019571399048484613, + "loss": 0.1019, + "step": 3999 + }, + { + "epoch": 0.0937662992199816, + "grad_norm": 0.6595910787582397, + "learning_rate": 0.0001957118567945669, + "loss": 0.0501, + "step": 4000 + }, + { + "epoch": 0.0937897407947866, + "grad_norm": 0.6981291174888611, + "learning_rate": 0.00019570972258495245, + "loss": 0.7163, + "step": 4001 + }, + { + "epoch": 0.09381318236959159, + "grad_norm": 0.3342951536178589, + "learning_rate": 0.00019570758785601434, + "loss": 0.1072, + "step": 4002 + }, + { + "epoch": 0.09383662394439658, + "grad_norm": 0.7870925664901733, + "learning_rate": 0.00019570545260776419, + "loss": 0.1956, + "step": 4003 + }, + { + "epoch": 0.09386006551920158, + "grad_norm": 0.7139207124710083, + "learning_rate": 0.00019570331684021355, + "loss": 0.1892, + "step": 4004 + }, + { + "epoch": 0.09388350709400657, + "grad_norm": 0.7881412506103516, + "learning_rate": 0.00019570118055337404, + "loss": 0.1284, + "step": 4005 + }, + { + "epoch": 0.09390694866881157, + "grad_norm": 0.8824299573898315, + "learning_rate": 0.00019569904374725717, + "loss": 0.2766, + "step": 4006 + }, + { + "epoch": 0.09393039024361656, + "grad_norm": 0.5379645228385925, + "learning_rate": 0.00019569690642187468, + "loss": 0.1946, + "step": 4007 + }, + { + "epoch": 0.09395383181842157, + "grad_norm": 0.47391968965530396, + "learning_rate": 0.00019569476857723804, + "loss": 0.1716, + "step": 4008 + }, + { + "epoch": 0.09397727339322656, + "grad_norm": 0.20518994331359863, + "learning_rate": 0.0001956926302133589, + "loss": 0.044, + "step": 4009 + }, + { + "epoch": 0.09400071496803156, + "grad_norm": 0.7448629140853882, + "learning_rate": 0.00019569049133024886, + "loss": 0.1509, + "step": 4010 + }, + { + "epoch": 0.09402415654283655, + "grad_norm": 0.48357537388801575, + "learning_rate": 0.00019568835192791952, + "loss": 0.1538, + "step": 4011 + }, + { + "epoch": 0.09404759811764155, + "grad_norm": 0.7118297219276428, + "learning_rate": 0.00019568621200638248, + "loss": 0.1743, + "step": 4012 + }, + { + "epoch": 0.09407103969244654, + "grad_norm": 0.1967829316854477, + "learning_rate": 0.00019568407156564938, + "loss": 0.0456, + "step": 4013 + }, + { + "epoch": 0.09409448126725153, + "grad_norm": 0.6830894947052002, + "learning_rate": 0.0001956819306057318, + "loss": 0.172, + "step": 4014 + }, + { + "epoch": 0.09411792284205653, + "grad_norm": 0.20338371396064758, + "learning_rate": 0.0001956797891266414, + "loss": 0.0519, + "step": 4015 + }, + { + "epoch": 0.09414136441686152, + "grad_norm": 1.051416277885437, + "learning_rate": 0.00019567764712838978, + "loss": 0.1395, + "step": 4016 + }, + { + "epoch": 0.09416480599166652, + "grad_norm": 0.5059745907783508, + "learning_rate": 0.00019567550461098855, + "loss": 0.8909, + "step": 4017 + }, + { + "epoch": 0.09418824756647151, + "grad_norm": 0.7467055916786194, + "learning_rate": 0.0001956733615744493, + "loss": 0.2125, + "step": 4018 + }, + { + "epoch": 0.0942116891412765, + "grad_norm": 0.8714766502380371, + "learning_rate": 0.00019567121801878376, + "loss": 0.1825, + "step": 4019 + }, + { + "epoch": 0.0942351307160815, + "grad_norm": 0.3462282419204712, + "learning_rate": 0.00019566907394400347, + "loss": 0.1335, + "step": 4020 + }, + { + "epoch": 0.09425857229088651, + "grad_norm": 0.25850480794906616, + "learning_rate": 0.0001956669293501201, + "loss": 0.0915, + "step": 4021 + }, + { + "epoch": 0.0942820138656915, + "grad_norm": 0.5490046143531799, + "learning_rate": 0.00019566478423714526, + "loss": 0.1294, + "step": 4022 + }, + { + "epoch": 0.0943054554404965, + "grad_norm": 0.6319834589958191, + "learning_rate": 0.00019566263860509064, + "loss": 0.1868, + "step": 4023 + }, + { + "epoch": 0.09432889701530149, + "grad_norm": 0.5944236516952515, + "learning_rate": 0.00019566049245396786, + "loss": 0.1195, + "step": 4024 + }, + { + "epoch": 0.09435233859010649, + "grad_norm": 0.6439743041992188, + "learning_rate": 0.00019565834578378851, + "loss": 0.1681, + "step": 4025 + }, + { + "epoch": 0.09437578016491148, + "grad_norm": 0.6140376925468445, + "learning_rate": 0.00019565619859456432, + "loss": 0.1474, + "step": 4026 + }, + { + "epoch": 0.09439922173971647, + "grad_norm": 0.34390902519226074, + "learning_rate": 0.0001956540508863069, + "loss": 0.1518, + "step": 4027 + }, + { + "epoch": 0.09442266331452147, + "grad_norm": 0.4831221103668213, + "learning_rate": 0.00019565190265902793, + "loss": 0.1485, + "step": 4028 + }, + { + "epoch": 0.09444610488932646, + "grad_norm": 0.47082939743995667, + "learning_rate": 0.00019564975391273904, + "loss": 0.0802, + "step": 4029 + }, + { + "epoch": 0.09446954646413146, + "grad_norm": 0.19239433109760284, + "learning_rate": 0.0001956476046474519, + "loss": 0.0558, + "step": 4030 + }, + { + "epoch": 0.09449298803893645, + "grad_norm": 0.49090808629989624, + "learning_rate": 0.00019564545486317814, + "loss": 0.186, + "step": 4031 + }, + { + "epoch": 0.09451642961374145, + "grad_norm": 0.5484843850135803, + "learning_rate": 0.00019564330455992948, + "loss": 0.2488, + "step": 4032 + }, + { + "epoch": 0.09453987118854644, + "grad_norm": 0.5549308657646179, + "learning_rate": 0.00019564115373771755, + "loss": 0.1364, + "step": 4033 + }, + { + "epoch": 0.09456331276335145, + "grad_norm": 0.23860840499401093, + "learning_rate": 0.000195639002396554, + "loss": 0.0648, + "step": 4034 + }, + { + "epoch": 0.09458675433815644, + "grad_norm": 0.789966881275177, + "learning_rate": 0.00019563685053645058, + "loss": 0.3194, + "step": 4035 + }, + { + "epoch": 0.09461019591296144, + "grad_norm": 0.8374842405319214, + "learning_rate": 0.0001956346981574189, + "loss": 0.1981, + "step": 4036 + }, + { + "epoch": 0.09463363748776643, + "grad_norm": 0.20069436728954315, + "learning_rate": 0.00019563254525947066, + "loss": 0.0467, + "step": 4037 + }, + { + "epoch": 0.09465707906257143, + "grad_norm": 0.12409088015556335, + "learning_rate": 0.00019563039184261754, + "loss": 0.0285, + "step": 4038 + }, + { + "epoch": 0.09468052063737642, + "grad_norm": 0.5185081958770752, + "learning_rate": 0.00019562823790687122, + "loss": 0.1445, + "step": 4039 + }, + { + "epoch": 0.09470396221218141, + "grad_norm": 0.5372198820114136, + "learning_rate": 0.0001956260834522434, + "loss": 0.1905, + "step": 4040 + }, + { + "epoch": 0.09472740378698641, + "grad_norm": 0.20208987593650818, + "learning_rate": 0.00019562392847874576, + "loss": 0.0688, + "step": 4041 + }, + { + "epoch": 0.0947508453617914, + "grad_norm": 0.43175479769706726, + "learning_rate": 0.00019562177298639, + "loss": 0.1237, + "step": 4042 + }, + { + "epoch": 0.0947742869365964, + "grad_norm": 0.5969201326370239, + "learning_rate": 0.0001956196169751878, + "loss": 0.27, + "step": 4043 + }, + { + "epoch": 0.09479772851140139, + "grad_norm": 0.663073718547821, + "learning_rate": 0.00019561746044515087, + "loss": 0.1625, + "step": 4044 + }, + { + "epoch": 0.09482117008620639, + "grad_norm": 0.3125869631767273, + "learning_rate": 0.00019561530339629092, + "loss": 0.0658, + "step": 4045 + }, + { + "epoch": 0.09484461166101138, + "grad_norm": 0.6836917400360107, + "learning_rate": 0.00019561314582861966, + "loss": 0.1311, + "step": 4046 + }, + { + "epoch": 0.09486805323581639, + "grad_norm": 0.34974369406700134, + "learning_rate": 0.00019561098774214877, + "loss": 0.0676, + "step": 4047 + }, + { + "epoch": 0.09489149481062138, + "grad_norm": 0.40684905648231506, + "learning_rate": 0.00019560882913688995, + "loss": 0.1378, + "step": 4048 + }, + { + "epoch": 0.09491493638542638, + "grad_norm": 0.7353321313858032, + "learning_rate": 0.00019560667001285495, + "loss": 0.1657, + "step": 4049 + }, + { + "epoch": 0.09493837796023137, + "grad_norm": 0.5553657412528992, + "learning_rate": 0.00019560451037005548, + "loss": 0.1712, + "step": 4050 + }, + { + "epoch": 0.09496181953503637, + "grad_norm": 0.22218331694602966, + "learning_rate": 0.00019560235020850324, + "loss": 0.0615, + "step": 4051 + }, + { + "epoch": 0.09498526110984136, + "grad_norm": 0.22153978049755096, + "learning_rate": 0.00019560018952820998, + "loss": 0.053, + "step": 4052 + }, + { + "epoch": 0.09500870268464635, + "grad_norm": 0.6938797831535339, + "learning_rate": 0.00019559802832918736, + "loss": 0.1889, + "step": 4053 + }, + { + "epoch": 0.09503214425945135, + "grad_norm": 0.8268341422080994, + "learning_rate": 0.0001955958666114472, + "loss": 0.1292, + "step": 4054 + }, + { + "epoch": 0.09505558583425634, + "grad_norm": 1.2020076513290405, + "learning_rate": 0.00019559370437500112, + "loss": 0.5572, + "step": 4055 + }, + { + "epoch": 0.09507902740906134, + "grad_norm": 0.47128626704216003, + "learning_rate": 0.00019559154161986094, + "loss": 0.1424, + "step": 4056 + }, + { + "epoch": 0.09510246898386633, + "grad_norm": 0.7076482772827148, + "learning_rate": 0.0001955893783460384, + "loss": 0.2138, + "step": 4057 + }, + { + "epoch": 0.09512591055867133, + "grad_norm": 0.24650359153747559, + "learning_rate": 0.00019558721455354517, + "loss": 0.1274, + "step": 4058 + }, + { + "epoch": 0.09514935213347633, + "grad_norm": 0.614686131477356, + "learning_rate": 0.00019558505024239303, + "loss": 0.2367, + "step": 4059 + }, + { + "epoch": 0.09517279370828133, + "grad_norm": 0.5234877467155457, + "learning_rate": 0.00019558288541259375, + "loss": 0.1952, + "step": 4060 + }, + { + "epoch": 0.09519623528308632, + "grad_norm": 0.3300509750843048, + "learning_rate": 0.000195580720064159, + "loss": 0.0594, + "step": 4061 + }, + { + "epoch": 0.09521967685789132, + "grad_norm": 0.544722855091095, + "learning_rate": 0.0001955785541971006, + "loss": 0.1214, + "step": 4062 + }, + { + "epoch": 0.09524311843269631, + "grad_norm": 0.20503224432468414, + "learning_rate": 0.00019557638781143028, + "loss": 0.0509, + "step": 4063 + }, + { + "epoch": 0.0952665600075013, + "grad_norm": 0.771754801273346, + "learning_rate": 0.0001955742209071598, + "loss": 0.1827, + "step": 4064 + }, + { + "epoch": 0.0952900015823063, + "grad_norm": 0.6976708769798279, + "learning_rate": 0.0001955720534843009, + "loss": 0.2076, + "step": 4065 + }, + { + "epoch": 0.0953134431571113, + "grad_norm": 0.38354504108428955, + "learning_rate": 0.00019556988554286534, + "loss": 0.1566, + "step": 4066 + }, + { + "epoch": 0.09533688473191629, + "grad_norm": 0.5879237651824951, + "learning_rate": 0.0001955677170828649, + "loss": 0.1414, + "step": 4067 + }, + { + "epoch": 0.09536032630672128, + "grad_norm": 0.2675374746322632, + "learning_rate": 0.00019556554810431137, + "loss": 0.0835, + "step": 4068 + }, + { + "epoch": 0.09538376788152628, + "grad_norm": 0.4770928621292114, + "learning_rate": 0.00019556337860721646, + "loss": 0.1692, + "step": 4069 + }, + { + "epoch": 0.09540720945633127, + "grad_norm": 0.6641725301742554, + "learning_rate": 0.00019556120859159194, + "loss": 0.1705, + "step": 4070 + }, + { + "epoch": 0.09543065103113627, + "grad_norm": 0.6420589685440063, + "learning_rate": 0.00019555903805744965, + "loss": 0.6883, + "step": 4071 + }, + { + "epoch": 0.09545409260594127, + "grad_norm": 0.31961214542388916, + "learning_rate": 0.00019555686700480134, + "loss": 0.0861, + "step": 4072 + }, + { + "epoch": 0.09547753418074627, + "grad_norm": 0.8285989165306091, + "learning_rate": 0.00019555469543365876, + "loss": 0.2792, + "step": 4073 + }, + { + "epoch": 0.09550097575555126, + "grad_norm": 0.6568976044654846, + "learning_rate": 0.00019555252334403373, + "loss": 0.2528, + "step": 4074 + }, + { + "epoch": 0.09552441733035626, + "grad_norm": 0.5078408122062683, + "learning_rate": 0.00019555035073593802, + "loss": 0.0973, + "step": 4075 + }, + { + "epoch": 0.09554785890516125, + "grad_norm": 0.5124058127403259, + "learning_rate": 0.00019554817760938342, + "loss": 0.1589, + "step": 4076 + }, + { + "epoch": 0.09557130047996625, + "grad_norm": 0.4837631583213806, + "learning_rate": 0.00019554600396438172, + "loss": 0.1334, + "step": 4077 + }, + { + "epoch": 0.09559474205477124, + "grad_norm": 0.5878453254699707, + "learning_rate": 0.00019554382980094473, + "loss": 0.0724, + "step": 4078 + }, + { + "epoch": 0.09561818362957623, + "grad_norm": 0.42870837450027466, + "learning_rate": 0.0001955416551190842, + "loss": 0.091, + "step": 4079 + }, + { + "epoch": 0.09564162520438123, + "grad_norm": 0.6146594285964966, + "learning_rate": 0.000195539479918812, + "loss": 0.156, + "step": 4080 + }, + { + "epoch": 0.09566506677918622, + "grad_norm": 0.5742204785346985, + "learning_rate": 0.00019553730420013987, + "loss": 0.1474, + "step": 4081 + }, + { + "epoch": 0.09568850835399122, + "grad_norm": 0.12386063486337662, + "learning_rate": 0.00019553512796307966, + "loss": 0.0404, + "step": 4082 + }, + { + "epoch": 0.09571194992879621, + "grad_norm": 0.2999393343925476, + "learning_rate": 0.00019553295120764314, + "loss": 0.0386, + "step": 4083 + }, + { + "epoch": 0.0957353915036012, + "grad_norm": 0.6356807947158813, + "learning_rate": 0.00019553077393384216, + "loss": 0.1998, + "step": 4084 + }, + { + "epoch": 0.09575883307840621, + "grad_norm": 1.0155527591705322, + "learning_rate": 0.0001955285961416885, + "loss": 0.304, + "step": 4085 + }, + { + "epoch": 0.09578227465321121, + "grad_norm": 0.5772472620010376, + "learning_rate": 0.000195526417831194, + "loss": 0.1444, + "step": 4086 + }, + { + "epoch": 0.0958057162280162, + "grad_norm": 0.7987528443336487, + "learning_rate": 0.00019552423900237047, + "loss": 0.1167, + "step": 4087 + }, + { + "epoch": 0.0958291578028212, + "grad_norm": 0.543113648891449, + "learning_rate": 0.0001955220596552297, + "loss": 0.135, + "step": 4088 + }, + { + "epoch": 0.09585259937762619, + "grad_norm": 0.6126608848571777, + "learning_rate": 0.00019551987978978362, + "loss": 0.1268, + "step": 4089 + }, + { + "epoch": 0.09587604095243119, + "grad_norm": 0.614715039730072, + "learning_rate": 0.0001955176994060439, + "loss": 0.1835, + "step": 4090 + }, + { + "epoch": 0.09589948252723618, + "grad_norm": 0.3781900703907013, + "learning_rate": 0.00019551551850402254, + "loss": 0.0811, + "step": 4091 + }, + { + "epoch": 0.09592292410204117, + "grad_norm": 0.48885366320610046, + "learning_rate": 0.00019551333708373126, + "loss": 0.1314, + "step": 4092 + }, + { + "epoch": 0.09594636567684617, + "grad_norm": 0.8411356806755066, + "learning_rate": 0.0001955111551451819, + "loss": 0.2092, + "step": 4093 + }, + { + "epoch": 0.09596980725165116, + "grad_norm": 0.5621843934059143, + "learning_rate": 0.00019550897268838635, + "loss": 0.1051, + "step": 4094 + }, + { + "epoch": 0.09599324882645616, + "grad_norm": 0.49518445134162903, + "learning_rate": 0.00019550678971335644, + "loss": 0.1304, + "step": 4095 + }, + { + "epoch": 0.09601669040126115, + "grad_norm": 0.5542749762535095, + "learning_rate": 0.00019550460622010397, + "loss": 0.2249, + "step": 4096 + }, + { + "epoch": 0.09604013197606615, + "grad_norm": 0.7297506332397461, + "learning_rate": 0.00019550242220864086, + "loss": 0.2168, + "step": 4097 + }, + { + "epoch": 0.09606357355087115, + "grad_norm": 0.3275330364704132, + "learning_rate": 0.00019550023767897888, + "loss": 0.1031, + "step": 4098 + }, + { + "epoch": 0.09608701512567615, + "grad_norm": 0.5919314622879028, + "learning_rate": 0.00019549805263112995, + "loss": 0.2142, + "step": 4099 + }, + { + "epoch": 0.09611045670048114, + "grad_norm": 0.6528933644294739, + "learning_rate": 0.00019549586706510592, + "loss": 0.8983, + "step": 4100 + }, + { + "epoch": 0.09613389827528614, + "grad_norm": 0.23511068522930145, + "learning_rate": 0.0001954936809809186, + "loss": 0.036, + "step": 4101 + }, + { + "epoch": 0.09615733985009113, + "grad_norm": 0.7569811344146729, + "learning_rate": 0.0001954914943785799, + "loss": 0.2323, + "step": 4102 + }, + { + "epoch": 0.09618078142489613, + "grad_norm": 0.5265446305274963, + "learning_rate": 0.00019548930725810166, + "loss": 0.1326, + "step": 4103 + }, + { + "epoch": 0.09620422299970112, + "grad_norm": 0.5428662896156311, + "learning_rate": 0.00019548711961949575, + "loss": 0.1153, + "step": 4104 + }, + { + "epoch": 0.09622766457450611, + "grad_norm": 0.24734285473823547, + "learning_rate": 0.00019548493146277402, + "loss": 0.074, + "step": 4105 + }, + { + "epoch": 0.09625110614931111, + "grad_norm": 0.4183102250099182, + "learning_rate": 0.00019548274278794841, + "loss": 0.1231, + "step": 4106 + }, + { + "epoch": 0.0962745477241161, + "grad_norm": 0.09307616949081421, + "learning_rate": 0.00019548055359503073, + "loss": 0.0261, + "step": 4107 + }, + { + "epoch": 0.0962979892989211, + "grad_norm": 0.5760511159896851, + "learning_rate": 0.00019547836388403292, + "loss": 0.1388, + "step": 4108 + }, + { + "epoch": 0.09632143087372609, + "grad_norm": 0.572966456413269, + "learning_rate": 0.00019547617365496675, + "loss": 0.5709, + "step": 4109 + }, + { + "epoch": 0.0963448724485311, + "grad_norm": 0.2656494081020355, + "learning_rate": 0.00019547398290784424, + "loss": 0.0573, + "step": 4110 + }, + { + "epoch": 0.0963683140233361, + "grad_norm": 0.5610581040382385, + "learning_rate": 0.0001954717916426772, + "loss": 0.1465, + "step": 4111 + }, + { + "epoch": 0.09639175559814109, + "grad_norm": 0.17865264415740967, + "learning_rate": 0.0001954695998594775, + "loss": 0.0454, + "step": 4112 + }, + { + "epoch": 0.09641519717294608, + "grad_norm": 0.26951685547828674, + "learning_rate": 0.0001954674075582571, + "loss": 0.0619, + "step": 4113 + }, + { + "epoch": 0.09643863874775108, + "grad_norm": 0.5509044528007507, + "learning_rate": 0.00019546521473902786, + "loss": 0.1538, + "step": 4114 + }, + { + "epoch": 0.09646208032255607, + "grad_norm": 0.3038719892501831, + "learning_rate": 0.00019546302140180166, + "loss": 0.11, + "step": 4115 + }, + { + "epoch": 0.09648552189736107, + "grad_norm": 0.6590532660484314, + "learning_rate": 0.00019546082754659045, + "loss": 0.208, + "step": 4116 + }, + { + "epoch": 0.09650896347216606, + "grad_norm": 0.6448440551757812, + "learning_rate": 0.0001954586331734061, + "loss": 0.1757, + "step": 4117 + }, + { + "epoch": 0.09653240504697105, + "grad_norm": 0.3829464912414551, + "learning_rate": 0.00019545643828226052, + "loss": 0.0927, + "step": 4118 + }, + { + "epoch": 0.09655584662177605, + "grad_norm": 0.6610739231109619, + "learning_rate": 0.0001954542428731656, + "loss": 0.1901, + "step": 4119 + }, + { + "epoch": 0.09657928819658104, + "grad_norm": 0.20454145967960358, + "learning_rate": 0.0001954520469461333, + "loss": 0.0313, + "step": 4120 + }, + { + "epoch": 0.09660272977138604, + "grad_norm": 0.5000278949737549, + "learning_rate": 0.0001954498505011755, + "loss": 0.1249, + "step": 4121 + }, + { + "epoch": 0.09662617134619103, + "grad_norm": 0.1295481026172638, + "learning_rate": 0.0001954476535383041, + "loss": 0.0166, + "step": 4122 + }, + { + "epoch": 0.09664961292099604, + "grad_norm": 0.4327109158039093, + "learning_rate": 0.0001954454560575311, + "loss": 0.074, + "step": 4123 + }, + { + "epoch": 0.09667305449580103, + "grad_norm": 0.516372799873352, + "learning_rate": 0.00019544325805886836, + "loss": 0.1116, + "step": 4124 + }, + { + "epoch": 0.09669649607060603, + "grad_norm": 0.5380634069442749, + "learning_rate": 0.0001954410595423278, + "loss": 0.3766, + "step": 4125 + }, + { + "epoch": 0.09671993764541102, + "grad_norm": 0.11568639427423477, + "learning_rate": 0.00019543886050792138, + "loss": 0.0193, + "step": 4126 + }, + { + "epoch": 0.09674337922021602, + "grad_norm": 0.21619904041290283, + "learning_rate": 0.000195436660955661, + "loss": 0.024, + "step": 4127 + }, + { + "epoch": 0.09676682079502101, + "grad_norm": 0.7163117527961731, + "learning_rate": 0.00019543446088555864, + "loss": 0.0634, + "step": 4128 + }, + { + "epoch": 0.096790262369826, + "grad_norm": 0.43075117468833923, + "learning_rate": 0.00019543226029762617, + "loss": 0.0439, + "step": 4129 + }, + { + "epoch": 0.096813703944631, + "grad_norm": 0.8228789567947388, + "learning_rate": 0.00019543005919187562, + "loss": 0.1446, + "step": 4130 + }, + { + "epoch": 0.096837145519436, + "grad_norm": 1.0815190076828003, + "learning_rate": 0.00019542785756831887, + "loss": 0.1486, + "step": 4131 + }, + { + "epoch": 0.09686058709424099, + "grad_norm": 0.7637614011764526, + "learning_rate": 0.00019542565542696787, + "loss": 0.2742, + "step": 4132 + }, + { + "epoch": 0.09688402866904598, + "grad_norm": 0.62015300989151, + "learning_rate": 0.0001954234527678346, + "loss": 0.2816, + "step": 4133 + }, + { + "epoch": 0.09690747024385098, + "grad_norm": 1.0958157777786255, + "learning_rate": 0.00019542124959093097, + "loss": 0.2599, + "step": 4134 + }, + { + "epoch": 0.09693091181865597, + "grad_norm": 0.6005454659461975, + "learning_rate": 0.00019541904589626897, + "loss": 0.2054, + "step": 4135 + }, + { + "epoch": 0.09695435339346098, + "grad_norm": 0.7742257714271545, + "learning_rate": 0.00019541684168386056, + "loss": 0.1839, + "step": 4136 + }, + { + "epoch": 0.09697779496826597, + "grad_norm": 0.6132038831710815, + "learning_rate": 0.00019541463695371767, + "loss": 0.1215, + "step": 4137 + }, + { + "epoch": 0.09700123654307097, + "grad_norm": 0.36026352643966675, + "learning_rate": 0.00019541243170585225, + "loss": 0.0887, + "step": 4138 + }, + { + "epoch": 0.09702467811787596, + "grad_norm": 0.5497832298278809, + "learning_rate": 0.0001954102259402763, + "loss": 0.1275, + "step": 4139 + }, + { + "epoch": 0.09704811969268096, + "grad_norm": 0.603851318359375, + "learning_rate": 0.00019540801965700185, + "loss": 0.1266, + "step": 4140 + }, + { + "epoch": 0.09707156126748595, + "grad_norm": 0.921351432800293, + "learning_rate": 0.00019540581285604073, + "loss": 0.2415, + "step": 4141 + }, + { + "epoch": 0.09709500284229095, + "grad_norm": 0.6817610859870911, + "learning_rate": 0.000195403605537405, + "loss": 0.8855, + "step": 4142 + }, + { + "epoch": 0.09711844441709594, + "grad_norm": 0.6888790130615234, + "learning_rate": 0.0001954013977011066, + "loss": 0.2369, + "step": 4143 + }, + { + "epoch": 0.09714188599190093, + "grad_norm": 0.18263976275920868, + "learning_rate": 0.0001953991893471576, + "loss": 0.0488, + "step": 4144 + }, + { + "epoch": 0.09716532756670593, + "grad_norm": 0.728890597820282, + "learning_rate": 0.0001953969804755699, + "loss": 0.1851, + "step": 4145 + }, + { + "epoch": 0.09718876914151092, + "grad_norm": 0.5864750742912292, + "learning_rate": 0.00019539477108635546, + "loss": 0.1526, + "step": 4146 + }, + { + "epoch": 0.09721221071631592, + "grad_norm": 0.554291307926178, + "learning_rate": 0.00019539256117952634, + "loss": 0.1208, + "step": 4147 + }, + { + "epoch": 0.09723565229112091, + "grad_norm": 0.4422803819179535, + "learning_rate": 0.0001953903507550945, + "loss": 0.113, + "step": 4148 + }, + { + "epoch": 0.09725909386592592, + "grad_norm": 0.41577672958374023, + "learning_rate": 0.0001953881398130719, + "loss": 0.0907, + "step": 4149 + }, + { + "epoch": 0.09728253544073091, + "grad_norm": 0.5782780051231384, + "learning_rate": 0.0001953859283534706, + "loss": 0.1931, + "step": 4150 + }, + { + "epoch": 0.09730597701553591, + "grad_norm": 0.3147791624069214, + "learning_rate": 0.0001953837163763026, + "loss": 0.0614, + "step": 4151 + }, + { + "epoch": 0.0973294185903409, + "grad_norm": 0.4480401873588562, + "learning_rate": 0.00019538150388157983, + "loss": 0.1363, + "step": 4152 + }, + { + "epoch": 0.0973528601651459, + "grad_norm": 1.2705785036087036, + "learning_rate": 0.00019537929086931434, + "loss": 0.2878, + "step": 4153 + }, + { + "epoch": 0.09737630173995089, + "grad_norm": 0.8070724606513977, + "learning_rate": 0.00019537707733951814, + "loss": 0.2696, + "step": 4154 + }, + { + "epoch": 0.09739974331475588, + "grad_norm": 0.5087562799453735, + "learning_rate": 0.00019537486329220324, + "loss": 0.1562, + "step": 4155 + }, + { + "epoch": 0.09742318488956088, + "grad_norm": 0.6334993243217468, + "learning_rate": 0.00019537264872738165, + "loss": 0.1849, + "step": 4156 + }, + { + "epoch": 0.09744662646436587, + "grad_norm": 0.7669090628623962, + "learning_rate": 0.0001953704336450654, + "loss": 0.1889, + "step": 4157 + }, + { + "epoch": 0.09747006803917087, + "grad_norm": 0.40592339634895325, + "learning_rate": 0.0001953682180452665, + "loss": 0.1055, + "step": 4158 + }, + { + "epoch": 0.09749350961397586, + "grad_norm": 0.5892220735549927, + "learning_rate": 0.00019536600192799695, + "loss": 0.2022, + "step": 4159 + }, + { + "epoch": 0.09751695118878086, + "grad_norm": 0.6118261814117432, + "learning_rate": 0.00019536378529326877, + "loss": 0.187, + "step": 4160 + }, + { + "epoch": 0.09754039276358585, + "grad_norm": 0.646239697933197, + "learning_rate": 0.00019536156814109405, + "loss": 0.153, + "step": 4161 + }, + { + "epoch": 0.09756383433839086, + "grad_norm": 0.5197877287864685, + "learning_rate": 0.00019535935047148477, + "loss": 0.1605, + "step": 4162 + }, + { + "epoch": 0.09758727591319585, + "grad_norm": 0.17431877553462982, + "learning_rate": 0.00019535713228445296, + "loss": 0.0616, + "step": 4163 + }, + { + "epoch": 0.09761071748800085, + "grad_norm": 0.15882320702075958, + "learning_rate": 0.00019535491358001067, + "loss": 0.0364, + "step": 4164 + }, + { + "epoch": 0.09763415906280584, + "grad_norm": 0.37771090865135193, + "learning_rate": 0.00019535269435816996, + "loss": 0.0759, + "step": 4165 + }, + { + "epoch": 0.09765760063761084, + "grad_norm": 0.17454960942268372, + "learning_rate": 0.0001953504746189428, + "loss": 0.0385, + "step": 4166 + }, + { + "epoch": 0.09768104221241583, + "grad_norm": 0.3614146113395691, + "learning_rate": 0.00019534825436234133, + "loss": 0.0996, + "step": 4167 + }, + { + "epoch": 0.09770448378722082, + "grad_norm": 0.20126238465309143, + "learning_rate": 0.00019534603358837755, + "loss": 0.0227, + "step": 4168 + }, + { + "epoch": 0.09772792536202582, + "grad_norm": 0.16078515350818634, + "learning_rate": 0.0001953438122970635, + "loss": 0.043, + "step": 4169 + }, + { + "epoch": 0.09775136693683081, + "grad_norm": 0.30929577350616455, + "learning_rate": 0.00019534159048841123, + "loss": 0.0989, + "step": 4170 + }, + { + "epoch": 0.09777480851163581, + "grad_norm": 0.28995758295059204, + "learning_rate": 0.00019533936816243283, + "loss": 0.0533, + "step": 4171 + }, + { + "epoch": 0.0977982500864408, + "grad_norm": 0.5843889713287354, + "learning_rate": 0.00019533714531914034, + "loss": 0.8451, + "step": 4172 + }, + { + "epoch": 0.0978216916612458, + "grad_norm": 0.20761266350746155, + "learning_rate": 0.0001953349219585458, + "loss": 0.0214, + "step": 4173 + }, + { + "epoch": 0.0978451332360508, + "grad_norm": 0.5389126539230347, + "learning_rate": 0.0001953326980806613, + "loss": 0.1505, + "step": 4174 + }, + { + "epoch": 0.0978685748108558, + "grad_norm": 0.47879931330680847, + "learning_rate": 0.0001953304736854989, + "loss": 0.1386, + "step": 4175 + }, + { + "epoch": 0.0978920163856608, + "grad_norm": 0.7218381762504578, + "learning_rate": 0.00019532824877307068, + "loss": 0.1445, + "step": 4176 + }, + { + "epoch": 0.09791545796046579, + "grad_norm": 0.3536318242549896, + "learning_rate": 0.00019532602334338867, + "loss": 0.059, + "step": 4177 + }, + { + "epoch": 0.09793889953527078, + "grad_norm": 0.15141277015209198, + "learning_rate": 0.000195323797396465, + "loss": 0.0363, + "step": 4178 + }, + { + "epoch": 0.09796234111007578, + "grad_norm": 0.916733980178833, + "learning_rate": 0.00019532157093231174, + "loss": 0.1517, + "step": 4179 + }, + { + "epoch": 0.09798578268488077, + "grad_norm": 0.35351717472076416, + "learning_rate": 0.00019531934395094095, + "loss": 0.0618, + "step": 4180 + }, + { + "epoch": 0.09800922425968576, + "grad_norm": 0.19741974771022797, + "learning_rate": 0.0001953171164523647, + "loss": 0.0342, + "step": 4181 + }, + { + "epoch": 0.09803266583449076, + "grad_norm": 0.9572240710258484, + "learning_rate": 0.00019531488843659508, + "loss": 0.7609, + "step": 4182 + }, + { + "epoch": 0.09805610740929575, + "grad_norm": 0.5569593906402588, + "learning_rate": 0.00019531265990364424, + "loss": 0.1366, + "step": 4183 + }, + { + "epoch": 0.09807954898410075, + "grad_norm": 0.6445264220237732, + "learning_rate": 0.00019531043085352422, + "loss": 0.1365, + "step": 4184 + }, + { + "epoch": 0.09810299055890574, + "grad_norm": 0.7353662252426147, + "learning_rate": 0.0001953082012862471, + "loss": 0.2553, + "step": 4185 + }, + { + "epoch": 0.09812643213371074, + "grad_norm": 0.5095255374908447, + "learning_rate": 0.00019530597120182502, + "loss": 0.1132, + "step": 4186 + }, + { + "epoch": 0.09814987370851574, + "grad_norm": 0.26974815130233765, + "learning_rate": 0.00019530374060027005, + "loss": 0.0512, + "step": 4187 + }, + { + "epoch": 0.09817331528332074, + "grad_norm": 0.5831702947616577, + "learning_rate": 0.0001953015094815943, + "loss": 0.1924, + "step": 4188 + }, + { + "epoch": 0.09819675685812573, + "grad_norm": 0.7845077514648438, + "learning_rate": 0.00019529927784580988, + "loss": 0.1504, + "step": 4189 + }, + { + "epoch": 0.09822019843293073, + "grad_norm": 0.6649360060691833, + "learning_rate": 0.00019529704569292892, + "loss": 0.1274, + "step": 4190 + }, + { + "epoch": 0.09824364000773572, + "grad_norm": 0.487766295671463, + "learning_rate": 0.00019529481302296348, + "loss": 0.1086, + "step": 4191 + }, + { + "epoch": 0.09826708158254072, + "grad_norm": 0.22090737521648407, + "learning_rate": 0.0001952925798359257, + "loss": 0.0419, + "step": 4192 + }, + { + "epoch": 0.09829052315734571, + "grad_norm": 0.19454112648963928, + "learning_rate": 0.00019529034613182774, + "loss": 0.0463, + "step": 4193 + }, + { + "epoch": 0.0983139647321507, + "grad_norm": 0.18549387156963348, + "learning_rate": 0.00019528811191068165, + "loss": 0.0704, + "step": 4194 + }, + { + "epoch": 0.0983374063069557, + "grad_norm": 0.20983931422233582, + "learning_rate": 0.0001952858771724996, + "loss": 0.044, + "step": 4195 + }, + { + "epoch": 0.0983608478817607, + "grad_norm": 0.7421796321868896, + "learning_rate": 0.00019528364191729371, + "loss": 0.1932, + "step": 4196 + }, + { + "epoch": 0.09838428945656569, + "grad_norm": 0.6336399912834167, + "learning_rate": 0.0001952814061450761, + "loss": 0.1834, + "step": 4197 + }, + { + "epoch": 0.09840773103137068, + "grad_norm": 0.09817273169755936, + "learning_rate": 0.00019527916985585887, + "loss": 0.0219, + "step": 4198 + }, + { + "epoch": 0.09843117260617568, + "grad_norm": 0.5713086724281311, + "learning_rate": 0.00019527693304965423, + "loss": 0.106, + "step": 4199 + }, + { + "epoch": 0.09845461418098068, + "grad_norm": 0.2306801974773407, + "learning_rate": 0.00019527469572647426, + "loss": 0.0244, + "step": 4200 + }, + { + "epoch": 0.09847805575578568, + "grad_norm": 0.18061509728431702, + "learning_rate": 0.0001952724578863311, + "loss": 0.0463, + "step": 4201 + }, + { + "epoch": 0.09850149733059067, + "grad_norm": 0.7806215286254883, + "learning_rate": 0.0001952702195292369, + "loss": 0.1738, + "step": 4202 + }, + { + "epoch": 0.09852493890539567, + "grad_norm": 0.6048714518547058, + "learning_rate": 0.0001952679806552038, + "loss": 0.6738, + "step": 4203 + }, + { + "epoch": 0.09854838048020066, + "grad_norm": 0.11905292421579361, + "learning_rate": 0.00019526574126424396, + "loss": 0.0292, + "step": 4204 + }, + { + "epoch": 0.09857182205500566, + "grad_norm": 0.6141340136528015, + "learning_rate": 0.00019526350135636955, + "loss": 0.1696, + "step": 4205 + }, + { + "epoch": 0.09859526362981065, + "grad_norm": 0.6078602075576782, + "learning_rate": 0.0001952612609315927, + "loss": 0.8637, + "step": 4206 + }, + { + "epoch": 0.09861870520461564, + "grad_norm": 0.3023436367511749, + "learning_rate": 0.00019525901998992558, + "loss": 0.0762, + "step": 4207 + }, + { + "epoch": 0.09864214677942064, + "grad_norm": 0.803372859954834, + "learning_rate": 0.0001952567785313803, + "loss": 0.2488, + "step": 4208 + }, + { + "epoch": 0.09866558835422563, + "grad_norm": 0.562396764755249, + "learning_rate": 0.00019525453655596913, + "loss": 0.1169, + "step": 4209 + }, + { + "epoch": 0.09868902992903063, + "grad_norm": 0.16103914380073547, + "learning_rate": 0.00019525229406370409, + "loss": 0.044, + "step": 4210 + }, + { + "epoch": 0.09871247150383562, + "grad_norm": 0.5973129868507385, + "learning_rate": 0.00019525005105459746, + "loss": 0.1142, + "step": 4211 + }, + { + "epoch": 0.09873591307864062, + "grad_norm": 0.2374761998653412, + "learning_rate": 0.00019524780752866138, + "loss": 0.0596, + "step": 4212 + }, + { + "epoch": 0.09875935465344562, + "grad_norm": 0.17027537524700165, + "learning_rate": 0.000195245563485908, + "loss": 0.0517, + "step": 4213 + }, + { + "epoch": 0.09878279622825062, + "grad_norm": 0.952194333076477, + "learning_rate": 0.00019524331892634952, + "loss": 0.3779, + "step": 4214 + }, + { + "epoch": 0.09880623780305561, + "grad_norm": 0.5134127140045166, + "learning_rate": 0.00019524107384999813, + "loss": 0.0586, + "step": 4215 + }, + { + "epoch": 0.09882967937786061, + "grad_norm": 0.45916101336479187, + "learning_rate": 0.00019523882825686598, + "loss": 0.1854, + "step": 4216 + }, + { + "epoch": 0.0988531209526656, + "grad_norm": 0.4108150601387024, + "learning_rate": 0.00019523658214696526, + "loss": 0.0967, + "step": 4217 + }, + { + "epoch": 0.0988765625274706, + "grad_norm": 0.43756532669067383, + "learning_rate": 0.0001952343355203082, + "loss": 0.0774, + "step": 4218 + }, + { + "epoch": 0.09890000410227559, + "grad_norm": 0.883324384689331, + "learning_rate": 0.00019523208837690694, + "loss": 0.1668, + "step": 4219 + }, + { + "epoch": 0.09892344567708058, + "grad_norm": 0.06424044072628021, + "learning_rate": 0.00019522984071677367, + "loss": 0.0275, + "step": 4220 + }, + { + "epoch": 0.09894688725188558, + "grad_norm": 0.4909697473049164, + "learning_rate": 0.00019522759253992063, + "loss": 0.1406, + "step": 4221 + }, + { + "epoch": 0.09897032882669057, + "grad_norm": 0.5075114965438843, + "learning_rate": 0.00019522534384636, + "loss": 0.1714, + "step": 4222 + }, + { + "epoch": 0.09899377040149557, + "grad_norm": 0.5469490885734558, + "learning_rate": 0.00019522309463610398, + "loss": 0.1108, + "step": 4223 + }, + { + "epoch": 0.09901721197630056, + "grad_norm": 0.5414002537727356, + "learning_rate": 0.00019522084490916476, + "loss": 0.1768, + "step": 4224 + }, + { + "epoch": 0.09904065355110556, + "grad_norm": 0.6407678723335266, + "learning_rate": 0.00019521859466555458, + "loss": 0.1141, + "step": 4225 + }, + { + "epoch": 0.09906409512591056, + "grad_norm": 0.6047683358192444, + "learning_rate": 0.00019521634390528558, + "loss": 0.7686, + "step": 4226 + }, + { + "epoch": 0.09908753670071556, + "grad_norm": 0.5063349604606628, + "learning_rate": 0.00019521409262837004, + "loss": 0.158, + "step": 4227 + }, + { + "epoch": 0.09911097827552055, + "grad_norm": 0.31972965598106384, + "learning_rate": 0.00019521184083482016, + "loss": 0.0636, + "step": 4228 + }, + { + "epoch": 0.09913441985032555, + "grad_norm": 0.4194682836532593, + "learning_rate": 0.00019520958852464816, + "loss": 0.1331, + "step": 4229 + }, + { + "epoch": 0.09915786142513054, + "grad_norm": 0.5380992293357849, + "learning_rate": 0.00019520733569786626, + "loss": 0.2308, + "step": 4230 + }, + { + "epoch": 0.09918130299993554, + "grad_norm": 0.5970016717910767, + "learning_rate": 0.00019520508235448667, + "loss": 0.8496, + "step": 4231 + }, + { + "epoch": 0.09920474457474053, + "grad_norm": 0.5628905892372131, + "learning_rate": 0.00019520282849452164, + "loss": 0.196, + "step": 4232 + }, + { + "epoch": 0.09922818614954552, + "grad_norm": 1.0105414390563965, + "learning_rate": 0.00019520057411798338, + "loss": 0.2576, + "step": 4233 + }, + { + "epoch": 0.09925162772435052, + "grad_norm": 0.5917713642120361, + "learning_rate": 0.00019519831922488413, + "loss": 0.115, + "step": 4234 + }, + { + "epoch": 0.09927506929915551, + "grad_norm": 0.13572923839092255, + "learning_rate": 0.0001951960638152361, + "loss": 0.049, + "step": 4235 + }, + { + "epoch": 0.09929851087396051, + "grad_norm": 0.5403391718864441, + "learning_rate": 0.00019519380788905158, + "loss": 0.2075, + "step": 4236 + }, + { + "epoch": 0.0993219524487655, + "grad_norm": 0.21205484867095947, + "learning_rate": 0.00019519155144634277, + "loss": 0.0499, + "step": 4237 + }, + { + "epoch": 0.09934539402357051, + "grad_norm": 0.5263277292251587, + "learning_rate": 0.0001951892944871219, + "loss": 0.0865, + "step": 4238 + }, + { + "epoch": 0.0993688355983755, + "grad_norm": 0.2063051015138626, + "learning_rate": 0.00019518703701140125, + "loss": 0.0614, + "step": 4239 + }, + { + "epoch": 0.0993922771731805, + "grad_norm": 0.49973994493484497, + "learning_rate": 0.0001951847790191931, + "loss": 0.1293, + "step": 4240 + }, + { + "epoch": 0.0994157187479855, + "grad_norm": 0.3416864573955536, + "learning_rate": 0.00019518252051050963, + "loss": 0.0884, + "step": 4241 + }, + { + "epoch": 0.09943916032279049, + "grad_norm": 0.68593829870224, + "learning_rate": 0.00019518026148536316, + "loss": 0.2265, + "step": 4242 + }, + { + "epoch": 0.09946260189759548, + "grad_norm": 1.094049096107483, + "learning_rate": 0.00019517800194376585, + "loss": 0.3063, + "step": 4243 + }, + { + "epoch": 0.09948604347240048, + "grad_norm": 0.14452117681503296, + "learning_rate": 0.00019517574188573008, + "loss": 0.0261, + "step": 4244 + }, + { + "epoch": 0.09950948504720547, + "grad_norm": 0.8971830010414124, + "learning_rate": 0.00019517348131126806, + "loss": 0.1589, + "step": 4245 + }, + { + "epoch": 0.09953292662201046, + "grad_norm": 0.5847511887550354, + "learning_rate": 0.00019517122022039203, + "loss": 0.1708, + "step": 4246 + }, + { + "epoch": 0.09955636819681546, + "grad_norm": 0.3813796937465668, + "learning_rate": 0.00019516895861311427, + "loss": 0.1649, + "step": 4247 + }, + { + "epoch": 0.09957980977162045, + "grad_norm": 0.5563102960586548, + "learning_rate": 0.0001951666964894471, + "loss": 0.1108, + "step": 4248 + }, + { + "epoch": 0.09960325134642545, + "grad_norm": 0.6155687570571899, + "learning_rate": 0.00019516443384940275, + "loss": 0.191, + "step": 4249 + }, + { + "epoch": 0.09962669292123044, + "grad_norm": 0.33528903126716614, + "learning_rate": 0.00019516217069299348, + "loss": 0.1005, + "step": 4250 + }, + { + "epoch": 0.09965013449603545, + "grad_norm": 0.7464436888694763, + "learning_rate": 0.00019515990702023163, + "loss": 0.2218, + "step": 4251 + }, + { + "epoch": 0.09967357607084044, + "grad_norm": 0.30715349316596985, + "learning_rate": 0.00019515764283112945, + "loss": 0.073, + "step": 4252 + }, + { + "epoch": 0.09969701764564544, + "grad_norm": 0.6551823616027832, + "learning_rate": 0.0001951553781256992, + "loss": 0.1352, + "step": 4253 + }, + { + "epoch": 0.09972045922045043, + "grad_norm": 0.7410808801651001, + "learning_rate": 0.00019515311290395322, + "loss": 0.2239, + "step": 4254 + }, + { + "epoch": 0.09974390079525543, + "grad_norm": 0.4800601005554199, + "learning_rate": 0.00019515084716590375, + "loss": 0.146, + "step": 4255 + }, + { + "epoch": 0.09976734237006042, + "grad_norm": 0.5421104431152344, + "learning_rate": 0.00019514858091156312, + "loss": 0.1773, + "step": 4256 + }, + { + "epoch": 0.09979078394486542, + "grad_norm": 0.7920913100242615, + "learning_rate": 0.0001951463141409436, + "loss": 0.1338, + "step": 4257 + }, + { + "epoch": 0.09981422551967041, + "grad_norm": 0.15559229254722595, + "learning_rate": 0.0001951440468540575, + "loss": 0.0482, + "step": 4258 + }, + { + "epoch": 0.0998376670944754, + "grad_norm": 0.3019532263278961, + "learning_rate": 0.00019514177905091717, + "loss": 0.0759, + "step": 4259 + }, + { + "epoch": 0.0998611086692804, + "grad_norm": 0.7349664568901062, + "learning_rate": 0.00019513951073153485, + "loss": 0.1725, + "step": 4260 + }, + { + "epoch": 0.0998845502440854, + "grad_norm": 0.6777001023292542, + "learning_rate": 0.00019513724189592285, + "loss": 0.1824, + "step": 4261 + }, + { + "epoch": 0.09990799181889039, + "grad_norm": 0.15168669819831848, + "learning_rate": 0.0001951349725440935, + "loss": 0.0473, + "step": 4262 + }, + { + "epoch": 0.09993143339369538, + "grad_norm": 0.46900802850723267, + "learning_rate": 0.00019513270267605915, + "loss": 0.1426, + "step": 4263 + }, + { + "epoch": 0.09995487496850039, + "grad_norm": 0.5750395059585571, + "learning_rate": 0.00019513043229183206, + "loss": 0.1795, + "step": 4264 + }, + { + "epoch": 0.09997831654330538, + "grad_norm": 0.9421368837356567, + "learning_rate": 0.00019512816139142456, + "loss": 0.1809, + "step": 4265 + }, + { + "epoch": 0.10000175811811038, + "grad_norm": 0.3806561827659607, + "learning_rate": 0.00019512588997484898, + "loss": 0.1005, + "step": 4266 + }, + { + "epoch": 0.10002519969291537, + "grad_norm": 0.6443682312965393, + "learning_rate": 0.00019512361804211766, + "loss": 0.1205, + "step": 4267 + }, + { + "epoch": 0.10004864126772037, + "grad_norm": 0.4419521987438202, + "learning_rate": 0.0001951213455932429, + "loss": 0.1672, + "step": 4268 + }, + { + "epoch": 0.10007208284252536, + "grad_norm": 0.5088673233985901, + "learning_rate": 0.00019511907262823706, + "loss": 0.1721, + "step": 4269 + }, + { + "epoch": 0.10009552441733036, + "grad_norm": 0.5282859206199646, + "learning_rate": 0.00019511679914711245, + "loss": 0.1719, + "step": 4270 + }, + { + "epoch": 0.10011896599213535, + "grad_norm": 0.7238675951957703, + "learning_rate": 0.00019511452514988143, + "loss": 0.222, + "step": 4271 + }, + { + "epoch": 0.10014240756694034, + "grad_norm": 0.14707940816879272, + "learning_rate": 0.0001951122506365563, + "loss": 0.0298, + "step": 4272 + }, + { + "epoch": 0.10016584914174534, + "grad_norm": 0.2908579707145691, + "learning_rate": 0.0001951099756071494, + "loss": 0.0457, + "step": 4273 + }, + { + "epoch": 0.10018929071655033, + "grad_norm": 0.5345767736434937, + "learning_rate": 0.00019510770006167313, + "loss": 0.0847, + "step": 4274 + }, + { + "epoch": 0.10021273229135533, + "grad_norm": 0.6399059891700745, + "learning_rate": 0.00019510542400013982, + "loss": 0.1609, + "step": 4275 + }, + { + "epoch": 0.10023617386616032, + "grad_norm": 0.45083191990852356, + "learning_rate": 0.0001951031474225618, + "loss": 0.1214, + "step": 4276 + }, + { + "epoch": 0.10025961544096533, + "grad_norm": 0.21299931406974792, + "learning_rate": 0.0001951008703289514, + "loss": 0.0653, + "step": 4277 + }, + { + "epoch": 0.10028305701577032, + "grad_norm": 0.6438643336296082, + "learning_rate": 0.00019509859271932102, + "loss": 0.2399, + "step": 4278 + }, + { + "epoch": 0.10030649859057532, + "grad_norm": 0.5364943146705627, + "learning_rate": 0.00019509631459368302, + "loss": 0.1343, + "step": 4279 + }, + { + "epoch": 0.10032994016538031, + "grad_norm": 0.4979501962661743, + "learning_rate": 0.00019509403595204973, + "loss": 0.1499, + "step": 4280 + }, + { + "epoch": 0.10035338174018531, + "grad_norm": 0.5357359647750854, + "learning_rate": 0.0001950917567944335, + "loss": 0.6392, + "step": 4281 + }, + { + "epoch": 0.1003768233149903, + "grad_norm": 0.8075301051139832, + "learning_rate": 0.00019508947712084675, + "loss": 0.2915, + "step": 4282 + }, + { + "epoch": 0.1004002648897953, + "grad_norm": 0.7383580207824707, + "learning_rate": 0.00019508719693130182, + "loss": 0.176, + "step": 4283 + }, + { + "epoch": 0.10042370646460029, + "grad_norm": 0.7377861738204956, + "learning_rate": 0.00019508491622581109, + "loss": 0.1191, + "step": 4284 + }, + { + "epoch": 0.10044714803940528, + "grad_norm": 0.5218890905380249, + "learning_rate": 0.00019508263500438692, + "loss": 0.1162, + "step": 4285 + }, + { + "epoch": 0.10047058961421028, + "grad_norm": 0.47927141189575195, + "learning_rate": 0.00019508035326704171, + "loss": 0.1242, + "step": 4286 + }, + { + "epoch": 0.10049403118901527, + "grad_norm": 0.6812264919281006, + "learning_rate": 0.00019507807101378785, + "loss": 0.654, + "step": 4287 + }, + { + "epoch": 0.10051747276382027, + "grad_norm": 0.6842209696769714, + "learning_rate": 0.00019507578824463762, + "loss": 0.2058, + "step": 4288 + }, + { + "epoch": 0.10054091433862528, + "grad_norm": 0.4788963198661804, + "learning_rate": 0.00019507350495960357, + "loss": 0.0792, + "step": 4289 + }, + { + "epoch": 0.10056435591343027, + "grad_norm": 0.18690598011016846, + "learning_rate": 0.000195071221158698, + "loss": 0.0679, + "step": 4290 + }, + { + "epoch": 0.10058779748823526, + "grad_norm": 0.23462751507759094, + "learning_rate": 0.0001950689368419333, + "loss": 0.063, + "step": 4291 + }, + { + "epoch": 0.10061123906304026, + "grad_norm": 1.1884474754333496, + "learning_rate": 0.00019506665200932186, + "loss": 0.1725, + "step": 4292 + }, + { + "epoch": 0.10063468063784525, + "grad_norm": 0.3272130787372589, + "learning_rate": 0.00019506436666087611, + "loss": 0.1208, + "step": 4293 + }, + { + "epoch": 0.10065812221265025, + "grad_norm": 0.6633151769638062, + "learning_rate": 0.00019506208079660841, + "loss": 0.6171, + "step": 4294 + }, + { + "epoch": 0.10068156378745524, + "grad_norm": 1.2584980726242065, + "learning_rate": 0.00019505979441653123, + "loss": 0.3189, + "step": 4295 + }, + { + "epoch": 0.10070500536226024, + "grad_norm": 0.7191201448440552, + "learning_rate": 0.0001950575075206569, + "loss": 0.105, + "step": 4296 + }, + { + "epoch": 0.10072844693706523, + "grad_norm": 0.5015390515327454, + "learning_rate": 0.00019505522010899786, + "loss": 0.1424, + "step": 4297 + }, + { + "epoch": 0.10075188851187022, + "grad_norm": 0.4734789729118347, + "learning_rate": 0.00019505293218156656, + "loss": 0.0843, + "step": 4298 + }, + { + "epoch": 0.10077533008667522, + "grad_norm": 0.6270214915275574, + "learning_rate": 0.00019505064373837535, + "loss": 0.1976, + "step": 4299 + }, + { + "epoch": 0.10079877166148021, + "grad_norm": 0.7529721260070801, + "learning_rate": 0.00019504835477943667, + "loss": 0.2746, + "step": 4300 + }, + { + "epoch": 0.10082221323628521, + "grad_norm": 0.15119698643684387, + "learning_rate": 0.00019504606530476294, + "loss": 0.0141, + "step": 4301 + }, + { + "epoch": 0.10084565481109022, + "grad_norm": 0.7135301232337952, + "learning_rate": 0.0001950437753143666, + "loss": 0.8249, + "step": 4302 + }, + { + "epoch": 0.10086909638589521, + "grad_norm": 0.6135610938072205, + "learning_rate": 0.00019504148480826004, + "loss": 0.1418, + "step": 4303 + }, + { + "epoch": 0.1008925379607002, + "grad_norm": 0.708949089050293, + "learning_rate": 0.00019503919378645575, + "loss": 0.1543, + "step": 4304 + }, + { + "epoch": 0.1009159795355052, + "grad_norm": 0.45043912529945374, + "learning_rate": 0.00019503690224896613, + "loss": 0.1164, + "step": 4305 + }, + { + "epoch": 0.10093942111031019, + "grad_norm": 0.8494976758956909, + "learning_rate": 0.00019503461019580357, + "loss": 0.1553, + "step": 4306 + }, + { + "epoch": 0.10096286268511519, + "grad_norm": 0.45303502678871155, + "learning_rate": 0.00019503231762698053, + "loss": 0.1379, + "step": 4307 + }, + { + "epoch": 0.10098630425992018, + "grad_norm": 0.6366778016090393, + "learning_rate": 0.0001950300245425095, + "loss": 0.1955, + "step": 4308 + }, + { + "epoch": 0.10100974583472518, + "grad_norm": 0.5864023566246033, + "learning_rate": 0.00019502773094240287, + "loss": 0.0727, + "step": 4309 + }, + { + "epoch": 0.10103318740953017, + "grad_norm": 0.4877522587776184, + "learning_rate": 0.0001950254368266731, + "loss": 0.1246, + "step": 4310 + }, + { + "epoch": 0.10105662898433516, + "grad_norm": 0.35994017124176025, + "learning_rate": 0.0001950231421953326, + "loss": 0.112, + "step": 4311 + }, + { + "epoch": 0.10108007055914016, + "grad_norm": 0.5871253609657288, + "learning_rate": 0.00019502084704839395, + "loss": 0.1708, + "step": 4312 + }, + { + "epoch": 0.10110351213394515, + "grad_norm": 0.44327419996261597, + "learning_rate": 0.00019501855138586943, + "loss": 0.0956, + "step": 4313 + }, + { + "epoch": 0.10112695370875015, + "grad_norm": 0.1495332419872284, + "learning_rate": 0.00019501625520777161, + "loss": 0.0529, + "step": 4314 + }, + { + "epoch": 0.10115039528355516, + "grad_norm": 0.5561747550964355, + "learning_rate": 0.0001950139585141129, + "loss": 0.1285, + "step": 4315 + }, + { + "epoch": 0.10117383685836015, + "grad_norm": 0.6331071853637695, + "learning_rate": 0.00019501166130490584, + "loss": 0.199, + "step": 4316 + }, + { + "epoch": 0.10119727843316514, + "grad_norm": 0.5922608971595764, + "learning_rate": 0.00019500936358016275, + "loss": 0.1735, + "step": 4317 + }, + { + "epoch": 0.10122072000797014, + "grad_norm": 0.7683104276657104, + "learning_rate": 0.00019500706533989623, + "loss": 0.2752, + "step": 4318 + }, + { + "epoch": 0.10124416158277513, + "grad_norm": 0.6167133450508118, + "learning_rate": 0.00019500476658411867, + "loss": 0.1097, + "step": 4319 + }, + { + "epoch": 0.10126760315758013, + "grad_norm": 0.6609891057014465, + "learning_rate": 0.0001950024673128426, + "loss": 0.2216, + "step": 4320 + }, + { + "epoch": 0.10129104473238512, + "grad_norm": 0.21541565656661987, + "learning_rate": 0.0001950001675260805, + "loss": 0.093, + "step": 4321 + }, + { + "epoch": 0.10131448630719012, + "grad_norm": 0.5072643160820007, + "learning_rate": 0.0001949978672238448, + "loss": 0.234, + "step": 4322 + }, + { + "epoch": 0.10133792788199511, + "grad_norm": 0.4457538425922394, + "learning_rate": 0.00019499556640614795, + "loss": 0.1115, + "step": 4323 + }, + { + "epoch": 0.1013613694568001, + "grad_norm": 0.29225948452949524, + "learning_rate": 0.00019499326507300255, + "loss": 0.0835, + "step": 4324 + }, + { + "epoch": 0.1013848110316051, + "grad_norm": 0.4191027581691742, + "learning_rate": 0.000194990963224421, + "loss": 0.0948, + "step": 4325 + }, + { + "epoch": 0.1014082526064101, + "grad_norm": 0.20705178380012512, + "learning_rate": 0.0001949886608604158, + "loss": 0.0584, + "step": 4326 + }, + { + "epoch": 0.10143169418121509, + "grad_norm": 0.49851688742637634, + "learning_rate": 0.00019498635798099944, + "loss": 0.2037, + "step": 4327 + }, + { + "epoch": 0.1014551357560201, + "grad_norm": 0.17300444841384888, + "learning_rate": 0.00019498405458618447, + "loss": 0.06, + "step": 4328 + }, + { + "epoch": 0.10147857733082509, + "grad_norm": 0.36958783864974976, + "learning_rate": 0.00019498175067598334, + "loss": 0.1028, + "step": 4329 + }, + { + "epoch": 0.10150201890563008, + "grad_norm": 0.515782356262207, + "learning_rate": 0.00019497944625040855, + "loss": 0.1001, + "step": 4330 + }, + { + "epoch": 0.10152546048043508, + "grad_norm": 0.18924899399280548, + "learning_rate": 0.00019497714130947263, + "loss": 0.0737, + "step": 4331 + }, + { + "epoch": 0.10154890205524007, + "grad_norm": 0.4078608751296997, + "learning_rate": 0.00019497483585318806, + "loss": 0.0517, + "step": 4332 + }, + { + "epoch": 0.10157234363004507, + "grad_norm": 0.38454189896583557, + "learning_rate": 0.00019497252988156735, + "loss": 0.1189, + "step": 4333 + }, + { + "epoch": 0.10159578520485006, + "grad_norm": 0.5130324363708496, + "learning_rate": 0.00019497022339462304, + "loss": 0.0975, + "step": 4334 + }, + { + "epoch": 0.10161922677965506, + "grad_norm": 0.5204680562019348, + "learning_rate": 0.00019496791639236763, + "loss": 0.182, + "step": 4335 + }, + { + "epoch": 0.10164266835446005, + "grad_norm": 0.7257456183433533, + "learning_rate": 0.00019496560887481362, + "loss": 0.1951, + "step": 4336 + }, + { + "epoch": 0.10166610992926504, + "grad_norm": 0.6947634816169739, + "learning_rate": 0.00019496330084197355, + "loss": 0.1123, + "step": 4337 + }, + { + "epoch": 0.10168955150407004, + "grad_norm": 0.6758540272712708, + "learning_rate": 0.00019496099229385994, + "loss": 0.1945, + "step": 4338 + }, + { + "epoch": 0.10171299307887503, + "grad_norm": 0.18235193192958832, + "learning_rate": 0.00019495868323048533, + "loss": 0.0403, + "step": 4339 + }, + { + "epoch": 0.10173643465368003, + "grad_norm": 0.8071116209030151, + "learning_rate": 0.0001949563736518622, + "loss": 0.1522, + "step": 4340 + }, + { + "epoch": 0.10175987622848504, + "grad_norm": 0.7422962188720703, + "learning_rate": 0.00019495406355800316, + "loss": 0.1938, + "step": 4341 + }, + { + "epoch": 0.10178331780329003, + "grad_norm": 0.19806276261806488, + "learning_rate": 0.00019495175294892067, + "loss": 0.0563, + "step": 4342 + }, + { + "epoch": 0.10180675937809502, + "grad_norm": 0.5566330552101135, + "learning_rate": 0.00019494944182462732, + "loss": 0.7249, + "step": 4343 + }, + { + "epoch": 0.10183020095290002, + "grad_norm": 0.547298014163971, + "learning_rate": 0.00019494713018513562, + "loss": 0.1629, + "step": 4344 + }, + { + "epoch": 0.10185364252770501, + "grad_norm": 0.947387158870697, + "learning_rate": 0.0001949448180304581, + "loss": 0.1672, + "step": 4345 + }, + { + "epoch": 0.10187708410251001, + "grad_norm": 0.5555575489997864, + "learning_rate": 0.00019494250536060737, + "loss": 0.0909, + "step": 4346 + }, + { + "epoch": 0.101900525677315, + "grad_norm": 0.6981589198112488, + "learning_rate": 0.0001949401921755959, + "loss": 0.1352, + "step": 4347 + }, + { + "epoch": 0.10192396725212, + "grad_norm": 0.6308282017707825, + "learning_rate": 0.0001949378784754363, + "loss": 0.1167, + "step": 4348 + }, + { + "epoch": 0.10194740882692499, + "grad_norm": 0.34135720133781433, + "learning_rate": 0.0001949355642601411, + "loss": 0.1162, + "step": 4349 + }, + { + "epoch": 0.10197085040172998, + "grad_norm": 0.7169831395149231, + "learning_rate": 0.00019493324952972284, + "loss": 0.7223, + "step": 4350 + }, + { + "epoch": 0.10199429197653498, + "grad_norm": 0.2766970694065094, + "learning_rate": 0.00019493093428419412, + "loss": 0.075, + "step": 4351 + }, + { + "epoch": 0.10201773355133997, + "grad_norm": 0.8458486795425415, + "learning_rate": 0.00019492861852356746, + "loss": 0.1866, + "step": 4352 + }, + { + "epoch": 0.10204117512614498, + "grad_norm": 0.588150680065155, + "learning_rate": 0.00019492630224785546, + "loss": 0.1709, + "step": 4353 + }, + { + "epoch": 0.10206461670094998, + "grad_norm": 0.5155200958251953, + "learning_rate": 0.00019492398545707066, + "loss": 0.0755, + "step": 4354 + }, + { + "epoch": 0.10208805827575497, + "grad_norm": 0.4312809407711029, + "learning_rate": 0.00019492166815122567, + "loss": 0.0767, + "step": 4355 + }, + { + "epoch": 0.10211149985055996, + "grad_norm": 0.843212366104126, + "learning_rate": 0.00019491935033033304, + "loss": 0.1146, + "step": 4356 + }, + { + "epoch": 0.10213494142536496, + "grad_norm": 0.507084310054779, + "learning_rate": 0.00019491703199440535, + "loss": 0.1719, + "step": 4357 + }, + { + "epoch": 0.10215838300016995, + "grad_norm": 0.22630292177200317, + "learning_rate": 0.00019491471314345515, + "loss": 0.0545, + "step": 4358 + }, + { + "epoch": 0.10218182457497495, + "grad_norm": 0.28101977705955505, + "learning_rate": 0.00019491239377749506, + "loss": 0.0657, + "step": 4359 + }, + { + "epoch": 0.10220526614977994, + "grad_norm": 0.9265746474266052, + "learning_rate": 0.00019491007389653767, + "loss": 0.1058, + "step": 4360 + }, + { + "epoch": 0.10222870772458494, + "grad_norm": 0.7487101554870605, + "learning_rate": 0.00019490775350059552, + "loss": 0.1284, + "step": 4361 + }, + { + "epoch": 0.10225214929938993, + "grad_norm": 0.5548025965690613, + "learning_rate": 0.00019490543258968124, + "loss": 0.1606, + "step": 4362 + }, + { + "epoch": 0.10227559087419492, + "grad_norm": 1.02643883228302, + "learning_rate": 0.00019490311116380742, + "loss": 0.1735, + "step": 4363 + }, + { + "epoch": 0.10229903244899992, + "grad_norm": 0.6867809295654297, + "learning_rate": 0.00019490078922298665, + "loss": 0.0887, + "step": 4364 + }, + { + "epoch": 0.10232247402380491, + "grad_norm": 0.36911237239837646, + "learning_rate": 0.00019489846676723152, + "loss": 0.1203, + "step": 4365 + }, + { + "epoch": 0.10234591559860992, + "grad_norm": 0.6933668851852417, + "learning_rate": 0.00019489614379655465, + "loss": 0.1136, + "step": 4366 + }, + { + "epoch": 0.10236935717341492, + "grad_norm": 0.25340649485588074, + "learning_rate": 0.00019489382031096864, + "loss": 0.1711, + "step": 4367 + }, + { + "epoch": 0.10239279874821991, + "grad_norm": 0.7893999218940735, + "learning_rate": 0.00019489149631048608, + "loss": 0.2316, + "step": 4368 + }, + { + "epoch": 0.1024162403230249, + "grad_norm": 0.3812544047832489, + "learning_rate": 0.00019488917179511962, + "loss": 0.0616, + "step": 4369 + }, + { + "epoch": 0.1024396818978299, + "grad_norm": 0.6835368275642395, + "learning_rate": 0.0001948868467648818, + "loss": 0.1972, + "step": 4370 + }, + { + "epoch": 0.10246312347263489, + "grad_norm": 0.5411664843559265, + "learning_rate": 0.0001948845212197853, + "loss": 0.1371, + "step": 4371 + }, + { + "epoch": 0.10248656504743989, + "grad_norm": 0.17507600784301758, + "learning_rate": 0.00019488219515984275, + "loss": 0.0363, + "step": 4372 + }, + { + "epoch": 0.10251000662224488, + "grad_norm": 1.0291284322738647, + "learning_rate": 0.0001948798685850667, + "loss": 0.2715, + "step": 4373 + }, + { + "epoch": 0.10253344819704988, + "grad_norm": 0.9253339767456055, + "learning_rate": 0.00019487754149546984, + "loss": 0.0932, + "step": 4374 + }, + { + "epoch": 0.10255688977185487, + "grad_norm": 0.14515328407287598, + "learning_rate": 0.00019487521389106478, + "loss": 0.0401, + "step": 4375 + }, + { + "epoch": 0.10258033134665986, + "grad_norm": 1.0942903757095337, + "learning_rate": 0.0001948728857718641, + "loss": 0.2041, + "step": 4376 + }, + { + "epoch": 0.10260377292146486, + "grad_norm": 0.7633674740791321, + "learning_rate": 0.00019487055713788052, + "loss": 0.1706, + "step": 4377 + }, + { + "epoch": 0.10262721449626985, + "grad_norm": 0.6236113905906677, + "learning_rate": 0.0001948682279891266, + "loss": 0.1728, + "step": 4378 + }, + { + "epoch": 0.10265065607107486, + "grad_norm": 1.1988211870193481, + "learning_rate": 0.00019486589832561503, + "loss": 0.2572, + "step": 4379 + }, + { + "epoch": 0.10267409764587986, + "grad_norm": 0.5242301225662231, + "learning_rate": 0.00019486356814735843, + "loss": 0.1809, + "step": 4380 + }, + { + "epoch": 0.10269753922068485, + "grad_norm": 0.6331515908241272, + "learning_rate": 0.0001948612374543694, + "loss": 0.234, + "step": 4381 + }, + { + "epoch": 0.10272098079548984, + "grad_norm": 0.4753817319869995, + "learning_rate": 0.00019485890624666068, + "loss": 0.1878, + "step": 4382 + }, + { + "epoch": 0.10274442237029484, + "grad_norm": 0.2881629765033722, + "learning_rate": 0.00019485657452424483, + "loss": 0.0905, + "step": 4383 + }, + { + "epoch": 0.10276786394509983, + "grad_norm": 0.09984491020441055, + "learning_rate": 0.00019485424228713456, + "loss": 0.0235, + "step": 4384 + }, + { + "epoch": 0.10279130551990483, + "grad_norm": 0.6585657000541687, + "learning_rate": 0.0001948519095353425, + "loss": 0.0788, + "step": 4385 + }, + { + "epoch": 0.10281474709470982, + "grad_norm": 0.8331627249717712, + "learning_rate": 0.0001948495762688813, + "loss": 0.297, + "step": 4386 + }, + { + "epoch": 0.10283818866951482, + "grad_norm": 0.39871370792388916, + "learning_rate": 0.00019484724248776365, + "loss": 0.0836, + "step": 4387 + }, + { + "epoch": 0.10286163024431981, + "grad_norm": 0.48736870288848877, + "learning_rate": 0.0001948449081920022, + "loss": 0.5086, + "step": 4388 + }, + { + "epoch": 0.1028850718191248, + "grad_norm": 0.40906551480293274, + "learning_rate": 0.00019484257338160958, + "loss": 0.062, + "step": 4389 + }, + { + "epoch": 0.1029085133939298, + "grad_norm": 0.6953468918800354, + "learning_rate": 0.00019484023805659852, + "loss": 0.1311, + "step": 4390 + }, + { + "epoch": 0.10293195496873479, + "grad_norm": 0.5054724812507629, + "learning_rate": 0.00019483790221698163, + "loss": 0.1916, + "step": 4391 + }, + { + "epoch": 0.1029553965435398, + "grad_norm": 0.29305240511894226, + "learning_rate": 0.00019483556586277165, + "loss": 0.3336, + "step": 4392 + }, + { + "epoch": 0.1029788381183448, + "grad_norm": 0.457590252161026, + "learning_rate": 0.0001948332289939812, + "loss": 0.1141, + "step": 4393 + }, + { + "epoch": 0.10300227969314979, + "grad_norm": 0.6602004766464233, + "learning_rate": 0.000194830891610623, + "loss": 0.1495, + "step": 4394 + }, + { + "epoch": 0.10302572126795478, + "grad_norm": 0.1336526721715927, + "learning_rate": 0.00019482855371270972, + "loss": 0.0474, + "step": 4395 + }, + { + "epoch": 0.10304916284275978, + "grad_norm": 0.5991193056106567, + "learning_rate": 0.00019482621530025402, + "loss": 0.0831, + "step": 4396 + }, + { + "epoch": 0.10307260441756477, + "grad_norm": 0.33957332372665405, + "learning_rate": 0.00019482387637326863, + "loss": 0.068, + "step": 4397 + }, + { + "epoch": 0.10309604599236977, + "grad_norm": 0.24385951459407806, + "learning_rate": 0.00019482153693176622, + "loss": 0.0276, + "step": 4398 + }, + { + "epoch": 0.10311948756717476, + "grad_norm": 0.6284312009811401, + "learning_rate": 0.00019481919697575945, + "loss": 0.14, + "step": 4399 + }, + { + "epoch": 0.10314292914197976, + "grad_norm": 0.29972323775291443, + "learning_rate": 0.00019481685650526108, + "loss": 0.087, + "step": 4400 + }, + { + "epoch": 0.10316637071678475, + "grad_norm": 0.7420600652694702, + "learning_rate": 0.0001948145155202838, + "loss": 0.1673, + "step": 4401 + }, + { + "epoch": 0.10318981229158974, + "grad_norm": 0.10444308072328568, + "learning_rate": 0.00019481217402084024, + "loss": 0.0303, + "step": 4402 + }, + { + "epoch": 0.10321325386639474, + "grad_norm": 0.3073175847530365, + "learning_rate": 0.00019480983200694318, + "loss": 0.0816, + "step": 4403 + }, + { + "epoch": 0.10323669544119975, + "grad_norm": 0.30560776591300964, + "learning_rate": 0.00019480748947860535, + "loss": 0.2707, + "step": 4404 + }, + { + "epoch": 0.10326013701600474, + "grad_norm": 0.3859415352344513, + "learning_rate": 0.00019480514643583935, + "loss": 0.1052, + "step": 4405 + }, + { + "epoch": 0.10328357859080974, + "grad_norm": 0.752676784992218, + "learning_rate": 0.000194802802878658, + "loss": 0.1701, + "step": 4406 + }, + { + "epoch": 0.10330702016561473, + "grad_norm": 0.9273086786270142, + "learning_rate": 0.00019480045880707395, + "loss": 0.2756, + "step": 4407 + }, + { + "epoch": 0.10333046174041972, + "grad_norm": 0.6383225917816162, + "learning_rate": 0.00019479811422109996, + "loss": 0.1502, + "step": 4408 + }, + { + "epoch": 0.10335390331522472, + "grad_norm": 0.5225962400436401, + "learning_rate": 0.00019479576912074875, + "loss": 0.0979, + "step": 4409 + }, + { + "epoch": 0.10337734489002971, + "grad_norm": 0.5405019521713257, + "learning_rate": 0.000194793423506033, + "loss": 0.7844, + "step": 4410 + }, + { + "epoch": 0.1034007864648347, + "grad_norm": 0.5342116355895996, + "learning_rate": 0.0001947910773769655, + "loss": 0.1582, + "step": 4411 + }, + { + "epoch": 0.1034242280396397, + "grad_norm": 0.3753044009208679, + "learning_rate": 0.00019478873073355891, + "loss": 0.1011, + "step": 4412 + }, + { + "epoch": 0.1034476696144447, + "grad_norm": 0.5145132541656494, + "learning_rate": 0.00019478638357582604, + "loss": 0.0873, + "step": 4413 + }, + { + "epoch": 0.10347111118924969, + "grad_norm": 0.22504164278507233, + "learning_rate": 0.00019478403590377958, + "loss": 0.066, + "step": 4414 + }, + { + "epoch": 0.10349455276405468, + "grad_norm": 0.38162946701049805, + "learning_rate": 0.00019478168771743228, + "loss": 0.1034, + "step": 4415 + }, + { + "epoch": 0.10351799433885968, + "grad_norm": 0.5910634994506836, + "learning_rate": 0.00019477933901679685, + "loss": 0.1599, + "step": 4416 + }, + { + "epoch": 0.10354143591366469, + "grad_norm": 0.5831109285354614, + "learning_rate": 0.0001947769898018861, + "loss": 0.1744, + "step": 4417 + }, + { + "epoch": 0.10356487748846968, + "grad_norm": 0.9172058701515198, + "learning_rate": 0.0001947746400727127, + "loss": 0.2054, + "step": 4418 + }, + { + "epoch": 0.10358831906327468, + "grad_norm": 0.5356972813606262, + "learning_rate": 0.00019477228982928946, + "loss": 0.1001, + "step": 4419 + }, + { + "epoch": 0.10361176063807967, + "grad_norm": 0.4262782037258148, + "learning_rate": 0.0001947699390716291, + "loss": 0.1066, + "step": 4420 + }, + { + "epoch": 0.10363520221288466, + "grad_norm": 0.5320981740951538, + "learning_rate": 0.0001947675877997444, + "loss": 0.0939, + "step": 4421 + }, + { + "epoch": 0.10365864378768966, + "grad_norm": 0.23414680361747742, + "learning_rate": 0.0001947652360136481, + "loss": 0.0461, + "step": 4422 + }, + { + "epoch": 0.10368208536249465, + "grad_norm": 0.6386762261390686, + "learning_rate": 0.00019476288371335297, + "loss": 0.2497, + "step": 4423 + }, + { + "epoch": 0.10370552693729965, + "grad_norm": 0.40079402923583984, + "learning_rate": 0.00019476053089887175, + "loss": 0.1237, + "step": 4424 + }, + { + "epoch": 0.10372896851210464, + "grad_norm": 0.6267813444137573, + "learning_rate": 0.00019475817757021725, + "loss": 0.1849, + "step": 4425 + }, + { + "epoch": 0.10375241008690964, + "grad_norm": 0.6387212872505188, + "learning_rate": 0.00019475582372740217, + "loss": 0.0865, + "step": 4426 + }, + { + "epoch": 0.10377585166171463, + "grad_norm": 0.5879784226417542, + "learning_rate": 0.00019475346937043936, + "loss": 0.1389, + "step": 4427 + }, + { + "epoch": 0.10379929323651962, + "grad_norm": 0.4101647734642029, + "learning_rate": 0.00019475111449934155, + "loss": 0.6577, + "step": 4428 + }, + { + "epoch": 0.10382273481132462, + "grad_norm": 0.11211680620908737, + "learning_rate": 0.00019474875911412156, + "loss": 0.0209, + "step": 4429 + }, + { + "epoch": 0.10384617638612963, + "grad_norm": 0.5842093229293823, + "learning_rate": 0.0001947464032147921, + "loss": 0.1957, + "step": 4430 + }, + { + "epoch": 0.10386961796093462, + "grad_norm": 0.529055655002594, + "learning_rate": 0.000194744046801366, + "loss": 0.6682, + "step": 4431 + }, + { + "epoch": 0.10389305953573962, + "grad_norm": 0.29677721858024597, + "learning_rate": 0.00019474168987385605, + "loss": 0.0897, + "step": 4432 + }, + { + "epoch": 0.10391650111054461, + "grad_norm": 0.3394126296043396, + "learning_rate": 0.00019473933243227505, + "loss": 0.0983, + "step": 4433 + }, + { + "epoch": 0.1039399426853496, + "grad_norm": 0.6878093481063843, + "learning_rate": 0.00019473697447663574, + "loss": 0.1803, + "step": 4434 + }, + { + "epoch": 0.1039633842601546, + "grad_norm": 0.4345320463180542, + "learning_rate": 0.00019473461600695094, + "loss": 0.0818, + "step": 4435 + }, + { + "epoch": 0.10398682583495959, + "grad_norm": 0.3889100253582001, + "learning_rate": 0.00019473225702323343, + "loss": 0.0882, + "step": 4436 + }, + { + "epoch": 0.10401026740976459, + "grad_norm": 0.5003756284713745, + "learning_rate": 0.00019472989752549605, + "loss": 0.6902, + "step": 4437 + }, + { + "epoch": 0.10403370898456958, + "grad_norm": 0.8347571492195129, + "learning_rate": 0.0001947275375137516, + "loss": 0.299, + "step": 4438 + }, + { + "epoch": 0.10405715055937458, + "grad_norm": 0.8949996829032898, + "learning_rate": 0.00019472517698801282, + "loss": 0.1515, + "step": 4439 + }, + { + "epoch": 0.10408059213417957, + "grad_norm": 0.7361933588981628, + "learning_rate": 0.0001947228159482926, + "loss": 0.2408, + "step": 4440 + }, + { + "epoch": 0.10410403370898456, + "grad_norm": 0.43797174096107483, + "learning_rate": 0.00019472045439460374, + "loss": 0.1122, + "step": 4441 + }, + { + "epoch": 0.10412747528378956, + "grad_norm": 0.26383811235427856, + "learning_rate": 0.00019471809232695901, + "loss": 0.0279, + "step": 4442 + }, + { + "epoch": 0.10415091685859457, + "grad_norm": 0.37469303607940674, + "learning_rate": 0.00019471572974537124, + "loss": 0.497, + "step": 4443 + }, + { + "epoch": 0.10417435843339956, + "grad_norm": 0.5365418195724487, + "learning_rate": 0.00019471336664985325, + "loss": 0.1808, + "step": 4444 + }, + { + "epoch": 0.10419780000820456, + "grad_norm": 0.2209341824054718, + "learning_rate": 0.00019471100304041788, + "loss": 0.0788, + "step": 4445 + }, + { + "epoch": 0.10422124158300955, + "grad_norm": 0.6699633002281189, + "learning_rate": 0.00019470863891707794, + "loss": 0.1038, + "step": 4446 + }, + { + "epoch": 0.10424468315781454, + "grad_norm": 0.8885058164596558, + "learning_rate": 0.00019470627427984626, + "loss": 0.2452, + "step": 4447 + }, + { + "epoch": 0.10426812473261954, + "grad_norm": 0.5218070149421692, + "learning_rate": 0.0001947039091287357, + "loss": 0.077, + "step": 4448 + }, + { + "epoch": 0.10429156630742453, + "grad_norm": 0.7096479535102844, + "learning_rate": 0.00019470154346375903, + "loss": 0.7928, + "step": 4449 + }, + { + "epoch": 0.10431500788222953, + "grad_norm": 0.47494685649871826, + "learning_rate": 0.00019469917728492914, + "loss": 0.1138, + "step": 4450 + }, + { + "epoch": 0.10433844945703452, + "grad_norm": 0.8690114617347717, + "learning_rate": 0.00019469681059225885, + "loss": 0.1748, + "step": 4451 + }, + { + "epoch": 0.10436189103183952, + "grad_norm": 0.6333012580871582, + "learning_rate": 0.00019469444338576103, + "loss": 0.9579, + "step": 4452 + }, + { + "epoch": 0.10438533260664451, + "grad_norm": 0.53826504945755, + "learning_rate": 0.00019469207566544846, + "loss": 0.1088, + "step": 4453 + }, + { + "epoch": 0.1044087741814495, + "grad_norm": 0.7201665639877319, + "learning_rate": 0.00019468970743133403, + "loss": 0.1241, + "step": 4454 + }, + { + "epoch": 0.1044322157562545, + "grad_norm": 0.7316722869873047, + "learning_rate": 0.0001946873386834306, + "loss": 0.0697, + "step": 4455 + }, + { + "epoch": 0.1044556573310595, + "grad_norm": 0.6396315097808838, + "learning_rate": 0.000194684969421751, + "loss": 0.1689, + "step": 4456 + }, + { + "epoch": 0.1044790989058645, + "grad_norm": 0.4256676137447357, + "learning_rate": 0.00019468259964630812, + "loss": 0.0805, + "step": 4457 + }, + { + "epoch": 0.1045025404806695, + "grad_norm": 0.27333948016166687, + "learning_rate": 0.00019468022935711473, + "loss": 0.0726, + "step": 4458 + }, + { + "epoch": 0.10452598205547449, + "grad_norm": 0.6021402478218079, + "learning_rate": 0.00019467785855418378, + "loss": 0.1367, + "step": 4459 + }, + { + "epoch": 0.10454942363027948, + "grad_norm": 0.8284604549407959, + "learning_rate": 0.00019467548723752813, + "loss": 0.1993, + "step": 4460 + }, + { + "epoch": 0.10457286520508448, + "grad_norm": 0.37320736050605774, + "learning_rate": 0.0001946731154071606, + "loss": 0.0568, + "step": 4461 + }, + { + "epoch": 0.10459630677988947, + "grad_norm": 0.43235349655151367, + "learning_rate": 0.0001946707430630941, + "loss": 0.1202, + "step": 4462 + }, + { + "epoch": 0.10461974835469447, + "grad_norm": 0.649739146232605, + "learning_rate": 0.00019466837020534148, + "loss": 0.1659, + "step": 4463 + }, + { + "epoch": 0.10464318992949946, + "grad_norm": 0.6052900552749634, + "learning_rate": 0.00019466599683391563, + "loss": 0.1282, + "step": 4464 + }, + { + "epoch": 0.10466663150430446, + "grad_norm": 0.33135220408439636, + "learning_rate": 0.0001946636229488294, + "loss": 0.0861, + "step": 4465 + }, + { + "epoch": 0.10469007307910945, + "grad_norm": 0.1994415521621704, + "learning_rate": 0.00019466124855009568, + "loss": 0.0405, + "step": 4466 + }, + { + "epoch": 0.10471351465391444, + "grad_norm": 0.6021029949188232, + "learning_rate": 0.0001946588736377274, + "loss": 0.1377, + "step": 4467 + }, + { + "epoch": 0.10473695622871945, + "grad_norm": 0.675441563129425, + "learning_rate": 0.0001946564982117374, + "loss": 0.1795, + "step": 4468 + }, + { + "epoch": 0.10476039780352445, + "grad_norm": 0.323388934135437, + "learning_rate": 0.00019465412227213856, + "loss": 0.0662, + "step": 4469 + }, + { + "epoch": 0.10478383937832944, + "grad_norm": 0.6214348077774048, + "learning_rate": 0.0001946517458189438, + "loss": 0.1732, + "step": 4470 + }, + { + "epoch": 0.10480728095313444, + "grad_norm": 0.6044795513153076, + "learning_rate": 0.00019464936885216604, + "loss": 0.1931, + "step": 4471 + }, + { + "epoch": 0.10483072252793943, + "grad_norm": 0.9175548553466797, + "learning_rate": 0.0001946469913718181, + "loss": 0.1588, + "step": 4472 + }, + { + "epoch": 0.10485416410274442, + "grad_norm": 0.3432771563529968, + "learning_rate": 0.00019464461337791294, + "loss": 0.0864, + "step": 4473 + }, + { + "epoch": 0.10487760567754942, + "grad_norm": 0.2597019076347351, + "learning_rate": 0.00019464223487046343, + "loss": 0.2673, + "step": 4474 + }, + { + "epoch": 0.10490104725235441, + "grad_norm": 0.5281209945678711, + "learning_rate": 0.00019463985584948252, + "loss": 0.0789, + "step": 4475 + }, + { + "epoch": 0.1049244888271594, + "grad_norm": 0.19843898713588715, + "learning_rate": 0.0001946374763149831, + "loss": 0.0457, + "step": 4476 + }, + { + "epoch": 0.1049479304019644, + "grad_norm": 0.6288020610809326, + "learning_rate": 0.00019463509626697803, + "loss": 0.1366, + "step": 4477 + }, + { + "epoch": 0.1049713719767694, + "grad_norm": 0.1993919461965561, + "learning_rate": 0.0001946327157054803, + "loss": 0.0578, + "step": 4478 + }, + { + "epoch": 0.10499481355157439, + "grad_norm": 0.6440058350563049, + "learning_rate": 0.00019463033463050278, + "loss": 0.1322, + "step": 4479 + }, + { + "epoch": 0.10501825512637938, + "grad_norm": 0.427574098110199, + "learning_rate": 0.00019462795304205842, + "loss": 0.1281, + "step": 4480 + }, + { + "epoch": 0.10504169670118439, + "grad_norm": 0.20794595777988434, + "learning_rate": 0.00019462557094016009, + "loss": 0.0841, + "step": 4481 + }, + { + "epoch": 0.10506513827598939, + "grad_norm": 0.5458052158355713, + "learning_rate": 0.00019462318832482077, + "loss": 0.1339, + "step": 4482 + }, + { + "epoch": 0.10508857985079438, + "grad_norm": 0.7350121736526489, + "learning_rate": 0.00019462080519605336, + "loss": 0.135, + "step": 4483 + }, + { + "epoch": 0.10511202142559938, + "grad_norm": 0.528266966342926, + "learning_rate": 0.00019461842155387083, + "loss": 0.1569, + "step": 4484 + }, + { + "epoch": 0.10513546300040437, + "grad_norm": 0.6032071709632874, + "learning_rate": 0.00019461603739828607, + "loss": 0.1594, + "step": 4485 + }, + { + "epoch": 0.10515890457520936, + "grad_norm": 0.5000777840614319, + "learning_rate": 0.000194613652729312, + "loss": 0.6891, + "step": 4486 + }, + { + "epoch": 0.10518234615001436, + "grad_norm": 0.5553989410400391, + "learning_rate": 0.0001946112675469616, + "loss": 0.1705, + "step": 4487 + }, + { + "epoch": 0.10520578772481935, + "grad_norm": 0.12631851434707642, + "learning_rate": 0.00019460888185124785, + "loss": 0.0248, + "step": 4488 + }, + { + "epoch": 0.10522922929962435, + "grad_norm": 0.34684452414512634, + "learning_rate": 0.00019460649564218363, + "loss": 0.1397, + "step": 4489 + }, + { + "epoch": 0.10525267087442934, + "grad_norm": 0.27589651942253113, + "learning_rate": 0.00019460410891978188, + "loss": 0.061, + "step": 4490 + }, + { + "epoch": 0.10527611244923434, + "grad_norm": 0.3998684883117676, + "learning_rate": 0.00019460172168405556, + "loss": 0.0901, + "step": 4491 + }, + { + "epoch": 0.10529955402403933, + "grad_norm": 0.5270940065383911, + "learning_rate": 0.0001945993339350177, + "loss": 0.0888, + "step": 4492 + }, + { + "epoch": 0.10532299559884432, + "grad_norm": 0.5771549940109253, + "learning_rate": 0.00019459694567268115, + "loss": 0.1835, + "step": 4493 + }, + { + "epoch": 0.10534643717364933, + "grad_norm": 0.5196640491485596, + "learning_rate": 0.0001945945568970589, + "loss": 0.1972, + "step": 4494 + }, + { + "epoch": 0.10536987874845433, + "grad_norm": 0.7039899826049805, + "learning_rate": 0.00019459216760816394, + "loss": 0.216, + "step": 4495 + }, + { + "epoch": 0.10539332032325932, + "grad_norm": 0.6300528049468994, + "learning_rate": 0.00019458977780600925, + "loss": 0.1195, + "step": 4496 + }, + { + "epoch": 0.10541676189806432, + "grad_norm": 0.7159882187843323, + "learning_rate": 0.00019458738749060771, + "loss": 0.1731, + "step": 4497 + }, + { + "epoch": 0.10544020347286931, + "grad_norm": 0.834570586681366, + "learning_rate": 0.00019458499666197239, + "loss": 0.1807, + "step": 4498 + }, + { + "epoch": 0.1054636450476743, + "grad_norm": 0.5526511073112488, + "learning_rate": 0.00019458260532011622, + "loss": 0.1518, + "step": 4499 + }, + { + "epoch": 0.1054870866224793, + "grad_norm": 0.32996201515197754, + "learning_rate": 0.00019458021346505216, + "loss": 0.078, + "step": 4500 + }, + { + "epoch": 0.10551052819728429, + "grad_norm": 0.45770594477653503, + "learning_rate": 0.00019457782109679318, + "loss": 0.1678, + "step": 4501 + }, + { + "epoch": 0.10553396977208929, + "grad_norm": 0.5497982501983643, + "learning_rate": 0.00019457542821535234, + "loss": 0.1432, + "step": 4502 + }, + { + "epoch": 0.10555741134689428, + "grad_norm": 0.3629191517829895, + "learning_rate": 0.00019457303482074253, + "loss": 0.08, + "step": 4503 + }, + { + "epoch": 0.10558085292169928, + "grad_norm": 1.0702539682388306, + "learning_rate": 0.00019457064091297675, + "loss": 0.1929, + "step": 4504 + }, + { + "epoch": 0.10560429449650427, + "grad_norm": 0.49134671688079834, + "learning_rate": 0.00019456824649206805, + "loss": 0.1966, + "step": 4505 + }, + { + "epoch": 0.10562773607130926, + "grad_norm": 0.5510765314102173, + "learning_rate": 0.00019456585155802936, + "loss": 0.1022, + "step": 4506 + }, + { + "epoch": 0.10565117764611427, + "grad_norm": 0.15609480440616608, + "learning_rate": 0.00019456345611087373, + "loss": 0.0758, + "step": 4507 + }, + { + "epoch": 0.10567461922091927, + "grad_norm": 0.5029918551445007, + "learning_rate": 0.0001945610601506141, + "loss": 0.1578, + "step": 4508 + }, + { + "epoch": 0.10569806079572426, + "grad_norm": 0.2839835584163666, + "learning_rate": 0.0001945586636772635, + "loss": 0.0483, + "step": 4509 + }, + { + "epoch": 0.10572150237052926, + "grad_norm": 0.24501806497573853, + "learning_rate": 0.00019455626669083497, + "loss": 0.0706, + "step": 4510 + }, + { + "epoch": 0.10574494394533425, + "grad_norm": 0.5809745192527771, + "learning_rate": 0.00019455386919134144, + "loss": 0.1245, + "step": 4511 + }, + { + "epoch": 0.10576838552013924, + "grad_norm": 0.7772834300994873, + "learning_rate": 0.000194551471178796, + "loss": 0.2069, + "step": 4512 + }, + { + "epoch": 0.10579182709494424, + "grad_norm": 0.9078623652458191, + "learning_rate": 0.00019454907265321157, + "loss": 0.1711, + "step": 4513 + }, + { + "epoch": 0.10581526866974923, + "grad_norm": 0.5024828910827637, + "learning_rate": 0.00019454667361460125, + "loss": 0.1481, + "step": 4514 + }, + { + "epoch": 0.10583871024455423, + "grad_norm": 0.17940682172775269, + "learning_rate": 0.00019454427406297796, + "loss": 0.06, + "step": 4515 + }, + { + "epoch": 0.10586215181935922, + "grad_norm": 0.5549818277359009, + "learning_rate": 0.00019454187399835484, + "loss": 0.2472, + "step": 4516 + }, + { + "epoch": 0.10588559339416422, + "grad_norm": 0.477268785238266, + "learning_rate": 0.00019453947342074483, + "loss": 0.4396, + "step": 4517 + }, + { + "epoch": 0.10590903496896921, + "grad_norm": 0.14992481470108032, + "learning_rate": 0.00019453707233016098, + "loss": 0.0426, + "step": 4518 + }, + { + "epoch": 0.1059324765437742, + "grad_norm": 0.22562260925769806, + "learning_rate": 0.00019453467072661634, + "loss": 0.0911, + "step": 4519 + }, + { + "epoch": 0.10595591811857921, + "grad_norm": 0.3308351933956146, + "learning_rate": 0.0001945322686101239, + "loss": 0.0921, + "step": 4520 + }, + { + "epoch": 0.1059793596933842, + "grad_norm": 0.5620218515396118, + "learning_rate": 0.00019452986598069672, + "loss": 0.1446, + "step": 4521 + }, + { + "epoch": 0.1060028012681892, + "grad_norm": 0.8887631297111511, + "learning_rate": 0.0001945274628383478, + "loss": 0.3797, + "step": 4522 + }, + { + "epoch": 0.1060262428429942, + "grad_norm": 0.8450334668159485, + "learning_rate": 0.00019452505918309023, + "loss": 0.1834, + "step": 4523 + }, + { + "epoch": 0.10604968441779919, + "grad_norm": 0.3120020925998688, + "learning_rate": 0.00019452265501493704, + "loss": 0.0882, + "step": 4524 + }, + { + "epoch": 0.10607312599260418, + "grad_norm": 0.786359429359436, + "learning_rate": 0.00019452025033390126, + "loss": 0.1401, + "step": 4525 + }, + { + "epoch": 0.10609656756740918, + "grad_norm": 0.2740801274776459, + "learning_rate": 0.00019451784513999594, + "loss": 0.075, + "step": 4526 + }, + { + "epoch": 0.10612000914221417, + "grad_norm": 0.6625960469245911, + "learning_rate": 0.00019451543943323416, + "loss": 0.1468, + "step": 4527 + }, + { + "epoch": 0.10614345071701917, + "grad_norm": 0.3065277934074402, + "learning_rate": 0.0001945130332136289, + "loss": 0.0731, + "step": 4528 + }, + { + "epoch": 0.10616689229182416, + "grad_norm": 0.5228333473205566, + "learning_rate": 0.0001945106264811933, + "loss": 0.1786, + "step": 4529 + }, + { + "epoch": 0.10619033386662916, + "grad_norm": 0.5770134329795837, + "learning_rate": 0.00019450821923594038, + "loss": 0.1091, + "step": 4530 + }, + { + "epoch": 0.10621377544143415, + "grad_norm": 0.7044620513916016, + "learning_rate": 0.0001945058114778832, + "loss": 0.1017, + "step": 4531 + }, + { + "epoch": 0.10623721701623916, + "grad_norm": 0.5255204439163208, + "learning_rate": 0.00019450340320703482, + "loss": 0.1483, + "step": 4532 + }, + { + "epoch": 0.10626065859104415, + "grad_norm": 0.6186519861221313, + "learning_rate": 0.00019450099442340832, + "loss": 0.1526, + "step": 4533 + }, + { + "epoch": 0.10628410016584915, + "grad_norm": 0.49215349555015564, + "learning_rate": 0.00019449858512701678, + "loss": 0.0793, + "step": 4534 + }, + { + "epoch": 0.10630754174065414, + "grad_norm": 0.6970916986465454, + "learning_rate": 0.00019449617531787326, + "loss": 0.1943, + "step": 4535 + }, + { + "epoch": 0.10633098331545913, + "grad_norm": 0.5398865342140198, + "learning_rate": 0.00019449376499599083, + "loss": 0.1334, + "step": 4536 + }, + { + "epoch": 0.10635442489026413, + "grad_norm": 0.7117979526519775, + "learning_rate": 0.00019449135416138254, + "loss": 0.2769, + "step": 4537 + }, + { + "epoch": 0.10637786646506912, + "grad_norm": 0.34178781509399414, + "learning_rate": 0.00019448894281406156, + "loss": 0.0936, + "step": 4538 + }, + { + "epoch": 0.10640130803987412, + "grad_norm": 0.3905903697013855, + "learning_rate": 0.00019448653095404091, + "loss": 0.0831, + "step": 4539 + }, + { + "epoch": 0.10642474961467911, + "grad_norm": 0.6091437339782715, + "learning_rate": 0.00019448411858133367, + "loss": 0.7595, + "step": 4540 + }, + { + "epoch": 0.1064481911894841, + "grad_norm": 0.5477787256240845, + "learning_rate": 0.00019448170569595294, + "loss": 0.1766, + "step": 4541 + }, + { + "epoch": 0.1064716327642891, + "grad_norm": 0.5716670751571655, + "learning_rate": 0.00019447929229791182, + "loss": 0.1643, + "step": 4542 + }, + { + "epoch": 0.1064950743390941, + "grad_norm": 0.5241855382919312, + "learning_rate": 0.0001944768783872234, + "loss": 0.138, + "step": 4543 + }, + { + "epoch": 0.10651851591389909, + "grad_norm": 0.3463185131549835, + "learning_rate": 0.0001944744639639008, + "loss": 0.0653, + "step": 4544 + }, + { + "epoch": 0.1065419574887041, + "grad_norm": 0.586405336856842, + "learning_rate": 0.0001944720490279571, + "loss": 0.1092, + "step": 4545 + }, + { + "epoch": 0.10656539906350909, + "grad_norm": 0.6436442732810974, + "learning_rate": 0.0001944696335794054, + "loss": 1.4897, + "step": 4546 + }, + { + "epoch": 0.10658884063831409, + "grad_norm": 0.4719747304916382, + "learning_rate": 0.00019446721761825884, + "loss": 0.1216, + "step": 4547 + }, + { + "epoch": 0.10661228221311908, + "grad_norm": 0.3040018379688263, + "learning_rate": 0.00019446480114453045, + "loss": 0.0771, + "step": 4548 + }, + { + "epoch": 0.10663572378792407, + "grad_norm": 0.8111423850059509, + "learning_rate": 0.00019446238415823342, + "loss": 0.1188, + "step": 4549 + }, + { + "epoch": 0.10665916536272907, + "grad_norm": 0.47749561071395874, + "learning_rate": 0.00019445996665938084, + "loss": 0.1306, + "step": 4550 + }, + { + "epoch": 0.10668260693753406, + "grad_norm": 0.6369925737380981, + "learning_rate": 0.00019445754864798583, + "loss": 0.2016, + "step": 4551 + }, + { + "epoch": 0.10670604851233906, + "grad_norm": 0.3041408061981201, + "learning_rate": 0.00019445513012406146, + "loss": 0.0608, + "step": 4552 + }, + { + "epoch": 0.10672949008714405, + "grad_norm": 0.2492433488368988, + "learning_rate": 0.00019445271108762096, + "loss": 0.0624, + "step": 4553 + }, + { + "epoch": 0.10675293166194905, + "grad_norm": 0.2310728132724762, + "learning_rate": 0.00019445029153867735, + "loss": 0.0323, + "step": 4554 + }, + { + "epoch": 0.10677637323675404, + "grad_norm": 0.44880950450897217, + "learning_rate": 0.00019444787147724382, + "loss": 0.1986, + "step": 4555 + }, + { + "epoch": 0.10679981481155904, + "grad_norm": 0.7667344212532043, + "learning_rate": 0.00019444545090333348, + "loss": 0.1338, + "step": 4556 + }, + { + "epoch": 0.10682325638636403, + "grad_norm": 0.5277459025382996, + "learning_rate": 0.00019444302981695948, + "loss": 0.0996, + "step": 4557 + }, + { + "epoch": 0.10684669796116904, + "grad_norm": 0.6199811100959778, + "learning_rate": 0.00019444060821813492, + "loss": 0.1713, + "step": 4558 + }, + { + "epoch": 0.10687013953597403, + "grad_norm": 0.5941827297210693, + "learning_rate": 0.00019443818610687298, + "loss": 0.134, + "step": 4559 + }, + { + "epoch": 0.10689358111077903, + "grad_norm": 0.9275622963905334, + "learning_rate": 0.0001944357634831868, + "loss": 0.1759, + "step": 4560 + }, + { + "epoch": 0.10691702268558402, + "grad_norm": 0.6137329936027527, + "learning_rate": 0.0001944333403470895, + "loss": 0.7928, + "step": 4561 + }, + { + "epoch": 0.10694046426038901, + "grad_norm": 0.7808348536491394, + "learning_rate": 0.00019443091669859422, + "loss": 0.2564, + "step": 4562 + }, + { + "epoch": 0.10696390583519401, + "grad_norm": 0.5775575637817383, + "learning_rate": 0.00019442849253771416, + "loss": 0.9049, + "step": 4563 + }, + { + "epoch": 0.106987347409999, + "grad_norm": 0.2251901477575302, + "learning_rate": 0.00019442606786446243, + "loss": 0.0994, + "step": 4564 + }, + { + "epoch": 0.107010788984804, + "grad_norm": 0.2684239149093628, + "learning_rate": 0.0001944236426788522, + "loss": 0.0596, + "step": 4565 + }, + { + "epoch": 0.10703423055960899, + "grad_norm": 0.34205782413482666, + "learning_rate": 0.00019442121698089664, + "loss": 0.0651, + "step": 4566 + }, + { + "epoch": 0.10705767213441399, + "grad_norm": 0.2318384051322937, + "learning_rate": 0.0001944187907706089, + "loss": 0.0959, + "step": 4567 + }, + { + "epoch": 0.10708111370921898, + "grad_norm": 0.3231886029243469, + "learning_rate": 0.00019441636404800213, + "loss": 0.1036, + "step": 4568 + }, + { + "epoch": 0.10710455528402398, + "grad_norm": 0.3125443160533905, + "learning_rate": 0.00019441393681308953, + "loss": 0.0701, + "step": 4569 + }, + { + "epoch": 0.10712799685882897, + "grad_norm": 0.2481469362974167, + "learning_rate": 0.00019441150906588425, + "loss": 0.0703, + "step": 4570 + }, + { + "epoch": 0.10715143843363398, + "grad_norm": 0.5705224871635437, + "learning_rate": 0.00019440908080639946, + "loss": 0.1566, + "step": 4571 + }, + { + "epoch": 0.10717488000843897, + "grad_norm": 0.40887343883514404, + "learning_rate": 0.00019440665203464836, + "loss": 0.087, + "step": 4572 + }, + { + "epoch": 0.10719832158324397, + "grad_norm": 0.5495659708976746, + "learning_rate": 0.0001944042227506441, + "loss": 0.1758, + "step": 4573 + }, + { + "epoch": 0.10722176315804896, + "grad_norm": 0.6098688244819641, + "learning_rate": 0.00019440179295439987, + "loss": 0.2319, + "step": 4574 + }, + { + "epoch": 0.10724520473285395, + "grad_norm": 0.6293767690658569, + "learning_rate": 0.00019439936264592885, + "loss": 0.1483, + "step": 4575 + }, + { + "epoch": 0.10726864630765895, + "grad_norm": 0.5350756049156189, + "learning_rate": 0.00019439693182524424, + "loss": 0.1108, + "step": 4576 + }, + { + "epoch": 0.10729208788246394, + "grad_norm": 0.6182426810264587, + "learning_rate": 0.00019439450049235925, + "loss": 0.1571, + "step": 4577 + }, + { + "epoch": 0.10731552945726894, + "grad_norm": 0.24024836719036102, + "learning_rate": 0.000194392068647287, + "loss": 0.0631, + "step": 4578 + }, + { + "epoch": 0.10733897103207393, + "grad_norm": 0.15898622572422028, + "learning_rate": 0.00019438963629004076, + "loss": 0.0604, + "step": 4579 + }, + { + "epoch": 0.10736241260687893, + "grad_norm": 0.14765658974647522, + "learning_rate": 0.0001943872034206337, + "loss": 0.0406, + "step": 4580 + }, + { + "epoch": 0.10738585418168392, + "grad_norm": 0.30945175886154175, + "learning_rate": 0.00019438477003907902, + "loss": 0.1285, + "step": 4581 + }, + { + "epoch": 0.10740929575648892, + "grad_norm": 0.6456810832023621, + "learning_rate": 0.00019438233614538995, + "loss": 0.2006, + "step": 4582 + }, + { + "epoch": 0.10743273733129392, + "grad_norm": 0.5686415433883667, + "learning_rate": 0.00019437990173957963, + "loss": 0.1774, + "step": 4583 + }, + { + "epoch": 0.10745617890609892, + "grad_norm": 0.5156419277191162, + "learning_rate": 0.00019437746682166132, + "loss": 0.1986, + "step": 4584 + }, + { + "epoch": 0.10747962048090391, + "grad_norm": 0.6564487218856812, + "learning_rate": 0.0001943750313916482, + "loss": 0.1624, + "step": 4585 + }, + { + "epoch": 0.1075030620557089, + "grad_norm": 0.5231313109397888, + "learning_rate": 0.00019437259544955352, + "loss": 0.141, + "step": 4586 + }, + { + "epoch": 0.1075265036305139, + "grad_norm": 0.4491168260574341, + "learning_rate": 0.00019437015899539053, + "loss": 0.116, + "step": 4587 + }, + { + "epoch": 0.1075499452053189, + "grad_norm": 0.15056872367858887, + "learning_rate": 0.00019436772202917235, + "loss": 0.0261, + "step": 4588 + }, + { + "epoch": 0.10757338678012389, + "grad_norm": 0.4082236886024475, + "learning_rate": 0.00019436528455091227, + "loss": 0.0884, + "step": 4589 + }, + { + "epoch": 0.10759682835492888, + "grad_norm": 0.4688475728034973, + "learning_rate": 0.0001943628465606235, + "loss": 0.1192, + "step": 4590 + }, + { + "epoch": 0.10762026992973388, + "grad_norm": 0.9499017596244812, + "learning_rate": 0.0001943604080583193, + "loss": 0.1151, + "step": 4591 + }, + { + "epoch": 0.10764371150453887, + "grad_norm": 0.4986504316329956, + "learning_rate": 0.00019435796904401285, + "loss": 0.1271, + "step": 4592 + }, + { + "epoch": 0.10766715307934387, + "grad_norm": 0.30064335465431213, + "learning_rate": 0.00019435552951771738, + "loss": 0.0683, + "step": 4593 + }, + { + "epoch": 0.10769059465414886, + "grad_norm": 0.4002862274646759, + "learning_rate": 0.00019435308947944618, + "loss": 0.0918, + "step": 4594 + }, + { + "epoch": 0.10771403622895385, + "grad_norm": 0.9569686651229858, + "learning_rate": 0.00019435064892921246, + "loss": 0.186, + "step": 4595 + }, + { + "epoch": 0.10773747780375886, + "grad_norm": 0.5675362944602966, + "learning_rate": 0.00019434820786702947, + "loss": 0.7034, + "step": 4596 + }, + { + "epoch": 0.10776091937856386, + "grad_norm": 0.2922826111316681, + "learning_rate": 0.00019434576629291046, + "loss": 0.0747, + "step": 4597 + }, + { + "epoch": 0.10778436095336885, + "grad_norm": 0.7296756505966187, + "learning_rate": 0.00019434332420686866, + "loss": 0.1145, + "step": 4598 + }, + { + "epoch": 0.10780780252817385, + "grad_norm": 0.9182863235473633, + "learning_rate": 0.0001943408816089173, + "loss": 0.2196, + "step": 4599 + }, + { + "epoch": 0.10783124410297884, + "grad_norm": 0.4998657703399658, + "learning_rate": 0.00019433843849906972, + "loss": 0.0993, + "step": 4600 + }, + { + "epoch": 0.10785468567778383, + "grad_norm": 0.42778292298316956, + "learning_rate": 0.00019433599487733908, + "loss": 0.0991, + "step": 4601 + }, + { + "epoch": 0.10787812725258883, + "grad_norm": 0.7434832453727722, + "learning_rate": 0.0001943335507437387, + "loss": 0.2017, + "step": 4602 + }, + { + "epoch": 0.10790156882739382, + "grad_norm": 0.7358222007751465, + "learning_rate": 0.00019433110609828183, + "loss": 0.1577, + "step": 4603 + }, + { + "epoch": 0.10792501040219882, + "grad_norm": 0.3370353579521179, + "learning_rate": 0.0001943286609409817, + "loss": 0.0917, + "step": 4604 + }, + { + "epoch": 0.10794845197700381, + "grad_norm": 0.2974768579006195, + "learning_rate": 0.0001943262152718516, + "loss": 0.086, + "step": 4605 + }, + { + "epoch": 0.1079718935518088, + "grad_norm": 0.7427977919578552, + "learning_rate": 0.00019432376909090483, + "loss": 0.1416, + "step": 4606 + }, + { + "epoch": 0.1079953351266138, + "grad_norm": 0.5344607830047607, + "learning_rate": 0.00019432132239815463, + "loss": 0.1024, + "step": 4607 + }, + { + "epoch": 0.1080187767014188, + "grad_norm": 0.09144431352615356, + "learning_rate": 0.00019431887519361427, + "loss": 0.0318, + "step": 4608 + }, + { + "epoch": 0.1080422182762238, + "grad_norm": 0.6567863821983337, + "learning_rate": 0.00019431642747729706, + "loss": 0.7885, + "step": 4609 + }, + { + "epoch": 0.1080656598510288, + "grad_norm": 0.8911720514297485, + "learning_rate": 0.00019431397924921624, + "loss": 0.1837, + "step": 4610 + }, + { + "epoch": 0.10808910142583379, + "grad_norm": 0.9323638677597046, + "learning_rate": 0.00019431153050938514, + "loss": 0.2129, + "step": 4611 + }, + { + "epoch": 0.10811254300063879, + "grad_norm": 0.3354238271713257, + "learning_rate": 0.000194309081257817, + "loss": 0.0554, + "step": 4612 + }, + { + "epoch": 0.10813598457544378, + "grad_norm": 0.7846522927284241, + "learning_rate": 0.00019430663149452512, + "loss": 0.1558, + "step": 4613 + }, + { + "epoch": 0.10815942615024877, + "grad_norm": 0.09157256036996841, + "learning_rate": 0.00019430418121952286, + "loss": 0.0172, + "step": 4614 + }, + { + "epoch": 0.10818286772505377, + "grad_norm": 0.7933543920516968, + "learning_rate": 0.00019430173043282343, + "loss": 0.0969, + "step": 4615 + }, + { + "epoch": 0.10820630929985876, + "grad_norm": 0.39844536781311035, + "learning_rate": 0.00019429927913444014, + "loss": 0.0854, + "step": 4616 + }, + { + "epoch": 0.10822975087466376, + "grad_norm": 0.4138292074203491, + "learning_rate": 0.00019429682732438635, + "loss": 0.0962, + "step": 4617 + }, + { + "epoch": 0.10825319244946875, + "grad_norm": 0.3709239065647125, + "learning_rate": 0.00019429437500267528, + "loss": 0.0538, + "step": 4618 + }, + { + "epoch": 0.10827663402427375, + "grad_norm": 1.0042896270751953, + "learning_rate": 0.00019429192216932032, + "loss": 0.1644, + "step": 4619 + }, + { + "epoch": 0.10830007559907874, + "grad_norm": 0.6383767127990723, + "learning_rate": 0.0001942894688243347, + "loss": 0.2007, + "step": 4620 + }, + { + "epoch": 0.10832351717388373, + "grad_norm": 0.3496202528476715, + "learning_rate": 0.00019428701496773183, + "loss": 0.0798, + "step": 4621 + }, + { + "epoch": 0.10834695874868874, + "grad_norm": 0.6372846364974976, + "learning_rate": 0.00019428456059952492, + "loss": 0.1915, + "step": 4622 + }, + { + "epoch": 0.10837040032349374, + "grad_norm": 0.8693512082099915, + "learning_rate": 0.00019428210571972735, + "loss": 0.1732, + "step": 4623 + }, + { + "epoch": 0.10839384189829873, + "grad_norm": 0.6547343134880066, + "learning_rate": 0.0001942796503283524, + "loss": 0.1316, + "step": 4624 + }, + { + "epoch": 0.10841728347310373, + "grad_norm": 0.43095386028289795, + "learning_rate": 0.00019427719442541343, + "loss": 0.0699, + "step": 4625 + }, + { + "epoch": 0.10844072504790872, + "grad_norm": 0.5163125991821289, + "learning_rate": 0.00019427473801092378, + "loss": 0.8224, + "step": 4626 + }, + { + "epoch": 0.10846416662271371, + "grad_norm": 0.3870822489261627, + "learning_rate": 0.0001942722810848967, + "loss": 0.1307, + "step": 4627 + }, + { + "epoch": 0.10848760819751871, + "grad_norm": 0.6923630237579346, + "learning_rate": 0.00019426982364734562, + "loss": 0.3678, + "step": 4628 + }, + { + "epoch": 0.1085110497723237, + "grad_norm": 0.6145868897438049, + "learning_rate": 0.00019426736569828378, + "loss": 0.1941, + "step": 4629 + }, + { + "epoch": 0.1085344913471287, + "grad_norm": 0.5604655146598816, + "learning_rate": 0.0001942649072377246, + "loss": 0.1287, + "step": 4630 + }, + { + "epoch": 0.10855793292193369, + "grad_norm": 0.2950195372104645, + "learning_rate": 0.00019426244826568136, + "loss": 0.4089, + "step": 4631 + }, + { + "epoch": 0.10858137449673869, + "grad_norm": 0.7252217531204224, + "learning_rate": 0.00019425998878216746, + "loss": 0.2205, + "step": 4632 + }, + { + "epoch": 0.10860481607154368, + "grad_norm": 0.33460623025894165, + "learning_rate": 0.00019425752878719617, + "loss": 0.0482, + "step": 4633 + }, + { + "epoch": 0.10862825764634867, + "grad_norm": 0.43981122970581055, + "learning_rate": 0.00019425506828078088, + "loss": 0.0546, + "step": 4634 + }, + { + "epoch": 0.10865169922115368, + "grad_norm": 0.8347533345222473, + "learning_rate": 0.00019425260726293496, + "loss": 0.5202, + "step": 4635 + }, + { + "epoch": 0.10867514079595868, + "grad_norm": 0.1949017345905304, + "learning_rate": 0.00019425014573367174, + "loss": 0.0487, + "step": 4636 + }, + { + "epoch": 0.10869858237076367, + "grad_norm": 0.8947609663009644, + "learning_rate": 0.00019424768369300457, + "loss": 0.1502, + "step": 4637 + }, + { + "epoch": 0.10872202394556867, + "grad_norm": 0.654971182346344, + "learning_rate": 0.00019424522114094684, + "loss": 0.1299, + "step": 4638 + }, + { + "epoch": 0.10874546552037366, + "grad_norm": 1.1451936960220337, + "learning_rate": 0.00019424275807751186, + "loss": 0.2549, + "step": 4639 + }, + { + "epoch": 0.10876890709517865, + "grad_norm": 0.5334569215774536, + "learning_rate": 0.00019424029450271303, + "loss": 0.098, + "step": 4640 + }, + { + "epoch": 0.10879234866998365, + "grad_norm": 0.5437631011009216, + "learning_rate": 0.00019423783041656374, + "loss": 0.1548, + "step": 4641 + }, + { + "epoch": 0.10881579024478864, + "grad_norm": 0.5405294895172119, + "learning_rate": 0.00019423536581907727, + "loss": 0.1365, + "step": 4642 + }, + { + "epoch": 0.10883923181959364, + "grad_norm": 0.5737701654434204, + "learning_rate": 0.00019423290071026713, + "loss": 0.2147, + "step": 4643 + }, + { + "epoch": 0.10886267339439863, + "grad_norm": 0.19694378972053528, + "learning_rate": 0.00019423043509014657, + "loss": 0.0664, + "step": 4644 + }, + { + "epoch": 0.10888611496920363, + "grad_norm": 0.07348666340112686, + "learning_rate": 0.00019422796895872906, + "loss": 0.0243, + "step": 4645 + }, + { + "epoch": 0.10890955654400862, + "grad_norm": 0.6677390933036804, + "learning_rate": 0.00019422550231602793, + "loss": 0.6488, + "step": 4646 + }, + { + "epoch": 0.10893299811881363, + "grad_norm": 0.6713370680809021, + "learning_rate": 0.00019422303516205655, + "loss": 0.528, + "step": 4647 + }, + { + "epoch": 0.10895643969361862, + "grad_norm": 0.6807405948638916, + "learning_rate": 0.00019422056749682836, + "loss": 0.2254, + "step": 4648 + }, + { + "epoch": 0.10897988126842362, + "grad_norm": 0.4083050489425659, + "learning_rate": 0.0001942180993203567, + "loss": 0.1042, + "step": 4649 + }, + { + "epoch": 0.10900332284322861, + "grad_norm": 0.7198930978775024, + "learning_rate": 0.000194215630632655, + "loss": 0.1521, + "step": 4650 + }, + { + "epoch": 0.1090267644180336, + "grad_norm": 0.5548511147499084, + "learning_rate": 0.00019421316143373662, + "loss": 0.1145, + "step": 4651 + }, + { + "epoch": 0.1090502059928386, + "grad_norm": 0.14284707605838776, + "learning_rate": 0.000194210691723615, + "loss": 0.0277, + "step": 4652 + }, + { + "epoch": 0.1090736475676436, + "grad_norm": 0.23005254566669464, + "learning_rate": 0.0001942082215023035, + "loss": 0.0357, + "step": 4653 + }, + { + "epoch": 0.10909708914244859, + "grad_norm": 0.791355550289154, + "learning_rate": 0.00019420575076981554, + "loss": 0.3336, + "step": 4654 + }, + { + "epoch": 0.10912053071725358, + "grad_norm": 0.44865524768829346, + "learning_rate": 0.00019420327952616453, + "loss": 0.178, + "step": 4655 + }, + { + "epoch": 0.10914397229205858, + "grad_norm": 0.34719324111938477, + "learning_rate": 0.00019420080777136388, + "loss": 0.0721, + "step": 4656 + }, + { + "epoch": 0.10916741386686357, + "grad_norm": 0.21229083836078644, + "learning_rate": 0.00019419833550542702, + "loss": 0.0922, + "step": 4657 + }, + { + "epoch": 0.10919085544166857, + "grad_norm": 0.7637500166893005, + "learning_rate": 0.00019419586272836732, + "loss": 0.1709, + "step": 4658 + }, + { + "epoch": 0.10921429701647356, + "grad_norm": 0.36122840642929077, + "learning_rate": 0.00019419338944019822, + "loss": 0.1455, + "step": 4659 + }, + { + "epoch": 0.10923773859127857, + "grad_norm": 0.5629582405090332, + "learning_rate": 0.00019419091564093315, + "loss": 0.5734, + "step": 4660 + }, + { + "epoch": 0.10926118016608356, + "grad_norm": 0.15658783912658691, + "learning_rate": 0.00019418844133058553, + "loss": 0.0511, + "step": 4661 + }, + { + "epoch": 0.10928462174088856, + "grad_norm": 0.46233412623405457, + "learning_rate": 0.00019418596650916877, + "loss": 0.1196, + "step": 4662 + }, + { + "epoch": 0.10930806331569355, + "grad_norm": 0.5333887934684753, + "learning_rate": 0.00019418349117669627, + "loss": 0.1376, + "step": 4663 + }, + { + "epoch": 0.10933150489049855, + "grad_norm": 0.5088186264038086, + "learning_rate": 0.00019418101533318156, + "loss": 0.1969, + "step": 4664 + }, + { + "epoch": 0.10935494646530354, + "grad_norm": 0.41141194105148315, + "learning_rate": 0.00019417853897863796, + "loss": 0.1033, + "step": 4665 + }, + { + "epoch": 0.10937838804010853, + "grad_norm": 0.5392824411392212, + "learning_rate": 0.00019417606211307898, + "loss": 0.7574, + "step": 4666 + }, + { + "epoch": 0.10940182961491353, + "grad_norm": 0.32145190238952637, + "learning_rate": 0.00019417358473651806, + "loss": 0.0801, + "step": 4667 + }, + { + "epoch": 0.10942527118971852, + "grad_norm": 0.17991741001605988, + "learning_rate": 0.0001941711068489686, + "loss": 0.0485, + "step": 4668 + }, + { + "epoch": 0.10944871276452352, + "grad_norm": 0.8593155741691589, + "learning_rate": 0.00019416862845044407, + "loss": 0.1618, + "step": 4669 + }, + { + "epoch": 0.10947215433932851, + "grad_norm": 0.6982229351997375, + "learning_rate": 0.00019416614954095792, + "loss": 0.704, + "step": 4670 + }, + { + "epoch": 0.1094955959141335, + "grad_norm": 0.549403190612793, + "learning_rate": 0.0001941636701205236, + "loss": 0.7693, + "step": 4671 + }, + { + "epoch": 0.1095190374889385, + "grad_norm": 0.15055514872074127, + "learning_rate": 0.00019416119018915453, + "loss": 0.0381, + "step": 4672 + }, + { + "epoch": 0.10954247906374351, + "grad_norm": 0.401947021484375, + "learning_rate": 0.00019415870974686422, + "loss": 0.0985, + "step": 4673 + }, + { + "epoch": 0.1095659206385485, + "grad_norm": 0.2638985216617584, + "learning_rate": 0.0001941562287936661, + "loss": 0.0671, + "step": 4674 + }, + { + "epoch": 0.1095893622133535, + "grad_norm": 0.6272874474525452, + "learning_rate": 0.00019415374732957363, + "loss": 0.1164, + "step": 4675 + }, + { + "epoch": 0.10961280378815849, + "grad_norm": 0.546650230884552, + "learning_rate": 0.00019415126535460026, + "loss": 0.6069, + "step": 4676 + }, + { + "epoch": 0.10963624536296349, + "grad_norm": 0.5671120285987854, + "learning_rate": 0.0001941487828687595, + "loss": 0.8389, + "step": 4677 + }, + { + "epoch": 0.10965968693776848, + "grad_norm": 0.32769477367401123, + "learning_rate": 0.0001941462998720648, + "loss": 0.0431, + "step": 4678 + }, + { + "epoch": 0.10968312851257347, + "grad_norm": 0.17094704508781433, + "learning_rate": 0.00019414381636452963, + "loss": 0.0604, + "step": 4679 + }, + { + "epoch": 0.10970657008737847, + "grad_norm": 0.20909839868545532, + "learning_rate": 0.00019414133234616748, + "loss": 0.0528, + "step": 4680 + }, + { + "epoch": 0.10973001166218346, + "grad_norm": 0.566421627998352, + "learning_rate": 0.0001941388478169918, + "loss": 0.1326, + "step": 4681 + }, + { + "epoch": 0.10975345323698846, + "grad_norm": 2.3216640949249268, + "learning_rate": 0.00019413636277701605, + "loss": 0.1555, + "step": 4682 + }, + { + "epoch": 0.10977689481179345, + "grad_norm": 0.19969962537288666, + "learning_rate": 0.00019413387722625377, + "loss": 0.0359, + "step": 4683 + }, + { + "epoch": 0.10980033638659845, + "grad_norm": 0.8113041520118713, + "learning_rate": 0.00019413139116471846, + "loss": 0.1918, + "step": 4684 + }, + { + "epoch": 0.10982377796140344, + "grad_norm": 0.8122677206993103, + "learning_rate": 0.00019412890459242355, + "loss": 0.1383, + "step": 4685 + }, + { + "epoch": 0.10984721953620845, + "grad_norm": 0.8692430853843689, + "learning_rate": 0.0001941264175093825, + "loss": 0.2995, + "step": 4686 + }, + { + "epoch": 0.10987066111101344, + "grad_norm": 0.637410044670105, + "learning_rate": 0.00019412392991560894, + "loss": 0.1032, + "step": 4687 + }, + { + "epoch": 0.10989410268581844, + "grad_norm": 0.29447826743125916, + "learning_rate": 0.00019412144181111626, + "loss": 0.0642, + "step": 4688 + }, + { + "epoch": 0.10991754426062343, + "grad_norm": 0.30423790216445923, + "learning_rate": 0.000194118953195918, + "loss": 0.071, + "step": 4689 + }, + { + "epoch": 0.10994098583542843, + "grad_norm": 0.5363950729370117, + "learning_rate": 0.00019411646407002767, + "loss": 0.1937, + "step": 4690 + }, + { + "epoch": 0.10996442741023342, + "grad_norm": 0.6468029618263245, + "learning_rate": 0.00019411397443345874, + "loss": 0.1402, + "step": 4691 + }, + { + "epoch": 0.10998786898503841, + "grad_norm": 0.1754930466413498, + "learning_rate": 0.00019411148428622474, + "loss": 0.0363, + "step": 4692 + }, + { + "epoch": 0.11001131055984341, + "grad_norm": 0.6681795716285706, + "learning_rate": 0.00019410899362833916, + "loss": 0.1605, + "step": 4693 + }, + { + "epoch": 0.1100347521346484, + "grad_norm": 0.7264685034751892, + "learning_rate": 0.00019410650245981558, + "loss": 0.1758, + "step": 4694 + }, + { + "epoch": 0.1100581937094534, + "grad_norm": 0.33742639422416687, + "learning_rate": 0.0001941040107806674, + "loss": 0.0714, + "step": 4695 + }, + { + "epoch": 0.11008163528425839, + "grad_norm": 0.4525356590747833, + "learning_rate": 0.00019410151859090827, + "loss": 0.0846, + "step": 4696 + }, + { + "epoch": 0.11010507685906339, + "grad_norm": 0.32647672295570374, + "learning_rate": 0.00019409902589055165, + "loss": 0.0457, + "step": 4697 + }, + { + "epoch": 0.11012851843386838, + "grad_norm": 0.8061145544052124, + "learning_rate": 0.00019409653267961105, + "loss": 0.2093, + "step": 4698 + }, + { + "epoch": 0.11015196000867339, + "grad_norm": 0.4769352972507477, + "learning_rate": 0.00019409403895810003, + "loss": 0.1019, + "step": 4699 + }, + { + "epoch": 0.11017540158347838, + "grad_norm": 0.6531544327735901, + "learning_rate": 0.0001940915447260321, + "loss": 0.1578, + "step": 4700 + }, + { + "epoch": 0.11019884315828338, + "grad_norm": 0.24453257024288177, + "learning_rate": 0.00019408904998342082, + "loss": 0.0456, + "step": 4701 + }, + { + "epoch": 0.11022228473308837, + "grad_norm": 0.3953584134578705, + "learning_rate": 0.0001940865547302797, + "loss": 0.0948, + "step": 4702 + }, + { + "epoch": 0.11024572630789337, + "grad_norm": 0.46093031764030457, + "learning_rate": 0.00019408405896662227, + "loss": 0.1239, + "step": 4703 + }, + { + "epoch": 0.11026916788269836, + "grad_norm": 0.7137247920036316, + "learning_rate": 0.00019408156269246212, + "loss": 0.1258, + "step": 4704 + }, + { + "epoch": 0.11029260945750335, + "grad_norm": 0.5459460020065308, + "learning_rate": 0.00019407906590781276, + "loss": 0.233, + "step": 4705 + }, + { + "epoch": 0.11031605103230835, + "grad_norm": 0.2030208259820938, + "learning_rate": 0.00019407656861268773, + "loss": 0.0559, + "step": 4706 + }, + { + "epoch": 0.11033949260711334, + "grad_norm": 0.6562537550926208, + "learning_rate": 0.00019407407080710058, + "loss": 0.1555, + "step": 4707 + }, + { + "epoch": 0.11036293418191834, + "grad_norm": 0.6018023490905762, + "learning_rate": 0.00019407157249106492, + "loss": 0.1596, + "step": 4708 + }, + { + "epoch": 0.11038637575672333, + "grad_norm": 0.49074500799179077, + "learning_rate": 0.00019406907366459423, + "loss": 0.1737, + "step": 4709 + }, + { + "epoch": 0.11040981733152833, + "grad_norm": 0.9116832613945007, + "learning_rate": 0.0001940665743277021, + "loss": 0.3587, + "step": 4710 + }, + { + "epoch": 0.11043325890633333, + "grad_norm": 0.5757728815078735, + "learning_rate": 0.0001940640744804021, + "loss": 0.0915, + "step": 4711 + }, + { + "epoch": 0.11045670048113833, + "grad_norm": 0.20538941025733948, + "learning_rate": 0.0001940615741227078, + "loss": 0.0666, + "step": 4712 + }, + { + "epoch": 0.11048014205594332, + "grad_norm": 0.42000752687454224, + "learning_rate": 0.00019405907325463275, + "loss": 0.0772, + "step": 4713 + }, + { + "epoch": 0.11050358363074832, + "grad_norm": 0.5926405191421509, + "learning_rate": 0.00019405657187619051, + "loss": 0.8107, + "step": 4714 + }, + { + "epoch": 0.11052702520555331, + "grad_norm": 0.6580383777618408, + "learning_rate": 0.0001940540699873947, + "loss": 0.2068, + "step": 4715 + }, + { + "epoch": 0.1105504667803583, + "grad_norm": 0.4426911473274231, + "learning_rate": 0.00019405156758825883, + "loss": 0.638, + "step": 4716 + }, + { + "epoch": 0.1105739083551633, + "grad_norm": 0.3081970810890198, + "learning_rate": 0.0001940490646787965, + "loss": 0.067, + "step": 4717 + }, + { + "epoch": 0.1105973499299683, + "grad_norm": 0.4948129653930664, + "learning_rate": 0.00019404656125902132, + "loss": 0.1215, + "step": 4718 + }, + { + "epoch": 0.11062079150477329, + "grad_norm": 0.33756884932518005, + "learning_rate": 0.00019404405732894686, + "loss": 0.1363, + "step": 4719 + }, + { + "epoch": 0.11064423307957828, + "grad_norm": 0.6432211399078369, + "learning_rate": 0.0001940415528885867, + "loss": 0.1291, + "step": 4720 + }, + { + "epoch": 0.11066767465438328, + "grad_norm": 0.20093952119350433, + "learning_rate": 0.0001940390479379544, + "loss": 0.072, + "step": 4721 + }, + { + "epoch": 0.11069111622918827, + "grad_norm": 0.639449417591095, + "learning_rate": 0.0001940365424770636, + "loss": 0.1372, + "step": 4722 + }, + { + "epoch": 0.11071455780399327, + "grad_norm": 0.6587502956390381, + "learning_rate": 0.00019403403650592788, + "loss": 0.1812, + "step": 4723 + }, + { + "epoch": 0.11073799937879827, + "grad_norm": 0.3623262345790863, + "learning_rate": 0.00019403153002456086, + "loss": 0.4362, + "step": 4724 + }, + { + "epoch": 0.11076144095360327, + "grad_norm": 0.6353287100791931, + "learning_rate": 0.00019402902303297605, + "loss": 0.1294, + "step": 4725 + }, + { + "epoch": 0.11078488252840826, + "grad_norm": 0.5291690230369568, + "learning_rate": 0.00019402651553118716, + "loss": 0.1546, + "step": 4726 + }, + { + "epoch": 0.11080832410321326, + "grad_norm": 0.5201575756072998, + "learning_rate": 0.00019402400751920775, + "loss": 0.1209, + "step": 4727 + }, + { + "epoch": 0.11083176567801825, + "grad_norm": 0.8764455914497375, + "learning_rate": 0.00019402149899705142, + "loss": 0.2102, + "step": 4728 + }, + { + "epoch": 0.11085520725282325, + "grad_norm": 0.8726580739021301, + "learning_rate": 0.0001940189899647318, + "loss": 0.209, + "step": 4729 + }, + { + "epoch": 0.11087864882762824, + "grad_norm": 0.40440455079078674, + "learning_rate": 0.00019401648042226253, + "loss": 0.1431, + "step": 4730 + }, + { + "epoch": 0.11090209040243323, + "grad_norm": 0.667154848575592, + "learning_rate": 0.00019401397036965715, + "loss": 0.2389, + "step": 4731 + }, + { + "epoch": 0.11092553197723823, + "grad_norm": 0.6777355670928955, + "learning_rate": 0.00019401145980692933, + "loss": 0.1095, + "step": 4732 + }, + { + "epoch": 0.11094897355204322, + "grad_norm": 0.6524519920349121, + "learning_rate": 0.0001940089487340927, + "loss": 0.2283, + "step": 4733 + }, + { + "epoch": 0.11097241512684822, + "grad_norm": 0.21996867656707764, + "learning_rate": 0.00019400643715116085, + "loss": 0.0559, + "step": 4734 + }, + { + "epoch": 0.11099585670165321, + "grad_norm": 0.5691244006156921, + "learning_rate": 0.00019400392505814745, + "loss": 0.7003, + "step": 4735 + }, + { + "epoch": 0.1110192982764582, + "grad_norm": 1.0165151357650757, + "learning_rate": 0.0001940014124550661, + "loss": 0.164, + "step": 4736 + }, + { + "epoch": 0.11104273985126321, + "grad_norm": 0.4440787136554718, + "learning_rate": 0.00019399889934193043, + "loss": 0.0582, + "step": 4737 + }, + { + "epoch": 0.11106618142606821, + "grad_norm": 0.21886686980724335, + "learning_rate": 0.00019399638571875413, + "loss": 0.0358, + "step": 4738 + }, + { + "epoch": 0.1110896230008732, + "grad_norm": 0.7244622707366943, + "learning_rate": 0.00019399387158555075, + "loss": 0.2196, + "step": 4739 + }, + { + "epoch": 0.1111130645756782, + "grad_norm": 0.6104933023452759, + "learning_rate": 0.000193991356942334, + "loss": 0.1268, + "step": 4740 + }, + { + "epoch": 0.11113650615048319, + "grad_norm": 0.5873619318008423, + "learning_rate": 0.0001939888417891175, + "loss": 0.1729, + "step": 4741 + }, + { + "epoch": 0.11115994772528819, + "grad_norm": 0.34572505950927734, + "learning_rate": 0.0001939863261259149, + "loss": 0.1211, + "step": 4742 + }, + { + "epoch": 0.11118338930009318, + "grad_norm": 0.27112069725990295, + "learning_rate": 0.0001939838099527399, + "loss": 0.089, + "step": 4743 + }, + { + "epoch": 0.11120683087489817, + "grad_norm": 0.5778934955596924, + "learning_rate": 0.00019398129326960603, + "loss": 0.1605, + "step": 4744 + }, + { + "epoch": 0.11123027244970317, + "grad_norm": 0.6858824491500854, + "learning_rate": 0.00019397877607652702, + "loss": 0.1898, + "step": 4745 + }, + { + "epoch": 0.11125371402450816, + "grad_norm": 0.5959123373031616, + "learning_rate": 0.0001939762583735166, + "loss": 0.1675, + "step": 4746 + }, + { + "epoch": 0.11127715559931316, + "grad_norm": 0.7692940831184387, + "learning_rate": 0.00019397374016058832, + "loss": 0.7239, + "step": 4747 + }, + { + "epoch": 0.11130059717411815, + "grad_norm": 0.6997921466827393, + "learning_rate": 0.00019397122143775588, + "loss": 0.1579, + "step": 4748 + }, + { + "epoch": 0.11132403874892315, + "grad_norm": 1.0977891683578491, + "learning_rate": 0.00019396870220503296, + "loss": 0.2095, + "step": 4749 + }, + { + "epoch": 0.11134748032372815, + "grad_norm": 0.1426767110824585, + "learning_rate": 0.00019396618246243322, + "loss": 0.0369, + "step": 4750 + }, + { + "epoch": 0.11137092189853315, + "grad_norm": 0.4336775839328766, + "learning_rate": 0.0001939636622099703, + "loss": 0.0991, + "step": 4751 + }, + { + "epoch": 0.11139436347333814, + "grad_norm": 0.20896939933300018, + "learning_rate": 0.00019396114144765796, + "loss": 0.0501, + "step": 4752 + }, + { + "epoch": 0.11141780504814314, + "grad_norm": 0.5244956612586975, + "learning_rate": 0.00019395862017550979, + "loss": 0.1005, + "step": 4753 + }, + { + "epoch": 0.11144124662294813, + "grad_norm": 0.654085099697113, + "learning_rate": 0.00019395609839353953, + "loss": 0.1415, + "step": 4754 + }, + { + "epoch": 0.11146468819775313, + "grad_norm": 0.23592881858348846, + "learning_rate": 0.0001939535761017608, + "loss": 0.0357, + "step": 4755 + }, + { + "epoch": 0.11148812977255812, + "grad_norm": 0.5087711811065674, + "learning_rate": 0.00019395105330018736, + "loss": 0.1807, + "step": 4756 + }, + { + "epoch": 0.11151157134736311, + "grad_norm": 0.7386165261268616, + "learning_rate": 0.00019394852998883284, + "loss": 0.7891, + "step": 4757 + }, + { + "epoch": 0.11153501292216811, + "grad_norm": 0.5553576350212097, + "learning_rate": 0.00019394600616771097, + "loss": 0.893, + "step": 4758 + }, + { + "epoch": 0.1115584544969731, + "grad_norm": 0.589718759059906, + "learning_rate": 0.00019394348183683543, + "loss": 0.1022, + "step": 4759 + }, + { + "epoch": 0.1115818960717781, + "grad_norm": 0.6170942783355713, + "learning_rate": 0.00019394095699621988, + "loss": 0.1411, + "step": 4760 + }, + { + "epoch": 0.11160533764658309, + "grad_norm": 0.7042184472084045, + "learning_rate": 0.0001939384316458781, + "loss": 0.221, + "step": 4761 + }, + { + "epoch": 0.1116287792213881, + "grad_norm": 0.7331459522247314, + "learning_rate": 0.0001939359057858237, + "loss": 0.6447, + "step": 4762 + }, + { + "epoch": 0.1116522207961931, + "grad_norm": 0.5616356730461121, + "learning_rate": 0.00019393337941607048, + "loss": 0.2012, + "step": 4763 + }, + { + "epoch": 0.11167566237099809, + "grad_norm": 0.148293137550354, + "learning_rate": 0.0001939308525366321, + "loss": 0.0326, + "step": 4764 + }, + { + "epoch": 0.11169910394580308, + "grad_norm": 0.3896613121032715, + "learning_rate": 0.00019392832514752227, + "loss": 0.1211, + "step": 4765 + }, + { + "epoch": 0.11172254552060808, + "grad_norm": 0.8193808197975159, + "learning_rate": 0.0001939257972487547, + "loss": 0.6965, + "step": 4766 + }, + { + "epoch": 0.11174598709541307, + "grad_norm": 0.13901491463184357, + "learning_rate": 0.00019392326884034311, + "loss": 0.025, + "step": 4767 + }, + { + "epoch": 0.11176942867021807, + "grad_norm": 0.19233089685440063, + "learning_rate": 0.00019392073992230123, + "loss": 0.0327, + "step": 4768 + }, + { + "epoch": 0.11179287024502306, + "grad_norm": 0.30877014994621277, + "learning_rate": 0.00019391821049464275, + "loss": 0.0428, + "step": 4769 + }, + { + "epoch": 0.11181631181982805, + "grad_norm": 0.37298646569252014, + "learning_rate": 0.00019391568055738146, + "loss": 0.1235, + "step": 4770 + }, + { + "epoch": 0.11183975339463305, + "grad_norm": 0.14729242026805878, + "learning_rate": 0.000193913150110531, + "loss": 0.0363, + "step": 4771 + }, + { + "epoch": 0.11186319496943804, + "grad_norm": 0.8936415314674377, + "learning_rate": 0.00019391061915410517, + "loss": 0.2834, + "step": 4772 + }, + { + "epoch": 0.11188663654424304, + "grad_norm": 0.6752088069915771, + "learning_rate": 0.0001939080876881177, + "loss": 0.1428, + "step": 4773 + }, + { + "epoch": 0.11191007811904803, + "grad_norm": 0.7584056854248047, + "learning_rate": 0.00019390555571258228, + "loss": 0.168, + "step": 4774 + }, + { + "epoch": 0.11193351969385304, + "grad_norm": 0.7336487770080566, + "learning_rate": 0.00019390302322751268, + "loss": 0.2254, + "step": 4775 + }, + { + "epoch": 0.11195696126865803, + "grad_norm": 0.7267651557922363, + "learning_rate": 0.00019390049023292264, + "loss": 0.3063, + "step": 4776 + }, + { + "epoch": 0.11198040284346303, + "grad_norm": 0.30565834045410156, + "learning_rate": 0.0001938979567288259, + "loss": 0.0676, + "step": 4777 + }, + { + "epoch": 0.11200384441826802, + "grad_norm": 0.8558627963066101, + "learning_rate": 0.00019389542271523621, + "loss": 0.3103, + "step": 4778 + }, + { + "epoch": 0.11202728599307302, + "grad_norm": 0.5311993956565857, + "learning_rate": 0.00019389288819216728, + "loss": 0.0974, + "step": 4779 + }, + { + "epoch": 0.11205072756787801, + "grad_norm": 0.4213391840457916, + "learning_rate": 0.00019389035315963294, + "loss": 0.088, + "step": 4780 + }, + { + "epoch": 0.112074169142683, + "grad_norm": 0.4271632134914398, + "learning_rate": 0.00019388781761764688, + "loss": 0.1006, + "step": 4781 + }, + { + "epoch": 0.112097610717488, + "grad_norm": 0.441895067691803, + "learning_rate": 0.00019388528156622292, + "loss": 0.1066, + "step": 4782 + }, + { + "epoch": 0.112121052292293, + "grad_norm": 0.5191879868507385, + "learning_rate": 0.00019388274500537476, + "loss": 0.1473, + "step": 4783 + }, + { + "epoch": 0.11214449386709799, + "grad_norm": 0.5025242567062378, + "learning_rate": 0.00019388020793511615, + "loss": 0.1291, + "step": 4784 + }, + { + "epoch": 0.11216793544190298, + "grad_norm": 0.18334415555000305, + "learning_rate": 0.00019387767035546096, + "loss": 0.0442, + "step": 4785 + }, + { + "epoch": 0.11219137701670798, + "grad_norm": 0.33490386605262756, + "learning_rate": 0.00019387513226642284, + "loss": 0.0672, + "step": 4786 + }, + { + "epoch": 0.11221481859151297, + "grad_norm": 0.7627081274986267, + "learning_rate": 0.00019387259366801564, + "loss": 0.2727, + "step": 4787 + }, + { + "epoch": 0.11223826016631798, + "grad_norm": 0.18186601996421814, + "learning_rate": 0.00019387005456025308, + "loss": 0.0459, + "step": 4788 + }, + { + "epoch": 0.11226170174112297, + "grad_norm": 0.3935736119747162, + "learning_rate": 0.00019386751494314899, + "loss": 0.0972, + "step": 4789 + }, + { + "epoch": 0.11228514331592797, + "grad_norm": 1.0655049085617065, + "learning_rate": 0.00019386497481671712, + "loss": 0.1307, + "step": 4790 + }, + { + "epoch": 0.11230858489073296, + "grad_norm": 0.49550381302833557, + "learning_rate": 0.00019386243418097125, + "loss": 0.1478, + "step": 4791 + }, + { + "epoch": 0.11233202646553796, + "grad_norm": 0.6349312663078308, + "learning_rate": 0.00019385989303592519, + "loss": 0.1121, + "step": 4792 + }, + { + "epoch": 0.11235546804034295, + "grad_norm": 0.8649066686630249, + "learning_rate": 0.00019385735138159266, + "loss": 0.2547, + "step": 4793 + }, + { + "epoch": 0.11237890961514795, + "grad_norm": 0.7121130228042603, + "learning_rate": 0.00019385480921798754, + "loss": 0.2066, + "step": 4794 + }, + { + "epoch": 0.11240235118995294, + "grad_norm": 0.16299161314964294, + "learning_rate": 0.0001938522665451236, + "loss": 0.0436, + "step": 4795 + }, + { + "epoch": 0.11242579276475793, + "grad_norm": 0.12612242996692657, + "learning_rate": 0.0001938497233630146, + "loss": 0.0222, + "step": 4796 + }, + { + "epoch": 0.11244923433956293, + "grad_norm": 0.4300075173377991, + "learning_rate": 0.00019384717967167435, + "loss": 0.1717, + "step": 4797 + }, + { + "epoch": 0.11247267591436792, + "grad_norm": 0.6844779849052429, + "learning_rate": 0.0001938446354711167, + "loss": 0.1654, + "step": 4798 + }, + { + "epoch": 0.11249611748917292, + "grad_norm": 0.3821423649787903, + "learning_rate": 0.00019384209076135539, + "loss": 0.0788, + "step": 4799 + }, + { + "epoch": 0.11251955906397791, + "grad_norm": 0.6219905018806458, + "learning_rate": 0.00019383954554240429, + "loss": 0.1785, + "step": 4800 + }, + { + "epoch": 0.11254300063878292, + "grad_norm": 0.5654222965240479, + "learning_rate": 0.00019383699981427715, + "loss": 0.1173, + "step": 4801 + }, + { + "epoch": 0.11256644221358791, + "grad_norm": 0.4501243233680725, + "learning_rate": 0.0001938344535769878, + "loss": 0.0561, + "step": 4802 + }, + { + "epoch": 0.11258988378839291, + "grad_norm": 0.49910539388656616, + "learning_rate": 0.00019383190683055013, + "loss": 0.1552, + "step": 4803 + }, + { + "epoch": 0.1126133253631979, + "grad_norm": 0.2969502806663513, + "learning_rate": 0.00019382935957497783, + "loss": 0.0894, + "step": 4804 + }, + { + "epoch": 0.1126367669380029, + "grad_norm": 0.20761460065841675, + "learning_rate": 0.00019382681181028482, + "loss": 0.0576, + "step": 4805 + }, + { + "epoch": 0.11266020851280789, + "grad_norm": 0.7970400452613831, + "learning_rate": 0.00019382426353648486, + "loss": 0.1519, + "step": 4806 + }, + { + "epoch": 0.11268365008761289, + "grad_norm": 0.4313945174217224, + "learning_rate": 0.00019382171475359184, + "loss": 0.1461, + "step": 4807 + }, + { + "epoch": 0.11270709166241788, + "grad_norm": 0.4478367567062378, + "learning_rate": 0.00019381916546161953, + "loss": 0.1391, + "step": 4808 + }, + { + "epoch": 0.11273053323722287, + "grad_norm": 0.5727171897888184, + "learning_rate": 0.0001938166156605818, + "loss": 0.0825, + "step": 4809 + }, + { + "epoch": 0.11275397481202787, + "grad_norm": 0.2175525724887848, + "learning_rate": 0.00019381406535049249, + "loss": 0.0484, + "step": 4810 + }, + { + "epoch": 0.11277741638683286, + "grad_norm": 0.6161749958992004, + "learning_rate": 0.0001938115145313654, + "loss": 0.1809, + "step": 4811 + }, + { + "epoch": 0.11280085796163786, + "grad_norm": 0.4681802988052368, + "learning_rate": 0.0001938089632032144, + "loss": 0.125, + "step": 4812 + }, + { + "epoch": 0.11282429953644285, + "grad_norm": 0.46600258350372314, + "learning_rate": 0.00019380641136605334, + "loss": 0.1488, + "step": 4813 + }, + { + "epoch": 0.11284774111124786, + "grad_norm": 0.8698940277099609, + "learning_rate": 0.00019380385901989606, + "loss": 0.2683, + "step": 4814 + }, + { + "epoch": 0.11287118268605285, + "grad_norm": 0.19901271164417267, + "learning_rate": 0.00019380130616475637, + "loss": 0.0392, + "step": 4815 + }, + { + "epoch": 0.11289462426085785, + "grad_norm": 0.29635393619537354, + "learning_rate": 0.0001937987528006482, + "loss": 0.0877, + "step": 4816 + }, + { + "epoch": 0.11291806583566284, + "grad_norm": 0.6490626931190491, + "learning_rate": 0.00019379619892758533, + "loss": 0.2309, + "step": 4817 + }, + { + "epoch": 0.11294150741046784, + "grad_norm": 0.6927595138549805, + "learning_rate": 0.00019379364454558167, + "loss": 0.1408, + "step": 4818 + }, + { + "epoch": 0.11296494898527283, + "grad_norm": 0.23962336778640747, + "learning_rate": 0.00019379108965465104, + "loss": 0.0831, + "step": 4819 + }, + { + "epoch": 0.11298839056007783, + "grad_norm": 0.45379820466041565, + "learning_rate": 0.00019378853425480733, + "loss": 0.1095, + "step": 4820 + }, + { + "epoch": 0.11301183213488282, + "grad_norm": 0.48084110021591187, + "learning_rate": 0.00019378597834606437, + "loss": 0.1266, + "step": 4821 + }, + { + "epoch": 0.11303527370968781, + "grad_norm": 1.2029368877410889, + "learning_rate": 0.0001937834219284361, + "loss": 0.28, + "step": 4822 + }, + { + "epoch": 0.11305871528449281, + "grad_norm": 0.5713449716567993, + "learning_rate": 0.0001937808650019363, + "loss": 0.1367, + "step": 4823 + }, + { + "epoch": 0.1130821568592978, + "grad_norm": 0.5169363021850586, + "learning_rate": 0.00019377830756657892, + "loss": 0.8326, + "step": 4824 + }, + { + "epoch": 0.1131055984341028, + "grad_norm": 0.5641379356384277, + "learning_rate": 0.0001937757496223778, + "loss": 0.1462, + "step": 4825 + }, + { + "epoch": 0.1131290400089078, + "grad_norm": 0.495940625667572, + "learning_rate": 0.0001937731911693468, + "loss": 0.177, + "step": 4826 + }, + { + "epoch": 0.1131524815837128, + "grad_norm": 0.4647684395313263, + "learning_rate": 0.00019377063220749988, + "loss": 0.1435, + "step": 4827 + }, + { + "epoch": 0.1131759231585178, + "grad_norm": 0.19104160368442535, + "learning_rate": 0.0001937680727368508, + "loss": 0.0544, + "step": 4828 + }, + { + "epoch": 0.11319936473332279, + "grad_norm": 0.6198276877403259, + "learning_rate": 0.00019376551275741358, + "loss": 0.1535, + "step": 4829 + }, + { + "epoch": 0.11322280630812778, + "grad_norm": 0.6660774350166321, + "learning_rate": 0.00019376295226920202, + "loss": 0.2065, + "step": 4830 + }, + { + "epoch": 0.11324624788293278, + "grad_norm": 0.25293922424316406, + "learning_rate": 0.00019376039127223007, + "loss": 0.0549, + "step": 4831 + }, + { + "epoch": 0.11326968945773777, + "grad_norm": 0.29865890741348267, + "learning_rate": 0.0001937578297665116, + "loss": 0.0636, + "step": 4832 + }, + { + "epoch": 0.11329313103254277, + "grad_norm": 0.6183738708496094, + "learning_rate": 0.00019375526775206048, + "loss": 0.246, + "step": 4833 + }, + { + "epoch": 0.11331657260734776, + "grad_norm": 0.6681657433509827, + "learning_rate": 0.0001937527052288907, + "loss": 0.2154, + "step": 4834 + }, + { + "epoch": 0.11334001418215275, + "grad_norm": 0.7029377222061157, + "learning_rate": 0.00019375014219701604, + "loss": 0.1163, + "step": 4835 + }, + { + "epoch": 0.11336345575695775, + "grad_norm": 0.2777594029903412, + "learning_rate": 0.00019374757865645047, + "loss": 0.096, + "step": 4836 + }, + { + "epoch": 0.11338689733176274, + "grad_norm": 0.411202609539032, + "learning_rate": 0.00019374501460720796, + "loss": 0.1282, + "step": 4837 + }, + { + "epoch": 0.11341033890656774, + "grad_norm": 0.4490339756011963, + "learning_rate": 0.00019374245004930233, + "loss": 0.5485, + "step": 4838 + }, + { + "epoch": 0.11343378048137275, + "grad_norm": 0.19188278913497925, + "learning_rate": 0.00019373988498274752, + "loss": 0.0492, + "step": 4839 + }, + { + "epoch": 0.11345722205617774, + "grad_norm": 0.7967156171798706, + "learning_rate": 0.00019373731940755748, + "loss": 0.1348, + "step": 4840 + }, + { + "epoch": 0.11348066363098273, + "grad_norm": 0.5181742906570435, + "learning_rate": 0.00019373475332374611, + "loss": 0.1212, + "step": 4841 + }, + { + "epoch": 0.11350410520578773, + "grad_norm": 0.4783012270927429, + "learning_rate": 0.00019373218673132732, + "loss": 0.0811, + "step": 4842 + }, + { + "epoch": 0.11352754678059272, + "grad_norm": 0.17915478348731995, + "learning_rate": 0.00019372961963031506, + "loss": 0.042, + "step": 4843 + }, + { + "epoch": 0.11355098835539772, + "grad_norm": 0.167355477809906, + "learning_rate": 0.00019372705202072324, + "loss": 0.045, + "step": 4844 + }, + { + "epoch": 0.11357442993020271, + "grad_norm": 0.2854059338569641, + "learning_rate": 0.00019372448390256582, + "loss": 0.094, + "step": 4845 + }, + { + "epoch": 0.1135978715050077, + "grad_norm": 0.47441980242729187, + "learning_rate": 0.00019372191527585668, + "loss": 0.1363, + "step": 4846 + }, + { + "epoch": 0.1136213130798127, + "grad_norm": 0.3488914966583252, + "learning_rate": 0.0001937193461406098, + "loss": 0.1257, + "step": 4847 + }, + { + "epoch": 0.1136447546546177, + "grad_norm": 0.7332471609115601, + "learning_rate": 0.00019371677649683916, + "loss": 0.1629, + "step": 4848 + }, + { + "epoch": 0.11366819622942269, + "grad_norm": 0.13585996627807617, + "learning_rate": 0.0001937142063445586, + "loss": 0.0405, + "step": 4849 + }, + { + "epoch": 0.11369163780422768, + "grad_norm": 0.7176643013954163, + "learning_rate": 0.00019371163568378212, + "loss": 0.1645, + "step": 4850 + }, + { + "epoch": 0.11371507937903268, + "grad_norm": 0.6131482124328613, + "learning_rate": 0.0001937090645145237, + "loss": 0.1417, + "step": 4851 + }, + { + "epoch": 0.11373852095383769, + "grad_norm": 0.8310525417327881, + "learning_rate": 0.00019370649283679728, + "loss": 0.1672, + "step": 4852 + }, + { + "epoch": 0.11376196252864268, + "grad_norm": 0.6360403895378113, + "learning_rate": 0.00019370392065061674, + "loss": 0.1258, + "step": 4853 + }, + { + "epoch": 0.11378540410344767, + "grad_norm": 0.1757609099149704, + "learning_rate": 0.00019370134795599614, + "loss": 0.0291, + "step": 4854 + }, + { + "epoch": 0.11380884567825267, + "grad_norm": 0.722216010093689, + "learning_rate": 0.00019369877475294936, + "loss": 0.17, + "step": 4855 + }, + { + "epoch": 0.11383228725305766, + "grad_norm": 0.43891966342926025, + "learning_rate": 0.0001936962010414904, + "loss": 0.1055, + "step": 4856 + }, + { + "epoch": 0.11385572882786266, + "grad_norm": 0.5937979221343994, + "learning_rate": 0.00019369362682163322, + "loss": 0.127, + "step": 4857 + }, + { + "epoch": 0.11387917040266765, + "grad_norm": 0.7957517504692078, + "learning_rate": 0.0001936910520933918, + "loss": 0.1942, + "step": 4858 + }, + { + "epoch": 0.11390261197747265, + "grad_norm": 0.6267497539520264, + "learning_rate": 0.00019368847685678008, + "loss": 0.1925, + "step": 4859 + }, + { + "epoch": 0.11392605355227764, + "grad_norm": 0.47181904315948486, + "learning_rate": 0.00019368590111181204, + "loss": 0.1207, + "step": 4860 + }, + { + "epoch": 0.11394949512708263, + "grad_norm": 0.4997454583644867, + "learning_rate": 0.0001936833248585017, + "loss": 0.2133, + "step": 4861 + }, + { + "epoch": 0.11397293670188763, + "grad_norm": 0.21135662496089935, + "learning_rate": 0.000193680748096863, + "loss": 0.0359, + "step": 4862 + }, + { + "epoch": 0.11399637827669262, + "grad_norm": 0.688575029373169, + "learning_rate": 0.00019367817082690988, + "loss": 0.7229, + "step": 4863 + }, + { + "epoch": 0.11401981985149762, + "grad_norm": 0.624396562576294, + "learning_rate": 0.0001936755930486564, + "loss": 0.1806, + "step": 4864 + }, + { + "epoch": 0.11404326142630263, + "grad_norm": 0.6723016500473022, + "learning_rate": 0.00019367301476211655, + "loss": 0.1818, + "step": 4865 + }, + { + "epoch": 0.11406670300110762, + "grad_norm": 0.6741052865982056, + "learning_rate": 0.00019367043596730426, + "loss": 0.1371, + "step": 4866 + }, + { + "epoch": 0.11409014457591261, + "grad_norm": 0.20694711804389954, + "learning_rate": 0.00019366785666423354, + "loss": 0.0485, + "step": 4867 + }, + { + "epoch": 0.11411358615071761, + "grad_norm": 0.5967862010002136, + "learning_rate": 0.0001936652768529184, + "loss": 0.6573, + "step": 4868 + }, + { + "epoch": 0.1141370277255226, + "grad_norm": 0.3679240643978119, + "learning_rate": 0.00019366269653337285, + "loss": 0.0487, + "step": 4869 + }, + { + "epoch": 0.1141604693003276, + "grad_norm": 0.6098951697349548, + "learning_rate": 0.00019366011570561088, + "loss": 0.1698, + "step": 4870 + }, + { + "epoch": 0.11418391087513259, + "grad_norm": 0.39497411251068115, + "learning_rate": 0.00019365753436964645, + "loss": 0.1058, + "step": 4871 + }, + { + "epoch": 0.11420735244993759, + "grad_norm": 0.6003655195236206, + "learning_rate": 0.00019365495252549365, + "loss": 0.1706, + "step": 4872 + }, + { + "epoch": 0.11423079402474258, + "grad_norm": 0.7420499324798584, + "learning_rate": 0.00019365237017316642, + "loss": 0.2272, + "step": 4873 + }, + { + "epoch": 0.11425423559954757, + "grad_norm": 0.6902353763580322, + "learning_rate": 0.00019364978731267879, + "loss": 0.1728, + "step": 4874 + }, + { + "epoch": 0.11427767717435257, + "grad_norm": 0.5637635588645935, + "learning_rate": 0.0001936472039440448, + "loss": 0.1232, + "step": 4875 + }, + { + "epoch": 0.11430111874915756, + "grad_norm": 0.4874906539916992, + "learning_rate": 0.00019364462006727842, + "loss": 0.0944, + "step": 4876 + }, + { + "epoch": 0.11432456032396257, + "grad_norm": 0.455695778131485, + "learning_rate": 0.00019364203568239375, + "loss": 0.0849, + "step": 4877 + }, + { + "epoch": 0.11434800189876757, + "grad_norm": 0.17657089233398438, + "learning_rate": 0.00019363945078940473, + "loss": 0.0435, + "step": 4878 + }, + { + "epoch": 0.11437144347357256, + "grad_norm": 0.1418565809726715, + "learning_rate": 0.0001936368653883254, + "loss": 0.0416, + "step": 4879 + }, + { + "epoch": 0.11439488504837755, + "grad_norm": 0.2236277014017105, + "learning_rate": 0.00019363427947916985, + "loss": 0.0458, + "step": 4880 + }, + { + "epoch": 0.11441832662318255, + "grad_norm": 0.2598627805709839, + "learning_rate": 0.00019363169306195203, + "loss": 0.0842, + "step": 4881 + }, + { + "epoch": 0.11444176819798754, + "grad_norm": 0.8887888193130493, + "learning_rate": 0.00019362910613668604, + "loss": 0.736, + "step": 4882 + }, + { + "epoch": 0.11446520977279254, + "grad_norm": 0.3296543061733246, + "learning_rate": 0.00019362651870338585, + "loss": 0.0616, + "step": 4883 + }, + { + "epoch": 0.11448865134759753, + "grad_norm": 0.8858222365379333, + "learning_rate": 0.00019362393076206556, + "loss": 0.1643, + "step": 4884 + }, + { + "epoch": 0.11451209292240253, + "grad_norm": 0.6996061205863953, + "learning_rate": 0.00019362134231273918, + "loss": 0.2003, + "step": 4885 + }, + { + "epoch": 0.11453553449720752, + "grad_norm": 0.26478442549705505, + "learning_rate": 0.00019361875335542077, + "loss": 0.1213, + "step": 4886 + }, + { + "epoch": 0.11455897607201251, + "grad_norm": 0.3479476571083069, + "learning_rate": 0.00019361616389012437, + "loss": 0.0979, + "step": 4887 + }, + { + "epoch": 0.11458241764681751, + "grad_norm": 0.6687260866165161, + "learning_rate": 0.00019361357391686404, + "loss": 0.1246, + "step": 4888 + }, + { + "epoch": 0.1146058592216225, + "grad_norm": 0.7853033542633057, + "learning_rate": 0.00019361098343565383, + "loss": 0.253, + "step": 4889 + }, + { + "epoch": 0.11462930079642751, + "grad_norm": 0.39463627338409424, + "learning_rate": 0.00019360839244650775, + "loss": 0.1035, + "step": 4890 + }, + { + "epoch": 0.1146527423712325, + "grad_norm": 0.17322373390197754, + "learning_rate": 0.00019360580094943995, + "loss": 0.0686, + "step": 4891 + }, + { + "epoch": 0.1146761839460375, + "grad_norm": 0.6279179453849792, + "learning_rate": 0.00019360320894446441, + "loss": 0.5486, + "step": 4892 + }, + { + "epoch": 0.1146996255208425, + "grad_norm": 0.4453895092010498, + "learning_rate": 0.00019360061643159524, + "loss": 0.1217, + "step": 4893 + }, + { + "epoch": 0.11472306709564749, + "grad_norm": 0.5819786190986633, + "learning_rate": 0.0001935980234108465, + "loss": 0.1934, + "step": 4894 + }, + { + "epoch": 0.11474650867045248, + "grad_norm": 0.42946022748947144, + "learning_rate": 0.00019359542988223226, + "loss": 0.0923, + "step": 4895 + }, + { + "epoch": 0.11476995024525748, + "grad_norm": 0.20567940175533295, + "learning_rate": 0.00019359283584576654, + "loss": 0.074, + "step": 4896 + }, + { + "epoch": 0.11479339182006247, + "grad_norm": 0.17433428764343262, + "learning_rate": 0.0001935902413014635, + "loss": 0.0507, + "step": 4897 + }, + { + "epoch": 0.11481683339486747, + "grad_norm": 0.4282066226005554, + "learning_rate": 0.0001935876462493372, + "loss": 0.0474, + "step": 4898 + }, + { + "epoch": 0.11484027496967246, + "grad_norm": 0.6014747023582458, + "learning_rate": 0.00019358505068940165, + "loss": 0.1807, + "step": 4899 + }, + { + "epoch": 0.11486371654447745, + "grad_norm": 0.49905824661254883, + "learning_rate": 0.000193582454621671, + "loss": 0.1152, + "step": 4900 + }, + { + "epoch": 0.11488715811928245, + "grad_norm": 0.8138443231582642, + "learning_rate": 0.00019357985804615934, + "loss": 0.3115, + "step": 4901 + }, + { + "epoch": 0.11491059969408744, + "grad_norm": 0.5408506989479065, + "learning_rate": 0.0001935772609628807, + "loss": 0.1581, + "step": 4902 + }, + { + "epoch": 0.11493404126889245, + "grad_norm": 0.9616414308547974, + "learning_rate": 0.00019357466337184925, + "loss": 0.29, + "step": 4903 + }, + { + "epoch": 0.11495748284369744, + "grad_norm": 0.30078041553497314, + "learning_rate": 0.00019357206527307902, + "loss": 0.0775, + "step": 4904 + }, + { + "epoch": 0.11498092441850244, + "grad_norm": 0.5094637870788574, + "learning_rate": 0.00019356946666658416, + "loss": 0.8277, + "step": 4905 + }, + { + "epoch": 0.11500436599330743, + "grad_norm": 0.931945264339447, + "learning_rate": 0.00019356686755237873, + "loss": 0.7058, + "step": 4906 + }, + { + "epoch": 0.11502780756811243, + "grad_norm": 0.49665138125419617, + "learning_rate": 0.00019356426793047684, + "loss": 0.1102, + "step": 4907 + }, + { + "epoch": 0.11505124914291742, + "grad_norm": 0.3174557387828827, + "learning_rate": 0.0001935616678008926, + "loss": 0.0806, + "step": 4908 + }, + { + "epoch": 0.11507469071772242, + "grad_norm": 0.5422606468200684, + "learning_rate": 0.00019355906716364012, + "loss": 0.1874, + "step": 4909 + }, + { + "epoch": 0.11509813229252741, + "grad_norm": 0.6631536483764648, + "learning_rate": 0.00019355646601873353, + "loss": 0.2669, + "step": 4910 + }, + { + "epoch": 0.1151215738673324, + "grad_norm": 0.1820083111524582, + "learning_rate": 0.0001935538643661869, + "loss": 0.0388, + "step": 4911 + }, + { + "epoch": 0.1151450154421374, + "grad_norm": 1.4845404624938965, + "learning_rate": 0.0001935512622060144, + "loss": 0.1377, + "step": 4912 + }, + { + "epoch": 0.1151684570169424, + "grad_norm": 0.811065673828125, + "learning_rate": 0.00019354865953823008, + "loss": 0.1972, + "step": 4913 + }, + { + "epoch": 0.11519189859174739, + "grad_norm": 0.17247846722602844, + "learning_rate": 0.00019354605636284812, + "loss": 0.02, + "step": 4914 + }, + { + "epoch": 0.11521534016655238, + "grad_norm": 0.18947085738182068, + "learning_rate": 0.00019354345267988263, + "loss": 0.0697, + "step": 4915 + }, + { + "epoch": 0.11523878174135739, + "grad_norm": 0.551422119140625, + "learning_rate": 0.00019354084848934776, + "loss": 0.1529, + "step": 4916 + }, + { + "epoch": 0.11526222331616238, + "grad_norm": 0.6556926369667053, + "learning_rate": 0.00019353824379125757, + "loss": 0.1843, + "step": 4917 + }, + { + "epoch": 0.11528566489096738, + "grad_norm": 0.8047510385513306, + "learning_rate": 0.00019353563858562627, + "loss": 0.2647, + "step": 4918 + }, + { + "epoch": 0.11530910646577237, + "grad_norm": 0.4048628509044647, + "learning_rate": 0.00019353303287246795, + "loss": 0.1193, + "step": 4919 + }, + { + "epoch": 0.11533254804057737, + "grad_norm": 0.43582218885421753, + "learning_rate": 0.0001935304266517968, + "loss": 0.1023, + "step": 4920 + }, + { + "epoch": 0.11535598961538236, + "grad_norm": 0.7295442223548889, + "learning_rate": 0.00019352781992362688, + "loss": 0.1644, + "step": 4921 + }, + { + "epoch": 0.11537943119018736, + "grad_norm": 0.7629011869430542, + "learning_rate": 0.00019352521268797238, + "loss": 0.2129, + "step": 4922 + }, + { + "epoch": 0.11540287276499235, + "grad_norm": 0.5105053782463074, + "learning_rate": 0.00019352260494484748, + "loss": 0.156, + "step": 4923 + }, + { + "epoch": 0.11542631433979735, + "grad_norm": 0.9037526249885559, + "learning_rate": 0.00019351999669426628, + "loss": 0.1405, + "step": 4924 + }, + { + "epoch": 0.11544975591460234, + "grad_norm": 0.6654301881790161, + "learning_rate": 0.00019351738793624295, + "loss": 0.6291, + "step": 4925 + }, + { + "epoch": 0.11547319748940733, + "grad_norm": 0.5586565136909485, + "learning_rate": 0.00019351477867079163, + "loss": 0.137, + "step": 4926 + }, + { + "epoch": 0.11549663906421233, + "grad_norm": 0.4794611632823944, + "learning_rate": 0.00019351216889792653, + "loss": 0.1721, + "step": 4927 + }, + { + "epoch": 0.11552008063901732, + "grad_norm": 0.4230114221572876, + "learning_rate": 0.00019350955861766172, + "loss": 0.0768, + "step": 4928 + }, + { + "epoch": 0.11554352221382233, + "grad_norm": 0.5964739322662354, + "learning_rate": 0.00019350694783001147, + "loss": 0.1493, + "step": 4929 + }, + { + "epoch": 0.11556696378862732, + "grad_norm": 0.7722912430763245, + "learning_rate": 0.00019350433653498984, + "loss": 0.1991, + "step": 4930 + }, + { + "epoch": 0.11559040536343232, + "grad_norm": 0.3038357198238373, + "learning_rate": 0.00019350172473261108, + "loss": 0.0643, + "step": 4931 + }, + { + "epoch": 0.11561384693823731, + "grad_norm": 0.48823341727256775, + "learning_rate": 0.00019349911242288933, + "loss": 0.1444, + "step": 4932 + }, + { + "epoch": 0.11563728851304231, + "grad_norm": 0.7652583718299866, + "learning_rate": 0.00019349649960583875, + "loss": 0.1121, + "step": 4933 + }, + { + "epoch": 0.1156607300878473, + "grad_norm": 0.36847829818725586, + "learning_rate": 0.00019349388628147357, + "loss": 0.0637, + "step": 4934 + }, + { + "epoch": 0.1156841716626523, + "grad_norm": 0.33951422572135925, + "learning_rate": 0.0001934912724498079, + "loss": 0.1003, + "step": 4935 + }, + { + "epoch": 0.11570761323745729, + "grad_norm": 0.2559436559677124, + "learning_rate": 0.00019348865811085597, + "loss": 0.0572, + "step": 4936 + }, + { + "epoch": 0.11573105481226229, + "grad_norm": 0.5978869795799255, + "learning_rate": 0.00019348604326463193, + "loss": 0.1302, + "step": 4937 + }, + { + "epoch": 0.11575449638706728, + "grad_norm": 0.5482823848724365, + "learning_rate": 0.00019348342791115003, + "loss": 0.1077, + "step": 4938 + }, + { + "epoch": 0.11577793796187227, + "grad_norm": 0.6073734760284424, + "learning_rate": 0.00019348081205042438, + "loss": 0.1689, + "step": 4939 + }, + { + "epoch": 0.11580137953667727, + "grad_norm": 0.7926615476608276, + "learning_rate": 0.00019347819568246924, + "loss": 0.2604, + "step": 4940 + }, + { + "epoch": 0.11582482111148228, + "grad_norm": 0.8328088521957397, + "learning_rate": 0.00019347557880729878, + "loss": 0.2629, + "step": 4941 + }, + { + "epoch": 0.11584826268628727, + "grad_norm": 0.5330125689506531, + "learning_rate": 0.0001934729614249272, + "loss": 0.1487, + "step": 4942 + }, + { + "epoch": 0.11587170426109226, + "grad_norm": 0.7514384984970093, + "learning_rate": 0.00019347034353536872, + "loss": 0.1559, + "step": 4943 + }, + { + "epoch": 0.11589514583589726, + "grad_norm": 0.5861029624938965, + "learning_rate": 0.00019346772513863751, + "loss": 0.2102, + "step": 4944 + }, + { + "epoch": 0.11591858741070225, + "grad_norm": 0.8110901117324829, + "learning_rate": 0.0001934651062347478, + "loss": 0.2249, + "step": 4945 + }, + { + "epoch": 0.11594202898550725, + "grad_norm": 0.46790629625320435, + "learning_rate": 0.00019346248682371379, + "loss": 0.2058, + "step": 4946 + }, + { + "epoch": 0.11596547056031224, + "grad_norm": 0.3686502277851105, + "learning_rate": 0.00019345986690554972, + "loss": 0.0885, + "step": 4947 + }, + { + "epoch": 0.11598891213511724, + "grad_norm": 0.5091930031776428, + "learning_rate": 0.00019345724648026975, + "loss": 0.1068, + "step": 4948 + }, + { + "epoch": 0.11601235370992223, + "grad_norm": 0.8453034162521362, + "learning_rate": 0.00019345462554788814, + "loss": 0.2093, + "step": 4949 + }, + { + "epoch": 0.11603579528472723, + "grad_norm": 0.5791153311729431, + "learning_rate": 0.00019345200410841913, + "loss": 0.7256, + "step": 4950 + }, + { + "epoch": 0.11605923685953222, + "grad_norm": 1.2396419048309326, + "learning_rate": 0.0001934493821618769, + "loss": 0.3791, + "step": 4951 + }, + { + "epoch": 0.11608267843433721, + "grad_norm": 0.5729357600212097, + "learning_rate": 0.00019344675970827568, + "loss": 0.1745, + "step": 4952 + }, + { + "epoch": 0.11610612000914221, + "grad_norm": 0.5952863693237305, + "learning_rate": 0.00019344413674762974, + "loss": 0.1302, + "step": 4953 + }, + { + "epoch": 0.11612956158394722, + "grad_norm": 0.5916410684585571, + "learning_rate": 0.00019344151327995327, + "loss": 0.7286, + "step": 4954 + }, + { + "epoch": 0.11615300315875221, + "grad_norm": 0.15512633323669434, + "learning_rate": 0.00019343888930526052, + "loss": 0.0655, + "step": 4955 + }, + { + "epoch": 0.1161764447335572, + "grad_norm": 0.12929442524909973, + "learning_rate": 0.00019343626482356573, + "loss": 0.0213, + "step": 4956 + }, + { + "epoch": 0.1161998863083622, + "grad_norm": 0.6657840609550476, + "learning_rate": 0.00019343363983488317, + "loss": 0.1232, + "step": 4957 + }, + { + "epoch": 0.1162233278831672, + "grad_norm": 0.7750276923179626, + "learning_rate": 0.000193431014339227, + "loss": 0.1347, + "step": 4958 + }, + { + "epoch": 0.11624676945797219, + "grad_norm": 0.3051101863384247, + "learning_rate": 0.00019342838833661156, + "loss": 0.0761, + "step": 4959 + }, + { + "epoch": 0.11627021103277718, + "grad_norm": 0.3722939193248749, + "learning_rate": 0.00019342576182705107, + "loss": 0.0768, + "step": 4960 + }, + { + "epoch": 0.11629365260758218, + "grad_norm": 0.5559911131858826, + "learning_rate": 0.00019342313481055974, + "loss": 0.6704, + "step": 4961 + }, + { + "epoch": 0.11631709418238717, + "grad_norm": 0.2942192852497101, + "learning_rate": 0.00019342050728715188, + "loss": 0.0532, + "step": 4962 + }, + { + "epoch": 0.11634053575719217, + "grad_norm": 0.14642664790153503, + "learning_rate": 0.0001934178792568417, + "loss": 0.0297, + "step": 4963 + }, + { + "epoch": 0.11636397733199716, + "grad_norm": 0.33432820439338684, + "learning_rate": 0.00019341525071964346, + "loss": 0.105, + "step": 4964 + }, + { + "epoch": 0.11638741890680215, + "grad_norm": 0.6049864292144775, + "learning_rate": 0.00019341262167557147, + "loss": 0.2619, + "step": 4965 + }, + { + "epoch": 0.11641086048160715, + "grad_norm": 0.575694739818573, + "learning_rate": 0.00019340999212463995, + "loss": 0.6297, + "step": 4966 + }, + { + "epoch": 0.11643430205641216, + "grad_norm": 0.7230942249298096, + "learning_rate": 0.0001934073620668632, + "loss": 0.2054, + "step": 4967 + }, + { + "epoch": 0.11645774363121715, + "grad_norm": 0.5958495736122131, + "learning_rate": 0.0001934047315022555, + "loss": 0.1459, + "step": 4968 + }, + { + "epoch": 0.11648118520602214, + "grad_norm": 0.09750469028949738, + "learning_rate": 0.00019340210043083107, + "loss": 0.0162, + "step": 4969 + }, + { + "epoch": 0.11650462678082714, + "grad_norm": 0.4112282395362854, + "learning_rate": 0.00019339946885260422, + "loss": 0.0445, + "step": 4970 + }, + { + "epoch": 0.11652806835563213, + "grad_norm": 0.5469250679016113, + "learning_rate": 0.00019339683676758922, + "loss": 0.135, + "step": 4971 + }, + { + "epoch": 0.11655150993043713, + "grad_norm": 0.12179195135831833, + "learning_rate": 0.0001933942041758004, + "loss": 0.0395, + "step": 4972 + }, + { + "epoch": 0.11657495150524212, + "grad_norm": 0.2768514156341553, + "learning_rate": 0.00019339157107725195, + "loss": 0.1015, + "step": 4973 + }, + { + "epoch": 0.11659839308004712, + "grad_norm": 0.5901384949684143, + "learning_rate": 0.00019338893747195823, + "loss": 0.9206, + "step": 4974 + }, + { + "epoch": 0.11662183465485211, + "grad_norm": 0.24992553889751434, + "learning_rate": 0.0001933863033599335, + "loss": 0.0818, + "step": 4975 + }, + { + "epoch": 0.1166452762296571, + "grad_norm": 0.7224879860877991, + "learning_rate": 0.00019338366874119208, + "loss": 0.1466, + "step": 4976 + }, + { + "epoch": 0.1166687178044621, + "grad_norm": 0.459356427192688, + "learning_rate": 0.00019338103361574822, + "loss": 0.1009, + "step": 4977 + }, + { + "epoch": 0.1166921593792671, + "grad_norm": 0.15957479178905487, + "learning_rate": 0.00019337839798361627, + "loss": 0.0461, + "step": 4978 + }, + { + "epoch": 0.11671560095407209, + "grad_norm": 0.7504222393035889, + "learning_rate": 0.0001933757618448105, + "loss": 0.8458, + "step": 4979 + }, + { + "epoch": 0.1167390425288771, + "grad_norm": 0.40280571579933167, + "learning_rate": 0.00019337312519934524, + "loss": 0.1122, + "step": 4980 + }, + { + "epoch": 0.11676248410368209, + "grad_norm": 0.5029394030570984, + "learning_rate": 0.00019337048804723473, + "loss": 0.1126, + "step": 4981 + }, + { + "epoch": 0.11678592567848708, + "grad_norm": 0.5986918210983276, + "learning_rate": 0.00019336785038849335, + "loss": 0.7083, + "step": 4982 + }, + { + "epoch": 0.11680936725329208, + "grad_norm": 0.692916750907898, + "learning_rate": 0.0001933652122231354, + "loss": 0.0914, + "step": 4983 + }, + { + "epoch": 0.11683280882809707, + "grad_norm": 0.7927285432815552, + "learning_rate": 0.00019336257355117518, + "loss": 0.2119, + "step": 4984 + }, + { + "epoch": 0.11685625040290207, + "grad_norm": 0.6001626253128052, + "learning_rate": 0.00019335993437262703, + "loss": 0.1432, + "step": 4985 + }, + { + "epoch": 0.11687969197770706, + "grad_norm": 0.9429384469985962, + "learning_rate": 0.0001933572946875052, + "loss": 0.2812, + "step": 4986 + }, + { + "epoch": 0.11690313355251206, + "grad_norm": 0.5501313805580139, + "learning_rate": 0.0001933546544958241, + "loss": 0.7225, + "step": 4987 + }, + { + "epoch": 0.11692657512731705, + "grad_norm": 0.2645573019981384, + "learning_rate": 0.000193352013797598, + "loss": 0.0499, + "step": 4988 + }, + { + "epoch": 0.11695001670212204, + "grad_norm": 0.5333468317985535, + "learning_rate": 0.00019334937259284129, + "loss": 0.1328, + "step": 4989 + }, + { + "epoch": 0.11697345827692704, + "grad_norm": 3.1759285926818848, + "learning_rate": 0.00019334673088156824, + "loss": 0.2671, + "step": 4990 + }, + { + "epoch": 0.11699689985173203, + "grad_norm": 0.3566672205924988, + "learning_rate": 0.00019334408866379316, + "loss": 0.0945, + "step": 4991 + }, + { + "epoch": 0.11702034142653703, + "grad_norm": 0.5494361519813538, + "learning_rate": 0.00019334144593953048, + "loss": 0.1491, + "step": 4992 + }, + { + "epoch": 0.11704378300134204, + "grad_norm": 0.6025683283805847, + "learning_rate": 0.00019333880270879446, + "loss": 0.1803, + "step": 4993 + }, + { + "epoch": 0.11706722457614703, + "grad_norm": 0.8519682884216309, + "learning_rate": 0.00019333615897159948, + "loss": 0.2765, + "step": 4994 + }, + { + "epoch": 0.11709066615095202, + "grad_norm": 0.16257262229919434, + "learning_rate": 0.0001933335147279599, + "loss": 0.0496, + "step": 4995 + }, + { + "epoch": 0.11711410772575702, + "grad_norm": 0.3568175435066223, + "learning_rate": 0.00019333086997789001, + "loss": 0.0821, + "step": 4996 + }, + { + "epoch": 0.11713754930056201, + "grad_norm": 0.167741060256958, + "learning_rate": 0.00019332822472140423, + "loss": 0.0483, + "step": 4997 + }, + { + "epoch": 0.11716099087536701, + "grad_norm": 0.8560519814491272, + "learning_rate": 0.00019332557895851687, + "loss": 0.1454, + "step": 4998 + }, + { + "epoch": 0.117184432450172, + "grad_norm": 0.4425429403781891, + "learning_rate": 0.00019332293268924228, + "loss": 0.0764, + "step": 4999 + }, + { + "epoch": 0.117207874024977, + "grad_norm": 0.6130775809288025, + "learning_rate": 0.00019332028591359483, + "loss": 0.1903, + "step": 5000 + }, + { + "epoch": 0.11723131559978199, + "grad_norm": 0.6001055836677551, + "learning_rate": 0.00019331763863158892, + "loss": 0.4914, + "step": 5001 + }, + { + "epoch": 0.11725475717458698, + "grad_norm": 0.2926148474216461, + "learning_rate": 0.00019331499084323882, + "loss": 0.0858, + "step": 5002 + }, + { + "epoch": 0.11727819874939198, + "grad_norm": 0.5526954531669617, + "learning_rate": 0.000193312342548559, + "loss": 0.1631, + "step": 5003 + }, + { + "epoch": 0.11730164032419697, + "grad_norm": 0.26925039291381836, + "learning_rate": 0.00019330969374756378, + "loss": 0.0826, + "step": 5004 + }, + { + "epoch": 0.11732508189900198, + "grad_norm": 0.7351076006889343, + "learning_rate": 0.00019330704444026756, + "loss": 0.191, + "step": 5005 + }, + { + "epoch": 0.11734852347380698, + "grad_norm": 0.44038864970207214, + "learning_rate": 0.00019330439462668465, + "loss": 0.135, + "step": 5006 + }, + { + "epoch": 0.11737196504861197, + "grad_norm": 0.7984957098960876, + "learning_rate": 0.0001933017443068295, + "loss": 0.6334, + "step": 5007 + }, + { + "epoch": 0.11739540662341696, + "grad_norm": 0.4411831200122833, + "learning_rate": 0.00019329909348071643, + "loss": 0.1395, + "step": 5008 + }, + { + "epoch": 0.11741884819822196, + "grad_norm": 1.1283355951309204, + "learning_rate": 0.0001932964421483599, + "loss": 0.2198, + "step": 5009 + }, + { + "epoch": 0.11744228977302695, + "grad_norm": 0.5613596439361572, + "learning_rate": 0.0001932937903097742, + "loss": 0.097, + "step": 5010 + }, + { + "epoch": 0.11746573134783195, + "grad_norm": 0.3914056718349457, + "learning_rate": 0.0001932911379649738, + "loss": 0.1194, + "step": 5011 + }, + { + "epoch": 0.11748917292263694, + "grad_norm": 0.6485000252723694, + "learning_rate": 0.00019328848511397306, + "loss": 0.2447, + "step": 5012 + }, + { + "epoch": 0.11751261449744194, + "grad_norm": 0.7877719402313232, + "learning_rate": 0.00019328583175678637, + "loss": 0.1685, + "step": 5013 + }, + { + "epoch": 0.11753605607224693, + "grad_norm": 0.8572128415107727, + "learning_rate": 0.00019328317789342814, + "loss": 0.2396, + "step": 5014 + }, + { + "epoch": 0.11755949764705192, + "grad_norm": 0.19659480452537537, + "learning_rate": 0.00019328052352391277, + "loss": 0.0587, + "step": 5015 + }, + { + "epoch": 0.11758293922185692, + "grad_norm": 0.7041271924972534, + "learning_rate": 0.00019327786864825464, + "loss": 0.2251, + "step": 5016 + }, + { + "epoch": 0.11760638079666191, + "grad_norm": 0.6544857621192932, + "learning_rate": 0.00019327521326646817, + "loss": 0.1418, + "step": 5017 + }, + { + "epoch": 0.11762982237146692, + "grad_norm": 1.1598812341690063, + "learning_rate": 0.0001932725573785678, + "loss": 0.1342, + "step": 5018 + }, + { + "epoch": 0.11765326394627192, + "grad_norm": 0.5325208902359009, + "learning_rate": 0.0001932699009845679, + "loss": 0.6177, + "step": 5019 + }, + { + "epoch": 0.11767670552107691, + "grad_norm": 0.14813606441020966, + "learning_rate": 0.00019326724408448288, + "loss": 0.0432, + "step": 5020 + }, + { + "epoch": 0.1177001470958819, + "grad_norm": 0.6246055364608765, + "learning_rate": 0.0001932645866783272, + "loss": 0.2302, + "step": 5021 + }, + { + "epoch": 0.1177235886706869, + "grad_norm": 0.4092569947242737, + "learning_rate": 0.00019326192876611522, + "loss": 0.1281, + "step": 5022 + }, + { + "epoch": 0.1177470302454919, + "grad_norm": 0.7166110873222351, + "learning_rate": 0.0001932592703478614, + "loss": 0.1501, + "step": 5023 + }, + { + "epoch": 0.11777047182029689, + "grad_norm": 0.37270474433898926, + "learning_rate": 0.00019325661142358016, + "loss": 0.1128, + "step": 5024 + }, + { + "epoch": 0.11779391339510188, + "grad_norm": 0.789855420589447, + "learning_rate": 0.00019325395199328593, + "loss": 0.1755, + "step": 5025 + }, + { + "epoch": 0.11781735496990688, + "grad_norm": 0.7009834051132202, + "learning_rate": 0.00019325129205699312, + "loss": 0.1622, + "step": 5026 + }, + { + "epoch": 0.11784079654471187, + "grad_norm": 0.7710626125335693, + "learning_rate": 0.00019324863161471618, + "loss": 0.1277, + "step": 5027 + }, + { + "epoch": 0.11786423811951686, + "grad_norm": 0.399604856967926, + "learning_rate": 0.00019324597066646955, + "loss": 0.0762, + "step": 5028 + }, + { + "epoch": 0.11788767969432186, + "grad_norm": 0.4946717619895935, + "learning_rate": 0.00019324330921226764, + "loss": 0.2107, + "step": 5029 + }, + { + "epoch": 0.11791112126912685, + "grad_norm": 0.3840550482273102, + "learning_rate": 0.00019324064725212496, + "loss": 0.0982, + "step": 5030 + }, + { + "epoch": 0.11793456284393186, + "grad_norm": 0.5050975680351257, + "learning_rate": 0.0001932379847860559, + "loss": 0.1144, + "step": 5031 + }, + { + "epoch": 0.11795800441873686, + "grad_norm": 0.09259575605392456, + "learning_rate": 0.00019323532181407486, + "loss": 0.0217, + "step": 5032 + }, + { + "epoch": 0.11798144599354185, + "grad_norm": 0.5216699838638306, + "learning_rate": 0.00019323265833619637, + "loss": 0.2145, + "step": 5033 + }, + { + "epoch": 0.11800488756834684, + "grad_norm": 0.602277934551239, + "learning_rate": 0.00019322999435243484, + "loss": 0.1917, + "step": 5034 + }, + { + "epoch": 0.11802832914315184, + "grad_norm": 0.2353886067867279, + "learning_rate": 0.00019322732986280477, + "loss": 0.0482, + "step": 5035 + }, + { + "epoch": 0.11805177071795683, + "grad_norm": 0.22792261838912964, + "learning_rate": 0.00019322466486732057, + "loss": 0.0403, + "step": 5036 + }, + { + "epoch": 0.11807521229276183, + "grad_norm": 0.26347216963768005, + "learning_rate": 0.0001932219993659967, + "loss": 0.0808, + "step": 5037 + }, + { + "epoch": 0.11809865386756682, + "grad_norm": 0.38420021533966064, + "learning_rate": 0.00019321933335884767, + "loss": 0.4834, + "step": 5038 + }, + { + "epoch": 0.11812209544237182, + "grad_norm": 0.4735865294933319, + "learning_rate": 0.00019321666684588793, + "loss": 0.4326, + "step": 5039 + }, + { + "epoch": 0.11814553701717681, + "grad_norm": 0.1324344426393509, + "learning_rate": 0.00019321399982713192, + "loss": 0.0322, + "step": 5040 + }, + { + "epoch": 0.1181689785919818, + "grad_norm": 0.46292412281036377, + "learning_rate": 0.0001932113323025941, + "loss": 0.0789, + "step": 5041 + }, + { + "epoch": 0.1181924201667868, + "grad_norm": 0.5427022576332092, + "learning_rate": 0.000193208664272289, + "loss": 0.1369, + "step": 5042 + }, + { + "epoch": 0.1182158617415918, + "grad_norm": 0.6594032645225525, + "learning_rate": 0.00019320599573623105, + "loss": 0.179, + "step": 5043 + }, + { + "epoch": 0.1182393033163968, + "grad_norm": 0.43548280000686646, + "learning_rate": 0.00019320332669443474, + "loss": 0.1153, + "step": 5044 + }, + { + "epoch": 0.1182627448912018, + "grad_norm": 0.30410999059677124, + "learning_rate": 0.00019320065714691458, + "loss": 0.0682, + "step": 5045 + }, + { + "epoch": 0.11828618646600679, + "grad_norm": 0.6365828514099121, + "learning_rate": 0.00019319798709368504, + "loss": 0.226, + "step": 5046 + }, + { + "epoch": 0.11830962804081178, + "grad_norm": 0.25068745017051697, + "learning_rate": 0.00019319531653476057, + "loss": 0.0682, + "step": 5047 + }, + { + "epoch": 0.11833306961561678, + "grad_norm": 0.23093834519386292, + "learning_rate": 0.00019319264547015572, + "loss": 0.0604, + "step": 5048 + }, + { + "epoch": 0.11835651119042177, + "grad_norm": 0.6342729926109314, + "learning_rate": 0.00019318997389988494, + "loss": 0.8249, + "step": 5049 + }, + { + "epoch": 0.11837995276522677, + "grad_norm": 0.45357397198677063, + "learning_rate": 0.0001931873018239627, + "loss": 0.144, + "step": 5050 + }, + { + "epoch": 0.11840339434003176, + "grad_norm": 0.306598424911499, + "learning_rate": 0.00019318462924240362, + "loss": 0.0908, + "step": 5051 + }, + { + "epoch": 0.11842683591483676, + "grad_norm": 0.7236713171005249, + "learning_rate": 0.0001931819561552221, + "loss": 0.1495, + "step": 5052 + }, + { + "epoch": 0.11845027748964175, + "grad_norm": 0.6431005597114563, + "learning_rate": 0.00019317928256243265, + "loss": 0.2076, + "step": 5053 + }, + { + "epoch": 0.11847371906444674, + "grad_norm": 0.5245051383972168, + "learning_rate": 0.0001931766084640498, + "loss": 0.15, + "step": 5054 + }, + { + "epoch": 0.11849716063925174, + "grad_norm": 0.6645878553390503, + "learning_rate": 0.00019317393386008802, + "loss": 0.2136, + "step": 5055 + }, + { + "epoch": 0.11852060221405675, + "grad_norm": 0.623523473739624, + "learning_rate": 0.0001931712587505619, + "loss": 0.1861, + "step": 5056 + }, + { + "epoch": 0.11854404378886174, + "grad_norm": 0.6580193638801575, + "learning_rate": 0.00019316858313548588, + "loss": 0.285, + "step": 5057 + }, + { + "epoch": 0.11856748536366674, + "grad_norm": 0.3581666648387909, + "learning_rate": 0.00019316590701487453, + "loss": 0.0811, + "step": 5058 + }, + { + "epoch": 0.11859092693847173, + "grad_norm": 0.7595141530036926, + "learning_rate": 0.00019316323038874234, + "loss": 0.153, + "step": 5059 + }, + { + "epoch": 0.11861436851327672, + "grad_norm": 0.5046383738517761, + "learning_rate": 0.0001931605532571038, + "loss": 0.2, + "step": 5060 + }, + { + "epoch": 0.11863781008808172, + "grad_norm": 0.534795880317688, + "learning_rate": 0.00019315787561997354, + "loss": 0.1379, + "step": 5061 + }, + { + "epoch": 0.11866125166288671, + "grad_norm": 0.6397818326950073, + "learning_rate": 0.00019315519747736603, + "loss": 0.1217, + "step": 5062 + }, + { + "epoch": 0.11868469323769171, + "grad_norm": 0.3467605412006378, + "learning_rate": 0.00019315251882929578, + "loss": 0.0942, + "step": 5063 + }, + { + "epoch": 0.1187081348124967, + "grad_norm": 0.46143436431884766, + "learning_rate": 0.00019314983967577733, + "loss": 0.0801, + "step": 5064 + }, + { + "epoch": 0.1187315763873017, + "grad_norm": 0.7359980940818787, + "learning_rate": 0.00019314716001682522, + "loss": 0.2246, + "step": 5065 + }, + { + "epoch": 0.11875501796210669, + "grad_norm": 0.24187862873077393, + "learning_rate": 0.00019314447985245404, + "loss": 0.0773, + "step": 5066 + }, + { + "epoch": 0.11877845953691168, + "grad_norm": 0.27644336223602295, + "learning_rate": 0.00019314179918267829, + "loss": 0.0888, + "step": 5067 + }, + { + "epoch": 0.11880190111171668, + "grad_norm": 0.18605558574199677, + "learning_rate": 0.0001931391180075125, + "loss": 0.0305, + "step": 5068 + }, + { + "epoch": 0.11882534268652169, + "grad_norm": 0.6233201622962952, + "learning_rate": 0.00019313643632697125, + "loss": 0.6743, + "step": 5069 + }, + { + "epoch": 0.11884878426132668, + "grad_norm": 0.44130757451057434, + "learning_rate": 0.00019313375414106907, + "loss": 0.1084, + "step": 5070 + }, + { + "epoch": 0.11887222583613168, + "grad_norm": 0.5565061569213867, + "learning_rate": 0.0001931310714498205, + "loss": 0.1583, + "step": 5071 + }, + { + "epoch": 0.11889566741093667, + "grad_norm": 0.3243198096752167, + "learning_rate": 0.00019312838825324014, + "loss": 0.0832, + "step": 5072 + }, + { + "epoch": 0.11891910898574166, + "grad_norm": 0.5794554948806763, + "learning_rate": 0.00019312570455134252, + "loss": 0.1978, + "step": 5073 + }, + { + "epoch": 0.11894255056054666, + "grad_norm": 0.6760635375976562, + "learning_rate": 0.00019312302034414224, + "loss": 0.1748, + "step": 5074 + }, + { + "epoch": 0.11896599213535165, + "grad_norm": 0.906816303730011, + "learning_rate": 0.00019312033563165376, + "loss": 0.2039, + "step": 5075 + }, + { + "epoch": 0.11898943371015665, + "grad_norm": 0.7243847846984863, + "learning_rate": 0.0001931176504138918, + "loss": 0.2749, + "step": 5076 + }, + { + "epoch": 0.11901287528496164, + "grad_norm": 0.2778832018375397, + "learning_rate": 0.00019311496469087081, + "loss": 0.107, + "step": 5077 + }, + { + "epoch": 0.11903631685976664, + "grad_norm": 0.2290181815624237, + "learning_rate": 0.00019311227846260544, + "loss": 0.0476, + "step": 5078 + }, + { + "epoch": 0.11905975843457163, + "grad_norm": 0.7346113324165344, + "learning_rate": 0.0001931095917291102, + "loss": 0.2538, + "step": 5079 + }, + { + "epoch": 0.11908320000937662, + "grad_norm": 0.3091098666191101, + "learning_rate": 0.00019310690449039964, + "loss": 0.0542, + "step": 5080 + }, + { + "epoch": 0.11910664158418162, + "grad_norm": 0.7446015477180481, + "learning_rate": 0.00019310421674648847, + "loss": 0.238, + "step": 5081 + }, + { + "epoch": 0.11913008315898663, + "grad_norm": 0.45357874035835266, + "learning_rate": 0.00019310152849739122, + "loss": 0.1207, + "step": 5082 + }, + { + "epoch": 0.11915352473379162, + "grad_norm": 0.5481953620910645, + "learning_rate": 0.00019309883974312242, + "loss": 0.704, + "step": 5083 + }, + { + "epoch": 0.11917696630859662, + "grad_norm": 0.33546459674835205, + "learning_rate": 0.00019309615048369668, + "loss": 0.1036, + "step": 5084 + }, + { + "epoch": 0.11920040788340161, + "grad_norm": 0.5485216379165649, + "learning_rate": 0.00019309346071912865, + "loss": 0.1012, + "step": 5085 + }, + { + "epoch": 0.1192238494582066, + "grad_norm": 0.7611598968505859, + "learning_rate": 0.00019309077044943286, + "loss": 0.1679, + "step": 5086 + }, + { + "epoch": 0.1192472910330116, + "grad_norm": 0.25541985034942627, + "learning_rate": 0.00019308807967462395, + "loss": 0.08, + "step": 5087 + }, + { + "epoch": 0.11927073260781659, + "grad_norm": 0.5387372374534607, + "learning_rate": 0.0001930853883947165, + "loss": 0.151, + "step": 5088 + }, + { + "epoch": 0.11929417418262159, + "grad_norm": 0.49248579144477844, + "learning_rate": 0.00019308269660972508, + "loss": 0.1301, + "step": 5089 + }, + { + "epoch": 0.11931761575742658, + "grad_norm": 0.20892658829689026, + "learning_rate": 0.00019308000431966438, + "loss": 0.0443, + "step": 5090 + }, + { + "epoch": 0.11934105733223158, + "grad_norm": 0.11111277341842651, + "learning_rate": 0.00019307731152454894, + "loss": 0.0166, + "step": 5091 + }, + { + "epoch": 0.11936449890703657, + "grad_norm": 0.5139471292495728, + "learning_rate": 0.0001930746182243934, + "loss": 0.1163, + "step": 5092 + }, + { + "epoch": 0.11938794048184156, + "grad_norm": 0.6360413432121277, + "learning_rate": 0.00019307192441921233, + "loss": 0.2002, + "step": 5093 + }, + { + "epoch": 0.11941138205664656, + "grad_norm": 0.5483053922653198, + "learning_rate": 0.0001930692301090204, + "loss": 0.6381, + "step": 5094 + }, + { + "epoch": 0.11943482363145157, + "grad_norm": 0.4143471419811249, + "learning_rate": 0.00019306653529383222, + "loss": 0.09, + "step": 5095 + }, + { + "epoch": 0.11945826520625656, + "grad_norm": 0.5626707077026367, + "learning_rate": 0.0001930638399736624, + "loss": 0.5705, + "step": 5096 + }, + { + "epoch": 0.11948170678106156, + "grad_norm": 1.3824199438095093, + "learning_rate": 0.0001930611441485256, + "loss": 0.128, + "step": 5097 + }, + { + "epoch": 0.11950514835586655, + "grad_norm": 0.517864465713501, + "learning_rate": 0.00019305844781843633, + "loss": 0.2461, + "step": 5098 + }, + { + "epoch": 0.11952858993067154, + "grad_norm": 0.32526400685310364, + "learning_rate": 0.00019305575098340938, + "loss": 0.0834, + "step": 5099 + }, + { + "epoch": 0.11955203150547654, + "grad_norm": 0.822178065776825, + "learning_rate": 0.0001930530536434593, + "loss": 0.1266, + "step": 5100 + }, + { + "epoch": 0.11957547308028153, + "grad_norm": 0.45751458406448364, + "learning_rate": 0.0001930503557986007, + "loss": 0.1018, + "step": 5101 + }, + { + "epoch": 0.11959891465508653, + "grad_norm": 0.6393406987190247, + "learning_rate": 0.00019304765744884828, + "loss": 0.1692, + "step": 5102 + }, + { + "epoch": 0.11962235622989152, + "grad_norm": 0.47173836827278137, + "learning_rate": 0.00019304495859421664, + "loss": 0.1004, + "step": 5103 + }, + { + "epoch": 0.11964579780469652, + "grad_norm": 0.6546781063079834, + "learning_rate": 0.00019304225923472043, + "loss": 0.4515, + "step": 5104 + }, + { + "epoch": 0.11966923937950151, + "grad_norm": 0.25394847989082336, + "learning_rate": 0.00019303955937037434, + "loss": 0.0539, + "step": 5105 + }, + { + "epoch": 0.1196926809543065, + "grad_norm": 0.9269647598266602, + "learning_rate": 0.00019303685900119298, + "loss": 0.113, + "step": 5106 + }, + { + "epoch": 0.1197161225291115, + "grad_norm": 0.453998327255249, + "learning_rate": 0.00019303415812719097, + "loss": 0.0708, + "step": 5107 + }, + { + "epoch": 0.11973956410391651, + "grad_norm": 0.4597581923007965, + "learning_rate": 0.00019303145674838304, + "loss": 0.2932, + "step": 5108 + }, + { + "epoch": 0.1197630056787215, + "grad_norm": 0.49524518847465515, + "learning_rate": 0.00019302875486478378, + "loss": 0.0701, + "step": 5109 + }, + { + "epoch": 0.1197864472535265, + "grad_norm": 0.20644128322601318, + "learning_rate": 0.0001930260524764079, + "loss": 0.059, + "step": 5110 + }, + { + "epoch": 0.11980988882833149, + "grad_norm": 0.25759807229042053, + "learning_rate": 0.00019302334958327003, + "loss": 0.0774, + "step": 5111 + }, + { + "epoch": 0.11983333040313648, + "grad_norm": 0.3974683880805969, + "learning_rate": 0.00019302064618538487, + "loss": 0.0612, + "step": 5112 + }, + { + "epoch": 0.11985677197794148, + "grad_norm": 0.38083967566490173, + "learning_rate": 0.00019301794228276706, + "loss": 0.0859, + "step": 5113 + }, + { + "epoch": 0.11988021355274647, + "grad_norm": 0.25734737515449524, + "learning_rate": 0.00019301523787543126, + "loss": 0.0749, + "step": 5114 + }, + { + "epoch": 0.11990365512755147, + "grad_norm": 0.44145143032073975, + "learning_rate": 0.0001930125329633922, + "loss": 0.1113, + "step": 5115 + }, + { + "epoch": 0.11992709670235646, + "grad_norm": 0.6733248233795166, + "learning_rate": 0.00019300982754666448, + "loss": 0.2444, + "step": 5116 + }, + { + "epoch": 0.11995053827716146, + "grad_norm": 0.6875708699226379, + "learning_rate": 0.00019300712162526287, + "loss": 0.1405, + "step": 5117 + }, + { + "epoch": 0.11997397985196645, + "grad_norm": 0.8257202506065369, + "learning_rate": 0.00019300441519920196, + "loss": 0.2802, + "step": 5118 + }, + { + "epoch": 0.11999742142677144, + "grad_norm": 0.6804804801940918, + "learning_rate": 0.0001930017082684965, + "loss": 0.1067, + "step": 5119 + }, + { + "epoch": 0.12002086300157645, + "grad_norm": 0.619547426700592, + "learning_rate": 0.00019299900083316114, + "loss": 0.1833, + "step": 5120 + }, + { + "epoch": 0.12004430457638145, + "grad_norm": 0.6598297357559204, + "learning_rate": 0.0001929962928932106, + "loss": 0.7483, + "step": 5121 + }, + { + "epoch": 0.12006774615118644, + "grad_norm": 0.23809164762496948, + "learning_rate": 0.00019299358444865953, + "loss": 0.0534, + "step": 5122 + }, + { + "epoch": 0.12009118772599144, + "grad_norm": 0.5493518710136414, + "learning_rate": 0.0001929908754995227, + "loss": 0.1269, + "step": 5123 + }, + { + "epoch": 0.12011462930079643, + "grad_norm": 0.9181116223335266, + "learning_rate": 0.00019298816604581474, + "loss": 0.1577, + "step": 5124 + }, + { + "epoch": 0.12013807087560142, + "grad_norm": 0.7451571822166443, + "learning_rate": 0.00019298545608755039, + "loss": 0.1765, + "step": 5125 + }, + { + "epoch": 0.12016151245040642, + "grad_norm": 0.5383606553077698, + "learning_rate": 0.00019298274562474434, + "loss": 0.5261, + "step": 5126 + }, + { + "epoch": 0.12018495402521141, + "grad_norm": 0.48138806223869324, + "learning_rate": 0.00019298003465741128, + "loss": 0.081, + "step": 5127 + }, + { + "epoch": 0.12020839560001641, + "grad_norm": 0.21310171484947205, + "learning_rate": 0.00019297732318556595, + "loss": 0.0464, + "step": 5128 + }, + { + "epoch": 0.1202318371748214, + "grad_norm": 0.1244264617562294, + "learning_rate": 0.00019297461120922304, + "loss": 0.0366, + "step": 5129 + }, + { + "epoch": 0.1202552787496264, + "grad_norm": 0.3044517934322357, + "learning_rate": 0.0001929718987283973, + "loss": 0.0823, + "step": 5130 + }, + { + "epoch": 0.12027872032443139, + "grad_norm": 0.6664391160011292, + "learning_rate": 0.00019296918574310338, + "loss": 0.1686, + "step": 5131 + }, + { + "epoch": 0.12030216189923638, + "grad_norm": 0.6119699478149414, + "learning_rate": 0.00019296647225335609, + "loss": 0.1718, + "step": 5132 + }, + { + "epoch": 0.12032560347404139, + "grad_norm": 0.24148228764533997, + "learning_rate": 0.00019296375825917007, + "loss": 0.0587, + "step": 5133 + }, + { + "epoch": 0.12034904504884639, + "grad_norm": 0.4678937792778015, + "learning_rate": 0.00019296104376056013, + "loss": 0.0982, + "step": 5134 + }, + { + "epoch": 0.12037248662365138, + "grad_norm": 0.6071341633796692, + "learning_rate": 0.00019295832875754093, + "loss": 0.173, + "step": 5135 + }, + { + "epoch": 0.12039592819845638, + "grad_norm": 0.2233961969614029, + "learning_rate": 0.00019295561325012722, + "loss": 0.0723, + "step": 5136 + }, + { + "epoch": 0.12041936977326137, + "grad_norm": 0.3302127420902252, + "learning_rate": 0.00019295289723833374, + "loss": 0.1201, + "step": 5137 + }, + { + "epoch": 0.12044281134806636, + "grad_norm": 0.40021589398384094, + "learning_rate": 0.0001929501807221752, + "loss": 0.3966, + "step": 5138 + }, + { + "epoch": 0.12046625292287136, + "grad_norm": 0.38679569959640503, + "learning_rate": 0.00019294746370166637, + "loss": 0.133, + "step": 5139 + }, + { + "epoch": 0.12048969449767635, + "grad_norm": 0.4983476996421814, + "learning_rate": 0.00019294474617682203, + "loss": 0.1418, + "step": 5140 + }, + { + "epoch": 0.12051313607248135, + "grad_norm": 0.4617307484149933, + "learning_rate": 0.00019294202814765688, + "loss": 0.1241, + "step": 5141 + }, + { + "epoch": 0.12053657764728634, + "grad_norm": 0.44005393981933594, + "learning_rate": 0.00019293930961418565, + "loss": 0.0793, + "step": 5142 + }, + { + "epoch": 0.12056001922209134, + "grad_norm": 0.7466046214103699, + "learning_rate": 0.0001929365905764231, + "loss": 0.1653, + "step": 5143 + }, + { + "epoch": 0.12058346079689633, + "grad_norm": 0.21471598744392395, + "learning_rate": 0.00019293387103438403, + "loss": 0.0504, + "step": 5144 + }, + { + "epoch": 0.12060690237170132, + "grad_norm": 0.4573913514614105, + "learning_rate": 0.00019293115098808315, + "loss": 0.201, + "step": 5145 + }, + { + "epoch": 0.12063034394650633, + "grad_norm": 0.6678661108016968, + "learning_rate": 0.0001929284304375352, + "loss": 0.1816, + "step": 5146 + }, + { + "epoch": 0.12065378552131133, + "grad_norm": 0.6297444701194763, + "learning_rate": 0.00019292570938275503, + "loss": 0.1735, + "step": 5147 + }, + { + "epoch": 0.12067722709611632, + "grad_norm": 0.513887882232666, + "learning_rate": 0.00019292298782375733, + "loss": 0.1396, + "step": 5148 + }, + { + "epoch": 0.12070066867092132, + "grad_norm": 0.6324220895767212, + "learning_rate": 0.00019292026576055686, + "loss": 0.1651, + "step": 5149 + }, + { + "epoch": 0.12072411024572631, + "grad_norm": 0.3417743444442749, + "learning_rate": 0.00019291754319316842, + "loss": 0.0874, + "step": 5150 + }, + { + "epoch": 0.1207475518205313, + "grad_norm": 0.18697841465473175, + "learning_rate": 0.0001929148201216068, + "loss": 0.044, + "step": 5151 + }, + { + "epoch": 0.1207709933953363, + "grad_norm": 0.522650957107544, + "learning_rate": 0.00019291209654588674, + "loss": 0.2109, + "step": 5152 + }, + { + "epoch": 0.12079443497014129, + "grad_norm": 0.5628789663314819, + "learning_rate": 0.00019290937246602306, + "loss": 0.1459, + "step": 5153 + }, + { + "epoch": 0.12081787654494629, + "grad_norm": 0.48276448249816895, + "learning_rate": 0.00019290664788203045, + "loss": 0.0955, + "step": 5154 + }, + { + "epoch": 0.12084131811975128, + "grad_norm": 0.8983280658721924, + "learning_rate": 0.00019290392279392381, + "loss": 0.2178, + "step": 5155 + }, + { + "epoch": 0.12086475969455628, + "grad_norm": 0.6469811201095581, + "learning_rate": 0.00019290119720171783, + "loss": 0.9186, + "step": 5156 + }, + { + "epoch": 0.12088820126936127, + "grad_norm": 0.3778000771999359, + "learning_rate": 0.0001928984711054274, + "loss": 0.0851, + "step": 5157 + }, + { + "epoch": 0.12091164284416626, + "grad_norm": 0.47568705677986145, + "learning_rate": 0.0001928957445050672, + "loss": 0.1399, + "step": 5158 + }, + { + "epoch": 0.12093508441897127, + "grad_norm": 0.48747870326042175, + "learning_rate": 0.0001928930174006521, + "loss": 0.0911, + "step": 5159 + }, + { + "epoch": 0.12095852599377627, + "grad_norm": 0.46357113122940063, + "learning_rate": 0.0001928902897921969, + "loss": 0.4815, + "step": 5160 + }, + { + "epoch": 0.12098196756858126, + "grad_norm": 1.944098711013794, + "learning_rate": 0.00019288756167971634, + "loss": 0.2325, + "step": 5161 + }, + { + "epoch": 0.12100540914338626, + "grad_norm": 0.5379078984260559, + "learning_rate": 0.00019288483306322527, + "loss": 0.73, + "step": 5162 + }, + { + "epoch": 0.12102885071819125, + "grad_norm": 0.243895024061203, + "learning_rate": 0.00019288210394273846, + "loss": 0.0465, + "step": 5163 + }, + { + "epoch": 0.12105229229299624, + "grad_norm": 0.1827118843793869, + "learning_rate": 0.0001928793743182708, + "loss": 0.0599, + "step": 5164 + }, + { + "epoch": 0.12107573386780124, + "grad_norm": 0.18641765415668488, + "learning_rate": 0.000192876644189837, + "loss": 0.0271, + "step": 5165 + }, + { + "epoch": 0.12109917544260623, + "grad_norm": 0.5235726237297058, + "learning_rate": 0.0001928739135574519, + "loss": 0.1146, + "step": 5166 + }, + { + "epoch": 0.12112261701741123, + "grad_norm": 0.08510436117649078, + "learning_rate": 0.00019287118242113037, + "loss": 0.0214, + "step": 5167 + }, + { + "epoch": 0.12114605859221622, + "grad_norm": 0.36803242564201355, + "learning_rate": 0.0001928684507808872, + "loss": 0.0493, + "step": 5168 + }, + { + "epoch": 0.12116950016702122, + "grad_norm": 0.5251390337944031, + "learning_rate": 0.00019286571863673716, + "loss": 0.1084, + "step": 5169 + }, + { + "epoch": 0.12119294174182621, + "grad_norm": 0.315143346786499, + "learning_rate": 0.00019286298598869517, + "loss": 0.0826, + "step": 5170 + }, + { + "epoch": 0.1212163833166312, + "grad_norm": 0.5971767902374268, + "learning_rate": 0.00019286025283677598, + "loss": 0.1052, + "step": 5171 + }, + { + "epoch": 0.12123982489143621, + "grad_norm": 0.41717013716697693, + "learning_rate": 0.00019285751918099445, + "loss": 0.1138, + "step": 5172 + }, + { + "epoch": 0.1212632664662412, + "grad_norm": 0.5828597545623779, + "learning_rate": 0.00019285478502136544, + "loss": 0.21, + "step": 5173 + }, + { + "epoch": 0.1212867080410462, + "grad_norm": 0.6040427088737488, + "learning_rate": 0.0001928520503579037, + "loss": 0.7992, + "step": 5174 + }, + { + "epoch": 0.1213101496158512, + "grad_norm": 1.079635500907898, + "learning_rate": 0.00019284931519062415, + "loss": 0.1556, + "step": 5175 + }, + { + "epoch": 0.12133359119065619, + "grad_norm": 0.5185387134552002, + "learning_rate": 0.00019284657951954158, + "loss": 0.0202, + "step": 5176 + }, + { + "epoch": 0.12135703276546118, + "grad_norm": 0.4466891288757324, + "learning_rate": 0.00019284384334467092, + "loss": 0.07, + "step": 5177 + }, + { + "epoch": 0.12138047434026618, + "grad_norm": 0.1549254059791565, + "learning_rate": 0.00019284110666602691, + "loss": 0.0419, + "step": 5178 + }, + { + "epoch": 0.12140391591507117, + "grad_norm": 1.0458208322525024, + "learning_rate": 0.00019283836948362447, + "loss": 0.2424, + "step": 5179 + }, + { + "epoch": 0.12142735748987617, + "grad_norm": 0.6324083209037781, + "learning_rate": 0.00019283563179747838, + "loss": 0.8419, + "step": 5180 + }, + { + "epoch": 0.12145079906468116, + "grad_norm": 0.29667577147483826, + "learning_rate": 0.0001928328936076036, + "loss": 0.0766, + "step": 5181 + }, + { + "epoch": 0.12147424063948616, + "grad_norm": 0.5566476583480835, + "learning_rate": 0.0001928301549140149, + "loss": 0.1655, + "step": 5182 + }, + { + "epoch": 0.12149768221429115, + "grad_norm": 0.6471428871154785, + "learning_rate": 0.00019282741571672717, + "loss": 0.1356, + "step": 5183 + }, + { + "epoch": 0.12152112378909616, + "grad_norm": 0.5106896758079529, + "learning_rate": 0.00019282467601575528, + "loss": 0.0843, + "step": 5184 + }, + { + "epoch": 0.12154456536390115, + "grad_norm": 0.6822691559791565, + "learning_rate": 0.00019282193581111407, + "loss": 0.1782, + "step": 5185 + }, + { + "epoch": 0.12156800693870615, + "grad_norm": 0.6325902938842773, + "learning_rate": 0.00019281919510281842, + "loss": 0.3519, + "step": 5186 + }, + { + "epoch": 0.12159144851351114, + "grad_norm": 0.7409558296203613, + "learning_rate": 0.00019281645389088324, + "loss": 0.1735, + "step": 5187 + }, + { + "epoch": 0.12161489008831614, + "grad_norm": 0.4337514340877533, + "learning_rate": 0.0001928137121753234, + "loss": 0.0904, + "step": 5188 + }, + { + "epoch": 0.12163833166312113, + "grad_norm": 0.6040026545524597, + "learning_rate": 0.0001928109699561537, + "loss": 0.1363, + "step": 5189 + }, + { + "epoch": 0.12166177323792612, + "grad_norm": 0.5851377248764038, + "learning_rate": 0.00019280822723338905, + "loss": 0.1866, + "step": 5190 + }, + { + "epoch": 0.12168521481273112, + "grad_norm": 0.8689699769020081, + "learning_rate": 0.00019280548400704437, + "loss": 0.151, + "step": 5191 + }, + { + "epoch": 0.12170865638753611, + "grad_norm": 0.2961560785770416, + "learning_rate": 0.00019280274027713452, + "loss": 0.0558, + "step": 5192 + }, + { + "epoch": 0.12173209796234111, + "grad_norm": 0.7515779137611389, + "learning_rate": 0.00019279999604367441, + "loss": 0.202, + "step": 5193 + }, + { + "epoch": 0.1217555395371461, + "grad_norm": 0.10345352441072464, + "learning_rate": 0.0001927972513066789, + "loss": 0.0283, + "step": 5194 + }, + { + "epoch": 0.1217789811119511, + "grad_norm": 0.5062533617019653, + "learning_rate": 0.0001927945060661629, + "loss": 0.1761, + "step": 5195 + }, + { + "epoch": 0.12180242268675609, + "grad_norm": 0.47883397340774536, + "learning_rate": 0.00019279176032214128, + "loss": 0.0838, + "step": 5196 + }, + { + "epoch": 0.1218258642615611, + "grad_norm": 0.8406347036361694, + "learning_rate": 0.00019278901407462898, + "loss": 0.3348, + "step": 5197 + }, + { + "epoch": 0.12184930583636609, + "grad_norm": 0.42309096455574036, + "learning_rate": 0.00019278626732364086, + "loss": 0.6207, + "step": 5198 + }, + { + "epoch": 0.12187274741117109, + "grad_norm": 0.5029279589653015, + "learning_rate": 0.00019278352006919187, + "loss": 0.1257, + "step": 5199 + }, + { + "epoch": 0.12189618898597608, + "grad_norm": 0.4653453230857849, + "learning_rate": 0.0001927807723112969, + "loss": 0.0941, + "step": 5200 + }, + { + "epoch": 0.12191963056078108, + "grad_norm": 0.7344146370887756, + "learning_rate": 0.00019277802404997084, + "loss": 0.2156, + "step": 5201 + }, + { + "epoch": 0.12194307213558607, + "grad_norm": 0.2928001284599304, + "learning_rate": 0.0001927752752852286, + "loss": 0.0585, + "step": 5202 + }, + { + "epoch": 0.12196651371039106, + "grad_norm": 0.4613032042980194, + "learning_rate": 0.0001927725260170851, + "loss": 0.1583, + "step": 5203 + }, + { + "epoch": 0.12198995528519606, + "grad_norm": 0.5803639888763428, + "learning_rate": 0.00019276977624555527, + "loss": 0.172, + "step": 5204 + }, + { + "epoch": 0.12201339686000105, + "grad_norm": 0.4246852397918701, + "learning_rate": 0.00019276702597065403, + "loss": 0.1177, + "step": 5205 + }, + { + "epoch": 0.12203683843480605, + "grad_norm": 0.6743556261062622, + "learning_rate": 0.00019276427519239632, + "loss": 0.135, + "step": 5206 + }, + { + "epoch": 0.12206028000961104, + "grad_norm": 0.20646719634532928, + "learning_rate": 0.00019276152391079702, + "loss": 0.0487, + "step": 5207 + }, + { + "epoch": 0.12208372158441604, + "grad_norm": 0.47490260004997253, + "learning_rate": 0.0001927587721258711, + "loss": 0.4826, + "step": 5208 + }, + { + "epoch": 0.12210716315922103, + "grad_norm": 0.5718293190002441, + "learning_rate": 0.00019275601983763347, + "loss": 0.2402, + "step": 5209 + }, + { + "epoch": 0.12213060473402604, + "grad_norm": 0.3222339451313019, + "learning_rate": 0.00019275326704609906, + "loss": 0.0275, + "step": 5210 + }, + { + "epoch": 0.12215404630883103, + "grad_norm": 0.8310773372650146, + "learning_rate": 0.00019275051375128283, + "loss": 0.1488, + "step": 5211 + }, + { + "epoch": 0.12217748788363603, + "grad_norm": 0.602134108543396, + "learning_rate": 0.00019274775995319966, + "loss": 0.0799, + "step": 5212 + }, + { + "epoch": 0.12220092945844102, + "grad_norm": 0.41003090143203735, + "learning_rate": 0.00019274500565186458, + "loss": 0.0819, + "step": 5213 + }, + { + "epoch": 0.12222437103324602, + "grad_norm": 0.4877302944660187, + "learning_rate": 0.0001927422508472925, + "loss": 0.6889, + "step": 5214 + }, + { + "epoch": 0.12224781260805101, + "grad_norm": 0.5526537299156189, + "learning_rate": 0.00019273949553949834, + "loss": 0.0575, + "step": 5215 + }, + { + "epoch": 0.122271254182856, + "grad_norm": 0.8397637009620667, + "learning_rate": 0.00019273673972849706, + "loss": 0.8015, + "step": 5216 + }, + { + "epoch": 0.122294695757661, + "grad_norm": 0.8178645968437195, + "learning_rate": 0.00019273398341430363, + "loss": 0.2127, + "step": 5217 + }, + { + "epoch": 0.12231813733246599, + "grad_norm": 0.20699205994606018, + "learning_rate": 0.00019273122659693303, + "loss": 0.0424, + "step": 5218 + }, + { + "epoch": 0.12234157890727099, + "grad_norm": 0.7112703919410706, + "learning_rate": 0.00019272846927640013, + "loss": 0.1564, + "step": 5219 + }, + { + "epoch": 0.12236502048207598, + "grad_norm": 0.7743474245071411, + "learning_rate": 0.00019272571145272, + "loss": 0.1723, + "step": 5220 + }, + { + "epoch": 0.12238846205688098, + "grad_norm": 0.5157688856124878, + "learning_rate": 0.0001927229531259075, + "loss": 0.1328, + "step": 5221 + }, + { + "epoch": 0.12241190363168597, + "grad_norm": 0.6226338148117065, + "learning_rate": 0.00019272019429597767, + "loss": 0.1771, + "step": 5222 + }, + { + "epoch": 0.12243534520649098, + "grad_norm": 0.6069114804267883, + "learning_rate": 0.00019271743496294548, + "loss": 0.1858, + "step": 5223 + }, + { + "epoch": 0.12245878678129597, + "grad_norm": 0.4121764302253723, + "learning_rate": 0.0001927146751268259, + "loss": 0.1253, + "step": 5224 + }, + { + "epoch": 0.12248222835610097, + "grad_norm": 0.5404337644577026, + "learning_rate": 0.0001927119147876338, + "loss": 0.1842, + "step": 5225 + }, + { + "epoch": 0.12250566993090596, + "grad_norm": 0.6094802618026733, + "learning_rate": 0.00019270915394538433, + "loss": 0.1152, + "step": 5226 + }, + { + "epoch": 0.12252911150571096, + "grad_norm": 0.3992633521556854, + "learning_rate": 0.00019270639260009236, + "loss": 0.1544, + "step": 5227 + }, + { + "epoch": 0.12255255308051595, + "grad_norm": 0.09415066242218018, + "learning_rate": 0.0001927036307517729, + "loss": 0.0176, + "step": 5228 + }, + { + "epoch": 0.12257599465532094, + "grad_norm": 0.5501013994216919, + "learning_rate": 0.0001927008684004409, + "loss": 0.1674, + "step": 5229 + }, + { + "epoch": 0.12259943623012594, + "grad_norm": 0.377213716506958, + "learning_rate": 0.00019269810554611142, + "loss": 0.1241, + "step": 5230 + }, + { + "epoch": 0.12262287780493093, + "grad_norm": 0.34779059886932373, + "learning_rate": 0.00019269534218879937, + "loss": 0.0647, + "step": 5231 + }, + { + "epoch": 0.12264631937973593, + "grad_norm": 0.6413613557815552, + "learning_rate": 0.00019269257832851984, + "loss": 0.2549, + "step": 5232 + }, + { + "epoch": 0.12266976095454092, + "grad_norm": 0.5577290058135986, + "learning_rate": 0.00019268981396528773, + "loss": 0.1247, + "step": 5233 + }, + { + "epoch": 0.12269320252934592, + "grad_norm": 0.6531746983528137, + "learning_rate": 0.0001926870490991181, + "loss": 0.1734, + "step": 5234 + }, + { + "epoch": 0.12271664410415092, + "grad_norm": 0.4452236592769623, + "learning_rate": 0.00019268428373002595, + "loss": 0.0993, + "step": 5235 + }, + { + "epoch": 0.12274008567895592, + "grad_norm": 0.5616729855537415, + "learning_rate": 0.00019268151785802623, + "loss": 0.1499, + "step": 5236 + }, + { + "epoch": 0.12276352725376091, + "grad_norm": 0.6243994235992432, + "learning_rate": 0.00019267875148313402, + "loss": 0.1399, + "step": 5237 + }, + { + "epoch": 0.1227869688285659, + "grad_norm": 0.5515998005867004, + "learning_rate": 0.0001926759846053643, + "loss": 0.8174, + "step": 5238 + }, + { + "epoch": 0.1228104104033709, + "grad_norm": 0.48306038975715637, + "learning_rate": 0.00019267321722473207, + "loss": 0.1274, + "step": 5239 + }, + { + "epoch": 0.1228338519781759, + "grad_norm": 0.6293735504150391, + "learning_rate": 0.00019267044934125237, + "loss": 0.2222, + "step": 5240 + }, + { + "epoch": 0.12285729355298089, + "grad_norm": 0.5771521329879761, + "learning_rate": 0.00019266768095494018, + "loss": 0.1231, + "step": 5241 + }, + { + "epoch": 0.12288073512778588, + "grad_norm": 0.14351458847522736, + "learning_rate": 0.00019266491206581057, + "loss": 0.0215, + "step": 5242 + }, + { + "epoch": 0.12290417670259088, + "grad_norm": 0.43109366297721863, + "learning_rate": 0.00019266214267387854, + "loss": 0.0769, + "step": 5243 + }, + { + "epoch": 0.12292761827739587, + "grad_norm": 0.47766631841659546, + "learning_rate": 0.0001926593727791591, + "loss": 0.152, + "step": 5244 + }, + { + "epoch": 0.12295105985220087, + "grad_norm": 0.37792888283729553, + "learning_rate": 0.00019265660238166734, + "loss": 0.0943, + "step": 5245 + }, + { + "epoch": 0.12297450142700586, + "grad_norm": 0.21353314816951752, + "learning_rate": 0.00019265383148141824, + "loss": 0.0442, + "step": 5246 + }, + { + "epoch": 0.12299794300181086, + "grad_norm": 0.30320584774017334, + "learning_rate": 0.0001926510600784268, + "loss": 0.0647, + "step": 5247 + }, + { + "epoch": 0.12302138457661586, + "grad_norm": 0.5180674195289612, + "learning_rate": 0.00019264828817270812, + "loss": 0.1364, + "step": 5248 + }, + { + "epoch": 0.12304482615142086, + "grad_norm": 0.7033361196517944, + "learning_rate": 0.00019264551576427726, + "loss": 0.1266, + "step": 5249 + }, + { + "epoch": 0.12306826772622585, + "grad_norm": 0.8656555414199829, + "learning_rate": 0.0001926427428531492, + "loss": 0.1446, + "step": 5250 + }, + { + "epoch": 0.12309170930103085, + "grad_norm": 0.23671473562717438, + "learning_rate": 0.000192639969439339, + "loss": 0.0751, + "step": 5251 + }, + { + "epoch": 0.12311515087583584, + "grad_norm": 0.7141321301460266, + "learning_rate": 0.00019263719552286171, + "loss": 0.1381, + "step": 5252 + }, + { + "epoch": 0.12313859245064084, + "grad_norm": 0.4670754373073578, + "learning_rate": 0.00019263442110373244, + "loss": 0.065, + "step": 5253 + }, + { + "epoch": 0.12316203402544583, + "grad_norm": 0.6721237301826477, + "learning_rate": 0.0001926316461819662, + "loss": 0.2439, + "step": 5254 + }, + { + "epoch": 0.12318547560025082, + "grad_norm": 0.38437843322753906, + "learning_rate": 0.000192628870757578, + "loss": 0.0943, + "step": 5255 + }, + { + "epoch": 0.12320891717505582, + "grad_norm": 0.5646991729736328, + "learning_rate": 0.00019262609483058297, + "loss": 0.1005, + "step": 5256 + }, + { + "epoch": 0.12323235874986081, + "grad_norm": 0.3010539412498474, + "learning_rate": 0.0001926233184009961, + "loss": 0.0654, + "step": 5257 + }, + { + "epoch": 0.1232558003246658, + "grad_norm": 0.6738088130950928, + "learning_rate": 0.00019262054146883256, + "loss": 0.0986, + "step": 5258 + }, + { + "epoch": 0.1232792418994708, + "grad_norm": 0.800136923789978, + "learning_rate": 0.00019261776403410732, + "loss": 0.3375, + "step": 5259 + }, + { + "epoch": 0.1233026834742758, + "grad_norm": 0.3000994622707367, + "learning_rate": 0.0001926149860968355, + "loss": 0.0875, + "step": 5260 + }, + { + "epoch": 0.1233261250490808, + "grad_norm": 0.2091541737318039, + "learning_rate": 0.00019261220765703216, + "loss": 0.035, + "step": 5261 + }, + { + "epoch": 0.1233495666238858, + "grad_norm": 0.42130523920059204, + "learning_rate": 0.00019260942871471238, + "loss": 0.081, + "step": 5262 + }, + { + "epoch": 0.12337300819869079, + "grad_norm": 0.33378785848617554, + "learning_rate": 0.00019260664926989125, + "loss": 0.0487, + "step": 5263 + }, + { + "epoch": 0.12339644977349579, + "grad_norm": 0.3551618158817291, + "learning_rate": 0.00019260386932258382, + "loss": 0.1307, + "step": 5264 + }, + { + "epoch": 0.12341989134830078, + "grad_norm": 0.5208948254585266, + "learning_rate": 0.0001926010888728052, + "loss": 0.1225, + "step": 5265 + }, + { + "epoch": 0.12344333292310578, + "grad_norm": 0.6218944787979126, + "learning_rate": 0.00019259830792057046, + "loss": 0.0748, + "step": 5266 + }, + { + "epoch": 0.12346677449791077, + "grad_norm": 0.7352637648582458, + "learning_rate": 0.0001925955264658947, + "loss": 0.2237, + "step": 5267 + }, + { + "epoch": 0.12349021607271576, + "grad_norm": 0.37346336245536804, + "learning_rate": 0.000192592744508793, + "loss": 0.0532, + "step": 5268 + }, + { + "epoch": 0.12351365764752076, + "grad_norm": 1.062689185142517, + "learning_rate": 0.00019258996204928047, + "loss": 0.1242, + "step": 5269 + }, + { + "epoch": 0.12353709922232575, + "grad_norm": 0.5303933620452881, + "learning_rate": 0.00019258717908737221, + "loss": 0.1662, + "step": 5270 + }, + { + "epoch": 0.12356054079713075, + "grad_norm": 0.3007787764072418, + "learning_rate": 0.00019258439562308332, + "loss": 0.0552, + "step": 5271 + }, + { + "epoch": 0.12358398237193574, + "grad_norm": 1.3681042194366455, + "learning_rate": 0.00019258161165642887, + "loss": 0.2816, + "step": 5272 + }, + { + "epoch": 0.12360742394674074, + "grad_norm": 0.2796531021595001, + "learning_rate": 0.00019257882718742402, + "loss": 0.0708, + "step": 5273 + }, + { + "epoch": 0.12363086552154574, + "grad_norm": 0.26708391308784485, + "learning_rate": 0.00019257604221608384, + "loss": 0.0953, + "step": 5274 + }, + { + "epoch": 0.12365430709635074, + "grad_norm": 0.577034592628479, + "learning_rate": 0.00019257325674242344, + "loss": 0.0962, + "step": 5275 + }, + { + "epoch": 0.12367774867115573, + "grad_norm": 0.472379595041275, + "learning_rate": 0.00019257047076645794, + "loss": 0.0973, + "step": 5276 + }, + { + "epoch": 0.12370119024596073, + "grad_norm": 0.6107684373855591, + "learning_rate": 0.00019256768428820248, + "loss": 0.0767, + "step": 5277 + }, + { + "epoch": 0.12372463182076572, + "grad_norm": 0.8256498575210571, + "learning_rate": 0.00019256489730767217, + "loss": 0.1794, + "step": 5278 + }, + { + "epoch": 0.12374807339557072, + "grad_norm": 0.7719846367835999, + "learning_rate": 0.00019256210982488212, + "loss": 0.1696, + "step": 5279 + }, + { + "epoch": 0.12377151497037571, + "grad_norm": 0.40988215804100037, + "learning_rate": 0.00019255932183984747, + "loss": 0.5091, + "step": 5280 + }, + { + "epoch": 0.1237949565451807, + "grad_norm": 0.6060330867767334, + "learning_rate": 0.0001925565333525833, + "loss": 0.1155, + "step": 5281 + }, + { + "epoch": 0.1238183981199857, + "grad_norm": 0.5218788981437683, + "learning_rate": 0.0001925537443631048, + "loss": 0.1331, + "step": 5282 + }, + { + "epoch": 0.12384183969479069, + "grad_norm": 0.3589623272418976, + "learning_rate": 0.0001925509548714271, + "loss": 0.0864, + "step": 5283 + }, + { + "epoch": 0.12386528126959569, + "grad_norm": 0.6469882130622864, + "learning_rate": 0.00019254816487756528, + "loss": 0.1817, + "step": 5284 + }, + { + "epoch": 0.12388872284440068, + "grad_norm": 0.6686276793479919, + "learning_rate": 0.00019254537438153456, + "loss": 0.1655, + "step": 5285 + }, + { + "epoch": 0.12391216441920568, + "grad_norm": 0.5209758281707764, + "learning_rate": 0.00019254258338334998, + "loss": 0.1275, + "step": 5286 + }, + { + "epoch": 0.12393560599401068, + "grad_norm": 0.6837011575698853, + "learning_rate": 0.00019253979188302677, + "loss": 0.2761, + "step": 5287 + }, + { + "epoch": 0.12395904756881568, + "grad_norm": 0.6197140216827393, + "learning_rate": 0.00019253699988058004, + "loss": 0.0923, + "step": 5288 + }, + { + "epoch": 0.12398248914362067, + "grad_norm": 0.4683173894882202, + "learning_rate": 0.00019253420737602497, + "loss": 0.09, + "step": 5289 + }, + { + "epoch": 0.12400593071842567, + "grad_norm": 0.44493821263313293, + "learning_rate": 0.00019253141436937666, + "loss": 0.1135, + "step": 5290 + }, + { + "epoch": 0.12402937229323066, + "grad_norm": 0.8701930046081543, + "learning_rate": 0.00019252862086065032, + "loss": 0.2632, + "step": 5291 + }, + { + "epoch": 0.12405281386803566, + "grad_norm": 0.536953330039978, + "learning_rate": 0.00019252582684986104, + "loss": 0.1636, + "step": 5292 + }, + { + "epoch": 0.12407625544284065, + "grad_norm": 0.5631737112998962, + "learning_rate": 0.00019252303233702406, + "loss": 0.1141, + "step": 5293 + }, + { + "epoch": 0.12409969701764564, + "grad_norm": 0.32279619574546814, + "learning_rate": 0.0001925202373221545, + "loss": 0.0484, + "step": 5294 + }, + { + "epoch": 0.12412313859245064, + "grad_norm": 0.3007888197898865, + "learning_rate": 0.00019251744180526754, + "loss": 0.0552, + "step": 5295 + }, + { + "epoch": 0.12414658016725563, + "grad_norm": 0.37057238817214966, + "learning_rate": 0.00019251464578637833, + "loss": 0.0865, + "step": 5296 + }, + { + "epoch": 0.12417002174206063, + "grad_norm": 0.6346404552459717, + "learning_rate": 0.00019251184926550203, + "loss": 0.131, + "step": 5297 + }, + { + "epoch": 0.12419346331686562, + "grad_norm": 0.37954774498939514, + "learning_rate": 0.00019250905224265386, + "loss": 0.0865, + "step": 5298 + }, + { + "epoch": 0.12421690489167063, + "grad_norm": 0.583644449710846, + "learning_rate": 0.00019250625471784895, + "loss": 0.0878, + "step": 5299 + }, + { + "epoch": 0.12424034646647562, + "grad_norm": 0.2047155499458313, + "learning_rate": 0.0001925034566911025, + "loss": 0.0566, + "step": 5300 + }, + { + "epoch": 0.12426378804128062, + "grad_norm": 0.9874632358551025, + "learning_rate": 0.00019250065816242974, + "loss": 0.1158, + "step": 5301 + }, + { + "epoch": 0.12428722961608561, + "grad_norm": 0.3574824035167694, + "learning_rate": 0.00019249785913184574, + "loss": 0.1057, + "step": 5302 + }, + { + "epoch": 0.1243106711908906, + "grad_norm": 0.538435161113739, + "learning_rate": 0.0001924950595993658, + "loss": 0.6374, + "step": 5303 + }, + { + "epoch": 0.1243341127656956, + "grad_norm": 0.8848696947097778, + "learning_rate": 0.00019249225956500504, + "loss": 0.1455, + "step": 5304 + }, + { + "epoch": 0.1243575543405006, + "grad_norm": 0.5053797364234924, + "learning_rate": 0.0001924894590287787, + "loss": 0.1177, + "step": 5305 + }, + { + "epoch": 0.12438099591530559, + "grad_norm": 0.6062100529670715, + "learning_rate": 0.00019248665799070195, + "loss": 0.1339, + "step": 5306 + }, + { + "epoch": 0.12440443749011058, + "grad_norm": 0.718043327331543, + "learning_rate": 0.00019248385645079, + "loss": 0.2121, + "step": 5307 + }, + { + "epoch": 0.12442787906491558, + "grad_norm": 1.3256498575210571, + "learning_rate": 0.00019248105440905803, + "loss": 0.069, + "step": 5308 + }, + { + "epoch": 0.12445132063972057, + "grad_norm": 0.512968122959137, + "learning_rate": 0.00019247825186552127, + "loss": 0.1263, + "step": 5309 + }, + { + "epoch": 0.12447476221452557, + "grad_norm": 0.3208520710468292, + "learning_rate": 0.0001924754488201949, + "loss": 0.0925, + "step": 5310 + }, + { + "epoch": 0.12449820378933056, + "grad_norm": 0.4141823649406433, + "learning_rate": 0.00019247264527309416, + "loss": 0.1147, + "step": 5311 + }, + { + "epoch": 0.12452164536413557, + "grad_norm": 0.408816397190094, + "learning_rate": 0.00019246984122423425, + "loss": 0.1299, + "step": 5312 + }, + { + "epoch": 0.12454508693894056, + "grad_norm": 0.6680583953857422, + "learning_rate": 0.00019246703667363035, + "loss": 0.1533, + "step": 5313 + }, + { + "epoch": 0.12456852851374556, + "grad_norm": 0.1978268176317215, + "learning_rate": 0.00019246423162129776, + "loss": 0.0573, + "step": 5314 + }, + { + "epoch": 0.12459197008855055, + "grad_norm": 0.5020516514778137, + "learning_rate": 0.00019246142606725163, + "loss": 0.1749, + "step": 5315 + }, + { + "epoch": 0.12461541166335555, + "grad_norm": 0.7985655069351196, + "learning_rate": 0.00019245862001150718, + "loss": 0.239, + "step": 5316 + }, + { + "epoch": 0.12463885323816054, + "grad_norm": 0.4740305542945862, + "learning_rate": 0.00019245581345407966, + "loss": 0.0848, + "step": 5317 + }, + { + "epoch": 0.12466229481296554, + "grad_norm": 0.6231048703193665, + "learning_rate": 0.00019245300639498432, + "loss": 0.112, + "step": 5318 + }, + { + "epoch": 0.12468573638777053, + "grad_norm": 0.6728546619415283, + "learning_rate": 0.00019245019883423634, + "loss": 0.1591, + "step": 5319 + }, + { + "epoch": 0.12470917796257552, + "grad_norm": 0.5483601689338684, + "learning_rate": 0.00019244739077185102, + "loss": 0.162, + "step": 5320 + }, + { + "epoch": 0.12473261953738052, + "grad_norm": 0.524666428565979, + "learning_rate": 0.00019244458220784352, + "loss": 0.4759, + "step": 5321 + }, + { + "epoch": 0.12475606111218551, + "grad_norm": 0.7948358654975891, + "learning_rate": 0.00019244177314222912, + "loss": 0.1463, + "step": 5322 + }, + { + "epoch": 0.1247795026869905, + "grad_norm": 0.19416488707065582, + "learning_rate": 0.0001924389635750231, + "loss": 0.0647, + "step": 5323 + }, + { + "epoch": 0.1248029442617955, + "grad_norm": 0.25975003838539124, + "learning_rate": 0.00019243615350624064, + "loss": 0.0692, + "step": 5324 + }, + { + "epoch": 0.12482638583660051, + "grad_norm": 0.10892990231513977, + "learning_rate": 0.00019243334293589698, + "loss": 0.0145, + "step": 5325 + }, + { + "epoch": 0.1248498274114055, + "grad_norm": 0.39105692505836487, + "learning_rate": 0.00019243053186400742, + "loss": 0.0829, + "step": 5326 + }, + { + "epoch": 0.1248732689862105, + "grad_norm": 0.7967643737792969, + "learning_rate": 0.00019242772029058722, + "loss": 0.1409, + "step": 5327 + }, + { + "epoch": 0.12489671056101549, + "grad_norm": 0.8023490309715271, + "learning_rate": 0.0001924249082156516, + "loss": 0.2488, + "step": 5328 + }, + { + "epoch": 0.12492015213582049, + "grad_norm": 0.7603198885917664, + "learning_rate": 0.00019242209563921585, + "loss": 0.1125, + "step": 5329 + }, + { + "epoch": 0.12494359371062548, + "grad_norm": 0.7050050497055054, + "learning_rate": 0.00019241928256129515, + "loss": 0.1823, + "step": 5330 + }, + { + "epoch": 0.12496703528543048, + "grad_norm": 0.39531272649765015, + "learning_rate": 0.00019241646898190487, + "loss": 0.0884, + "step": 5331 + }, + { + "epoch": 0.12499047686023547, + "grad_norm": 0.3083620071411133, + "learning_rate": 0.00019241365490106022, + "loss": 0.1303, + "step": 5332 + }, + { + "epoch": 0.12501391843504048, + "grad_norm": 0.18770688772201538, + "learning_rate": 0.00019241084031877649, + "loss": 0.041, + "step": 5333 + }, + { + "epoch": 0.12503736000984547, + "grad_norm": 0.24640047550201416, + "learning_rate": 0.00019240802523506892, + "loss": 0.0415, + "step": 5334 + }, + { + "epoch": 0.12506080158465047, + "grad_norm": 0.4373776316642761, + "learning_rate": 0.00019240520964995285, + "loss": 0.1259, + "step": 5335 + }, + { + "epoch": 0.12508424315945546, + "grad_norm": 0.7047239542007446, + "learning_rate": 0.00019240239356344346, + "loss": 0.1376, + "step": 5336 + }, + { + "epoch": 0.12510768473426045, + "grad_norm": 0.4816850423812866, + "learning_rate": 0.00019239957697555612, + "loss": 0.0745, + "step": 5337 + }, + { + "epoch": 0.12513112630906545, + "grad_norm": 0.5414191484451294, + "learning_rate": 0.00019239675988630607, + "loss": 0.1501, + "step": 5338 + }, + { + "epoch": 0.12515456788387044, + "grad_norm": 0.36607202887535095, + "learning_rate": 0.00019239394229570862, + "loss": 0.1024, + "step": 5339 + }, + { + "epoch": 0.12517800945867544, + "grad_norm": 0.7094634771347046, + "learning_rate": 0.000192391124203779, + "loss": 0.1312, + "step": 5340 + }, + { + "epoch": 0.12520145103348043, + "grad_norm": 0.21229097247123718, + "learning_rate": 0.00019238830561053258, + "loss": 0.0394, + "step": 5341 + }, + { + "epoch": 0.12522489260828543, + "grad_norm": 0.26475653052330017, + "learning_rate": 0.0001923854865159846, + "loss": 0.0736, + "step": 5342 + }, + { + "epoch": 0.12524833418309042, + "grad_norm": 0.577642023563385, + "learning_rate": 0.00019238266692015037, + "loss": 0.1175, + "step": 5343 + }, + { + "epoch": 0.12527177575789541, + "grad_norm": 0.8807522058486938, + "learning_rate": 0.0001923798468230452, + "loss": 0.2114, + "step": 5344 + }, + { + "epoch": 0.1252952173327004, + "grad_norm": 0.6641179919242859, + "learning_rate": 0.00019237702622468438, + "loss": 0.0462, + "step": 5345 + }, + { + "epoch": 0.1253186589075054, + "grad_norm": 0.7598125338554382, + "learning_rate": 0.0001923742051250832, + "loss": 0.1843, + "step": 5346 + }, + { + "epoch": 0.1253421004823104, + "grad_norm": 0.5924615859985352, + "learning_rate": 0.00019237138352425702, + "loss": 0.1118, + "step": 5347 + }, + { + "epoch": 0.1253655420571154, + "grad_norm": 0.20410674810409546, + "learning_rate": 0.00019236856142222108, + "loss": 0.054, + "step": 5348 + }, + { + "epoch": 0.1253889836319204, + "grad_norm": 0.5882973670959473, + "learning_rate": 0.00019236573881899075, + "loss": 0.6583, + "step": 5349 + }, + { + "epoch": 0.12541242520672538, + "grad_norm": 0.5627157092094421, + "learning_rate": 0.0001923629157145813, + "loss": 0.1154, + "step": 5350 + }, + { + "epoch": 0.12543586678153038, + "grad_norm": 0.2724912464618683, + "learning_rate": 0.00019236009210900812, + "loss": 0.1079, + "step": 5351 + }, + { + "epoch": 0.12545930835633537, + "grad_norm": 0.7831101417541504, + "learning_rate": 0.00019235726800228645, + "loss": 0.1764, + "step": 5352 + }, + { + "epoch": 0.12548274993114036, + "grad_norm": 0.1979042887687683, + "learning_rate": 0.00019235444339443163, + "loss": 0.0708, + "step": 5353 + }, + { + "epoch": 0.12550619150594536, + "grad_norm": 0.100710429251194, + "learning_rate": 0.00019235161828545905, + "loss": 0.024, + "step": 5354 + }, + { + "epoch": 0.12552963308075035, + "grad_norm": 0.38550713658332825, + "learning_rate": 0.00019234879267538395, + "loss": 0.0778, + "step": 5355 + }, + { + "epoch": 0.12555307465555535, + "grad_norm": 0.15829601883888245, + "learning_rate": 0.00019234596656422172, + "loss": 0.0279, + "step": 5356 + }, + { + "epoch": 0.12557651623036037, + "grad_norm": 0.45256227254867554, + "learning_rate": 0.00019234313995198767, + "loss": 0.0953, + "step": 5357 + }, + { + "epoch": 0.12559995780516536, + "grad_norm": 1.1176419258117676, + "learning_rate": 0.00019234031283869719, + "loss": 0.2248, + "step": 5358 + }, + { + "epoch": 0.12562339937997036, + "grad_norm": 0.7502868175506592, + "learning_rate": 0.00019233748522436551, + "loss": 0.2351, + "step": 5359 + }, + { + "epoch": 0.12564684095477535, + "grad_norm": 0.1883973479270935, + "learning_rate": 0.00019233465710900808, + "loss": 0.031, + "step": 5360 + }, + { + "epoch": 0.12567028252958035, + "grad_norm": 0.43529120087623596, + "learning_rate": 0.00019233182849264023, + "loss": 0.0998, + "step": 5361 + }, + { + "epoch": 0.12569372410438534, + "grad_norm": 0.6419172883033752, + "learning_rate": 0.00019232899937527721, + "loss": 0.1871, + "step": 5362 + }, + { + "epoch": 0.12571716567919033, + "grad_norm": 0.24008843302726746, + "learning_rate": 0.0001923261697569345, + "loss": 0.0328, + "step": 5363 + }, + { + "epoch": 0.12574060725399533, + "grad_norm": 0.7823966145515442, + "learning_rate": 0.0001923233396376274, + "loss": 0.2587, + "step": 5364 + }, + { + "epoch": 0.12576404882880032, + "grad_norm": 0.4590373635292053, + "learning_rate": 0.00019232050901737122, + "loss": 0.0741, + "step": 5365 + }, + { + "epoch": 0.12578749040360532, + "grad_norm": 0.20165276527404785, + "learning_rate": 0.0001923176778961814, + "loss": 0.0428, + "step": 5366 + }, + { + "epoch": 0.1258109319784103, + "grad_norm": 0.1393086016178131, + "learning_rate": 0.00019231484627407324, + "loss": 0.036, + "step": 5367 + }, + { + "epoch": 0.1258343735532153, + "grad_norm": 0.4024730324745178, + "learning_rate": 0.00019231201415106212, + "loss": 0.8191, + "step": 5368 + }, + { + "epoch": 0.1258578151280203, + "grad_norm": 0.43099287152290344, + "learning_rate": 0.00019230918152716344, + "loss": 0.0744, + "step": 5369 + }, + { + "epoch": 0.1258812567028253, + "grad_norm": 0.522578775882721, + "learning_rate": 0.00019230634840239255, + "loss": 0.1376, + "step": 5370 + }, + { + "epoch": 0.1259046982776303, + "grad_norm": 0.7728557586669922, + "learning_rate": 0.0001923035147767648, + "loss": 0.1105, + "step": 5371 + }, + { + "epoch": 0.12592813985243528, + "grad_norm": 0.7125654220581055, + "learning_rate": 0.00019230068065029558, + "loss": 0.127, + "step": 5372 + }, + { + "epoch": 0.12595158142724028, + "grad_norm": 0.6867541074752808, + "learning_rate": 0.0001922978460230003, + "loss": 0.5702, + "step": 5373 + }, + { + "epoch": 0.12597502300204527, + "grad_norm": 0.5944532155990601, + "learning_rate": 0.0001922950108948943, + "loss": 0.1758, + "step": 5374 + }, + { + "epoch": 0.12599846457685027, + "grad_norm": 0.6252401471138, + "learning_rate": 0.00019229217526599296, + "loss": 0.126, + "step": 5375 + }, + { + "epoch": 0.12602190615165526, + "grad_norm": 0.27609360218048096, + "learning_rate": 0.00019228933913631167, + "loss": 0.0577, + "step": 5376 + }, + { + "epoch": 0.12604534772646026, + "grad_norm": 0.36872830986976624, + "learning_rate": 0.00019228650250586587, + "loss": 0.0902, + "step": 5377 + }, + { + "epoch": 0.12606878930126525, + "grad_norm": 0.354900985956192, + "learning_rate": 0.0001922836653746709, + "loss": 0.0667, + "step": 5378 + }, + { + "epoch": 0.12609223087607024, + "grad_norm": 0.7933406233787537, + "learning_rate": 0.00019228082774274216, + "loss": 0.1621, + "step": 5379 + }, + { + "epoch": 0.12611567245087524, + "grad_norm": 0.7732988595962524, + "learning_rate": 0.00019227798961009507, + "loss": 0.1865, + "step": 5380 + }, + { + "epoch": 0.12613911402568023, + "grad_norm": 0.7383318543434143, + "learning_rate": 0.00019227515097674497, + "loss": 0.0787, + "step": 5381 + }, + { + "epoch": 0.12616255560048525, + "grad_norm": 0.6446529030799866, + "learning_rate": 0.00019227231184270734, + "loss": 0.1855, + "step": 5382 + }, + { + "epoch": 0.12618599717529025, + "grad_norm": 0.17542266845703125, + "learning_rate": 0.00019226947220799754, + "loss": 0.042, + "step": 5383 + }, + { + "epoch": 0.12620943875009524, + "grad_norm": 0.6874074935913086, + "learning_rate": 0.00019226663207263102, + "loss": 0.1518, + "step": 5384 + }, + { + "epoch": 0.12623288032490024, + "grad_norm": 0.33352670073509216, + "learning_rate": 0.00019226379143662313, + "loss": 0.1098, + "step": 5385 + }, + { + "epoch": 0.12625632189970523, + "grad_norm": 0.2948526442050934, + "learning_rate": 0.0001922609502999893, + "loss": 0.0399, + "step": 5386 + }, + { + "epoch": 0.12627976347451023, + "grad_norm": 0.6431112289428711, + "learning_rate": 0.00019225810866274498, + "loss": 0.1513, + "step": 5387 + }, + { + "epoch": 0.12630320504931522, + "grad_norm": 0.9546235799789429, + "learning_rate": 0.0001922552665249056, + "loss": 0.249, + "step": 5388 + }, + { + "epoch": 0.12632664662412021, + "grad_norm": 0.1486416459083557, + "learning_rate": 0.0001922524238864865, + "loss": 0.0287, + "step": 5389 + }, + { + "epoch": 0.1263500881989252, + "grad_norm": 0.5273579955101013, + "learning_rate": 0.0001922495807475032, + "loss": 0.1229, + "step": 5390 + }, + { + "epoch": 0.1263735297737302, + "grad_norm": 0.33756017684936523, + "learning_rate": 0.00019224673710797104, + "loss": 0.0626, + "step": 5391 + }, + { + "epoch": 0.1263969713485352, + "grad_norm": 0.35007795691490173, + "learning_rate": 0.0001922438929679055, + "loss": 0.0481, + "step": 5392 + }, + { + "epoch": 0.1264204129233402, + "grad_norm": 0.29406481981277466, + "learning_rate": 0.000192241048327322, + "loss": 0.033, + "step": 5393 + }, + { + "epoch": 0.1264438544981452, + "grad_norm": 0.2216564267873764, + "learning_rate": 0.00019223820318623597, + "loss": 0.047, + "step": 5394 + }, + { + "epoch": 0.12646729607295018, + "grad_norm": 0.2734454274177551, + "learning_rate": 0.0001922353575446629, + "loss": 0.0796, + "step": 5395 + }, + { + "epoch": 0.12649073764775517, + "grad_norm": 0.5556796193122864, + "learning_rate": 0.00019223251140261814, + "loss": 0.1384, + "step": 5396 + }, + { + "epoch": 0.12651417922256017, + "grad_norm": 0.22672779858112335, + "learning_rate": 0.0001922296647601172, + "loss": 0.0404, + "step": 5397 + }, + { + "epoch": 0.12653762079736516, + "grad_norm": 0.580760657787323, + "learning_rate": 0.0001922268176171755, + "loss": 0.0918, + "step": 5398 + }, + { + "epoch": 0.12656106237217016, + "grad_norm": 1.0596144199371338, + "learning_rate": 0.00019222396997380851, + "loss": 0.3006, + "step": 5399 + }, + { + "epoch": 0.12658450394697515, + "grad_norm": 0.2814246118068695, + "learning_rate": 0.00019222112183003165, + "loss": 0.0866, + "step": 5400 + }, + { + "epoch": 0.12660794552178015, + "grad_norm": 0.3816803991794586, + "learning_rate": 0.00019221827318586038, + "loss": 0.0765, + "step": 5401 + }, + { + "epoch": 0.12663138709658514, + "grad_norm": 0.4310418963432312, + "learning_rate": 0.00019221542404131018, + "loss": 0.1028, + "step": 5402 + }, + { + "epoch": 0.12665482867139014, + "grad_norm": 0.29537996649742126, + "learning_rate": 0.0001922125743963965, + "loss": 0.0382, + "step": 5403 + }, + { + "epoch": 0.12667827024619513, + "grad_norm": 0.48627328872680664, + "learning_rate": 0.0001922097242511348, + "loss": 0.0714, + "step": 5404 + }, + { + "epoch": 0.12670171182100012, + "grad_norm": 0.5994488000869751, + "learning_rate": 0.00019220687360554055, + "loss": 0.1719, + "step": 5405 + }, + { + "epoch": 0.12672515339580512, + "grad_norm": 0.6540833711624146, + "learning_rate": 0.00019220402245962917, + "loss": 0.102, + "step": 5406 + }, + { + "epoch": 0.1267485949706101, + "grad_norm": 0.4961029291152954, + "learning_rate": 0.00019220117081341623, + "loss": 0.1126, + "step": 5407 + }, + { + "epoch": 0.12677203654541513, + "grad_norm": 0.6628702282905579, + "learning_rate": 0.0001921983186669171, + "loss": 0.2008, + "step": 5408 + }, + { + "epoch": 0.12679547812022013, + "grad_norm": 0.690980851650238, + "learning_rate": 0.00019219546602014732, + "loss": 0.2033, + "step": 5409 + }, + { + "epoch": 0.12681891969502512, + "grad_norm": 0.4762322008609772, + "learning_rate": 0.00019219261287312236, + "loss": 0.0596, + "step": 5410 + }, + { + "epoch": 0.12684236126983012, + "grad_norm": 0.6854256987571716, + "learning_rate": 0.00019218975922585766, + "loss": 0.2019, + "step": 5411 + }, + { + "epoch": 0.1268658028446351, + "grad_norm": 0.6446226835250854, + "learning_rate": 0.00019218690507836876, + "loss": 0.1458, + "step": 5412 + }, + { + "epoch": 0.1268892444194401, + "grad_norm": 0.4366174638271332, + "learning_rate": 0.00019218405043067112, + "loss": 0.0802, + "step": 5413 + }, + { + "epoch": 0.1269126859942451, + "grad_norm": 0.5668319463729858, + "learning_rate": 0.0001921811952827802, + "loss": 0.2167, + "step": 5414 + }, + { + "epoch": 0.1269361275690501, + "grad_norm": 0.23880372941493988, + "learning_rate": 0.00019217833963471154, + "loss": 0.0343, + "step": 5415 + }, + { + "epoch": 0.1269595691438551, + "grad_norm": 0.488459050655365, + "learning_rate": 0.0001921754834864806, + "loss": 0.0556, + "step": 5416 + }, + { + "epoch": 0.12698301071866008, + "grad_norm": 0.4095788300037384, + "learning_rate": 0.00019217262683810297, + "loss": 0.0742, + "step": 5417 + }, + { + "epoch": 0.12700645229346508, + "grad_norm": 0.33881303668022156, + "learning_rate": 0.00019216976968959402, + "loss": 0.0402, + "step": 5418 + }, + { + "epoch": 0.12702989386827007, + "grad_norm": 0.7162553071975708, + "learning_rate": 0.0001921669120409693, + "loss": 0.1242, + "step": 5419 + }, + { + "epoch": 0.12705333544307507, + "grad_norm": 0.25315389037132263, + "learning_rate": 0.00019216405389224433, + "loss": 0.0404, + "step": 5420 + }, + { + "epoch": 0.12707677701788006, + "grad_norm": 0.6697626113891602, + "learning_rate": 0.0001921611952434346, + "loss": 0.2482, + "step": 5421 + }, + { + "epoch": 0.12710021859268505, + "grad_norm": 0.3340182900428772, + "learning_rate": 0.00019215833609455567, + "loss": 0.0842, + "step": 5422 + }, + { + "epoch": 0.12712366016749005, + "grad_norm": 0.7517386078834534, + "learning_rate": 0.000192155476445623, + "loss": 0.1594, + "step": 5423 + }, + { + "epoch": 0.12714710174229504, + "grad_norm": 1.129943609237671, + "learning_rate": 0.0001921526162966521, + "loss": 0.2274, + "step": 5424 + }, + { + "epoch": 0.12717054331710004, + "grad_norm": 0.40678995847702026, + "learning_rate": 0.00019214975564765853, + "loss": 0.0694, + "step": 5425 + }, + { + "epoch": 0.12719398489190503, + "grad_norm": 0.6579654216766357, + "learning_rate": 0.0001921468944986578, + "loss": 0.2005, + "step": 5426 + }, + { + "epoch": 0.12721742646671003, + "grad_norm": 0.6291545033454895, + "learning_rate": 0.00019214403284966542, + "loss": 0.127, + "step": 5427 + }, + { + "epoch": 0.12724086804151502, + "grad_norm": 0.6209110617637634, + "learning_rate": 0.00019214117070069696, + "loss": 0.1637, + "step": 5428 + }, + { + "epoch": 0.12726430961632001, + "grad_norm": 1.1923601627349854, + "learning_rate": 0.00019213830805176786, + "loss": 0.1867, + "step": 5429 + }, + { + "epoch": 0.127287751191125, + "grad_norm": 0.2982250154018402, + "learning_rate": 0.00019213544490289376, + "loss": 0.1094, + "step": 5430 + }, + { + "epoch": 0.12731119276593, + "grad_norm": 0.6008378863334656, + "learning_rate": 0.00019213258125409013, + "loss": 0.1275, + "step": 5431 + }, + { + "epoch": 0.127334634340735, + "grad_norm": 0.39613714814186096, + "learning_rate": 0.0001921297171053725, + "loss": 0.0481, + "step": 5432 + }, + { + "epoch": 0.12735807591554, + "grad_norm": 0.1533796638250351, + "learning_rate": 0.00019212685245675645, + "loss": 0.0296, + "step": 5433 + }, + { + "epoch": 0.12738151749034501, + "grad_norm": 0.1642487347126007, + "learning_rate": 0.00019212398730825752, + "loss": 0.0416, + "step": 5434 + }, + { + "epoch": 0.12740495906515, + "grad_norm": 0.731430172920227, + "learning_rate": 0.0001921211216598912, + "loss": 0.1286, + "step": 5435 + }, + { + "epoch": 0.127428400639955, + "grad_norm": 0.5707637071609497, + "learning_rate": 0.00019211825551167315, + "loss": 0.1467, + "step": 5436 + }, + { + "epoch": 0.12745184221476, + "grad_norm": 0.7076404690742493, + "learning_rate": 0.00019211538886361882, + "loss": 0.8859, + "step": 5437 + }, + { + "epoch": 0.127475283789565, + "grad_norm": 0.23199507594108582, + "learning_rate": 0.0001921125217157438, + "loss": 0.0637, + "step": 5438 + }, + { + "epoch": 0.12749872536436999, + "grad_norm": 0.4441787004470825, + "learning_rate": 0.00019210965406806367, + "loss": 0.095, + "step": 5439 + }, + { + "epoch": 0.12752216693917498, + "grad_norm": 0.5696524977684021, + "learning_rate": 0.00019210678592059395, + "loss": 0.3006, + "step": 5440 + }, + { + "epoch": 0.12754560851397997, + "grad_norm": 0.212777778506279, + "learning_rate": 0.0001921039172733502, + "loss": 0.0409, + "step": 5441 + }, + { + "epoch": 0.12756905008878497, + "grad_norm": 0.5239803194999695, + "learning_rate": 0.00019210104812634804, + "loss": 0.1447, + "step": 5442 + }, + { + "epoch": 0.12759249166358996, + "grad_norm": 0.17628850042819977, + "learning_rate": 0.00019209817847960298, + "loss": 0.0471, + "step": 5443 + }, + { + "epoch": 0.12761593323839496, + "grad_norm": 0.4248546361923218, + "learning_rate": 0.00019209530833313063, + "loss": 0.1484, + "step": 5444 + }, + { + "epoch": 0.12763937481319995, + "grad_norm": 1.0930883884429932, + "learning_rate": 0.00019209243768694654, + "loss": 0.2295, + "step": 5445 + }, + { + "epoch": 0.12766281638800495, + "grad_norm": 0.5874789357185364, + "learning_rate": 0.00019208956654106628, + "loss": 0.142, + "step": 5446 + }, + { + "epoch": 0.12768625796280994, + "grad_norm": 0.9726516604423523, + "learning_rate": 0.00019208669489550548, + "loss": 0.2127, + "step": 5447 + }, + { + "epoch": 0.12770969953761493, + "grad_norm": 0.6966570615768433, + "learning_rate": 0.00019208382275027965, + "loss": 0.1376, + "step": 5448 + }, + { + "epoch": 0.12773314111241993, + "grad_norm": 1.013843059539795, + "learning_rate": 0.00019208095010540444, + "loss": 0.1744, + "step": 5449 + }, + { + "epoch": 0.12775658268722492, + "grad_norm": 0.3868230879306793, + "learning_rate": 0.00019207807696089538, + "loss": 0.1211, + "step": 5450 + }, + { + "epoch": 0.12778002426202992, + "grad_norm": 0.9655842185020447, + "learning_rate": 0.00019207520331676806, + "loss": 0.2007, + "step": 5451 + }, + { + "epoch": 0.1278034658368349, + "grad_norm": 0.7235747575759888, + "learning_rate": 0.00019207232917303814, + "loss": 0.2104, + "step": 5452 + }, + { + "epoch": 0.1278269074116399, + "grad_norm": 0.4975389540195465, + "learning_rate": 0.00019206945452972116, + "loss": 0.0902, + "step": 5453 + }, + { + "epoch": 0.1278503489864449, + "grad_norm": 0.4622918963432312, + "learning_rate": 0.00019206657938683271, + "loss": 0.1246, + "step": 5454 + }, + { + "epoch": 0.1278737905612499, + "grad_norm": 0.5217340588569641, + "learning_rate": 0.00019206370374438847, + "loss": 0.0756, + "step": 5455 + }, + { + "epoch": 0.1278972321360549, + "grad_norm": 0.6683793067932129, + "learning_rate": 0.00019206082760240393, + "loss": 0.3177, + "step": 5456 + }, + { + "epoch": 0.12792067371085988, + "grad_norm": 0.6532400846481323, + "learning_rate": 0.00019205795096089475, + "loss": 0.0909, + "step": 5457 + }, + { + "epoch": 0.12794411528566488, + "grad_norm": 0.634517252445221, + "learning_rate": 0.00019205507381987657, + "loss": 0.1483, + "step": 5458 + }, + { + "epoch": 0.1279675568604699, + "grad_norm": 0.14644606411457062, + "learning_rate": 0.00019205219617936497, + "loss": 0.036, + "step": 5459 + }, + { + "epoch": 0.1279909984352749, + "grad_norm": 0.192566379904747, + "learning_rate": 0.00019204931803937556, + "loss": 0.0495, + "step": 5460 + }, + { + "epoch": 0.1280144400100799, + "grad_norm": 0.43099862337112427, + "learning_rate": 0.00019204643939992394, + "loss": 0.1315, + "step": 5461 + }, + { + "epoch": 0.12803788158488488, + "grad_norm": 1.4446911811828613, + "learning_rate": 0.00019204356026102576, + "loss": 0.2639, + "step": 5462 + }, + { + "epoch": 0.12806132315968988, + "grad_norm": 0.6080420613288879, + "learning_rate": 0.00019204068062269667, + "loss": 0.1181, + "step": 5463 + }, + { + "epoch": 0.12808476473449487, + "grad_norm": 0.7086613774299622, + "learning_rate": 0.0001920378004849522, + "loss": 0.1536, + "step": 5464 + }, + { + "epoch": 0.12810820630929987, + "grad_norm": 0.9008617997169495, + "learning_rate": 0.0001920349198478081, + "loss": 0.1776, + "step": 5465 + }, + { + "epoch": 0.12813164788410486, + "grad_norm": 0.4156549870967865, + "learning_rate": 0.00019203203871127992, + "loss": 0.1231, + "step": 5466 + }, + { + "epoch": 0.12815508945890985, + "grad_norm": 0.15781910717487335, + "learning_rate": 0.00019202915707538327, + "loss": 0.0455, + "step": 5467 + }, + { + "epoch": 0.12817853103371485, + "grad_norm": 0.7393582463264465, + "learning_rate": 0.00019202627494013385, + "loss": 0.2506, + "step": 5468 + }, + { + "epoch": 0.12820197260851984, + "grad_norm": 0.4081532955169678, + "learning_rate": 0.00019202339230554727, + "loss": 0.1084, + "step": 5469 + }, + { + "epoch": 0.12822541418332484, + "grad_norm": 0.5018624663352966, + "learning_rate": 0.0001920205091716392, + "loss": 0.1527, + "step": 5470 + }, + { + "epoch": 0.12824885575812983, + "grad_norm": 0.8122196197509766, + "learning_rate": 0.00019201762553842523, + "loss": 0.1819, + "step": 5471 + }, + { + "epoch": 0.12827229733293483, + "grad_norm": 0.701198399066925, + "learning_rate": 0.00019201474140592107, + "loss": 0.1946, + "step": 5472 + }, + { + "epoch": 0.12829573890773982, + "grad_norm": 0.1422864943742752, + "learning_rate": 0.0001920118567741423, + "loss": 0.0367, + "step": 5473 + }, + { + "epoch": 0.12831918048254481, + "grad_norm": 0.3510405421257019, + "learning_rate": 0.00019200897164310462, + "loss": 0.0818, + "step": 5474 + }, + { + "epoch": 0.1283426220573498, + "grad_norm": 0.2936546504497528, + "learning_rate": 0.00019200608601282367, + "loss": 0.0856, + "step": 5475 + }, + { + "epoch": 0.1283660636321548, + "grad_norm": 0.42851021885871887, + "learning_rate": 0.0001920031998833151, + "loss": 0.1134, + "step": 5476 + }, + { + "epoch": 0.1283895052069598, + "grad_norm": 0.2524735927581787, + "learning_rate": 0.00019200031325459463, + "loss": 0.0395, + "step": 5477 + }, + { + "epoch": 0.1284129467817648, + "grad_norm": 0.506443440914154, + "learning_rate": 0.00019199742612667785, + "loss": 0.1531, + "step": 5478 + }, + { + "epoch": 0.1284363883565698, + "grad_norm": 0.4858253300189972, + "learning_rate": 0.0001919945384995804, + "loss": 0.1247, + "step": 5479 + }, + { + "epoch": 0.12845982993137478, + "grad_norm": 1.2919026613235474, + "learning_rate": 0.00019199165037331802, + "loss": 0.1849, + "step": 5480 + }, + { + "epoch": 0.12848327150617977, + "grad_norm": 0.2930174767971039, + "learning_rate": 0.0001919887617479064, + "loss": 0.0855, + "step": 5481 + }, + { + "epoch": 0.12850671308098477, + "grad_norm": 0.6353545784950256, + "learning_rate": 0.00019198587262336117, + "loss": 0.2742, + "step": 5482 + }, + { + "epoch": 0.12853015465578976, + "grad_norm": 0.6972396969795227, + "learning_rate": 0.00019198298299969796, + "loss": 0.2548, + "step": 5483 + }, + { + "epoch": 0.12855359623059476, + "grad_norm": 0.6789432168006897, + "learning_rate": 0.00019198009287693253, + "loss": 0.6143, + "step": 5484 + }, + { + "epoch": 0.12857703780539978, + "grad_norm": 0.530250072479248, + "learning_rate": 0.00019197720225508054, + "loss": 0.1573, + "step": 5485 + }, + { + "epoch": 0.12860047938020477, + "grad_norm": 0.592706024646759, + "learning_rate": 0.00019197431113415764, + "loss": 0.1723, + "step": 5486 + }, + { + "epoch": 0.12862392095500977, + "grad_norm": 0.6823065280914307, + "learning_rate": 0.00019197141951417953, + "loss": 0.1617, + "step": 5487 + }, + { + "epoch": 0.12864736252981476, + "grad_norm": 0.7039656043052673, + "learning_rate": 0.00019196852739516195, + "loss": 0.1695, + "step": 5488 + }, + { + "epoch": 0.12867080410461976, + "grad_norm": 0.5220221281051636, + "learning_rate": 0.00019196563477712054, + "loss": 0.56, + "step": 5489 + }, + { + "epoch": 0.12869424567942475, + "grad_norm": 1.027693271636963, + "learning_rate": 0.000191962741660071, + "loss": 0.285, + "step": 5490 + }, + { + "epoch": 0.12871768725422975, + "grad_norm": 0.15059231221675873, + "learning_rate": 0.00019195984804402905, + "loss": 0.0377, + "step": 5491 + }, + { + "epoch": 0.12874112882903474, + "grad_norm": 0.5000175833702087, + "learning_rate": 0.00019195695392901034, + "loss": 0.126, + "step": 5492 + }, + { + "epoch": 0.12876457040383973, + "grad_norm": 0.2513093054294586, + "learning_rate": 0.00019195405931503068, + "loss": 0.0536, + "step": 5493 + }, + { + "epoch": 0.12878801197864473, + "grad_norm": 0.5432225465774536, + "learning_rate": 0.00019195116420210565, + "loss": 0.155, + "step": 5494 + }, + { + "epoch": 0.12881145355344972, + "grad_norm": 0.5044050812721252, + "learning_rate": 0.00019194826859025103, + "loss": 0.1545, + "step": 5495 + }, + { + "epoch": 0.12883489512825472, + "grad_norm": 0.6131700277328491, + "learning_rate": 0.00019194537247948254, + "loss": 0.1196, + "step": 5496 + }, + { + "epoch": 0.1288583367030597, + "grad_norm": 0.15609689056873322, + "learning_rate": 0.0001919424758698159, + "loss": 0.0335, + "step": 5497 + }, + { + "epoch": 0.1288817782778647, + "grad_norm": 0.4721537232398987, + "learning_rate": 0.00019193957876126674, + "loss": 0.1127, + "step": 5498 + }, + { + "epoch": 0.1289052198526697, + "grad_norm": 0.5158931016921997, + "learning_rate": 0.00019193668115385085, + "loss": 0.0972, + "step": 5499 + }, + { + "epoch": 0.1289286614274747, + "grad_norm": 0.1783408373594284, + "learning_rate": 0.00019193378304758398, + "loss": 0.045, + "step": 5500 + }, + { + "epoch": 0.1289521030022797, + "grad_norm": 0.7787634134292603, + "learning_rate": 0.00019193088444248178, + "loss": 0.2298, + "step": 5501 + }, + { + "epoch": 0.12897554457708468, + "grad_norm": 0.5881511569023132, + "learning_rate": 0.00019192798533856003, + "loss": 0.2179, + "step": 5502 + }, + { + "epoch": 0.12899898615188968, + "grad_norm": 0.30246680974960327, + "learning_rate": 0.00019192508573583446, + "loss": 0.1579, + "step": 5503 + }, + { + "epoch": 0.12902242772669467, + "grad_norm": 0.6367906928062439, + "learning_rate": 0.0001919221856343208, + "loss": 0.1372, + "step": 5504 + }, + { + "epoch": 0.12904586930149967, + "grad_norm": 0.3406086266040802, + "learning_rate": 0.00019191928503403474, + "loss": 0.1465, + "step": 5505 + }, + { + "epoch": 0.12906931087630466, + "grad_norm": 0.16937832534313202, + "learning_rate": 0.00019191638393499204, + "loss": 0.0289, + "step": 5506 + }, + { + "epoch": 0.12909275245110965, + "grad_norm": 0.41439780592918396, + "learning_rate": 0.0001919134823372085, + "loss": 0.1083, + "step": 5507 + }, + { + "epoch": 0.12911619402591465, + "grad_norm": 0.8018065094947815, + "learning_rate": 0.0001919105802406998, + "loss": 0.1297, + "step": 5508 + }, + { + "epoch": 0.12913963560071964, + "grad_norm": 0.513583242893219, + "learning_rate": 0.00019190767764548175, + "loss": 0.1399, + "step": 5509 + }, + { + "epoch": 0.12916307717552467, + "grad_norm": 0.5522946715354919, + "learning_rate": 0.00019190477455157, + "loss": 0.2236, + "step": 5510 + }, + { + "epoch": 0.12918651875032966, + "grad_norm": 0.22134177386760712, + "learning_rate": 0.00019190187095898038, + "loss": 0.0673, + "step": 5511 + }, + { + "epoch": 0.12920996032513465, + "grad_norm": 0.6210094094276428, + "learning_rate": 0.00019189896686772865, + "loss": 0.1547, + "step": 5512 + }, + { + "epoch": 0.12923340189993965, + "grad_norm": 0.6164183616638184, + "learning_rate": 0.00019189606227783048, + "loss": 0.6853, + "step": 5513 + }, + { + "epoch": 0.12925684347474464, + "grad_norm": 0.31994563341140747, + "learning_rate": 0.00019189315718930175, + "loss": 0.0935, + "step": 5514 + }, + { + "epoch": 0.12928028504954964, + "grad_norm": 0.2999821901321411, + "learning_rate": 0.00019189025160215811, + "loss": 0.033, + "step": 5515 + }, + { + "epoch": 0.12930372662435463, + "grad_norm": 0.6619710922241211, + "learning_rate": 0.00019188734551641546, + "loss": 0.1641, + "step": 5516 + }, + { + "epoch": 0.12932716819915963, + "grad_norm": 0.5503281950950623, + "learning_rate": 0.00019188443893208943, + "loss": 0.1575, + "step": 5517 + }, + { + "epoch": 0.12935060977396462, + "grad_norm": 0.6445028781890869, + "learning_rate": 0.00019188153184919587, + "loss": 0.1228, + "step": 5518 + }, + { + "epoch": 0.12937405134876961, + "grad_norm": 0.7907410860061646, + "learning_rate": 0.0001918786242677505, + "loss": 0.1313, + "step": 5519 + }, + { + "epoch": 0.1293974929235746, + "grad_norm": 1.1177806854248047, + "learning_rate": 0.00019187571618776918, + "loss": 0.234, + "step": 5520 + }, + { + "epoch": 0.1294209344983796, + "grad_norm": 0.48553305864334106, + "learning_rate": 0.00019187280760926762, + "loss": 0.0709, + "step": 5521 + }, + { + "epoch": 0.1294443760731846, + "grad_norm": 0.8482956290245056, + "learning_rate": 0.00019186989853226162, + "loss": 0.2173, + "step": 5522 + }, + { + "epoch": 0.1294678176479896, + "grad_norm": 0.15273989737033844, + "learning_rate": 0.000191866988956767, + "loss": 0.0486, + "step": 5523 + }, + { + "epoch": 0.12949125922279459, + "grad_norm": 0.8126285076141357, + "learning_rate": 0.00019186407888279947, + "loss": 0.2292, + "step": 5524 + }, + { + "epoch": 0.12951470079759958, + "grad_norm": 0.5539083480834961, + "learning_rate": 0.00019186116831037487, + "loss": 0.1777, + "step": 5525 + }, + { + "epoch": 0.12953814237240457, + "grad_norm": 0.4785238206386566, + "learning_rate": 0.000191858257239509, + "loss": 0.0897, + "step": 5526 + }, + { + "epoch": 0.12956158394720957, + "grad_norm": 0.2527446150779724, + "learning_rate": 0.00019185534567021763, + "loss": 0.0546, + "step": 5527 + }, + { + "epoch": 0.12958502552201456, + "grad_norm": 0.8960402011871338, + "learning_rate": 0.0001918524336025166, + "loss": 0.1358, + "step": 5528 + }, + { + "epoch": 0.12960846709681956, + "grad_norm": 0.8361531496047974, + "learning_rate": 0.00019184952103642168, + "loss": 0.1878, + "step": 5529 + }, + { + "epoch": 0.12963190867162455, + "grad_norm": 0.6155869960784912, + "learning_rate": 0.00019184660797194866, + "loss": 0.147, + "step": 5530 + }, + { + "epoch": 0.12965535024642955, + "grad_norm": 0.813796877861023, + "learning_rate": 0.00019184369440911335, + "loss": 0.1431, + "step": 5531 + }, + { + "epoch": 0.12967879182123454, + "grad_norm": 0.5577207207679749, + "learning_rate": 0.0001918407803479316, + "loss": 0.8266, + "step": 5532 + }, + { + "epoch": 0.12970223339603953, + "grad_norm": 0.22103749215602875, + "learning_rate": 0.0001918378657884192, + "loss": 0.058, + "step": 5533 + }, + { + "epoch": 0.12972567497084453, + "grad_norm": 0.3860591948032379, + "learning_rate": 0.00019183495073059193, + "loss": 0.0803, + "step": 5534 + }, + { + "epoch": 0.12974911654564952, + "grad_norm": 0.38964027166366577, + "learning_rate": 0.00019183203517446562, + "loss": 0.152, + "step": 5535 + }, + { + "epoch": 0.12977255812045455, + "grad_norm": 0.32543009519577026, + "learning_rate": 0.00019182911912005615, + "loss": 0.0935, + "step": 5536 + }, + { + "epoch": 0.12979599969525954, + "grad_norm": 0.6206031441688538, + "learning_rate": 0.0001918262025673793, + "loss": 0.2611, + "step": 5537 + }, + { + "epoch": 0.12981944127006453, + "grad_norm": 0.6132276058197021, + "learning_rate": 0.00019182328551645082, + "loss": 0.1831, + "step": 5538 + }, + { + "epoch": 0.12984288284486953, + "grad_norm": 0.395865261554718, + "learning_rate": 0.00019182036796728668, + "loss": 0.0796, + "step": 5539 + }, + { + "epoch": 0.12986632441967452, + "grad_norm": 0.22640638053417206, + "learning_rate": 0.00019181744991990263, + "loss": 0.0414, + "step": 5540 + }, + { + "epoch": 0.12988976599447952, + "grad_norm": 0.5781745910644531, + "learning_rate": 0.0001918145313743145, + "loss": 0.2097, + "step": 5541 + }, + { + "epoch": 0.1299132075692845, + "grad_norm": 0.5126654505729675, + "learning_rate": 0.00019181161233053815, + "loss": 0.1543, + "step": 5542 + }, + { + "epoch": 0.1299366491440895, + "grad_norm": 0.510229766368866, + "learning_rate": 0.00019180869278858943, + "loss": 0.1653, + "step": 5543 + }, + { + "epoch": 0.1299600907188945, + "grad_norm": 0.4523560404777527, + "learning_rate": 0.00019180577274848412, + "loss": 0.1363, + "step": 5544 + }, + { + "epoch": 0.1299835322936995, + "grad_norm": 0.5683102011680603, + "learning_rate": 0.00019180285221023812, + "loss": 0.2696, + "step": 5545 + }, + { + "epoch": 0.1300069738685045, + "grad_norm": 0.3458428382873535, + "learning_rate": 0.00019179993117386727, + "loss": 0.0902, + "step": 5546 + }, + { + "epoch": 0.13003041544330948, + "grad_norm": 0.3865189552307129, + "learning_rate": 0.0001917970096393874, + "loss": 0.0708, + "step": 5547 + }, + { + "epoch": 0.13005385701811448, + "grad_norm": 0.5504473447799683, + "learning_rate": 0.00019179408760681442, + "loss": 0.2454, + "step": 5548 + }, + { + "epoch": 0.13007729859291947, + "grad_norm": 0.44126465916633606, + "learning_rate": 0.00019179116507616408, + "loss": 0.1541, + "step": 5549 + }, + { + "epoch": 0.13010074016772447, + "grad_norm": 0.2845028042793274, + "learning_rate": 0.00019178824204745234, + "loss": 0.0549, + "step": 5550 + }, + { + "epoch": 0.13012418174252946, + "grad_norm": 0.42912545800209045, + "learning_rate": 0.000191785318520695, + "loss": 0.1059, + "step": 5551 + }, + { + "epoch": 0.13014762331733445, + "grad_norm": 0.5141563415527344, + "learning_rate": 0.00019178239449590793, + "loss": 0.2391, + "step": 5552 + }, + { + "epoch": 0.13017106489213945, + "grad_norm": 0.5067538619041443, + "learning_rate": 0.00019177946997310703, + "loss": 0.1474, + "step": 5553 + }, + { + "epoch": 0.13019450646694444, + "grad_norm": 0.484304279088974, + "learning_rate": 0.00019177654495230813, + "loss": 0.098, + "step": 5554 + }, + { + "epoch": 0.13021794804174944, + "grad_norm": 0.8236412405967712, + "learning_rate": 0.00019177361943352713, + "loss": 0.2157, + "step": 5555 + }, + { + "epoch": 0.13024138961655443, + "grad_norm": 0.8646347522735596, + "learning_rate": 0.00019177069341677988, + "loss": 0.2993, + "step": 5556 + }, + { + "epoch": 0.13026483119135943, + "grad_norm": 0.5444441437721252, + "learning_rate": 0.00019176776690208228, + "loss": 0.125, + "step": 5557 + }, + { + "epoch": 0.13028827276616442, + "grad_norm": 0.498295396566391, + "learning_rate": 0.00019176483988945017, + "loss": 0.1298, + "step": 5558 + }, + { + "epoch": 0.13031171434096941, + "grad_norm": 0.5998585224151611, + "learning_rate": 0.00019176191237889946, + "loss": 0.1449, + "step": 5559 + }, + { + "epoch": 0.1303351559157744, + "grad_norm": 0.39493685960769653, + "learning_rate": 0.00019175898437044607, + "loss": 0.1237, + "step": 5560 + }, + { + "epoch": 0.13035859749057943, + "grad_norm": 0.5636532306671143, + "learning_rate": 0.0001917560558641058, + "loss": 0.1461, + "step": 5561 + }, + { + "epoch": 0.13038203906538443, + "grad_norm": 0.14970889687538147, + "learning_rate": 0.00019175312685989462, + "loss": 0.0334, + "step": 5562 + }, + { + "epoch": 0.13040548064018942, + "grad_norm": 0.43927499651908875, + "learning_rate": 0.0001917501973578284, + "loss": 0.4615, + "step": 5563 + }, + { + "epoch": 0.1304289222149944, + "grad_norm": 0.4086213707923889, + "learning_rate": 0.00019174726735792302, + "loss": 0.0583, + "step": 5564 + }, + { + "epoch": 0.1304523637897994, + "grad_norm": 0.5045574307441711, + "learning_rate": 0.00019174433686019437, + "loss": 0.1427, + "step": 5565 + }, + { + "epoch": 0.1304758053646044, + "grad_norm": 0.7007167935371399, + "learning_rate": 0.00019174140586465842, + "loss": 0.1605, + "step": 5566 + }, + { + "epoch": 0.1304992469394094, + "grad_norm": 0.5935521721839905, + "learning_rate": 0.00019173847437133098, + "loss": 0.7222, + "step": 5567 + }, + { + "epoch": 0.1305226885142144, + "grad_norm": 0.2668466866016388, + "learning_rate": 0.000191735542380228, + "loss": 0.0632, + "step": 5568 + }, + { + "epoch": 0.13054613008901939, + "grad_norm": 0.6972464919090271, + "learning_rate": 0.0001917326098913654, + "loss": 0.2114, + "step": 5569 + }, + { + "epoch": 0.13056957166382438, + "grad_norm": 0.5260108113288879, + "learning_rate": 0.00019172967690475907, + "loss": 0.1546, + "step": 5570 + }, + { + "epoch": 0.13059301323862937, + "grad_norm": 0.9882737994194031, + "learning_rate": 0.00019172674342042494, + "loss": 0.223, + "step": 5571 + }, + { + "epoch": 0.13061645481343437, + "grad_norm": 0.5058031678199768, + "learning_rate": 0.0001917238094383789, + "loss": 0.072, + "step": 5572 + }, + { + "epoch": 0.13063989638823936, + "grad_norm": 0.39893296360969543, + "learning_rate": 0.00019172087495863694, + "loss": 0.1041, + "step": 5573 + }, + { + "epoch": 0.13066333796304436, + "grad_norm": 0.8104521036148071, + "learning_rate": 0.00019171793998121487, + "loss": 0.1786, + "step": 5574 + }, + { + "epoch": 0.13068677953784935, + "grad_norm": 0.1556311994791031, + "learning_rate": 0.00019171500450612875, + "loss": 0.0313, + "step": 5575 + }, + { + "epoch": 0.13071022111265435, + "grad_norm": 1.1329270601272583, + "learning_rate": 0.0001917120685333944, + "loss": 0.2265, + "step": 5576 + }, + { + "epoch": 0.13073366268745934, + "grad_norm": 0.36280640959739685, + "learning_rate": 0.00019170913206302775, + "loss": 0.2501, + "step": 5577 + }, + { + "epoch": 0.13075710426226433, + "grad_norm": 0.6188695430755615, + "learning_rate": 0.00019170619509504484, + "loss": 0.8023, + "step": 5578 + }, + { + "epoch": 0.13078054583706933, + "grad_norm": 0.7656279802322388, + "learning_rate": 0.0001917032576294615, + "loss": 0.1655, + "step": 5579 + }, + { + "epoch": 0.13080398741187432, + "grad_norm": 0.48937058448791504, + "learning_rate": 0.0001917003196662937, + "loss": 0.0659, + "step": 5580 + }, + { + "epoch": 0.13082742898667932, + "grad_norm": 0.38404789566993713, + "learning_rate": 0.00019169738120555741, + "loss": 0.0754, + "step": 5581 + }, + { + "epoch": 0.1308508705614843, + "grad_norm": 0.394162654876709, + "learning_rate": 0.00019169444224726855, + "loss": 0.1072, + "step": 5582 + }, + { + "epoch": 0.1308743121362893, + "grad_norm": 0.20417724549770355, + "learning_rate": 0.00019169150279144305, + "loss": 0.0511, + "step": 5583 + }, + { + "epoch": 0.1308977537110943, + "grad_norm": 0.33603623509407043, + "learning_rate": 0.00019168856283809687, + "loss": 0.0576, + "step": 5584 + }, + { + "epoch": 0.1309211952858993, + "grad_norm": 0.1337730586528778, + "learning_rate": 0.000191685622387246, + "loss": 0.0269, + "step": 5585 + }, + { + "epoch": 0.1309446368607043, + "grad_norm": 0.22262489795684814, + "learning_rate": 0.00019168268143890632, + "loss": 0.0477, + "step": 5586 + }, + { + "epoch": 0.1309680784355093, + "grad_norm": 0.1707533299922943, + "learning_rate": 0.0001916797399930939, + "loss": 0.037, + "step": 5587 + }, + { + "epoch": 0.1309915200103143, + "grad_norm": 0.40170666575431824, + "learning_rate": 0.0001916767980498246, + "loss": 0.0635, + "step": 5588 + }, + { + "epoch": 0.1310149615851193, + "grad_norm": 0.7525162100791931, + "learning_rate": 0.0001916738556091144, + "loss": 0.2453, + "step": 5589 + }, + { + "epoch": 0.1310384031599243, + "grad_norm": 0.4775580167770386, + "learning_rate": 0.0001916709126709793, + "loss": 0.1935, + "step": 5590 + }, + { + "epoch": 0.1310618447347293, + "grad_norm": 0.8253698348999023, + "learning_rate": 0.00019166796923543523, + "loss": 0.2681, + "step": 5591 + }, + { + "epoch": 0.13108528630953428, + "grad_norm": 0.5412485599517822, + "learning_rate": 0.0001916650253024982, + "loss": 0.1271, + "step": 5592 + }, + { + "epoch": 0.13110872788433928, + "grad_norm": 0.574082612991333, + "learning_rate": 0.00019166208087218416, + "loss": 0.1821, + "step": 5593 + }, + { + "epoch": 0.13113216945914427, + "grad_norm": 0.8946244120597839, + "learning_rate": 0.0001916591359445091, + "loss": 0.1393, + "step": 5594 + }, + { + "epoch": 0.13115561103394927, + "grad_norm": 0.5164310336112976, + "learning_rate": 0.00019165619051948897, + "loss": 0.1622, + "step": 5595 + }, + { + "epoch": 0.13117905260875426, + "grad_norm": 0.623349130153656, + "learning_rate": 0.0001916532445971398, + "loss": 0.071, + "step": 5596 + }, + { + "epoch": 0.13120249418355925, + "grad_norm": 0.33540335297584534, + "learning_rate": 0.00019165029817747752, + "loss": 0.1139, + "step": 5597 + }, + { + "epoch": 0.13122593575836425, + "grad_norm": 0.3768281042575836, + "learning_rate": 0.00019164735126051815, + "loss": 0.0956, + "step": 5598 + }, + { + "epoch": 0.13124937733316924, + "grad_norm": 0.640990138053894, + "learning_rate": 0.0001916444038462777, + "loss": 0.266, + "step": 5599 + }, + { + "epoch": 0.13127281890797424, + "grad_norm": 0.21036013960838318, + "learning_rate": 0.00019164145593477213, + "loss": 0.0424, + "step": 5600 + }, + { + "epoch": 0.13129626048277923, + "grad_norm": 1.047577977180481, + "learning_rate": 0.0001916385075260174, + "loss": 0.1464, + "step": 5601 + }, + { + "epoch": 0.13131970205758423, + "grad_norm": 0.5251909494400024, + "learning_rate": 0.00019163555862002963, + "loss": 0.0744, + "step": 5602 + }, + { + "epoch": 0.13134314363238922, + "grad_norm": 0.49889880418777466, + "learning_rate": 0.00019163260921682468, + "loss": 0.1016, + "step": 5603 + }, + { + "epoch": 0.13136658520719421, + "grad_norm": 1.4814033508300781, + "learning_rate": 0.00019162965931641862, + "loss": 0.2078, + "step": 5604 + }, + { + "epoch": 0.1313900267819992, + "grad_norm": 0.4241439402103424, + "learning_rate": 0.0001916267089188275, + "loss": 0.0971, + "step": 5605 + }, + { + "epoch": 0.1314134683568042, + "grad_norm": 0.5754828453063965, + "learning_rate": 0.0001916237580240672, + "loss": 0.1601, + "step": 5606 + }, + { + "epoch": 0.1314369099316092, + "grad_norm": 0.48439082503318787, + "learning_rate": 0.0001916208066321539, + "loss": 0.1134, + "step": 5607 + }, + { + "epoch": 0.1314603515064142, + "grad_norm": 0.18569745123386383, + "learning_rate": 0.00019161785474310346, + "loss": 0.0452, + "step": 5608 + }, + { + "epoch": 0.13148379308121919, + "grad_norm": 0.6800441145896912, + "learning_rate": 0.000191614902356932, + "loss": 0.1522, + "step": 5609 + }, + { + "epoch": 0.13150723465602418, + "grad_norm": 0.8549872040748596, + "learning_rate": 0.0001916119494736555, + "loss": 0.2233, + "step": 5610 + }, + { + "epoch": 0.13153067623082917, + "grad_norm": 0.9567316770553589, + "learning_rate": 0.00019160899609329, + "loss": 0.1811, + "step": 5611 + }, + { + "epoch": 0.13155411780563417, + "grad_norm": 0.7172710299491882, + "learning_rate": 0.0001916060422158515, + "loss": 0.2052, + "step": 5612 + }, + { + "epoch": 0.1315775593804392, + "grad_norm": 0.672437310218811, + "learning_rate": 0.00019160308784135603, + "loss": 0.1794, + "step": 5613 + }, + { + "epoch": 0.13160100095524419, + "grad_norm": 0.12982416152954102, + "learning_rate": 0.00019160013296981964, + "loss": 0.0219, + "step": 5614 + }, + { + "epoch": 0.13162444253004918, + "grad_norm": 0.6378021836280823, + "learning_rate": 0.00019159717760125835, + "loss": 0.1909, + "step": 5615 + }, + { + "epoch": 0.13164788410485417, + "grad_norm": 0.7584196329116821, + "learning_rate": 0.00019159422173568817, + "loss": 0.0805, + "step": 5616 + }, + { + "epoch": 0.13167132567965917, + "grad_norm": 0.37222033739089966, + "learning_rate": 0.00019159126537312518, + "loss": 0.0968, + "step": 5617 + }, + { + "epoch": 0.13169476725446416, + "grad_norm": 0.6182650327682495, + "learning_rate": 0.00019158830851358544, + "loss": 0.1463, + "step": 5618 + }, + { + "epoch": 0.13171820882926916, + "grad_norm": 0.5820327401161194, + "learning_rate": 0.00019158535115708497, + "loss": 0.758, + "step": 5619 + }, + { + "epoch": 0.13174165040407415, + "grad_norm": 0.7238723039627075, + "learning_rate": 0.00019158239330363977, + "loss": 0.1673, + "step": 5620 + }, + { + "epoch": 0.13176509197887915, + "grad_norm": 0.49012771248817444, + "learning_rate": 0.00019157943495326594, + "loss": 0.0566, + "step": 5621 + }, + { + "epoch": 0.13178853355368414, + "grad_norm": 0.35568922758102417, + "learning_rate": 0.0001915764761059795, + "loss": 0.1029, + "step": 5622 + }, + { + "epoch": 0.13181197512848913, + "grad_norm": 0.6180596947669983, + "learning_rate": 0.00019157351676179655, + "loss": 0.1771, + "step": 5623 + }, + { + "epoch": 0.13183541670329413, + "grad_norm": 0.6083436012268066, + "learning_rate": 0.00019157055692073315, + "loss": 0.0729, + "step": 5624 + }, + { + "epoch": 0.13185885827809912, + "grad_norm": 0.8926347494125366, + "learning_rate": 0.0001915675965828053, + "loss": 0.2485, + "step": 5625 + }, + { + "epoch": 0.13188229985290412, + "grad_norm": 0.4060634672641754, + "learning_rate": 0.0001915646357480291, + "loss": 0.0862, + "step": 5626 + }, + { + "epoch": 0.1319057414277091, + "grad_norm": 0.680759072303772, + "learning_rate": 0.0001915616744164206, + "loss": 0.106, + "step": 5627 + }, + { + "epoch": 0.1319291830025141, + "grad_norm": 0.27163633704185486, + "learning_rate": 0.0001915587125879959, + "loss": 0.0372, + "step": 5628 + }, + { + "epoch": 0.1319526245773191, + "grad_norm": 0.44910991191864014, + "learning_rate": 0.00019155575026277104, + "loss": 0.0802, + "step": 5629 + }, + { + "epoch": 0.1319760661521241, + "grad_norm": 0.5127878785133362, + "learning_rate": 0.00019155278744076212, + "loss": 0.1607, + "step": 5630 + }, + { + "epoch": 0.1319995077269291, + "grad_norm": 0.6593365669250488, + "learning_rate": 0.00019154982412198518, + "loss": 0.1416, + "step": 5631 + }, + { + "epoch": 0.13202294930173408, + "grad_norm": 0.7719193696975708, + "learning_rate": 0.00019154686030645635, + "loss": 0.1842, + "step": 5632 + }, + { + "epoch": 0.13204639087653908, + "grad_norm": 0.21539849042892456, + "learning_rate": 0.00019154389599419167, + "loss": 0.0626, + "step": 5633 + }, + { + "epoch": 0.13206983245134407, + "grad_norm": 0.6188259720802307, + "learning_rate": 0.00019154093118520723, + "loss": 0.1466, + "step": 5634 + }, + { + "epoch": 0.13209327402614907, + "grad_norm": 0.8281932473182678, + "learning_rate": 0.0001915379658795191, + "loss": 0.101, + "step": 5635 + }, + { + "epoch": 0.13211671560095406, + "grad_norm": 0.7204368710517883, + "learning_rate": 0.00019153500007714343, + "loss": 0.1953, + "step": 5636 + }, + { + "epoch": 0.13214015717575905, + "grad_norm": 0.17190596461296082, + "learning_rate": 0.00019153203377809627, + "loss": 0.0417, + "step": 5637 + }, + { + "epoch": 0.13216359875056408, + "grad_norm": 0.31978824734687805, + "learning_rate": 0.00019152906698239371, + "loss": 0.0433, + "step": 5638 + }, + { + "epoch": 0.13218704032536907, + "grad_norm": 0.9154897332191467, + "learning_rate": 0.00019152609969005186, + "loss": 0.2059, + "step": 5639 + }, + { + "epoch": 0.13221048190017406, + "grad_norm": 0.6316845417022705, + "learning_rate": 0.00019152313190108682, + "loss": 0.768, + "step": 5640 + }, + { + "epoch": 0.13223392347497906, + "grad_norm": 0.7034017443656921, + "learning_rate": 0.0001915201636155147, + "loss": 0.0859, + "step": 5641 + }, + { + "epoch": 0.13225736504978405, + "grad_norm": 0.6716580390930176, + "learning_rate": 0.0001915171948333516, + "loss": 0.0628, + "step": 5642 + }, + { + "epoch": 0.13228080662458905, + "grad_norm": 0.660463809967041, + "learning_rate": 0.0001915142255546136, + "loss": 0.1002, + "step": 5643 + }, + { + "epoch": 0.13230424819939404, + "grad_norm": 0.5589186549186707, + "learning_rate": 0.00019151125577931687, + "loss": 0.1068, + "step": 5644 + }, + { + "epoch": 0.13232768977419904, + "grad_norm": 0.5404201745986938, + "learning_rate": 0.00019150828550747748, + "loss": 0.1142, + "step": 5645 + }, + { + "epoch": 0.13235113134900403, + "grad_norm": 0.8437836766242981, + "learning_rate": 0.00019150531473911158, + "loss": 0.2449, + "step": 5646 + }, + { + "epoch": 0.13237457292380903, + "grad_norm": 0.5974132418632507, + "learning_rate": 0.00019150234347423522, + "loss": 0.1109, + "step": 5647 + }, + { + "epoch": 0.13239801449861402, + "grad_norm": 0.5824565291404724, + "learning_rate": 0.00019149937171286462, + "loss": 0.0885, + "step": 5648 + }, + { + "epoch": 0.132421456073419, + "grad_norm": 0.3380241096019745, + "learning_rate": 0.00019149639945501584, + "loss": 0.0318, + "step": 5649 + }, + { + "epoch": 0.132444897648224, + "grad_norm": 0.25709468126296997, + "learning_rate": 0.000191493426700705, + "loss": 0.0418, + "step": 5650 + }, + { + "epoch": 0.132468339223029, + "grad_norm": 0.5986593961715698, + "learning_rate": 0.0001914904534499483, + "loss": 0.173, + "step": 5651 + }, + { + "epoch": 0.132491780797834, + "grad_norm": 0.11441398411989212, + "learning_rate": 0.00019148747970276178, + "loss": 0.0304, + "step": 5652 + }, + { + "epoch": 0.132515222372639, + "grad_norm": 0.23416990041732788, + "learning_rate": 0.00019148450545916168, + "loss": 0.1048, + "step": 5653 + }, + { + "epoch": 0.13253866394744399, + "grad_norm": 0.8028340935707092, + "learning_rate": 0.00019148153071916403, + "loss": 0.2029, + "step": 5654 + }, + { + "epoch": 0.13256210552224898, + "grad_norm": 0.7615124583244324, + "learning_rate": 0.00019147855548278507, + "loss": 0.156, + "step": 5655 + }, + { + "epoch": 0.13258554709705397, + "grad_norm": 0.3028900623321533, + "learning_rate": 0.00019147557975004084, + "loss": 0.0834, + "step": 5656 + }, + { + "epoch": 0.13260898867185897, + "grad_norm": 0.758773684501648, + "learning_rate": 0.00019147260352094759, + "loss": 0.1689, + "step": 5657 + }, + { + "epoch": 0.13263243024666396, + "grad_norm": 0.567668080329895, + "learning_rate": 0.00019146962679552138, + "loss": 0.1577, + "step": 5658 + }, + { + "epoch": 0.13265587182146896, + "grad_norm": 0.7125945687294006, + "learning_rate": 0.00019146664957377842, + "loss": 0.1019, + "step": 5659 + }, + { + "epoch": 0.13267931339627395, + "grad_norm": 0.3225789964199066, + "learning_rate": 0.00019146367185573487, + "loss": 0.0859, + "step": 5660 + }, + { + "epoch": 0.13270275497107895, + "grad_norm": 0.4827589988708496, + "learning_rate": 0.00019146069364140683, + "loss": 0.7038, + "step": 5661 + }, + { + "epoch": 0.13272619654588394, + "grad_norm": 0.24696066975593567, + "learning_rate": 0.0001914577149308105, + "loss": 0.0773, + "step": 5662 + }, + { + "epoch": 0.13274963812068893, + "grad_norm": 0.508111834526062, + "learning_rate": 0.00019145473572396206, + "loss": 0.1138, + "step": 5663 + }, + { + "epoch": 0.13277307969549396, + "grad_norm": 0.22393028438091278, + "learning_rate": 0.00019145175602087764, + "loss": 0.0786, + "step": 5664 + }, + { + "epoch": 0.13279652127029895, + "grad_norm": 0.3276944160461426, + "learning_rate": 0.00019144877582157342, + "loss": 0.0863, + "step": 5665 + }, + { + "epoch": 0.13281996284510394, + "grad_norm": 0.38588088750839233, + "learning_rate": 0.00019144579512606557, + "loss": 0.1101, + "step": 5666 + }, + { + "epoch": 0.13284340441990894, + "grad_norm": 0.6115149855613708, + "learning_rate": 0.0001914428139343703, + "loss": 0.5725, + "step": 5667 + }, + { + "epoch": 0.13286684599471393, + "grad_norm": 0.7017603516578674, + "learning_rate": 0.00019143983224650368, + "loss": 0.6125, + "step": 5668 + }, + { + "epoch": 0.13289028756951893, + "grad_norm": 0.6135273575782776, + "learning_rate": 0.000191436850062482, + "loss": 0.1608, + "step": 5669 + }, + { + "epoch": 0.13291372914432392, + "grad_norm": 0.42269366979599, + "learning_rate": 0.0001914338673823214, + "loss": 0.0794, + "step": 5670 + }, + { + "epoch": 0.13293717071912892, + "grad_norm": 0.19758225977420807, + "learning_rate": 0.0001914308842060381, + "loss": 0.0708, + "step": 5671 + }, + { + "epoch": 0.1329606122939339, + "grad_norm": 0.6496872305870056, + "learning_rate": 0.00019142790053364818, + "loss": 0.1072, + "step": 5672 + }, + { + "epoch": 0.1329840538687389, + "grad_norm": 0.5113241076469421, + "learning_rate": 0.00019142491636516793, + "loss": 0.1106, + "step": 5673 + }, + { + "epoch": 0.1330074954435439, + "grad_norm": 0.8979988694190979, + "learning_rate": 0.00019142193170061352, + "loss": 0.1135, + "step": 5674 + }, + { + "epoch": 0.1330309370183489, + "grad_norm": 0.5873823165893555, + "learning_rate": 0.00019141894654000114, + "loss": 0.1593, + "step": 5675 + }, + { + "epoch": 0.1330543785931539, + "grad_norm": 0.3715650737285614, + "learning_rate": 0.00019141596088334697, + "loss": 0.1073, + "step": 5676 + }, + { + "epoch": 0.13307782016795888, + "grad_norm": 0.7450442314147949, + "learning_rate": 0.00019141297473066724, + "loss": 0.2027, + "step": 5677 + }, + { + "epoch": 0.13310126174276388, + "grad_norm": 0.18140217661857605, + "learning_rate": 0.00019140998808197813, + "loss": 0.0479, + "step": 5678 + }, + { + "epoch": 0.13312470331756887, + "grad_norm": 0.5389111638069153, + "learning_rate": 0.00019140700093729586, + "loss": 0.1426, + "step": 5679 + }, + { + "epoch": 0.13314814489237387, + "grad_norm": 0.3686136305332184, + "learning_rate": 0.00019140401329663663, + "loss": 0.1085, + "step": 5680 + }, + { + "epoch": 0.13317158646717886, + "grad_norm": 0.591879665851593, + "learning_rate": 0.00019140102516001668, + "loss": 0.185, + "step": 5681 + }, + { + "epoch": 0.13319502804198385, + "grad_norm": 0.8037941455841064, + "learning_rate": 0.00019139803652745216, + "loss": 0.227, + "step": 5682 + }, + { + "epoch": 0.13321846961678885, + "grad_norm": 0.5263047218322754, + "learning_rate": 0.00019139504739895932, + "loss": 0.0834, + "step": 5683 + }, + { + "epoch": 0.13324191119159384, + "grad_norm": 0.5298687219619751, + "learning_rate": 0.0001913920577745544, + "loss": 0.1251, + "step": 5684 + }, + { + "epoch": 0.13326535276639884, + "grad_norm": 0.6051421165466309, + "learning_rate": 0.00019138906765425363, + "loss": 0.1625, + "step": 5685 + }, + { + "epoch": 0.13328879434120383, + "grad_norm": 0.4795573055744171, + "learning_rate": 0.00019138607703807317, + "loss": 0.1121, + "step": 5686 + }, + { + "epoch": 0.13331223591600883, + "grad_norm": 0.6081002354621887, + "learning_rate": 0.0001913830859260293, + "loss": 0.1392, + "step": 5687 + }, + { + "epoch": 0.13333567749081382, + "grad_norm": 0.4221321642398834, + "learning_rate": 0.00019138009431813826, + "loss": 0.0913, + "step": 5688 + }, + { + "epoch": 0.13335911906561884, + "grad_norm": 0.24860531091690063, + "learning_rate": 0.00019137710221441622, + "loss": 0.0781, + "step": 5689 + }, + { + "epoch": 0.13338256064042384, + "grad_norm": 0.5476186871528625, + "learning_rate": 0.00019137410961487947, + "loss": 0.0934, + "step": 5690 + }, + { + "epoch": 0.13340600221522883, + "grad_norm": 0.7506556510925293, + "learning_rate": 0.00019137111651954425, + "loss": 0.2402, + "step": 5691 + }, + { + "epoch": 0.13342944379003382, + "grad_norm": 0.6422134041786194, + "learning_rate": 0.00019136812292842677, + "loss": 0.1315, + "step": 5692 + }, + { + "epoch": 0.13345288536483882, + "grad_norm": 0.2117587924003601, + "learning_rate": 0.00019136512884154326, + "loss": 0.0556, + "step": 5693 + }, + { + "epoch": 0.1334763269396438, + "grad_norm": 0.7030360102653503, + "learning_rate": 0.00019136213425891002, + "loss": 0.2471, + "step": 5694 + }, + { + "epoch": 0.1334997685144488, + "grad_norm": 0.44013693928718567, + "learning_rate": 0.00019135913918054325, + "loss": 0.0786, + "step": 5695 + }, + { + "epoch": 0.1335232100892538, + "grad_norm": 0.35028696060180664, + "learning_rate": 0.00019135614360645925, + "loss": 0.1042, + "step": 5696 + }, + { + "epoch": 0.1335466516640588, + "grad_norm": 0.7212689518928528, + "learning_rate": 0.00019135314753667424, + "loss": 0.1492, + "step": 5697 + }, + { + "epoch": 0.1335700932388638, + "grad_norm": 0.43306615948677063, + "learning_rate": 0.00019135015097120447, + "loss": 0.0839, + "step": 5698 + }, + { + "epoch": 0.13359353481366879, + "grad_norm": 0.21800129115581512, + "learning_rate": 0.0001913471539100662, + "loss": 0.0489, + "step": 5699 + }, + { + "epoch": 0.13361697638847378, + "grad_norm": 0.7119736671447754, + "learning_rate": 0.0001913441563532757, + "loss": 0.1666, + "step": 5700 + }, + { + "epoch": 0.13364041796327877, + "grad_norm": 0.28365471959114075, + "learning_rate": 0.0001913411583008493, + "loss": 0.0448, + "step": 5701 + }, + { + "epoch": 0.13366385953808377, + "grad_norm": 0.7287699580192566, + "learning_rate": 0.00019133815975280313, + "loss": 0.2156, + "step": 5702 + }, + { + "epoch": 0.13368730111288876, + "grad_norm": 0.3877710700035095, + "learning_rate": 0.00019133516070915358, + "loss": 0.0791, + "step": 5703 + }, + { + "epoch": 0.13371074268769376, + "grad_norm": 0.5249022245407104, + "learning_rate": 0.0001913321611699169, + "loss": 0.1502, + "step": 5704 + }, + { + "epoch": 0.13373418426249875, + "grad_norm": 0.5204666256904602, + "learning_rate": 0.0001913291611351093, + "loss": 0.166, + "step": 5705 + }, + { + "epoch": 0.13375762583730375, + "grad_norm": 0.5781721472740173, + "learning_rate": 0.00019132616060474713, + "loss": 0.2022, + "step": 5706 + }, + { + "epoch": 0.13378106741210874, + "grad_norm": 0.5931090116500854, + "learning_rate": 0.00019132315957884664, + "loss": 0.6106, + "step": 5707 + }, + { + "epoch": 0.13380450898691373, + "grad_norm": 0.707216203212738, + "learning_rate": 0.00019132015805742411, + "loss": 0.8224, + "step": 5708 + }, + { + "epoch": 0.13382795056171873, + "grad_norm": 0.7581456303596497, + "learning_rate": 0.00019131715604049586, + "loss": 0.1371, + "step": 5709 + }, + { + "epoch": 0.13385139213652372, + "grad_norm": 0.44068554043769836, + "learning_rate": 0.0001913141535280781, + "loss": 0.1425, + "step": 5710 + }, + { + "epoch": 0.13387483371132872, + "grad_norm": 0.6149005889892578, + "learning_rate": 0.00019131115052018721, + "loss": 0.1151, + "step": 5711 + }, + { + "epoch": 0.1338982752861337, + "grad_norm": 0.5517605543136597, + "learning_rate": 0.00019130814701683946, + "loss": 0.1513, + "step": 5712 + }, + { + "epoch": 0.1339217168609387, + "grad_norm": 0.34718120098114014, + "learning_rate": 0.00019130514301805114, + "loss": 0.0482, + "step": 5713 + }, + { + "epoch": 0.1339451584357437, + "grad_norm": 0.16342392563819885, + "learning_rate": 0.00019130213852383852, + "loss": 0.0366, + "step": 5714 + }, + { + "epoch": 0.13396860001054872, + "grad_norm": 0.47614786028862, + "learning_rate": 0.00019129913353421794, + "loss": 0.6649, + "step": 5715 + }, + { + "epoch": 0.13399204158535372, + "grad_norm": 0.15540871024131775, + "learning_rate": 0.0001912961280492057, + "loss": 0.0372, + "step": 5716 + }, + { + "epoch": 0.1340154831601587, + "grad_norm": 0.4965563118457794, + "learning_rate": 0.00019129312206881812, + "loss": 0.106, + "step": 5717 + }, + { + "epoch": 0.1340389247349637, + "grad_norm": 0.3009815514087677, + "learning_rate": 0.00019129011559307146, + "loss": 0.083, + "step": 5718 + }, + { + "epoch": 0.1340623663097687, + "grad_norm": 0.7699943780899048, + "learning_rate": 0.00019128710862198207, + "loss": 0.2301, + "step": 5719 + }, + { + "epoch": 0.1340858078845737, + "grad_norm": 1.040202021598816, + "learning_rate": 0.00019128410115556627, + "loss": 0.2539, + "step": 5720 + }, + { + "epoch": 0.1341092494593787, + "grad_norm": 0.59761643409729, + "learning_rate": 0.00019128109319384038, + "loss": 0.1642, + "step": 5721 + }, + { + "epoch": 0.13413269103418368, + "grad_norm": 0.43942418694496155, + "learning_rate": 0.0001912780847368207, + "loss": 0.1354, + "step": 5722 + }, + { + "epoch": 0.13415613260898868, + "grad_norm": 0.2964972257614136, + "learning_rate": 0.00019127507578452357, + "loss": 0.0817, + "step": 5723 + }, + { + "epoch": 0.13417957418379367, + "grad_norm": 0.2797893285751343, + "learning_rate": 0.0001912720663369653, + "loss": 0.0694, + "step": 5724 + }, + { + "epoch": 0.13420301575859866, + "grad_norm": 0.7519292235374451, + "learning_rate": 0.00019126905639416225, + "loss": 0.1703, + "step": 5725 + }, + { + "epoch": 0.13422645733340366, + "grad_norm": 0.8468036651611328, + "learning_rate": 0.00019126604595613071, + "loss": 0.181, + "step": 5726 + }, + { + "epoch": 0.13424989890820865, + "grad_norm": 0.2619486451148987, + "learning_rate": 0.00019126303502288707, + "loss": 0.0569, + "step": 5727 + }, + { + "epoch": 0.13427334048301365, + "grad_norm": 0.21433134377002716, + "learning_rate": 0.00019126002359444764, + "loss": 0.0839, + "step": 5728 + }, + { + "epoch": 0.13429678205781864, + "grad_norm": 0.381171315908432, + "learning_rate": 0.00019125701167082873, + "loss": 0.1361, + "step": 5729 + }, + { + "epoch": 0.13432022363262364, + "grad_norm": 0.8697596788406372, + "learning_rate": 0.0001912539992520467, + "loss": 0.1455, + "step": 5730 + }, + { + "epoch": 0.13434366520742863, + "grad_norm": 0.43730536103248596, + "learning_rate": 0.0001912509863381179, + "loss": 0.1208, + "step": 5731 + }, + { + "epoch": 0.13436710678223363, + "grad_norm": 0.696049153804779, + "learning_rate": 0.0001912479729290587, + "loss": 0.1971, + "step": 5732 + }, + { + "epoch": 0.13439054835703862, + "grad_norm": 0.4349110722541809, + "learning_rate": 0.00019124495902488547, + "loss": 0.5714, + "step": 5733 + }, + { + "epoch": 0.1344139899318436, + "grad_norm": 0.5058250427246094, + "learning_rate": 0.0001912419446256145, + "loss": 0.1144, + "step": 5734 + }, + { + "epoch": 0.1344374315066486, + "grad_norm": 0.7198127508163452, + "learning_rate": 0.00019123892973126215, + "loss": 0.1473, + "step": 5735 + }, + { + "epoch": 0.1344608730814536, + "grad_norm": 0.6757160425186157, + "learning_rate": 0.0001912359143418448, + "loss": 0.1396, + "step": 5736 + }, + { + "epoch": 0.1344843146562586, + "grad_norm": 0.12641862034797668, + "learning_rate": 0.00019123289845737884, + "loss": 0.0296, + "step": 5737 + }, + { + "epoch": 0.1345077562310636, + "grad_norm": 0.5795274376869202, + "learning_rate": 0.0001912298820778806, + "loss": 0.6808, + "step": 5738 + }, + { + "epoch": 0.13453119780586859, + "grad_norm": 0.6523303985595703, + "learning_rate": 0.00019122686520336646, + "loss": 0.144, + "step": 5739 + }, + { + "epoch": 0.1345546393806736, + "grad_norm": 0.43756312131881714, + "learning_rate": 0.00019122384783385276, + "loss": 0.1135, + "step": 5740 + }, + { + "epoch": 0.1345780809554786, + "grad_norm": 0.6331928968429565, + "learning_rate": 0.0001912208299693559, + "loss": 0.1705, + "step": 5741 + }, + { + "epoch": 0.1346015225302836, + "grad_norm": 0.25863388180732727, + "learning_rate": 0.00019121781160989228, + "loss": 0.0671, + "step": 5742 + }, + { + "epoch": 0.1346249641050886, + "grad_norm": 0.5726792216300964, + "learning_rate": 0.00019121479275547825, + "loss": 0.7876, + "step": 5743 + }, + { + "epoch": 0.13464840567989358, + "grad_norm": 0.6719364523887634, + "learning_rate": 0.00019121177340613015, + "loss": 0.1268, + "step": 5744 + }, + { + "epoch": 0.13467184725469858, + "grad_norm": 0.6984656453132629, + "learning_rate": 0.00019120875356186444, + "loss": 0.11, + "step": 5745 + }, + { + "epoch": 0.13469528882950357, + "grad_norm": 0.3907843828201294, + "learning_rate": 0.00019120573322269744, + "loss": 0.065, + "step": 5746 + }, + { + "epoch": 0.13471873040430857, + "grad_norm": 0.09464084357023239, + "learning_rate": 0.0001912027123886456, + "loss": 0.0129, + "step": 5747 + }, + { + "epoch": 0.13474217197911356, + "grad_norm": 0.15803375840187073, + "learning_rate": 0.00019119969105972526, + "loss": 0.0267, + "step": 5748 + }, + { + "epoch": 0.13476561355391856, + "grad_norm": 0.5468818545341492, + "learning_rate": 0.00019119666923595282, + "loss": 0.142, + "step": 5749 + }, + { + "epoch": 0.13478905512872355, + "grad_norm": 0.47093847393989563, + "learning_rate": 0.0001911936469173447, + "loss": 0.0984, + "step": 5750 + }, + { + "epoch": 0.13481249670352854, + "grad_norm": 0.4574374556541443, + "learning_rate": 0.00019119062410391727, + "loss": 0.1489, + "step": 5751 + }, + { + "epoch": 0.13483593827833354, + "grad_norm": 0.5617175102233887, + "learning_rate": 0.000191187600795687, + "loss": 0.1342, + "step": 5752 + }, + { + "epoch": 0.13485937985313853, + "grad_norm": 0.9450098276138306, + "learning_rate": 0.0001911845769926702, + "loss": 0.1301, + "step": 5753 + }, + { + "epoch": 0.13488282142794353, + "grad_norm": 0.45831069350242615, + "learning_rate": 0.00019118155269488333, + "loss": 0.1186, + "step": 5754 + }, + { + "epoch": 0.13490626300274852, + "grad_norm": 0.534153401851654, + "learning_rate": 0.00019117852790234282, + "loss": 0.1446, + "step": 5755 + }, + { + "epoch": 0.13492970457755352, + "grad_norm": 0.4007713794708252, + "learning_rate": 0.00019117550261506502, + "loss": 0.1115, + "step": 5756 + }, + { + "epoch": 0.1349531461523585, + "grad_norm": 0.5195124745368958, + "learning_rate": 0.0001911724768330664, + "loss": 0.0982, + "step": 5757 + }, + { + "epoch": 0.1349765877271635, + "grad_norm": 0.6450272798538208, + "learning_rate": 0.00019116945055636334, + "loss": 0.1405, + "step": 5758 + }, + { + "epoch": 0.1350000293019685, + "grad_norm": 0.5056310892105103, + "learning_rate": 0.00019116642378497228, + "loss": 0.1533, + "step": 5759 + }, + { + "epoch": 0.1350234708767735, + "grad_norm": 0.5690203905105591, + "learning_rate": 0.00019116339651890964, + "loss": 0.1174, + "step": 5760 + }, + { + "epoch": 0.1350469124515785, + "grad_norm": 0.7776669263839722, + "learning_rate": 0.00019116036875819185, + "loss": 0.1918, + "step": 5761 + }, + { + "epoch": 0.13507035402638348, + "grad_norm": 0.181594580411911, + "learning_rate": 0.00019115734050283533, + "loss": 0.0348, + "step": 5762 + }, + { + "epoch": 0.13509379560118848, + "grad_norm": 0.4995880126953125, + "learning_rate": 0.00019115431175285651, + "loss": 0.1031, + "step": 5763 + }, + { + "epoch": 0.13511723717599347, + "grad_norm": 0.7445613741874695, + "learning_rate": 0.00019115128250827185, + "loss": 0.2097, + "step": 5764 + }, + { + "epoch": 0.13514067875079847, + "grad_norm": 0.7971057295799255, + "learning_rate": 0.00019114825276909776, + "loss": 0.1418, + "step": 5765 + }, + { + "epoch": 0.1351641203256035, + "grad_norm": 0.46290650963783264, + "learning_rate": 0.0001911452225353507, + "loss": 0.1033, + "step": 5766 + }, + { + "epoch": 0.13518756190040848, + "grad_norm": 0.7746226787567139, + "learning_rate": 0.0001911421918070471, + "loss": 0.5518, + "step": 5767 + }, + { + "epoch": 0.13521100347521348, + "grad_norm": 0.1755896508693695, + "learning_rate": 0.0001911391605842034, + "loss": 0.0584, + "step": 5768 + }, + { + "epoch": 0.13523444505001847, + "grad_norm": 0.34431979060173035, + "learning_rate": 0.00019113612886683602, + "loss": 0.0464, + "step": 5769 + }, + { + "epoch": 0.13525788662482346, + "grad_norm": 0.6802372336387634, + "learning_rate": 0.00019113309665496147, + "loss": 0.1184, + "step": 5770 + }, + { + "epoch": 0.13528132819962846, + "grad_norm": 0.11307163536548615, + "learning_rate": 0.0001911300639485962, + "loss": 0.0242, + "step": 5771 + }, + { + "epoch": 0.13530476977443345, + "grad_norm": 0.48446837067604065, + "learning_rate": 0.0001911270307477566, + "loss": 0.1297, + "step": 5772 + }, + { + "epoch": 0.13532821134923845, + "grad_norm": 0.6221198439598083, + "learning_rate": 0.0001911239970524592, + "loss": 0.1494, + "step": 5773 + }, + { + "epoch": 0.13535165292404344, + "grad_norm": 0.1468968689441681, + "learning_rate": 0.0001911209628627204, + "loss": 0.0252, + "step": 5774 + }, + { + "epoch": 0.13537509449884844, + "grad_norm": 0.6514798402786255, + "learning_rate": 0.00019111792817855673, + "loss": 0.1354, + "step": 5775 + }, + { + "epoch": 0.13539853607365343, + "grad_norm": 0.7344433069229126, + "learning_rate": 0.0001911148929999846, + "loss": 0.1594, + "step": 5776 + }, + { + "epoch": 0.13542197764845842, + "grad_norm": 0.5462332367897034, + "learning_rate": 0.00019111185732702049, + "loss": 0.1279, + "step": 5777 + }, + { + "epoch": 0.13544541922326342, + "grad_norm": 0.7302651405334473, + "learning_rate": 0.00019110882115968086, + "loss": 0.2065, + "step": 5778 + }, + { + "epoch": 0.1354688607980684, + "grad_norm": 0.2341184765100479, + "learning_rate": 0.00019110578449798226, + "loss": 0.0464, + "step": 5779 + }, + { + "epoch": 0.1354923023728734, + "grad_norm": 0.5587164759635925, + "learning_rate": 0.00019110274734194106, + "loss": 0.1286, + "step": 5780 + }, + { + "epoch": 0.1355157439476784, + "grad_norm": 0.2210274338722229, + "learning_rate": 0.0001910997096915738, + "loss": 0.0543, + "step": 5781 + }, + { + "epoch": 0.1355391855224834, + "grad_norm": 1.3473083972930908, + "learning_rate": 0.000191096671546897, + "loss": 0.1635, + "step": 5782 + }, + { + "epoch": 0.1355626270972884, + "grad_norm": 0.5416290163993835, + "learning_rate": 0.00019109363290792706, + "loss": 0.1069, + "step": 5783 + }, + { + "epoch": 0.13558606867209338, + "grad_norm": 0.4160126745700836, + "learning_rate": 0.00019109059377468054, + "loss": 0.0887, + "step": 5784 + }, + { + "epoch": 0.13560951024689838, + "grad_norm": 0.7924141883850098, + "learning_rate": 0.00019108755414717386, + "loss": 0.6312, + "step": 5785 + }, + { + "epoch": 0.13563295182170337, + "grad_norm": 0.7541924715042114, + "learning_rate": 0.00019108451402542355, + "loss": 0.1734, + "step": 5786 + }, + { + "epoch": 0.13565639339650837, + "grad_norm": 0.4015858471393585, + "learning_rate": 0.00019108147340944612, + "loss": 0.1478, + "step": 5787 + }, + { + "epoch": 0.13567983497131336, + "grad_norm": 0.6500323414802551, + "learning_rate": 0.00019107843229925804, + "loss": 0.1896, + "step": 5788 + }, + { + "epoch": 0.13570327654611836, + "grad_norm": 0.5677887201309204, + "learning_rate": 0.00019107539069487586, + "loss": 0.2195, + "step": 5789 + }, + { + "epoch": 0.13572671812092335, + "grad_norm": 0.4809390604496002, + "learning_rate": 0.000191072348596316, + "loss": 0.1545, + "step": 5790 + }, + { + "epoch": 0.13575015969572835, + "grad_norm": 0.6122188568115234, + "learning_rate": 0.00019106930600359506, + "loss": 0.1332, + "step": 5791 + }, + { + "epoch": 0.13577360127053337, + "grad_norm": 0.27907034754753113, + "learning_rate": 0.00019106626291672947, + "loss": 0.0293, + "step": 5792 + }, + { + "epoch": 0.13579704284533836, + "grad_norm": 0.5649412870407104, + "learning_rate": 0.0001910632193357358, + "loss": 0.1635, + "step": 5793 + }, + { + "epoch": 0.13582048442014336, + "grad_norm": 0.2817741930484772, + "learning_rate": 0.0001910601752606305, + "loss": 0.0534, + "step": 5794 + }, + { + "epoch": 0.13584392599494835, + "grad_norm": 0.7445122599601746, + "learning_rate": 0.0001910571306914302, + "loss": 0.0991, + "step": 5795 + }, + { + "epoch": 0.13586736756975334, + "grad_norm": 0.24242858588695526, + "learning_rate": 0.00019105408562815132, + "loss": 0.0317, + "step": 5796 + }, + { + "epoch": 0.13589080914455834, + "grad_norm": 0.5426859259605408, + "learning_rate": 0.00019105104007081038, + "loss": 0.1561, + "step": 5797 + }, + { + "epoch": 0.13591425071936333, + "grad_norm": 0.5307457447052002, + "learning_rate": 0.00019104799401942397, + "loss": 0.8157, + "step": 5798 + }, + { + "epoch": 0.13593769229416833, + "grad_norm": 0.7795862555503845, + "learning_rate": 0.0001910449474740086, + "loss": 0.1462, + "step": 5799 + }, + { + "epoch": 0.13596113386897332, + "grad_norm": 0.6607405543327332, + "learning_rate": 0.00019104190043458074, + "loss": 0.1323, + "step": 5800 + }, + { + "epoch": 0.13598457544377832, + "grad_norm": 0.3361104130744934, + "learning_rate": 0.00019103885290115696, + "loss": 0.0639, + "step": 5801 + }, + { + "epoch": 0.1360080170185833, + "grad_norm": 0.5371630787849426, + "learning_rate": 0.00019103580487375384, + "loss": 0.1245, + "step": 5802 + }, + { + "epoch": 0.1360314585933883, + "grad_norm": 0.4638861119747162, + "learning_rate": 0.0001910327563523879, + "loss": 0.1059, + "step": 5803 + }, + { + "epoch": 0.1360549001681933, + "grad_norm": 0.6501160264015198, + "learning_rate": 0.0001910297073370756, + "loss": 0.2311, + "step": 5804 + }, + { + "epoch": 0.1360783417429983, + "grad_norm": 0.39314785599708557, + "learning_rate": 0.0001910266578278336, + "loss": 0.1095, + "step": 5805 + }, + { + "epoch": 0.1361017833178033, + "grad_norm": 0.7238527536392212, + "learning_rate": 0.00019102360782467838, + "loss": 0.1979, + "step": 5806 + }, + { + "epoch": 0.13612522489260828, + "grad_norm": 0.6097061634063721, + "learning_rate": 0.0001910205573276265, + "loss": 0.1306, + "step": 5807 + }, + { + "epoch": 0.13614866646741328, + "grad_norm": 0.6048839092254639, + "learning_rate": 0.00019101750633669453, + "loss": 0.1342, + "step": 5808 + }, + { + "epoch": 0.13617210804221827, + "grad_norm": 0.48528048396110535, + "learning_rate": 0.00019101445485189902, + "loss": 0.1377, + "step": 5809 + }, + { + "epoch": 0.13619554961702326, + "grad_norm": 0.4858405888080597, + "learning_rate": 0.0001910114028732565, + "loss": 0.1373, + "step": 5810 + }, + { + "epoch": 0.13621899119182826, + "grad_norm": 0.36225780844688416, + "learning_rate": 0.00019100835040078355, + "loss": 0.0561, + "step": 5811 + }, + { + "epoch": 0.13624243276663325, + "grad_norm": 0.6073256134986877, + "learning_rate": 0.00019100529743449674, + "loss": 0.3839, + "step": 5812 + }, + { + "epoch": 0.13626587434143825, + "grad_norm": 0.8643395900726318, + "learning_rate": 0.00019100224397441262, + "loss": 0.2924, + "step": 5813 + }, + { + "epoch": 0.13628931591624324, + "grad_norm": 0.3958791196346283, + "learning_rate": 0.00019099919002054776, + "loss": 0.1058, + "step": 5814 + }, + { + "epoch": 0.13631275749104824, + "grad_norm": 1.5487252473831177, + "learning_rate": 0.00019099613557291875, + "loss": 0.1597, + "step": 5815 + }, + { + "epoch": 0.13633619906585323, + "grad_norm": 0.6324335932731628, + "learning_rate": 0.00019099308063154214, + "loss": 0.2094, + "step": 5816 + }, + { + "epoch": 0.13635964064065825, + "grad_norm": 0.6887726187705994, + "learning_rate": 0.00019099002519643452, + "loss": 0.8866, + "step": 5817 + }, + { + "epoch": 0.13638308221546325, + "grad_norm": 0.5746175050735474, + "learning_rate": 0.00019098696926761246, + "loss": 0.0788, + "step": 5818 + }, + { + "epoch": 0.13640652379026824, + "grad_norm": 0.43803244829177856, + "learning_rate": 0.00019098391284509255, + "loss": 0.11, + "step": 5819 + }, + { + "epoch": 0.13642996536507324, + "grad_norm": 0.3880695104598999, + "learning_rate": 0.0001909808559288914, + "loss": 0.0919, + "step": 5820 + }, + { + "epoch": 0.13645340693987823, + "grad_norm": 0.5631226897239685, + "learning_rate": 0.00019097779851902553, + "loss": 0.1336, + "step": 5821 + }, + { + "epoch": 0.13647684851468322, + "grad_norm": 0.3859885036945343, + "learning_rate": 0.00019097474061551158, + "loss": 0.0779, + "step": 5822 + }, + { + "epoch": 0.13650029008948822, + "grad_norm": 0.4395730793476105, + "learning_rate": 0.00019097168221836612, + "loss": 0.1061, + "step": 5823 + }, + { + "epoch": 0.1365237316642932, + "grad_norm": 0.13975702226161957, + "learning_rate": 0.00019096862332760575, + "loss": 0.0252, + "step": 5824 + }, + { + "epoch": 0.1365471732390982, + "grad_norm": 0.5119225382804871, + "learning_rate": 0.0001909655639432471, + "loss": 0.2034, + "step": 5825 + }, + { + "epoch": 0.1365706148139032, + "grad_norm": 0.5414130687713623, + "learning_rate": 0.0001909625040653067, + "loss": 0.8695, + "step": 5826 + }, + { + "epoch": 0.1365940563887082, + "grad_norm": 0.36151519417762756, + "learning_rate": 0.00019095944369380122, + "loss": 0.0827, + "step": 5827 + }, + { + "epoch": 0.1366174979635132, + "grad_norm": 0.5370573401451111, + "learning_rate": 0.00019095638282874724, + "loss": 0.1377, + "step": 5828 + }, + { + "epoch": 0.13664093953831818, + "grad_norm": 0.4978106915950775, + "learning_rate": 0.00019095332147016136, + "loss": 0.1106, + "step": 5829 + }, + { + "epoch": 0.13666438111312318, + "grad_norm": 0.3728780448436737, + "learning_rate": 0.00019095025961806024, + "loss": 0.1425, + "step": 5830 + }, + { + "epoch": 0.13668782268792817, + "grad_norm": 0.6268336772918701, + "learning_rate": 0.0001909471972724604, + "loss": 0.1696, + "step": 5831 + }, + { + "epoch": 0.13671126426273317, + "grad_norm": 0.2885759174823761, + "learning_rate": 0.00019094413443337854, + "loss": 0.0737, + "step": 5832 + }, + { + "epoch": 0.13673470583753816, + "grad_norm": 0.11327942460775375, + "learning_rate": 0.00019094107110083125, + "loss": 0.019, + "step": 5833 + }, + { + "epoch": 0.13675814741234316, + "grad_norm": 0.3291715383529663, + "learning_rate": 0.00019093800727483515, + "loss": 0.1096, + "step": 5834 + }, + { + "epoch": 0.13678158898714815, + "grad_norm": 0.28285688161849976, + "learning_rate": 0.00019093494295540687, + "loss": 0.0581, + "step": 5835 + }, + { + "epoch": 0.13680503056195314, + "grad_norm": 0.6387301683425903, + "learning_rate": 0.000190931878142563, + "loss": 0.1933, + "step": 5836 + }, + { + "epoch": 0.13682847213675814, + "grad_norm": 0.6701120138168335, + "learning_rate": 0.00019092881283632024, + "loss": 0.8527, + "step": 5837 + }, + { + "epoch": 0.13685191371156313, + "grad_norm": 0.8709423542022705, + "learning_rate": 0.00019092574703669516, + "loss": 0.3687, + "step": 5838 + }, + { + "epoch": 0.13687535528636813, + "grad_norm": 0.49759766459465027, + "learning_rate": 0.00019092268074370443, + "loss": 0.8145, + "step": 5839 + }, + { + "epoch": 0.13689879686117312, + "grad_norm": 1.3889025449752808, + "learning_rate": 0.0001909196139573647, + "loss": 0.1383, + "step": 5840 + }, + { + "epoch": 0.13692223843597812, + "grad_norm": 0.23376226425170898, + "learning_rate": 0.00019091654667769255, + "loss": 0.0465, + "step": 5841 + }, + { + "epoch": 0.1369456800107831, + "grad_norm": 0.5918636322021484, + "learning_rate": 0.00019091347890470468, + "loss": 0.1539, + "step": 5842 + }, + { + "epoch": 0.13696912158558813, + "grad_norm": 0.4335293769836426, + "learning_rate": 0.00019091041063841772, + "loss": 0.0846, + "step": 5843 + }, + { + "epoch": 0.13699256316039313, + "grad_norm": 0.3892264664173126, + "learning_rate": 0.00019090734187884832, + "loss": 0.1191, + "step": 5844 + }, + { + "epoch": 0.13701600473519812, + "grad_norm": 0.3140595853328705, + "learning_rate": 0.00019090427262601311, + "loss": 0.1103, + "step": 5845 + }, + { + "epoch": 0.13703944631000312, + "grad_norm": 0.3495230972766876, + "learning_rate": 0.00019090120287992875, + "loss": 0.0667, + "step": 5846 + }, + { + "epoch": 0.1370628878848081, + "grad_norm": 0.1581612229347229, + "learning_rate": 0.00019089813264061193, + "loss": 0.025, + "step": 5847 + }, + { + "epoch": 0.1370863294596131, + "grad_norm": 0.41234302520751953, + "learning_rate": 0.00019089506190807928, + "loss": 0.0864, + "step": 5848 + }, + { + "epoch": 0.1371097710344181, + "grad_norm": 0.5266339182853699, + "learning_rate": 0.0001908919906823475, + "loss": 0.1712, + "step": 5849 + }, + { + "epoch": 0.1371332126092231, + "grad_norm": 0.6489524245262146, + "learning_rate": 0.00019088891896343316, + "loss": 0.1502, + "step": 5850 + }, + { + "epoch": 0.1371566541840281, + "grad_norm": 0.7199462652206421, + "learning_rate": 0.00019088584675135305, + "loss": 0.2464, + "step": 5851 + }, + { + "epoch": 0.13718009575883308, + "grad_norm": 0.21045023202896118, + "learning_rate": 0.00019088277404612374, + "loss": 0.0279, + "step": 5852 + }, + { + "epoch": 0.13720353733363808, + "grad_norm": 0.7560389041900635, + "learning_rate": 0.00019087970084776196, + "loss": 0.4278, + "step": 5853 + }, + { + "epoch": 0.13722697890844307, + "grad_norm": 0.228938028216362, + "learning_rate": 0.00019087662715628438, + "loss": 0.065, + "step": 5854 + }, + { + "epoch": 0.13725042048324806, + "grad_norm": 0.40723785758018494, + "learning_rate": 0.00019087355297170764, + "loss": 0.0508, + "step": 5855 + }, + { + "epoch": 0.13727386205805306, + "grad_norm": 0.4404120147228241, + "learning_rate": 0.00019087047829404846, + "loss": 0.1081, + "step": 5856 + }, + { + "epoch": 0.13729730363285805, + "grad_norm": 0.23387819528579712, + "learning_rate": 0.0001908674031233235, + "loss": 0.0543, + "step": 5857 + }, + { + "epoch": 0.13732074520766305, + "grad_norm": 0.5464262366294861, + "learning_rate": 0.00019086432745954945, + "loss": 0.1596, + "step": 5858 + }, + { + "epoch": 0.13734418678246804, + "grad_norm": 0.5167922973632812, + "learning_rate": 0.000190861251302743, + "loss": 0.1304, + "step": 5859 + }, + { + "epoch": 0.13736762835727304, + "grad_norm": 0.5791605114936829, + "learning_rate": 0.0001908581746529209, + "loss": 0.8643, + "step": 5860 + }, + { + "epoch": 0.13739106993207803, + "grad_norm": 0.34157493710517883, + "learning_rate": 0.00019085509751009974, + "loss": 0.1108, + "step": 5861 + }, + { + "epoch": 0.13741451150688302, + "grad_norm": 0.6169381141662598, + "learning_rate": 0.00019085201987429628, + "loss": 0.1901, + "step": 5862 + }, + { + "epoch": 0.13743795308168802, + "grad_norm": 0.12844599783420563, + "learning_rate": 0.0001908489417455272, + "loss": 0.0285, + "step": 5863 + }, + { + "epoch": 0.137461394656493, + "grad_norm": 0.6162275075912476, + "learning_rate": 0.00019084586312380922, + "loss": 0.1701, + "step": 5864 + }, + { + "epoch": 0.137484836231298, + "grad_norm": 0.834966242313385, + "learning_rate": 0.000190842784009159, + "loss": 0.1584, + "step": 5865 + }, + { + "epoch": 0.137508277806103, + "grad_norm": 0.19390861690044403, + "learning_rate": 0.0001908397044015933, + "loss": 0.0695, + "step": 5866 + }, + { + "epoch": 0.137531719380908, + "grad_norm": 0.6264896988868713, + "learning_rate": 0.00019083662430112883, + "loss": 0.1181, + "step": 5867 + }, + { + "epoch": 0.13755516095571302, + "grad_norm": 0.4555320739746094, + "learning_rate": 0.00019083354370778225, + "loss": 0.1397, + "step": 5868 + }, + { + "epoch": 0.137578602530518, + "grad_norm": 0.5788241624832153, + "learning_rate": 0.00019083046262157033, + "loss": 0.628, + "step": 5869 + }, + { + "epoch": 0.137602044105323, + "grad_norm": 0.36689096689224243, + "learning_rate": 0.00019082738104250977, + "loss": 0.1118, + "step": 5870 + }, + { + "epoch": 0.137625485680128, + "grad_norm": 0.5326579809188843, + "learning_rate": 0.00019082429897061725, + "loss": 0.0775, + "step": 5871 + }, + { + "epoch": 0.137648927254933, + "grad_norm": 0.4426817297935486, + "learning_rate": 0.00019082121640590958, + "loss": 0.1211, + "step": 5872 + }, + { + "epoch": 0.137672368829738, + "grad_norm": 0.5002864599227905, + "learning_rate": 0.0001908181333484034, + "loss": 0.0908, + "step": 5873 + }, + { + "epoch": 0.13769581040454298, + "grad_norm": 0.6866104602813721, + "learning_rate": 0.0001908150497981155, + "loss": 0.165, + "step": 5874 + }, + { + "epoch": 0.13771925197934798, + "grad_norm": 0.10062633454799652, + "learning_rate": 0.00019081196575506258, + "loss": 0.0224, + "step": 5875 + }, + { + "epoch": 0.13774269355415297, + "grad_norm": 0.7788085341453552, + "learning_rate": 0.00019080888121926137, + "loss": 0.133, + "step": 5876 + }, + { + "epoch": 0.13776613512895797, + "grad_norm": 0.3949412405490875, + "learning_rate": 0.00019080579619072864, + "loss": 0.1043, + "step": 5877 + }, + { + "epoch": 0.13778957670376296, + "grad_norm": 0.6746712327003479, + "learning_rate": 0.00019080271066948106, + "loss": 0.1718, + "step": 5878 + }, + { + "epoch": 0.13781301827856796, + "grad_norm": 0.8888312578201294, + "learning_rate": 0.00019079962465553545, + "loss": 0.1899, + "step": 5879 + }, + { + "epoch": 0.13783645985337295, + "grad_norm": 0.42366087436676025, + "learning_rate": 0.0001907965381489085, + "loss": 0.417, + "step": 5880 + }, + { + "epoch": 0.13785990142817794, + "grad_norm": 0.39104992151260376, + "learning_rate": 0.00019079345114961701, + "loss": 0.1046, + "step": 5881 + }, + { + "epoch": 0.13788334300298294, + "grad_norm": 0.19133567810058594, + "learning_rate": 0.0001907903636576777, + "loss": 0.0424, + "step": 5882 + }, + { + "epoch": 0.13790678457778793, + "grad_norm": 0.5725039839744568, + "learning_rate": 0.00019078727567310732, + "loss": 0.1267, + "step": 5883 + }, + { + "epoch": 0.13793022615259293, + "grad_norm": 0.5737932324409485, + "learning_rate": 0.0001907841871959226, + "loss": 0.1933, + "step": 5884 + }, + { + "epoch": 0.13795366772739792, + "grad_norm": 0.239521324634552, + "learning_rate": 0.00019078109822614039, + "loss": 0.0563, + "step": 5885 + }, + { + "epoch": 0.13797710930220292, + "grad_norm": 0.15178048610687256, + "learning_rate": 0.00019077800876377734, + "loss": 0.031, + "step": 5886 + }, + { + "epoch": 0.1380005508770079, + "grad_norm": 0.3964998722076416, + "learning_rate": 0.00019077491880885027, + "loss": 0.1232, + "step": 5887 + }, + { + "epoch": 0.1380239924518129, + "grad_norm": 0.5672482848167419, + "learning_rate": 0.00019077182836137595, + "loss": 0.2222, + "step": 5888 + }, + { + "epoch": 0.1380474340266179, + "grad_norm": 0.49362248182296753, + "learning_rate": 0.00019076873742137113, + "loss": 0.699, + "step": 5889 + }, + { + "epoch": 0.1380708756014229, + "grad_norm": 0.5005885362625122, + "learning_rate": 0.00019076564598885257, + "loss": 0.0831, + "step": 5890 + }, + { + "epoch": 0.1380943171762279, + "grad_norm": 0.3838890492916107, + "learning_rate": 0.0001907625540638371, + "loss": 0.1286, + "step": 5891 + }, + { + "epoch": 0.13811775875103288, + "grad_norm": 0.5262749195098877, + "learning_rate": 0.00019075946164634141, + "loss": 0.1234, + "step": 5892 + }, + { + "epoch": 0.13814120032583788, + "grad_norm": 0.2457820475101471, + "learning_rate": 0.00019075636873638237, + "loss": 0.0725, + "step": 5893 + }, + { + "epoch": 0.1381646419006429, + "grad_norm": 0.17641395330429077, + "learning_rate": 0.00019075327533397672, + "loss": 0.0439, + "step": 5894 + }, + { + "epoch": 0.1381880834754479, + "grad_norm": 0.31645822525024414, + "learning_rate": 0.00019075018143914125, + "loss": 0.055, + "step": 5895 + }, + { + "epoch": 0.1382115250502529, + "grad_norm": 0.32229670882225037, + "learning_rate": 0.00019074708705189272, + "loss": 0.0796, + "step": 5896 + }, + { + "epoch": 0.13823496662505788, + "grad_norm": 0.5393668413162231, + "learning_rate": 0.00019074399217224795, + "loss": 0.1626, + "step": 5897 + }, + { + "epoch": 0.13825840819986288, + "grad_norm": 0.10499467700719833, + "learning_rate": 0.00019074089680022374, + "loss": 0.0267, + "step": 5898 + }, + { + "epoch": 0.13828184977466787, + "grad_norm": 0.3126498758792877, + "learning_rate": 0.00019073780093583685, + "loss": 0.0809, + "step": 5899 + }, + { + "epoch": 0.13830529134947286, + "grad_norm": 0.9921278953552246, + "learning_rate": 0.00019073470457910412, + "loss": 0.2257, + "step": 5900 + }, + { + "epoch": 0.13832873292427786, + "grad_norm": 0.10635191947221756, + "learning_rate": 0.00019073160773004232, + "loss": 0.0182, + "step": 5901 + }, + { + "epoch": 0.13835217449908285, + "grad_norm": 0.8243650794029236, + "learning_rate": 0.00019072851038866827, + "loss": 0.1688, + "step": 5902 + }, + { + "epoch": 0.13837561607388785, + "grad_norm": 0.40251824259757996, + "learning_rate": 0.00019072541255499878, + "loss": 0.0965, + "step": 5903 + }, + { + "epoch": 0.13839905764869284, + "grad_norm": 0.43564048409461975, + "learning_rate": 0.00019072231422905065, + "loss": 0.5758, + "step": 5904 + }, + { + "epoch": 0.13842249922349784, + "grad_norm": 0.44088214635849, + "learning_rate": 0.00019071921541084067, + "loss": 0.0392, + "step": 5905 + }, + { + "epoch": 0.13844594079830283, + "grad_norm": 0.6327665448188782, + "learning_rate": 0.0001907161161003857, + "loss": 0.1529, + "step": 5906 + }, + { + "epoch": 0.13846938237310782, + "grad_norm": 0.4340084195137024, + "learning_rate": 0.00019071301629770254, + "loss": 0.1601, + "step": 5907 + }, + { + "epoch": 0.13849282394791282, + "grad_norm": 0.4565422832965851, + "learning_rate": 0.00019070991600280798, + "loss": 0.0999, + "step": 5908 + }, + { + "epoch": 0.1385162655227178, + "grad_norm": 0.6055148839950562, + "learning_rate": 0.00019070681521571886, + "loss": 0.1944, + "step": 5909 + }, + { + "epoch": 0.1385397070975228, + "grad_norm": 0.23459076881408691, + "learning_rate": 0.00019070371393645202, + "loss": 0.0443, + "step": 5910 + }, + { + "epoch": 0.1385631486723278, + "grad_norm": 0.1433952897787094, + "learning_rate": 0.0001907006121650243, + "loss": 0.0391, + "step": 5911 + }, + { + "epoch": 0.1385865902471328, + "grad_norm": 0.19088266789913177, + "learning_rate": 0.00019069750990145248, + "loss": 0.0385, + "step": 5912 + }, + { + "epoch": 0.1386100318219378, + "grad_norm": 0.5967620611190796, + "learning_rate": 0.00019069440714575343, + "loss": 0.1784, + "step": 5913 + }, + { + "epoch": 0.13863347339674278, + "grad_norm": 0.5451310276985168, + "learning_rate": 0.000190691303897944, + "loss": 0.0888, + "step": 5914 + }, + { + "epoch": 0.13865691497154778, + "grad_norm": 0.558199405670166, + "learning_rate": 0.00019068820015804097, + "loss": 0.1135, + "step": 5915 + }, + { + "epoch": 0.13868035654635277, + "grad_norm": 0.6118086576461792, + "learning_rate": 0.00019068509592606124, + "loss": 0.8574, + "step": 5916 + }, + { + "epoch": 0.13870379812115777, + "grad_norm": 0.1568400114774704, + "learning_rate": 0.00019068199120202158, + "loss": 0.0525, + "step": 5917 + }, + { + "epoch": 0.13872723969596276, + "grad_norm": 0.5225921273231506, + "learning_rate": 0.00019067888598593894, + "loss": 0.151, + "step": 5918 + }, + { + "epoch": 0.13875068127076778, + "grad_norm": 0.574431836605072, + "learning_rate": 0.0001906757802778301, + "loss": 0.0967, + "step": 5919 + }, + { + "epoch": 0.13877412284557278, + "grad_norm": 0.5210288763046265, + "learning_rate": 0.0001906726740777119, + "loss": 0.1091, + "step": 5920 + }, + { + "epoch": 0.13879756442037777, + "grad_norm": 0.34316006302833557, + "learning_rate": 0.00019066956738560128, + "loss": 0.1088, + "step": 5921 + }, + { + "epoch": 0.13882100599518277, + "grad_norm": 0.6204622983932495, + "learning_rate": 0.000190666460201515, + "loss": 0.1491, + "step": 5922 + }, + { + "epoch": 0.13884444756998776, + "grad_norm": 0.1885031908750534, + "learning_rate": 0.00019066335252546996, + "loss": 0.0738, + "step": 5923 + }, + { + "epoch": 0.13886788914479276, + "grad_norm": 0.7449998259544373, + "learning_rate": 0.00019066024435748303, + "loss": 0.2548, + "step": 5924 + }, + { + "epoch": 0.13889133071959775, + "grad_norm": 1.6117733716964722, + "learning_rate": 0.00019065713569757105, + "loss": 0.1809, + "step": 5925 + }, + { + "epoch": 0.13891477229440274, + "grad_norm": 0.6252404451370239, + "learning_rate": 0.0001906540265457509, + "loss": 0.7168, + "step": 5926 + }, + { + "epoch": 0.13893821386920774, + "grad_norm": 1.4032604694366455, + "learning_rate": 0.00019065091690203948, + "loss": 0.117, + "step": 5927 + }, + { + "epoch": 0.13896165544401273, + "grad_norm": 0.938271164894104, + "learning_rate": 0.00019064780676645362, + "loss": 0.2199, + "step": 5928 + }, + { + "epoch": 0.13898509701881773, + "grad_norm": 0.6909931302070618, + "learning_rate": 0.00019064469613901022, + "loss": 0.1787, + "step": 5929 + }, + { + "epoch": 0.13900853859362272, + "grad_norm": 0.23576028645038605, + "learning_rate": 0.00019064158501972612, + "loss": 0.0873, + "step": 5930 + }, + { + "epoch": 0.13903198016842772, + "grad_norm": 0.632098376750946, + "learning_rate": 0.00019063847340861827, + "loss": 0.1644, + "step": 5931 + }, + { + "epoch": 0.1390554217432327, + "grad_norm": 0.6682330369949341, + "learning_rate": 0.0001906353613057035, + "loss": 0.1855, + "step": 5932 + }, + { + "epoch": 0.1390788633180377, + "grad_norm": 0.13165231049060822, + "learning_rate": 0.0001906322487109987, + "loss": 0.024, + "step": 5933 + }, + { + "epoch": 0.1391023048928427, + "grad_norm": 0.6565446853637695, + "learning_rate": 0.00019062913562452078, + "loss": 0.196, + "step": 5934 + }, + { + "epoch": 0.1391257464676477, + "grad_norm": 0.1576703041791916, + "learning_rate": 0.00019062602204628664, + "loss": 0.0319, + "step": 5935 + }, + { + "epoch": 0.1391491880424527, + "grad_norm": 0.13507702946662903, + "learning_rate": 0.0001906229079763131, + "loss": 0.0519, + "step": 5936 + }, + { + "epoch": 0.13917262961725768, + "grad_norm": 0.38944175839424133, + "learning_rate": 0.00019061979341461718, + "loss": 0.1128, + "step": 5937 + }, + { + "epoch": 0.13919607119206268, + "grad_norm": 0.6035330295562744, + "learning_rate": 0.00019061667836121565, + "loss": 0.1379, + "step": 5938 + }, + { + "epoch": 0.13921951276686767, + "grad_norm": 0.4714129865169525, + "learning_rate": 0.00019061356281612553, + "loss": 0.1875, + "step": 5939 + }, + { + "epoch": 0.13924295434167266, + "grad_norm": 0.26230522990226746, + "learning_rate": 0.00019061044677936363, + "loss": 0.0516, + "step": 5940 + }, + { + "epoch": 0.13926639591647766, + "grad_norm": 0.6051207184791565, + "learning_rate": 0.00019060733025094692, + "loss": 0.1456, + "step": 5941 + }, + { + "epoch": 0.13928983749128265, + "grad_norm": 0.6942124366760254, + "learning_rate": 0.0001906042132308923, + "loss": 0.2363, + "step": 5942 + }, + { + "epoch": 0.13931327906608765, + "grad_norm": 0.6119300127029419, + "learning_rate": 0.00019060109571921664, + "loss": 0.4835, + "step": 5943 + }, + { + "epoch": 0.13933672064089264, + "grad_norm": 0.34956756234169006, + "learning_rate": 0.00019059797771593692, + "loss": 0.0848, + "step": 5944 + }, + { + "epoch": 0.13936016221569766, + "grad_norm": 0.27364400029182434, + "learning_rate": 0.00019059485922107, + "loss": 0.0868, + "step": 5945 + }, + { + "epoch": 0.13938360379050266, + "grad_norm": 0.7796478271484375, + "learning_rate": 0.00019059174023463283, + "loss": 0.2106, + "step": 5946 + }, + { + "epoch": 0.13940704536530765, + "grad_norm": 0.4336390495300293, + "learning_rate": 0.00019058862075664235, + "loss": 0.4976, + "step": 5947 + }, + { + "epoch": 0.13943048694011265, + "grad_norm": 0.6770291924476624, + "learning_rate": 0.00019058550078711542, + "loss": 0.1549, + "step": 5948 + }, + { + "epoch": 0.13945392851491764, + "grad_norm": 0.8904136419296265, + "learning_rate": 0.00019058238032606907, + "loss": 0.2641, + "step": 5949 + }, + { + "epoch": 0.13947737008972264, + "grad_norm": 1.061906099319458, + "learning_rate": 0.00019057925937352016, + "loss": 0.2666, + "step": 5950 + }, + { + "epoch": 0.13950081166452763, + "grad_norm": 0.5532277226448059, + "learning_rate": 0.00019057613792948562, + "loss": 0.1404, + "step": 5951 + }, + { + "epoch": 0.13952425323933262, + "grad_norm": 0.3995380103588104, + "learning_rate": 0.00019057301599398243, + "loss": 0.1299, + "step": 5952 + }, + { + "epoch": 0.13954769481413762, + "grad_norm": 0.5907908082008362, + "learning_rate": 0.00019056989356702752, + "loss": 0.0907, + "step": 5953 + }, + { + "epoch": 0.1395711363889426, + "grad_norm": 0.37709859013557434, + "learning_rate": 0.0001905667706486378, + "loss": 0.0663, + "step": 5954 + }, + { + "epoch": 0.1395945779637476, + "grad_norm": 0.4518117606639862, + "learning_rate": 0.00019056364723883022, + "loss": 0.1529, + "step": 5955 + }, + { + "epoch": 0.1396180195385526, + "grad_norm": 0.13570600748062134, + "learning_rate": 0.00019056052333762177, + "loss": 0.0323, + "step": 5956 + }, + { + "epoch": 0.1396414611133576, + "grad_norm": 0.7283958196640015, + "learning_rate": 0.00019055739894502938, + "loss": 0.1915, + "step": 5957 + }, + { + "epoch": 0.1396649026881626, + "grad_norm": 0.9026922583580017, + "learning_rate": 0.00019055427406106999, + "loss": 0.2866, + "step": 5958 + }, + { + "epoch": 0.13968834426296758, + "grad_norm": 0.5810667276382446, + "learning_rate": 0.00019055114868576055, + "loss": 0.1837, + "step": 5959 + }, + { + "epoch": 0.13971178583777258, + "grad_norm": 0.3030453026294708, + "learning_rate": 0.00019054802281911803, + "loss": 0.0614, + "step": 5960 + }, + { + "epoch": 0.13973522741257757, + "grad_norm": 0.12709268927574158, + "learning_rate": 0.0001905448964611594, + "loss": 0.0335, + "step": 5961 + }, + { + "epoch": 0.13975866898738257, + "grad_norm": 0.566169261932373, + "learning_rate": 0.00019054176961190162, + "loss": 0.1068, + "step": 5962 + }, + { + "epoch": 0.13978211056218756, + "grad_norm": 0.4237300455570221, + "learning_rate": 0.00019053864227136166, + "loss": 0.1268, + "step": 5963 + }, + { + "epoch": 0.13980555213699256, + "grad_norm": 0.23756545782089233, + "learning_rate": 0.00019053551443955647, + "loss": 0.0723, + "step": 5964 + }, + { + "epoch": 0.13982899371179755, + "grad_norm": 0.4589858055114746, + "learning_rate": 0.00019053238611650304, + "loss": 0.1844, + "step": 5965 + }, + { + "epoch": 0.13985243528660254, + "grad_norm": 0.45786428451538086, + "learning_rate": 0.00019052925730221834, + "loss": 0.5479, + "step": 5966 + }, + { + "epoch": 0.13987587686140754, + "grad_norm": 0.2094022035598755, + "learning_rate": 0.00019052612799671931, + "loss": 0.0544, + "step": 5967 + }, + { + "epoch": 0.13989931843621253, + "grad_norm": 0.3926270604133606, + "learning_rate": 0.000190522998200023, + "loss": 0.1046, + "step": 5968 + }, + { + "epoch": 0.13992276001101753, + "grad_norm": 0.5151132941246033, + "learning_rate": 0.00019051986791214636, + "loss": 0.1065, + "step": 5969 + }, + { + "epoch": 0.13994620158582255, + "grad_norm": 0.46460169553756714, + "learning_rate": 0.00019051673713310635, + "loss": 0.101, + "step": 5970 + }, + { + "epoch": 0.13996964316062754, + "grad_norm": 0.9042111039161682, + "learning_rate": 0.00019051360586291998, + "loss": 0.69, + "step": 5971 + }, + { + "epoch": 0.13999308473543254, + "grad_norm": 0.11823637783527374, + "learning_rate": 0.00019051047410160427, + "loss": 0.0467, + "step": 5972 + }, + { + "epoch": 0.14001652631023753, + "grad_norm": 0.5509101152420044, + "learning_rate": 0.00019050734184917616, + "loss": 0.1888, + "step": 5973 + }, + { + "epoch": 0.14003996788504253, + "grad_norm": 0.5545378923416138, + "learning_rate": 0.00019050420910565267, + "loss": 0.1588, + "step": 5974 + }, + { + "epoch": 0.14006340945984752, + "grad_norm": 0.2487071454524994, + "learning_rate": 0.00019050107587105077, + "loss": 0.0525, + "step": 5975 + }, + { + "epoch": 0.14008685103465252, + "grad_norm": 0.23638202250003815, + "learning_rate": 0.00019049794214538753, + "loss": 0.0647, + "step": 5976 + }, + { + "epoch": 0.1401102926094575, + "grad_norm": 0.5138981938362122, + "learning_rate": 0.00019049480792867989, + "loss": 0.0912, + "step": 5977 + }, + { + "epoch": 0.1401337341842625, + "grad_norm": 0.5661172270774841, + "learning_rate": 0.00019049167322094488, + "loss": 0.1365, + "step": 5978 + }, + { + "epoch": 0.1401571757590675, + "grad_norm": 0.3106982111930847, + "learning_rate": 0.00019048853802219952, + "loss": 0.0859, + "step": 5979 + }, + { + "epoch": 0.1401806173338725, + "grad_norm": 0.3399451673030853, + "learning_rate": 0.00019048540233246078, + "loss": 0.0657, + "step": 5980 + }, + { + "epoch": 0.1402040589086775, + "grad_norm": 0.24899634718894958, + "learning_rate": 0.00019048226615174573, + "loss": 0.0951, + "step": 5981 + }, + { + "epoch": 0.14022750048348248, + "grad_norm": 0.5489285588264465, + "learning_rate": 0.00019047912948007132, + "loss": 0.1497, + "step": 5982 + }, + { + "epoch": 0.14025094205828748, + "grad_norm": 1.1069862842559814, + "learning_rate": 0.00019047599231745464, + "loss": 0.33, + "step": 5983 + }, + { + "epoch": 0.14027438363309247, + "grad_norm": 0.45669814944267273, + "learning_rate": 0.00019047285466391268, + "loss": 0.0997, + "step": 5984 + }, + { + "epoch": 0.14029782520789746, + "grad_norm": 0.7244880199432373, + "learning_rate": 0.0001904697165194625, + "loss": 0.1134, + "step": 5985 + }, + { + "epoch": 0.14032126678270246, + "grad_norm": 0.22668863832950592, + "learning_rate": 0.00019046657788412104, + "loss": 0.0154, + "step": 5986 + }, + { + "epoch": 0.14034470835750745, + "grad_norm": 0.6536818146705627, + "learning_rate": 0.0001904634387579054, + "loss": 0.6511, + "step": 5987 + }, + { + "epoch": 0.14036814993231245, + "grad_norm": 1.044212818145752, + "learning_rate": 0.00019046029914083262, + "loss": 0.2403, + "step": 5988 + }, + { + "epoch": 0.14039159150711744, + "grad_norm": 0.3136846721172333, + "learning_rate": 0.0001904571590329197, + "loss": 0.049, + "step": 5989 + }, + { + "epoch": 0.14041503308192244, + "grad_norm": 0.6150447726249695, + "learning_rate": 0.00019045401843418368, + "loss": 0.2305, + "step": 5990 + }, + { + "epoch": 0.14043847465672743, + "grad_norm": 0.7492564916610718, + "learning_rate": 0.00019045087734464162, + "loss": 0.2217, + "step": 5991 + }, + { + "epoch": 0.14046191623153242, + "grad_norm": 0.5889996290206909, + "learning_rate": 0.00019044773576431054, + "loss": 0.1446, + "step": 5992 + }, + { + "epoch": 0.14048535780633742, + "grad_norm": 0.7264955639839172, + "learning_rate": 0.00019044459369320753, + "loss": 0.3295, + "step": 5993 + }, + { + "epoch": 0.1405087993811424, + "grad_norm": 0.9205420613288879, + "learning_rate": 0.0001904414511313496, + "loss": 0.1521, + "step": 5994 + }, + { + "epoch": 0.1405322409559474, + "grad_norm": 0.45761173963546753, + "learning_rate": 0.00019043830807875383, + "loss": 0.1392, + "step": 5995 + }, + { + "epoch": 0.14055568253075243, + "grad_norm": 0.41176339983940125, + "learning_rate": 0.00019043516453543725, + "loss": 0.0832, + "step": 5996 + }, + { + "epoch": 0.14057912410555742, + "grad_norm": 0.413678377866745, + "learning_rate": 0.0001904320205014169, + "loss": 0.1401, + "step": 5997 + }, + { + "epoch": 0.14060256568036242, + "grad_norm": 0.5828797221183777, + "learning_rate": 0.0001904288759767099, + "loss": 0.1691, + "step": 5998 + }, + { + "epoch": 0.1406260072551674, + "grad_norm": 0.1512313038110733, + "learning_rate": 0.00019042573096133322, + "loss": 0.0346, + "step": 5999 + }, + { + "epoch": 0.1406494488299724, + "grad_norm": 0.38205191493034363, + "learning_rate": 0.00019042258545530402, + "loss": 0.1559, + "step": 6000 + }, + { + "epoch": 0.1406728904047774, + "grad_norm": 0.34872040152549744, + "learning_rate": 0.00019041943945863934, + "loss": 0.0787, + "step": 6001 + }, + { + "epoch": 0.1406963319795824, + "grad_norm": 0.23532089591026306, + "learning_rate": 0.00019041629297135621, + "loss": 0.0624, + "step": 6002 + }, + { + "epoch": 0.1407197735543874, + "grad_norm": 0.34757572412490845, + "learning_rate": 0.00019041314599347174, + "loss": 0.0614, + "step": 6003 + }, + { + "epoch": 0.14074321512919238, + "grad_norm": 0.5670728087425232, + "learning_rate": 0.00019040999852500302, + "loss": 0.1138, + "step": 6004 + }, + { + "epoch": 0.14076665670399738, + "grad_norm": 0.5372018814086914, + "learning_rate": 0.0001904068505659671, + "loss": 0.1756, + "step": 6005 + }, + { + "epoch": 0.14079009827880237, + "grad_norm": 0.38775765895843506, + "learning_rate": 0.00019040370211638105, + "loss": 0.1176, + "step": 6006 + }, + { + "epoch": 0.14081353985360737, + "grad_norm": 0.6348205804824829, + "learning_rate": 0.000190400553176262, + "loss": 0.2635, + "step": 6007 + }, + { + "epoch": 0.14083698142841236, + "grad_norm": 0.15677668154239655, + "learning_rate": 0.00019039740374562696, + "loss": 0.033, + "step": 6008 + }, + { + "epoch": 0.14086042300321736, + "grad_norm": 0.2784712016582489, + "learning_rate": 0.0001903942538244931, + "loss": 0.0786, + "step": 6009 + }, + { + "epoch": 0.14088386457802235, + "grad_norm": 0.20976071059703827, + "learning_rate": 0.00019039110341287748, + "loss": 0.0495, + "step": 6010 + }, + { + "epoch": 0.14090730615282734, + "grad_norm": 0.16243800520896912, + "learning_rate": 0.00019038795251079718, + "loss": 0.0194, + "step": 6011 + }, + { + "epoch": 0.14093074772763234, + "grad_norm": 0.5783772468566895, + "learning_rate": 0.0001903848011182693, + "loss": 0.1441, + "step": 6012 + }, + { + "epoch": 0.14095418930243733, + "grad_norm": 0.31369638442993164, + "learning_rate": 0.00019038164923531093, + "loss": 0.0718, + "step": 6013 + }, + { + "epoch": 0.14097763087724233, + "grad_norm": 0.6048810482025146, + "learning_rate": 0.00019037849686193923, + "loss": 0.0803, + "step": 6014 + }, + { + "epoch": 0.14100107245204732, + "grad_norm": 0.47051918506622314, + "learning_rate": 0.00019037534399817126, + "loss": 0.1388, + "step": 6015 + }, + { + "epoch": 0.14102451402685232, + "grad_norm": 0.43853116035461426, + "learning_rate": 0.00019037219064402413, + "loss": 0.0835, + "step": 6016 + }, + { + "epoch": 0.1410479556016573, + "grad_norm": 1.1131490468978882, + "learning_rate": 0.00019036903679951493, + "loss": 0.2278, + "step": 6017 + }, + { + "epoch": 0.1410713971764623, + "grad_norm": 0.15103617310523987, + "learning_rate": 0.00019036588246466082, + "loss": 0.0285, + "step": 6018 + }, + { + "epoch": 0.1410948387512673, + "grad_norm": 0.5937040448188782, + "learning_rate": 0.00019036272763947885, + "loss": 0.146, + "step": 6019 + }, + { + "epoch": 0.1411182803260723, + "grad_norm": 0.8148579597473145, + "learning_rate": 0.00019035957232398624, + "loss": 0.2735, + "step": 6020 + }, + { + "epoch": 0.1411417219008773, + "grad_norm": 0.5390664935112, + "learning_rate": 0.0001903564165182, + "loss": 0.1477, + "step": 6021 + }, + { + "epoch": 0.1411651634756823, + "grad_norm": 0.6568225622177124, + "learning_rate": 0.00019035326022213736, + "loss": 0.1376, + "step": 6022 + }, + { + "epoch": 0.1411886050504873, + "grad_norm": 0.24856336414813995, + "learning_rate": 0.00019035010343581532, + "loss": 0.044, + "step": 6023 + }, + { + "epoch": 0.1412120466252923, + "grad_norm": 0.5297638773918152, + "learning_rate": 0.00019034694615925114, + "loss": 0.1496, + "step": 6024 + }, + { + "epoch": 0.1412354882000973, + "grad_norm": 0.4437825083732605, + "learning_rate": 0.00019034378839246182, + "loss": 0.1544, + "step": 6025 + }, + { + "epoch": 0.1412589297749023, + "grad_norm": 0.3374846875667572, + "learning_rate": 0.0001903406301354646, + "loss": 0.0531, + "step": 6026 + }, + { + "epoch": 0.14128237134970728, + "grad_norm": 0.4657328724861145, + "learning_rate": 0.00019033747138827662, + "loss": 0.0828, + "step": 6027 + }, + { + "epoch": 0.14130581292451228, + "grad_norm": 0.21836675703525543, + "learning_rate": 0.00019033431215091496, + "loss": 0.0583, + "step": 6028 + }, + { + "epoch": 0.14132925449931727, + "grad_norm": 0.6398928165435791, + "learning_rate": 0.00019033115242339673, + "loss": 0.0483, + "step": 6029 + }, + { + "epoch": 0.14135269607412226, + "grad_norm": 0.7566096186637878, + "learning_rate": 0.00019032799220573918, + "loss": 0.2315, + "step": 6030 + }, + { + "epoch": 0.14137613764892726, + "grad_norm": 0.9923942685127258, + "learning_rate": 0.00019032483149795942, + "loss": 0.2147, + "step": 6031 + }, + { + "epoch": 0.14139957922373225, + "grad_norm": 1.0274882316589355, + "learning_rate": 0.00019032167030007453, + "loss": 0.1973, + "step": 6032 + }, + { + "epoch": 0.14142302079853725, + "grad_norm": 0.5555745363235474, + "learning_rate": 0.00019031850861210174, + "loss": 0.1931, + "step": 6033 + }, + { + "epoch": 0.14144646237334224, + "grad_norm": 0.5712507963180542, + "learning_rate": 0.00019031534643405818, + "loss": 0.1695, + "step": 6034 + }, + { + "epoch": 0.14146990394814724, + "grad_norm": 0.681348979473114, + "learning_rate": 0.00019031218376596102, + "loss": 0.1838, + "step": 6035 + }, + { + "epoch": 0.14149334552295223, + "grad_norm": 0.6832557320594788, + "learning_rate": 0.0001903090206078274, + "loss": 0.2324, + "step": 6036 + }, + { + "epoch": 0.14151678709775722, + "grad_norm": 0.4764489531517029, + "learning_rate": 0.00019030585695967452, + "loss": 0.5949, + "step": 6037 + }, + { + "epoch": 0.14154022867256222, + "grad_norm": 0.23773548007011414, + "learning_rate": 0.0001903026928215195, + "loss": 0.0479, + "step": 6038 + }, + { + "epoch": 0.1415636702473672, + "grad_norm": 0.5477513670921326, + "learning_rate": 0.00019029952819337953, + "loss": 0.1738, + "step": 6039 + }, + { + "epoch": 0.1415871118221722, + "grad_norm": 0.3958067297935486, + "learning_rate": 0.0001902963630752718, + "loss": 0.0753, + "step": 6040 + }, + { + "epoch": 0.1416105533969772, + "grad_norm": 0.7061850428581238, + "learning_rate": 0.00019029319746721344, + "loss": 0.231, + "step": 6041 + }, + { + "epoch": 0.1416339949717822, + "grad_norm": 0.3707088828086853, + "learning_rate": 0.00019029003136922164, + "loss": 0.0874, + "step": 6042 + }, + { + "epoch": 0.1416574365465872, + "grad_norm": 0.3262605667114258, + "learning_rate": 0.00019028686478131363, + "loss": 0.0771, + "step": 6043 + }, + { + "epoch": 0.14168087812139218, + "grad_norm": 0.4688214957714081, + "learning_rate": 0.0001902836977035065, + "loss": 0.1286, + "step": 6044 + }, + { + "epoch": 0.14170431969619718, + "grad_norm": 0.7138878703117371, + "learning_rate": 0.00019028053013581752, + "loss": 0.2068, + "step": 6045 + }, + { + "epoch": 0.14172776127100217, + "grad_norm": 0.4075721800327301, + "learning_rate": 0.00019027736207826386, + "loss": 0.1192, + "step": 6046 + }, + { + "epoch": 0.1417512028458072, + "grad_norm": 0.3024711310863495, + "learning_rate": 0.00019027419353086268, + "loss": 0.3642, + "step": 6047 + }, + { + "epoch": 0.1417746444206122, + "grad_norm": 0.4835458993911743, + "learning_rate": 0.0001902710244936312, + "loss": 0.1212, + "step": 6048 + }, + { + "epoch": 0.14179808599541718, + "grad_norm": 0.724997878074646, + "learning_rate": 0.00019026785496658658, + "loss": 0.2212, + "step": 6049 + }, + { + "epoch": 0.14182152757022218, + "grad_norm": 0.7062482237815857, + "learning_rate": 0.00019026468494974608, + "loss": 0.1753, + "step": 6050 + }, + { + "epoch": 0.14184496914502717, + "grad_norm": 0.6878710389137268, + "learning_rate": 0.00019026151444312682, + "loss": 0.584, + "step": 6051 + }, + { + "epoch": 0.14186841071983217, + "grad_norm": 0.2755690813064575, + "learning_rate": 0.00019025834344674607, + "loss": 0.0989, + "step": 6052 + }, + { + "epoch": 0.14189185229463716, + "grad_norm": 0.2608139216899872, + "learning_rate": 0.000190255171960621, + "loss": 0.0748, + "step": 6053 + }, + { + "epoch": 0.14191529386944216, + "grad_norm": 0.6851416826248169, + "learning_rate": 0.00019025199998476882, + "loss": 0.2084, + "step": 6054 + }, + { + "epoch": 0.14193873544424715, + "grad_norm": 0.39840608835220337, + "learning_rate": 0.00019024882751920676, + "loss": 0.107, + "step": 6055 + }, + { + "epoch": 0.14196217701905214, + "grad_norm": 0.6088863015174866, + "learning_rate": 0.00019024565456395202, + "loss": 0.2107, + "step": 6056 + }, + { + "epoch": 0.14198561859385714, + "grad_norm": 0.3724347949028015, + "learning_rate": 0.00019024248111902182, + "loss": 0.0926, + "step": 6057 + }, + { + "epoch": 0.14200906016866213, + "grad_norm": 0.4771249294281006, + "learning_rate": 0.0001902393071844334, + "loss": 0.1703, + "step": 6058 + }, + { + "epoch": 0.14203250174346713, + "grad_norm": 0.37086552381515503, + "learning_rate": 0.00019023613276020394, + "loss": 0.3763, + "step": 6059 + }, + { + "epoch": 0.14205594331827212, + "grad_norm": 0.23363399505615234, + "learning_rate": 0.0001902329578463507, + "loss": 0.0661, + "step": 6060 + }, + { + "epoch": 0.14207938489307712, + "grad_norm": 0.6796140670776367, + "learning_rate": 0.00019022978244289087, + "loss": 0.7883, + "step": 6061 + }, + { + "epoch": 0.1421028264678821, + "grad_norm": 0.2705962657928467, + "learning_rate": 0.00019022660654984176, + "loss": 0.0901, + "step": 6062 + }, + { + "epoch": 0.1421262680426871, + "grad_norm": 0.710403561592102, + "learning_rate": 0.00019022343016722047, + "loss": 0.2511, + "step": 6063 + }, + { + "epoch": 0.1421497096174921, + "grad_norm": 0.7145451903343201, + "learning_rate": 0.00019022025329504437, + "loss": 0.1638, + "step": 6064 + }, + { + "epoch": 0.1421731511922971, + "grad_norm": 0.4554133117198944, + "learning_rate": 0.0001902170759333306, + "loss": 0.0976, + "step": 6065 + }, + { + "epoch": 0.1421965927671021, + "grad_norm": 0.7360729575157166, + "learning_rate": 0.00019021389808209646, + "loss": 0.1842, + "step": 6066 + }, + { + "epoch": 0.14222003434190708, + "grad_norm": 0.15821343660354614, + "learning_rate": 0.0001902107197413592, + "loss": 0.0311, + "step": 6067 + }, + { + "epoch": 0.14224347591671208, + "grad_norm": 0.439601868391037, + "learning_rate": 0.00019020754091113597, + "loss": 0.0839, + "step": 6068 + }, + { + "epoch": 0.14226691749151707, + "grad_norm": 0.3839534521102905, + "learning_rate": 0.00019020436159144412, + "loss": 0.085, + "step": 6069 + }, + { + "epoch": 0.14229035906632206, + "grad_norm": 0.6661440134048462, + "learning_rate": 0.00019020118178230089, + "loss": 0.1104, + "step": 6070 + }, + { + "epoch": 0.14231380064112706, + "grad_norm": 0.45743057131767273, + "learning_rate": 0.00019019800148372347, + "loss": 0.1693, + "step": 6071 + }, + { + "epoch": 0.14233724221593205, + "grad_norm": 0.3809828758239746, + "learning_rate": 0.00019019482069572918, + "loss": 0.1204, + "step": 6072 + }, + { + "epoch": 0.14236068379073707, + "grad_norm": 0.6088724136352539, + "learning_rate": 0.00019019163941833526, + "loss": 0.2057, + "step": 6073 + }, + { + "epoch": 0.14238412536554207, + "grad_norm": 0.2640458047389984, + "learning_rate": 0.00019018845765155897, + "loss": 0.0568, + "step": 6074 + }, + { + "epoch": 0.14240756694034706, + "grad_norm": 0.20732802152633667, + "learning_rate": 0.00019018527539541754, + "loss": 0.0496, + "step": 6075 + }, + { + "epoch": 0.14243100851515206, + "grad_norm": 0.4603261351585388, + "learning_rate": 0.00019018209264992831, + "loss": 0.1079, + "step": 6076 + }, + { + "epoch": 0.14245445008995705, + "grad_norm": 0.5685482025146484, + "learning_rate": 0.0001901789094151085, + "loss": 0.1964, + "step": 6077 + }, + { + "epoch": 0.14247789166476205, + "grad_norm": 0.12365163117647171, + "learning_rate": 0.00019017572569097538, + "loss": 0.0328, + "step": 6078 + }, + { + "epoch": 0.14250133323956704, + "grad_norm": 0.38184472918510437, + "learning_rate": 0.00019017254147754627, + "loss": 0.1016, + "step": 6079 + }, + { + "epoch": 0.14252477481437203, + "grad_norm": 0.5560317635536194, + "learning_rate": 0.0001901693567748384, + "loss": 0.1562, + "step": 6080 + }, + { + "epoch": 0.14254821638917703, + "grad_norm": 0.9535935521125793, + "learning_rate": 0.00019016617158286906, + "loss": 0.2438, + "step": 6081 + }, + { + "epoch": 0.14257165796398202, + "grad_norm": 0.8070653676986694, + "learning_rate": 0.00019016298590165552, + "loss": 0.1045, + "step": 6082 + }, + { + "epoch": 0.14259509953878702, + "grad_norm": 0.41426634788513184, + "learning_rate": 0.00019015979973121508, + "loss": 0.1022, + "step": 6083 + }, + { + "epoch": 0.142618541113592, + "grad_norm": 0.42773497104644775, + "learning_rate": 0.00019015661307156506, + "loss": 0.0986, + "step": 6084 + }, + { + "epoch": 0.142641982688397, + "grad_norm": 0.6831638216972351, + "learning_rate": 0.00019015342592272272, + "loss": 0.2065, + "step": 6085 + }, + { + "epoch": 0.142665424263202, + "grad_norm": 0.557022750377655, + "learning_rate": 0.00019015023828470535, + "loss": 0.1643, + "step": 6086 + }, + { + "epoch": 0.142688865838007, + "grad_norm": 0.34534451365470886, + "learning_rate": 0.00019014705015753025, + "loss": 0.0962, + "step": 6087 + }, + { + "epoch": 0.142712307412812, + "grad_norm": 0.6757498979568481, + "learning_rate": 0.0001901438615412147, + "loss": 0.1738, + "step": 6088 + }, + { + "epoch": 0.14273574898761698, + "grad_norm": 0.2397674173116684, + "learning_rate": 0.00019014067243577605, + "loss": 0.0813, + "step": 6089 + }, + { + "epoch": 0.14275919056242198, + "grad_norm": 0.7129843235015869, + "learning_rate": 0.0001901374828412316, + "loss": 0.1211, + "step": 6090 + }, + { + "epoch": 0.14278263213722697, + "grad_norm": 0.17193706333637238, + "learning_rate": 0.00019013429275759855, + "loss": 0.0349, + "step": 6091 + }, + { + "epoch": 0.14280607371203197, + "grad_norm": 0.46288174390792847, + "learning_rate": 0.00019013110218489438, + "loss": 0.1476, + "step": 6092 + }, + { + "epoch": 0.14282951528683696, + "grad_norm": 0.5320079922676086, + "learning_rate": 0.00019012791112313625, + "loss": 0.5035, + "step": 6093 + }, + { + "epoch": 0.14285295686164196, + "grad_norm": 0.36039242148399353, + "learning_rate": 0.00019012471957234158, + "loss": 0.0912, + "step": 6094 + }, + { + "epoch": 0.14287639843644695, + "grad_norm": 0.4243229031562805, + "learning_rate": 0.00019012152753252766, + "loss": 0.1392, + "step": 6095 + }, + { + "epoch": 0.14289984001125194, + "grad_norm": 0.36516958475112915, + "learning_rate": 0.00019011833500371175, + "loss": 0.0711, + "step": 6096 + }, + { + "epoch": 0.14292328158605694, + "grad_norm": 0.6534271240234375, + "learning_rate": 0.00019011514198591126, + "loss": 0.8275, + "step": 6097 + }, + { + "epoch": 0.14294672316086196, + "grad_norm": 0.699104368686676, + "learning_rate": 0.00019011194847914347, + "loss": 0.1272, + "step": 6098 + }, + { + "epoch": 0.14297016473566695, + "grad_norm": 0.5428085923194885, + "learning_rate": 0.0001901087544834257, + "loss": 0.9317, + "step": 6099 + }, + { + "epoch": 0.14299360631047195, + "grad_norm": 0.26833170652389526, + "learning_rate": 0.0001901055599987753, + "loss": 0.0906, + "step": 6100 + }, + { + "epoch": 0.14301704788527694, + "grad_norm": 0.17820997536182404, + "learning_rate": 0.00019010236502520962, + "loss": 0.0423, + "step": 6101 + }, + { + "epoch": 0.14304048946008194, + "grad_norm": 0.4728848338127136, + "learning_rate": 0.00019009916956274595, + "loss": 0.1036, + "step": 6102 + }, + { + "epoch": 0.14306393103488693, + "grad_norm": 1.5160306692123413, + "learning_rate": 0.00019009597361140164, + "loss": 0.1751, + "step": 6103 + }, + { + "epoch": 0.14308737260969193, + "grad_norm": 0.42843833565711975, + "learning_rate": 0.00019009277717119404, + "loss": 0.0793, + "step": 6104 + }, + { + "epoch": 0.14311081418449692, + "grad_norm": 0.5980677604675293, + "learning_rate": 0.00019008958024214055, + "loss": 0.1136, + "step": 6105 + }, + { + "epoch": 0.14313425575930191, + "grad_norm": 0.2898414731025696, + "learning_rate": 0.00019008638282425841, + "loss": 0.0615, + "step": 6106 + }, + { + "epoch": 0.1431576973341069, + "grad_norm": 0.5608275532722473, + "learning_rate": 0.00019008318491756505, + "loss": 0.1376, + "step": 6107 + }, + { + "epoch": 0.1431811389089119, + "grad_norm": 0.6216612458229065, + "learning_rate": 0.0001900799865220778, + "loss": 0.216, + "step": 6108 + }, + { + "epoch": 0.1432045804837169, + "grad_norm": 0.11298571527004242, + "learning_rate": 0.000190076787637814, + "loss": 0.0206, + "step": 6109 + }, + { + "epoch": 0.1432280220585219, + "grad_norm": 0.4206097722053528, + "learning_rate": 0.000190073588264791, + "loss": 0.0861, + "step": 6110 + }, + { + "epoch": 0.1432514636333269, + "grad_norm": 0.6606231927871704, + "learning_rate": 0.00019007038840302621, + "loss": 0.1908, + "step": 6111 + }, + { + "epoch": 0.14327490520813188, + "grad_norm": 0.5780471563339233, + "learning_rate": 0.00019006718805253696, + "loss": 0.2022, + "step": 6112 + }, + { + "epoch": 0.14329834678293688, + "grad_norm": 0.45060062408447266, + "learning_rate": 0.0001900639872133406, + "loss": 0.1701, + "step": 6113 + }, + { + "epoch": 0.14332178835774187, + "grad_norm": 0.7358613014221191, + "learning_rate": 0.0001900607858854545, + "loss": 0.1331, + "step": 6114 + }, + { + "epoch": 0.14334522993254686, + "grad_norm": 0.49378305673599243, + "learning_rate": 0.00019005758406889607, + "loss": 0.0749, + "step": 6115 + }, + { + "epoch": 0.14336867150735186, + "grad_norm": 0.170782208442688, + "learning_rate": 0.00019005438176368264, + "loss": 0.0259, + "step": 6116 + }, + { + "epoch": 0.14339211308215685, + "grad_norm": 0.43185678124427795, + "learning_rate": 0.0001900511789698316, + "loss": 0.1169, + "step": 6117 + }, + { + "epoch": 0.14341555465696185, + "grad_norm": 0.6806376576423645, + "learning_rate": 0.00019004797568736035, + "loss": 0.1583, + "step": 6118 + }, + { + "epoch": 0.14343899623176684, + "grad_norm": 0.7043593525886536, + "learning_rate": 0.00019004477191628623, + "loss": 0.1553, + "step": 6119 + }, + { + "epoch": 0.14346243780657184, + "grad_norm": 0.6870680451393127, + "learning_rate": 0.00019004156765662668, + "loss": 0.2047, + "step": 6120 + }, + { + "epoch": 0.14348587938137683, + "grad_norm": 0.7095122337341309, + "learning_rate": 0.000190038362908399, + "loss": 0.7753, + "step": 6121 + }, + { + "epoch": 0.14350932095618182, + "grad_norm": 0.2696634531021118, + "learning_rate": 0.00019003515767162068, + "loss": 0.0794, + "step": 6122 + }, + { + "epoch": 0.14353276253098682, + "grad_norm": 0.6992700695991516, + "learning_rate": 0.00019003195194630906, + "loss": 0.1205, + "step": 6123 + }, + { + "epoch": 0.14355620410579184, + "grad_norm": 0.6837976574897766, + "learning_rate": 0.0001900287457324815, + "loss": 0.2874, + "step": 6124 + }, + { + "epoch": 0.14357964568059683, + "grad_norm": 0.766730010509491, + "learning_rate": 0.00019002553903015544, + "loss": 0.167, + "step": 6125 + }, + { + "epoch": 0.14360308725540183, + "grad_norm": 0.3081222474575043, + "learning_rate": 0.0001900223318393483, + "loss": 0.0785, + "step": 6126 + }, + { + "epoch": 0.14362652883020682, + "grad_norm": 0.5382353663444519, + "learning_rate": 0.00019001912416007747, + "loss": 0.1221, + "step": 6127 + }, + { + "epoch": 0.14364997040501182, + "grad_norm": 0.5707526206970215, + "learning_rate": 0.00019001591599236028, + "loss": 0.1072, + "step": 6128 + }, + { + "epoch": 0.1436734119798168, + "grad_norm": 0.2008078694343567, + "learning_rate": 0.00019001270733621427, + "loss": 0.0497, + "step": 6129 + }, + { + "epoch": 0.1436968535546218, + "grad_norm": 0.3407547175884247, + "learning_rate": 0.0001900094981916567, + "loss": 0.0645, + "step": 6130 + }, + { + "epoch": 0.1437202951294268, + "grad_norm": 0.5662685632705688, + "learning_rate": 0.00019000628855870514, + "loss": 0.1549, + "step": 6131 + }, + { + "epoch": 0.1437437367042318, + "grad_norm": 0.737597644329071, + "learning_rate": 0.00019000307843737688, + "loss": 0.16, + "step": 6132 + }, + { + "epoch": 0.1437671782790368, + "grad_norm": 0.8684943914413452, + "learning_rate": 0.0001899998678276894, + "loss": 0.7707, + "step": 6133 + }, + { + "epoch": 0.14379061985384178, + "grad_norm": 0.5312809944152832, + "learning_rate": 0.0001899966567296601, + "loss": 0.1188, + "step": 6134 + }, + { + "epoch": 0.14381406142864678, + "grad_norm": 0.9245336651802063, + "learning_rate": 0.00018999344514330642, + "loss": 0.1033, + "step": 6135 + }, + { + "epoch": 0.14383750300345177, + "grad_norm": 0.4886225163936615, + "learning_rate": 0.00018999023306864575, + "loss": 0.1141, + "step": 6136 + }, + { + "epoch": 0.14386094457825677, + "grad_norm": 0.809787392616272, + "learning_rate": 0.00018998702050569558, + "loss": 0.3951, + "step": 6137 + }, + { + "epoch": 0.14388438615306176, + "grad_norm": 1.0173364877700806, + "learning_rate": 0.0001899838074544733, + "loss": 0.1403, + "step": 6138 + }, + { + "epoch": 0.14390782772786676, + "grad_norm": 0.7293844819068909, + "learning_rate": 0.00018998059391499633, + "loss": 0.1642, + "step": 6139 + }, + { + "epoch": 0.14393126930267175, + "grad_norm": 0.14407159388065338, + "learning_rate": 0.00018997737988728214, + "loss": 0.0188, + "step": 6140 + }, + { + "epoch": 0.14395471087747674, + "grad_norm": 0.3458007574081421, + "learning_rate": 0.0001899741653713482, + "loss": 0.0213, + "step": 6141 + }, + { + "epoch": 0.14397815245228174, + "grad_norm": 0.6228263974189758, + "learning_rate": 0.00018997095036721185, + "loss": 0.2022, + "step": 6142 + }, + { + "epoch": 0.14400159402708673, + "grad_norm": 1.7561330795288086, + "learning_rate": 0.0001899677348748906, + "loss": 0.1625, + "step": 6143 + }, + { + "epoch": 0.14402503560189173, + "grad_norm": 0.33120572566986084, + "learning_rate": 0.0001899645188944019, + "loss": 0.0348, + "step": 6144 + }, + { + "epoch": 0.14404847717669672, + "grad_norm": 0.637673020362854, + "learning_rate": 0.0001899613024257632, + "loss": 0.2035, + "step": 6145 + }, + { + "epoch": 0.14407191875150172, + "grad_norm": 0.47305023670196533, + "learning_rate": 0.0001899580854689919, + "loss": 0.1167, + "step": 6146 + }, + { + "epoch": 0.1440953603263067, + "grad_norm": 0.9935901761054993, + "learning_rate": 0.00018995486802410555, + "loss": 0.2087, + "step": 6147 + }, + { + "epoch": 0.1441188019011117, + "grad_norm": 0.5625693202018738, + "learning_rate": 0.00018995165009112156, + "loss": 0.6421, + "step": 6148 + }, + { + "epoch": 0.14414224347591673, + "grad_norm": 0.7222112417221069, + "learning_rate": 0.00018994843167005737, + "loss": 0.1588, + "step": 6149 + }, + { + "epoch": 0.14416568505072172, + "grad_norm": 0.5757148265838623, + "learning_rate": 0.00018994521276093044, + "loss": 0.1989, + "step": 6150 + }, + { + "epoch": 0.14418912662552671, + "grad_norm": 0.3314206600189209, + "learning_rate": 0.00018994199336375828, + "loss": 0.0593, + "step": 6151 + }, + { + "epoch": 0.1442125682003317, + "grad_norm": 0.6449065804481506, + "learning_rate": 0.00018993877347855834, + "loss": 0.1338, + "step": 6152 + }, + { + "epoch": 0.1442360097751367, + "grad_norm": 0.29078081250190735, + "learning_rate": 0.00018993555310534808, + "loss": 0.0564, + "step": 6153 + }, + { + "epoch": 0.1442594513499417, + "grad_norm": 0.7503832578659058, + "learning_rate": 0.00018993233224414495, + "loss": 0.1562, + "step": 6154 + }, + { + "epoch": 0.1442828929247467, + "grad_norm": 0.5784512162208557, + "learning_rate": 0.0001899291108949665, + "loss": 0.7379, + "step": 6155 + }, + { + "epoch": 0.1443063344995517, + "grad_norm": 0.3602547347545624, + "learning_rate": 0.00018992588905783014, + "loss": 0.0844, + "step": 6156 + }, + { + "epoch": 0.14432977607435668, + "grad_norm": 0.5382815599441528, + "learning_rate": 0.00018992266673275337, + "loss": 0.1826, + "step": 6157 + }, + { + "epoch": 0.14435321764916167, + "grad_norm": 0.5058600306510925, + "learning_rate": 0.00018991944391975372, + "loss": 0.0874, + "step": 6158 + }, + { + "epoch": 0.14437665922396667, + "grad_norm": 0.1447630375623703, + "learning_rate": 0.00018991622061884862, + "loss": 0.0351, + "step": 6159 + }, + { + "epoch": 0.14440010079877166, + "grad_norm": 0.6369964480400085, + "learning_rate": 0.00018991299683005557, + "loss": 0.2332, + "step": 6160 + }, + { + "epoch": 0.14442354237357666, + "grad_norm": 0.21385328471660614, + "learning_rate": 0.0001899097725533921, + "loss": 0.0631, + "step": 6161 + }, + { + "epoch": 0.14444698394838165, + "grad_norm": 0.8109645247459412, + "learning_rate": 0.00018990654778887564, + "loss": 0.1909, + "step": 6162 + }, + { + "epoch": 0.14447042552318665, + "grad_norm": 0.4314221441745758, + "learning_rate": 0.00018990332253652375, + "loss": 0.4543, + "step": 6163 + }, + { + "epoch": 0.14449386709799164, + "grad_norm": 0.22674649953842163, + "learning_rate": 0.00018990009679635387, + "loss": 0.0612, + "step": 6164 + }, + { + "epoch": 0.14451730867279663, + "grad_norm": 0.5984475612640381, + "learning_rate": 0.00018989687056838358, + "loss": 0.1716, + "step": 6165 + }, + { + "epoch": 0.14454075024760163, + "grad_norm": 0.49162837862968445, + "learning_rate": 0.00018989364385263032, + "loss": 0.1196, + "step": 6166 + }, + { + "epoch": 0.14456419182240662, + "grad_norm": 0.18227837979793549, + "learning_rate": 0.0001898904166491116, + "loss": 0.0416, + "step": 6167 + }, + { + "epoch": 0.14458763339721162, + "grad_norm": 0.5352467894554138, + "learning_rate": 0.000189887188957845, + "loss": 0.1447, + "step": 6168 + }, + { + "epoch": 0.1446110749720166, + "grad_norm": 0.3572659194469452, + "learning_rate": 0.00018988396077884793, + "loss": 0.0818, + "step": 6169 + }, + { + "epoch": 0.1446345165468216, + "grad_norm": 1.2472611665725708, + "learning_rate": 0.000189880732112138, + "loss": 0.2381, + "step": 6170 + }, + { + "epoch": 0.1446579581216266, + "grad_norm": 0.4206973910331726, + "learning_rate": 0.0001898775029577327, + "loss": 0.0858, + "step": 6171 + }, + { + "epoch": 0.1446813996964316, + "grad_norm": 0.6264958381652832, + "learning_rate": 0.0001898742733156495, + "loss": 0.0936, + "step": 6172 + }, + { + "epoch": 0.1447048412712366, + "grad_norm": 1.1436165571212769, + "learning_rate": 0.000189871043185906, + "loss": 0.2055, + "step": 6173 + }, + { + "epoch": 0.14472828284604158, + "grad_norm": 0.27700236439704895, + "learning_rate": 0.0001898678125685197, + "loss": 0.0316, + "step": 6174 + }, + { + "epoch": 0.1447517244208466, + "grad_norm": 0.582500696182251, + "learning_rate": 0.0001898645814635081, + "loss": 0.128, + "step": 6175 + }, + { + "epoch": 0.1447751659956516, + "grad_norm": 0.5723477005958557, + "learning_rate": 0.00018986134987088873, + "loss": 0.146, + "step": 6176 + }, + { + "epoch": 0.1447986075704566, + "grad_norm": 0.5739715099334717, + "learning_rate": 0.00018985811779067917, + "loss": 0.1532, + "step": 6177 + }, + { + "epoch": 0.1448220491452616, + "grad_norm": 0.6125586628913879, + "learning_rate": 0.00018985488522289695, + "loss": 0.0715, + "step": 6178 + }, + { + "epoch": 0.14484549072006658, + "grad_norm": 0.31908828020095825, + "learning_rate": 0.0001898516521675596, + "loss": 0.0884, + "step": 6179 + }, + { + "epoch": 0.14486893229487158, + "grad_norm": 0.589844822883606, + "learning_rate": 0.00018984841862468464, + "loss": 0.1886, + "step": 6180 + }, + { + "epoch": 0.14489237386967657, + "grad_norm": 0.5546556115150452, + "learning_rate": 0.00018984518459428962, + "loss": 0.5918, + "step": 6181 + }, + { + "epoch": 0.14491581544448157, + "grad_norm": 0.19893893599510193, + "learning_rate": 0.00018984195007639215, + "loss": 0.0449, + "step": 6182 + }, + { + "epoch": 0.14493925701928656, + "grad_norm": 0.7699630856513977, + "learning_rate": 0.0001898387150710097, + "loss": 0.1521, + "step": 6183 + }, + { + "epoch": 0.14496269859409155, + "grad_norm": 0.27399706840515137, + "learning_rate": 0.0001898354795781599, + "loss": 0.039, + "step": 6184 + }, + { + "epoch": 0.14498614016889655, + "grad_norm": 0.4140751361846924, + "learning_rate": 0.0001898322435978602, + "loss": 0.0585, + "step": 6185 + }, + { + "epoch": 0.14500958174370154, + "grad_norm": 0.30190277099609375, + "learning_rate": 0.00018982900713012825, + "loss": 0.0954, + "step": 6186 + }, + { + "epoch": 0.14503302331850654, + "grad_norm": 0.7839354276657104, + "learning_rate": 0.0001898257701749816, + "loss": 0.1645, + "step": 6187 + }, + { + "epoch": 0.14505646489331153, + "grad_norm": 0.2076377421617508, + "learning_rate": 0.00018982253273243775, + "loss": 0.0385, + "step": 6188 + }, + { + "epoch": 0.14507990646811653, + "grad_norm": 0.4333295524120331, + "learning_rate": 0.00018981929480251433, + "loss": 0.0912, + "step": 6189 + }, + { + "epoch": 0.14510334804292152, + "grad_norm": 0.27706485986709595, + "learning_rate": 0.00018981605638522888, + "loss": 0.0377, + "step": 6190 + }, + { + "epoch": 0.14512678961772651, + "grad_norm": 0.35284292697906494, + "learning_rate": 0.000189812817480599, + "loss": 0.0806, + "step": 6191 + }, + { + "epoch": 0.1451502311925315, + "grad_norm": 0.32741379737854004, + "learning_rate": 0.00018980957808864223, + "loss": 0.0894, + "step": 6192 + }, + { + "epoch": 0.1451736727673365, + "grad_norm": 0.6584396362304688, + "learning_rate": 0.00018980633820937621, + "loss": 0.1884, + "step": 6193 + }, + { + "epoch": 0.1451971143421415, + "grad_norm": 1.1586534976959229, + "learning_rate": 0.00018980309784281843, + "loss": 0.2191, + "step": 6194 + }, + { + "epoch": 0.1452205559169465, + "grad_norm": 0.4672541618347168, + "learning_rate": 0.00018979985698898655, + "loss": 0.0534, + "step": 6195 + }, + { + "epoch": 0.1452439974917515, + "grad_norm": 0.23391559720039368, + "learning_rate": 0.00018979661564789808, + "loss": 0.0445, + "step": 6196 + }, + { + "epoch": 0.14526743906655648, + "grad_norm": 0.5861639380455017, + "learning_rate": 0.00018979337381957065, + "loss": 0.1591, + "step": 6197 + }, + { + "epoch": 0.14529088064136148, + "grad_norm": 0.6802726984024048, + "learning_rate": 0.0001897901315040219, + "loss": 0.7128, + "step": 6198 + }, + { + "epoch": 0.14531432221616647, + "grad_norm": 0.6484432816505432, + "learning_rate": 0.00018978688870126932, + "loss": 0.1042, + "step": 6199 + }, + { + "epoch": 0.14533776379097146, + "grad_norm": 0.9425653219223022, + "learning_rate": 0.0001897836454113306, + "loss": 0.3779, + "step": 6200 + }, + { + "epoch": 0.14536120536577649, + "grad_norm": 0.3176623582839966, + "learning_rate": 0.00018978040163422326, + "loss": 0.2306, + "step": 6201 + }, + { + "epoch": 0.14538464694058148, + "grad_norm": 0.5159093737602234, + "learning_rate": 0.00018977715736996493, + "loss": 0.1088, + "step": 6202 + }, + { + "epoch": 0.14540808851538647, + "grad_norm": 0.2705397605895996, + "learning_rate": 0.00018977391261857326, + "loss": 0.0549, + "step": 6203 + }, + { + "epoch": 0.14543153009019147, + "grad_norm": 0.44972914457321167, + "learning_rate": 0.00018977066738006578, + "loss": 0.1119, + "step": 6204 + }, + { + "epoch": 0.14545497166499646, + "grad_norm": 0.6354265213012695, + "learning_rate": 0.00018976742165446016, + "loss": 0.1305, + "step": 6205 + }, + { + "epoch": 0.14547841323980146, + "grad_norm": 0.19365237653255463, + "learning_rate": 0.00018976417544177396, + "loss": 0.04, + "step": 6206 + }, + { + "epoch": 0.14550185481460645, + "grad_norm": 0.5895678400993347, + "learning_rate": 0.00018976092874202485, + "loss": 0.6299, + "step": 6207 + }, + { + "epoch": 0.14552529638941145, + "grad_norm": 0.2891083061695099, + "learning_rate": 0.0001897576815552304, + "loss": 0.0605, + "step": 6208 + }, + { + "epoch": 0.14554873796421644, + "grad_norm": 0.28556743264198303, + "learning_rate": 0.00018975443388140825, + "loss": 0.086, + "step": 6209 + }, + { + "epoch": 0.14557217953902143, + "grad_norm": 0.5431203246116638, + "learning_rate": 0.000189751185720576, + "loss": 0.1456, + "step": 6210 + }, + { + "epoch": 0.14559562111382643, + "grad_norm": 0.15809296071529388, + "learning_rate": 0.0001897479370727513, + "loss": 0.0362, + "step": 6211 + }, + { + "epoch": 0.14561906268863142, + "grad_norm": 0.6168155074119568, + "learning_rate": 0.0001897446879379518, + "loss": 0.5197, + "step": 6212 + }, + { + "epoch": 0.14564250426343642, + "grad_norm": 0.17600664496421814, + "learning_rate": 0.00018974143831619505, + "loss": 0.028, + "step": 6213 + }, + { + "epoch": 0.1456659458382414, + "grad_norm": 0.7096483707427979, + "learning_rate": 0.00018973818820749877, + "loss": 0.1981, + "step": 6214 + }, + { + "epoch": 0.1456893874130464, + "grad_norm": 0.5888804197311401, + "learning_rate": 0.00018973493761188057, + "loss": 0.0669, + "step": 6215 + }, + { + "epoch": 0.1457128289878514, + "grad_norm": 0.5312600135803223, + "learning_rate": 0.00018973168652935805, + "loss": 0.1698, + "step": 6216 + }, + { + "epoch": 0.1457362705626564, + "grad_norm": 0.5564470291137695, + "learning_rate": 0.00018972843495994887, + "loss": 0.1447, + "step": 6217 + }, + { + "epoch": 0.1457597121374614, + "grad_norm": 0.43944016098976135, + "learning_rate": 0.0001897251829036707, + "loss": 0.1104, + "step": 6218 + }, + { + "epoch": 0.14578315371226638, + "grad_norm": 0.480033278465271, + "learning_rate": 0.00018972193036054114, + "loss": 0.1338, + "step": 6219 + }, + { + "epoch": 0.14580659528707138, + "grad_norm": 0.3816237449645996, + "learning_rate": 0.00018971867733057787, + "loss": 0.3474, + "step": 6220 + }, + { + "epoch": 0.14583003686187637, + "grad_norm": 0.8394709229469299, + "learning_rate": 0.00018971542381379853, + "loss": 0.1774, + "step": 6221 + }, + { + "epoch": 0.14585347843668137, + "grad_norm": 0.4204491376876831, + "learning_rate": 0.0001897121698102208, + "loss": 0.0813, + "step": 6222 + }, + { + "epoch": 0.14587692001148636, + "grad_norm": 0.5787673592567444, + "learning_rate": 0.0001897089153198623, + "loss": 0.1732, + "step": 6223 + }, + { + "epoch": 0.14590036158629135, + "grad_norm": 0.6009409427642822, + "learning_rate": 0.00018970566034274072, + "loss": 0.131, + "step": 6224 + }, + { + "epoch": 0.14592380316109635, + "grad_norm": 0.8904398083686829, + "learning_rate": 0.00018970240487887368, + "loss": 0.7286, + "step": 6225 + }, + { + "epoch": 0.14594724473590137, + "grad_norm": 0.1061154305934906, + "learning_rate": 0.0001896991489282789, + "loss": 0.0322, + "step": 6226 + }, + { + "epoch": 0.14597068631070637, + "grad_norm": 0.6049712896347046, + "learning_rate": 0.00018969589249097398, + "loss": 0.1737, + "step": 6227 + }, + { + "epoch": 0.14599412788551136, + "grad_norm": 0.4641425907611847, + "learning_rate": 0.00018969263556697664, + "loss": 0.1498, + "step": 6228 + }, + { + "epoch": 0.14601756946031635, + "grad_norm": 0.509366512298584, + "learning_rate": 0.00018968937815630455, + "loss": 0.5274, + "step": 6229 + }, + { + "epoch": 0.14604101103512135, + "grad_norm": 0.1667369306087494, + "learning_rate": 0.00018968612025897533, + "loss": 0.037, + "step": 6230 + }, + { + "epoch": 0.14606445260992634, + "grad_norm": 0.514068603515625, + "learning_rate": 0.00018968286187500674, + "loss": 0.1637, + "step": 6231 + }, + { + "epoch": 0.14608789418473134, + "grad_norm": 0.39273640513420105, + "learning_rate": 0.00018967960300441642, + "loss": 0.0944, + "step": 6232 + }, + { + "epoch": 0.14611133575953633, + "grad_norm": 0.1370335966348648, + "learning_rate": 0.00018967634364722204, + "loss": 0.0155, + "step": 6233 + }, + { + "epoch": 0.14613477733434133, + "grad_norm": 0.5691137909889221, + "learning_rate": 0.00018967308380344128, + "loss": 0.1955, + "step": 6234 + }, + { + "epoch": 0.14615821890914632, + "grad_norm": 0.7118719816207886, + "learning_rate": 0.00018966982347309188, + "loss": 0.1321, + "step": 6235 + }, + { + "epoch": 0.14618166048395131, + "grad_norm": 0.8135843276977539, + "learning_rate": 0.0001896665626561915, + "loss": 0.1774, + "step": 6236 + }, + { + "epoch": 0.1462051020587563, + "grad_norm": 0.945244550704956, + "learning_rate": 0.0001896633013527578, + "loss": 0.1867, + "step": 6237 + }, + { + "epoch": 0.1462285436335613, + "grad_norm": 0.1592172384262085, + "learning_rate": 0.0001896600395628085, + "loss": 0.0164, + "step": 6238 + }, + { + "epoch": 0.1462519852083663, + "grad_norm": 0.2197091430425644, + "learning_rate": 0.00018965677728636132, + "loss": 0.045, + "step": 6239 + }, + { + "epoch": 0.1462754267831713, + "grad_norm": 0.40033188462257385, + "learning_rate": 0.00018965351452343397, + "loss": 0.0831, + "step": 6240 + }, + { + "epoch": 0.14629886835797629, + "grad_norm": 0.4708442986011505, + "learning_rate": 0.0001896502512740441, + "loss": 0.1279, + "step": 6241 + }, + { + "epoch": 0.14632230993278128, + "grad_norm": 0.345679372549057, + "learning_rate": 0.00018964698753820946, + "loss": 0.1077, + "step": 6242 + }, + { + "epoch": 0.14634575150758627, + "grad_norm": 0.19564175605773926, + "learning_rate": 0.00018964372331594772, + "loss": 0.0455, + "step": 6243 + }, + { + "epoch": 0.14636919308239127, + "grad_norm": 0.5477762818336487, + "learning_rate": 0.00018964045860727665, + "loss": 0.1988, + "step": 6244 + }, + { + "epoch": 0.14639263465719626, + "grad_norm": 0.6158508062362671, + "learning_rate": 0.0001896371934122139, + "loss": 0.2229, + "step": 6245 + }, + { + "epoch": 0.14641607623200126, + "grad_norm": 0.8382636308670044, + "learning_rate": 0.00018963392773077727, + "loss": 0.1629, + "step": 6246 + }, + { + "epoch": 0.14643951780680625, + "grad_norm": 0.48442161083221436, + "learning_rate": 0.00018963066156298437, + "loss": 0.0784, + "step": 6247 + }, + { + "epoch": 0.14646295938161125, + "grad_norm": 0.6324441432952881, + "learning_rate": 0.000189627394908853, + "loss": 0.2057, + "step": 6248 + }, + { + "epoch": 0.14648640095641624, + "grad_norm": 0.5374578833580017, + "learning_rate": 0.00018962412776840087, + "loss": 0.1937, + "step": 6249 + }, + { + "epoch": 0.14650984253122123, + "grad_norm": 0.6543413996696472, + "learning_rate": 0.0001896208601416457, + "loss": 0.2434, + "step": 6250 + }, + { + "epoch": 0.14653328410602623, + "grad_norm": 0.798316240310669, + "learning_rate": 0.00018961759202860523, + "loss": 0.7776, + "step": 6251 + }, + { + "epoch": 0.14655672568083125, + "grad_norm": 0.20740529894828796, + "learning_rate": 0.00018961432342929716, + "loss": 0.0317, + "step": 6252 + }, + { + "epoch": 0.14658016725563625, + "grad_norm": 0.25994765758514404, + "learning_rate": 0.00018961105434373926, + "loss": 0.0614, + "step": 6253 + }, + { + "epoch": 0.14660360883044124, + "grad_norm": 0.40420010685920715, + "learning_rate": 0.0001896077847719493, + "loss": 0.1303, + "step": 6254 + }, + { + "epoch": 0.14662705040524623, + "grad_norm": 0.26332613825798035, + "learning_rate": 0.00018960451471394492, + "loss": 0.0749, + "step": 6255 + }, + { + "epoch": 0.14665049198005123, + "grad_norm": 0.10348295420408249, + "learning_rate": 0.00018960124416974398, + "loss": 0.0221, + "step": 6256 + }, + { + "epoch": 0.14667393355485622, + "grad_norm": 0.48813989758491516, + "learning_rate": 0.00018959797313936413, + "loss": 0.1084, + "step": 6257 + }, + { + "epoch": 0.14669737512966122, + "grad_norm": 0.48004427552223206, + "learning_rate": 0.00018959470162282317, + "loss": 0.1263, + "step": 6258 + }, + { + "epoch": 0.1467208167044662, + "grad_norm": 0.4717961549758911, + "learning_rate": 0.00018959142962013886, + "loss": 0.2016, + "step": 6259 + }, + { + "epoch": 0.1467442582792712, + "grad_norm": 0.4509887397289276, + "learning_rate": 0.0001895881571313289, + "loss": 0.6188, + "step": 6260 + }, + { + "epoch": 0.1467676998540762, + "grad_norm": 0.4882364571094513, + "learning_rate": 0.00018958488415641108, + "loss": 0.1242, + "step": 6261 + }, + { + "epoch": 0.1467911414288812, + "grad_norm": 0.5354409217834473, + "learning_rate": 0.00018958161069540318, + "loss": 0.1994, + "step": 6262 + }, + { + "epoch": 0.1468145830036862, + "grad_norm": 0.9787826538085938, + "learning_rate": 0.00018957833674832295, + "loss": 0.2366, + "step": 6263 + }, + { + "epoch": 0.14683802457849118, + "grad_norm": 0.6215789318084717, + "learning_rate": 0.00018957506231518813, + "loss": 0.1377, + "step": 6264 + }, + { + "epoch": 0.14686146615329618, + "grad_norm": 0.6030758619308472, + "learning_rate": 0.0001895717873960165, + "loss": 0.1124, + "step": 6265 + }, + { + "epoch": 0.14688490772810117, + "grad_norm": 0.6100962162017822, + "learning_rate": 0.00018956851199082583, + "loss": 0.4468, + "step": 6266 + }, + { + "epoch": 0.14690834930290617, + "grad_norm": 0.40360692143440247, + "learning_rate": 0.00018956523609963392, + "loss": 0.1146, + "step": 6267 + }, + { + "epoch": 0.14693179087771116, + "grad_norm": 0.43163207173347473, + "learning_rate": 0.0001895619597224585, + "loss": 0.1297, + "step": 6268 + }, + { + "epoch": 0.14695523245251615, + "grad_norm": 0.43321502208709717, + "learning_rate": 0.0001895586828593174, + "loss": 0.118, + "step": 6269 + }, + { + "epoch": 0.14697867402732115, + "grad_norm": 0.44749221205711365, + "learning_rate": 0.0001895554055102283, + "loss": 0.1185, + "step": 6270 + }, + { + "epoch": 0.14700211560212614, + "grad_norm": 0.47361522912979126, + "learning_rate": 0.0001895521276752091, + "loss": 0.1567, + "step": 6271 + }, + { + "epoch": 0.14702555717693114, + "grad_norm": 0.5282322764396667, + "learning_rate": 0.00018954884935427755, + "loss": 0.1366, + "step": 6272 + }, + { + "epoch": 0.14704899875173613, + "grad_norm": 0.2068149745464325, + "learning_rate": 0.0001895455705474514, + "loss": 0.0476, + "step": 6273 + }, + { + "epoch": 0.14707244032654113, + "grad_norm": 0.31493595242500305, + "learning_rate": 0.00018954229125474846, + "loss": 0.0732, + "step": 6274 + }, + { + "epoch": 0.14709588190134612, + "grad_norm": 1.1543142795562744, + "learning_rate": 0.00018953901147618655, + "loss": 0.2048, + "step": 6275 + }, + { + "epoch": 0.14711932347615111, + "grad_norm": 0.621058464050293, + "learning_rate": 0.00018953573121178346, + "loss": 0.158, + "step": 6276 + }, + { + "epoch": 0.14714276505095614, + "grad_norm": 0.5842419266700745, + "learning_rate": 0.00018953245046155693, + "loss": 0.6739, + "step": 6277 + }, + { + "epoch": 0.14716620662576113, + "grad_norm": 0.5018057227134705, + "learning_rate": 0.00018952916922552484, + "loss": 0.1449, + "step": 6278 + }, + { + "epoch": 0.14718964820056613, + "grad_norm": 0.3899982273578644, + "learning_rate": 0.00018952588750370495, + "loss": 0.4965, + "step": 6279 + }, + { + "epoch": 0.14721308977537112, + "grad_norm": 0.5935701131820679, + "learning_rate": 0.00018952260529611506, + "loss": 0.1288, + "step": 6280 + }, + { + "epoch": 0.14723653135017611, + "grad_norm": 0.6219072937965393, + "learning_rate": 0.00018951932260277303, + "loss": 0.2273, + "step": 6281 + }, + { + "epoch": 0.1472599729249811, + "grad_norm": 0.3051478862762451, + "learning_rate": 0.00018951603942369663, + "loss": 0.0845, + "step": 6282 + }, + { + "epoch": 0.1472834144997861, + "grad_norm": 0.3595830500125885, + "learning_rate": 0.00018951275575890367, + "loss": 0.1593, + "step": 6283 + }, + { + "epoch": 0.1473068560745911, + "grad_norm": 0.6258861422538757, + "learning_rate": 0.000189509471608412, + "loss": 0.4944, + "step": 6284 + }, + { + "epoch": 0.1473302976493961, + "grad_norm": 0.4086894392967224, + "learning_rate": 0.00018950618697223938, + "loss": 0.3277, + "step": 6285 + }, + { + "epoch": 0.14735373922420109, + "grad_norm": 0.5789048671722412, + "learning_rate": 0.0001895029018504037, + "loss": 0.1586, + "step": 6286 + }, + { + "epoch": 0.14737718079900608, + "grad_norm": 0.606671154499054, + "learning_rate": 0.00018949961624292278, + "loss": 0.1845, + "step": 6287 + }, + { + "epoch": 0.14740062237381107, + "grad_norm": 0.200893834233284, + "learning_rate": 0.0001894963301498144, + "loss": 0.0338, + "step": 6288 + }, + { + "epoch": 0.14742406394861607, + "grad_norm": 0.6008874177932739, + "learning_rate": 0.00018949304357109643, + "loss": 0.563, + "step": 6289 + }, + { + "epoch": 0.14744750552342106, + "grad_norm": 0.5187680721282959, + "learning_rate": 0.00018948975650678667, + "loss": 0.125, + "step": 6290 + }, + { + "epoch": 0.14747094709822606, + "grad_norm": 0.346358984708786, + "learning_rate": 0.00018948646895690295, + "loss": 0.5087, + "step": 6291 + }, + { + "epoch": 0.14749438867303105, + "grad_norm": 0.706204354763031, + "learning_rate": 0.0001894831809214632, + "loss": 0.0941, + "step": 6292 + }, + { + "epoch": 0.14751783024783605, + "grad_norm": 0.4558933675289154, + "learning_rate": 0.00018947989240048515, + "loss": 0.1365, + "step": 6293 + }, + { + "epoch": 0.14754127182264104, + "grad_norm": 0.6789003610610962, + "learning_rate": 0.0001894766033939867, + "loss": 0.1485, + "step": 6294 + }, + { + "epoch": 0.14756471339744603, + "grad_norm": 0.6374213695526123, + "learning_rate": 0.0001894733139019857, + "loss": 0.6589, + "step": 6295 + }, + { + "epoch": 0.14758815497225103, + "grad_norm": 0.30911189317703247, + "learning_rate": 0.00018947002392449995, + "loss": 0.0476, + "step": 6296 + }, + { + "epoch": 0.14761159654705602, + "grad_norm": 0.11422283202409744, + "learning_rate": 0.00018946673346154736, + "loss": 0.0199, + "step": 6297 + }, + { + "epoch": 0.14763503812186102, + "grad_norm": 0.19742383062839508, + "learning_rate": 0.00018946344251314577, + "loss": 0.0426, + "step": 6298 + }, + { + "epoch": 0.147658479696666, + "grad_norm": 0.47891366481781006, + "learning_rate": 0.00018946015107931298, + "loss": 0.1511, + "step": 6299 + }, + { + "epoch": 0.147681921271471, + "grad_norm": 0.13634547591209412, + "learning_rate": 0.0001894568591600669, + "loss": 0.0307, + "step": 6300 + }, + { + "epoch": 0.147705362846276, + "grad_norm": 0.5206576585769653, + "learning_rate": 0.00018945356675542543, + "loss": 0.18, + "step": 6301 + }, + { + "epoch": 0.147728804421081, + "grad_norm": 0.562019407749176, + "learning_rate": 0.00018945027386540637, + "loss": 0.167, + "step": 6302 + }, + { + "epoch": 0.14775224599588602, + "grad_norm": 0.8155651092529297, + "learning_rate": 0.0001894469804900276, + "loss": 0.8417, + "step": 6303 + }, + { + "epoch": 0.147775687570691, + "grad_norm": 0.2868097722530365, + "learning_rate": 0.000189443686629307, + "loss": 0.0678, + "step": 6304 + }, + { + "epoch": 0.147799129145496, + "grad_norm": 0.75325608253479, + "learning_rate": 0.00018944039228326246, + "loss": 0.7212, + "step": 6305 + }, + { + "epoch": 0.147822570720301, + "grad_norm": 0.2413099855184555, + "learning_rate": 0.00018943709745191184, + "loss": 0.0679, + "step": 6306 + }, + { + "epoch": 0.147846012295106, + "grad_norm": 0.6832599639892578, + "learning_rate": 0.00018943380213527297, + "loss": 0.1559, + "step": 6307 + }, + { + "epoch": 0.147869453869911, + "grad_norm": 0.7110723853111267, + "learning_rate": 0.0001894305063333638, + "loss": 0.6833, + "step": 6308 + }, + { + "epoch": 0.14789289544471598, + "grad_norm": 0.11288414150476456, + "learning_rate": 0.0001894272100462022, + "loss": 0.0148, + "step": 6309 + }, + { + "epoch": 0.14791633701952098, + "grad_norm": 0.5696084499359131, + "learning_rate": 0.00018942391327380603, + "loss": 0.1689, + "step": 6310 + }, + { + "epoch": 0.14793977859432597, + "grad_norm": 0.38263770937919617, + "learning_rate": 0.00018942061601619317, + "loss": 0.0781, + "step": 6311 + }, + { + "epoch": 0.14796322016913097, + "grad_norm": 0.1880982667207718, + "learning_rate": 0.0001894173182733816, + "loss": 0.0568, + "step": 6312 + }, + { + "epoch": 0.14798666174393596, + "grad_norm": 0.3341318666934967, + "learning_rate": 0.0001894140200453891, + "loss": 0.0712, + "step": 6313 + }, + { + "epoch": 0.14801010331874095, + "grad_norm": 0.11358880251646042, + "learning_rate": 0.00018941072133223363, + "loss": 0.0206, + "step": 6314 + }, + { + "epoch": 0.14803354489354595, + "grad_norm": 0.5498273372650146, + "learning_rate": 0.00018940742213393306, + "loss": 0.1603, + "step": 6315 + }, + { + "epoch": 0.14805698646835094, + "grad_norm": 0.4565960168838501, + "learning_rate": 0.0001894041224505053, + "loss": 0.1114, + "step": 6316 + }, + { + "epoch": 0.14808042804315594, + "grad_norm": 0.35538938641548157, + "learning_rate": 0.00018940082228196826, + "loss": 0.0969, + "step": 6317 + }, + { + "epoch": 0.14810386961796093, + "grad_norm": 0.5352493524551392, + "learning_rate": 0.00018939752162833986, + "loss": 0.714, + "step": 6318 + }, + { + "epoch": 0.14812731119276593, + "grad_norm": 0.8296027183532715, + "learning_rate": 0.00018939422048963798, + "loss": 0.2001, + "step": 6319 + }, + { + "epoch": 0.14815075276757092, + "grad_norm": 0.44435104727745056, + "learning_rate": 0.00018939091886588055, + "loss": 0.066, + "step": 6320 + }, + { + "epoch": 0.14817419434237591, + "grad_norm": 0.632613480091095, + "learning_rate": 0.00018938761675708546, + "loss": 0.1244, + "step": 6321 + }, + { + "epoch": 0.1481976359171809, + "grad_norm": 0.5183263421058655, + "learning_rate": 0.0001893843141632707, + "loss": 0.1093, + "step": 6322 + }, + { + "epoch": 0.1482210774919859, + "grad_norm": 0.44785362482070923, + "learning_rate": 0.0001893810110844541, + "loss": 0.12, + "step": 6323 + }, + { + "epoch": 0.1482445190667909, + "grad_norm": 0.4447363018989563, + "learning_rate": 0.00018937770752065362, + "loss": 0.2053, + "step": 6324 + }, + { + "epoch": 0.1482679606415959, + "grad_norm": 0.6736703515052795, + "learning_rate": 0.0001893744034718872, + "loss": 0.1513, + "step": 6325 + }, + { + "epoch": 0.14829140221640089, + "grad_norm": 0.4812237322330475, + "learning_rate": 0.00018937109893817273, + "loss": 0.1366, + "step": 6326 + }, + { + "epoch": 0.14831484379120588, + "grad_norm": 0.6886353492736816, + "learning_rate": 0.0001893677939195282, + "loss": 0.1444, + "step": 6327 + }, + { + "epoch": 0.1483382853660109, + "grad_norm": 0.4960653781890869, + "learning_rate": 0.00018936448841597146, + "loss": 0.1571, + "step": 6328 + }, + { + "epoch": 0.1483617269408159, + "grad_norm": 0.2770587205886841, + "learning_rate": 0.00018936118242752055, + "loss": 0.0692, + "step": 6329 + }, + { + "epoch": 0.1483851685156209, + "grad_norm": 0.5901471376419067, + "learning_rate": 0.0001893578759541933, + "loss": 0.1043, + "step": 6330 + }, + { + "epoch": 0.14840861009042589, + "grad_norm": 0.11565019190311432, + "learning_rate": 0.00018935456899600773, + "loss": 0.018, + "step": 6331 + }, + { + "epoch": 0.14843205166523088, + "grad_norm": 0.5065077543258667, + "learning_rate": 0.00018935126155298176, + "loss": 0.1666, + "step": 6332 + }, + { + "epoch": 0.14845549324003587, + "grad_norm": 0.516981303691864, + "learning_rate": 0.00018934795362513332, + "loss": 0.1131, + "step": 6333 + }, + { + "epoch": 0.14847893481484087, + "grad_norm": 0.5198931097984314, + "learning_rate": 0.00018934464521248038, + "loss": 0.1248, + "step": 6334 + }, + { + "epoch": 0.14850237638964586, + "grad_norm": 0.6579111218452454, + "learning_rate": 0.00018934133631504088, + "loss": 0.1386, + "step": 6335 + }, + { + "epoch": 0.14852581796445086, + "grad_norm": 0.5737887620925903, + "learning_rate": 0.00018933802693283275, + "loss": 0.0957, + "step": 6336 + }, + { + "epoch": 0.14854925953925585, + "grad_norm": 0.35489577054977417, + "learning_rate": 0.000189334717065874, + "loss": 0.0769, + "step": 6337 + }, + { + "epoch": 0.14857270111406085, + "grad_norm": 0.4796897768974304, + "learning_rate": 0.00018933140671418256, + "loss": 0.0788, + "step": 6338 + }, + { + "epoch": 0.14859614268886584, + "grad_norm": 0.06643225252628326, + "learning_rate": 0.00018932809587777642, + "loss": 0.0088, + "step": 6339 + }, + { + "epoch": 0.14861958426367083, + "grad_norm": 1.0803852081298828, + "learning_rate": 0.00018932478455667347, + "loss": 0.1906, + "step": 6340 + }, + { + "epoch": 0.14864302583847583, + "grad_norm": 0.21960033476352692, + "learning_rate": 0.00018932147275089176, + "loss": 0.0307, + "step": 6341 + }, + { + "epoch": 0.14866646741328082, + "grad_norm": 0.721204936504364, + "learning_rate": 0.00018931816046044924, + "loss": 0.1116, + "step": 6342 + }, + { + "epoch": 0.14868990898808582, + "grad_norm": 0.6589672565460205, + "learning_rate": 0.00018931484768536385, + "loss": 0.1684, + "step": 6343 + }, + { + "epoch": 0.1487133505628908, + "grad_norm": 0.43282485008239746, + "learning_rate": 0.00018931153442565356, + "loss": 0.101, + "step": 6344 + }, + { + "epoch": 0.1487367921376958, + "grad_norm": 0.6683513522148132, + "learning_rate": 0.00018930822068133642, + "loss": 0.1917, + "step": 6345 + }, + { + "epoch": 0.1487602337125008, + "grad_norm": 0.5681314468383789, + "learning_rate": 0.00018930490645243034, + "loss": 0.1475, + "step": 6346 + }, + { + "epoch": 0.1487836752873058, + "grad_norm": 0.36228057742118835, + "learning_rate": 0.00018930159173895334, + "loss": 0.1069, + "step": 6347 + }, + { + "epoch": 0.1488071168621108, + "grad_norm": 0.5641838908195496, + "learning_rate": 0.00018929827654092337, + "loss": 0.4627, + "step": 6348 + }, + { + "epoch": 0.14883055843691578, + "grad_norm": 0.7869707345962524, + "learning_rate": 0.00018929496085835846, + "loss": 0.2056, + "step": 6349 + }, + { + "epoch": 0.14885400001172078, + "grad_norm": 0.23832309246063232, + "learning_rate": 0.0001892916446912766, + "loss": 0.0716, + "step": 6350 + }, + { + "epoch": 0.14887744158652577, + "grad_norm": 0.45109477639198303, + "learning_rate": 0.00018928832803969572, + "loss": 0.1236, + "step": 6351 + }, + { + "epoch": 0.14890088316133077, + "grad_norm": 0.24775207042694092, + "learning_rate": 0.00018928501090363387, + "loss": 0.0276, + "step": 6352 + }, + { + "epoch": 0.14892432473613576, + "grad_norm": 0.47912371158599854, + "learning_rate": 0.00018928169328310905, + "loss": 0.1329, + "step": 6353 + }, + { + "epoch": 0.14894776631094078, + "grad_norm": 0.7060585618019104, + "learning_rate": 0.00018927837517813926, + "loss": 0.1095, + "step": 6354 + }, + { + "epoch": 0.14897120788574578, + "grad_norm": 0.560736358165741, + "learning_rate": 0.0001892750565887425, + "loss": 0.1317, + "step": 6355 + }, + { + "epoch": 0.14899464946055077, + "grad_norm": 0.7881178855895996, + "learning_rate": 0.00018927173751493676, + "loss": 0.7634, + "step": 6356 + }, + { + "epoch": 0.14901809103535577, + "grad_norm": 0.45014217495918274, + "learning_rate": 0.00018926841795674005, + "loss": 0.0686, + "step": 6357 + }, + { + "epoch": 0.14904153261016076, + "grad_norm": 0.2892800569534302, + "learning_rate": 0.0001892650979141704, + "loss": 0.0452, + "step": 6358 + }, + { + "epoch": 0.14906497418496575, + "grad_norm": 0.4027005732059479, + "learning_rate": 0.00018926177738724586, + "loss": 0.1152, + "step": 6359 + }, + { + "epoch": 0.14908841575977075, + "grad_norm": 0.0775633379817009, + "learning_rate": 0.00018925845637598435, + "loss": 0.0133, + "step": 6360 + }, + { + "epoch": 0.14911185733457574, + "grad_norm": 0.616970956325531, + "learning_rate": 0.00018925513488040398, + "loss": 0.2229, + "step": 6361 + }, + { + "epoch": 0.14913529890938074, + "grad_norm": 0.13064956665039062, + "learning_rate": 0.00018925181290052273, + "loss": 0.0099, + "step": 6362 + }, + { + "epoch": 0.14915874048418573, + "grad_norm": 0.10487546771764755, + "learning_rate": 0.00018924849043635864, + "loss": 0.0357, + "step": 6363 + }, + { + "epoch": 0.14918218205899073, + "grad_norm": 0.7319186925888062, + "learning_rate": 0.00018924516748792974, + "loss": 0.2012, + "step": 6364 + }, + { + "epoch": 0.14920562363379572, + "grad_norm": 0.32156485319137573, + "learning_rate": 0.000189241844055254, + "loss": 0.0899, + "step": 6365 + }, + { + "epoch": 0.14922906520860071, + "grad_norm": 0.7083336114883423, + "learning_rate": 0.00018923852013834954, + "loss": 0.1403, + "step": 6366 + }, + { + "epoch": 0.1492525067834057, + "grad_norm": 0.5891680121421814, + "learning_rate": 0.0001892351957372344, + "loss": 0.1872, + "step": 6367 + }, + { + "epoch": 0.1492759483582107, + "grad_norm": 0.33684614300727844, + "learning_rate": 0.0001892318708519265, + "loss": 0.0874, + "step": 6368 + }, + { + "epoch": 0.1492993899330157, + "grad_norm": 0.22734573483467102, + "learning_rate": 0.00018922854548244402, + "loss": 0.039, + "step": 6369 + }, + { + "epoch": 0.1493228315078207, + "grad_norm": 0.4250650703907013, + "learning_rate": 0.0001892252196288049, + "loss": 0.1072, + "step": 6370 + }, + { + "epoch": 0.14934627308262569, + "grad_norm": 0.22391770780086517, + "learning_rate": 0.00018922189329102727, + "loss": 0.0638, + "step": 6371 + }, + { + "epoch": 0.14936971465743068, + "grad_norm": 0.24776123464107513, + "learning_rate": 0.00018921856646912912, + "loss": 0.0476, + "step": 6372 + }, + { + "epoch": 0.14939315623223567, + "grad_norm": 0.7599626779556274, + "learning_rate": 0.00018921523916312852, + "loss": 0.7352, + "step": 6373 + }, + { + "epoch": 0.14941659780704067, + "grad_norm": 0.41164642572402954, + "learning_rate": 0.00018921191137304352, + "loss": 0.1398, + "step": 6374 + }, + { + "epoch": 0.14944003938184566, + "grad_norm": 0.39118751883506775, + "learning_rate": 0.0001892085830988922, + "loss": 0.0613, + "step": 6375 + }, + { + "epoch": 0.14946348095665066, + "grad_norm": 0.30296647548675537, + "learning_rate": 0.0001892052543406926, + "loss": 0.0676, + "step": 6376 + }, + { + "epoch": 0.14948692253145565, + "grad_norm": 0.8190886974334717, + "learning_rate": 0.00018920192509846274, + "loss": 0.2034, + "step": 6377 + }, + { + "epoch": 0.14951036410626065, + "grad_norm": 0.4500477910041809, + "learning_rate": 0.00018919859537222075, + "loss": 0.0795, + "step": 6378 + }, + { + "epoch": 0.14953380568106564, + "grad_norm": 0.42288148403167725, + "learning_rate": 0.0001891952651619847, + "loss": 0.14, + "step": 6379 + }, + { + "epoch": 0.14955724725587066, + "grad_norm": 0.6011150479316711, + "learning_rate": 0.0001891919344677726, + "loss": 0.1607, + "step": 6380 + }, + { + "epoch": 0.14958068883067566, + "grad_norm": 0.24941104650497437, + "learning_rate": 0.00018918860328960254, + "loss": 0.0707, + "step": 6381 + }, + { + "epoch": 0.14960413040548065, + "grad_norm": 0.49194854497909546, + "learning_rate": 0.00018918527162749266, + "loss": 0.1638, + "step": 6382 + }, + { + "epoch": 0.14962757198028565, + "grad_norm": 0.5738462209701538, + "learning_rate": 0.00018918193948146093, + "loss": 0.1283, + "step": 6383 + }, + { + "epoch": 0.14965101355509064, + "grad_norm": 0.5218638777732849, + "learning_rate": 0.00018917860685152554, + "loss": 0.1037, + "step": 6384 + }, + { + "epoch": 0.14967445512989563, + "grad_norm": 0.5932311415672302, + "learning_rate": 0.00018917527373770452, + "loss": 0.1972, + "step": 6385 + }, + { + "epoch": 0.14969789670470063, + "grad_norm": 0.4566747844219208, + "learning_rate": 0.00018917194014001594, + "loss": 0.135, + "step": 6386 + }, + { + "epoch": 0.14972133827950562, + "grad_norm": 0.873294472694397, + "learning_rate": 0.0001891686060584779, + "loss": 0.1314, + "step": 6387 + }, + { + "epoch": 0.14974477985431062, + "grad_norm": 0.666894793510437, + "learning_rate": 0.00018916527149310852, + "loss": 0.2002, + "step": 6388 + }, + { + "epoch": 0.1497682214291156, + "grad_norm": 0.8261436820030212, + "learning_rate": 0.00018916193644392585, + "loss": 0.1823, + "step": 6389 + }, + { + "epoch": 0.1497916630039206, + "grad_norm": 0.5116464495658875, + "learning_rate": 0.000189158600910948, + "loss": 0.144, + "step": 6390 + }, + { + "epoch": 0.1498151045787256, + "grad_norm": 0.49607619643211365, + "learning_rate": 0.00018915526489419307, + "loss": 0.1339, + "step": 6391 + }, + { + "epoch": 0.1498385461535306, + "grad_norm": 0.4675305485725403, + "learning_rate": 0.0001891519283936792, + "loss": 0.1541, + "step": 6392 + }, + { + "epoch": 0.1498619877283356, + "grad_norm": 0.5183549523353577, + "learning_rate": 0.00018914859140942443, + "loss": 0.0831, + "step": 6393 + }, + { + "epoch": 0.14988542930314058, + "grad_norm": 0.3888775408267975, + "learning_rate": 0.0001891452539414469, + "loss": 0.103, + "step": 6394 + }, + { + "epoch": 0.14990887087794558, + "grad_norm": 0.4293849468231201, + "learning_rate": 0.0001891419159897647, + "loss": 0.5694, + "step": 6395 + }, + { + "epoch": 0.14993231245275057, + "grad_norm": 0.7741116881370544, + "learning_rate": 0.00018913857755439598, + "loss": 0.1314, + "step": 6396 + }, + { + "epoch": 0.14995575402755557, + "grad_norm": 0.7114346027374268, + "learning_rate": 0.00018913523863535887, + "loss": 0.187, + "step": 6397 + }, + { + "epoch": 0.14997919560236056, + "grad_norm": 0.7271052598953247, + "learning_rate": 0.0001891318992326714, + "loss": 0.1724, + "step": 6398 + }, + { + "epoch": 0.15000263717716555, + "grad_norm": 1.0278384685516357, + "learning_rate": 0.00018912855934635173, + "loss": 0.1421, + "step": 6399 + }, + { + "epoch": 0.15002607875197055, + "grad_norm": 0.36159607768058777, + "learning_rate": 0.00018912521897641803, + "loss": 0.0987, + "step": 6400 + }, + { + "epoch": 0.15004952032677554, + "grad_norm": 0.6789456009864807, + "learning_rate": 0.00018912187812288836, + "loss": 0.1146, + "step": 6401 + }, + { + "epoch": 0.15007296190158054, + "grad_norm": 1.041595697402954, + "learning_rate": 0.0001891185367857809, + "loss": 0.2071, + "step": 6402 + }, + { + "epoch": 0.15009640347638553, + "grad_norm": 0.4007265865802765, + "learning_rate": 0.00018911519496511372, + "loss": 0.1055, + "step": 6403 + }, + { + "epoch": 0.15011984505119053, + "grad_norm": 0.4703895151615143, + "learning_rate": 0.000189111852660905, + "loss": 0.1053, + "step": 6404 + }, + { + "epoch": 0.15014328662599555, + "grad_norm": 0.4644508361816406, + "learning_rate": 0.00018910850987317285, + "loss": 0.1615, + "step": 6405 + }, + { + "epoch": 0.15016672820080054, + "grad_norm": 0.523662805557251, + "learning_rate": 0.00018910516660193544, + "loss": 0.1137, + "step": 6406 + }, + { + "epoch": 0.15019016977560554, + "grad_norm": 0.44373399019241333, + "learning_rate": 0.00018910182284721086, + "loss": 0.1106, + "step": 6407 + }, + { + "epoch": 0.15021361135041053, + "grad_norm": 0.20548640191555023, + "learning_rate": 0.0001890984786090173, + "loss": 0.0526, + "step": 6408 + }, + { + "epoch": 0.15023705292521553, + "grad_norm": 0.42198997735977173, + "learning_rate": 0.00018909513388737289, + "loss": 0.1571, + "step": 6409 + }, + { + "epoch": 0.15026049450002052, + "grad_norm": 0.7710381746292114, + "learning_rate": 0.0001890917886822958, + "loss": 0.2079, + "step": 6410 + }, + { + "epoch": 0.1502839360748255, + "grad_norm": 0.36610332131385803, + "learning_rate": 0.00018908844299380415, + "loss": 0.0628, + "step": 6411 + }, + { + "epoch": 0.1503073776496305, + "grad_norm": 0.39063021540641785, + "learning_rate": 0.0001890850968219161, + "loss": 0.1035, + "step": 6412 + }, + { + "epoch": 0.1503308192244355, + "grad_norm": 0.10686293244361877, + "learning_rate": 0.0001890817501666498, + "loss": 0.0231, + "step": 6413 + }, + { + "epoch": 0.1503542607992405, + "grad_norm": 0.19998949766159058, + "learning_rate": 0.0001890784030280234, + "loss": 0.0407, + "step": 6414 + }, + { + "epoch": 0.1503777023740455, + "grad_norm": 0.6785634756088257, + "learning_rate": 0.00018907505540605513, + "loss": 0.1585, + "step": 6415 + }, + { + "epoch": 0.15040114394885049, + "grad_norm": 0.7255882620811462, + "learning_rate": 0.00018907170730076306, + "loss": 0.6253, + "step": 6416 + }, + { + "epoch": 0.15042458552365548, + "grad_norm": 0.7185360193252563, + "learning_rate": 0.00018906835871216544, + "loss": 0.1514, + "step": 6417 + }, + { + "epoch": 0.15044802709846047, + "grad_norm": 0.10431347787380219, + "learning_rate": 0.0001890650096402804, + "loss": 0.0206, + "step": 6418 + }, + { + "epoch": 0.15047146867326547, + "grad_norm": 0.5070340633392334, + "learning_rate": 0.0001890616600851261, + "loss": 0.1558, + "step": 6419 + }, + { + "epoch": 0.15049491024807046, + "grad_norm": 0.1735791563987732, + "learning_rate": 0.0001890583100467207, + "loss": 0.0346, + "step": 6420 + }, + { + "epoch": 0.15051835182287546, + "grad_norm": 0.6048739552497864, + "learning_rate": 0.00018905495952508243, + "loss": 0.2008, + "step": 6421 + }, + { + "epoch": 0.15054179339768045, + "grad_norm": 0.6137261986732483, + "learning_rate": 0.00018905160852022946, + "loss": 0.1406, + "step": 6422 + }, + { + "epoch": 0.15056523497248545, + "grad_norm": 0.4733929932117462, + "learning_rate": 0.00018904825703217992, + "loss": 0.1349, + "step": 6423 + }, + { + "epoch": 0.15058867654729044, + "grad_norm": 0.31657111644744873, + "learning_rate": 0.00018904490506095207, + "loss": 0.0778, + "step": 6424 + }, + { + "epoch": 0.15061211812209543, + "grad_norm": 0.5673781633377075, + "learning_rate": 0.00018904155260656403, + "loss": 0.1716, + "step": 6425 + }, + { + "epoch": 0.15063555969690043, + "grad_norm": 0.5762863159179688, + "learning_rate": 0.00018903819966903406, + "loss": 0.7051, + "step": 6426 + }, + { + "epoch": 0.15065900127170542, + "grad_norm": 0.14017337560653687, + "learning_rate": 0.00018903484624838027, + "loss": 0.0324, + "step": 6427 + }, + { + "epoch": 0.15068244284651042, + "grad_norm": 0.5749315023422241, + "learning_rate": 0.00018903149234462095, + "loss": 0.1458, + "step": 6428 + }, + { + "epoch": 0.1507058844213154, + "grad_norm": 1.1388108730316162, + "learning_rate": 0.00018902813795777424, + "loss": 0.3714, + "step": 6429 + }, + { + "epoch": 0.1507293259961204, + "grad_norm": 0.26646873354911804, + "learning_rate": 0.00018902478308785835, + "loss": 0.097, + "step": 6430 + }, + { + "epoch": 0.15075276757092543, + "grad_norm": 0.4199231266975403, + "learning_rate": 0.00018902142773489144, + "loss": 0.1595, + "step": 6431 + }, + { + "epoch": 0.15077620914573042, + "grad_norm": 0.5645947456359863, + "learning_rate": 0.0001890180718988918, + "loss": 0.2077, + "step": 6432 + }, + { + "epoch": 0.15079965072053542, + "grad_norm": 0.14997486770153046, + "learning_rate": 0.0001890147155798776, + "loss": 0.0449, + "step": 6433 + }, + { + "epoch": 0.1508230922953404, + "grad_norm": 0.4112611413002014, + "learning_rate": 0.00018901135877786703, + "loss": 0.0832, + "step": 6434 + }, + { + "epoch": 0.1508465338701454, + "grad_norm": 0.347827285528183, + "learning_rate": 0.00018900800149287833, + "loss": 0.0934, + "step": 6435 + }, + { + "epoch": 0.1508699754449504, + "grad_norm": 0.493827223777771, + "learning_rate": 0.00018900464372492972, + "loss": 0.1612, + "step": 6436 + }, + { + "epoch": 0.1508934170197554, + "grad_norm": 0.47959214448928833, + "learning_rate": 0.0001890012854740394, + "loss": 0.1355, + "step": 6437 + }, + { + "epoch": 0.1509168585945604, + "grad_norm": 0.3464865982532501, + "learning_rate": 0.0001889979267402256, + "loss": 0.1641, + "step": 6438 + }, + { + "epoch": 0.15094030016936538, + "grad_norm": 0.13887587189674377, + "learning_rate": 0.00018899456752350657, + "loss": 0.0282, + "step": 6439 + }, + { + "epoch": 0.15096374174417038, + "grad_norm": 0.189180389046669, + "learning_rate": 0.00018899120782390048, + "loss": 0.0472, + "step": 6440 + }, + { + "epoch": 0.15098718331897537, + "grad_norm": 0.45214614272117615, + "learning_rate": 0.0001889878476414256, + "loss": 0.1046, + "step": 6441 + }, + { + "epoch": 0.15101062489378037, + "grad_norm": 0.1456882804632187, + "learning_rate": 0.0001889844869761002, + "loss": 0.0276, + "step": 6442 + }, + { + "epoch": 0.15103406646858536, + "grad_norm": 0.7083305716514587, + "learning_rate": 0.00018898112582794242, + "loss": 0.2151, + "step": 6443 + }, + { + "epoch": 0.15105750804339035, + "grad_norm": 0.3272738456726074, + "learning_rate": 0.0001889777641969706, + "loss": 0.0994, + "step": 6444 + }, + { + "epoch": 0.15108094961819535, + "grad_norm": 0.6471911668777466, + "learning_rate": 0.0001889744020832029, + "loss": 0.17, + "step": 6445 + }, + { + "epoch": 0.15110439119300034, + "grad_norm": 0.468473345041275, + "learning_rate": 0.00018897103948665759, + "loss": 0.9086, + "step": 6446 + }, + { + "epoch": 0.15112783276780534, + "grad_norm": 0.40627235174179077, + "learning_rate": 0.0001889676764073529, + "loss": 0.078, + "step": 6447 + }, + { + "epoch": 0.15115127434261033, + "grad_norm": 0.10998660326004028, + "learning_rate": 0.00018896431284530713, + "loss": 0.0159, + "step": 6448 + }, + { + "epoch": 0.15117471591741533, + "grad_norm": 0.25392064452171326, + "learning_rate": 0.00018896094880053847, + "loss": 0.0792, + "step": 6449 + }, + { + "epoch": 0.15119815749222032, + "grad_norm": 0.4141767621040344, + "learning_rate": 0.00018895758427306522, + "loss": 0.1479, + "step": 6450 + }, + { + "epoch": 0.15122159906702531, + "grad_norm": 0.7785126566886902, + "learning_rate": 0.00018895421926290562, + "loss": 0.1333, + "step": 6451 + }, + { + "epoch": 0.1512450406418303, + "grad_norm": 0.504914402961731, + "learning_rate": 0.00018895085377007792, + "loss": 0.1141, + "step": 6452 + }, + { + "epoch": 0.1512684822166353, + "grad_norm": 0.49130064249038696, + "learning_rate": 0.00018894748779460039, + "loss": 0.1311, + "step": 6453 + }, + { + "epoch": 0.1512919237914403, + "grad_norm": 0.3456307649612427, + "learning_rate": 0.00018894412133649125, + "loss": 0.0747, + "step": 6454 + }, + { + "epoch": 0.1513153653662453, + "grad_norm": 0.5844643712043762, + "learning_rate": 0.00018894075439576885, + "loss": 0.1538, + "step": 6455 + }, + { + "epoch": 0.1513388069410503, + "grad_norm": 0.2842559218406677, + "learning_rate": 0.00018893738697245142, + "loss": 0.0614, + "step": 6456 + }, + { + "epoch": 0.1513622485158553, + "grad_norm": 0.7565361261367798, + "learning_rate": 0.0001889340190665572, + "loss": 0.4586, + "step": 6457 + }, + { + "epoch": 0.1513856900906603, + "grad_norm": 0.379155695438385, + "learning_rate": 0.00018893065067810452, + "loss": 0.1061, + "step": 6458 + }, + { + "epoch": 0.1514091316654653, + "grad_norm": 0.782687246799469, + "learning_rate": 0.0001889272818071116, + "loss": 0.3748, + "step": 6459 + }, + { + "epoch": 0.1514325732402703, + "grad_norm": 0.49088573455810547, + "learning_rate": 0.00018892391245359678, + "loss": 0.0896, + "step": 6460 + }, + { + "epoch": 0.15145601481507528, + "grad_norm": 0.4904979169368744, + "learning_rate": 0.00018892054261757828, + "loss": 0.0589, + "step": 6461 + }, + { + "epoch": 0.15147945638988028, + "grad_norm": 0.39430004358291626, + "learning_rate": 0.00018891717229907443, + "loss": 0.123, + "step": 6462 + }, + { + "epoch": 0.15150289796468527, + "grad_norm": 0.6835745573043823, + "learning_rate": 0.0001889138014981035, + "loss": 0.1662, + "step": 6463 + }, + { + "epoch": 0.15152633953949027, + "grad_norm": 0.6066717505455017, + "learning_rate": 0.0001889104302146838, + "loss": 0.1539, + "step": 6464 + }, + { + "epoch": 0.15154978111429526, + "grad_norm": 0.5313868522644043, + "learning_rate": 0.0001889070584488336, + "loss": 0.1143, + "step": 6465 + }, + { + "epoch": 0.15157322268910026, + "grad_norm": 0.20887966454029083, + "learning_rate": 0.00018890368620057117, + "loss": 0.0459, + "step": 6466 + }, + { + "epoch": 0.15159666426390525, + "grad_norm": 0.27384936809539795, + "learning_rate": 0.00018890031346991486, + "loss": 0.0818, + "step": 6467 + }, + { + "epoch": 0.15162010583871025, + "grad_norm": 0.7449420690536499, + "learning_rate": 0.00018889694025688296, + "loss": 0.1462, + "step": 6468 + }, + { + "epoch": 0.15164354741351524, + "grad_norm": 0.3977442979812622, + "learning_rate": 0.00018889356656149375, + "loss": 0.0973, + "step": 6469 + }, + { + "epoch": 0.15166698898832023, + "grad_norm": 0.37322449684143066, + "learning_rate": 0.00018889019238376556, + "loss": 0.0786, + "step": 6470 + }, + { + "epoch": 0.15169043056312523, + "grad_norm": 0.6133217215538025, + "learning_rate": 0.00018888681772371668, + "loss": 0.1391, + "step": 6471 + }, + { + "epoch": 0.15171387213793022, + "grad_norm": 0.5340884923934937, + "learning_rate": 0.00018888344258136542, + "loss": 0.1347, + "step": 6472 + }, + { + "epoch": 0.15173731371273522, + "grad_norm": 0.4109053909778595, + "learning_rate": 0.0001888800669567301, + "loss": 0.0586, + "step": 6473 + }, + { + "epoch": 0.1517607552875402, + "grad_norm": 0.1891183704137802, + "learning_rate": 0.00018887669084982904, + "loss": 0.0188, + "step": 6474 + }, + { + "epoch": 0.1517841968623452, + "grad_norm": 0.7658421993255615, + "learning_rate": 0.00018887331426068057, + "loss": 0.2404, + "step": 6475 + }, + { + "epoch": 0.1518076384371502, + "grad_norm": 0.6881532669067383, + "learning_rate": 0.00018886993718930298, + "loss": 0.2057, + "step": 6476 + }, + { + "epoch": 0.1518310800119552, + "grad_norm": 0.23836258053779602, + "learning_rate": 0.00018886655963571466, + "loss": 0.0697, + "step": 6477 + }, + { + "epoch": 0.1518545215867602, + "grad_norm": 0.4540603458881378, + "learning_rate": 0.00018886318159993382, + "loss": 0.0958, + "step": 6478 + }, + { + "epoch": 0.15187796316156518, + "grad_norm": 0.6543919444084167, + "learning_rate": 0.0001888598030819789, + "loss": 0.1607, + "step": 6479 + }, + { + "epoch": 0.15190140473637018, + "grad_norm": 0.8105286955833435, + "learning_rate": 0.00018885642408186817, + "loss": 0.1437, + "step": 6480 + }, + { + "epoch": 0.15192484631117517, + "grad_norm": 0.7271211743354797, + "learning_rate": 0.00018885304459961998, + "loss": 0.3466, + "step": 6481 + }, + { + "epoch": 0.1519482878859802, + "grad_norm": 0.9278488159179688, + "learning_rate": 0.0001888496646352527, + "loss": 0.1691, + "step": 6482 + }, + { + "epoch": 0.1519717294607852, + "grad_norm": 0.14840759336948395, + "learning_rate": 0.00018884628418878462, + "loss": 0.0477, + "step": 6483 + }, + { + "epoch": 0.15199517103559018, + "grad_norm": 0.17634116113185883, + "learning_rate": 0.0001888429032602341, + "loss": 0.0479, + "step": 6484 + }, + { + "epoch": 0.15201861261039518, + "grad_norm": 0.5316860675811768, + "learning_rate": 0.0001888395218496195, + "loss": 0.4911, + "step": 6485 + }, + { + "epoch": 0.15204205418520017, + "grad_norm": 0.5239275693893433, + "learning_rate": 0.00018883613995695914, + "loss": 0.1135, + "step": 6486 + }, + { + "epoch": 0.15206549576000516, + "grad_norm": 0.5633696913719177, + "learning_rate": 0.00018883275758227144, + "loss": 0.1349, + "step": 6487 + }, + { + "epoch": 0.15208893733481016, + "grad_norm": 0.6969264149665833, + "learning_rate": 0.00018882937472557466, + "loss": 0.6252, + "step": 6488 + }, + { + "epoch": 0.15211237890961515, + "grad_norm": 0.6180074214935303, + "learning_rate": 0.00018882599138688718, + "loss": 0.2218, + "step": 6489 + }, + { + "epoch": 0.15213582048442015, + "grad_norm": 0.32583165168762207, + "learning_rate": 0.0001888226075662274, + "loss": 0.1073, + "step": 6490 + }, + { + "epoch": 0.15215926205922514, + "grad_norm": 0.5683456659317017, + "learning_rate": 0.00018881922326361363, + "loss": 0.5127, + "step": 6491 + }, + { + "epoch": 0.15218270363403014, + "grad_norm": 0.7250357270240784, + "learning_rate": 0.00018881583847906428, + "loss": 0.1716, + "step": 6492 + }, + { + "epoch": 0.15220614520883513, + "grad_norm": 0.5433035492897034, + "learning_rate": 0.00018881245321259767, + "loss": 0.1481, + "step": 6493 + }, + { + "epoch": 0.15222958678364013, + "grad_norm": 0.34785765409469604, + "learning_rate": 0.0001888090674642322, + "loss": 0.1236, + "step": 6494 + }, + { + "epoch": 0.15225302835844512, + "grad_norm": 0.22638480365276337, + "learning_rate": 0.00018880568123398622, + "loss": 0.0449, + "step": 6495 + }, + { + "epoch": 0.1522764699332501, + "grad_norm": 0.6732659935951233, + "learning_rate": 0.00018880229452187817, + "loss": 0.8679, + "step": 6496 + }, + { + "epoch": 0.1522999115080551, + "grad_norm": 0.6904654502868652, + "learning_rate": 0.00018879890732792634, + "loss": 0.1082, + "step": 6497 + }, + { + "epoch": 0.1523233530828601, + "grad_norm": 0.3869714140892029, + "learning_rate": 0.00018879551965214912, + "loss": 0.0865, + "step": 6498 + }, + { + "epoch": 0.1523467946576651, + "grad_norm": 0.4819958209991455, + "learning_rate": 0.00018879213149456493, + "loss": 0.1153, + "step": 6499 + }, + { + "epoch": 0.1523702362324701, + "grad_norm": 0.5193347930908203, + "learning_rate": 0.00018878874285519216, + "loss": 0.6183, + "step": 6500 + }, + { + "epoch": 0.15239367780727509, + "grad_norm": 0.4150405526161194, + "learning_rate": 0.00018878535373404915, + "loss": 0.0997, + "step": 6501 + }, + { + "epoch": 0.15241711938208008, + "grad_norm": 0.5979143977165222, + "learning_rate": 0.00018878196413115431, + "loss": 0.8011, + "step": 6502 + }, + { + "epoch": 0.15244056095688507, + "grad_norm": 0.5658357739448547, + "learning_rate": 0.00018877857404652607, + "loss": 0.1116, + "step": 6503 + }, + { + "epoch": 0.15246400253169007, + "grad_norm": 0.6542536020278931, + "learning_rate": 0.00018877518348018276, + "loss": 0.3171, + "step": 6504 + }, + { + "epoch": 0.15248744410649506, + "grad_norm": 0.13190700113773346, + "learning_rate": 0.0001887717924321428, + "loss": 0.042, + "step": 6505 + }, + { + "epoch": 0.15251088568130006, + "grad_norm": 0.4390162229537964, + "learning_rate": 0.0001887684009024246, + "loss": 0.0949, + "step": 6506 + }, + { + "epoch": 0.15253432725610508, + "grad_norm": 0.18062768876552582, + "learning_rate": 0.0001887650088910466, + "loss": 0.0375, + "step": 6507 + }, + { + "epoch": 0.15255776883091007, + "grad_norm": 1.1004447937011719, + "learning_rate": 0.00018876161639802715, + "loss": 0.1358, + "step": 6508 + }, + { + "epoch": 0.15258121040571507, + "grad_norm": 0.41648754477500916, + "learning_rate": 0.00018875822342338466, + "loss": 0.0869, + "step": 6509 + }, + { + "epoch": 0.15260465198052006, + "grad_norm": 0.1625819355249405, + "learning_rate": 0.00018875482996713757, + "loss": 0.0122, + "step": 6510 + }, + { + "epoch": 0.15262809355532506, + "grad_norm": 0.5804142355918884, + "learning_rate": 0.00018875143602930424, + "loss": 0.1037, + "step": 6511 + }, + { + "epoch": 0.15265153513013005, + "grad_norm": 0.7825366854667664, + "learning_rate": 0.00018874804160990316, + "loss": 0.1598, + "step": 6512 + }, + { + "epoch": 0.15267497670493504, + "grad_norm": 0.4982525408267975, + "learning_rate": 0.0001887446467089527, + "loss": 0.1269, + "step": 6513 + }, + { + "epoch": 0.15269841827974004, + "grad_norm": 0.7198958396911621, + "learning_rate": 0.0001887412513264713, + "loss": 0.276, + "step": 6514 + }, + { + "epoch": 0.15272185985454503, + "grad_norm": 0.5519269108772278, + "learning_rate": 0.00018873785546247735, + "loss": 0.144, + "step": 6515 + }, + { + "epoch": 0.15274530142935003, + "grad_norm": 0.7246097326278687, + "learning_rate": 0.00018873445911698933, + "loss": 0.2418, + "step": 6516 + }, + { + "epoch": 0.15276874300415502, + "grad_norm": 0.7309103012084961, + "learning_rate": 0.00018873106229002563, + "loss": 0.1867, + "step": 6517 + }, + { + "epoch": 0.15279218457896002, + "grad_norm": 0.5653204321861267, + "learning_rate": 0.0001887276649816047, + "loss": 0.1709, + "step": 6518 + }, + { + "epoch": 0.152815626153765, + "grad_norm": 0.7423766851425171, + "learning_rate": 0.00018872426719174492, + "loss": 0.2026, + "step": 6519 + }, + { + "epoch": 0.15283906772857, + "grad_norm": 0.48512908816337585, + "learning_rate": 0.00018872086892046482, + "loss": 0.1157, + "step": 6520 + }, + { + "epoch": 0.152862509303375, + "grad_norm": 0.5993993282318115, + "learning_rate": 0.00018871747016778277, + "loss": 0.1075, + "step": 6521 + }, + { + "epoch": 0.15288595087818, + "grad_norm": 0.23056839406490326, + "learning_rate": 0.00018871407093371722, + "loss": 0.0696, + "step": 6522 + }, + { + "epoch": 0.152909392452985, + "grad_norm": 0.18194544315338135, + "learning_rate": 0.00018871067121828665, + "loss": 0.0408, + "step": 6523 + }, + { + "epoch": 0.15293283402778998, + "grad_norm": 0.6245230436325073, + "learning_rate": 0.00018870727102150947, + "loss": 0.2504, + "step": 6524 + }, + { + "epoch": 0.15295627560259498, + "grad_norm": 0.5382835268974304, + "learning_rate": 0.00018870387034340417, + "loss": 0.0819, + "step": 6525 + }, + { + "epoch": 0.15297971717739997, + "grad_norm": 0.1819489300251007, + "learning_rate": 0.00018870046918398914, + "loss": 0.0837, + "step": 6526 + }, + { + "epoch": 0.15300315875220497, + "grad_norm": 0.4035559296607971, + "learning_rate": 0.0001886970675432829, + "loss": 0.0915, + "step": 6527 + }, + { + "epoch": 0.15302660032700996, + "grad_norm": 0.5286772847175598, + "learning_rate": 0.00018869366542130386, + "loss": 0.134, + "step": 6528 + }, + { + "epoch": 0.15305004190181495, + "grad_norm": 0.133855938911438, + "learning_rate": 0.00018869026281807049, + "loss": 0.0389, + "step": 6529 + }, + { + "epoch": 0.15307348347661995, + "grad_norm": 0.1926535964012146, + "learning_rate": 0.0001886868597336013, + "loss": 0.0197, + "step": 6530 + }, + { + "epoch": 0.15309692505142494, + "grad_norm": 0.6229070425033569, + "learning_rate": 0.00018868345616791468, + "loss": 0.1948, + "step": 6531 + }, + { + "epoch": 0.15312036662622994, + "grad_norm": 0.4782978594303131, + "learning_rate": 0.00018868005212102913, + "loss": 0.0838, + "step": 6532 + }, + { + "epoch": 0.15314380820103496, + "grad_norm": 0.5489671230316162, + "learning_rate": 0.00018867664759296315, + "loss": 0.5724, + "step": 6533 + }, + { + "epoch": 0.15316724977583995, + "grad_norm": 0.2427992969751358, + "learning_rate": 0.00018867324258373515, + "loss": 0.0793, + "step": 6534 + }, + { + "epoch": 0.15319069135064495, + "grad_norm": 0.39294499158859253, + "learning_rate": 0.00018866983709336368, + "loss": 0.1241, + "step": 6535 + }, + { + "epoch": 0.15321413292544994, + "grad_norm": 0.7496275901794434, + "learning_rate": 0.00018866643112186714, + "loss": 0.0939, + "step": 6536 + }, + { + "epoch": 0.15323757450025494, + "grad_norm": 0.16665950417518616, + "learning_rate": 0.0001886630246692641, + "loss": 0.0263, + "step": 6537 + }, + { + "epoch": 0.15326101607505993, + "grad_norm": 0.7124923467636108, + "learning_rate": 0.00018865961773557296, + "loss": 0.1255, + "step": 6538 + }, + { + "epoch": 0.15328445764986492, + "grad_norm": 0.2378113567829132, + "learning_rate": 0.00018865621032081227, + "loss": 0.0651, + "step": 6539 + }, + { + "epoch": 0.15330789922466992, + "grad_norm": 0.2572391927242279, + "learning_rate": 0.00018865280242500045, + "loss": 0.0533, + "step": 6540 + }, + { + "epoch": 0.1533313407994749, + "grad_norm": 0.5820692181587219, + "learning_rate": 0.00018864939404815605, + "loss": 0.1329, + "step": 6541 + }, + { + "epoch": 0.1533547823742799, + "grad_norm": 0.34537264704704285, + "learning_rate": 0.00018864598519029757, + "loss": 0.0827, + "step": 6542 + }, + { + "epoch": 0.1533782239490849, + "grad_norm": 0.7532632350921631, + "learning_rate": 0.00018864257585144345, + "loss": 0.1431, + "step": 6543 + }, + { + "epoch": 0.1534016655238899, + "grad_norm": 0.25563907623291016, + "learning_rate": 0.00018863916603161223, + "loss": 0.0702, + "step": 6544 + }, + { + "epoch": 0.1534251070986949, + "grad_norm": 0.21549132466316223, + "learning_rate": 0.0001886357557308224, + "loss": 0.0637, + "step": 6545 + }, + { + "epoch": 0.15344854867349988, + "grad_norm": 0.6092728972434998, + "learning_rate": 0.00018863234494909248, + "loss": 0.1218, + "step": 6546 + }, + { + "epoch": 0.15347199024830488, + "grad_norm": 0.34257790446281433, + "learning_rate": 0.00018862893368644094, + "loss": 0.0331, + "step": 6547 + }, + { + "epoch": 0.15349543182310987, + "grad_norm": 0.17298997938632965, + "learning_rate": 0.00018862552194288634, + "loss": 0.0573, + "step": 6548 + }, + { + "epoch": 0.15351887339791487, + "grad_norm": 0.6419896483421326, + "learning_rate": 0.00018862210971844716, + "loss": 0.0863, + "step": 6549 + }, + { + "epoch": 0.15354231497271986, + "grad_norm": 0.502442479133606, + "learning_rate": 0.0001886186970131419, + "loss": 0.1637, + "step": 6550 + }, + { + "epoch": 0.15356575654752486, + "grad_norm": 0.6313626766204834, + "learning_rate": 0.00018861528382698914, + "loss": 0.2162, + "step": 6551 + }, + { + "epoch": 0.15358919812232985, + "grad_norm": 0.3504679799079895, + "learning_rate": 0.00018861187016000732, + "loss": 0.1024, + "step": 6552 + }, + { + "epoch": 0.15361263969713485, + "grad_norm": 0.2879413366317749, + "learning_rate": 0.00018860845601221502, + "loss": 0.068, + "step": 6553 + }, + { + "epoch": 0.15363608127193984, + "grad_norm": 0.7589139342308044, + "learning_rate": 0.00018860504138363076, + "loss": 0.8768, + "step": 6554 + }, + { + "epoch": 0.15365952284674483, + "grad_norm": 0.25977984070777893, + "learning_rate": 0.000188601626274273, + "loss": 0.0566, + "step": 6555 + }, + { + "epoch": 0.15368296442154983, + "grad_norm": 0.835725724697113, + "learning_rate": 0.00018859821068416036, + "loss": 0.1933, + "step": 6556 + }, + { + "epoch": 0.15370640599635482, + "grad_norm": 0.6620938777923584, + "learning_rate": 0.00018859479461331137, + "loss": 0.168, + "step": 6557 + }, + { + "epoch": 0.15372984757115982, + "grad_norm": 0.9637582898139954, + "learning_rate": 0.0001885913780617445, + "loss": 0.152, + "step": 6558 + }, + { + "epoch": 0.15375328914596484, + "grad_norm": 0.4273906648159027, + "learning_rate": 0.0001885879610294783, + "loss": 0.0939, + "step": 6559 + }, + { + "epoch": 0.15377673072076983, + "grad_norm": 0.18148483335971832, + "learning_rate": 0.00018858454351653134, + "loss": 0.0605, + "step": 6560 + }, + { + "epoch": 0.15380017229557483, + "grad_norm": 0.5307674407958984, + "learning_rate": 0.00018858112552292216, + "loss": 0.121, + "step": 6561 + }, + { + "epoch": 0.15382361387037982, + "grad_norm": 0.1572244018316269, + "learning_rate": 0.00018857770704866936, + "loss": 0.0495, + "step": 6562 + }, + { + "epoch": 0.15384705544518482, + "grad_norm": 0.20306973159313202, + "learning_rate": 0.00018857428809379136, + "loss": 0.055, + "step": 6563 + }, + { + "epoch": 0.1538704970199898, + "grad_norm": 0.6964028477668762, + "learning_rate": 0.0001885708686583068, + "loss": 0.6112, + "step": 6564 + }, + { + "epoch": 0.1538939385947948, + "grad_norm": 0.20130790770053864, + "learning_rate": 0.00018856744874223423, + "loss": 0.0548, + "step": 6565 + }, + { + "epoch": 0.1539173801695998, + "grad_norm": 0.7175066471099854, + "learning_rate": 0.00018856402834559218, + "loss": 0.1705, + "step": 6566 + }, + { + "epoch": 0.1539408217444048, + "grad_norm": 0.17389298975467682, + "learning_rate": 0.00018856060746839923, + "loss": 0.0358, + "step": 6567 + }, + { + "epoch": 0.1539642633192098, + "grad_norm": 0.34764471650123596, + "learning_rate": 0.00018855718611067394, + "loss": 0.1115, + "step": 6568 + }, + { + "epoch": 0.15398770489401478, + "grad_norm": 0.6708670854568481, + "learning_rate": 0.00018855376427243486, + "loss": 0.183, + "step": 6569 + }, + { + "epoch": 0.15401114646881978, + "grad_norm": 0.5928683876991272, + "learning_rate": 0.00018855034195370055, + "loss": 0.6505, + "step": 6570 + }, + { + "epoch": 0.15403458804362477, + "grad_norm": 0.42326459288597107, + "learning_rate": 0.0001885469191544896, + "loss": 0.759, + "step": 6571 + }, + { + "epoch": 0.15405802961842976, + "grad_norm": 0.15652257204055786, + "learning_rate": 0.0001885434958748206, + "loss": 0.0436, + "step": 6572 + }, + { + "epoch": 0.15408147119323476, + "grad_norm": 0.47644877433776855, + "learning_rate": 0.0001885400721147121, + "loss": 0.2053, + "step": 6573 + }, + { + "epoch": 0.15410491276803975, + "grad_norm": 1.1230367422103882, + "learning_rate": 0.00018853664787418266, + "loss": 0.195, + "step": 6574 + }, + { + "epoch": 0.15412835434284475, + "grad_norm": 0.5368688106536865, + "learning_rate": 0.00018853322315325087, + "loss": 0.1746, + "step": 6575 + }, + { + "epoch": 0.15415179591764974, + "grad_norm": 0.11012569814920425, + "learning_rate": 0.00018852979795193535, + "loss": 0.0192, + "step": 6576 + }, + { + "epoch": 0.15417523749245474, + "grad_norm": 0.7417455911636353, + "learning_rate": 0.00018852637227025464, + "loss": 0.19, + "step": 6577 + }, + { + "epoch": 0.15419867906725973, + "grad_norm": 0.45843809843063354, + "learning_rate": 0.00018852294610822735, + "loss": 0.1332, + "step": 6578 + }, + { + "epoch": 0.15422212064206473, + "grad_norm": 0.6881541013717651, + "learning_rate": 0.00018851951946587207, + "loss": 0.1946, + "step": 6579 + }, + { + "epoch": 0.15424556221686972, + "grad_norm": 0.6564016342163086, + "learning_rate": 0.0001885160923432074, + "loss": 0.188, + "step": 6580 + }, + { + "epoch": 0.1542690037916747, + "grad_norm": 0.1945473998785019, + "learning_rate": 0.00018851266474025188, + "loss": 0.0242, + "step": 6581 + }, + { + "epoch": 0.1542924453664797, + "grad_norm": 0.5084395408630371, + "learning_rate": 0.0001885092366570242, + "loss": 0.1418, + "step": 6582 + }, + { + "epoch": 0.1543158869412847, + "grad_norm": 0.7208613157272339, + "learning_rate": 0.00018850580809354287, + "loss": 0.2032, + "step": 6583 + }, + { + "epoch": 0.15433932851608972, + "grad_norm": 0.5877615213394165, + "learning_rate": 0.00018850237904982658, + "loss": 0.1304, + "step": 6584 + }, + { + "epoch": 0.15436277009089472, + "grad_norm": 0.5785815715789795, + "learning_rate": 0.00018849894952589386, + "loss": 0.1872, + "step": 6585 + }, + { + "epoch": 0.1543862116656997, + "grad_norm": 0.44653645157814026, + "learning_rate": 0.00018849551952176337, + "loss": 0.0852, + "step": 6586 + }, + { + "epoch": 0.1544096532405047, + "grad_norm": 0.7739850878715515, + "learning_rate": 0.0001884920890374537, + "loss": 0.1625, + "step": 6587 + }, + { + "epoch": 0.1544330948153097, + "grad_norm": 0.29667407274246216, + "learning_rate": 0.00018848865807298345, + "loss": 0.0758, + "step": 6588 + }, + { + "epoch": 0.1544565363901147, + "grad_norm": 0.25716283917427063, + "learning_rate": 0.0001884852266283713, + "loss": 0.0614, + "step": 6589 + }, + { + "epoch": 0.1544799779649197, + "grad_norm": 0.1710757166147232, + "learning_rate": 0.00018848179470363577, + "loss": 0.0486, + "step": 6590 + }, + { + "epoch": 0.15450341953972468, + "grad_norm": 0.49425581097602844, + "learning_rate": 0.00018847836229879557, + "loss": 0.1241, + "step": 6591 + }, + { + "epoch": 0.15452686111452968, + "grad_norm": 0.7983647584915161, + "learning_rate": 0.00018847492941386924, + "loss": 0.283, + "step": 6592 + }, + { + "epoch": 0.15455030268933467, + "grad_norm": 0.4484703242778778, + "learning_rate": 0.0001884714960488755, + "loss": 0.1982, + "step": 6593 + }, + { + "epoch": 0.15457374426413967, + "grad_norm": 0.7329010367393494, + "learning_rate": 0.00018846806220383292, + "loss": 0.1889, + "step": 6594 + }, + { + "epoch": 0.15459718583894466, + "grad_norm": 0.32152917981147766, + "learning_rate": 0.00018846462787876016, + "loss": 0.0625, + "step": 6595 + }, + { + "epoch": 0.15462062741374966, + "grad_norm": 0.3760320544242859, + "learning_rate": 0.00018846119307367582, + "loss": 0.1143, + "step": 6596 + }, + { + "epoch": 0.15464406898855465, + "grad_norm": 0.46336954832077026, + "learning_rate": 0.00018845775778859856, + "loss": 0.1064, + "step": 6597 + }, + { + "epoch": 0.15466751056335964, + "grad_norm": 0.6060660481452942, + "learning_rate": 0.00018845432202354702, + "loss": 0.1373, + "step": 6598 + }, + { + "epoch": 0.15469095213816464, + "grad_norm": 0.9844868183135986, + "learning_rate": 0.00018845088577853987, + "loss": 0.7381, + "step": 6599 + }, + { + "epoch": 0.15471439371296963, + "grad_norm": 0.1529807597398758, + "learning_rate": 0.0001884474490535957, + "loss": 0.0185, + "step": 6600 + }, + { + "epoch": 0.15473783528777463, + "grad_norm": 0.6818338632583618, + "learning_rate": 0.0001884440118487332, + "loss": 0.1988, + "step": 6601 + }, + { + "epoch": 0.15476127686257962, + "grad_norm": 0.29689112305641174, + "learning_rate": 0.00018844057416397098, + "loss": 0.0662, + "step": 6602 + }, + { + "epoch": 0.15478471843738462, + "grad_norm": 0.7393508553504944, + "learning_rate": 0.00018843713599932775, + "loss": 0.2003, + "step": 6603 + }, + { + "epoch": 0.1548081600121896, + "grad_norm": 0.5104969143867493, + "learning_rate": 0.0001884336973548221, + "loss": 0.4723, + "step": 6604 + }, + { + "epoch": 0.1548316015869946, + "grad_norm": 0.17451755702495575, + "learning_rate": 0.00018843025823047277, + "loss": 0.0417, + "step": 6605 + }, + { + "epoch": 0.1548550431617996, + "grad_norm": 0.4431343078613281, + "learning_rate": 0.00018842681862629834, + "loss": 0.0892, + "step": 6606 + }, + { + "epoch": 0.1548784847366046, + "grad_norm": 0.5300533771514893, + "learning_rate": 0.0001884233785423175, + "loss": 0.667, + "step": 6607 + }, + { + "epoch": 0.1549019263114096, + "grad_norm": 0.6109365820884705, + "learning_rate": 0.00018841993797854892, + "loss": 0.1444, + "step": 6608 + }, + { + "epoch": 0.15492536788621458, + "grad_norm": 0.4099035859107971, + "learning_rate": 0.0001884164969350113, + "loss": 0.0919, + "step": 6609 + }, + { + "epoch": 0.1549488094610196, + "grad_norm": 0.6166794300079346, + "learning_rate": 0.00018841305541172324, + "loss": 0.0701, + "step": 6610 + }, + { + "epoch": 0.1549722510358246, + "grad_norm": 0.847571849822998, + "learning_rate": 0.0001884096134087035, + "loss": 0.231, + "step": 6611 + }, + { + "epoch": 0.1549956926106296, + "grad_norm": 0.4566146731376648, + "learning_rate": 0.00018840617092597065, + "loss": 0.0813, + "step": 6612 + }, + { + "epoch": 0.1550191341854346, + "grad_norm": 0.452371209859848, + "learning_rate": 0.0001884027279635435, + "loss": 0.1484, + "step": 6613 + }, + { + "epoch": 0.15504257576023958, + "grad_norm": 0.5564781427383423, + "learning_rate": 0.00018839928452144062, + "loss": 0.2152, + "step": 6614 + }, + { + "epoch": 0.15506601733504458, + "grad_norm": 0.5058563947677612, + "learning_rate": 0.00018839584059968077, + "loss": 0.0908, + "step": 6615 + }, + { + "epoch": 0.15508945890984957, + "grad_norm": 0.4800446331501007, + "learning_rate": 0.00018839239619828258, + "loss": 0.1522, + "step": 6616 + }, + { + "epoch": 0.15511290048465456, + "grad_norm": 0.5658813118934631, + "learning_rate": 0.00018838895131726476, + "loss": 0.1042, + "step": 6617 + }, + { + "epoch": 0.15513634205945956, + "grad_norm": 0.4539608955383301, + "learning_rate": 0.000188385505956646, + "loss": 0.099, + "step": 6618 + }, + { + "epoch": 0.15515978363426455, + "grad_norm": 0.2400745004415512, + "learning_rate": 0.00018838206011644498, + "loss": 0.0641, + "step": 6619 + }, + { + "epoch": 0.15518322520906955, + "grad_norm": 0.4516720175743103, + "learning_rate": 0.00018837861379668048, + "loss": 0.1533, + "step": 6620 + }, + { + "epoch": 0.15520666678387454, + "grad_norm": 0.4045311510562897, + "learning_rate": 0.0001883751669973711, + "loss": 0.1336, + "step": 6621 + }, + { + "epoch": 0.15523010835867954, + "grad_norm": 0.5165786147117615, + "learning_rate": 0.00018837171971853559, + "loss": 0.154, + "step": 6622 + }, + { + "epoch": 0.15525354993348453, + "grad_norm": 0.6111255288124084, + "learning_rate": 0.00018836827196019263, + "loss": 0.1187, + "step": 6623 + }, + { + "epoch": 0.15527699150828952, + "grad_norm": 0.6637106537818909, + "learning_rate": 0.00018836482372236094, + "loss": 0.2419, + "step": 6624 + }, + { + "epoch": 0.15530043308309452, + "grad_norm": 0.5788123607635498, + "learning_rate": 0.00018836137500505924, + "loss": 0.1794, + "step": 6625 + }, + { + "epoch": 0.1553238746578995, + "grad_norm": 0.5193964242935181, + "learning_rate": 0.0001883579258083062, + "loss": 0.1259, + "step": 6626 + }, + { + "epoch": 0.1553473162327045, + "grad_norm": 0.600354790687561, + "learning_rate": 0.00018835447613212063, + "loss": 0.2358, + "step": 6627 + }, + { + "epoch": 0.1553707578075095, + "grad_norm": 0.23194949328899384, + "learning_rate": 0.00018835102597652116, + "loss": 0.051, + "step": 6628 + }, + { + "epoch": 0.1553941993823145, + "grad_norm": 0.44358009099960327, + "learning_rate": 0.00018834757534152654, + "loss": 0.1001, + "step": 6629 + }, + { + "epoch": 0.1554176409571195, + "grad_norm": 0.20811058580875397, + "learning_rate": 0.00018834412422715548, + "loss": 0.0271, + "step": 6630 + }, + { + "epoch": 0.15544108253192448, + "grad_norm": 0.5526291131973267, + "learning_rate": 0.00018834067263342674, + "loss": 0.6919, + "step": 6631 + }, + { + "epoch": 0.15546452410672948, + "grad_norm": 0.6031337976455688, + "learning_rate": 0.00018833722056035902, + "loss": 0.2126, + "step": 6632 + }, + { + "epoch": 0.15548796568153447, + "grad_norm": 0.5963051319122314, + "learning_rate": 0.00018833376800797105, + "loss": 0.2019, + "step": 6633 + }, + { + "epoch": 0.15551140725633947, + "grad_norm": 0.2577495276927948, + "learning_rate": 0.00018833031497628158, + "loss": 0.1023, + "step": 6634 + }, + { + "epoch": 0.1555348488311445, + "grad_norm": 0.7766934633255005, + "learning_rate": 0.00018832686146530932, + "loss": 0.217, + "step": 6635 + }, + { + "epoch": 0.15555829040594948, + "grad_norm": 0.2588687241077423, + "learning_rate": 0.00018832340747507303, + "loss": 0.0344, + "step": 6636 + }, + { + "epoch": 0.15558173198075448, + "grad_norm": 0.695566713809967, + "learning_rate": 0.00018831995300559149, + "loss": 0.1672, + "step": 6637 + }, + { + "epoch": 0.15560517355555947, + "grad_norm": 0.35094621777534485, + "learning_rate": 0.00018831649805688336, + "loss": 0.0484, + "step": 6638 + }, + { + "epoch": 0.15562861513036447, + "grad_norm": 0.3709966540336609, + "learning_rate": 0.00018831304262896742, + "loss": 0.0626, + "step": 6639 + }, + { + "epoch": 0.15565205670516946, + "grad_norm": 0.4728020131587982, + "learning_rate": 0.00018830958672186245, + "loss": 0.109, + "step": 6640 + }, + { + "epoch": 0.15567549827997446, + "grad_norm": 0.36588767170906067, + "learning_rate": 0.00018830613033558717, + "loss": 0.0934, + "step": 6641 + }, + { + "epoch": 0.15569893985477945, + "grad_norm": 0.6611272692680359, + "learning_rate": 0.00018830267347016036, + "loss": 0.1633, + "step": 6642 + }, + { + "epoch": 0.15572238142958444, + "grad_norm": 0.3631906807422638, + "learning_rate": 0.00018829921612560074, + "loss": 0.0514, + "step": 6643 + }, + { + "epoch": 0.15574582300438944, + "grad_norm": 0.3224336504936218, + "learning_rate": 0.00018829575830192708, + "loss": 0.0538, + "step": 6644 + }, + { + "epoch": 0.15576926457919443, + "grad_norm": 0.3639262914657593, + "learning_rate": 0.0001882922999991582, + "loss": 0.0708, + "step": 6645 + }, + { + "epoch": 0.15579270615399943, + "grad_norm": 0.708806574344635, + "learning_rate": 0.00018828884121731282, + "loss": 0.1358, + "step": 6646 + }, + { + "epoch": 0.15581614772880442, + "grad_norm": 0.7912581562995911, + "learning_rate": 0.00018828538195640965, + "loss": 0.2133, + "step": 6647 + }, + { + "epoch": 0.15583958930360942, + "grad_norm": 0.7392577528953552, + "learning_rate": 0.00018828192221646752, + "loss": 0.1358, + "step": 6648 + }, + { + "epoch": 0.1558630308784144, + "grad_norm": 0.6583675742149353, + "learning_rate": 0.00018827846199750522, + "loss": 0.7563, + "step": 6649 + }, + { + "epoch": 0.1558864724532194, + "grad_norm": 0.5432679653167725, + "learning_rate": 0.0001882750012995415, + "loss": 0.6137, + "step": 6650 + }, + { + "epoch": 0.1559099140280244, + "grad_norm": 0.6877303719520569, + "learning_rate": 0.00018827154012259517, + "loss": 0.1855, + "step": 6651 + }, + { + "epoch": 0.1559333556028294, + "grad_norm": 0.39024779200553894, + "learning_rate": 0.00018826807846668495, + "loss": 0.1139, + "step": 6652 + }, + { + "epoch": 0.1559567971776344, + "grad_norm": 0.6391832828521729, + "learning_rate": 0.00018826461633182965, + "loss": 0.1213, + "step": 6653 + }, + { + "epoch": 0.15598023875243938, + "grad_norm": 0.42085129022598267, + "learning_rate": 0.00018826115371804803, + "loss": 0.0632, + "step": 6654 + }, + { + "epoch": 0.15600368032724438, + "grad_norm": 0.18680019676685333, + "learning_rate": 0.00018825769062535898, + "loss": 0.0282, + "step": 6655 + }, + { + "epoch": 0.15602712190204937, + "grad_norm": 0.33097681403160095, + "learning_rate": 0.00018825422705378116, + "loss": 0.0794, + "step": 6656 + }, + { + "epoch": 0.15605056347685436, + "grad_norm": 0.41968461871147156, + "learning_rate": 0.00018825076300333347, + "loss": 0.1163, + "step": 6657 + }, + { + "epoch": 0.15607400505165936, + "grad_norm": 0.5258558392524719, + "learning_rate": 0.0001882472984740346, + "loss": 0.1096, + "step": 6658 + }, + { + "epoch": 0.15609744662646435, + "grad_norm": 0.4750383198261261, + "learning_rate": 0.0001882438334659035, + "loss": 0.1215, + "step": 6659 + }, + { + "epoch": 0.15612088820126935, + "grad_norm": 0.5000057220458984, + "learning_rate": 0.0001882403679789588, + "loss": 0.1149, + "step": 6660 + }, + { + "epoch": 0.15614432977607437, + "grad_norm": 0.6260259747505188, + "learning_rate": 0.0001882369020132194, + "loss": 0.1343, + "step": 6661 + }, + { + "epoch": 0.15616777135087936, + "grad_norm": 0.3130440413951874, + "learning_rate": 0.0001882334355687041, + "loss": 0.054, + "step": 6662 + }, + { + "epoch": 0.15619121292568436, + "grad_norm": 0.632876455783844, + "learning_rate": 0.00018822996864543168, + "loss": 0.2197, + "step": 6663 + }, + { + "epoch": 0.15621465450048935, + "grad_norm": 0.42453518509864807, + "learning_rate": 0.00018822650124342098, + "loss": 0.0698, + "step": 6664 + }, + { + "epoch": 0.15623809607529435, + "grad_norm": 0.39459481835365295, + "learning_rate": 0.0001882230333626908, + "loss": 0.113, + "step": 6665 + }, + { + "epoch": 0.15626153765009934, + "grad_norm": 0.4631701111793518, + "learning_rate": 0.00018821956500325997, + "loss": 0.1826, + "step": 6666 + }, + { + "epoch": 0.15628497922490434, + "grad_norm": 0.3054182827472687, + "learning_rate": 0.00018821609616514727, + "loss": 0.0581, + "step": 6667 + }, + { + "epoch": 0.15630842079970933, + "grad_norm": 0.4971144497394562, + "learning_rate": 0.00018821262684837157, + "loss": 0.6742, + "step": 6668 + }, + { + "epoch": 0.15633186237451432, + "grad_norm": 0.6023528575897217, + "learning_rate": 0.0001882091570529517, + "loss": 0.1402, + "step": 6669 + }, + { + "epoch": 0.15635530394931932, + "grad_norm": 0.5526311993598938, + "learning_rate": 0.00018820568677890644, + "loss": 0.5874, + "step": 6670 + }, + { + "epoch": 0.1563787455241243, + "grad_norm": 0.3445347845554352, + "learning_rate": 0.00018820221602625466, + "loss": 0.0956, + "step": 6671 + }, + { + "epoch": 0.1564021870989293, + "grad_norm": 0.5560101270675659, + "learning_rate": 0.00018819874479501515, + "loss": 0.1435, + "step": 6672 + }, + { + "epoch": 0.1564256286737343, + "grad_norm": 0.6848194003105164, + "learning_rate": 0.0001881952730852068, + "loss": 0.1442, + "step": 6673 + }, + { + "epoch": 0.1564490702485393, + "grad_norm": 0.5507950782775879, + "learning_rate": 0.0001881918008968484, + "loss": 0.2109, + "step": 6674 + }, + { + "epoch": 0.1564725118233443, + "grad_norm": 0.6272192597389221, + "learning_rate": 0.0001881883282299588, + "loss": 0.2366, + "step": 6675 + }, + { + "epoch": 0.15649595339814928, + "grad_norm": 0.2140696495771408, + "learning_rate": 0.00018818485508455686, + "loss": 0.0749, + "step": 6676 + }, + { + "epoch": 0.15651939497295428, + "grad_norm": 0.47675859928131104, + "learning_rate": 0.00018818138146066142, + "loss": 0.1728, + "step": 6677 + }, + { + "epoch": 0.15654283654775927, + "grad_norm": 0.34340009093284607, + "learning_rate": 0.00018817790735829132, + "loss": 0.1059, + "step": 6678 + }, + { + "epoch": 0.15656627812256427, + "grad_norm": 0.5716361403465271, + "learning_rate": 0.0001881744327774654, + "loss": 0.133, + "step": 6679 + }, + { + "epoch": 0.15658971969736926, + "grad_norm": 0.3258032500743866, + "learning_rate": 0.00018817095771820258, + "loss": 0.0757, + "step": 6680 + }, + { + "epoch": 0.15661316127217426, + "grad_norm": 0.39236417412757874, + "learning_rate": 0.00018816748218052163, + "loss": 0.0817, + "step": 6681 + }, + { + "epoch": 0.15663660284697925, + "grad_norm": 0.2791128158569336, + "learning_rate": 0.00018816400616444146, + "loss": 0.1172, + "step": 6682 + }, + { + "epoch": 0.15666004442178424, + "grad_norm": 1.0088415145874023, + "learning_rate": 0.00018816052966998088, + "loss": 0.1674, + "step": 6683 + }, + { + "epoch": 0.15668348599658924, + "grad_norm": 0.4226864278316498, + "learning_rate": 0.00018815705269715883, + "loss": 0.1252, + "step": 6684 + }, + { + "epoch": 0.15670692757139423, + "grad_norm": 0.5018945336341858, + "learning_rate": 0.00018815357524599408, + "loss": 0.1147, + "step": 6685 + }, + { + "epoch": 0.15673036914619926, + "grad_norm": 0.6152146458625793, + "learning_rate": 0.0001881500973165056, + "loss": 0.1091, + "step": 6686 + }, + { + "epoch": 0.15675381072100425, + "grad_norm": 0.46020573377609253, + "learning_rate": 0.00018814661890871216, + "loss": 0.1013, + "step": 6687 + }, + { + "epoch": 0.15677725229580924, + "grad_norm": 0.37882092595100403, + "learning_rate": 0.00018814314002263275, + "loss": 0.0917, + "step": 6688 + }, + { + "epoch": 0.15680069387061424, + "grad_norm": 0.9742274284362793, + "learning_rate": 0.00018813966065828614, + "loss": 0.1519, + "step": 6689 + }, + { + "epoch": 0.15682413544541923, + "grad_norm": 0.4939063787460327, + "learning_rate": 0.00018813618081569127, + "loss": 0.1009, + "step": 6690 + }, + { + "epoch": 0.15684757702022423, + "grad_norm": 0.3389990031719208, + "learning_rate": 0.000188132700494867, + "loss": 0.5315, + "step": 6691 + }, + { + "epoch": 0.15687101859502922, + "grad_norm": 0.3150341808795929, + "learning_rate": 0.00018812921969583223, + "loss": 0.1007, + "step": 6692 + }, + { + "epoch": 0.15689446016983422, + "grad_norm": 0.5132047533988953, + "learning_rate": 0.00018812573841860583, + "loss": 0.1368, + "step": 6693 + }, + { + "epoch": 0.1569179017446392, + "grad_norm": 0.7168307900428772, + "learning_rate": 0.00018812225666320668, + "loss": 0.2081, + "step": 6694 + }, + { + "epoch": 0.1569413433194442, + "grad_norm": 0.29742228984832764, + "learning_rate": 0.0001881187744296537, + "loss": 0.083, + "step": 6695 + }, + { + "epoch": 0.1569647848942492, + "grad_norm": 0.5821608304977417, + "learning_rate": 0.0001881152917179658, + "loss": 0.1754, + "step": 6696 + }, + { + "epoch": 0.1569882264690542, + "grad_norm": 0.5398024320602417, + "learning_rate": 0.00018811180852816182, + "loss": 0.1639, + "step": 6697 + }, + { + "epoch": 0.1570116680438592, + "grad_norm": 0.40599244832992554, + "learning_rate": 0.0001881083248602607, + "loss": 0.0631, + "step": 6698 + }, + { + "epoch": 0.15703510961866418, + "grad_norm": 0.515255868434906, + "learning_rate": 0.0001881048407142813, + "loss": 0.1268, + "step": 6699 + }, + { + "epoch": 0.15705855119346918, + "grad_norm": 0.5363599061965942, + "learning_rate": 0.00018810135609024259, + "loss": 0.8284, + "step": 6700 + }, + { + "epoch": 0.15708199276827417, + "grad_norm": 0.38756313920021057, + "learning_rate": 0.00018809787098816343, + "loss": 0.0574, + "step": 6701 + }, + { + "epoch": 0.15710543434307916, + "grad_norm": 0.13484160602092743, + "learning_rate": 0.00018809438540806276, + "loss": 0.0429, + "step": 6702 + }, + { + "epoch": 0.15712887591788416, + "grad_norm": 0.9237530827522278, + "learning_rate": 0.00018809089934995946, + "loss": 0.1309, + "step": 6703 + }, + { + "epoch": 0.15715231749268915, + "grad_norm": 0.5864646434783936, + "learning_rate": 0.0001880874128138725, + "loss": 0.209, + "step": 6704 + }, + { + "epoch": 0.15717575906749415, + "grad_norm": 0.8530954122543335, + "learning_rate": 0.00018808392579982073, + "loss": 0.1551, + "step": 6705 + }, + { + "epoch": 0.15719920064229914, + "grad_norm": 0.4058782756328583, + "learning_rate": 0.0001880804383078231, + "loss": 0.094, + "step": 6706 + }, + { + "epoch": 0.15722264221710414, + "grad_norm": 0.7310055494308472, + "learning_rate": 0.00018807695033789854, + "loss": 0.2232, + "step": 6707 + }, + { + "epoch": 0.15724608379190913, + "grad_norm": 0.5233128666877747, + "learning_rate": 0.00018807346189006596, + "loss": 0.1238, + "step": 6708 + }, + { + "epoch": 0.15726952536671412, + "grad_norm": 0.38194090127944946, + "learning_rate": 0.0001880699729643443, + "loss": 0.1297, + "step": 6709 + }, + { + "epoch": 0.15729296694151912, + "grad_norm": 0.8335458636283875, + "learning_rate": 0.00018806648356075253, + "loss": 0.2174, + "step": 6710 + }, + { + "epoch": 0.1573164085163241, + "grad_norm": 0.4867798984050751, + "learning_rate": 0.0001880629936793095, + "loss": 0.1256, + "step": 6711 + }, + { + "epoch": 0.15733985009112914, + "grad_norm": 0.5333802103996277, + "learning_rate": 0.00018805950332003422, + "loss": 0.1553, + "step": 6712 + }, + { + "epoch": 0.15736329166593413, + "grad_norm": 0.3515830934047699, + "learning_rate": 0.0001880560124829456, + "loss": 0.082, + "step": 6713 + }, + { + "epoch": 0.15738673324073912, + "grad_norm": 0.5879266262054443, + "learning_rate": 0.00018805252116806253, + "loss": 0.1283, + "step": 6714 + }, + { + "epoch": 0.15741017481554412, + "grad_norm": 0.6151152849197388, + "learning_rate": 0.00018804902937540402, + "loss": 0.1618, + "step": 6715 + }, + { + "epoch": 0.1574336163903491, + "grad_norm": 0.5086082816123962, + "learning_rate": 0.00018804553710498906, + "loss": 0.1272, + "step": 6716 + }, + { + "epoch": 0.1574570579651541, + "grad_norm": 0.28815746307373047, + "learning_rate": 0.00018804204435683648, + "loss": 0.0925, + "step": 6717 + }, + { + "epoch": 0.1574804995399591, + "grad_norm": 0.6257137060165405, + "learning_rate": 0.00018803855113096532, + "loss": 0.8251, + "step": 6718 + }, + { + "epoch": 0.1575039411147641, + "grad_norm": 0.9174325466156006, + "learning_rate": 0.0001880350574273945, + "loss": 0.213, + "step": 6719 + }, + { + "epoch": 0.1575273826895691, + "grad_norm": 0.22198407351970673, + "learning_rate": 0.00018803156324614297, + "loss": 0.0626, + "step": 6720 + }, + { + "epoch": 0.15755082426437408, + "grad_norm": 0.5417484641075134, + "learning_rate": 0.00018802806858722972, + "loss": 0.1933, + "step": 6721 + }, + { + "epoch": 0.15757426583917908, + "grad_norm": 0.3409513533115387, + "learning_rate": 0.00018802457345067368, + "loss": 0.0828, + "step": 6722 + }, + { + "epoch": 0.15759770741398407, + "grad_norm": 0.2545602023601532, + "learning_rate": 0.00018802107783649384, + "loss": 0.0459, + "step": 6723 + }, + { + "epoch": 0.15762114898878907, + "grad_norm": 0.6968624591827393, + "learning_rate": 0.00018801758174470917, + "loss": 0.1719, + "step": 6724 + }, + { + "epoch": 0.15764459056359406, + "grad_norm": 0.45207086205482483, + "learning_rate": 0.00018801408517533862, + "loss": 0.0758, + "step": 6725 + }, + { + "epoch": 0.15766803213839906, + "grad_norm": 0.388533353805542, + "learning_rate": 0.00018801058812840116, + "loss": 0.0807, + "step": 6726 + }, + { + "epoch": 0.15769147371320405, + "grad_norm": 0.5932191014289856, + "learning_rate": 0.00018800709060391575, + "loss": 0.1372, + "step": 6727 + }, + { + "epoch": 0.15771491528800904, + "grad_norm": 0.48585882782936096, + "learning_rate": 0.00018800359260190142, + "loss": 0.1201, + "step": 6728 + }, + { + "epoch": 0.15773835686281404, + "grad_norm": 0.6598438024520874, + "learning_rate": 0.00018800009412237714, + "loss": 0.1419, + "step": 6729 + }, + { + "epoch": 0.15776179843761903, + "grad_norm": 0.8605594038963318, + "learning_rate": 0.00018799659516536183, + "loss": 0.1202, + "step": 6730 + }, + { + "epoch": 0.15778524001242403, + "grad_norm": 0.1549370288848877, + "learning_rate": 0.00018799309573087455, + "loss": 0.0426, + "step": 6731 + }, + { + "epoch": 0.15780868158722902, + "grad_norm": 0.17640061676502228, + "learning_rate": 0.00018798959581893426, + "loss": 0.0464, + "step": 6732 + }, + { + "epoch": 0.15783212316203402, + "grad_norm": 0.7434650659561157, + "learning_rate": 0.00018798609542955994, + "loss": 0.1498, + "step": 6733 + }, + { + "epoch": 0.157855564736839, + "grad_norm": 0.5691967606544495, + "learning_rate": 0.00018798259456277058, + "loss": 0.139, + "step": 6734 + }, + { + "epoch": 0.157879006311644, + "grad_norm": 0.1217309832572937, + "learning_rate": 0.0001879790932185852, + "loss": 0.0274, + "step": 6735 + }, + { + "epoch": 0.157902447886449, + "grad_norm": 0.40961945056915283, + "learning_rate": 0.0001879755913970228, + "loss": 0.0922, + "step": 6736 + }, + { + "epoch": 0.157925889461254, + "grad_norm": 0.7312790751457214, + "learning_rate": 0.00018797208909810236, + "loss": 0.8207, + "step": 6737 + }, + { + "epoch": 0.15794933103605902, + "grad_norm": 0.8608194589614868, + "learning_rate": 0.00018796858632184286, + "loss": 0.0583, + "step": 6738 + }, + { + "epoch": 0.157972772610864, + "grad_norm": 0.27082112431526184, + "learning_rate": 0.00018796508306826338, + "loss": 0.0542, + "step": 6739 + }, + { + "epoch": 0.157996214185669, + "grad_norm": 0.6167426109313965, + "learning_rate": 0.0001879615793373829, + "loss": 0.1289, + "step": 6740 + }, + { + "epoch": 0.158019655760474, + "grad_norm": 0.36456403136253357, + "learning_rate": 0.00018795807512922034, + "loss": 0.0964, + "step": 6741 + }, + { + "epoch": 0.158043097335279, + "grad_norm": 0.38036996126174927, + "learning_rate": 0.00018795457044379484, + "loss": 0.0861, + "step": 6742 + }, + { + "epoch": 0.158066538910084, + "grad_norm": 0.3155044615268707, + "learning_rate": 0.00018795106528112537, + "loss": 0.0518, + "step": 6743 + }, + { + "epoch": 0.15808998048488898, + "grad_norm": 0.5999167561531067, + "learning_rate": 0.00018794755964123098, + "loss": 0.1243, + "step": 6744 + }, + { + "epoch": 0.15811342205969398, + "grad_norm": 0.45833709836006165, + "learning_rate": 0.00018794405352413058, + "loss": 0.1065, + "step": 6745 + }, + { + "epoch": 0.15813686363449897, + "grad_norm": 0.45995259284973145, + "learning_rate": 0.00018794054692984334, + "loss": 0.1089, + "step": 6746 + }, + { + "epoch": 0.15816030520930396, + "grad_norm": 0.15296339988708496, + "learning_rate": 0.00018793703985838818, + "loss": 0.0425, + "step": 6747 + }, + { + "epoch": 0.15818374678410896, + "grad_norm": 0.6158807277679443, + "learning_rate": 0.00018793353230978416, + "loss": 0.1648, + "step": 6748 + }, + { + "epoch": 0.15820718835891395, + "grad_norm": 0.9529522657394409, + "learning_rate": 0.00018793002428405035, + "loss": 0.2072, + "step": 6749 + }, + { + "epoch": 0.15823062993371895, + "grad_norm": 0.2505687475204468, + "learning_rate": 0.00018792651578120574, + "loss": 0.0875, + "step": 6750 + }, + { + "epoch": 0.15825407150852394, + "grad_norm": 0.7558417916297913, + "learning_rate": 0.00018792300680126942, + "loss": 0.1365, + "step": 6751 + }, + { + "epoch": 0.15827751308332894, + "grad_norm": 0.43059849739074707, + "learning_rate": 0.00018791949734426035, + "loss": 0.1327, + "step": 6752 + }, + { + "epoch": 0.15830095465813393, + "grad_norm": 0.6781091690063477, + "learning_rate": 0.0001879159874101976, + "loss": 0.6455, + "step": 6753 + }, + { + "epoch": 0.15832439623293892, + "grad_norm": 0.6013475656509399, + "learning_rate": 0.0001879124769991003, + "loss": 0.1326, + "step": 6754 + }, + { + "epoch": 0.15834783780774392, + "grad_norm": 0.6473913788795471, + "learning_rate": 0.0001879089661109874, + "loss": 0.1361, + "step": 6755 + }, + { + "epoch": 0.1583712793825489, + "grad_norm": 0.3179303705692291, + "learning_rate": 0.00018790545474587798, + "loss": 0.0655, + "step": 6756 + }, + { + "epoch": 0.1583947209573539, + "grad_norm": 0.18079152703285217, + "learning_rate": 0.00018790194290379107, + "loss": 0.0673, + "step": 6757 + }, + { + "epoch": 0.1584181625321589, + "grad_norm": 0.3737170994281769, + "learning_rate": 0.00018789843058474574, + "loss": 0.0772, + "step": 6758 + }, + { + "epoch": 0.1584416041069639, + "grad_norm": 0.8496583104133606, + "learning_rate": 0.00018789491778876106, + "loss": 0.1215, + "step": 6759 + }, + { + "epoch": 0.1584650456817689, + "grad_norm": 0.5573914051055908, + "learning_rate": 0.0001878914045158561, + "loss": 0.1182, + "step": 6760 + }, + { + "epoch": 0.15848848725657388, + "grad_norm": 0.17211028933525085, + "learning_rate": 0.00018788789076604992, + "loss": 0.0357, + "step": 6761 + }, + { + "epoch": 0.15851192883137888, + "grad_norm": 0.3162453770637512, + "learning_rate": 0.00018788437653936155, + "loss": 0.0722, + "step": 6762 + }, + { + "epoch": 0.1585353704061839, + "grad_norm": 1.0313550233840942, + "learning_rate": 0.00018788086183581007, + "loss": 0.6344, + "step": 6763 + }, + { + "epoch": 0.1585588119809889, + "grad_norm": 0.21354782581329346, + "learning_rate": 0.00018787734665541458, + "loss": 0.0583, + "step": 6764 + }, + { + "epoch": 0.1585822535557939, + "grad_norm": 0.24712055921554565, + "learning_rate": 0.00018787383099819413, + "loss": 0.0832, + "step": 6765 + }, + { + "epoch": 0.15860569513059888, + "grad_norm": 0.458389014005661, + "learning_rate": 0.0001878703148641678, + "loss": 0.074, + "step": 6766 + }, + { + "epoch": 0.15862913670540388, + "grad_norm": 0.644793689250946, + "learning_rate": 0.00018786679825335469, + "loss": 0.2237, + "step": 6767 + }, + { + "epoch": 0.15865257828020887, + "grad_norm": 0.11807826906442642, + "learning_rate": 0.00018786328116577385, + "loss": 0.0327, + "step": 6768 + }, + { + "epoch": 0.15867601985501387, + "grad_norm": 0.817305326461792, + "learning_rate": 0.00018785976360144438, + "loss": 0.2487, + "step": 6769 + }, + { + "epoch": 0.15869946142981886, + "grad_norm": 0.41517719626426697, + "learning_rate": 0.00018785624556038534, + "loss": 0.0718, + "step": 6770 + }, + { + "epoch": 0.15872290300462386, + "grad_norm": 0.6934635043144226, + "learning_rate": 0.0001878527270426159, + "loss": 0.4633, + "step": 6771 + }, + { + "epoch": 0.15874634457942885, + "grad_norm": 0.5757923126220703, + "learning_rate": 0.00018784920804815504, + "loss": 0.0947, + "step": 6772 + }, + { + "epoch": 0.15876978615423384, + "grad_norm": 0.8156081438064575, + "learning_rate": 0.00018784568857702191, + "loss": 0.2273, + "step": 6773 + }, + { + "epoch": 0.15879322772903884, + "grad_norm": 0.5518758296966553, + "learning_rate": 0.00018784216862923562, + "loss": 0.11, + "step": 6774 + }, + { + "epoch": 0.15881666930384383, + "grad_norm": 0.7732176780700684, + "learning_rate": 0.00018783864820481525, + "loss": 0.0985, + "step": 6775 + }, + { + "epoch": 0.15884011087864883, + "grad_norm": 0.3949122428894043, + "learning_rate": 0.0001878351273037799, + "loss": 0.1198, + "step": 6776 + }, + { + "epoch": 0.15886355245345382, + "grad_norm": 0.5022432208061218, + "learning_rate": 0.00018783160592614868, + "loss": 0.1443, + "step": 6777 + }, + { + "epoch": 0.15888699402825882, + "grad_norm": 0.6896098852157593, + "learning_rate": 0.00018782808407194067, + "loss": 0.2444, + "step": 6778 + }, + { + "epoch": 0.1589104356030638, + "grad_norm": 0.7467806935310364, + "learning_rate": 0.00018782456174117504, + "loss": 0.7031, + "step": 6779 + }, + { + "epoch": 0.1589338771778688, + "grad_norm": 0.5923810601234436, + "learning_rate": 0.00018782103893387087, + "loss": 0.1725, + "step": 6780 + }, + { + "epoch": 0.1589573187526738, + "grad_norm": 0.49548983573913574, + "learning_rate": 0.00018781751565004727, + "loss": 0.1023, + "step": 6781 + }, + { + "epoch": 0.1589807603274788, + "grad_norm": 0.3583187162876129, + "learning_rate": 0.00018781399188972335, + "loss": 0.0772, + "step": 6782 + }, + { + "epoch": 0.1590042019022838, + "grad_norm": 0.5250473022460938, + "learning_rate": 0.00018781046765291824, + "loss": 0.1732, + "step": 6783 + }, + { + "epoch": 0.15902764347708878, + "grad_norm": 0.1125989630818367, + "learning_rate": 0.00018780694293965108, + "loss": 0.0292, + "step": 6784 + }, + { + "epoch": 0.15905108505189378, + "grad_norm": 0.6092655658721924, + "learning_rate": 0.00018780341774994098, + "loss": 0.1128, + "step": 6785 + }, + { + "epoch": 0.15907452662669877, + "grad_norm": 0.42585691809654236, + "learning_rate": 0.00018779989208380705, + "loss": 0.1256, + "step": 6786 + }, + { + "epoch": 0.15909796820150376, + "grad_norm": 0.4090464413166046, + "learning_rate": 0.00018779636594126845, + "loss": 0.0801, + "step": 6787 + }, + { + "epoch": 0.15912140977630876, + "grad_norm": 0.30435073375701904, + "learning_rate": 0.0001877928393223443, + "loss": 0.0437, + "step": 6788 + }, + { + "epoch": 0.15914485135111378, + "grad_norm": 0.6221624612808228, + "learning_rate": 0.00018778931222705373, + "loss": 0.6451, + "step": 6789 + }, + { + "epoch": 0.15916829292591878, + "grad_norm": 0.5732249021530151, + "learning_rate": 0.00018778578465541588, + "loss": 0.1794, + "step": 6790 + }, + { + "epoch": 0.15919173450072377, + "grad_norm": 1.023145318031311, + "learning_rate": 0.00018778225660744991, + "loss": 0.2018, + "step": 6791 + }, + { + "epoch": 0.15921517607552876, + "grad_norm": 0.3535621762275696, + "learning_rate": 0.0001877787280831749, + "loss": 0.1082, + "step": 6792 + }, + { + "epoch": 0.15923861765033376, + "grad_norm": 0.4077555537223816, + "learning_rate": 0.00018777519908261007, + "loss": 0.0523, + "step": 6793 + }, + { + "epoch": 0.15926205922513875, + "grad_norm": 0.5091016292572021, + "learning_rate": 0.00018777166960577457, + "loss": 0.1646, + "step": 6794 + }, + { + "epoch": 0.15928550079994375, + "grad_norm": 0.41320767998695374, + "learning_rate": 0.0001877681396526875, + "loss": 0.0789, + "step": 6795 + }, + { + "epoch": 0.15930894237474874, + "grad_norm": 0.6239379644393921, + "learning_rate": 0.00018776460922336803, + "loss": 0.6756, + "step": 6796 + }, + { + "epoch": 0.15933238394955374, + "grad_norm": 0.7036024928092957, + "learning_rate": 0.00018776107831783534, + "loss": 0.1728, + "step": 6797 + }, + { + "epoch": 0.15935582552435873, + "grad_norm": 0.6400532126426697, + "learning_rate": 0.00018775754693610856, + "loss": 0.1299, + "step": 6798 + }, + { + "epoch": 0.15937926709916372, + "grad_norm": 0.6092662215232849, + "learning_rate": 0.00018775401507820688, + "loss": 0.1458, + "step": 6799 + }, + { + "epoch": 0.15940270867396872, + "grad_norm": 0.48450982570648193, + "learning_rate": 0.0001877504827441494, + "loss": 0.0923, + "step": 6800 + }, + { + "epoch": 0.1594261502487737, + "grad_norm": 0.45356056094169617, + "learning_rate": 0.0001877469499339554, + "loss": 0.5598, + "step": 6801 + }, + { + "epoch": 0.1594495918235787, + "grad_norm": 0.5797001123428345, + "learning_rate": 0.0001877434166476439, + "loss": 0.5654, + "step": 6802 + }, + { + "epoch": 0.1594730333983837, + "grad_norm": 0.1410662680864334, + "learning_rate": 0.00018773988288523424, + "loss": 0.0245, + "step": 6803 + }, + { + "epoch": 0.1594964749731887, + "grad_norm": 0.46525466442108154, + "learning_rate": 0.00018773634864674549, + "loss": 0.1109, + "step": 6804 + }, + { + "epoch": 0.1595199165479937, + "grad_norm": 0.4058215022087097, + "learning_rate": 0.00018773281393219679, + "loss": 0.0808, + "step": 6805 + }, + { + "epoch": 0.15954335812279868, + "grad_norm": 0.37076041102409363, + "learning_rate": 0.0001877292787416074, + "loss": 0.0715, + "step": 6806 + }, + { + "epoch": 0.15956679969760368, + "grad_norm": 0.190049946308136, + "learning_rate": 0.0001877257430749965, + "loss": 0.0506, + "step": 6807 + }, + { + "epoch": 0.15959024127240867, + "grad_norm": 0.6344842910766602, + "learning_rate": 0.00018772220693238323, + "loss": 0.0801, + "step": 6808 + }, + { + "epoch": 0.15961368284721367, + "grad_norm": 0.585986852645874, + "learning_rate": 0.0001877186703137868, + "loss": 0.1433, + "step": 6809 + }, + { + "epoch": 0.15963712442201866, + "grad_norm": 1.7551881074905396, + "learning_rate": 0.00018771513321922642, + "loss": 0.1461, + "step": 6810 + }, + { + "epoch": 0.15966056599682366, + "grad_norm": 0.18865956366062164, + "learning_rate": 0.00018771159564872126, + "loss": 0.0285, + "step": 6811 + }, + { + "epoch": 0.15968400757162865, + "grad_norm": 0.761063814163208, + "learning_rate": 0.00018770805760229048, + "loss": 0.1443, + "step": 6812 + }, + { + "epoch": 0.15970744914643364, + "grad_norm": 0.49601686000823975, + "learning_rate": 0.00018770451907995335, + "loss": 0.095, + "step": 6813 + }, + { + "epoch": 0.15973089072123867, + "grad_norm": 0.6089324355125427, + "learning_rate": 0.000187700980081729, + "loss": 0.1596, + "step": 6814 + }, + { + "epoch": 0.15975433229604366, + "grad_norm": 0.8901655077934265, + "learning_rate": 0.0001876974406076367, + "loss": 0.2892, + "step": 6815 + }, + { + "epoch": 0.15977777387084866, + "grad_norm": 0.7764641642570496, + "learning_rate": 0.00018769390065769562, + "loss": 0.8056, + "step": 6816 + }, + { + "epoch": 0.15980121544565365, + "grad_norm": 0.30768921971321106, + "learning_rate": 0.00018769036023192492, + "loss": 0.0716, + "step": 6817 + }, + { + "epoch": 0.15982465702045864, + "grad_norm": 0.19183145463466644, + "learning_rate": 0.00018768681933034392, + "loss": 0.0284, + "step": 6818 + }, + { + "epoch": 0.15984809859526364, + "grad_norm": 0.6846641898155212, + "learning_rate": 0.00018768327795297177, + "loss": 0.1827, + "step": 6819 + }, + { + "epoch": 0.15987154017006863, + "grad_norm": 0.16286161541938782, + "learning_rate": 0.00018767973609982766, + "loss": 0.0347, + "step": 6820 + }, + { + "epoch": 0.15989498174487363, + "grad_norm": 0.39921966195106506, + "learning_rate": 0.00018767619377093082, + "loss": 0.0569, + "step": 6821 + }, + { + "epoch": 0.15991842331967862, + "grad_norm": 0.15579383075237274, + "learning_rate": 0.0001876726509663005, + "loss": 0.0447, + "step": 6822 + }, + { + "epoch": 0.15994186489448362, + "grad_norm": 0.17965930700302124, + "learning_rate": 0.00018766910768595595, + "loss": 0.0406, + "step": 6823 + }, + { + "epoch": 0.1599653064692886, + "grad_norm": 0.7223326563835144, + "learning_rate": 0.0001876655639299163, + "loss": 0.7262, + "step": 6824 + }, + { + "epoch": 0.1599887480440936, + "grad_norm": 0.6054543256759644, + "learning_rate": 0.00018766201969820087, + "loss": 0.0756, + "step": 6825 + }, + { + "epoch": 0.1600121896188986, + "grad_norm": 0.5608171820640564, + "learning_rate": 0.00018765847499082884, + "loss": 0.0906, + "step": 6826 + }, + { + "epoch": 0.1600356311937036, + "grad_norm": 0.5172512531280518, + "learning_rate": 0.0001876549298078195, + "loss": 0.1867, + "step": 6827 + }, + { + "epoch": 0.1600590727685086, + "grad_norm": 0.3798580765724182, + "learning_rate": 0.000187651384149192, + "loss": 0.0928, + "step": 6828 + }, + { + "epoch": 0.16008251434331358, + "grad_norm": 0.6981317400932312, + "learning_rate": 0.00018764783801496565, + "loss": 0.2207, + "step": 6829 + }, + { + "epoch": 0.16010595591811858, + "grad_norm": 0.48209601640701294, + "learning_rate": 0.00018764429140515965, + "loss": 0.1355, + "step": 6830 + }, + { + "epoch": 0.16012939749292357, + "grad_norm": 0.1987864226102829, + "learning_rate": 0.00018764074431979326, + "loss": 0.0429, + "step": 6831 + }, + { + "epoch": 0.16015283906772856, + "grad_norm": 0.615087628364563, + "learning_rate": 0.00018763719675888573, + "loss": 0.6454, + "step": 6832 + }, + { + "epoch": 0.16017628064253356, + "grad_norm": 0.19125038385391235, + "learning_rate": 0.00018763364872245633, + "loss": 0.0457, + "step": 6833 + }, + { + "epoch": 0.16019972221733855, + "grad_norm": 0.518677294254303, + "learning_rate": 0.00018763010021052425, + "loss": 0.0883, + "step": 6834 + }, + { + "epoch": 0.16022316379214355, + "grad_norm": 0.556326687335968, + "learning_rate": 0.00018762655122310878, + "loss": 0.1489, + "step": 6835 + }, + { + "epoch": 0.16024660536694854, + "grad_norm": 0.12247483432292938, + "learning_rate": 0.0001876230017602292, + "loss": 0.0148, + "step": 6836 + }, + { + "epoch": 0.16027004694175354, + "grad_norm": 1.2212767601013184, + "learning_rate": 0.00018761945182190477, + "loss": 0.2481, + "step": 6837 + }, + { + "epoch": 0.16029348851655853, + "grad_norm": 0.4563119113445282, + "learning_rate": 0.00018761590140815471, + "loss": 0.1645, + "step": 6838 + }, + { + "epoch": 0.16031693009136352, + "grad_norm": 0.24816004931926727, + "learning_rate": 0.0001876123505189983, + "loss": 0.0521, + "step": 6839 + }, + { + "epoch": 0.16034037166616855, + "grad_norm": 0.37167125940322876, + "learning_rate": 0.00018760879915445485, + "loss": 0.0662, + "step": 6840 + }, + { + "epoch": 0.16036381324097354, + "grad_norm": 0.2792658805847168, + "learning_rate": 0.00018760524731454355, + "loss": 0.0623, + "step": 6841 + }, + { + "epoch": 0.16038725481577853, + "grad_norm": 0.18703149259090424, + "learning_rate": 0.0001876016949992837, + "loss": 0.0386, + "step": 6842 + }, + { + "epoch": 0.16041069639058353, + "grad_norm": 0.9560617804527283, + "learning_rate": 0.00018759814220869464, + "loss": 0.2332, + "step": 6843 + }, + { + "epoch": 0.16043413796538852, + "grad_norm": 0.47154051065444946, + "learning_rate": 0.00018759458894279557, + "loss": 0.0972, + "step": 6844 + }, + { + "epoch": 0.16045757954019352, + "grad_norm": 0.5408695936203003, + "learning_rate": 0.0001875910352016058, + "loss": 0.1986, + "step": 6845 + }, + { + "epoch": 0.1604810211149985, + "grad_norm": 0.5921871066093445, + "learning_rate": 0.0001875874809851446, + "loss": 0.1647, + "step": 6846 + }, + { + "epoch": 0.1605044626898035, + "grad_norm": 0.13128751516342163, + "learning_rate": 0.00018758392629343126, + "loss": 0.0285, + "step": 6847 + }, + { + "epoch": 0.1605279042646085, + "grad_norm": 0.5102279782295227, + "learning_rate": 0.0001875803711264851, + "loss": 0.139, + "step": 6848 + }, + { + "epoch": 0.1605513458394135, + "grad_norm": 0.3146857023239136, + "learning_rate": 0.00018757681548432537, + "loss": 0.0531, + "step": 6849 + }, + { + "epoch": 0.1605747874142185, + "grad_norm": 0.5522395968437195, + "learning_rate": 0.00018757325936697138, + "loss": 0.1738, + "step": 6850 + }, + { + "epoch": 0.16059822898902348, + "grad_norm": 0.29981300234794617, + "learning_rate": 0.00018756970277444243, + "loss": 0.0771, + "step": 6851 + }, + { + "epoch": 0.16062167056382848, + "grad_norm": 0.554253339767456, + "learning_rate": 0.00018756614570675782, + "loss": 0.2614, + "step": 6852 + }, + { + "epoch": 0.16064511213863347, + "grad_norm": 0.36659497022628784, + "learning_rate": 0.0001875625881639368, + "loss": 0.1055, + "step": 6853 + }, + { + "epoch": 0.16066855371343847, + "grad_norm": 0.5275478959083557, + "learning_rate": 0.00018755903014599875, + "loss": 0.1438, + "step": 6854 + }, + { + "epoch": 0.16069199528824346, + "grad_norm": 0.7678778767585754, + "learning_rate": 0.00018755547165296292, + "loss": 0.2051, + "step": 6855 + }, + { + "epoch": 0.16071543686304846, + "grad_norm": 0.6898640990257263, + "learning_rate": 0.00018755191268484864, + "loss": 0.1197, + "step": 6856 + }, + { + "epoch": 0.16073887843785345, + "grad_norm": 0.5820766687393188, + "learning_rate": 0.00018754835324167522, + "loss": 0.1846, + "step": 6857 + }, + { + "epoch": 0.16076232001265844, + "grad_norm": 0.08255664259195328, + "learning_rate": 0.00018754479332346198, + "loss": 0.0171, + "step": 6858 + }, + { + "epoch": 0.16078576158746344, + "grad_norm": 0.48308074474334717, + "learning_rate": 0.00018754123293022822, + "loss": 0.1333, + "step": 6859 + }, + { + "epoch": 0.16080920316226843, + "grad_norm": 0.16505685448646545, + "learning_rate": 0.00018753767206199329, + "loss": 0.0337, + "step": 6860 + }, + { + "epoch": 0.16083264473707343, + "grad_norm": 0.19024287164211273, + "learning_rate": 0.00018753411071877648, + "loss": 0.0816, + "step": 6861 + }, + { + "epoch": 0.16085608631187842, + "grad_norm": 0.36601871252059937, + "learning_rate": 0.00018753054890059714, + "loss": 0.088, + "step": 6862 + }, + { + "epoch": 0.16087952788668342, + "grad_norm": 0.592159628868103, + "learning_rate": 0.00018752698660747457, + "loss": 0.7855, + "step": 6863 + }, + { + "epoch": 0.1609029694614884, + "grad_norm": 0.3696102797985077, + "learning_rate": 0.00018752342383942809, + "loss": 0.1231, + "step": 6864 + }, + { + "epoch": 0.16092641103629343, + "grad_norm": 0.3126394748687744, + "learning_rate": 0.00018751986059647706, + "loss": 0.093, + "step": 6865 + }, + { + "epoch": 0.16094985261109843, + "grad_norm": 0.4303453862667084, + "learning_rate": 0.0001875162968786408, + "loss": 0.1089, + "step": 6866 + }, + { + "epoch": 0.16097329418590342, + "grad_norm": 0.323895126581192, + "learning_rate": 0.00018751273268593867, + "loss": 0.0958, + "step": 6867 + }, + { + "epoch": 0.16099673576070841, + "grad_norm": 0.5924549102783203, + "learning_rate": 0.00018750916801838998, + "loss": 0.097, + "step": 6868 + }, + { + "epoch": 0.1610201773355134, + "grad_norm": 0.4673294126987457, + "learning_rate": 0.00018750560287601409, + "loss": 0.146, + "step": 6869 + }, + { + "epoch": 0.1610436189103184, + "grad_norm": 0.7098835706710815, + "learning_rate": 0.00018750203725883036, + "loss": 0.1679, + "step": 6870 + }, + { + "epoch": 0.1610670604851234, + "grad_norm": 0.16038557887077332, + "learning_rate": 0.00018749847116685808, + "loss": 0.024, + "step": 6871 + }, + { + "epoch": 0.1610905020599284, + "grad_norm": 0.26255008578300476, + "learning_rate": 0.00018749490460011665, + "loss": 0.0972, + "step": 6872 + }, + { + "epoch": 0.1611139436347334, + "grad_norm": 0.5517424941062927, + "learning_rate": 0.0001874913375586254, + "loss": 0.1414, + "step": 6873 + }, + { + "epoch": 0.16113738520953838, + "grad_norm": 0.5192638635635376, + "learning_rate": 0.0001874877700424037, + "loss": 0.8135, + "step": 6874 + }, + { + "epoch": 0.16116082678434338, + "grad_norm": 0.4021015465259552, + "learning_rate": 0.00018748420205147087, + "loss": 0.0881, + "step": 6875 + }, + { + "epoch": 0.16118426835914837, + "grad_norm": 0.688843846321106, + "learning_rate": 0.00018748063358584635, + "loss": 0.1534, + "step": 6876 + }, + { + "epoch": 0.16120770993395336, + "grad_norm": 0.35086360573768616, + "learning_rate": 0.00018747706464554943, + "loss": 0.1243, + "step": 6877 + }, + { + "epoch": 0.16123115150875836, + "grad_norm": 0.9556066393852234, + "learning_rate": 0.00018747349523059952, + "loss": 0.1378, + "step": 6878 + }, + { + "epoch": 0.16125459308356335, + "grad_norm": 0.8680687546730042, + "learning_rate": 0.00018746992534101592, + "loss": 0.2069, + "step": 6879 + }, + { + "epoch": 0.16127803465836835, + "grad_norm": 0.451853483915329, + "learning_rate": 0.00018746635497681808, + "loss": 0.0978, + "step": 6880 + }, + { + "epoch": 0.16130147623317334, + "grad_norm": 0.5011263489723206, + "learning_rate": 0.00018746278413802532, + "loss": 0.1188, + "step": 6881 + }, + { + "epoch": 0.16132491780797834, + "grad_norm": 0.4764561951160431, + "learning_rate": 0.00018745921282465705, + "loss": 0.078, + "step": 6882 + }, + { + "epoch": 0.16134835938278333, + "grad_norm": 0.5702246427536011, + "learning_rate": 0.00018745564103673262, + "loss": 0.1646, + "step": 6883 + }, + { + "epoch": 0.16137180095758832, + "grad_norm": 0.12385294586420059, + "learning_rate": 0.00018745206877427141, + "loss": 0.0269, + "step": 6884 + }, + { + "epoch": 0.16139524253239332, + "grad_norm": 0.6838204860687256, + "learning_rate": 0.0001874484960372928, + "loss": 0.1506, + "step": 6885 + }, + { + "epoch": 0.1614186841071983, + "grad_norm": 0.4694569408893585, + "learning_rate": 0.00018744492282581624, + "loss": 0.1183, + "step": 6886 + }, + { + "epoch": 0.1614421256820033, + "grad_norm": 0.698679506778717, + "learning_rate": 0.00018744134913986106, + "loss": 0.1915, + "step": 6887 + }, + { + "epoch": 0.1614655672568083, + "grad_norm": 0.7156009674072266, + "learning_rate": 0.00018743777497944665, + "loss": 0.1578, + "step": 6888 + }, + { + "epoch": 0.1614890088316133, + "grad_norm": 0.5107437968254089, + "learning_rate": 0.00018743420034459245, + "loss": 0.1287, + "step": 6889 + }, + { + "epoch": 0.1615124504064183, + "grad_norm": 0.6076609492301941, + "learning_rate": 0.00018743062523531777, + "loss": 0.2376, + "step": 6890 + }, + { + "epoch": 0.1615358919812233, + "grad_norm": 0.45034292340278625, + "learning_rate": 0.00018742704965164208, + "loss": 0.0712, + "step": 6891 + }, + { + "epoch": 0.1615593335560283, + "grad_norm": 0.48074689507484436, + "learning_rate": 0.00018742347359358477, + "loss": 0.1939, + "step": 6892 + }, + { + "epoch": 0.1615827751308333, + "grad_norm": 0.4866468906402588, + "learning_rate": 0.0001874198970611652, + "loss": 0.0904, + "step": 6893 + }, + { + "epoch": 0.1616062167056383, + "grad_norm": 0.5255922675132751, + "learning_rate": 0.00018741632005440283, + "loss": 0.104, + "step": 6894 + }, + { + "epoch": 0.1616296582804433, + "grad_norm": 0.3782905340194702, + "learning_rate": 0.00018741274257331707, + "loss": 0.0927, + "step": 6895 + }, + { + "epoch": 0.16165309985524828, + "grad_norm": 0.6673352718353271, + "learning_rate": 0.0001874091646179273, + "loss": 0.841, + "step": 6896 + }, + { + "epoch": 0.16167654143005328, + "grad_norm": 0.583557665348053, + "learning_rate": 0.0001874055861882529, + "loss": 0.1163, + "step": 6897 + }, + { + "epoch": 0.16169998300485827, + "grad_norm": 0.3949233293533325, + "learning_rate": 0.0001874020072843134, + "loss": 0.0602, + "step": 6898 + }, + { + "epoch": 0.16172342457966327, + "grad_norm": 0.5336226224899292, + "learning_rate": 0.00018739842790612812, + "loss": 0.1863, + "step": 6899 + }, + { + "epoch": 0.16174686615446826, + "grad_norm": 0.4848705232143402, + "learning_rate": 0.00018739484805371651, + "loss": 0.0587, + "step": 6900 + }, + { + "epoch": 0.16177030772927325, + "grad_norm": 0.2378518283367157, + "learning_rate": 0.000187391267727098, + "loss": 0.0673, + "step": 6901 + }, + { + "epoch": 0.16179374930407825, + "grad_norm": 0.5078369379043579, + "learning_rate": 0.000187387686926292, + "loss": 0.0911, + "step": 6902 + }, + { + "epoch": 0.16181719087888324, + "grad_norm": 0.6527730226516724, + "learning_rate": 0.00018738410565131798, + "loss": 0.2067, + "step": 6903 + }, + { + "epoch": 0.16184063245368824, + "grad_norm": 0.26655423641204834, + "learning_rate": 0.00018738052390219535, + "loss": 0.0552, + "step": 6904 + }, + { + "epoch": 0.16186407402849323, + "grad_norm": 0.26433706283569336, + "learning_rate": 0.00018737694167894354, + "loss": 0.0675, + "step": 6905 + }, + { + "epoch": 0.16188751560329823, + "grad_norm": 0.6142370104789734, + "learning_rate": 0.00018737335898158194, + "loss": 0.1462, + "step": 6906 + }, + { + "epoch": 0.16191095717810322, + "grad_norm": 0.14159248769283295, + "learning_rate": 0.0001873697758101301, + "loss": 0.0326, + "step": 6907 + }, + { + "epoch": 0.16193439875290822, + "grad_norm": 0.62735915184021, + "learning_rate": 0.00018736619216460737, + "loss": 0.2025, + "step": 6908 + }, + { + "epoch": 0.1619578403277132, + "grad_norm": 0.3512662351131439, + "learning_rate": 0.00018736260804503321, + "loss": 0.114, + "step": 6909 + }, + { + "epoch": 0.1619812819025182, + "grad_norm": 0.936380922794342, + "learning_rate": 0.00018735902345142709, + "loss": 0.1585, + "step": 6910 + }, + { + "epoch": 0.1620047234773232, + "grad_norm": 0.45281633734703064, + "learning_rate": 0.00018735543838380845, + "loss": 0.5706, + "step": 6911 + }, + { + "epoch": 0.1620281650521282, + "grad_norm": 0.232364684343338, + "learning_rate": 0.00018735185284219675, + "loss": 0.0582, + "step": 6912 + }, + { + "epoch": 0.1620516066269332, + "grad_norm": 0.1670110672712326, + "learning_rate": 0.00018734826682661147, + "loss": 0.0346, + "step": 6913 + }, + { + "epoch": 0.16207504820173818, + "grad_norm": 0.9722400903701782, + "learning_rate": 0.000187344680337072, + "loss": 0.1594, + "step": 6914 + }, + { + "epoch": 0.16209848977654318, + "grad_norm": 1.2575910091400146, + "learning_rate": 0.00018734109337359786, + "loss": 0.1542, + "step": 6915 + }, + { + "epoch": 0.1621219313513482, + "grad_norm": 0.27829551696777344, + "learning_rate": 0.0001873375059362085, + "loss": 0.0527, + "step": 6916 + }, + { + "epoch": 0.1621453729261532, + "grad_norm": 0.701982319355011, + "learning_rate": 0.00018733391802492334, + "loss": 0.0704, + "step": 6917 + }, + { + "epoch": 0.16216881450095819, + "grad_norm": 0.5221447944641113, + "learning_rate": 0.0001873303296397619, + "loss": 0.1479, + "step": 6918 + }, + { + "epoch": 0.16219225607576318, + "grad_norm": 0.5621776580810547, + "learning_rate": 0.00018732674078074366, + "loss": 0.1301, + "step": 6919 + }, + { + "epoch": 0.16221569765056817, + "grad_norm": 0.7723535299301147, + "learning_rate": 0.00018732315144788804, + "loss": 0.1224, + "step": 6920 + }, + { + "epoch": 0.16223913922537317, + "grad_norm": 0.4723324477672577, + "learning_rate": 0.00018731956164121455, + "loss": 0.1488, + "step": 6921 + }, + { + "epoch": 0.16226258080017816, + "grad_norm": 0.5794104337692261, + "learning_rate": 0.00018731597136074267, + "loss": 0.2083, + "step": 6922 + }, + { + "epoch": 0.16228602237498316, + "grad_norm": 0.5434218049049377, + "learning_rate": 0.00018731238060649185, + "loss": 0.0945, + "step": 6923 + }, + { + "epoch": 0.16230946394978815, + "grad_norm": 0.6833059787750244, + "learning_rate": 0.00018730878937848165, + "loss": 0.1535, + "step": 6924 + }, + { + "epoch": 0.16233290552459315, + "grad_norm": 0.46902865171432495, + "learning_rate": 0.00018730519767673148, + "loss": 0.1182, + "step": 6925 + }, + { + "epoch": 0.16235634709939814, + "grad_norm": 0.33560261130332947, + "learning_rate": 0.00018730160550126083, + "loss": 0.0714, + "step": 6926 + }, + { + "epoch": 0.16237978867420313, + "grad_norm": 0.4277994930744171, + "learning_rate": 0.00018729801285208923, + "loss": 0.0867, + "step": 6927 + }, + { + "epoch": 0.16240323024900813, + "grad_norm": 0.7363865971565247, + "learning_rate": 0.00018729441972923616, + "loss": 0.1435, + "step": 6928 + }, + { + "epoch": 0.16242667182381312, + "grad_norm": 0.6209157109260559, + "learning_rate": 0.0001872908261327211, + "loss": 0.7947, + "step": 6929 + }, + { + "epoch": 0.16245011339861812, + "grad_norm": 0.16044507920742035, + "learning_rate": 0.0001872872320625636, + "loss": 0.0277, + "step": 6930 + }, + { + "epoch": 0.1624735549734231, + "grad_norm": 0.7256117463111877, + "learning_rate": 0.00018728363751878307, + "loss": 0.1961, + "step": 6931 + }, + { + "epoch": 0.1624969965482281, + "grad_norm": 0.470335990190506, + "learning_rate": 0.0001872800425013991, + "loss": 0.0773, + "step": 6932 + }, + { + "epoch": 0.1625204381230331, + "grad_norm": 0.5664422512054443, + "learning_rate": 0.00018727644701043117, + "loss": 0.0813, + "step": 6933 + }, + { + "epoch": 0.1625438796978381, + "grad_norm": 0.676813006401062, + "learning_rate": 0.0001872728510458988, + "loss": 0.1813, + "step": 6934 + }, + { + "epoch": 0.1625673212726431, + "grad_norm": 0.12728700041770935, + "learning_rate": 0.00018726925460782145, + "loss": 0.0353, + "step": 6935 + }, + { + "epoch": 0.16259076284744808, + "grad_norm": 0.4162920117378235, + "learning_rate": 0.00018726565769621868, + "loss": 0.0751, + "step": 6936 + }, + { + "epoch": 0.16261420442225308, + "grad_norm": 0.30121684074401855, + "learning_rate": 0.00018726206031111003, + "loss": 0.0676, + "step": 6937 + }, + { + "epoch": 0.16263764599705807, + "grad_norm": 0.5528425574302673, + "learning_rate": 0.00018725846245251494, + "loss": 0.7486, + "step": 6938 + }, + { + "epoch": 0.16266108757186307, + "grad_norm": 0.6591818928718567, + "learning_rate": 0.00018725486412045301, + "loss": 0.2165, + "step": 6939 + }, + { + "epoch": 0.16268452914666806, + "grad_norm": 0.6115097403526306, + "learning_rate": 0.00018725126531494372, + "loss": 0.1958, + "step": 6940 + }, + { + "epoch": 0.16270797072147306, + "grad_norm": 0.5970703959465027, + "learning_rate": 0.0001872476660360066, + "loss": 0.1553, + "step": 6941 + }, + { + "epoch": 0.16273141229627808, + "grad_norm": 0.616477906703949, + "learning_rate": 0.0001872440662836612, + "loss": 0.1467, + "step": 6942 + }, + { + "epoch": 0.16275485387108307, + "grad_norm": 0.676270067691803, + "learning_rate": 0.00018724046605792706, + "loss": 0.1569, + "step": 6943 + }, + { + "epoch": 0.16277829544588807, + "grad_norm": 0.61137455701828, + "learning_rate": 0.0001872368653588237, + "loss": 0.2127, + "step": 6944 + }, + { + "epoch": 0.16280173702069306, + "grad_norm": 0.5785744190216064, + "learning_rate": 0.00018723326418637064, + "loss": 0.0731, + "step": 6945 + }, + { + "epoch": 0.16282517859549805, + "grad_norm": 0.7279897332191467, + "learning_rate": 0.00018722966254058744, + "loss": 0.1345, + "step": 6946 + }, + { + "epoch": 0.16284862017030305, + "grad_norm": 0.3786972165107727, + "learning_rate": 0.00018722606042149363, + "loss": 0.1338, + "step": 6947 + }, + { + "epoch": 0.16287206174510804, + "grad_norm": 0.8311602473258972, + "learning_rate": 0.00018722245782910876, + "loss": 0.1864, + "step": 6948 + }, + { + "epoch": 0.16289550331991304, + "grad_norm": 0.20079323649406433, + "learning_rate": 0.0001872188547634524, + "loss": 0.0268, + "step": 6949 + }, + { + "epoch": 0.16291894489471803, + "grad_norm": 0.5664246082305908, + "learning_rate": 0.00018721525122454406, + "loss": 0.1142, + "step": 6950 + }, + { + "epoch": 0.16294238646952303, + "grad_norm": 0.4853242039680481, + "learning_rate": 0.00018721164721240333, + "loss": 0.1335, + "step": 6951 + }, + { + "epoch": 0.16296582804432802, + "grad_norm": 0.5278905034065247, + "learning_rate": 0.00018720804272704978, + "loss": 0.0659, + "step": 6952 + }, + { + "epoch": 0.16298926961913301, + "grad_norm": 0.7114537358283997, + "learning_rate": 0.0001872044377685029, + "loss": 0.1668, + "step": 6953 + }, + { + "epoch": 0.163012711193938, + "grad_norm": 0.34361234307289124, + "learning_rate": 0.00018720083233678228, + "loss": 0.0864, + "step": 6954 + }, + { + "epoch": 0.163036152768743, + "grad_norm": 0.5420557856559753, + "learning_rate": 0.00018719722643190752, + "loss": 0.1246, + "step": 6955 + }, + { + "epoch": 0.163059594343548, + "grad_norm": 0.7248944640159607, + "learning_rate": 0.00018719362005389814, + "loss": 0.1869, + "step": 6956 + }, + { + "epoch": 0.163083035918353, + "grad_norm": 0.7285752296447754, + "learning_rate": 0.00018719001320277373, + "loss": 0.2356, + "step": 6957 + }, + { + "epoch": 0.163106477493158, + "grad_norm": 0.6578269600868225, + "learning_rate": 0.00018718640587855385, + "loss": 0.2166, + "step": 6958 + }, + { + "epoch": 0.16312991906796298, + "grad_norm": 0.48192691802978516, + "learning_rate": 0.0001871827980812581, + "loss": 0.0933, + "step": 6959 + }, + { + "epoch": 0.16315336064276798, + "grad_norm": 0.3656095564365387, + "learning_rate": 0.000187179189810906, + "loss": 0.1613, + "step": 6960 + }, + { + "epoch": 0.16317680221757297, + "grad_norm": 0.9741886258125305, + "learning_rate": 0.0001871755810675172, + "loss": 0.2843, + "step": 6961 + }, + { + "epoch": 0.16320024379237796, + "grad_norm": 0.5235047936439514, + "learning_rate": 0.00018717197185111125, + "loss": 0.0974, + "step": 6962 + }, + { + "epoch": 0.16322368536718296, + "grad_norm": 0.5418724417686462, + "learning_rate": 0.0001871683621617077, + "loss": 0.6962, + "step": 6963 + }, + { + "epoch": 0.16324712694198795, + "grad_norm": 0.30983883142471313, + "learning_rate": 0.00018716475199932617, + "loss": 0.0871, + "step": 6964 + }, + { + "epoch": 0.16327056851679295, + "grad_norm": 0.5002550482749939, + "learning_rate": 0.00018716114136398625, + "loss": 0.1541, + "step": 6965 + }, + { + "epoch": 0.16329401009159794, + "grad_norm": 0.1603536456823349, + "learning_rate": 0.00018715753025570756, + "loss": 0.0227, + "step": 6966 + }, + { + "epoch": 0.16331745166640294, + "grad_norm": 0.2991400361061096, + "learning_rate": 0.0001871539186745096, + "loss": 0.0743, + "step": 6967 + }, + { + "epoch": 0.16334089324120796, + "grad_norm": 0.5989474654197693, + "learning_rate": 0.00018715030662041205, + "loss": 0.6931, + "step": 6968 + }, + { + "epoch": 0.16336433481601295, + "grad_norm": 0.21805506944656372, + "learning_rate": 0.00018714669409343452, + "loss": 0.055, + "step": 6969 + }, + { + "epoch": 0.16338777639081795, + "grad_norm": 0.4143700897693634, + "learning_rate": 0.00018714308109359654, + "loss": 0.1275, + "step": 6970 + }, + { + "epoch": 0.16341121796562294, + "grad_norm": 0.4844168722629547, + "learning_rate": 0.00018713946762091778, + "loss": 0.115, + "step": 6971 + }, + { + "epoch": 0.16343465954042793, + "grad_norm": 0.6232503056526184, + "learning_rate": 0.0001871358536754178, + "loss": 0.163, + "step": 6972 + }, + { + "epoch": 0.16345810111523293, + "grad_norm": 0.2729151248931885, + "learning_rate": 0.00018713223925711622, + "loss": 0.0505, + "step": 6973 + }, + { + "epoch": 0.16348154269003792, + "grad_norm": 0.6622239947319031, + "learning_rate": 0.00018712862436603267, + "loss": 0.1253, + "step": 6974 + }, + { + "epoch": 0.16350498426484292, + "grad_norm": 0.6073845028877258, + "learning_rate": 0.00018712500900218678, + "loss": 0.1133, + "step": 6975 + }, + { + "epoch": 0.1635284258396479, + "grad_norm": 0.6826187372207642, + "learning_rate": 0.0001871213931655981, + "loss": 0.2339, + "step": 6976 + }, + { + "epoch": 0.1635518674144529, + "grad_norm": 0.8882203102111816, + "learning_rate": 0.00018711777685628632, + "loss": 0.2686, + "step": 6977 + }, + { + "epoch": 0.1635753089892579, + "grad_norm": 0.45299097895622253, + "learning_rate": 0.000187114160074271, + "loss": 0.0766, + "step": 6978 + }, + { + "epoch": 0.1635987505640629, + "grad_norm": 0.6980862021446228, + "learning_rate": 0.0001871105428195718, + "loss": 0.1661, + "step": 6979 + }, + { + "epoch": 0.1636221921388679, + "grad_norm": 0.7317929267883301, + "learning_rate": 0.00018710692509220838, + "loss": 0.1275, + "step": 6980 + }, + { + "epoch": 0.16364563371367288, + "grad_norm": 0.4638831615447998, + "learning_rate": 0.0001871033068922003, + "loss": 0.1058, + "step": 6981 + }, + { + "epoch": 0.16366907528847788, + "grad_norm": 0.2190205156803131, + "learning_rate": 0.00018709968821956725, + "loss": 0.0587, + "step": 6982 + }, + { + "epoch": 0.16369251686328287, + "grad_norm": 0.5395455360412598, + "learning_rate": 0.00018709606907432887, + "loss": 0.1138, + "step": 6983 + }, + { + "epoch": 0.16371595843808787, + "grad_norm": 0.35437798500061035, + "learning_rate": 0.00018709244945650473, + "loss": 0.0855, + "step": 6984 + }, + { + "epoch": 0.16373940001289286, + "grad_norm": 0.9191519618034363, + "learning_rate": 0.00018708882936611452, + "loss": 0.218, + "step": 6985 + }, + { + "epoch": 0.16376284158769785, + "grad_norm": 0.82950359582901, + "learning_rate": 0.00018708520880317785, + "loss": 0.1833, + "step": 6986 + }, + { + "epoch": 0.16378628316250285, + "grad_norm": 1.3977596759796143, + "learning_rate": 0.00018708158776771443, + "loss": 0.1763, + "step": 6987 + }, + { + "epoch": 0.16380972473730784, + "grad_norm": 0.9271551370620728, + "learning_rate": 0.00018707796625974383, + "loss": 0.1588, + "step": 6988 + }, + { + "epoch": 0.16383316631211284, + "grad_norm": 0.5926432013511658, + "learning_rate": 0.00018707434427928576, + "loss": 0.4634, + "step": 6989 + }, + { + "epoch": 0.16385660788691783, + "grad_norm": 0.3745008111000061, + "learning_rate": 0.00018707072182635984, + "loss": 0.1262, + "step": 6990 + }, + { + "epoch": 0.16388004946172283, + "grad_norm": 0.7302194237709045, + "learning_rate": 0.00018706709890098573, + "loss": 0.7179, + "step": 6991 + }, + { + "epoch": 0.16390349103652782, + "grad_norm": 0.847789466381073, + "learning_rate": 0.00018706347550318311, + "loss": 0.2041, + "step": 6992 + }, + { + "epoch": 0.16392693261133284, + "grad_norm": 0.5068238973617554, + "learning_rate": 0.0001870598516329716, + "loss": 0.1373, + "step": 6993 + }, + { + "epoch": 0.16395037418613784, + "grad_norm": 0.5678148865699768, + "learning_rate": 0.0001870562272903709, + "loss": 0.1243, + "step": 6994 + }, + { + "epoch": 0.16397381576094283, + "grad_norm": 0.36569681763648987, + "learning_rate": 0.00018705260247540067, + "loss": 0.1414, + "step": 6995 + }, + { + "epoch": 0.16399725733574783, + "grad_norm": 0.08986152708530426, + "learning_rate": 0.00018704897718808058, + "loss": 0.0274, + "step": 6996 + }, + { + "epoch": 0.16402069891055282, + "grad_norm": 0.17815804481506348, + "learning_rate": 0.00018704535142843027, + "loss": 0.0452, + "step": 6997 + }, + { + "epoch": 0.16404414048535781, + "grad_norm": 0.5079019665718079, + "learning_rate": 0.00018704172519646946, + "loss": 0.1728, + "step": 6998 + }, + { + "epoch": 0.1640675820601628, + "grad_norm": 0.6957825422286987, + "learning_rate": 0.0001870380984922178, + "loss": 0.6822, + "step": 6999 + }, + { + "epoch": 0.1640910236349678, + "grad_norm": 0.3189159333705902, + "learning_rate": 0.00018703447131569493, + "loss": 0.0669, + "step": 7000 + }, + { + "epoch": 0.1641144652097728, + "grad_norm": 0.46518757939338684, + "learning_rate": 0.00018703084366692062, + "loss": 0.1516, + "step": 7001 + }, + { + "epoch": 0.1641379067845778, + "grad_norm": 0.19339419901371002, + "learning_rate": 0.00018702721554591446, + "loss": 0.0454, + "step": 7002 + }, + { + "epoch": 0.16416134835938279, + "grad_norm": 0.5686987042427063, + "learning_rate": 0.00018702358695269622, + "loss": 0.1557, + "step": 7003 + }, + { + "epoch": 0.16418478993418778, + "grad_norm": 0.16658514738082886, + "learning_rate": 0.00018701995788728552, + "loss": 0.0672, + "step": 7004 + }, + { + "epoch": 0.16420823150899277, + "grad_norm": 0.4426751136779785, + "learning_rate": 0.00018701632834970208, + "loss": 0.0875, + "step": 7005 + }, + { + "epoch": 0.16423167308379777, + "grad_norm": 0.4679994583129883, + "learning_rate": 0.00018701269833996565, + "loss": 0.1126, + "step": 7006 + }, + { + "epoch": 0.16425511465860276, + "grad_norm": 0.3835046589374542, + "learning_rate": 0.00018700906785809582, + "loss": 0.0675, + "step": 7007 + }, + { + "epoch": 0.16427855623340776, + "grad_norm": 0.609982430934906, + "learning_rate": 0.00018700543690411235, + "loss": 0.1552, + "step": 7008 + }, + { + "epoch": 0.16430199780821275, + "grad_norm": 0.19875094294548035, + "learning_rate": 0.00018700180547803495, + "loss": 0.0803, + "step": 7009 + }, + { + "epoch": 0.16432543938301775, + "grad_norm": 0.3785466253757477, + "learning_rate": 0.00018699817357988328, + "loss": 0.1047, + "step": 7010 + }, + { + "epoch": 0.16434888095782274, + "grad_norm": 0.3152969479560852, + "learning_rate": 0.0001869945412096771, + "loss": 0.072, + "step": 7011 + }, + { + "epoch": 0.16437232253262773, + "grad_norm": 0.4427037239074707, + "learning_rate": 0.00018699090836743606, + "loss": 0.0782, + "step": 7012 + }, + { + "epoch": 0.16439576410743273, + "grad_norm": 0.6188299655914307, + "learning_rate": 0.00018698727505317994, + "loss": 0.1303, + "step": 7013 + }, + { + "epoch": 0.16441920568223772, + "grad_norm": 0.6060221791267395, + "learning_rate": 0.00018698364126692842, + "loss": 0.1666, + "step": 7014 + }, + { + "epoch": 0.16444264725704272, + "grad_norm": 0.5101192593574524, + "learning_rate": 0.00018698000700870118, + "loss": 0.0917, + "step": 7015 + }, + { + "epoch": 0.1644660888318477, + "grad_norm": 0.5996703505516052, + "learning_rate": 0.00018697637227851804, + "loss": 0.2198, + "step": 7016 + }, + { + "epoch": 0.1644895304066527, + "grad_norm": 0.6314209699630737, + "learning_rate": 0.0001869727370763986, + "loss": 0.1518, + "step": 7017 + }, + { + "epoch": 0.1645129719814577, + "grad_norm": 0.5200343132019043, + "learning_rate": 0.00018696910140236268, + "loss": 0.0339, + "step": 7018 + }, + { + "epoch": 0.16453641355626272, + "grad_norm": 0.5397595167160034, + "learning_rate": 0.00018696546525642994, + "loss": 0.1194, + "step": 7019 + }, + { + "epoch": 0.16455985513106772, + "grad_norm": 1.2538318634033203, + "learning_rate": 0.00018696182863862015, + "loss": 0.2021, + "step": 7020 + }, + { + "epoch": 0.1645832967058727, + "grad_norm": 0.21283873915672302, + "learning_rate": 0.00018695819154895303, + "loss": 0.0626, + "step": 7021 + }, + { + "epoch": 0.1646067382806777, + "grad_norm": 0.602046549320221, + "learning_rate": 0.00018695455398744834, + "loss": 0.2399, + "step": 7022 + }, + { + "epoch": 0.1646301798554827, + "grad_norm": 0.805239200592041, + "learning_rate": 0.00018695091595412577, + "loss": 0.1802, + "step": 7023 + }, + { + "epoch": 0.1646536214302877, + "grad_norm": 0.18264292180538177, + "learning_rate": 0.0001869472774490051, + "loss": 0.0437, + "step": 7024 + }, + { + "epoch": 0.1646770630050927, + "grad_norm": 0.5437478423118591, + "learning_rate": 0.00018694363847210607, + "loss": 0.1812, + "step": 7025 + }, + { + "epoch": 0.16470050457989768, + "grad_norm": 0.4459090232849121, + "learning_rate": 0.0001869399990234484, + "loss": 0.061, + "step": 7026 + }, + { + "epoch": 0.16472394615470268, + "grad_norm": 0.6428391337394714, + "learning_rate": 0.00018693635910305183, + "loss": 0.1227, + "step": 7027 + }, + { + "epoch": 0.16474738772950767, + "grad_norm": 0.4352739453315735, + "learning_rate": 0.00018693271871093615, + "loss": 0.0914, + "step": 7028 + }, + { + "epoch": 0.16477082930431267, + "grad_norm": 0.5170860886573792, + "learning_rate": 0.00018692907784712107, + "loss": 0.101, + "step": 7029 + }, + { + "epoch": 0.16479427087911766, + "grad_norm": 0.20151612162590027, + "learning_rate": 0.00018692543651162638, + "loss": 0.0336, + "step": 7030 + }, + { + "epoch": 0.16481771245392265, + "grad_norm": 0.6275483965873718, + "learning_rate": 0.00018692179470447187, + "loss": 0.2031, + "step": 7031 + }, + { + "epoch": 0.16484115402872765, + "grad_norm": 0.8995509147644043, + "learning_rate": 0.0001869181524256772, + "loss": 0.1136, + "step": 7032 + }, + { + "epoch": 0.16486459560353264, + "grad_norm": 0.6680006980895996, + "learning_rate": 0.00018691450967526222, + "loss": 0.1684, + "step": 7033 + }, + { + "epoch": 0.16488803717833764, + "grad_norm": 0.5788986682891846, + "learning_rate": 0.00018691086645324667, + "loss": 0.0918, + "step": 7034 + }, + { + "epoch": 0.16491147875314263, + "grad_norm": 0.8088842034339905, + "learning_rate": 0.00018690722275965032, + "loss": 0.1456, + "step": 7035 + }, + { + "epoch": 0.16493492032794763, + "grad_norm": 0.6174768209457397, + "learning_rate": 0.0001869035785944929, + "loss": 0.1364, + "step": 7036 + }, + { + "epoch": 0.16495836190275262, + "grad_norm": 0.22681789100170135, + "learning_rate": 0.00018689993395779424, + "loss": 0.051, + "step": 7037 + }, + { + "epoch": 0.16498180347755761, + "grad_norm": 0.22517699003219604, + "learning_rate": 0.00018689628884957412, + "loss": 0.0581, + "step": 7038 + }, + { + "epoch": 0.1650052450523626, + "grad_norm": 0.8285951018333435, + "learning_rate": 0.00018689264326985226, + "loss": 0.1059, + "step": 7039 + }, + { + "epoch": 0.1650286866271676, + "grad_norm": 0.18244898319244385, + "learning_rate": 0.0001868889972186485, + "loss": 0.0276, + "step": 7040 + }, + { + "epoch": 0.1650521282019726, + "grad_norm": 0.6211318969726562, + "learning_rate": 0.00018688535069598259, + "loss": 0.1504, + "step": 7041 + }, + { + "epoch": 0.1650755697767776, + "grad_norm": 0.4667258858680725, + "learning_rate": 0.0001868817037018743, + "loss": 0.1358, + "step": 7042 + }, + { + "epoch": 0.1650990113515826, + "grad_norm": 0.3546099066734314, + "learning_rate": 0.00018687805623634346, + "loss": 0.0706, + "step": 7043 + }, + { + "epoch": 0.1651224529263876, + "grad_norm": 0.7664965391159058, + "learning_rate": 0.00018687440829940984, + "loss": 0.1812, + "step": 7044 + }, + { + "epoch": 0.1651458945011926, + "grad_norm": 0.3142024576663971, + "learning_rate": 0.00018687075989109325, + "loss": 0.0681, + "step": 7045 + }, + { + "epoch": 0.1651693360759976, + "grad_norm": 0.5601349472999573, + "learning_rate": 0.00018686711101141348, + "loss": 0.1987, + "step": 7046 + }, + { + "epoch": 0.1651927776508026, + "grad_norm": 0.37678542733192444, + "learning_rate": 0.00018686346166039029, + "loss": 0.0858, + "step": 7047 + }, + { + "epoch": 0.16521621922560759, + "grad_norm": 0.3262762725353241, + "learning_rate": 0.00018685981183804354, + "loss": 0.0798, + "step": 7048 + }, + { + "epoch": 0.16523966080041258, + "grad_norm": 0.48551326990127563, + "learning_rate": 0.00018685616154439302, + "loss": 0.1051, + "step": 7049 + }, + { + "epoch": 0.16526310237521757, + "grad_norm": 0.22923581302165985, + "learning_rate": 0.0001868525107794585, + "loss": 0.0718, + "step": 7050 + }, + { + "epoch": 0.16528654395002257, + "grad_norm": 0.503893256187439, + "learning_rate": 0.00018684885954325982, + "loss": 0.1236, + "step": 7051 + }, + { + "epoch": 0.16530998552482756, + "grad_norm": 0.4617311656475067, + "learning_rate": 0.00018684520783581677, + "loss": 0.1247, + "step": 7052 + }, + { + "epoch": 0.16533342709963256, + "grad_norm": 1.1668776273727417, + "learning_rate": 0.0001868415556571492, + "loss": 0.1168, + "step": 7053 + }, + { + "epoch": 0.16535686867443755, + "grad_norm": 0.6589359641075134, + "learning_rate": 0.00018683790300727692, + "loss": 0.1301, + "step": 7054 + }, + { + "epoch": 0.16538031024924255, + "grad_norm": 0.5928488969802856, + "learning_rate": 0.00018683424988621972, + "loss": 0.0703, + "step": 7055 + }, + { + "epoch": 0.16540375182404754, + "grad_norm": 0.2611190378665924, + "learning_rate": 0.00018683059629399744, + "loss": 0.0479, + "step": 7056 + }, + { + "epoch": 0.16542719339885253, + "grad_norm": 0.5989179611206055, + "learning_rate": 0.0001868269422306299, + "loss": 0.7973, + "step": 7057 + }, + { + "epoch": 0.16545063497365753, + "grad_norm": 0.5610138177871704, + "learning_rate": 0.00018682328769613693, + "loss": 0.1133, + "step": 7058 + }, + { + "epoch": 0.16547407654846252, + "grad_norm": 0.4122369587421417, + "learning_rate": 0.0001868196326905384, + "loss": 0.0806, + "step": 7059 + }, + { + "epoch": 0.16549751812326752, + "grad_norm": 0.6133387684822083, + "learning_rate": 0.00018681597721385403, + "loss": 0.1214, + "step": 7060 + }, + { + "epoch": 0.1655209596980725, + "grad_norm": 0.3164174258708954, + "learning_rate": 0.00018681232126610377, + "loss": 0.0568, + "step": 7061 + }, + { + "epoch": 0.1655444012728775, + "grad_norm": 0.19957880675792694, + "learning_rate": 0.00018680866484730742, + "loss": 0.0345, + "step": 7062 + }, + { + "epoch": 0.1655678428476825, + "grad_norm": 0.7890961766242981, + "learning_rate": 0.00018680500795748478, + "loss": 0.1098, + "step": 7063 + }, + { + "epoch": 0.1655912844224875, + "grad_norm": 0.26811325550079346, + "learning_rate": 0.00018680135059665577, + "loss": 0.0825, + "step": 7064 + }, + { + "epoch": 0.1656147259972925, + "grad_norm": 0.6697849631309509, + "learning_rate": 0.00018679769276484017, + "loss": 0.13, + "step": 7065 + }, + { + "epoch": 0.16563816757209748, + "grad_norm": 0.7465050220489502, + "learning_rate": 0.00018679403446205783, + "loss": 0.138, + "step": 7066 + }, + { + "epoch": 0.16566160914690248, + "grad_norm": 0.11144320666790009, + "learning_rate": 0.00018679037568832864, + "loss": 0.0163, + "step": 7067 + }, + { + "epoch": 0.16568505072170747, + "grad_norm": 0.679355263710022, + "learning_rate": 0.00018678671644367244, + "loss": 0.1338, + "step": 7068 + }, + { + "epoch": 0.16570849229651247, + "grad_norm": 0.6762435436248779, + "learning_rate": 0.00018678305672810902, + "loss": 0.228, + "step": 7069 + }, + { + "epoch": 0.1657319338713175, + "grad_norm": 0.5376651883125305, + "learning_rate": 0.00018677939654165835, + "loss": 0.0849, + "step": 7070 + }, + { + "epoch": 0.16575537544612248, + "grad_norm": 0.49673065543174744, + "learning_rate": 0.00018677573588434023, + "loss": 0.1899, + "step": 7071 + }, + { + "epoch": 0.16577881702092748, + "grad_norm": 0.7275193333625793, + "learning_rate": 0.00018677207475617454, + "loss": 0.1383, + "step": 7072 + }, + { + "epoch": 0.16580225859573247, + "grad_norm": 0.5049650073051453, + "learning_rate": 0.0001867684131571811, + "loss": 0.2281, + "step": 7073 + }, + { + "epoch": 0.16582570017053747, + "grad_norm": 0.7918320894241333, + "learning_rate": 0.0001867647510873798, + "loss": 0.2262, + "step": 7074 + }, + { + "epoch": 0.16584914174534246, + "grad_norm": 0.5072224140167236, + "learning_rate": 0.00018676108854679056, + "loss": 0.114, + "step": 7075 + }, + { + "epoch": 0.16587258332014745, + "grad_norm": 0.5407446026802063, + "learning_rate": 0.00018675742553543319, + "loss": 0.1888, + "step": 7076 + }, + { + "epoch": 0.16589602489495245, + "grad_norm": 0.2796923518180847, + "learning_rate": 0.0001867537620533276, + "loss": 0.0902, + "step": 7077 + }, + { + "epoch": 0.16591946646975744, + "grad_norm": 0.9934393167495728, + "learning_rate": 0.00018675009810049367, + "loss": 0.1808, + "step": 7078 + }, + { + "epoch": 0.16594290804456244, + "grad_norm": 0.5196599960327148, + "learning_rate": 0.00018674643367695123, + "loss": 0.1161, + "step": 7079 + }, + { + "epoch": 0.16596634961936743, + "grad_norm": 0.6096166372299194, + "learning_rate": 0.0001867427687827202, + "loss": 0.1331, + "step": 7080 + }, + { + "epoch": 0.16598979119417243, + "grad_norm": 0.8152663707733154, + "learning_rate": 0.00018673910341782053, + "loss": 0.1746, + "step": 7081 + }, + { + "epoch": 0.16601323276897742, + "grad_norm": 0.4787577688694, + "learning_rate": 0.00018673543758227198, + "loss": 0.1254, + "step": 7082 + }, + { + "epoch": 0.16603667434378241, + "grad_norm": 0.22846442461013794, + "learning_rate": 0.00018673177127609454, + "loss": 0.0769, + "step": 7083 + }, + { + "epoch": 0.1660601159185874, + "grad_norm": 0.14655379951000214, + "learning_rate": 0.00018672810449930807, + "loss": 0.0338, + "step": 7084 + }, + { + "epoch": 0.1660835574933924, + "grad_norm": 0.4333442449569702, + "learning_rate": 0.00018672443725193248, + "loss": 0.0906, + "step": 7085 + }, + { + "epoch": 0.1661069990681974, + "grad_norm": 0.5185546278953552, + "learning_rate": 0.0001867207695339876, + "loss": 0.1498, + "step": 7086 + }, + { + "epoch": 0.1661304406430024, + "grad_norm": 0.38315120339393616, + "learning_rate": 0.00018671710134549345, + "loss": 0.0768, + "step": 7087 + }, + { + "epoch": 0.16615388221780739, + "grad_norm": 0.43518179655075073, + "learning_rate": 0.00018671343268646982, + "loss": 0.069, + "step": 7088 + }, + { + "epoch": 0.16617732379261238, + "grad_norm": 0.713311493396759, + "learning_rate": 0.0001867097635569367, + "loss": 0.1293, + "step": 7089 + }, + { + "epoch": 0.16620076536741737, + "grad_norm": 0.21047714352607727, + "learning_rate": 0.00018670609395691394, + "loss": 0.0577, + "step": 7090 + }, + { + "epoch": 0.16622420694222237, + "grad_norm": 0.5125988125801086, + "learning_rate": 0.0001867024238864215, + "loss": 0.1939, + "step": 7091 + }, + { + "epoch": 0.16624764851702736, + "grad_norm": 0.4535394310951233, + "learning_rate": 0.00018669875334547927, + "loss": 0.1199, + "step": 7092 + }, + { + "epoch": 0.16627109009183236, + "grad_norm": 0.20619353652000427, + "learning_rate": 0.00018669508233410716, + "loss": 0.0484, + "step": 7093 + }, + { + "epoch": 0.16629453166663735, + "grad_norm": 0.13177305459976196, + "learning_rate": 0.0001866914108523251, + "loss": 0.0277, + "step": 7094 + }, + { + "epoch": 0.16631797324144237, + "grad_norm": 0.5410699248313904, + "learning_rate": 0.00018668773890015297, + "loss": 0.1371, + "step": 7095 + }, + { + "epoch": 0.16634141481624737, + "grad_norm": 0.4920395314693451, + "learning_rate": 0.00018668406647761077, + "loss": 0.1582, + "step": 7096 + }, + { + "epoch": 0.16636485639105236, + "grad_norm": 0.5023159384727478, + "learning_rate": 0.0001866803935847184, + "loss": 0.155, + "step": 7097 + }, + { + "epoch": 0.16638829796585736, + "grad_norm": 0.6395843029022217, + "learning_rate": 0.0001866767202214957, + "loss": 0.147, + "step": 7098 + }, + { + "epoch": 0.16641173954066235, + "grad_norm": 0.47093722224235535, + "learning_rate": 0.00018667304638796277, + "loss": 0.5367, + "step": 7099 + }, + { + "epoch": 0.16643518111546735, + "grad_norm": 0.6324458122253418, + "learning_rate": 0.0001866693720841394, + "loss": 0.6578, + "step": 7100 + }, + { + "epoch": 0.16645862269027234, + "grad_norm": 0.32038626074790955, + "learning_rate": 0.00018666569731004562, + "loss": 0.064, + "step": 7101 + }, + { + "epoch": 0.16648206426507733, + "grad_norm": 0.4958682954311371, + "learning_rate": 0.0001866620220657013, + "loss": 0.1401, + "step": 7102 + }, + { + "epoch": 0.16650550583988233, + "grad_norm": 0.6284269094467163, + "learning_rate": 0.0001866583463511264, + "loss": 0.1788, + "step": 7103 + }, + { + "epoch": 0.16652894741468732, + "grad_norm": 0.7378327250480652, + "learning_rate": 0.0001866546701663409, + "loss": 0.1859, + "step": 7104 + }, + { + "epoch": 0.16655238898949232, + "grad_norm": 0.36393794417381287, + "learning_rate": 0.0001866509935113647, + "loss": 0.0953, + "step": 7105 + }, + { + "epoch": 0.1665758305642973, + "grad_norm": 0.4675121605396271, + "learning_rate": 0.0001866473163862178, + "loss": 0.077, + "step": 7106 + }, + { + "epoch": 0.1665992721391023, + "grad_norm": 0.42860162258148193, + "learning_rate": 0.00018664363879092013, + "loss": 0.0495, + "step": 7107 + }, + { + "epoch": 0.1666227137139073, + "grad_norm": 0.4625133275985718, + "learning_rate": 0.00018663996072549162, + "loss": 0.3954, + "step": 7108 + }, + { + "epoch": 0.1666461552887123, + "grad_norm": 0.5037506818771362, + "learning_rate": 0.00018663628218995224, + "loss": 0.0756, + "step": 7109 + }, + { + "epoch": 0.1666695968635173, + "grad_norm": 0.5472611784934998, + "learning_rate": 0.000186632603184322, + "loss": 0.0585, + "step": 7110 + }, + { + "epoch": 0.16669303843832228, + "grad_norm": 0.25995662808418274, + "learning_rate": 0.00018662892370862079, + "loss": 0.079, + "step": 7111 + }, + { + "epoch": 0.16671648001312728, + "grad_norm": 0.5415239930152893, + "learning_rate": 0.00018662524376286858, + "loss": 0.0956, + "step": 7112 + }, + { + "epoch": 0.16673992158793227, + "grad_norm": 0.6822202801704407, + "learning_rate": 0.0001866215633470854, + "loss": 0.237, + "step": 7113 + }, + { + "epoch": 0.16676336316273727, + "grad_norm": 0.3569190800189972, + "learning_rate": 0.00018661788246129114, + "loss": 0.1004, + "step": 7114 + }, + { + "epoch": 0.16678680473754226, + "grad_norm": 0.5612393617630005, + "learning_rate": 0.00018661420110550585, + "loss": 0.0973, + "step": 7115 + }, + { + "epoch": 0.16681024631234725, + "grad_norm": 0.599284291267395, + "learning_rate": 0.00018661051927974947, + "loss": 0.1712, + "step": 7116 + }, + { + "epoch": 0.16683368788715225, + "grad_norm": 0.4140212833881378, + "learning_rate": 0.00018660683698404197, + "loss": 0.0501, + "step": 7117 + }, + { + "epoch": 0.16685712946195724, + "grad_norm": 0.4845835268497467, + "learning_rate": 0.00018660315421840332, + "loss": 0.1491, + "step": 7118 + }, + { + "epoch": 0.16688057103676224, + "grad_norm": 0.21133127808570862, + "learning_rate": 0.00018659947098285354, + "loss": 0.0442, + "step": 7119 + }, + { + "epoch": 0.16690401261156723, + "grad_norm": 0.440148264169693, + "learning_rate": 0.00018659578727741257, + "loss": 0.0946, + "step": 7120 + }, + { + "epoch": 0.16692745418637225, + "grad_norm": 0.35464441776275635, + "learning_rate": 0.00018659210310210047, + "loss": 0.0373, + "step": 7121 + }, + { + "epoch": 0.16695089576117725, + "grad_norm": 0.14486736059188843, + "learning_rate": 0.00018658841845693716, + "loss": 0.0177, + "step": 7122 + }, + { + "epoch": 0.16697433733598224, + "grad_norm": 0.7557530403137207, + "learning_rate": 0.00018658473334194265, + "loss": 0.185, + "step": 7123 + }, + { + "epoch": 0.16699777891078724, + "grad_norm": 0.44748133420944214, + "learning_rate": 0.00018658104775713693, + "loss": 0.131, + "step": 7124 + }, + { + "epoch": 0.16702122048559223, + "grad_norm": 0.30091628432273865, + "learning_rate": 0.00018657736170254004, + "loss": 0.0365, + "step": 7125 + }, + { + "epoch": 0.16704466206039723, + "grad_norm": 0.5139893889427185, + "learning_rate": 0.00018657367517817193, + "loss": 0.0859, + "step": 7126 + }, + { + "epoch": 0.16706810363520222, + "grad_norm": 0.630373477935791, + "learning_rate": 0.00018656998818405263, + "loss": 0.1154, + "step": 7127 + }, + { + "epoch": 0.16709154521000721, + "grad_norm": 0.5671384930610657, + "learning_rate": 0.00018656630072020215, + "loss": 0.1838, + "step": 7128 + }, + { + "epoch": 0.1671149867848122, + "grad_norm": 0.6212090849876404, + "learning_rate": 0.0001865626127866405, + "loss": 0.5797, + "step": 7129 + }, + { + "epoch": 0.1671384283596172, + "grad_norm": 0.6193507313728333, + "learning_rate": 0.00018655892438338762, + "loss": 0.1685, + "step": 7130 + }, + { + "epoch": 0.1671618699344222, + "grad_norm": 0.6944594383239746, + "learning_rate": 0.00018655523551046362, + "loss": 0.74, + "step": 7131 + }, + { + "epoch": 0.1671853115092272, + "grad_norm": 0.19204284250736237, + "learning_rate": 0.0001865515461678885, + "loss": 0.027, + "step": 7132 + }, + { + "epoch": 0.16720875308403219, + "grad_norm": 0.516349732875824, + "learning_rate": 0.0001865478563556822, + "loss": 0.1265, + "step": 7133 + }, + { + "epoch": 0.16723219465883718, + "grad_norm": 1.0567842721939087, + "learning_rate": 0.00018654416607386486, + "loss": 0.2077, + "step": 7134 + }, + { + "epoch": 0.16725563623364217, + "grad_norm": 0.09959665685892105, + "learning_rate": 0.00018654047532245643, + "loss": 0.0145, + "step": 7135 + }, + { + "epoch": 0.16727907780844717, + "grad_norm": 0.13096754252910614, + "learning_rate": 0.00018653678410147692, + "loss": 0.0509, + "step": 7136 + }, + { + "epoch": 0.16730251938325216, + "grad_norm": 0.5960225462913513, + "learning_rate": 0.00018653309241094642, + "loss": 0.1323, + "step": 7137 + }, + { + "epoch": 0.16732596095805716, + "grad_norm": 0.6397585868835449, + "learning_rate": 0.0001865294002508849, + "loss": 0.1894, + "step": 7138 + }, + { + "epoch": 0.16734940253286215, + "grad_norm": 0.30539578199386597, + "learning_rate": 0.0001865257076213124, + "loss": 0.0535, + "step": 7139 + }, + { + "epoch": 0.16737284410766715, + "grad_norm": 0.2610195279121399, + "learning_rate": 0.00018652201452224905, + "loss": 0.0477, + "step": 7140 + }, + { + "epoch": 0.16739628568247214, + "grad_norm": 0.3127722144126892, + "learning_rate": 0.00018651832095371474, + "loss": 0.0651, + "step": 7141 + }, + { + "epoch": 0.16741972725727713, + "grad_norm": 0.6421046853065491, + "learning_rate": 0.00018651462691572963, + "loss": 0.1809, + "step": 7142 + }, + { + "epoch": 0.16744316883208213, + "grad_norm": 0.5778199434280396, + "learning_rate": 0.00018651093240831373, + "loss": 0.1061, + "step": 7143 + }, + { + "epoch": 0.16746661040688712, + "grad_norm": 0.7125733494758606, + "learning_rate": 0.00018650723743148702, + "loss": 0.2412, + "step": 7144 + }, + { + "epoch": 0.16749005198169212, + "grad_norm": 0.38409703969955444, + "learning_rate": 0.00018650354198526968, + "loss": 0.0815, + "step": 7145 + }, + { + "epoch": 0.1675134935564971, + "grad_norm": 0.44847336411476135, + "learning_rate": 0.00018649984606968165, + "loss": 0.1048, + "step": 7146 + }, + { + "epoch": 0.16753693513130213, + "grad_norm": 0.6289908289909363, + "learning_rate": 0.00018649614968474303, + "loss": 0.167, + "step": 7147 + }, + { + "epoch": 0.16756037670610713, + "grad_norm": 0.682898759841919, + "learning_rate": 0.00018649245283047387, + "loss": 0.1741, + "step": 7148 + }, + { + "epoch": 0.16758381828091212, + "grad_norm": 0.3949597179889679, + "learning_rate": 0.00018648875550689425, + "loss": 0.0668, + "step": 7149 + }, + { + "epoch": 0.16760725985571712, + "grad_norm": 0.4953908920288086, + "learning_rate": 0.00018648505771402417, + "loss": 0.1103, + "step": 7150 + }, + { + "epoch": 0.1676307014305221, + "grad_norm": 0.4546149969100952, + "learning_rate": 0.00018648135945188376, + "loss": 0.0827, + "step": 7151 + }, + { + "epoch": 0.1676541430053271, + "grad_norm": 0.19082704186439514, + "learning_rate": 0.00018647766072049304, + "loss": 0.0638, + "step": 7152 + }, + { + "epoch": 0.1676775845801321, + "grad_norm": 0.4809293746948242, + "learning_rate": 0.00018647396151987212, + "loss": 0.1564, + "step": 7153 + }, + { + "epoch": 0.1677010261549371, + "grad_norm": 0.6353076100349426, + "learning_rate": 0.00018647026185004105, + "loss": 0.2724, + "step": 7154 + }, + { + "epoch": 0.1677244677297421, + "grad_norm": 0.8198917508125305, + "learning_rate": 0.00018646656171101991, + "loss": 0.341, + "step": 7155 + }, + { + "epoch": 0.16774790930454708, + "grad_norm": 0.13321326673030853, + "learning_rate": 0.00018646286110282878, + "loss": 0.0568, + "step": 7156 + }, + { + "epoch": 0.16777135087935208, + "grad_norm": 0.3069111406803131, + "learning_rate": 0.00018645916002548775, + "loss": 0.0885, + "step": 7157 + }, + { + "epoch": 0.16779479245415707, + "grad_norm": 0.48046064376831055, + "learning_rate": 0.00018645545847901685, + "loss": 0.1711, + "step": 7158 + }, + { + "epoch": 0.16781823402896207, + "grad_norm": 0.24421752989292145, + "learning_rate": 0.0001864517564634362, + "loss": 0.1015, + "step": 7159 + }, + { + "epoch": 0.16784167560376706, + "grad_norm": 0.4216803014278412, + "learning_rate": 0.00018644805397876593, + "loss": 0.0982, + "step": 7160 + }, + { + "epoch": 0.16786511717857205, + "grad_norm": 0.3147313892841339, + "learning_rate": 0.00018644435102502604, + "loss": 0.0714, + "step": 7161 + }, + { + "epoch": 0.16788855875337705, + "grad_norm": 0.574037492275238, + "learning_rate": 0.0001864406476022367, + "loss": 0.1382, + "step": 7162 + }, + { + "epoch": 0.16791200032818204, + "grad_norm": 0.13499784469604492, + "learning_rate": 0.00018643694371041797, + "loss": 0.0321, + "step": 7163 + }, + { + "epoch": 0.16793544190298704, + "grad_norm": 0.5884706377983093, + "learning_rate": 0.00018643323934958996, + "loss": 0.6993, + "step": 7164 + }, + { + "epoch": 0.16795888347779203, + "grad_norm": 0.2201375961303711, + "learning_rate": 0.00018642953451977276, + "loss": 0.0649, + "step": 7165 + }, + { + "epoch": 0.16798232505259703, + "grad_norm": 0.6601713299751282, + "learning_rate": 0.0001864258292209865, + "loss": 0.2312, + "step": 7166 + }, + { + "epoch": 0.16800576662740202, + "grad_norm": 0.4844978153705597, + "learning_rate": 0.00018642212345325117, + "loss": 0.1582, + "step": 7167 + }, + { + "epoch": 0.16802920820220701, + "grad_norm": 0.5006577968597412, + "learning_rate": 0.00018641841721658706, + "loss": 0.1047, + "step": 7168 + }, + { + "epoch": 0.168052649777012, + "grad_norm": 0.170850470662117, + "learning_rate": 0.00018641471051101415, + "loss": 0.0546, + "step": 7169 + }, + { + "epoch": 0.168076091351817, + "grad_norm": 0.33017295598983765, + "learning_rate": 0.00018641100333655262, + "loss": 0.0936, + "step": 7170 + }, + { + "epoch": 0.168099532926622, + "grad_norm": 0.5349203944206238, + "learning_rate": 0.00018640729569322252, + "loss": 0.2246, + "step": 7171 + }, + { + "epoch": 0.16812297450142702, + "grad_norm": 0.34066855907440186, + "learning_rate": 0.000186403587581044, + "loss": 0.0515, + "step": 7172 + }, + { + "epoch": 0.168146416076232, + "grad_norm": 0.33118587732315063, + "learning_rate": 0.00018639987900003724, + "loss": 0.0933, + "step": 7173 + }, + { + "epoch": 0.168169857651037, + "grad_norm": 0.9057707190513611, + "learning_rate": 0.00018639616995022227, + "loss": 0.2222, + "step": 7174 + }, + { + "epoch": 0.168193299225842, + "grad_norm": 0.33639225363731384, + "learning_rate": 0.00018639246043161925, + "loss": 0.2749, + "step": 7175 + }, + { + "epoch": 0.168216740800647, + "grad_norm": 0.6029232144355774, + "learning_rate": 0.0001863887504442483, + "loss": 0.1407, + "step": 7176 + }, + { + "epoch": 0.168240182375452, + "grad_norm": 0.6390389800071716, + "learning_rate": 0.0001863850399881296, + "loss": 0.1337, + "step": 7177 + }, + { + "epoch": 0.16826362395025699, + "grad_norm": 0.24353334307670593, + "learning_rate": 0.00018638132906328322, + "loss": 0.0773, + "step": 7178 + }, + { + "epoch": 0.16828706552506198, + "grad_norm": 0.5622529983520508, + "learning_rate": 0.0001863776176697293, + "loss": 0.1503, + "step": 7179 + }, + { + "epoch": 0.16831050709986697, + "grad_norm": 1.0873489379882812, + "learning_rate": 0.000186373905807488, + "loss": 0.0807, + "step": 7180 + }, + { + "epoch": 0.16833394867467197, + "grad_norm": 0.5034977793693542, + "learning_rate": 0.0001863701934765795, + "loss": 0.1781, + "step": 7181 + }, + { + "epoch": 0.16835739024947696, + "grad_norm": 0.197564035654068, + "learning_rate": 0.00018636648067702386, + "loss": 0.0425, + "step": 7182 + }, + { + "epoch": 0.16838083182428196, + "grad_norm": 0.05869300290942192, + "learning_rate": 0.00018636276740884128, + "loss": 0.013, + "step": 7183 + }, + { + "epoch": 0.16840427339908695, + "grad_norm": 0.47377434372901917, + "learning_rate": 0.00018635905367205192, + "loss": 0.1337, + "step": 7184 + }, + { + "epoch": 0.16842771497389195, + "grad_norm": 0.6269667148590088, + "learning_rate": 0.00018635533946667588, + "loss": 0.1406, + "step": 7185 + }, + { + "epoch": 0.16845115654869694, + "grad_norm": 0.3468743860721588, + "learning_rate": 0.00018635162479273336, + "loss": 0.0828, + "step": 7186 + }, + { + "epoch": 0.16847459812350193, + "grad_norm": 0.3614303767681122, + "learning_rate": 0.00018634790965024446, + "loss": 0.1132, + "step": 7187 + }, + { + "epoch": 0.16849803969830693, + "grad_norm": 0.5447949171066284, + "learning_rate": 0.0001863441940392294, + "loss": 0.1017, + "step": 7188 + }, + { + "epoch": 0.16852148127311192, + "grad_norm": 0.8263357281684875, + "learning_rate": 0.00018634047795970833, + "loss": 0.1651, + "step": 7189 + }, + { + "epoch": 0.16854492284791692, + "grad_norm": 0.5296017527580261, + "learning_rate": 0.00018633676141170137, + "loss": 0.1293, + "step": 7190 + }, + { + "epoch": 0.1685683644227219, + "grad_norm": 0.49151259660720825, + "learning_rate": 0.00018633304439522874, + "loss": 0.0985, + "step": 7191 + }, + { + "epoch": 0.1685918059975269, + "grad_norm": 0.6578601598739624, + "learning_rate": 0.0001863293269103106, + "loss": 0.1882, + "step": 7192 + }, + { + "epoch": 0.1686152475723319, + "grad_norm": 0.3969271779060364, + "learning_rate": 0.00018632560895696703, + "loss": 0.0992, + "step": 7193 + }, + { + "epoch": 0.1686386891471369, + "grad_norm": 0.3870350420475006, + "learning_rate": 0.00018632189053521833, + "loss": 0.0672, + "step": 7194 + }, + { + "epoch": 0.1686621307219419, + "grad_norm": 0.4813297986984253, + "learning_rate": 0.00018631817164508463, + "loss": 0.167, + "step": 7195 + }, + { + "epoch": 0.16868557229674688, + "grad_norm": 0.36704790592193604, + "learning_rate": 0.0001863144522865861, + "loss": 0.084, + "step": 7196 + }, + { + "epoch": 0.16870901387155188, + "grad_norm": 0.5656521916389465, + "learning_rate": 0.0001863107324597429, + "loss": 0.1416, + "step": 7197 + }, + { + "epoch": 0.1687324554463569, + "grad_norm": 0.6311070322990417, + "learning_rate": 0.00018630701216457523, + "loss": 0.081, + "step": 7198 + }, + { + "epoch": 0.1687558970211619, + "grad_norm": 0.7136773467063904, + "learning_rate": 0.00018630329140110333, + "loss": 0.1652, + "step": 7199 + }, + { + "epoch": 0.1687793385959669, + "grad_norm": 0.35555005073547363, + "learning_rate": 0.00018629957016934735, + "loss": 0.1261, + "step": 7200 + }, + { + "epoch": 0.16880278017077188, + "grad_norm": 0.5764634609222412, + "learning_rate": 0.00018629584846932743, + "loss": 0.1653, + "step": 7201 + }, + { + "epoch": 0.16882622174557688, + "grad_norm": 0.6026821136474609, + "learning_rate": 0.00018629212630106383, + "loss": 0.1358, + "step": 7202 + }, + { + "epoch": 0.16884966332038187, + "grad_norm": 0.782097578048706, + "learning_rate": 0.00018628840366457673, + "loss": 0.2026, + "step": 7203 + }, + { + "epoch": 0.16887310489518687, + "grad_norm": 0.41740840673446655, + "learning_rate": 0.0001862846805598863, + "loss": 0.3481, + "step": 7204 + }, + { + "epoch": 0.16889654646999186, + "grad_norm": 0.3642044961452484, + "learning_rate": 0.0001862809569870128, + "loss": 0.1106, + "step": 7205 + }, + { + "epoch": 0.16891998804479685, + "grad_norm": 0.6549464464187622, + "learning_rate": 0.0001862772329459764, + "loss": 0.1632, + "step": 7206 + }, + { + "epoch": 0.16894342961960185, + "grad_norm": 0.5181927680969238, + "learning_rate": 0.00018627350843679729, + "loss": 0.1597, + "step": 7207 + }, + { + "epoch": 0.16896687119440684, + "grad_norm": 0.8305312395095825, + "learning_rate": 0.0001862697834594957, + "loss": 0.3245, + "step": 7208 + }, + { + "epoch": 0.16899031276921184, + "grad_norm": 0.55898517370224, + "learning_rate": 0.00018626605801409186, + "loss": 0.1428, + "step": 7209 + }, + { + "epoch": 0.16901375434401683, + "grad_norm": 0.3118499219417572, + "learning_rate": 0.00018626233210060594, + "loss": 0.0611, + "step": 7210 + }, + { + "epoch": 0.16903719591882183, + "grad_norm": 0.6117348074913025, + "learning_rate": 0.0001862586057190582, + "loss": 0.195, + "step": 7211 + }, + { + "epoch": 0.16906063749362682, + "grad_norm": 0.16531479358673096, + "learning_rate": 0.00018625487886946883, + "loss": 0.0183, + "step": 7212 + }, + { + "epoch": 0.16908407906843181, + "grad_norm": 0.4213234484195709, + "learning_rate": 0.00018625115155185805, + "loss": 0.0948, + "step": 7213 + }, + { + "epoch": 0.1691075206432368, + "grad_norm": 0.5389026403427124, + "learning_rate": 0.00018624742376624611, + "loss": 0.1234, + "step": 7214 + }, + { + "epoch": 0.1691309622180418, + "grad_norm": 0.46820947527885437, + "learning_rate": 0.00018624369551265325, + "loss": 0.1149, + "step": 7215 + }, + { + "epoch": 0.1691544037928468, + "grad_norm": 0.8042625784873962, + "learning_rate": 0.00018623996679109964, + "loss": 0.2097, + "step": 7216 + }, + { + "epoch": 0.1691778453676518, + "grad_norm": 0.17449522018432617, + "learning_rate": 0.00018623623760160554, + "loss": 0.0445, + "step": 7217 + }, + { + "epoch": 0.16920128694245679, + "grad_norm": 0.5115266442298889, + "learning_rate": 0.0001862325079441912, + "loss": 0.1492, + "step": 7218 + }, + { + "epoch": 0.16922472851726178, + "grad_norm": 0.40694659948349, + "learning_rate": 0.00018622877781887683, + "loss": 0.0988, + "step": 7219 + }, + { + "epoch": 0.16924817009206677, + "grad_norm": 0.6593654155731201, + "learning_rate": 0.00018622504722568272, + "loss": 0.1356, + "step": 7220 + }, + { + "epoch": 0.16927161166687177, + "grad_norm": 0.4960746169090271, + "learning_rate": 0.00018622131616462907, + "loss": 0.0727, + "step": 7221 + }, + { + "epoch": 0.16929505324167676, + "grad_norm": 0.3585556745529175, + "learning_rate": 0.00018621758463573614, + "loss": 0.0741, + "step": 7222 + }, + { + "epoch": 0.16931849481648178, + "grad_norm": 0.2254813313484192, + "learning_rate": 0.00018621385263902415, + "loss": 0.0618, + "step": 7223 + }, + { + "epoch": 0.16934193639128678, + "grad_norm": 0.570652186870575, + "learning_rate": 0.00018621012017451337, + "loss": 0.1448, + "step": 7224 + }, + { + "epoch": 0.16936537796609177, + "grad_norm": 0.15718412399291992, + "learning_rate": 0.00018620638724222409, + "loss": 0.0203, + "step": 7225 + }, + { + "epoch": 0.16938881954089677, + "grad_norm": 0.6945658922195435, + "learning_rate": 0.00018620265384217648, + "loss": 0.1553, + "step": 7226 + }, + { + "epoch": 0.16941226111570176, + "grad_norm": 0.07083947956562042, + "learning_rate": 0.0001861989199743909, + "loss": 0.0121, + "step": 7227 + }, + { + "epoch": 0.16943570269050676, + "grad_norm": 0.5558444261550903, + "learning_rate": 0.00018619518563888748, + "loss": 0.1731, + "step": 7228 + }, + { + "epoch": 0.16945914426531175, + "grad_norm": 0.509013831615448, + "learning_rate": 0.00018619145083568662, + "loss": 0.1023, + "step": 7229 + }, + { + "epoch": 0.16948258584011675, + "grad_norm": 0.18659937381744385, + "learning_rate": 0.00018618771556480854, + "loss": 0.0375, + "step": 7230 + }, + { + "epoch": 0.16950602741492174, + "grad_norm": 0.14160647988319397, + "learning_rate": 0.00018618397982627345, + "loss": 0.0312, + "step": 7231 + }, + { + "epoch": 0.16952946898972673, + "grad_norm": 0.5400662422180176, + "learning_rate": 0.00018618024362010167, + "loss": 0.1472, + "step": 7232 + }, + { + "epoch": 0.16955291056453173, + "grad_norm": 0.7308022975921631, + "learning_rate": 0.00018617650694631346, + "loss": 0.7282, + "step": 7233 + }, + { + "epoch": 0.16957635213933672, + "grad_norm": 0.5013003349304199, + "learning_rate": 0.0001861727698049291, + "loss": 0.0965, + "step": 7234 + }, + { + "epoch": 0.16959979371414172, + "grad_norm": 0.6565532088279724, + "learning_rate": 0.0001861690321959689, + "loss": 0.1213, + "step": 7235 + }, + { + "epoch": 0.1696232352889467, + "grad_norm": 0.36166393756866455, + "learning_rate": 0.00018616529411945307, + "loss": 0.0895, + "step": 7236 + }, + { + "epoch": 0.1696466768637517, + "grad_norm": 0.24696704745292664, + "learning_rate": 0.00018616155557540196, + "loss": 0.0584, + "step": 7237 + }, + { + "epoch": 0.1696701184385567, + "grad_norm": 0.495206743478775, + "learning_rate": 0.0001861578165638358, + "loss": 0.0987, + "step": 7238 + }, + { + "epoch": 0.1696935600133617, + "grad_norm": 0.6900332570075989, + "learning_rate": 0.00018615407708477496, + "loss": 0.1958, + "step": 7239 + }, + { + "epoch": 0.1697170015881667, + "grad_norm": 0.6422266364097595, + "learning_rate": 0.0001861503371382396, + "loss": 0.1891, + "step": 7240 + }, + { + "epoch": 0.16974044316297168, + "grad_norm": 0.8490719795227051, + "learning_rate": 0.00018614659672425013, + "loss": 0.2507, + "step": 7241 + }, + { + "epoch": 0.16976388473777668, + "grad_norm": 0.6368823647499084, + "learning_rate": 0.0001861428558428268, + "loss": 0.1238, + "step": 7242 + }, + { + "epoch": 0.16978732631258167, + "grad_norm": 0.1759035438299179, + "learning_rate": 0.0001861391144939899, + "loss": 0.0336, + "step": 7243 + }, + { + "epoch": 0.16981076788738667, + "grad_norm": 0.3921866714954376, + "learning_rate": 0.00018613537267775975, + "loss": 0.0615, + "step": 7244 + }, + { + "epoch": 0.16983420946219166, + "grad_norm": 0.7488810420036316, + "learning_rate": 0.00018613163039415666, + "loss": 0.1777, + "step": 7245 + }, + { + "epoch": 0.16985765103699665, + "grad_norm": 0.755702793598175, + "learning_rate": 0.0001861278876432009, + "loss": 0.1135, + "step": 7246 + }, + { + "epoch": 0.16988109261180165, + "grad_norm": 0.5731063485145569, + "learning_rate": 0.0001861241444249128, + "loss": 0.1157, + "step": 7247 + }, + { + "epoch": 0.16990453418660664, + "grad_norm": 0.5435763001441956, + "learning_rate": 0.0001861204007393127, + "loss": 0.1265, + "step": 7248 + }, + { + "epoch": 0.16992797576141166, + "grad_norm": 0.4004881680011749, + "learning_rate": 0.00018611665658642084, + "loss": 0.0814, + "step": 7249 + }, + { + "epoch": 0.16995141733621666, + "grad_norm": 0.7992649674415588, + "learning_rate": 0.0001861129119662576, + "loss": 0.1534, + "step": 7250 + }, + { + "epoch": 0.16997485891102165, + "grad_norm": 0.7113029360771179, + "learning_rate": 0.00018610916687884332, + "loss": 0.1494, + "step": 7251 + }, + { + "epoch": 0.16999830048582665, + "grad_norm": 0.557433009147644, + "learning_rate": 0.0001861054213241982, + "loss": 0.0992, + "step": 7252 + }, + { + "epoch": 0.17002174206063164, + "grad_norm": 0.6143291592597961, + "learning_rate": 0.0001861016753023427, + "loss": 0.1761, + "step": 7253 + }, + { + "epoch": 0.17004518363543664, + "grad_norm": 0.5330901145935059, + "learning_rate": 0.00018609792881329708, + "loss": 0.1473, + "step": 7254 + }, + { + "epoch": 0.17006862521024163, + "grad_norm": 0.5125857591629028, + "learning_rate": 0.00018609418185708168, + "loss": 0.1394, + "step": 7255 + }, + { + "epoch": 0.17009206678504663, + "grad_norm": 0.6466652154922485, + "learning_rate": 0.00018609043443371684, + "loss": 0.1764, + "step": 7256 + }, + { + "epoch": 0.17011550835985162, + "grad_norm": 0.7963827252388, + "learning_rate": 0.00018608668654322285, + "loss": 0.1211, + "step": 7257 + }, + { + "epoch": 0.1701389499346566, + "grad_norm": 0.40503206849098206, + "learning_rate": 0.00018608293818562008, + "loss": 0.0809, + "step": 7258 + }, + { + "epoch": 0.1701623915094616, + "grad_norm": 0.4265243411064148, + "learning_rate": 0.00018607918936092887, + "loss": 0.072, + "step": 7259 + }, + { + "epoch": 0.1701858330842666, + "grad_norm": 1.0822815895080566, + "learning_rate": 0.00018607544006916956, + "loss": 0.2606, + "step": 7260 + }, + { + "epoch": 0.1702092746590716, + "grad_norm": 0.37333500385284424, + "learning_rate": 0.0001860716903103625, + "loss": 0.1407, + "step": 7261 + }, + { + "epoch": 0.1702327162338766, + "grad_norm": 0.49635767936706543, + "learning_rate": 0.00018606794008452804, + "loss": 0.101, + "step": 7262 + }, + { + "epoch": 0.17025615780868159, + "grad_norm": 0.9342376589775085, + "learning_rate": 0.0001860641893916865, + "loss": 0.1655, + "step": 7263 + }, + { + "epoch": 0.17027959938348658, + "grad_norm": 0.7403984069824219, + "learning_rate": 0.00018606043823185826, + "loss": 0.1755, + "step": 7264 + }, + { + "epoch": 0.17030304095829157, + "grad_norm": 0.5350860953330994, + "learning_rate": 0.00018605668660506367, + "loss": 0.1109, + "step": 7265 + }, + { + "epoch": 0.17032648253309657, + "grad_norm": 0.6852744817733765, + "learning_rate": 0.000186052934511323, + "loss": 0.1717, + "step": 7266 + }, + { + "epoch": 0.17034992410790156, + "grad_norm": 0.7522061467170715, + "learning_rate": 0.00018604918195065678, + "loss": 0.2523, + "step": 7267 + }, + { + "epoch": 0.17037336568270656, + "grad_norm": 0.7986385822296143, + "learning_rate": 0.00018604542892308525, + "loss": 0.1316, + "step": 7268 + }, + { + "epoch": 0.17039680725751155, + "grad_norm": 0.8743409514427185, + "learning_rate": 0.00018604167542862882, + "loss": 0.2864, + "step": 7269 + }, + { + "epoch": 0.17042024883231655, + "grad_norm": 0.6186668872833252, + "learning_rate": 0.00018603792146730782, + "loss": 0.0881, + "step": 7270 + }, + { + "epoch": 0.17044369040712154, + "grad_norm": 0.40489280223846436, + "learning_rate": 0.00018603416703914262, + "loss": 0.142, + "step": 7271 + }, + { + "epoch": 0.17046713198192653, + "grad_norm": 0.3876563012599945, + "learning_rate": 0.0001860304121441536, + "loss": 0.0685, + "step": 7272 + }, + { + "epoch": 0.17049057355673153, + "grad_norm": 0.3351331651210785, + "learning_rate": 0.0001860266567823612, + "loss": 0.0582, + "step": 7273 + }, + { + "epoch": 0.17051401513153655, + "grad_norm": 0.7228589653968811, + "learning_rate": 0.0001860229009537857, + "loss": 0.5933, + "step": 7274 + }, + { + "epoch": 0.17053745670634154, + "grad_norm": 0.6178679466247559, + "learning_rate": 0.00018601914465844757, + "loss": 0.1036, + "step": 7275 + }, + { + "epoch": 0.17056089828114654, + "grad_norm": 0.2077976018190384, + "learning_rate": 0.0001860153878963671, + "loss": 0.0545, + "step": 7276 + }, + { + "epoch": 0.17058433985595153, + "grad_norm": 0.6647411584854126, + "learning_rate": 0.00018601163066756473, + "loss": 0.109, + "step": 7277 + }, + { + "epoch": 0.17060778143075653, + "grad_norm": 0.14480946958065033, + "learning_rate": 0.00018600787297206085, + "loss": 0.0303, + "step": 7278 + }, + { + "epoch": 0.17063122300556152, + "grad_norm": 0.7515875101089478, + "learning_rate": 0.0001860041148098758, + "loss": 0.2369, + "step": 7279 + }, + { + "epoch": 0.17065466458036652, + "grad_norm": 0.2213079184293747, + "learning_rate": 0.00018600035618103003, + "loss": 0.0749, + "step": 7280 + }, + { + "epoch": 0.1706781061551715, + "grad_norm": 0.20127516984939575, + "learning_rate": 0.0001859965970855439, + "loss": 0.0286, + "step": 7281 + }, + { + "epoch": 0.1707015477299765, + "grad_norm": 0.5882177948951721, + "learning_rate": 0.00018599283752343779, + "loss": 0.1502, + "step": 7282 + }, + { + "epoch": 0.1707249893047815, + "grad_norm": 0.35220569372177124, + "learning_rate": 0.00018598907749473216, + "loss": 0.0731, + "step": 7283 + }, + { + "epoch": 0.1707484308795865, + "grad_norm": 0.20252537727355957, + "learning_rate": 0.0001859853169994474, + "loss": 0.0506, + "step": 7284 + }, + { + "epoch": 0.1707718724543915, + "grad_norm": 0.6082836389541626, + "learning_rate": 0.00018598155603760384, + "loss": 0.1212, + "step": 7285 + }, + { + "epoch": 0.17079531402919648, + "grad_norm": 0.5812417268753052, + "learning_rate": 0.00018597779460922197, + "loss": 0.2434, + "step": 7286 + }, + { + "epoch": 0.17081875560400148, + "grad_norm": 0.5226956009864807, + "learning_rate": 0.00018597403271432216, + "loss": 0.2998, + "step": 7287 + }, + { + "epoch": 0.17084219717880647, + "grad_norm": 0.536309003829956, + "learning_rate": 0.00018597027035292482, + "loss": 0.4428, + "step": 7288 + }, + { + "epoch": 0.17086563875361147, + "grad_norm": 0.4921988844871521, + "learning_rate": 0.0001859665075250504, + "loss": 0.5765, + "step": 7289 + }, + { + "epoch": 0.17088908032841646, + "grad_norm": 0.3504329025745392, + "learning_rate": 0.00018596274423071925, + "loss": 0.0559, + "step": 7290 + }, + { + "epoch": 0.17091252190322145, + "grad_norm": 0.34958726167678833, + "learning_rate": 0.00018595898046995185, + "loss": 0.1274, + "step": 7291 + }, + { + "epoch": 0.17093596347802645, + "grad_norm": 0.2723616063594818, + "learning_rate": 0.00018595521624276864, + "loss": 0.0686, + "step": 7292 + }, + { + "epoch": 0.17095940505283144, + "grad_norm": 0.2686370015144348, + "learning_rate": 0.00018595145154918998, + "loss": 0.0733, + "step": 7293 + }, + { + "epoch": 0.17098284662763644, + "grad_norm": 0.34288373589515686, + "learning_rate": 0.00018594768638923633, + "loss": 0.0721, + "step": 7294 + }, + { + "epoch": 0.17100628820244143, + "grad_norm": 0.6885205507278442, + "learning_rate": 0.0001859439207629281, + "loss": 0.2084, + "step": 7295 + }, + { + "epoch": 0.17102972977724643, + "grad_norm": 0.6144954562187195, + "learning_rate": 0.00018594015467028577, + "loss": 0.1853, + "step": 7296 + }, + { + "epoch": 0.17105317135205142, + "grad_norm": 0.5016771554946899, + "learning_rate": 0.00018593638811132974, + "loss": 0.6495, + "step": 7297 + }, + { + "epoch": 0.17107661292685641, + "grad_norm": 0.6700524091720581, + "learning_rate": 0.00018593262108608044, + "loss": 0.8991, + "step": 7298 + }, + { + "epoch": 0.1711000545016614, + "grad_norm": 0.4472377300262451, + "learning_rate": 0.00018592885359455834, + "loss": 0.156, + "step": 7299 + }, + { + "epoch": 0.17112349607646643, + "grad_norm": 0.5283384919166565, + "learning_rate": 0.00018592508563678383, + "loss": 0.1335, + "step": 7300 + }, + { + "epoch": 0.17114693765127142, + "grad_norm": 0.5526187419891357, + "learning_rate": 0.00018592131721277745, + "loss": 0.1871, + "step": 7301 + }, + { + "epoch": 0.17117037922607642, + "grad_norm": 0.4097871482372284, + "learning_rate": 0.00018591754832255952, + "loss": 0.1152, + "step": 7302 + }, + { + "epoch": 0.1711938208008814, + "grad_norm": 0.23060598969459534, + "learning_rate": 0.0001859137789661506, + "loss": 0.0452, + "step": 7303 + }, + { + "epoch": 0.1712172623756864, + "grad_norm": 0.3913421332836151, + "learning_rate": 0.0001859100091435711, + "loss": 0.3545, + "step": 7304 + }, + { + "epoch": 0.1712407039504914, + "grad_norm": 0.3635166883468628, + "learning_rate": 0.00018590623885484143, + "loss": 0.411, + "step": 7305 + }, + { + "epoch": 0.1712641455252964, + "grad_norm": 0.3969447612762451, + "learning_rate": 0.00018590246809998214, + "loss": 0.0708, + "step": 7306 + }, + { + "epoch": 0.1712875871001014, + "grad_norm": 0.23411917686462402, + "learning_rate": 0.0001858986968790136, + "loss": 0.0427, + "step": 7307 + }, + { + "epoch": 0.17131102867490638, + "grad_norm": 0.43869802355766296, + "learning_rate": 0.00018589492519195634, + "loss": 0.7028, + "step": 7308 + }, + { + "epoch": 0.17133447024971138, + "grad_norm": 0.4035457372665405, + "learning_rate": 0.0001858911530388308, + "loss": 0.0986, + "step": 7309 + }, + { + "epoch": 0.17135791182451637, + "grad_norm": 0.1274019479751587, + "learning_rate": 0.00018588738041965748, + "loss": 0.0299, + "step": 7310 + }, + { + "epoch": 0.17138135339932137, + "grad_norm": 0.6543773412704468, + "learning_rate": 0.00018588360733445676, + "loss": 0.1747, + "step": 7311 + }, + { + "epoch": 0.17140479497412636, + "grad_norm": 0.25326165556907654, + "learning_rate": 0.0001858798337832492, + "loss": 0.0722, + "step": 7312 + }, + { + "epoch": 0.17142823654893136, + "grad_norm": 0.37529945373535156, + "learning_rate": 0.00018587605976605524, + "loss": 0.0582, + "step": 7313 + }, + { + "epoch": 0.17145167812373635, + "grad_norm": 0.3698737323284149, + "learning_rate": 0.0001858722852828954, + "loss": 0.0897, + "step": 7314 + }, + { + "epoch": 0.17147511969854135, + "grad_norm": 0.7504061460494995, + "learning_rate": 0.0001858685103337901, + "loss": 0.1481, + "step": 7315 + }, + { + "epoch": 0.17149856127334634, + "grad_norm": 0.37353581190109253, + "learning_rate": 0.00018586473491875985, + "loss": 0.092, + "step": 7316 + }, + { + "epoch": 0.17152200284815133, + "grad_norm": 0.5512000322341919, + "learning_rate": 0.00018586095903782513, + "loss": 0.1689, + "step": 7317 + }, + { + "epoch": 0.17154544442295633, + "grad_norm": 0.20060117542743683, + "learning_rate": 0.0001858571826910064, + "loss": 0.0397, + "step": 7318 + }, + { + "epoch": 0.17156888599776132, + "grad_norm": 0.4359113574028015, + "learning_rate": 0.00018585340587832423, + "loss": 0.1579, + "step": 7319 + }, + { + "epoch": 0.17159232757256632, + "grad_norm": 0.608471155166626, + "learning_rate": 0.00018584962859979903, + "loss": 0.1503, + "step": 7320 + }, + { + "epoch": 0.1716157691473713, + "grad_norm": 0.37048232555389404, + "learning_rate": 0.00018584585085545138, + "loss": 0.1126, + "step": 7321 + }, + { + "epoch": 0.1716392107221763, + "grad_norm": 0.33379948139190674, + "learning_rate": 0.0001858420726453017, + "loss": 0.0919, + "step": 7322 + }, + { + "epoch": 0.1716626522969813, + "grad_norm": 0.489343523979187, + "learning_rate": 0.0001858382939693705, + "loss": 0.5619, + "step": 7323 + }, + { + "epoch": 0.1716860938717863, + "grad_norm": 0.886835515499115, + "learning_rate": 0.00018583451482767833, + "loss": 0.2006, + "step": 7324 + }, + { + "epoch": 0.1717095354465913, + "grad_norm": 0.2529012858867645, + "learning_rate": 0.0001858307352202457, + "loss": 0.0542, + "step": 7325 + }, + { + "epoch": 0.1717329770213963, + "grad_norm": 0.1835561841726303, + "learning_rate": 0.00018582695514709303, + "loss": 0.0567, + "step": 7326 + }, + { + "epoch": 0.1717564185962013, + "grad_norm": 0.5570096969604492, + "learning_rate": 0.00018582317460824092, + "loss": 0.77, + "step": 7327 + }, + { + "epoch": 0.1717798601710063, + "grad_norm": 0.4517579674720764, + "learning_rate": 0.0001858193936037098, + "loss": 0.1649, + "step": 7328 + }, + { + "epoch": 0.1718033017458113, + "grad_norm": 0.35719069838523865, + "learning_rate": 0.00018581561213352031, + "loss": 0.081, + "step": 7329 + }, + { + "epoch": 0.1718267433206163, + "grad_norm": 0.14655552804470062, + "learning_rate": 0.00018581183019769287, + "loss": 0.0207, + "step": 7330 + }, + { + "epoch": 0.17185018489542128, + "grad_norm": 0.6358156204223633, + "learning_rate": 0.000185808047796248, + "loss": 0.0893, + "step": 7331 + }, + { + "epoch": 0.17187362647022628, + "grad_norm": 0.5691938400268555, + "learning_rate": 0.00018580426492920628, + "loss": 0.1343, + "step": 7332 + }, + { + "epoch": 0.17189706804503127, + "grad_norm": 0.09468837827444077, + "learning_rate": 0.00018580048159658817, + "loss": 0.0234, + "step": 7333 + }, + { + "epoch": 0.17192050961983626, + "grad_norm": 0.6747286915779114, + "learning_rate": 0.0001857966977984143, + "loss": 0.1507, + "step": 7334 + }, + { + "epoch": 0.17194395119464126, + "grad_norm": 0.5801804065704346, + "learning_rate": 0.0001857929135347051, + "loss": 0.7759, + "step": 7335 + }, + { + "epoch": 0.17196739276944625, + "grad_norm": 0.7983790636062622, + "learning_rate": 0.00018578912880548113, + "loss": 0.1056, + "step": 7336 + }, + { + "epoch": 0.17199083434425125, + "grad_norm": 0.7310202717781067, + "learning_rate": 0.00018578534361076291, + "loss": 0.175, + "step": 7337 + }, + { + "epoch": 0.17201427591905624, + "grad_norm": 0.3808025121688843, + "learning_rate": 0.00018578155795057107, + "loss": 0.1155, + "step": 7338 + }, + { + "epoch": 0.17203771749386124, + "grad_norm": 0.18849250674247742, + "learning_rate": 0.00018577777182492603, + "loss": 0.0346, + "step": 7339 + }, + { + "epoch": 0.17206115906866623, + "grad_norm": 0.644521951675415, + "learning_rate": 0.0001857739852338484, + "loss": 0.1756, + "step": 7340 + }, + { + "epoch": 0.17208460064347122, + "grad_norm": 0.7498167753219604, + "learning_rate": 0.00018577019817735872, + "loss": 0.1907, + "step": 7341 + }, + { + "epoch": 0.17210804221827622, + "grad_norm": 0.47232550382614136, + "learning_rate": 0.00018576641065547756, + "loss": 0.1029, + "step": 7342 + }, + { + "epoch": 0.1721314837930812, + "grad_norm": 0.6847161650657654, + "learning_rate": 0.0001857626226682254, + "loss": 0.1945, + "step": 7343 + }, + { + "epoch": 0.1721549253678862, + "grad_norm": 0.3171011805534363, + "learning_rate": 0.00018575883421562286, + "loss": 0.0521, + "step": 7344 + }, + { + "epoch": 0.1721783669426912, + "grad_norm": 0.2934843599796295, + "learning_rate": 0.00018575504529769047, + "loss": 0.0566, + "step": 7345 + }, + { + "epoch": 0.1722018085174962, + "grad_norm": 0.5611146092414856, + "learning_rate": 0.0001857512559144488, + "loss": 0.1369, + "step": 7346 + }, + { + "epoch": 0.1722252500923012, + "grad_norm": 0.18908125162124634, + "learning_rate": 0.00018574746606591838, + "loss": 0.059, + "step": 7347 + }, + { + "epoch": 0.17224869166710619, + "grad_norm": 0.5386728048324585, + "learning_rate": 0.00018574367575211983, + "loss": 0.1855, + "step": 7348 + }, + { + "epoch": 0.17227213324191118, + "grad_norm": 0.6532708406448364, + "learning_rate": 0.00018573988497307367, + "loss": 0.2513, + "step": 7349 + }, + { + "epoch": 0.17229557481671617, + "grad_norm": 0.5220661163330078, + "learning_rate": 0.00018573609372880048, + "loss": 0.0687, + "step": 7350 + }, + { + "epoch": 0.1723190163915212, + "grad_norm": 0.6005072593688965, + "learning_rate": 0.00018573230201932082, + "loss": 0.1335, + "step": 7351 + }, + { + "epoch": 0.1723424579663262, + "grad_norm": 0.5453731417655945, + "learning_rate": 0.0001857285098446553, + "loss": 0.1629, + "step": 7352 + }, + { + "epoch": 0.17236589954113118, + "grad_norm": 0.27270206809043884, + "learning_rate": 0.00018572471720482446, + "loss": 0.0667, + "step": 7353 + }, + { + "epoch": 0.17238934111593618, + "grad_norm": 0.4606037735939026, + "learning_rate": 0.00018572092409984895, + "loss": 0.1796, + "step": 7354 + }, + { + "epoch": 0.17241278269074117, + "grad_norm": 0.1630886197090149, + "learning_rate": 0.00018571713052974924, + "loss": 0.038, + "step": 7355 + }, + { + "epoch": 0.17243622426554617, + "grad_norm": 0.46139273047447205, + "learning_rate": 0.000185713336494546, + "loss": 0.0775, + "step": 7356 + }, + { + "epoch": 0.17245966584035116, + "grad_norm": 0.2311343252658844, + "learning_rate": 0.00018570954199425974, + "loss": 0.0324, + "step": 7357 + }, + { + "epoch": 0.17248310741515616, + "grad_norm": 0.252017617225647, + "learning_rate": 0.0001857057470289111, + "loss": 0.0608, + "step": 7358 + }, + { + "epoch": 0.17250654898996115, + "grad_norm": 0.5804604291915894, + "learning_rate": 0.00018570195159852072, + "loss": 0.1398, + "step": 7359 + }, + { + "epoch": 0.17252999056476614, + "grad_norm": 0.723820149898529, + "learning_rate": 0.00018569815570310908, + "loss": 0.2008, + "step": 7360 + }, + { + "epoch": 0.17255343213957114, + "grad_norm": 0.1477431356906891, + "learning_rate": 0.00018569435934269688, + "loss": 0.0166, + "step": 7361 + }, + { + "epoch": 0.17257687371437613, + "grad_norm": 0.2824326157569885, + "learning_rate": 0.00018569056251730465, + "loss": 0.055, + "step": 7362 + }, + { + "epoch": 0.17260031528918113, + "grad_norm": 0.4646925628185272, + "learning_rate": 0.00018568676522695305, + "loss": 0.1659, + "step": 7363 + }, + { + "epoch": 0.17262375686398612, + "grad_norm": 0.3437279462814331, + "learning_rate": 0.00018568296747166265, + "loss": 0.039, + "step": 7364 + }, + { + "epoch": 0.17264719843879112, + "grad_norm": 0.8291356563568115, + "learning_rate": 0.00018567916925145403, + "loss": 0.2435, + "step": 7365 + }, + { + "epoch": 0.1726706400135961, + "grad_norm": 0.4403074085712433, + "learning_rate": 0.00018567537056634782, + "loss": 0.1157, + "step": 7366 + }, + { + "epoch": 0.1726940815884011, + "grad_norm": 0.3968246877193451, + "learning_rate": 0.00018567157141636468, + "loss": 0.0967, + "step": 7367 + }, + { + "epoch": 0.1727175231632061, + "grad_norm": 0.42724609375, + "learning_rate": 0.00018566777180152518, + "loss": 0.1464, + "step": 7368 + }, + { + "epoch": 0.1727409647380111, + "grad_norm": 0.680668294429779, + "learning_rate": 0.00018566397172184993, + "loss": 0.6667, + "step": 7369 + }, + { + "epoch": 0.1727644063128161, + "grad_norm": 0.5778806209564209, + "learning_rate": 0.00018566017117735957, + "loss": 0.1786, + "step": 7370 + }, + { + "epoch": 0.17278784788762108, + "grad_norm": 0.39064034819602966, + "learning_rate": 0.00018565637016807469, + "loss": 0.1434, + "step": 7371 + }, + { + "epoch": 0.17281128946242608, + "grad_norm": 0.46267250180244446, + "learning_rate": 0.00018565256869401594, + "loss": 0.1218, + "step": 7372 + }, + { + "epoch": 0.17283473103723107, + "grad_norm": 0.4673093855381012, + "learning_rate": 0.00018564876675520396, + "loss": 0.1124, + "step": 7373 + }, + { + "epoch": 0.17285817261203607, + "grad_norm": 0.4714944660663605, + "learning_rate": 0.00018564496435165935, + "loss": 0.1129, + "step": 7374 + }, + { + "epoch": 0.17288161418684106, + "grad_norm": 0.38210928440093994, + "learning_rate": 0.00018564116148340277, + "loss": 0.0668, + "step": 7375 + }, + { + "epoch": 0.17290505576164605, + "grad_norm": 0.23784151673316956, + "learning_rate": 0.00018563735815045485, + "loss": 0.0579, + "step": 7376 + }, + { + "epoch": 0.17292849733645108, + "grad_norm": 0.8704209923744202, + "learning_rate": 0.0001856335543528362, + "loss": 0.1276, + "step": 7377 + }, + { + "epoch": 0.17295193891125607, + "grad_norm": 0.3197523355484009, + "learning_rate": 0.00018562975009056747, + "loss": 0.1022, + "step": 7378 + }, + { + "epoch": 0.17297538048606106, + "grad_norm": 0.612673819065094, + "learning_rate": 0.00018562594536366932, + "loss": 0.0934, + "step": 7379 + }, + { + "epoch": 0.17299882206086606, + "grad_norm": 0.9181567430496216, + "learning_rate": 0.00018562214017216238, + "loss": 0.2086, + "step": 7380 + }, + { + "epoch": 0.17302226363567105, + "grad_norm": 0.1864294856786728, + "learning_rate": 0.00018561833451606727, + "loss": 0.0264, + "step": 7381 + }, + { + "epoch": 0.17304570521047605, + "grad_norm": 0.6481404900550842, + "learning_rate": 0.00018561452839540473, + "loss": 0.1174, + "step": 7382 + }, + { + "epoch": 0.17306914678528104, + "grad_norm": 0.6923366785049438, + "learning_rate": 0.00018561072181019528, + "loss": 0.1897, + "step": 7383 + }, + { + "epoch": 0.17309258836008604, + "grad_norm": 0.21441826224327087, + "learning_rate": 0.00018560691476045968, + "loss": 0.0337, + "step": 7384 + }, + { + "epoch": 0.17311602993489103, + "grad_norm": 0.5199514627456665, + "learning_rate": 0.00018560310724621858, + "loss": 0.6544, + "step": 7385 + }, + { + "epoch": 0.17313947150969602, + "grad_norm": 0.6379154324531555, + "learning_rate": 0.0001855992992674926, + "loss": 0.1872, + "step": 7386 + }, + { + "epoch": 0.17316291308450102, + "grad_norm": 0.18929800391197205, + "learning_rate": 0.00018559549082430239, + "loss": 0.042, + "step": 7387 + }, + { + "epoch": 0.173186354659306, + "grad_norm": 0.5321101546287537, + "learning_rate": 0.00018559168191666867, + "loss": 0.1122, + "step": 7388 + }, + { + "epoch": 0.173209796234111, + "grad_norm": 0.674210786819458, + "learning_rate": 0.00018558787254461205, + "loss": 0.5512, + "step": 7389 + }, + { + "epoch": 0.173233237808916, + "grad_norm": 0.41194695234298706, + "learning_rate": 0.00018558406270815324, + "loss": 0.1001, + "step": 7390 + }, + { + "epoch": 0.173256679383721, + "grad_norm": 0.22949600219726562, + "learning_rate": 0.00018558025240731293, + "loss": 0.0327, + "step": 7391 + }, + { + "epoch": 0.173280120958526, + "grad_norm": 0.5888392925262451, + "learning_rate": 0.0001855764416421117, + "loss": 0.1283, + "step": 7392 + }, + { + "epoch": 0.17330356253333098, + "grad_norm": 0.6089851260185242, + "learning_rate": 0.00018557263041257032, + "loss": 0.1373, + "step": 7393 + }, + { + "epoch": 0.17332700410813598, + "grad_norm": 0.2204003632068634, + "learning_rate": 0.00018556881871870947, + "loss": 0.0504, + "step": 7394 + }, + { + "epoch": 0.17335044568294097, + "grad_norm": 0.1845255047082901, + "learning_rate": 0.00018556500656054977, + "loss": 0.0507, + "step": 7395 + }, + { + "epoch": 0.17337388725774597, + "grad_norm": 0.34240445494651794, + "learning_rate": 0.00018556119393811193, + "loss": 0.0859, + "step": 7396 + }, + { + "epoch": 0.17339732883255096, + "grad_norm": 0.6097816824913025, + "learning_rate": 0.0001855573808514167, + "loss": 0.506, + "step": 7397 + }, + { + "epoch": 0.17342077040735596, + "grad_norm": 0.34360936284065247, + "learning_rate": 0.00018555356730048464, + "loss": 0.1079, + "step": 7398 + }, + { + "epoch": 0.17344421198216095, + "grad_norm": 0.20077458024024963, + "learning_rate": 0.00018554975328533658, + "loss": 0.0242, + "step": 7399 + }, + { + "epoch": 0.17346765355696595, + "grad_norm": 0.6183021068572998, + "learning_rate": 0.00018554593880599314, + "loss": 0.2253, + "step": 7400 + }, + { + "epoch": 0.17349109513177094, + "grad_norm": 0.6016367077827454, + "learning_rate": 0.000185542123862475, + "loss": 0.2178, + "step": 7401 + }, + { + "epoch": 0.17351453670657596, + "grad_norm": 0.622941255569458, + "learning_rate": 0.00018553830845480294, + "loss": 0.7967, + "step": 7402 + }, + { + "epoch": 0.17353797828138096, + "grad_norm": 0.21477705240249634, + "learning_rate": 0.0001855344925829976, + "loss": 0.0583, + "step": 7403 + }, + { + "epoch": 0.17356141985618595, + "grad_norm": 0.5210747122764587, + "learning_rate": 0.00018553067624707965, + "loss": 0.1506, + "step": 7404 + }, + { + "epoch": 0.17358486143099094, + "grad_norm": 1.0410789251327515, + "learning_rate": 0.0001855268594470699, + "loss": 0.1874, + "step": 7405 + }, + { + "epoch": 0.17360830300579594, + "grad_norm": 0.44314658641815186, + "learning_rate": 0.00018552304218298897, + "loss": 0.1586, + "step": 7406 + }, + { + "epoch": 0.17363174458060093, + "grad_norm": 0.4737854599952698, + "learning_rate": 0.00018551922445485762, + "loss": 0.0677, + "step": 7407 + }, + { + "epoch": 0.17365518615540593, + "grad_norm": 0.46306300163269043, + "learning_rate": 0.00018551540626269655, + "loss": 0.1215, + "step": 7408 + }, + { + "epoch": 0.17367862773021092, + "grad_norm": 0.6485884785652161, + "learning_rate": 0.0001855115876065265, + "loss": 0.2098, + "step": 7409 + }, + { + "epoch": 0.17370206930501592, + "grad_norm": 0.23744039237499237, + "learning_rate": 0.00018550776848636814, + "loss": 0.0409, + "step": 7410 + }, + { + "epoch": 0.1737255108798209, + "grad_norm": 0.6346874237060547, + "learning_rate": 0.00018550394890224227, + "loss": 0.6946, + "step": 7411 + }, + { + "epoch": 0.1737489524546259, + "grad_norm": 0.47637948393821716, + "learning_rate": 0.00018550012885416953, + "loss": 0.1163, + "step": 7412 + }, + { + "epoch": 0.1737723940294309, + "grad_norm": 0.5875145792961121, + "learning_rate": 0.0001854963083421707, + "loss": 0.1231, + "step": 7413 + }, + { + "epoch": 0.1737958356042359, + "grad_norm": 0.35830751061439514, + "learning_rate": 0.0001854924873662665, + "loss": 0.0523, + "step": 7414 + }, + { + "epoch": 0.1738192771790409, + "grad_norm": 0.2821299433708191, + "learning_rate": 0.00018548866592647765, + "loss": 0.0384, + "step": 7415 + }, + { + "epoch": 0.17384271875384588, + "grad_norm": 0.7035308480262756, + "learning_rate": 0.0001854848440228249, + "loss": 0.2237, + "step": 7416 + }, + { + "epoch": 0.17386616032865088, + "grad_norm": 0.5324121117591858, + "learning_rate": 0.00018548102165532901, + "loss": 0.2319, + "step": 7417 + }, + { + "epoch": 0.17388960190345587, + "grad_norm": 0.28633618354797363, + "learning_rate": 0.00018547719882401067, + "loss": 0.0736, + "step": 7418 + }, + { + "epoch": 0.17391304347826086, + "grad_norm": 0.19425560534000397, + "learning_rate": 0.00018547337552889064, + "loss": 0.0485, + "step": 7419 + }, + { + "epoch": 0.17393648505306586, + "grad_norm": 0.26148632168769836, + "learning_rate": 0.00018546955176998968, + "loss": 0.0466, + "step": 7420 + }, + { + "epoch": 0.17395992662787085, + "grad_norm": 0.4516453146934509, + "learning_rate": 0.0001854657275473285, + "loss": 0.0707, + "step": 7421 + }, + { + "epoch": 0.17398336820267585, + "grad_norm": 0.6151387691497803, + "learning_rate": 0.0001854619028609279, + "loss": 0.185, + "step": 7422 + }, + { + "epoch": 0.17400680977748084, + "grad_norm": 0.7180807590484619, + "learning_rate": 0.00018545807771080863, + "loss": 0.1916, + "step": 7423 + }, + { + "epoch": 0.17403025135228584, + "grad_norm": 0.487381249666214, + "learning_rate": 0.00018545425209699144, + "loss": 0.0797, + "step": 7424 + }, + { + "epoch": 0.17405369292709083, + "grad_norm": 0.6150439977645874, + "learning_rate": 0.00018545042601949703, + "loss": 0.4742, + "step": 7425 + }, + { + "epoch": 0.17407713450189582, + "grad_norm": 0.48248884081840515, + "learning_rate": 0.00018544659947834622, + "loss": 0.1297, + "step": 7426 + }, + { + "epoch": 0.17410057607670082, + "grad_norm": 0.2887405455112457, + "learning_rate": 0.00018544277247355977, + "loss": 0.12, + "step": 7427 + }, + { + "epoch": 0.17412401765150584, + "grad_norm": 0.3301381468772888, + "learning_rate": 0.00018543894500515842, + "loss": 0.0791, + "step": 7428 + }, + { + "epoch": 0.17414745922631084, + "grad_norm": 0.606726348400116, + "learning_rate": 0.00018543511707316297, + "loss": 0.1385, + "step": 7429 + }, + { + "epoch": 0.17417090080111583, + "grad_norm": 0.538982629776001, + "learning_rate": 0.00018543128867759415, + "loss": 0.1016, + "step": 7430 + }, + { + "epoch": 0.17419434237592082, + "grad_norm": 0.7999199628829956, + "learning_rate": 0.00018542745981847277, + "loss": 0.1918, + "step": 7431 + }, + { + "epoch": 0.17421778395072582, + "grad_norm": 0.5047276020050049, + "learning_rate": 0.00018542363049581958, + "loss": 0.0595, + "step": 7432 + }, + { + "epoch": 0.1742412255255308, + "grad_norm": 0.43873536586761475, + "learning_rate": 0.00018541980070965538, + "loss": 0.0932, + "step": 7433 + }, + { + "epoch": 0.1742646671003358, + "grad_norm": 0.40905240178108215, + "learning_rate": 0.00018541597046000093, + "loss": 0.597, + "step": 7434 + }, + { + "epoch": 0.1742881086751408, + "grad_norm": 0.2692873179912567, + "learning_rate": 0.00018541213974687703, + "loss": 0.0636, + "step": 7435 + }, + { + "epoch": 0.1743115502499458, + "grad_norm": 0.3982955515384674, + "learning_rate": 0.00018540830857030445, + "loss": 0.0795, + "step": 7436 + }, + { + "epoch": 0.1743349918247508, + "grad_norm": 0.7269759178161621, + "learning_rate": 0.000185404476930304, + "loss": 0.7377, + "step": 7437 + }, + { + "epoch": 0.17435843339955578, + "grad_norm": 0.47275006771087646, + "learning_rate": 0.00018540064482689645, + "loss": 0.0788, + "step": 7438 + }, + { + "epoch": 0.17438187497436078, + "grad_norm": 0.409854531288147, + "learning_rate": 0.0001853968122601026, + "loss": 0.1122, + "step": 7439 + }, + { + "epoch": 0.17440531654916577, + "grad_norm": 0.2151332050561905, + "learning_rate": 0.00018539297922994324, + "loss": 0.0368, + "step": 7440 + }, + { + "epoch": 0.17442875812397077, + "grad_norm": 0.14236538112163544, + "learning_rate": 0.00018538914573643916, + "loss": 0.0231, + "step": 7441 + }, + { + "epoch": 0.17445219969877576, + "grad_norm": 0.3973069489002228, + "learning_rate": 0.00018538531177961123, + "loss": 0.101, + "step": 7442 + }, + { + "epoch": 0.17447564127358076, + "grad_norm": 0.5776017308235168, + "learning_rate": 0.00018538147735948016, + "loss": 0.1564, + "step": 7443 + }, + { + "epoch": 0.17449908284838575, + "grad_norm": 0.6452957987785339, + "learning_rate": 0.00018537764247606678, + "loss": 0.1087, + "step": 7444 + }, + { + "epoch": 0.17452252442319074, + "grad_norm": 1.0034732818603516, + "learning_rate": 0.00018537380712939193, + "loss": 0.1422, + "step": 7445 + }, + { + "epoch": 0.17454596599799574, + "grad_norm": 0.588900625705719, + "learning_rate": 0.0001853699713194764, + "loss": 0.0908, + "step": 7446 + }, + { + "epoch": 0.17456940757280073, + "grad_norm": 0.478998601436615, + "learning_rate": 0.00018536613504634099, + "loss": 0.1059, + "step": 7447 + }, + { + "epoch": 0.17459284914760573, + "grad_norm": 0.7172098755836487, + "learning_rate": 0.00018536229831000656, + "loss": 0.4128, + "step": 7448 + }, + { + "epoch": 0.17461629072241072, + "grad_norm": 0.20677687227725983, + "learning_rate": 0.0001853584611104939, + "loss": 0.0391, + "step": 7449 + }, + { + "epoch": 0.17463973229721572, + "grad_norm": 0.5595411062240601, + "learning_rate": 0.0001853546234478238, + "loss": 0.1016, + "step": 7450 + }, + { + "epoch": 0.1746631738720207, + "grad_norm": 0.44000014662742615, + "learning_rate": 0.0001853507853220171, + "loss": 0.4032, + "step": 7451 + }, + { + "epoch": 0.1746866154468257, + "grad_norm": 0.6394870281219482, + "learning_rate": 0.00018534694673309467, + "loss": 0.1506, + "step": 7452 + }, + { + "epoch": 0.17471005702163073, + "grad_norm": 0.6014655828475952, + "learning_rate": 0.0001853431076810773, + "loss": 0.2091, + "step": 7453 + }, + { + "epoch": 0.17473349859643572, + "grad_norm": 0.7039527297019958, + "learning_rate": 0.00018533926816598583, + "loss": 0.155, + "step": 7454 + }, + { + "epoch": 0.17475694017124072, + "grad_norm": 0.6948651075363159, + "learning_rate": 0.0001853354281878411, + "loss": 0.1769, + "step": 7455 + }, + { + "epoch": 0.1747803817460457, + "grad_norm": 0.6528434753417969, + "learning_rate": 0.00018533158774666392, + "loss": 0.2177, + "step": 7456 + }, + { + "epoch": 0.1748038233208507, + "grad_norm": 0.5681542754173279, + "learning_rate": 0.00018532774684247517, + "loss": 0.0955, + "step": 7457 + }, + { + "epoch": 0.1748272648956557, + "grad_norm": 0.5792772769927979, + "learning_rate": 0.00018532390547529564, + "loss": 0.1762, + "step": 7458 + }, + { + "epoch": 0.1748507064704607, + "grad_norm": 0.43051677942276, + "learning_rate": 0.0001853200636451462, + "loss": 0.0925, + "step": 7459 + }, + { + "epoch": 0.1748741480452657, + "grad_norm": 0.3307151198387146, + "learning_rate": 0.0001853162213520477, + "loss": 0.0642, + "step": 7460 + }, + { + "epoch": 0.17489758962007068, + "grad_norm": 0.29242366552352905, + "learning_rate": 0.00018531237859602102, + "loss": 0.0658, + "step": 7461 + }, + { + "epoch": 0.17492103119487568, + "grad_norm": 0.3808744251728058, + "learning_rate": 0.00018530853537708692, + "loss": 0.0704, + "step": 7462 + }, + { + "epoch": 0.17494447276968067, + "grad_norm": 0.36315619945526123, + "learning_rate": 0.00018530469169526637, + "loss": 0.1088, + "step": 7463 + }, + { + "epoch": 0.17496791434448566, + "grad_norm": 0.12063944339752197, + "learning_rate": 0.0001853008475505801, + "loss": 0.0244, + "step": 7464 + }, + { + "epoch": 0.17499135591929066, + "grad_norm": 0.9017153382301331, + "learning_rate": 0.00018529700294304906, + "loss": 0.1908, + "step": 7465 + }, + { + "epoch": 0.17501479749409565, + "grad_norm": 0.6141842007637024, + "learning_rate": 0.00018529315787269411, + "loss": 0.2254, + "step": 7466 + }, + { + "epoch": 0.17503823906890065, + "grad_norm": 0.2662367820739746, + "learning_rate": 0.00018528931233953606, + "loss": 0.0742, + "step": 7467 + }, + { + "epoch": 0.17506168064370564, + "grad_norm": 0.5125517845153809, + "learning_rate": 0.0001852854663435958, + "loss": 0.086, + "step": 7468 + }, + { + "epoch": 0.17508512221851064, + "grad_norm": 0.8219767212867737, + "learning_rate": 0.0001852816198848942, + "loss": 0.8077, + "step": 7469 + }, + { + "epoch": 0.17510856379331563, + "grad_norm": 0.2490508109331131, + "learning_rate": 0.00018527777296345214, + "loss": 0.0491, + "step": 7470 + }, + { + "epoch": 0.17513200536812062, + "grad_norm": 0.3591480255126953, + "learning_rate": 0.0001852739255792905, + "loss": 0.0968, + "step": 7471 + }, + { + "epoch": 0.17515544694292562, + "grad_norm": 0.4669296443462372, + "learning_rate": 0.0001852700777324301, + "loss": 0.1365, + "step": 7472 + }, + { + "epoch": 0.1751788885177306, + "grad_norm": 0.30692747235298157, + "learning_rate": 0.0001852662294228919, + "loss": 0.1063, + "step": 7473 + }, + { + "epoch": 0.1752023300925356, + "grad_norm": 0.15450452268123627, + "learning_rate": 0.00018526238065069673, + "loss": 0.0276, + "step": 7474 + }, + { + "epoch": 0.1752257716673406, + "grad_norm": 0.4646235406398773, + "learning_rate": 0.0001852585314158655, + "loss": 0.1715, + "step": 7475 + }, + { + "epoch": 0.1752492132421456, + "grad_norm": 1.0773050785064697, + "learning_rate": 0.00018525468171841903, + "loss": 0.1136, + "step": 7476 + }, + { + "epoch": 0.1752726548169506, + "grad_norm": 0.32597339153289795, + "learning_rate": 0.0001852508315583783, + "loss": 0.0661, + "step": 7477 + }, + { + "epoch": 0.17529609639175558, + "grad_norm": 0.4549053907394409, + "learning_rate": 0.00018524698093576417, + "loss": 0.1334, + "step": 7478 + }, + { + "epoch": 0.1753195379665606, + "grad_norm": 0.22185483574867249, + "learning_rate": 0.0001852431298505975, + "loss": 0.0725, + "step": 7479 + }, + { + "epoch": 0.1753429795413656, + "grad_norm": 0.6735120415687561, + "learning_rate": 0.00018523927830289925, + "loss": 0.7598, + "step": 7480 + }, + { + "epoch": 0.1753664211161706, + "grad_norm": 1.084596872329712, + "learning_rate": 0.00018523542629269023, + "loss": 0.0867, + "step": 7481 + }, + { + "epoch": 0.1753898626909756, + "grad_norm": 0.41799819469451904, + "learning_rate": 0.00018523157381999142, + "loss": 0.0492, + "step": 7482 + }, + { + "epoch": 0.17541330426578058, + "grad_norm": 0.4388532340526581, + "learning_rate": 0.0001852277208848237, + "loss": 0.1289, + "step": 7483 + }, + { + "epoch": 0.17543674584058558, + "grad_norm": 0.38500791788101196, + "learning_rate": 0.00018522386748720795, + "loss": 0.066, + "step": 7484 + }, + { + "epoch": 0.17546018741539057, + "grad_norm": 0.6394324898719788, + "learning_rate": 0.0001852200136271651, + "loss": 0.1698, + "step": 7485 + }, + { + "epoch": 0.17548362899019557, + "grad_norm": 0.1879613697528839, + "learning_rate": 0.00018521615930471605, + "loss": 0.0628, + "step": 7486 + }, + { + "epoch": 0.17550707056500056, + "grad_norm": 0.5508020520210266, + "learning_rate": 0.00018521230451988175, + "loss": 0.5716, + "step": 7487 + }, + { + "epoch": 0.17553051213980556, + "grad_norm": 0.4604242444038391, + "learning_rate": 0.00018520844927268307, + "loss": 0.1594, + "step": 7488 + }, + { + "epoch": 0.17555395371461055, + "grad_norm": 0.1816866397857666, + "learning_rate": 0.00018520459356314095, + "loss": 0.0287, + "step": 7489 + }, + { + "epoch": 0.17557739528941554, + "grad_norm": 0.4927252531051636, + "learning_rate": 0.0001852007373912763, + "loss": 0.0927, + "step": 7490 + }, + { + "epoch": 0.17560083686422054, + "grad_norm": 0.289246141910553, + "learning_rate": 0.00018519688075711011, + "loss": 0.0692, + "step": 7491 + }, + { + "epoch": 0.17562427843902553, + "grad_norm": 0.2110338807106018, + "learning_rate": 0.0001851930236606632, + "loss": 0.0173, + "step": 7492 + }, + { + "epoch": 0.17564772001383053, + "grad_norm": 0.826214075088501, + "learning_rate": 0.00018518916610195654, + "loss": 0.1463, + "step": 7493 + }, + { + "epoch": 0.17567116158863552, + "grad_norm": 0.4221503436565399, + "learning_rate": 0.0001851853080810111, + "loss": 0.1235, + "step": 7494 + }, + { + "epoch": 0.17569460316344052, + "grad_norm": 0.2359515130519867, + "learning_rate": 0.00018518144959784775, + "loss": 0.0435, + "step": 7495 + }, + { + "epoch": 0.1757180447382455, + "grad_norm": 0.18556098639965057, + "learning_rate": 0.00018517759065248747, + "loss": 0.0405, + "step": 7496 + }, + { + "epoch": 0.1757414863130505, + "grad_norm": 0.2868919372558594, + "learning_rate": 0.0001851737312449512, + "loss": 0.0821, + "step": 7497 + }, + { + "epoch": 0.1757649278878555, + "grad_norm": 0.6836004257202148, + "learning_rate": 0.00018516987137525987, + "loss": 0.1642, + "step": 7498 + }, + { + "epoch": 0.1757883694626605, + "grad_norm": 0.44125989079475403, + "learning_rate": 0.00018516601104343443, + "loss": 0.0901, + "step": 7499 + }, + { + "epoch": 0.1758118110374655, + "grad_norm": 0.2837117314338684, + "learning_rate": 0.0001851621502494958, + "loss": 0.0813, + "step": 7500 + }, + { + "epoch": 0.17583525261227048, + "grad_norm": 0.27462995052337646, + "learning_rate": 0.00018515828899346497, + "loss": 0.0985, + "step": 7501 + }, + { + "epoch": 0.17585869418707548, + "grad_norm": 0.1339501142501831, + "learning_rate": 0.00018515442727536286, + "loss": 0.02, + "step": 7502 + }, + { + "epoch": 0.17588213576188047, + "grad_norm": 0.12865066528320312, + "learning_rate": 0.00018515056509521043, + "loss": 0.0147, + "step": 7503 + }, + { + "epoch": 0.17590557733668546, + "grad_norm": 0.4798547923564911, + "learning_rate": 0.00018514670245302863, + "loss": 0.0863, + "step": 7504 + }, + { + "epoch": 0.1759290189114905, + "grad_norm": 0.22494272887706757, + "learning_rate": 0.00018514283934883844, + "loss": 0.0676, + "step": 7505 + }, + { + "epoch": 0.17595246048629548, + "grad_norm": 0.7478399872779846, + "learning_rate": 0.00018513897578266078, + "loss": 0.2428, + "step": 7506 + }, + { + "epoch": 0.17597590206110048, + "grad_norm": 0.34832635521888733, + "learning_rate": 0.00018513511175451667, + "loss": 0.0477, + "step": 7507 + }, + { + "epoch": 0.17599934363590547, + "grad_norm": 0.7811594605445862, + "learning_rate": 0.00018513124726442706, + "loss": 0.6951, + "step": 7508 + }, + { + "epoch": 0.17602278521071046, + "grad_norm": 0.40656179189682007, + "learning_rate": 0.00018512738231241292, + "loss": 0.076, + "step": 7509 + }, + { + "epoch": 0.17604622678551546, + "grad_norm": 0.44303131103515625, + "learning_rate": 0.00018512351689849518, + "loss": 0.1087, + "step": 7510 + }, + { + "epoch": 0.17606966836032045, + "grad_norm": 0.7945590019226074, + "learning_rate": 0.00018511965102269487, + "loss": 0.2366, + "step": 7511 + }, + { + "epoch": 0.17609310993512545, + "grad_norm": 0.6936621069908142, + "learning_rate": 0.0001851157846850329, + "loss": 0.2771, + "step": 7512 + }, + { + "epoch": 0.17611655150993044, + "grad_norm": 0.9812513589859009, + "learning_rate": 0.00018511191788553032, + "loss": 0.1521, + "step": 7513 + }, + { + "epoch": 0.17613999308473544, + "grad_norm": 0.3125937581062317, + "learning_rate": 0.00018510805062420805, + "loss": 0.1416, + "step": 7514 + }, + { + "epoch": 0.17616343465954043, + "grad_norm": 0.13797084987163544, + "learning_rate": 0.00018510418290108716, + "loss": 0.0383, + "step": 7515 + }, + { + "epoch": 0.17618687623434542, + "grad_norm": 0.24209623038768768, + "learning_rate": 0.00018510031471618855, + "loss": 0.0718, + "step": 7516 + }, + { + "epoch": 0.17621031780915042, + "grad_norm": 0.4499163031578064, + "learning_rate": 0.0001850964460695332, + "loss": 0.0848, + "step": 7517 + }, + { + "epoch": 0.1762337593839554, + "grad_norm": 0.1356833279132843, + "learning_rate": 0.0001850925769611422, + "loss": 0.0401, + "step": 7518 + }, + { + "epoch": 0.1762572009587604, + "grad_norm": 0.34450918436050415, + "learning_rate": 0.00018508870739103645, + "loss": 0.0522, + "step": 7519 + }, + { + "epoch": 0.1762806425335654, + "grad_norm": 0.5087942481040955, + "learning_rate": 0.000185084837359237, + "loss": 0.0492, + "step": 7520 + }, + { + "epoch": 0.1763040841083704, + "grad_norm": 0.7908738851547241, + "learning_rate": 0.00018508096686576484, + "loss": 0.1987, + "step": 7521 + }, + { + "epoch": 0.1763275256831754, + "grad_norm": 0.40032631158828735, + "learning_rate": 0.00018507709591064094, + "loss": 0.4764, + "step": 7522 + }, + { + "epoch": 0.17635096725798038, + "grad_norm": 0.3344307541847229, + "learning_rate": 0.00018507322449388633, + "loss": 0.0772, + "step": 7523 + }, + { + "epoch": 0.17637440883278538, + "grad_norm": 0.5520614981651306, + "learning_rate": 0.000185069352615522, + "loss": 0.172, + "step": 7524 + }, + { + "epoch": 0.17639785040759037, + "grad_norm": 0.3004385530948639, + "learning_rate": 0.00018506548027556898, + "loss": 0.0462, + "step": 7525 + }, + { + "epoch": 0.17642129198239537, + "grad_norm": 0.35226312279701233, + "learning_rate": 0.0001850616074740483, + "loss": 0.0962, + "step": 7526 + }, + { + "epoch": 0.17644473355720036, + "grad_norm": 0.6362230777740479, + "learning_rate": 0.00018505773421098089, + "loss": 0.1635, + "step": 7527 + }, + { + "epoch": 0.17646817513200536, + "grad_norm": 0.6944090723991394, + "learning_rate": 0.0001850538604863879, + "loss": 0.187, + "step": 7528 + }, + { + "epoch": 0.17649161670681035, + "grad_norm": 0.3664611876010895, + "learning_rate": 0.0001850499863002902, + "loss": 0.5496, + "step": 7529 + }, + { + "epoch": 0.17651505828161537, + "grad_norm": 0.6397160887718201, + "learning_rate": 0.0001850461116527089, + "loss": 0.1583, + "step": 7530 + }, + { + "epoch": 0.17653849985642037, + "grad_norm": 0.6467212438583374, + "learning_rate": 0.00018504223654366503, + "loss": 0.1631, + "step": 7531 + }, + { + "epoch": 0.17656194143122536, + "grad_norm": 0.6485815644264221, + "learning_rate": 0.00018503836097317957, + "loss": 0.1869, + "step": 7532 + }, + { + "epoch": 0.17658538300603036, + "grad_norm": 0.5884918570518494, + "learning_rate": 0.00018503448494127357, + "loss": 0.1875, + "step": 7533 + }, + { + "epoch": 0.17660882458083535, + "grad_norm": 0.49926453828811646, + "learning_rate": 0.0001850306084479681, + "loss": 0.4766, + "step": 7534 + }, + { + "epoch": 0.17663226615564034, + "grad_norm": 0.5392001271247864, + "learning_rate": 0.00018502673149328413, + "loss": 0.169, + "step": 7535 + }, + { + "epoch": 0.17665570773044534, + "grad_norm": 0.3961862027645111, + "learning_rate": 0.0001850228540772427, + "loss": 0.0377, + "step": 7536 + }, + { + "epoch": 0.17667914930525033, + "grad_norm": 0.3830231726169586, + "learning_rate": 0.00018501897619986493, + "loss": 0.0743, + "step": 7537 + }, + { + "epoch": 0.17670259088005533, + "grad_norm": 0.7005025148391724, + "learning_rate": 0.00018501509786117177, + "loss": 0.2125, + "step": 7538 + }, + { + "epoch": 0.17672603245486032, + "grad_norm": 0.42352449893951416, + "learning_rate": 0.0001850112190611843, + "loss": 0.074, + "step": 7539 + }, + { + "epoch": 0.17674947402966532, + "grad_norm": 0.3236844539642334, + "learning_rate": 0.00018500733979992355, + "loss": 0.0774, + "step": 7540 + }, + { + "epoch": 0.1767729156044703, + "grad_norm": 0.16473540663719177, + "learning_rate": 0.0001850034600774106, + "loss": 0.0234, + "step": 7541 + }, + { + "epoch": 0.1767963571792753, + "grad_norm": 0.44723206758499146, + "learning_rate": 0.00018499957989366651, + "loss": 0.1297, + "step": 7542 + }, + { + "epoch": 0.1768197987540803, + "grad_norm": 0.4710826873779297, + "learning_rate": 0.0001849956992487123, + "loss": 0.1165, + "step": 7543 + }, + { + "epoch": 0.1768432403288853, + "grad_norm": 0.5699921250343323, + "learning_rate": 0.000184991818142569, + "loss": 0.1055, + "step": 7544 + }, + { + "epoch": 0.1768666819036903, + "grad_norm": 0.6941460371017456, + "learning_rate": 0.00018498793657525773, + "loss": 0.1654, + "step": 7545 + }, + { + "epoch": 0.17689012347849528, + "grad_norm": 0.2003440260887146, + "learning_rate": 0.00018498405454679952, + "loss": 0.0406, + "step": 7546 + }, + { + "epoch": 0.17691356505330028, + "grad_norm": 0.5147202014923096, + "learning_rate": 0.00018498017205721546, + "loss": 0.0661, + "step": 7547 + }, + { + "epoch": 0.17693700662810527, + "grad_norm": 0.8713290095329285, + "learning_rate": 0.0001849762891065266, + "loss": 0.2462, + "step": 7548 + }, + { + "epoch": 0.17696044820291026, + "grad_norm": 0.1849461793899536, + "learning_rate": 0.00018497240569475399, + "loss": 0.0528, + "step": 7549 + }, + { + "epoch": 0.17698388977771526, + "grad_norm": 0.3792017996311188, + "learning_rate": 0.0001849685218219187, + "loss": 0.1021, + "step": 7550 + }, + { + "epoch": 0.17700733135252025, + "grad_norm": 1.0962854623794556, + "learning_rate": 0.00018496463748804182, + "loss": 0.1519, + "step": 7551 + }, + { + "epoch": 0.17703077292732525, + "grad_norm": 0.21504813432693481, + "learning_rate": 0.00018496075269314446, + "loss": 0.0549, + "step": 7552 + }, + { + "epoch": 0.17705421450213024, + "grad_norm": 0.6837963461875916, + "learning_rate": 0.00018495686743724764, + "loss": 0.1592, + "step": 7553 + }, + { + "epoch": 0.17707765607693524, + "grad_norm": 0.5671080946922302, + "learning_rate": 0.00018495298172037248, + "loss": 0.1259, + "step": 7554 + }, + { + "epoch": 0.17710109765174023, + "grad_norm": 0.4957760274410248, + "learning_rate": 0.00018494909554254006, + "loss": 0.1358, + "step": 7555 + }, + { + "epoch": 0.17712453922654525, + "grad_norm": 0.5186753869056702, + "learning_rate": 0.00018494520890377145, + "loss": 0.1241, + "step": 7556 + }, + { + "epoch": 0.17714798080135025, + "grad_norm": 0.4433806240558624, + "learning_rate": 0.00018494132180408776, + "loss": 0.1036, + "step": 7557 + }, + { + "epoch": 0.17717142237615524, + "grad_norm": 0.5061827898025513, + "learning_rate": 0.00018493743424351005, + "loss": 0.0833, + "step": 7558 + }, + { + "epoch": 0.17719486395096024, + "grad_norm": 0.71573805809021, + "learning_rate": 0.00018493354622205944, + "loss": 0.1324, + "step": 7559 + }, + { + "epoch": 0.17721830552576523, + "grad_norm": 0.8219785094261169, + "learning_rate": 0.00018492965773975705, + "loss": 0.2088, + "step": 7560 + }, + { + "epoch": 0.17724174710057022, + "grad_norm": 0.8399453163146973, + "learning_rate": 0.00018492576879662392, + "loss": 0.1027, + "step": 7561 + }, + { + "epoch": 0.17726518867537522, + "grad_norm": 0.6020707488059998, + "learning_rate": 0.00018492187939268115, + "loss": 0.0976, + "step": 7562 + }, + { + "epoch": 0.1772886302501802, + "grad_norm": 0.38853999972343445, + "learning_rate": 0.00018491798952794991, + "loss": 0.1248, + "step": 7563 + }, + { + "epoch": 0.1773120718249852, + "grad_norm": 0.23460440337657928, + "learning_rate": 0.0001849140992024513, + "loss": 0.0625, + "step": 7564 + }, + { + "epoch": 0.1773355133997902, + "grad_norm": 0.5752995014190674, + "learning_rate": 0.00018491020841620635, + "loss": 0.1331, + "step": 7565 + }, + { + "epoch": 0.1773589549745952, + "grad_norm": 0.22611665725708008, + "learning_rate": 0.00018490631716923622, + "loss": 0.0624, + "step": 7566 + }, + { + "epoch": 0.1773823965494002, + "grad_norm": 0.5184323191642761, + "learning_rate": 0.00018490242546156206, + "loss": 0.0938, + "step": 7567 + }, + { + "epoch": 0.17740583812420518, + "grad_norm": 0.5081623196601868, + "learning_rate": 0.00018489853329320494, + "loss": 0.1199, + "step": 7568 + }, + { + "epoch": 0.17742927969901018, + "grad_norm": 0.08924291282892227, + "learning_rate": 0.000184894640664186, + "loss": 0.0083, + "step": 7569 + }, + { + "epoch": 0.17745272127381517, + "grad_norm": 0.9433789253234863, + "learning_rate": 0.00018489074757452634, + "loss": 0.6659, + "step": 7570 + }, + { + "epoch": 0.17747616284862017, + "grad_norm": 0.4996302127838135, + "learning_rate": 0.0001848868540242471, + "loss": 0.0924, + "step": 7571 + }, + { + "epoch": 0.17749960442342516, + "grad_norm": 0.24181708693504333, + "learning_rate": 0.00018488296001336943, + "loss": 0.0564, + "step": 7572 + }, + { + "epoch": 0.17752304599823016, + "grad_norm": 0.5958655476570129, + "learning_rate": 0.0001848790655419144, + "loss": 0.1342, + "step": 7573 + }, + { + "epoch": 0.17754648757303515, + "grad_norm": 0.2125936895608902, + "learning_rate": 0.00018487517060990317, + "loss": 0.044, + "step": 7574 + }, + { + "epoch": 0.17756992914784014, + "grad_norm": 0.5018717646598816, + "learning_rate": 0.00018487127521735688, + "loss": 0.0849, + "step": 7575 + }, + { + "epoch": 0.17759337072264514, + "grad_norm": 0.6213692426681519, + "learning_rate": 0.00018486737936429667, + "loss": 0.1569, + "step": 7576 + }, + { + "epoch": 0.17761681229745013, + "grad_norm": 0.2786349058151245, + "learning_rate": 0.00018486348305074367, + "loss": 0.0733, + "step": 7577 + }, + { + "epoch": 0.17764025387225513, + "grad_norm": 0.12599731981754303, + "learning_rate": 0.000184859586276719, + "loss": 0.026, + "step": 7578 + }, + { + "epoch": 0.17766369544706012, + "grad_norm": 0.6001585125923157, + "learning_rate": 0.00018485568904224388, + "loss": 0.2266, + "step": 7579 + }, + { + "epoch": 0.17768713702186512, + "grad_norm": 0.5248948931694031, + "learning_rate": 0.0001848517913473394, + "loss": 0.1045, + "step": 7580 + }, + { + "epoch": 0.17771057859667014, + "grad_norm": 0.4901806116104126, + "learning_rate": 0.0001848478931920267, + "loss": 0.1473, + "step": 7581 + }, + { + "epoch": 0.17773402017147513, + "grad_norm": 0.17639997601509094, + "learning_rate": 0.00018484399457632694, + "loss": 0.0304, + "step": 7582 + }, + { + "epoch": 0.17775746174628013, + "grad_norm": 0.25076788663864136, + "learning_rate": 0.0001848400955002613, + "loss": 0.0665, + "step": 7583 + }, + { + "epoch": 0.17778090332108512, + "grad_norm": 0.39078617095947266, + "learning_rate": 0.00018483619596385088, + "loss": 0.0823, + "step": 7584 + }, + { + "epoch": 0.17780434489589012, + "grad_norm": 0.4833388924598694, + "learning_rate": 0.0001848322959671169, + "loss": 0.0981, + "step": 7585 + }, + { + "epoch": 0.1778277864706951, + "grad_norm": 0.5297741293907166, + "learning_rate": 0.00018482839551008047, + "loss": 0.0987, + "step": 7586 + }, + { + "epoch": 0.1778512280455001, + "grad_norm": 0.7327091097831726, + "learning_rate": 0.00018482449459276281, + "loss": 0.793, + "step": 7587 + }, + { + "epoch": 0.1778746696203051, + "grad_norm": 1.2381181716918945, + "learning_rate": 0.00018482059321518505, + "loss": 0.1497, + "step": 7588 + }, + { + "epoch": 0.1778981111951101, + "grad_norm": 0.4541827440261841, + "learning_rate": 0.00018481669137736836, + "loss": 0.1006, + "step": 7589 + }, + { + "epoch": 0.1779215527699151, + "grad_norm": 0.37487512826919556, + "learning_rate": 0.00018481278907933396, + "loss": 0.0876, + "step": 7590 + }, + { + "epoch": 0.17794499434472008, + "grad_norm": 0.5759239792823792, + "learning_rate": 0.00018480888632110292, + "loss": 0.7496, + "step": 7591 + }, + { + "epoch": 0.17796843591952508, + "grad_norm": 0.6312023401260376, + "learning_rate": 0.00018480498310269653, + "loss": 0.1524, + "step": 7592 + }, + { + "epoch": 0.17799187749433007, + "grad_norm": 0.35753270983695984, + "learning_rate": 0.00018480107942413587, + "loss": 0.0538, + "step": 7593 + }, + { + "epoch": 0.17801531906913506, + "grad_norm": 0.7501944303512573, + "learning_rate": 0.0001847971752854422, + "loss": 0.1335, + "step": 7594 + }, + { + "epoch": 0.17803876064394006, + "grad_norm": 0.7084853053092957, + "learning_rate": 0.00018479327068663664, + "loss": 0.8252, + "step": 7595 + }, + { + "epoch": 0.17806220221874505, + "grad_norm": 1.2777209281921387, + "learning_rate": 0.00018478936562774044, + "loss": 0.2951, + "step": 7596 + }, + { + "epoch": 0.17808564379355005, + "grad_norm": 0.8749207854270935, + "learning_rate": 0.00018478546010877476, + "loss": 0.1603, + "step": 7597 + }, + { + "epoch": 0.17810908536835504, + "grad_norm": 0.7221179008483887, + "learning_rate": 0.00018478155412976076, + "loss": 0.182, + "step": 7598 + }, + { + "epoch": 0.17813252694316004, + "grad_norm": 0.6073198914527893, + "learning_rate": 0.0001847776476907197, + "loss": 0.1068, + "step": 7599 + }, + { + "epoch": 0.17815596851796503, + "grad_norm": 0.7684563994407654, + "learning_rate": 0.00018477374079167272, + "loss": 0.2174, + "step": 7600 + }, + { + "epoch": 0.17817941009277002, + "grad_norm": 0.9145589470863342, + "learning_rate": 0.00018476983343264104, + "loss": 0.1638, + "step": 7601 + }, + { + "epoch": 0.17820285166757502, + "grad_norm": 0.6445520520210266, + "learning_rate": 0.00018476592561364585, + "loss": 0.124, + "step": 7602 + }, + { + "epoch": 0.17822629324238, + "grad_norm": 0.4988126754760742, + "learning_rate": 0.0001847620173347084, + "loss": 0.1618, + "step": 7603 + }, + { + "epoch": 0.178249734817185, + "grad_norm": 0.5094658732414246, + "learning_rate": 0.00018475810859584982, + "loss": 0.1371, + "step": 7604 + }, + { + "epoch": 0.17827317639199, + "grad_norm": 0.2071087807416916, + "learning_rate": 0.00018475419939709137, + "loss": 0.0416, + "step": 7605 + }, + { + "epoch": 0.178296617966795, + "grad_norm": 0.7229111790657043, + "learning_rate": 0.00018475028973845425, + "loss": 0.1571, + "step": 7606 + }, + { + "epoch": 0.17832005954160002, + "grad_norm": 0.29828616976737976, + "learning_rate": 0.0001847463796199597, + "loss": 0.0679, + "step": 7607 + }, + { + "epoch": 0.178343501116405, + "grad_norm": 0.1910124123096466, + "learning_rate": 0.0001847424690416289, + "loss": 0.0552, + "step": 7608 + }, + { + "epoch": 0.17836694269121, + "grad_norm": 0.3243410885334015, + "learning_rate": 0.00018473855800348307, + "loss": 0.0671, + "step": 7609 + }, + { + "epoch": 0.178390384266015, + "grad_norm": 0.11489306390285492, + "learning_rate": 0.00018473464650554342, + "loss": 0.0274, + "step": 7610 + }, + { + "epoch": 0.17841382584082, + "grad_norm": 0.8524791598320007, + "learning_rate": 0.00018473073454783123, + "loss": 0.1802, + "step": 7611 + }, + { + "epoch": 0.178437267415625, + "grad_norm": 0.6314359903335571, + "learning_rate": 0.00018472682213036768, + "loss": 0.2234, + "step": 7612 + }, + { + "epoch": 0.17846070899042998, + "grad_norm": 0.47830817103385925, + "learning_rate": 0.000184722909253174, + "loss": 0.1137, + "step": 7613 + }, + { + "epoch": 0.17848415056523498, + "grad_norm": 0.2177724540233612, + "learning_rate": 0.00018471899591627144, + "loss": 0.0374, + "step": 7614 + }, + { + "epoch": 0.17850759214003997, + "grad_norm": 0.4022240936756134, + "learning_rate": 0.00018471508211968124, + "loss": 0.1025, + "step": 7615 + }, + { + "epoch": 0.17853103371484497, + "grad_norm": 0.5276557803153992, + "learning_rate": 0.0001847111678634246, + "loss": 0.075, + "step": 7616 + }, + { + "epoch": 0.17855447528964996, + "grad_norm": 0.6494592428207397, + "learning_rate": 0.00018470725314752282, + "loss": 0.1542, + "step": 7617 + }, + { + "epoch": 0.17857791686445496, + "grad_norm": 0.8084151744842529, + "learning_rate": 0.00018470333797199708, + "loss": 0.1723, + "step": 7618 + }, + { + "epoch": 0.17860135843925995, + "grad_norm": 0.48507407307624817, + "learning_rate": 0.00018469942233686863, + "loss": 0.8222, + "step": 7619 + }, + { + "epoch": 0.17862480001406494, + "grad_norm": 0.4067218601703644, + "learning_rate": 0.00018469550624215875, + "loss": 0.1008, + "step": 7620 + }, + { + "epoch": 0.17864824158886994, + "grad_norm": 0.5419210195541382, + "learning_rate": 0.0001846915896878887, + "loss": 0.0799, + "step": 7621 + }, + { + "epoch": 0.17867168316367493, + "grad_norm": 0.5745722651481628, + "learning_rate": 0.00018468767267407963, + "loss": 0.6595, + "step": 7622 + }, + { + "epoch": 0.17869512473847993, + "grad_norm": 0.5671544075012207, + "learning_rate": 0.00018468375520075295, + "loss": 0.1738, + "step": 7623 + }, + { + "epoch": 0.17871856631328492, + "grad_norm": 0.3734149634838104, + "learning_rate": 0.00018467983726792978, + "loss": 0.0793, + "step": 7624 + }, + { + "epoch": 0.17874200788808992, + "grad_norm": 0.4759626090526581, + "learning_rate": 0.00018467591887563143, + "loss": 0.7994, + "step": 7625 + }, + { + "epoch": 0.1787654494628949, + "grad_norm": 0.16957040131092072, + "learning_rate": 0.00018467200002387916, + "loss": 0.0385, + "step": 7626 + }, + { + "epoch": 0.1787888910376999, + "grad_norm": 0.7808760404586792, + "learning_rate": 0.0001846680807126943, + "loss": 0.1025, + "step": 7627 + }, + { + "epoch": 0.1788123326125049, + "grad_norm": 0.5291102528572083, + "learning_rate": 0.000184664160942098, + "loss": 0.1306, + "step": 7628 + }, + { + "epoch": 0.1788357741873099, + "grad_norm": 0.43730002641677856, + "learning_rate": 0.00018466024071211158, + "loss": 0.1087, + "step": 7629 + }, + { + "epoch": 0.1788592157621149, + "grad_norm": 0.6328936219215393, + "learning_rate": 0.00018465632002275632, + "loss": 0.1123, + "step": 7630 + }, + { + "epoch": 0.17888265733691988, + "grad_norm": 0.8341935873031616, + "learning_rate": 0.0001846523988740535, + "loss": 0.3071, + "step": 7631 + }, + { + "epoch": 0.1789060989117249, + "grad_norm": 0.5054419040679932, + "learning_rate": 0.00018464847726602437, + "loss": 0.136, + "step": 7632 + }, + { + "epoch": 0.1789295404865299, + "grad_norm": 0.09020298719406128, + "learning_rate": 0.00018464455519869022, + "loss": 0.0265, + "step": 7633 + }, + { + "epoch": 0.1789529820613349, + "grad_norm": 0.6749717593193054, + "learning_rate": 0.00018464063267207233, + "loss": 0.0797, + "step": 7634 + }, + { + "epoch": 0.1789764236361399, + "grad_norm": 0.6286429166793823, + "learning_rate": 0.000184636709686192, + "loss": 0.7138, + "step": 7635 + }, + { + "epoch": 0.17899986521094488, + "grad_norm": 0.5310869216918945, + "learning_rate": 0.0001846327862410705, + "loss": 0.5559, + "step": 7636 + }, + { + "epoch": 0.17902330678574987, + "grad_norm": 0.9163084030151367, + "learning_rate": 0.00018462886233672915, + "loss": 0.2362, + "step": 7637 + }, + { + "epoch": 0.17904674836055487, + "grad_norm": 0.6748580932617188, + "learning_rate": 0.00018462493797318915, + "loss": 0.1593, + "step": 7638 + }, + { + "epoch": 0.17907018993535986, + "grad_norm": 0.829136073589325, + "learning_rate": 0.00018462101315047192, + "loss": 0.2453, + "step": 7639 + }, + { + "epoch": 0.17909363151016486, + "grad_norm": 0.48848992586135864, + "learning_rate": 0.00018461708786859867, + "loss": 0.6388, + "step": 7640 + }, + { + "epoch": 0.17911707308496985, + "grad_norm": 0.3001995086669922, + "learning_rate": 0.0001846131621275907, + "loss": 0.0517, + "step": 7641 + }, + { + "epoch": 0.17914051465977485, + "grad_norm": 0.7320228815078735, + "learning_rate": 0.00018460923592746937, + "loss": 0.6269, + "step": 7642 + }, + { + "epoch": 0.17916395623457984, + "grad_norm": 0.647135317325592, + "learning_rate": 0.0001846053092682559, + "loss": 0.1463, + "step": 7643 + }, + { + "epoch": 0.17918739780938484, + "grad_norm": 0.3947334289550781, + "learning_rate": 0.0001846013821499717, + "loss": 0.0976, + "step": 7644 + }, + { + "epoch": 0.17921083938418983, + "grad_norm": 0.7461673021316528, + "learning_rate": 0.00018459745457263796, + "loss": 0.1638, + "step": 7645 + }, + { + "epoch": 0.17923428095899482, + "grad_norm": 0.6458653807640076, + "learning_rate": 0.0001845935265362761, + "loss": 0.6966, + "step": 7646 + }, + { + "epoch": 0.17925772253379982, + "grad_norm": 0.25045496225357056, + "learning_rate": 0.00018458959804090735, + "loss": 0.0382, + "step": 7647 + }, + { + "epoch": 0.1792811641086048, + "grad_norm": 0.494687557220459, + "learning_rate": 0.0001845856690865531, + "loss": 0.1762, + "step": 7648 + }, + { + "epoch": 0.1793046056834098, + "grad_norm": 0.7342650890350342, + "learning_rate": 0.00018458173967323465, + "loss": 0.1791, + "step": 7649 + }, + { + "epoch": 0.1793280472582148, + "grad_norm": 0.3303140699863434, + "learning_rate": 0.00018457780980097324, + "loss": 0.085, + "step": 7650 + }, + { + "epoch": 0.1793514888330198, + "grad_norm": 0.25624629855155945, + "learning_rate": 0.0001845738794697903, + "loss": 0.0584, + "step": 7651 + }, + { + "epoch": 0.1793749304078248, + "grad_norm": 0.15328319370746613, + "learning_rate": 0.0001845699486797071, + "loss": 0.0263, + "step": 7652 + }, + { + "epoch": 0.17939837198262978, + "grad_norm": 0.653701663017273, + "learning_rate": 0.00018456601743074497, + "loss": 0.2046, + "step": 7653 + }, + { + "epoch": 0.17942181355743478, + "grad_norm": 0.3169473111629486, + "learning_rate": 0.00018456208572292524, + "loss": 0.0802, + "step": 7654 + }, + { + "epoch": 0.17944525513223977, + "grad_norm": 0.21963843703269958, + "learning_rate": 0.0001845581535562693, + "loss": 0.08, + "step": 7655 + }, + { + "epoch": 0.17946869670704477, + "grad_norm": 0.38738641142845154, + "learning_rate": 0.0001845542209307984, + "loss": 0.0955, + "step": 7656 + }, + { + "epoch": 0.17949213828184976, + "grad_norm": 0.7626962661743164, + "learning_rate": 0.00018455028784653395, + "loss": 0.1334, + "step": 7657 + }, + { + "epoch": 0.17951557985665478, + "grad_norm": 0.7014874219894409, + "learning_rate": 0.00018454635430349722, + "loss": 0.2618, + "step": 7658 + }, + { + "epoch": 0.17953902143145978, + "grad_norm": 0.1665605902671814, + "learning_rate": 0.00018454242030170964, + "loss": 0.0537, + "step": 7659 + }, + { + "epoch": 0.17956246300626477, + "grad_norm": 0.15347857773303986, + "learning_rate": 0.0001845384858411925, + "loss": 0.0338, + "step": 7660 + }, + { + "epoch": 0.17958590458106977, + "grad_norm": 0.5511149168014526, + "learning_rate": 0.00018453455092196715, + "loss": 0.1099, + "step": 7661 + }, + { + "epoch": 0.17960934615587476, + "grad_norm": 0.8172655701637268, + "learning_rate": 0.00018453061554405493, + "loss": 0.1246, + "step": 7662 + }, + { + "epoch": 0.17963278773067975, + "grad_norm": 0.6667346358299255, + "learning_rate": 0.00018452667970747725, + "loss": 0.6091, + "step": 7663 + }, + { + "epoch": 0.17965622930548475, + "grad_norm": 0.5431780815124512, + "learning_rate": 0.0001845227434122554, + "loss": 0.1044, + "step": 7664 + }, + { + "epoch": 0.17967967088028974, + "grad_norm": 0.621415913105011, + "learning_rate": 0.00018451880665841077, + "loss": 0.1557, + "step": 7665 + }, + { + "epoch": 0.17970311245509474, + "grad_norm": 0.5357124209403992, + "learning_rate": 0.00018451486944596473, + "loss": 0.1302, + "step": 7666 + }, + { + "epoch": 0.17972655402989973, + "grad_norm": 0.7907086610794067, + "learning_rate": 0.00018451093177493863, + "loss": 0.1899, + "step": 7667 + }, + { + "epoch": 0.17974999560470473, + "grad_norm": 0.6703487038612366, + "learning_rate": 0.00018450699364535385, + "loss": 0.1462, + "step": 7668 + }, + { + "epoch": 0.17977343717950972, + "grad_norm": 0.3591451644897461, + "learning_rate": 0.0001845030550572317, + "loss": 0.1214, + "step": 7669 + }, + { + "epoch": 0.17979687875431472, + "grad_norm": 0.6669412851333618, + "learning_rate": 0.00018449911601059365, + "loss": 0.2126, + "step": 7670 + }, + { + "epoch": 0.1798203203291197, + "grad_norm": 1.1426477432250977, + "learning_rate": 0.000184495176505461, + "loss": 0.1571, + "step": 7671 + }, + { + "epoch": 0.1798437619039247, + "grad_norm": 0.7809411287307739, + "learning_rate": 0.00018449123654185515, + "loss": 0.2588, + "step": 7672 + }, + { + "epoch": 0.1798672034787297, + "grad_norm": 0.46883150935173035, + "learning_rate": 0.00018448729611979746, + "loss": 0.0949, + "step": 7673 + }, + { + "epoch": 0.1798906450535347, + "grad_norm": 0.6106113195419312, + "learning_rate": 0.00018448335523930937, + "loss": 0.5422, + "step": 7674 + }, + { + "epoch": 0.1799140866283397, + "grad_norm": 0.5417326092720032, + "learning_rate": 0.0001844794139004122, + "loss": 0.6304, + "step": 7675 + }, + { + "epoch": 0.17993752820314468, + "grad_norm": 0.14598117768764496, + "learning_rate": 0.00018447547210312731, + "loss": 0.0378, + "step": 7676 + }, + { + "epoch": 0.17996096977794968, + "grad_norm": 0.39967963099479675, + "learning_rate": 0.00018447152984747619, + "loss": 0.1059, + "step": 7677 + }, + { + "epoch": 0.17998441135275467, + "grad_norm": 0.5712972283363342, + "learning_rate": 0.00018446758713348013, + "loss": 0.2451, + "step": 7678 + }, + { + "epoch": 0.18000785292755966, + "grad_norm": 0.46573108434677124, + "learning_rate": 0.00018446364396116057, + "loss": 0.1577, + "step": 7679 + }, + { + "epoch": 0.18003129450236466, + "grad_norm": 0.3548228144645691, + "learning_rate": 0.00018445970033053894, + "loss": 0.088, + "step": 7680 + }, + { + "epoch": 0.18005473607716965, + "grad_norm": 0.5494627356529236, + "learning_rate": 0.00018445575624163657, + "loss": 0.139, + "step": 7681 + }, + { + "epoch": 0.18007817765197465, + "grad_norm": 0.21974749863147736, + "learning_rate": 0.00018445181169447493, + "loss": 0.0597, + "step": 7682 + }, + { + "epoch": 0.18010161922677964, + "grad_norm": 0.5710932016372681, + "learning_rate": 0.00018444786668907534, + "loss": 0.1612, + "step": 7683 + }, + { + "epoch": 0.18012506080158466, + "grad_norm": 0.19393865764141083, + "learning_rate": 0.00018444392122545926, + "loss": 0.056, + "step": 7684 + }, + { + "epoch": 0.18014850237638966, + "grad_norm": 0.5494901537895203, + "learning_rate": 0.00018443997530364812, + "loss": 0.0876, + "step": 7685 + }, + { + "epoch": 0.18017194395119465, + "grad_norm": 0.4552406072616577, + "learning_rate": 0.00018443602892366328, + "loss": 0.132, + "step": 7686 + }, + { + "epoch": 0.18019538552599965, + "grad_norm": 0.5096695423126221, + "learning_rate": 0.00018443208208552615, + "loss": 0.1352, + "step": 7687 + }, + { + "epoch": 0.18021882710080464, + "grad_norm": 0.4938361346721649, + "learning_rate": 0.0001844281347892582, + "loss": 0.1032, + "step": 7688 + }, + { + "epoch": 0.18024226867560963, + "grad_norm": 0.5525330901145935, + "learning_rate": 0.0001844241870348808, + "loss": 0.081, + "step": 7689 + }, + { + "epoch": 0.18026571025041463, + "grad_norm": 0.35191601514816284, + "learning_rate": 0.0001844202388224154, + "loss": 0.0725, + "step": 7690 + }, + { + "epoch": 0.18028915182521962, + "grad_norm": 0.6712844967842102, + "learning_rate": 0.0001844162901518834, + "loss": 0.2062, + "step": 7691 + }, + { + "epoch": 0.18031259340002462, + "grad_norm": 0.21460434794425964, + "learning_rate": 0.00018441234102330626, + "loss": 0.0469, + "step": 7692 + }, + { + "epoch": 0.1803360349748296, + "grad_norm": 0.33570343255996704, + "learning_rate": 0.00018440839143670537, + "loss": 0.091, + "step": 7693 + }, + { + "epoch": 0.1803594765496346, + "grad_norm": 0.35397472977638245, + "learning_rate": 0.00018440444139210214, + "loss": 0.1006, + "step": 7694 + }, + { + "epoch": 0.1803829181244396, + "grad_norm": 0.14036476612091064, + "learning_rate": 0.00018440049088951806, + "loss": 0.0285, + "step": 7695 + }, + { + "epoch": 0.1804063596992446, + "grad_norm": 0.48240548372268677, + "learning_rate": 0.00018439653992897458, + "loss": 0.1428, + "step": 7696 + }, + { + "epoch": 0.1804298012740496, + "grad_norm": 0.30279338359832764, + "learning_rate": 0.00018439258851049307, + "loss": 0.0467, + "step": 7697 + }, + { + "epoch": 0.18045324284885458, + "grad_norm": 0.4266464412212372, + "learning_rate": 0.000184388636634095, + "loss": 0.1107, + "step": 7698 + }, + { + "epoch": 0.18047668442365958, + "grad_norm": 0.45509210228919983, + "learning_rate": 0.0001843846842998018, + "loss": 0.084, + "step": 7699 + }, + { + "epoch": 0.18050012599846457, + "grad_norm": 0.443765789270401, + "learning_rate": 0.00018438073150763495, + "loss": 0.1093, + "step": 7700 + }, + { + "epoch": 0.18052356757326957, + "grad_norm": 0.7713013887405396, + "learning_rate": 0.0001843767782576159, + "loss": 0.2333, + "step": 7701 + }, + { + "epoch": 0.18054700914807456, + "grad_norm": 0.5052270293235779, + "learning_rate": 0.00018437282454976603, + "loss": 0.1054, + "step": 7702 + }, + { + "epoch": 0.18057045072287956, + "grad_norm": 0.6293527483940125, + "learning_rate": 0.00018436887038410688, + "loss": 0.1624, + "step": 7703 + }, + { + "epoch": 0.18059389229768455, + "grad_norm": 0.5075452327728271, + "learning_rate": 0.00018436491576065986, + "loss": 0.1247, + "step": 7704 + }, + { + "epoch": 0.18061733387248954, + "grad_norm": 0.5297377109527588, + "learning_rate": 0.00018436096067944642, + "loss": 0.157, + "step": 7705 + }, + { + "epoch": 0.18064077544729454, + "grad_norm": 0.5384724140167236, + "learning_rate": 0.00018435700514048803, + "loss": 0.1244, + "step": 7706 + }, + { + "epoch": 0.18066421702209953, + "grad_norm": 0.41333866119384766, + "learning_rate": 0.0001843530491438062, + "loss": 0.0629, + "step": 7707 + }, + { + "epoch": 0.18068765859690453, + "grad_norm": 0.6127368807792664, + "learning_rate": 0.00018434909268942232, + "loss": 0.1483, + "step": 7708 + }, + { + "epoch": 0.18071110017170955, + "grad_norm": 0.7671463489532471, + "learning_rate": 0.00018434513577735787, + "loss": 0.4715, + "step": 7709 + }, + { + "epoch": 0.18073454174651454, + "grad_norm": 0.5268280506134033, + "learning_rate": 0.0001843411784076344, + "loss": 0.1214, + "step": 7710 + }, + { + "epoch": 0.18075798332131954, + "grad_norm": 0.6169238686561584, + "learning_rate": 0.00018433722058027328, + "loss": 0.0727, + "step": 7711 + }, + { + "epoch": 0.18078142489612453, + "grad_norm": 0.7632600665092468, + "learning_rate": 0.00018433326229529606, + "loss": 0.1487, + "step": 7712 + }, + { + "epoch": 0.18080486647092953, + "grad_norm": 0.7809204459190369, + "learning_rate": 0.00018432930355272417, + "loss": 0.2239, + "step": 7713 + }, + { + "epoch": 0.18082830804573452, + "grad_norm": 0.16168904304504395, + "learning_rate": 0.00018432534435257908, + "loss": 0.0321, + "step": 7714 + }, + { + "epoch": 0.18085174962053951, + "grad_norm": 0.7745987772941589, + "learning_rate": 0.00018432138469488233, + "loss": 0.8934, + "step": 7715 + }, + { + "epoch": 0.1808751911953445, + "grad_norm": 0.46868202090263367, + "learning_rate": 0.00018431742457965537, + "loss": 0.0932, + "step": 7716 + }, + { + "epoch": 0.1808986327701495, + "grad_norm": 0.7285307049751282, + "learning_rate": 0.00018431346400691972, + "loss": 0.46, + "step": 7717 + }, + { + "epoch": 0.1809220743449545, + "grad_norm": 0.34165212512016296, + "learning_rate": 0.0001843095029766968, + "loss": 0.0921, + "step": 7718 + }, + { + "epoch": 0.1809455159197595, + "grad_norm": 0.3672334849834442, + "learning_rate": 0.00018430554148900818, + "loss": 0.0482, + "step": 7719 + }, + { + "epoch": 0.1809689574945645, + "grad_norm": 0.26807156205177307, + "learning_rate": 0.00018430157954387532, + "loss": 0.037, + "step": 7720 + }, + { + "epoch": 0.18099239906936948, + "grad_norm": 0.565215528011322, + "learning_rate": 0.0001842976171413197, + "loss": 0.7739, + "step": 7721 + }, + { + "epoch": 0.18101584064417447, + "grad_norm": 0.7121895551681519, + "learning_rate": 0.00018429365428136285, + "loss": 0.1713, + "step": 7722 + }, + { + "epoch": 0.18103928221897947, + "grad_norm": 1.246735692024231, + "learning_rate": 0.00018428969096402625, + "loss": 0.1163, + "step": 7723 + }, + { + "epoch": 0.18106272379378446, + "grad_norm": 0.23240430653095245, + "learning_rate": 0.00018428572718933144, + "loss": 0.0623, + "step": 7724 + }, + { + "epoch": 0.18108616536858946, + "grad_norm": 0.5874930620193481, + "learning_rate": 0.00018428176295729988, + "loss": 0.2149, + "step": 7725 + }, + { + "epoch": 0.18110960694339445, + "grad_norm": 0.9831304550170898, + "learning_rate": 0.0001842777982679531, + "loss": 0.1332, + "step": 7726 + }, + { + "epoch": 0.18113304851819945, + "grad_norm": 0.5907866954803467, + "learning_rate": 0.00018427383312131265, + "loss": 0.1673, + "step": 7727 + }, + { + "epoch": 0.18115649009300444, + "grad_norm": 0.4440954625606537, + "learning_rate": 0.00018426986751740002, + "loss": 0.1175, + "step": 7728 + }, + { + "epoch": 0.18117993166780944, + "grad_norm": 0.47908705472946167, + "learning_rate": 0.00018426590145623667, + "loss": 0.1037, + "step": 7729 + }, + { + "epoch": 0.18120337324261443, + "grad_norm": 0.23442552983760834, + "learning_rate": 0.00018426193493784417, + "loss": 0.0824, + "step": 7730 + }, + { + "epoch": 0.18122681481741942, + "grad_norm": 0.3832928240299225, + "learning_rate": 0.00018425796796224408, + "loss": 0.1271, + "step": 7731 + }, + { + "epoch": 0.18125025639222442, + "grad_norm": 0.30309802293777466, + "learning_rate": 0.00018425400052945787, + "loss": 0.0636, + "step": 7732 + }, + { + "epoch": 0.1812736979670294, + "grad_norm": 0.36618858575820923, + "learning_rate": 0.0001842500326395071, + "loss": 0.0738, + "step": 7733 + }, + { + "epoch": 0.1812971395418344, + "grad_norm": 0.4105913043022156, + "learning_rate": 0.00018424606429241327, + "loss": 0.128, + "step": 7734 + }, + { + "epoch": 0.18132058111663943, + "grad_norm": 0.16054277122020721, + "learning_rate": 0.0001842420954881979, + "loss": 0.023, + "step": 7735 + }, + { + "epoch": 0.18134402269144442, + "grad_norm": 1.0301034450531006, + "learning_rate": 0.00018423812622688258, + "loss": 0.2442, + "step": 7736 + }, + { + "epoch": 0.18136746426624942, + "grad_norm": 0.20097801089286804, + "learning_rate": 0.00018423415650848887, + "loss": 0.0448, + "step": 7737 + }, + { + "epoch": 0.1813909058410544, + "grad_norm": 0.13747426867485046, + "learning_rate": 0.00018423018633303818, + "loss": 0.0238, + "step": 7738 + }, + { + "epoch": 0.1814143474158594, + "grad_norm": 0.35184574127197266, + "learning_rate": 0.00018422621570055215, + "loss": 0.1332, + "step": 7739 + }, + { + "epoch": 0.1814377889906644, + "grad_norm": 0.2657695710659027, + "learning_rate": 0.0001842222446110523, + "loss": 0.3517, + "step": 7740 + }, + { + "epoch": 0.1814612305654694, + "grad_norm": 0.14707475900650024, + "learning_rate": 0.0001842182730645602, + "loss": 0.0497, + "step": 7741 + }, + { + "epoch": 0.1814846721402744, + "grad_norm": 0.23950296640396118, + "learning_rate": 0.00018421430106109737, + "loss": 0.0627, + "step": 7742 + }, + { + "epoch": 0.18150811371507938, + "grad_norm": 0.4917319118976593, + "learning_rate": 0.00018421032860068537, + "loss": 0.0896, + "step": 7743 + }, + { + "epoch": 0.18153155528988438, + "grad_norm": 0.4448402523994446, + "learning_rate": 0.00018420635568334578, + "loss": 0.115, + "step": 7744 + }, + { + "epoch": 0.18155499686468937, + "grad_norm": 0.36436349153518677, + "learning_rate": 0.00018420238230910013, + "loss": 0.1229, + "step": 7745 + }, + { + "epoch": 0.18157843843949437, + "grad_norm": 0.19531197845935822, + "learning_rate": 0.00018419840847796998, + "loss": 0.0317, + "step": 7746 + }, + { + "epoch": 0.18160188001429936, + "grad_norm": 0.5713063478469849, + "learning_rate": 0.00018419443418997688, + "loss": 0.2125, + "step": 7747 + }, + { + "epoch": 0.18162532158910435, + "grad_norm": 0.5016030669212341, + "learning_rate": 0.00018419045944514245, + "loss": 0.1463, + "step": 7748 + }, + { + "epoch": 0.18164876316390935, + "grad_norm": 0.19356362521648407, + "learning_rate": 0.00018418648424348822, + "loss": 0.0511, + "step": 7749 + }, + { + "epoch": 0.18167220473871434, + "grad_norm": 0.31939104199409485, + "learning_rate": 0.0001841825085850357, + "loss": 0.051, + "step": 7750 + }, + { + "epoch": 0.18169564631351934, + "grad_norm": 0.8227411508560181, + "learning_rate": 0.00018417853246980657, + "loss": 0.164, + "step": 7751 + }, + { + "epoch": 0.18171908788832433, + "grad_norm": 0.22493961453437805, + "learning_rate": 0.00018417455589782235, + "loss": 0.0616, + "step": 7752 + }, + { + "epoch": 0.18174252946312933, + "grad_norm": 0.7055990695953369, + "learning_rate": 0.00018417057886910463, + "loss": 0.1512, + "step": 7753 + }, + { + "epoch": 0.18176597103793432, + "grad_norm": 0.5480656027793884, + "learning_rate": 0.00018416660138367497, + "loss": 0.1758, + "step": 7754 + }, + { + "epoch": 0.18178941261273932, + "grad_norm": 0.4519806504249573, + "learning_rate": 0.00018416262344155498, + "loss": 0.1028, + "step": 7755 + }, + { + "epoch": 0.1818128541875443, + "grad_norm": 0.5423775315284729, + "learning_rate": 0.0001841586450427662, + "loss": 0.1914, + "step": 7756 + }, + { + "epoch": 0.1818362957623493, + "grad_norm": 1.424468994140625, + "learning_rate": 0.00018415466618733025, + "loss": 0.123, + "step": 7757 + }, + { + "epoch": 0.1818597373371543, + "grad_norm": 0.6511371731758118, + "learning_rate": 0.00018415068687526874, + "loss": 0.1754, + "step": 7758 + }, + { + "epoch": 0.1818831789119593, + "grad_norm": 0.3200088143348694, + "learning_rate": 0.00018414670710660325, + "loss": 0.0905, + "step": 7759 + }, + { + "epoch": 0.18190662048676431, + "grad_norm": 0.4868009388446808, + "learning_rate": 0.00018414272688135533, + "loss": 0.091, + "step": 7760 + }, + { + "epoch": 0.1819300620615693, + "grad_norm": 0.5042424201965332, + "learning_rate": 0.0001841387461995466, + "loss": 0.0945, + "step": 7761 + }, + { + "epoch": 0.1819535036363743, + "grad_norm": 0.2661791741847992, + "learning_rate": 0.0001841347650611987, + "loss": 0.0768, + "step": 7762 + }, + { + "epoch": 0.1819769452111793, + "grad_norm": 0.5134760737419128, + "learning_rate": 0.00018413078346633316, + "loss": 0.1207, + "step": 7763 + }, + { + "epoch": 0.1820003867859843, + "grad_norm": 0.3853660523891449, + "learning_rate": 0.00018412680141497166, + "loss": 0.368, + "step": 7764 + }, + { + "epoch": 0.18202382836078929, + "grad_norm": 0.4318907856941223, + "learning_rate": 0.00018412281890713573, + "loss": 0.152, + "step": 7765 + }, + { + "epoch": 0.18204726993559428, + "grad_norm": 0.791029155254364, + "learning_rate": 0.00018411883594284705, + "loss": 0.1313, + "step": 7766 + }, + { + "epoch": 0.18207071151039927, + "grad_norm": 0.8916715383529663, + "learning_rate": 0.0001841148525221272, + "loss": 0.2049, + "step": 7767 + }, + { + "epoch": 0.18209415308520427, + "grad_norm": 0.3507769703865051, + "learning_rate": 0.0001841108686449978, + "loss": 0.0686, + "step": 7768 + }, + { + "epoch": 0.18211759466000926, + "grad_norm": 0.15618589520454407, + "learning_rate": 0.00018410688431148042, + "loss": 0.0285, + "step": 7769 + }, + { + "epoch": 0.18214103623481426, + "grad_norm": 0.48222559690475464, + "learning_rate": 0.00018410289952159676, + "loss": 0.1064, + "step": 7770 + }, + { + "epoch": 0.18216447780961925, + "grad_norm": 0.6676312685012817, + "learning_rate": 0.0001840989142753684, + "loss": 0.6654, + "step": 7771 + }, + { + "epoch": 0.18218791938442425, + "grad_norm": 0.26242873072624207, + "learning_rate": 0.00018409492857281696, + "loss": 0.0411, + "step": 7772 + }, + { + "epoch": 0.18221136095922924, + "grad_norm": 0.5092819929122925, + "learning_rate": 0.00018409094241396407, + "loss": 0.1279, + "step": 7773 + }, + { + "epoch": 0.18223480253403423, + "grad_norm": 0.6347038149833679, + "learning_rate": 0.00018408695579883138, + "loss": 0.13, + "step": 7774 + }, + { + "epoch": 0.18225824410883923, + "grad_norm": 1.4213144779205322, + "learning_rate": 0.00018408296872744049, + "loss": 0.1023, + "step": 7775 + }, + { + "epoch": 0.18228168568364422, + "grad_norm": 0.8441895842552185, + "learning_rate": 0.00018407898119981303, + "loss": 0.2336, + "step": 7776 + }, + { + "epoch": 0.18230512725844922, + "grad_norm": 0.5875049233436584, + "learning_rate": 0.00018407499321597068, + "loss": 0.0894, + "step": 7777 + }, + { + "epoch": 0.1823285688332542, + "grad_norm": 0.7566311359405518, + "learning_rate": 0.00018407100477593505, + "loss": 0.777, + "step": 7778 + }, + { + "epoch": 0.1823520104080592, + "grad_norm": 0.460846871137619, + "learning_rate": 0.00018406701587972776, + "loss": 0.5579, + "step": 7779 + }, + { + "epoch": 0.1823754519828642, + "grad_norm": 0.17911139130592346, + "learning_rate": 0.0001840630265273705, + "loss": 0.0361, + "step": 7780 + }, + { + "epoch": 0.1823988935576692, + "grad_norm": 0.8536090850830078, + "learning_rate": 0.00018405903671888488, + "loss": 0.1737, + "step": 7781 + }, + { + "epoch": 0.1824223351324742, + "grad_norm": 0.6413509249687195, + "learning_rate": 0.0001840550464542926, + "loss": 0.6836, + "step": 7782 + }, + { + "epoch": 0.18244577670727918, + "grad_norm": 0.35972118377685547, + "learning_rate": 0.00018405105573361525, + "loss": 0.0872, + "step": 7783 + }, + { + "epoch": 0.18246921828208418, + "grad_norm": 1.079451322555542, + "learning_rate": 0.0001840470645568745, + "loss": 0.2281, + "step": 7784 + }, + { + "epoch": 0.18249265985688917, + "grad_norm": 0.16339580714702606, + "learning_rate": 0.00018404307292409203, + "loss": 0.0344, + "step": 7785 + }, + { + "epoch": 0.1825161014316942, + "grad_norm": 0.6184564828872681, + "learning_rate": 0.00018403908083528946, + "loss": 0.6659, + "step": 7786 + }, + { + "epoch": 0.1825395430064992, + "grad_norm": 0.6712220907211304, + "learning_rate": 0.00018403508829048848, + "loss": 0.1411, + "step": 7787 + }, + { + "epoch": 0.18256298458130418, + "grad_norm": 0.5555512309074402, + "learning_rate": 0.00018403109528971077, + "loss": 0.1511, + "step": 7788 + }, + { + "epoch": 0.18258642615610918, + "grad_norm": 0.5268332362174988, + "learning_rate": 0.00018402710183297797, + "loss": 0.1251, + "step": 7789 + }, + { + "epoch": 0.18260986773091417, + "grad_norm": 0.12460042536258698, + "learning_rate": 0.00018402310792031172, + "loss": 0.0334, + "step": 7790 + }, + { + "epoch": 0.18263330930571917, + "grad_norm": 0.6557725667953491, + "learning_rate": 0.00018401911355173375, + "loss": 0.1575, + "step": 7791 + }, + { + "epoch": 0.18265675088052416, + "grad_norm": 0.5362340211868286, + "learning_rate": 0.00018401511872726572, + "loss": 0.1813, + "step": 7792 + }, + { + "epoch": 0.18268019245532915, + "grad_norm": 0.5777924656867981, + "learning_rate": 0.00018401112344692928, + "loss": 0.1294, + "step": 7793 + }, + { + "epoch": 0.18270363403013415, + "grad_norm": 0.88604336977005, + "learning_rate": 0.00018400712771074611, + "loss": 0.0919, + "step": 7794 + }, + { + "epoch": 0.18272707560493914, + "grad_norm": 0.39499491453170776, + "learning_rate": 0.00018400313151873792, + "loss": 0.1072, + "step": 7795 + }, + { + "epoch": 0.18275051717974414, + "grad_norm": 0.4455626606941223, + "learning_rate": 0.00018399913487092637, + "loss": 0.092, + "step": 7796 + }, + { + "epoch": 0.18277395875454913, + "grad_norm": 0.4958915114402771, + "learning_rate": 0.00018399513776733313, + "loss": 0.1543, + "step": 7797 + }, + { + "epoch": 0.18279740032935413, + "grad_norm": 0.6154912710189819, + "learning_rate": 0.00018399114020797993, + "loss": 0.1205, + "step": 7798 + }, + { + "epoch": 0.18282084190415912, + "grad_norm": 0.5918036699295044, + "learning_rate": 0.00018398714219288844, + "loss": 0.102, + "step": 7799 + }, + { + "epoch": 0.18284428347896411, + "grad_norm": 0.22470928728580475, + "learning_rate": 0.00018398314372208038, + "loss": 0.0568, + "step": 7800 + }, + { + "epoch": 0.1828677250537691, + "grad_norm": 0.4099430739879608, + "learning_rate": 0.00018397914479557738, + "loss": 0.0819, + "step": 7801 + }, + { + "epoch": 0.1828911666285741, + "grad_norm": 0.4960528016090393, + "learning_rate": 0.00018397514541340122, + "loss": 0.1006, + "step": 7802 + }, + { + "epoch": 0.1829146082033791, + "grad_norm": 0.21913200616836548, + "learning_rate": 0.00018397114557557355, + "loss": 0.0698, + "step": 7803 + }, + { + "epoch": 0.1829380497781841, + "grad_norm": 0.377022385597229, + "learning_rate": 0.00018396714528211606, + "loss": 0.0778, + "step": 7804 + }, + { + "epoch": 0.1829614913529891, + "grad_norm": 0.5933921933174133, + "learning_rate": 0.00018396314453305047, + "loss": 0.76, + "step": 7805 + }, + { + "epoch": 0.18298493292779408, + "grad_norm": 0.4306478798389435, + "learning_rate": 0.00018395914332839855, + "loss": 0.4042, + "step": 7806 + }, + { + "epoch": 0.18300837450259907, + "grad_norm": 0.7933269739151001, + "learning_rate": 0.0001839551416681819, + "loss": 0.1935, + "step": 7807 + }, + { + "epoch": 0.18303181607740407, + "grad_norm": 0.20473209023475647, + "learning_rate": 0.0001839511395524223, + "loss": 0.0469, + "step": 7808 + }, + { + "epoch": 0.18305525765220906, + "grad_norm": 0.39103806018829346, + "learning_rate": 0.0001839471369811415, + "loss": 0.2472, + "step": 7809 + }, + { + "epoch": 0.18307869922701406, + "grad_norm": 0.6343073844909668, + "learning_rate": 0.00018394313395436114, + "loss": 0.1484, + "step": 7810 + }, + { + "epoch": 0.18310214080181908, + "grad_norm": 0.5725985765457153, + "learning_rate": 0.00018393913047210297, + "loss": 0.1653, + "step": 7811 + }, + { + "epoch": 0.18312558237662407, + "grad_norm": 0.8160611987113953, + "learning_rate": 0.00018393512653438877, + "loss": 0.1878, + "step": 7812 + }, + { + "epoch": 0.18314902395142907, + "grad_norm": 0.7904089093208313, + "learning_rate": 0.00018393112214124015, + "loss": 0.1605, + "step": 7813 + }, + { + "epoch": 0.18317246552623406, + "grad_norm": 0.4372826814651489, + "learning_rate": 0.0001839271172926789, + "loss": 0.1189, + "step": 7814 + }, + { + "epoch": 0.18319590710103906, + "grad_norm": 0.2332017421722412, + "learning_rate": 0.0001839231119887268, + "loss": 0.0343, + "step": 7815 + }, + { + "epoch": 0.18321934867584405, + "grad_norm": 0.13227099180221558, + "learning_rate": 0.00018391910622940548, + "loss": 0.0334, + "step": 7816 + }, + { + "epoch": 0.18324279025064905, + "grad_norm": 0.29066187143325806, + "learning_rate": 0.00018391510001473677, + "loss": 0.1243, + "step": 7817 + }, + { + "epoch": 0.18326623182545404, + "grad_norm": 0.4115495979785919, + "learning_rate": 0.00018391109334474237, + "loss": 0.0924, + "step": 7818 + }, + { + "epoch": 0.18328967340025903, + "grad_norm": 0.6376981735229492, + "learning_rate": 0.00018390708621944402, + "loss": 0.0672, + "step": 7819 + }, + { + "epoch": 0.18331311497506403, + "grad_norm": 0.632804811000824, + "learning_rate": 0.00018390307863886345, + "loss": 0.1949, + "step": 7820 + }, + { + "epoch": 0.18333655654986902, + "grad_norm": 0.6520270705223083, + "learning_rate": 0.0001838990706030224, + "loss": 0.2027, + "step": 7821 + }, + { + "epoch": 0.18335999812467402, + "grad_norm": 0.689024031162262, + "learning_rate": 0.00018389506211194265, + "loss": 0.1763, + "step": 7822 + }, + { + "epoch": 0.183383439699479, + "grad_norm": 0.6206474900245667, + "learning_rate": 0.00018389105316564592, + "loss": 0.1809, + "step": 7823 + }, + { + "epoch": 0.183406881274284, + "grad_norm": 0.464765340089798, + "learning_rate": 0.00018388704376415399, + "loss": 0.134, + "step": 7824 + }, + { + "epoch": 0.183430322849089, + "grad_norm": 0.5785513520240784, + "learning_rate": 0.00018388303390748863, + "loss": 0.1403, + "step": 7825 + }, + { + "epoch": 0.183453764423894, + "grad_norm": 0.8205791711807251, + "learning_rate": 0.0001838790235956715, + "loss": 0.2799, + "step": 7826 + }, + { + "epoch": 0.183477205998699, + "grad_norm": 0.6809906363487244, + "learning_rate": 0.00018387501282872448, + "loss": 0.162, + "step": 7827 + }, + { + "epoch": 0.18350064757350398, + "grad_norm": 0.8375452756881714, + "learning_rate": 0.00018387100160666928, + "loss": 0.1562, + "step": 7828 + }, + { + "epoch": 0.18352408914830898, + "grad_norm": 0.35745969414711, + "learning_rate": 0.00018386698992952765, + "loss": 0.1183, + "step": 7829 + }, + { + "epoch": 0.18354753072311397, + "grad_norm": 0.18564021587371826, + "learning_rate": 0.0001838629777973214, + "loss": 0.0561, + "step": 7830 + }, + { + "epoch": 0.18357097229791897, + "grad_norm": 0.4831562936306, + "learning_rate": 0.00018385896521007226, + "loss": 0.1593, + "step": 7831 + }, + { + "epoch": 0.18359441387272396, + "grad_norm": 0.7023791670799255, + "learning_rate": 0.000183854952167802, + "loss": 0.2753, + "step": 7832 + }, + { + "epoch": 0.18361785544752895, + "grad_norm": 0.56514573097229, + "learning_rate": 0.0001838509386705324, + "loss": 0.184, + "step": 7833 + }, + { + "epoch": 0.18364129702233395, + "grad_norm": 0.2888402044773102, + "learning_rate": 0.0001838469247182853, + "loss": 0.06, + "step": 7834 + }, + { + "epoch": 0.18366473859713894, + "grad_norm": 0.489055871963501, + "learning_rate": 0.0001838429103110824, + "loss": 0.1172, + "step": 7835 + }, + { + "epoch": 0.18368818017194394, + "grad_norm": 0.3780592978000641, + "learning_rate": 0.0001838388954489455, + "loss": 0.0956, + "step": 7836 + }, + { + "epoch": 0.18371162174674896, + "grad_norm": 0.5575092434883118, + "learning_rate": 0.00018383488013189642, + "loss": 0.4123, + "step": 7837 + }, + { + "epoch": 0.18373506332155395, + "grad_norm": 0.595035970211029, + "learning_rate": 0.00018383086435995693, + "loss": 0.1734, + "step": 7838 + }, + { + "epoch": 0.18375850489635895, + "grad_norm": 0.1759868860244751, + "learning_rate": 0.00018382684813314878, + "loss": 0.0478, + "step": 7839 + }, + { + "epoch": 0.18378194647116394, + "grad_norm": 0.6442163586616516, + "learning_rate": 0.00018382283145149383, + "loss": 0.763, + "step": 7840 + }, + { + "epoch": 0.18380538804596894, + "grad_norm": 0.40254342555999756, + "learning_rate": 0.0001838188143150138, + "loss": 0.1002, + "step": 7841 + }, + { + "epoch": 0.18382882962077393, + "grad_norm": 0.7044575810432434, + "learning_rate": 0.00018381479672373055, + "loss": 0.2872, + "step": 7842 + }, + { + "epoch": 0.18385227119557893, + "grad_norm": 0.6260743141174316, + "learning_rate": 0.00018381077867766585, + "loss": 0.1425, + "step": 7843 + }, + { + "epoch": 0.18387571277038392, + "grad_norm": 0.5888274312019348, + "learning_rate": 0.00018380676017684154, + "loss": 0.4203, + "step": 7844 + }, + { + "epoch": 0.18389915434518891, + "grad_norm": 0.5394538044929504, + "learning_rate": 0.00018380274122127937, + "loss": 0.1386, + "step": 7845 + }, + { + "epoch": 0.1839225959199939, + "grad_norm": 0.18490485846996307, + "learning_rate": 0.00018379872181100115, + "loss": 0.0453, + "step": 7846 + }, + { + "epoch": 0.1839460374947989, + "grad_norm": 0.42393919825553894, + "learning_rate": 0.00018379470194602874, + "loss": 0.1199, + "step": 7847 + }, + { + "epoch": 0.1839694790696039, + "grad_norm": 0.5056720972061157, + "learning_rate": 0.0001837906816263839, + "loss": 0.1431, + "step": 7848 + }, + { + "epoch": 0.1839929206444089, + "grad_norm": 0.4685363173484802, + "learning_rate": 0.00018378666085208843, + "loss": 0.1372, + "step": 7849 + }, + { + "epoch": 0.18401636221921389, + "grad_norm": 0.4839577376842499, + "learning_rate": 0.00018378263962316423, + "loss": 0.1201, + "step": 7850 + }, + { + "epoch": 0.18403980379401888, + "grad_norm": 0.18874482810497284, + "learning_rate": 0.00018377861793963309, + "loss": 0.0415, + "step": 7851 + }, + { + "epoch": 0.18406324536882387, + "grad_norm": 0.38861313462257385, + "learning_rate": 0.00018377459580151679, + "loss": 0.1226, + "step": 7852 + }, + { + "epoch": 0.18408668694362887, + "grad_norm": 0.19123223423957825, + "learning_rate": 0.0001837705732088372, + "loss": 0.0213, + "step": 7853 + }, + { + "epoch": 0.18411012851843386, + "grad_norm": 0.60038822889328, + "learning_rate": 0.00018376655016161605, + "loss": 0.0935, + "step": 7854 + }, + { + "epoch": 0.18413357009323886, + "grad_norm": 0.10991685837507248, + "learning_rate": 0.00018376252665987532, + "loss": 0.028, + "step": 7855 + }, + { + "epoch": 0.18415701166804385, + "grad_norm": 0.4020353853702545, + "learning_rate": 0.00018375850270363673, + "loss": 0.1111, + "step": 7856 + }, + { + "epoch": 0.18418045324284885, + "grad_norm": 0.8007220029830933, + "learning_rate": 0.00018375447829292217, + "loss": 0.15, + "step": 7857 + }, + { + "epoch": 0.18420389481765384, + "grad_norm": 0.5078946352005005, + "learning_rate": 0.00018375045342775343, + "loss": 0.1253, + "step": 7858 + }, + { + "epoch": 0.18422733639245883, + "grad_norm": 0.5183929800987244, + "learning_rate": 0.0001837464281081524, + "loss": 0.174, + "step": 7859 + }, + { + "epoch": 0.18425077796726383, + "grad_norm": 0.3258698880672455, + "learning_rate": 0.0001837424023341409, + "loss": 0.0479, + "step": 7860 + }, + { + "epoch": 0.18427421954206882, + "grad_norm": 0.8814918398857117, + "learning_rate": 0.00018373837610574073, + "loss": 0.198, + "step": 7861 + }, + { + "epoch": 0.18429766111687385, + "grad_norm": 0.5898449420928955, + "learning_rate": 0.00018373434942297383, + "loss": 0.1518, + "step": 7862 + }, + { + "epoch": 0.18432110269167884, + "grad_norm": 0.19811876118183136, + "learning_rate": 0.00018373032228586196, + "loss": 0.0335, + "step": 7863 + }, + { + "epoch": 0.18434454426648383, + "grad_norm": 0.48773816227912903, + "learning_rate": 0.000183726294694427, + "loss": 0.0671, + "step": 7864 + }, + { + "epoch": 0.18436798584128883, + "grad_norm": 1.0500065088272095, + "learning_rate": 0.00018372226664869083, + "loss": 0.2168, + "step": 7865 + }, + { + "epoch": 0.18439142741609382, + "grad_norm": 0.8786992430686951, + "learning_rate": 0.0001837182381486753, + "loss": 0.17, + "step": 7866 + }, + { + "epoch": 0.18441486899089882, + "grad_norm": 0.7224990129470825, + "learning_rate": 0.00018371420919440222, + "loss": 0.6089, + "step": 7867 + }, + { + "epoch": 0.1844383105657038, + "grad_norm": 0.6296294927597046, + "learning_rate": 0.00018371017978589352, + "loss": 0.116, + "step": 7868 + }, + { + "epoch": 0.1844617521405088, + "grad_norm": 0.6784953474998474, + "learning_rate": 0.00018370614992317102, + "loss": 0.6852, + "step": 7869 + }, + { + "epoch": 0.1844851937153138, + "grad_norm": 0.6887684464454651, + "learning_rate": 0.0001837021196062566, + "loss": 0.2275, + "step": 7870 + }, + { + "epoch": 0.1845086352901188, + "grad_norm": 0.47185635566711426, + "learning_rate": 0.00018369808883517212, + "loss": 0.141, + "step": 7871 + }, + { + "epoch": 0.1845320768649238, + "grad_norm": 1.2928329706192017, + "learning_rate": 0.00018369405760993942, + "loss": 0.2595, + "step": 7872 + }, + { + "epoch": 0.18455551843972878, + "grad_norm": 0.4450065493583679, + "learning_rate": 0.00018369002593058046, + "loss": 0.0896, + "step": 7873 + }, + { + "epoch": 0.18457896001453378, + "grad_norm": 0.3974403142929077, + "learning_rate": 0.00018368599379711706, + "loss": 0.1027, + "step": 7874 + }, + { + "epoch": 0.18460240158933877, + "grad_norm": 0.7693203091621399, + "learning_rate": 0.00018368196120957108, + "loss": 0.1688, + "step": 7875 + }, + { + "epoch": 0.18462584316414377, + "grad_norm": 0.09004011005163193, + "learning_rate": 0.00018367792816796445, + "loss": 0.0289, + "step": 7876 + }, + { + "epoch": 0.18464928473894876, + "grad_norm": 0.5130283236503601, + "learning_rate": 0.00018367389467231903, + "loss": 0.1608, + "step": 7877 + }, + { + "epoch": 0.18467272631375375, + "grad_norm": 0.5679873824119568, + "learning_rate": 0.00018366986072265671, + "loss": 0.1345, + "step": 7878 + }, + { + "epoch": 0.18469616788855875, + "grad_norm": 0.29392415285110474, + "learning_rate": 0.00018366582631899935, + "loss": 0.0771, + "step": 7879 + }, + { + "epoch": 0.18471960946336374, + "grad_norm": 0.26370108127593994, + "learning_rate": 0.00018366179146136888, + "loss": 0.047, + "step": 7880 + }, + { + "epoch": 0.18474305103816874, + "grad_norm": 0.6362859010696411, + "learning_rate": 0.00018365775614978715, + "loss": 0.157, + "step": 7881 + }, + { + "epoch": 0.18476649261297373, + "grad_norm": 0.11917565762996674, + "learning_rate": 0.00018365372038427614, + "loss": 0.0414, + "step": 7882 + }, + { + "epoch": 0.18478993418777873, + "grad_norm": 0.4841688871383667, + "learning_rate": 0.0001836496841648576, + "loss": 0.1802, + "step": 7883 + }, + { + "epoch": 0.18481337576258372, + "grad_norm": 0.43360480666160583, + "learning_rate": 0.00018364564749155359, + "loss": 0.0943, + "step": 7884 + }, + { + "epoch": 0.18483681733738871, + "grad_norm": 0.20588679611682892, + "learning_rate": 0.00018364161036438594, + "loss": 0.0426, + "step": 7885 + }, + { + "epoch": 0.1848602589121937, + "grad_norm": 0.23612815141677856, + "learning_rate": 0.00018363757278337656, + "loss": 0.0356, + "step": 7886 + }, + { + "epoch": 0.1848837004869987, + "grad_norm": 0.33269381523132324, + "learning_rate": 0.00018363353474854733, + "loss": 0.0788, + "step": 7887 + }, + { + "epoch": 0.18490714206180373, + "grad_norm": 0.2114061713218689, + "learning_rate": 0.00018362949625992024, + "loss": 0.0468, + "step": 7888 + }, + { + "epoch": 0.18493058363660872, + "grad_norm": 0.6356387734413147, + "learning_rate": 0.0001836254573175171, + "loss": 0.5619, + "step": 7889 + }, + { + "epoch": 0.18495402521141371, + "grad_norm": 0.7915436029434204, + "learning_rate": 0.0001836214179213599, + "loss": 0.0673, + "step": 7890 + }, + { + "epoch": 0.1849774667862187, + "grad_norm": 0.6088650822639465, + "learning_rate": 0.00018361737807147055, + "loss": 0.0972, + "step": 7891 + }, + { + "epoch": 0.1850009083610237, + "grad_norm": 0.12336068600416183, + "learning_rate": 0.00018361333776787092, + "loss": 0.0163, + "step": 7892 + }, + { + "epoch": 0.1850243499358287, + "grad_norm": 0.4732952117919922, + "learning_rate": 0.00018360929701058298, + "loss": 0.0964, + "step": 7893 + }, + { + "epoch": 0.1850477915106337, + "grad_norm": 0.26315170526504517, + "learning_rate": 0.00018360525579962867, + "loss": 0.0621, + "step": 7894 + }, + { + "epoch": 0.18507123308543869, + "grad_norm": 1.3135948181152344, + "learning_rate": 0.00018360121413502984, + "loss": 0.7267, + "step": 7895 + }, + { + "epoch": 0.18509467466024368, + "grad_norm": 0.40903541445732117, + "learning_rate": 0.0001835971720168085, + "loss": 0.104, + "step": 7896 + }, + { + "epoch": 0.18511811623504867, + "grad_norm": 0.3312675356864929, + "learning_rate": 0.0001835931294449866, + "loss": 0.0652, + "step": 7897 + }, + { + "epoch": 0.18514155780985367, + "grad_norm": 0.5384608507156372, + "learning_rate": 0.00018358908641958596, + "loss": 0.6257, + "step": 7898 + }, + { + "epoch": 0.18516499938465866, + "grad_norm": 0.7563415765762329, + "learning_rate": 0.00018358504294062862, + "loss": 0.1886, + "step": 7899 + }, + { + "epoch": 0.18518844095946366, + "grad_norm": 0.8215303421020508, + "learning_rate": 0.00018358099900813647, + "loss": 0.4375, + "step": 7900 + }, + { + "epoch": 0.18521188253426865, + "grad_norm": 0.48281243443489075, + "learning_rate": 0.0001835769546221315, + "loss": 0.7843, + "step": 7901 + }, + { + "epoch": 0.18523532410907365, + "grad_norm": 0.5056141018867493, + "learning_rate": 0.00018357290978263558, + "loss": 0.1321, + "step": 7902 + }, + { + "epoch": 0.18525876568387864, + "grad_norm": 0.4278387427330017, + "learning_rate": 0.00018356886448967074, + "loss": 0.1182, + "step": 7903 + }, + { + "epoch": 0.18528220725868363, + "grad_norm": 0.4721543490886688, + "learning_rate": 0.00018356481874325887, + "loss": 0.1283, + "step": 7904 + }, + { + "epoch": 0.18530564883348863, + "grad_norm": 0.15734028816223145, + "learning_rate": 0.00018356077254342193, + "loss": 0.0222, + "step": 7905 + }, + { + "epoch": 0.18532909040829362, + "grad_norm": 0.6147198677062988, + "learning_rate": 0.00018355672589018192, + "loss": 0.1976, + "step": 7906 + }, + { + "epoch": 0.18535253198309862, + "grad_norm": 0.6628341674804688, + "learning_rate": 0.00018355267878356077, + "loss": 0.1814, + "step": 7907 + }, + { + "epoch": 0.1853759735579036, + "grad_norm": 0.38073569536209106, + "learning_rate": 0.0001835486312235804, + "loss": 0.0798, + "step": 7908 + }, + { + "epoch": 0.1853994151327086, + "grad_norm": 0.13731661438941956, + "learning_rate": 0.00018354458321026286, + "loss": 0.0311, + "step": 7909 + }, + { + "epoch": 0.1854228567075136, + "grad_norm": 0.6043486595153809, + "learning_rate": 0.00018354053474362999, + "loss": 0.8817, + "step": 7910 + }, + { + "epoch": 0.1854462982823186, + "grad_norm": 0.16058611869812012, + "learning_rate": 0.0001835364858237039, + "loss": 0.0346, + "step": 7911 + }, + { + "epoch": 0.1854697398571236, + "grad_norm": 0.5868579745292664, + "learning_rate": 0.00018353243645050644, + "loss": 0.1497, + "step": 7912 + }, + { + "epoch": 0.18549318143192858, + "grad_norm": 0.6947867274284363, + "learning_rate": 0.00018352838662405967, + "loss": 0.143, + "step": 7913 + }, + { + "epoch": 0.1855166230067336, + "grad_norm": 0.471019983291626, + "learning_rate": 0.0001835243363443855, + "loss": 0.1225, + "step": 7914 + }, + { + "epoch": 0.1855400645815386, + "grad_norm": 0.6225643754005432, + "learning_rate": 0.00018352028561150594, + "loss": 0.1559, + "step": 7915 + }, + { + "epoch": 0.1855635061563436, + "grad_norm": 0.9420234560966492, + "learning_rate": 0.00018351623442544295, + "loss": 0.2857, + "step": 7916 + }, + { + "epoch": 0.1855869477311486, + "grad_norm": 0.5089851021766663, + "learning_rate": 0.00018351218278621855, + "loss": 0.1085, + "step": 7917 + }, + { + "epoch": 0.18561038930595358, + "grad_norm": 0.5473029017448425, + "learning_rate": 0.0001835081306938547, + "loss": 0.14, + "step": 7918 + }, + { + "epoch": 0.18563383088075858, + "grad_norm": 0.15851786732673645, + "learning_rate": 0.00018350407814837335, + "loss": 0.0336, + "step": 7919 + }, + { + "epoch": 0.18565727245556357, + "grad_norm": 0.12385442107915878, + "learning_rate": 0.0001835000251497965, + "loss": 0.0235, + "step": 7920 + }, + { + "epoch": 0.18568071403036857, + "grad_norm": 0.5719663500785828, + "learning_rate": 0.00018349597169814623, + "loss": 0.1424, + "step": 7921 + }, + { + "epoch": 0.18570415560517356, + "grad_norm": 0.6703444719314575, + "learning_rate": 0.00018349191779344447, + "loss": 0.1758, + "step": 7922 + }, + { + "epoch": 0.18572759717997855, + "grad_norm": 0.4897042512893677, + "learning_rate": 0.00018348786343571317, + "loss": 0.1306, + "step": 7923 + }, + { + "epoch": 0.18575103875478355, + "grad_norm": 0.6481527090072632, + "learning_rate": 0.00018348380862497437, + "loss": 0.2286, + "step": 7924 + }, + { + "epoch": 0.18577448032958854, + "grad_norm": 0.3219852149486542, + "learning_rate": 0.00018347975336125013, + "loss": 0.0425, + "step": 7925 + }, + { + "epoch": 0.18579792190439354, + "grad_norm": 0.42814505100250244, + "learning_rate": 0.00018347569764456237, + "loss": 0.1008, + "step": 7926 + }, + { + "epoch": 0.18582136347919853, + "grad_norm": 0.23110128939151764, + "learning_rate": 0.00018347164147493314, + "loss": 0.0685, + "step": 7927 + }, + { + "epoch": 0.18584480505400353, + "grad_norm": 0.39626389741897583, + "learning_rate": 0.00018346758485238444, + "loss": 0.0891, + "step": 7928 + }, + { + "epoch": 0.18586824662880852, + "grad_norm": 0.209116593003273, + "learning_rate": 0.00018346352777693828, + "loss": 0.0649, + "step": 7929 + }, + { + "epoch": 0.18589168820361351, + "grad_norm": 0.5875134468078613, + "learning_rate": 0.00018345947024861665, + "loss": 0.1829, + "step": 7930 + }, + { + "epoch": 0.1859151297784185, + "grad_norm": 0.3390915095806122, + "learning_rate": 0.00018345541226744158, + "loss": 0.0425, + "step": 7931 + }, + { + "epoch": 0.1859385713532235, + "grad_norm": 0.5607396364212036, + "learning_rate": 0.00018345135383343513, + "loss": 0.1417, + "step": 7932 + }, + { + "epoch": 0.1859620129280285, + "grad_norm": 0.7433653473854065, + "learning_rate": 0.0001834472949466193, + "loss": 0.1431, + "step": 7933 + }, + { + "epoch": 0.1859854545028335, + "grad_norm": 0.35023775696754456, + "learning_rate": 0.00018344323560701603, + "loss": 0.3829, + "step": 7934 + }, + { + "epoch": 0.18600889607763849, + "grad_norm": 0.725634753704071, + "learning_rate": 0.00018343917581464746, + "loss": 0.2028, + "step": 7935 + }, + { + "epoch": 0.18603233765244348, + "grad_norm": 0.4925183355808258, + "learning_rate": 0.0001834351155695356, + "loss": 0.136, + "step": 7936 + }, + { + "epoch": 0.18605577922724847, + "grad_norm": 0.3943015933036804, + "learning_rate": 0.00018343105487170244, + "loss": 0.0814, + "step": 7937 + }, + { + "epoch": 0.18607922080205347, + "grad_norm": 0.5073683261871338, + "learning_rate": 0.00018342699372117, + "loss": 0.1711, + "step": 7938 + }, + { + "epoch": 0.1861026623768585, + "grad_norm": 0.3550027906894684, + "learning_rate": 0.00018342293211796035, + "loss": 0.0979, + "step": 7939 + }, + { + "epoch": 0.18612610395166349, + "grad_norm": 0.15950919687747955, + "learning_rate": 0.00018341887006209558, + "loss": 0.021, + "step": 7940 + }, + { + "epoch": 0.18614954552646848, + "grad_norm": 0.4254983365535736, + "learning_rate": 0.0001834148075535976, + "loss": 0.1087, + "step": 7941 + }, + { + "epoch": 0.18617298710127347, + "grad_norm": 0.3438663184642792, + "learning_rate": 0.00018341074459248859, + "loss": 0.1209, + "step": 7942 + }, + { + "epoch": 0.18619642867607847, + "grad_norm": 0.3642116189002991, + "learning_rate": 0.0001834066811787905, + "loss": 0.4075, + "step": 7943 + }, + { + "epoch": 0.18621987025088346, + "grad_norm": 0.2862793505191803, + "learning_rate": 0.00018340261731252544, + "loss": 0.1135, + "step": 7944 + }, + { + "epoch": 0.18624331182568846, + "grad_norm": 0.5203298926353455, + "learning_rate": 0.00018339855299371542, + "loss": 0.1151, + "step": 7945 + }, + { + "epoch": 0.18626675340049345, + "grad_norm": 0.7024027705192566, + "learning_rate": 0.0001833944882223825, + "loss": 0.1378, + "step": 7946 + }, + { + "epoch": 0.18629019497529845, + "grad_norm": 0.6762769818305969, + "learning_rate": 0.00018339042299854875, + "loss": 0.1211, + "step": 7947 + }, + { + "epoch": 0.18631363655010344, + "grad_norm": 1.9476046562194824, + "learning_rate": 0.0001833863573222362, + "loss": 0.1471, + "step": 7948 + }, + { + "epoch": 0.18633707812490843, + "grad_norm": 0.7616816759109497, + "learning_rate": 0.00018338229119346696, + "loss": 0.1224, + "step": 7949 + }, + { + "epoch": 0.18636051969971343, + "grad_norm": 0.40653330087661743, + "learning_rate": 0.00018337822461226305, + "loss": 0.0459, + "step": 7950 + }, + { + "epoch": 0.18638396127451842, + "grad_norm": 0.45589032769203186, + "learning_rate": 0.00018337415757864657, + "loss": 0.1485, + "step": 7951 + }, + { + "epoch": 0.18640740284932342, + "grad_norm": 0.42379552125930786, + "learning_rate": 0.00018337009009263954, + "loss": 0.0981, + "step": 7952 + }, + { + "epoch": 0.1864308444241284, + "grad_norm": 0.7994393706321716, + "learning_rate": 0.0001833660221542641, + "loss": 0.0789, + "step": 7953 + }, + { + "epoch": 0.1864542859989334, + "grad_norm": 0.31034421920776367, + "learning_rate": 0.0001833619537635422, + "loss": 0.0771, + "step": 7954 + }, + { + "epoch": 0.1864777275737384, + "grad_norm": 0.389854371547699, + "learning_rate": 0.00018335788492049602, + "loss": 0.1406, + "step": 7955 + }, + { + "epoch": 0.1865011691485434, + "grad_norm": 0.4353603720664978, + "learning_rate": 0.00018335381562514767, + "loss": 0.0771, + "step": 7956 + }, + { + "epoch": 0.1865246107233484, + "grad_norm": 0.4603060483932495, + "learning_rate": 0.0001833497458775191, + "loss": 0.1166, + "step": 7957 + }, + { + "epoch": 0.18654805229815338, + "grad_norm": 0.7463860511779785, + "learning_rate": 0.0001833456756776325, + "loss": 0.2306, + "step": 7958 + }, + { + "epoch": 0.18657149387295838, + "grad_norm": 0.6598794460296631, + "learning_rate": 0.00018334160502550995, + "loss": 0.1481, + "step": 7959 + }, + { + "epoch": 0.18659493544776337, + "grad_norm": 0.6035804152488708, + "learning_rate": 0.00018333753392117346, + "loss": 0.1814, + "step": 7960 + }, + { + "epoch": 0.18661837702256837, + "grad_norm": 0.23134402930736542, + "learning_rate": 0.0001833334623646452, + "loss": 0.0429, + "step": 7961 + }, + { + "epoch": 0.18664181859737336, + "grad_norm": 0.6369250416755676, + "learning_rate": 0.0001833293903559472, + "loss": 0.135, + "step": 7962 + }, + { + "epoch": 0.18666526017217835, + "grad_norm": 1.402462363243103, + "learning_rate": 0.00018332531789510162, + "loss": 0.3005, + "step": 7963 + }, + { + "epoch": 0.18668870174698335, + "grad_norm": 0.618297278881073, + "learning_rate": 0.0001833212449821305, + "loss": 0.081, + "step": 7964 + }, + { + "epoch": 0.18671214332178837, + "grad_norm": 0.3279470205307007, + "learning_rate": 0.00018331717161705594, + "loss": 0.052, + "step": 7965 + }, + { + "epoch": 0.18673558489659337, + "grad_norm": 0.41213443875312805, + "learning_rate": 0.00018331309779990012, + "loss": 0.0846, + "step": 7966 + }, + { + "epoch": 0.18675902647139836, + "grad_norm": 0.7656111717224121, + "learning_rate": 0.00018330902353068504, + "loss": 0.2035, + "step": 7967 + }, + { + "epoch": 0.18678246804620335, + "grad_norm": 0.08758345991373062, + "learning_rate": 0.0001833049488094329, + "loss": 0.0187, + "step": 7968 + }, + { + "epoch": 0.18680590962100835, + "grad_norm": 0.1424657553434372, + "learning_rate": 0.0001833008736361657, + "loss": 0.0246, + "step": 7969 + }, + { + "epoch": 0.18682935119581334, + "grad_norm": 0.24875947833061218, + "learning_rate": 0.00018329679801090568, + "loss": 0.0585, + "step": 7970 + }, + { + "epoch": 0.18685279277061834, + "grad_norm": 0.3663043975830078, + "learning_rate": 0.00018329272193367486, + "loss": 0.117, + "step": 7971 + }, + { + "epoch": 0.18687623434542333, + "grad_norm": 0.3555755615234375, + "learning_rate": 0.0001832886454044954, + "loss": 0.0681, + "step": 7972 + }, + { + "epoch": 0.18689967592022833, + "grad_norm": 1.0338208675384521, + "learning_rate": 0.0001832845684233894, + "loss": 0.1089, + "step": 7973 + }, + { + "epoch": 0.18692311749503332, + "grad_norm": 0.5320590734481812, + "learning_rate": 0.00018328049099037902, + "loss": 0.8827, + "step": 7974 + }, + { + "epoch": 0.18694655906983831, + "grad_norm": 0.7402088046073914, + "learning_rate": 0.0001832764131054863, + "loss": 0.1599, + "step": 7975 + }, + { + "epoch": 0.1869700006446433, + "grad_norm": 0.7613990306854248, + "learning_rate": 0.00018327233476873346, + "loss": 0.1635, + "step": 7976 + }, + { + "epoch": 0.1869934422194483, + "grad_norm": 0.41305553913116455, + "learning_rate": 0.0001832682559801426, + "loss": 0.1767, + "step": 7977 + }, + { + "epoch": 0.1870168837942533, + "grad_norm": 0.3482719957828522, + "learning_rate": 0.0001832641767397358, + "loss": 0.3569, + "step": 7978 + }, + { + "epoch": 0.1870403253690583, + "grad_norm": 0.2788352072238922, + "learning_rate": 0.00018326009704753527, + "loss": 0.0695, + "step": 7979 + }, + { + "epoch": 0.18706376694386329, + "grad_norm": 0.5279846787452698, + "learning_rate": 0.0001832560169035631, + "loss": 0.7169, + "step": 7980 + }, + { + "epoch": 0.18708720851866828, + "grad_norm": 0.6725841164588928, + "learning_rate": 0.00018325193630784142, + "loss": 0.5752, + "step": 7981 + }, + { + "epoch": 0.18711065009347327, + "grad_norm": 0.6013022065162659, + "learning_rate": 0.00018324785526039244, + "loss": 0.1639, + "step": 7982 + }, + { + "epoch": 0.18713409166827827, + "grad_norm": 0.27238941192626953, + "learning_rate": 0.00018324377376123822, + "loss": 0.0705, + "step": 7983 + }, + { + "epoch": 0.18715753324308326, + "grad_norm": 0.4722166657447815, + "learning_rate": 0.00018323969181040096, + "loss": 0.166, + "step": 7984 + }, + { + "epoch": 0.18718097481788826, + "grad_norm": 0.6144513487815857, + "learning_rate": 0.0001832356094079028, + "loss": 0.1787, + "step": 7985 + }, + { + "epoch": 0.18720441639269325, + "grad_norm": 0.45857423543930054, + "learning_rate": 0.00018323152655376585, + "loss": 0.0827, + "step": 7986 + }, + { + "epoch": 0.18722785796749825, + "grad_norm": 0.6392245888710022, + "learning_rate": 0.00018322744324801232, + "loss": 0.3691, + "step": 7987 + }, + { + "epoch": 0.18725129954230324, + "grad_norm": 0.2963353097438812, + "learning_rate": 0.00018322335949066433, + "loss": 0.1058, + "step": 7988 + }, + { + "epoch": 0.18727474111710823, + "grad_norm": 0.39235103130340576, + "learning_rate": 0.00018321927528174405, + "loss": 0.1104, + "step": 7989 + }, + { + "epoch": 0.18729818269191326, + "grad_norm": 0.38193267583847046, + "learning_rate": 0.00018321519062127365, + "loss": 0.054, + "step": 7990 + }, + { + "epoch": 0.18732162426671825, + "grad_norm": 0.6126484274864197, + "learning_rate": 0.0001832111055092753, + "loss": 0.0781, + "step": 7991 + }, + { + "epoch": 0.18734506584152325, + "grad_norm": 0.6126119494438171, + "learning_rate": 0.00018320701994577115, + "loss": 0.1914, + "step": 7992 + }, + { + "epoch": 0.18736850741632824, + "grad_norm": 0.43895360827445984, + "learning_rate": 0.00018320293393078336, + "loss": 0.1253, + "step": 7993 + }, + { + "epoch": 0.18739194899113323, + "grad_norm": 0.4603468179702759, + "learning_rate": 0.00018319884746433412, + "loss": 0.1008, + "step": 7994 + }, + { + "epoch": 0.18741539056593823, + "grad_norm": 0.39412063360214233, + "learning_rate": 0.00018319476054644559, + "loss": 0.0698, + "step": 7995 + }, + { + "epoch": 0.18743883214074322, + "grad_norm": 0.5382204055786133, + "learning_rate": 0.0001831906731771399, + "loss": 0.1052, + "step": 7996 + }, + { + "epoch": 0.18746227371554822, + "grad_norm": 0.1788424402475357, + "learning_rate": 0.00018318658535643935, + "loss": 0.0451, + "step": 7997 + }, + { + "epoch": 0.1874857152903532, + "grad_norm": 0.5106342434883118, + "learning_rate": 0.00018318249708436604, + "loss": 0.1296, + "step": 7998 + }, + { + "epoch": 0.1875091568651582, + "grad_norm": 0.5146291851997375, + "learning_rate": 0.00018317840836094213, + "loss": 0.1592, + "step": 7999 + }, + { + "epoch": 0.1875325984399632, + "grad_norm": 0.6887013912200928, + "learning_rate": 0.00018317431918618983, + "loss": 0.2382, + "step": 8000 + }, + { + "epoch": 0.1875560400147682, + "grad_norm": 0.42018961906433105, + "learning_rate": 0.00018317022956013137, + "loss": 0.1263, + "step": 8001 + }, + { + "epoch": 0.1875794815895732, + "grad_norm": 0.3544459640979767, + "learning_rate": 0.0001831661394827889, + "loss": 0.0669, + "step": 8002 + }, + { + "epoch": 0.18760292316437818, + "grad_norm": 0.7406161427497864, + "learning_rate": 0.0001831620489541846, + "loss": 0.2451, + "step": 8003 + }, + { + "epoch": 0.18762636473918318, + "grad_norm": 0.6257345676422119, + "learning_rate": 0.00018315795797434073, + "loss": 0.1495, + "step": 8004 + }, + { + "epoch": 0.18764980631398817, + "grad_norm": 0.6268408894538879, + "learning_rate": 0.0001831538665432794, + "loss": 0.1845, + "step": 8005 + }, + { + "epoch": 0.18767324788879317, + "grad_norm": 0.5944838523864746, + "learning_rate": 0.00018314977466102285, + "loss": 0.4616, + "step": 8006 + }, + { + "epoch": 0.18769668946359816, + "grad_norm": 0.1699419617652893, + "learning_rate": 0.0001831456823275933, + "loss": 0.0303, + "step": 8007 + }, + { + "epoch": 0.18772013103840315, + "grad_norm": 0.4158686399459839, + "learning_rate": 0.00018314158954301293, + "loss": 0.087, + "step": 8008 + }, + { + "epoch": 0.18774357261320815, + "grad_norm": 0.723851203918457, + "learning_rate": 0.00018313749630730395, + "loss": 0.1882, + "step": 8009 + }, + { + "epoch": 0.18776701418801314, + "grad_norm": 0.22376999258995056, + "learning_rate": 0.00018313340262048858, + "loss": 0.0666, + "step": 8010 + }, + { + "epoch": 0.18779045576281814, + "grad_norm": 0.434334397315979, + "learning_rate": 0.00018312930848258906, + "loss": 0.1106, + "step": 8011 + }, + { + "epoch": 0.18781389733762313, + "grad_norm": 0.39943212270736694, + "learning_rate": 0.00018312521389362755, + "loss": 0.1012, + "step": 8012 + }, + { + "epoch": 0.18783733891242813, + "grad_norm": 0.6000465750694275, + "learning_rate": 0.00018312111885362628, + "loss": 0.0944, + "step": 8013 + }, + { + "epoch": 0.18786078048723312, + "grad_norm": 0.9993109703063965, + "learning_rate": 0.00018311702336260753, + "loss": 0.2202, + "step": 8014 + }, + { + "epoch": 0.18788422206203811, + "grad_norm": 0.5772984027862549, + "learning_rate": 0.00018311292742059342, + "loss": 0.1216, + "step": 8015 + }, + { + "epoch": 0.18790766363684314, + "grad_norm": 0.8773107528686523, + "learning_rate": 0.00018310883102760626, + "loss": 0.6661, + "step": 8016 + }, + { + "epoch": 0.18793110521164813, + "grad_norm": 0.1784355640411377, + "learning_rate": 0.00018310473418366825, + "loss": 0.0386, + "step": 8017 + }, + { + "epoch": 0.18795454678645312, + "grad_norm": 0.5569331049919128, + "learning_rate": 0.00018310063688880158, + "loss": 0.1081, + "step": 8018 + }, + { + "epoch": 0.18797798836125812, + "grad_norm": 0.5340721011161804, + "learning_rate": 0.00018309653914302856, + "loss": 0.1215, + "step": 8019 + }, + { + "epoch": 0.1880014299360631, + "grad_norm": 0.5142626762390137, + "learning_rate": 0.00018309244094637134, + "loss": 0.1104, + "step": 8020 + }, + { + "epoch": 0.1880248715108681, + "grad_norm": 1.2207409143447876, + "learning_rate": 0.00018308834229885222, + "loss": 0.3346, + "step": 8021 + }, + { + "epoch": 0.1880483130856731, + "grad_norm": 0.745625913143158, + "learning_rate": 0.00018308424320049343, + "loss": 0.1594, + "step": 8022 + }, + { + "epoch": 0.1880717546604781, + "grad_norm": 0.20156675577163696, + "learning_rate": 0.0001830801436513172, + "loss": 0.0489, + "step": 8023 + }, + { + "epoch": 0.1880951962352831, + "grad_norm": 0.6603800058364868, + "learning_rate": 0.00018307604365134576, + "loss": 0.1056, + "step": 8024 + }, + { + "epoch": 0.18811863781008809, + "grad_norm": 0.22943605482578278, + "learning_rate": 0.00018307194320060137, + "loss": 0.0475, + "step": 8025 + }, + { + "epoch": 0.18814207938489308, + "grad_norm": 0.6822717189788818, + "learning_rate": 0.0001830678422991063, + "loss": 0.1775, + "step": 8026 + }, + { + "epoch": 0.18816552095969807, + "grad_norm": 0.9846447706222534, + "learning_rate": 0.00018306374094688278, + "loss": 0.1488, + "step": 8027 + }, + { + "epoch": 0.18818896253450307, + "grad_norm": 0.1753820776939392, + "learning_rate": 0.00018305963914395302, + "loss": 0.0558, + "step": 8028 + }, + { + "epoch": 0.18821240410930806, + "grad_norm": 0.34220442175865173, + "learning_rate": 0.00018305553689033937, + "loss": 0.0564, + "step": 8029 + }, + { + "epoch": 0.18823584568411306, + "grad_norm": 0.15252147614955902, + "learning_rate": 0.00018305143418606402, + "loss": 0.0213, + "step": 8030 + }, + { + "epoch": 0.18825928725891805, + "grad_norm": 0.5840504765510559, + "learning_rate": 0.00018304733103114925, + "loss": 0.1461, + "step": 8031 + }, + { + "epoch": 0.18828272883372305, + "grad_norm": 0.124555304646492, + "learning_rate": 0.00018304322742561735, + "loss": 0.0387, + "step": 8032 + }, + { + "epoch": 0.18830617040852804, + "grad_norm": 0.555683434009552, + "learning_rate": 0.00018303912336949053, + "loss": 0.1344, + "step": 8033 + }, + { + "epoch": 0.18832961198333303, + "grad_norm": 0.2758602797985077, + "learning_rate": 0.00018303501886279114, + "loss": 0.0384, + "step": 8034 + }, + { + "epoch": 0.18835305355813803, + "grad_norm": 1.0276501178741455, + "learning_rate": 0.00018303091390554137, + "loss": 0.1625, + "step": 8035 + }, + { + "epoch": 0.18837649513294302, + "grad_norm": 0.9940207600593567, + "learning_rate": 0.00018302680849776354, + "loss": 0.1807, + "step": 8036 + }, + { + "epoch": 0.18839993670774802, + "grad_norm": 0.4631074070930481, + "learning_rate": 0.0001830227026394799, + "loss": 0.1085, + "step": 8037 + }, + { + "epoch": 0.188423378282553, + "grad_norm": 0.2977472245693207, + "learning_rate": 0.00018301859633071274, + "loss": 0.0762, + "step": 8038 + }, + { + "epoch": 0.188446819857358, + "grad_norm": 0.6184436678886414, + "learning_rate": 0.00018301448957148435, + "loss": 0.1264, + "step": 8039 + }, + { + "epoch": 0.188470261432163, + "grad_norm": 0.26255857944488525, + "learning_rate": 0.000183010382361817, + "loss": 0.0276, + "step": 8040 + }, + { + "epoch": 0.18849370300696802, + "grad_norm": 0.2905712425708771, + "learning_rate": 0.000183006274701733, + "loss": 0.0686, + "step": 8041 + }, + { + "epoch": 0.18851714458177302, + "grad_norm": 0.5726886987686157, + "learning_rate": 0.0001830021665912546, + "loss": 0.0895, + "step": 8042 + }, + { + "epoch": 0.188540586156578, + "grad_norm": 0.8017542362213135, + "learning_rate": 0.0001829980580304041, + "loss": 0.9345, + "step": 8043 + }, + { + "epoch": 0.188564027731383, + "grad_norm": 0.4746580123901367, + "learning_rate": 0.00018299394901920383, + "loss": 0.147, + "step": 8044 + }, + { + "epoch": 0.188587469306188, + "grad_norm": 0.4627555310726166, + "learning_rate": 0.00018298983955767603, + "loss": 0.1054, + "step": 8045 + }, + { + "epoch": 0.188610910880993, + "grad_norm": 0.5520317554473877, + "learning_rate": 0.00018298572964584305, + "loss": 0.1275, + "step": 8046 + }, + { + "epoch": 0.188634352455798, + "grad_norm": 0.6809138059616089, + "learning_rate": 0.00018298161928372717, + "loss": 0.1513, + "step": 8047 + }, + { + "epoch": 0.18865779403060298, + "grad_norm": 0.8735474944114685, + "learning_rate": 0.00018297750847135066, + "loss": 0.2352, + "step": 8048 + }, + { + "epoch": 0.18868123560540798, + "grad_norm": 0.7437834739685059, + "learning_rate": 0.0001829733972087359, + "loss": 0.216, + "step": 8049 + }, + { + "epoch": 0.18870467718021297, + "grad_norm": 0.21963782608509064, + "learning_rate": 0.00018296928549590513, + "loss": 0.049, + "step": 8050 + }, + { + "epoch": 0.18872811875501797, + "grad_norm": 0.5155584216117859, + "learning_rate": 0.00018296517333288068, + "loss": 0.1118, + "step": 8051 + }, + { + "epoch": 0.18875156032982296, + "grad_norm": 0.5167748928070068, + "learning_rate": 0.00018296106071968488, + "loss": 0.0939, + "step": 8052 + }, + { + "epoch": 0.18877500190462795, + "grad_norm": 0.5241914987564087, + "learning_rate": 0.00018295694765634003, + "loss": 0.072, + "step": 8053 + }, + { + "epoch": 0.18879844347943295, + "grad_norm": 0.5633573532104492, + "learning_rate": 0.00018295283414286844, + "loss": 0.0996, + "step": 8054 + }, + { + "epoch": 0.18882188505423794, + "grad_norm": 0.2626652419567108, + "learning_rate": 0.00018294872017929245, + "loss": 0.0372, + "step": 8055 + }, + { + "epoch": 0.18884532662904294, + "grad_norm": 1.0041048526763916, + "learning_rate": 0.00018294460576563437, + "loss": 0.2688, + "step": 8056 + }, + { + "epoch": 0.18886876820384793, + "grad_norm": 0.2758469581604004, + "learning_rate": 0.0001829404909019165, + "loss": 0.1654, + "step": 8057 + }, + { + "epoch": 0.18889220977865293, + "grad_norm": 0.14182542264461517, + "learning_rate": 0.00018293637558816124, + "loss": 0.0235, + "step": 8058 + }, + { + "epoch": 0.18891565135345792, + "grad_norm": 0.6538909077644348, + "learning_rate": 0.00018293225982439083, + "loss": 0.2446, + "step": 8059 + }, + { + "epoch": 0.18893909292826291, + "grad_norm": 0.24531802535057068, + "learning_rate": 0.00018292814361062768, + "loss": 0.0426, + "step": 8060 + }, + { + "epoch": 0.1889625345030679, + "grad_norm": 0.7239654064178467, + "learning_rate": 0.00018292402694689407, + "loss": 0.6257, + "step": 8061 + }, + { + "epoch": 0.1889859760778729, + "grad_norm": 0.5105712413787842, + "learning_rate": 0.00018291990983321238, + "loss": 0.1158, + "step": 8062 + }, + { + "epoch": 0.1890094176526779, + "grad_norm": 0.5121939778327942, + "learning_rate": 0.00018291579226960492, + "loss": 0.6654, + "step": 8063 + }, + { + "epoch": 0.1890328592274829, + "grad_norm": 0.5304354429244995, + "learning_rate": 0.00018291167425609403, + "loss": 0.0759, + "step": 8064 + }, + { + "epoch": 0.18905630080228789, + "grad_norm": 0.687252402305603, + "learning_rate": 0.00018290755579270207, + "loss": 0.1506, + "step": 8065 + }, + { + "epoch": 0.18907974237709288, + "grad_norm": 0.2358570247888565, + "learning_rate": 0.00018290343687945136, + "loss": 0.0604, + "step": 8066 + }, + { + "epoch": 0.1891031839518979, + "grad_norm": 0.28472691774368286, + "learning_rate": 0.0001828993175163643, + "loss": 0.0492, + "step": 8067 + }, + { + "epoch": 0.1891266255267029, + "grad_norm": 0.2713087797164917, + "learning_rate": 0.0001828951977034632, + "loss": 0.0633, + "step": 8068 + }, + { + "epoch": 0.1891500671015079, + "grad_norm": 0.3622820973396301, + "learning_rate": 0.00018289107744077042, + "loss": 0.0716, + "step": 8069 + }, + { + "epoch": 0.18917350867631288, + "grad_norm": 0.5480299592018127, + "learning_rate": 0.00018288695672830834, + "loss": 0.123, + "step": 8070 + }, + { + "epoch": 0.18919695025111788, + "grad_norm": 0.3831466734409332, + "learning_rate": 0.0001828828355660993, + "loss": 0.0555, + "step": 8071 + }, + { + "epoch": 0.18922039182592287, + "grad_norm": 0.3847202658653259, + "learning_rate": 0.00018287871395416568, + "loss": 0.0719, + "step": 8072 + }, + { + "epoch": 0.18924383340072787, + "grad_norm": 0.719404399394989, + "learning_rate": 0.00018287459189252978, + "loss": 0.1603, + "step": 8073 + }, + { + "epoch": 0.18926727497553286, + "grad_norm": 0.5788927674293518, + "learning_rate": 0.00018287046938121403, + "loss": 0.1315, + "step": 8074 + }, + { + "epoch": 0.18929071655033786, + "grad_norm": 0.29967406392097473, + "learning_rate": 0.0001828663464202408, + "loss": 0.0455, + "step": 8075 + }, + { + "epoch": 0.18931415812514285, + "grad_norm": 0.4190027713775635, + "learning_rate": 0.00018286222300963246, + "loss": 0.4552, + "step": 8076 + }, + { + "epoch": 0.18933759969994785, + "grad_norm": 0.4232204556465149, + "learning_rate": 0.00018285809914941133, + "loss": 0.5757, + "step": 8077 + }, + { + "epoch": 0.18936104127475284, + "grad_norm": 0.29494979977607727, + "learning_rate": 0.00018285397483959982, + "loss": 0.0685, + "step": 8078 + }, + { + "epoch": 0.18938448284955783, + "grad_norm": 0.2428644895553589, + "learning_rate": 0.00018284985008022037, + "loss": 0.0474, + "step": 8079 + }, + { + "epoch": 0.18940792442436283, + "grad_norm": 0.7015787363052368, + "learning_rate": 0.00018284572487129523, + "loss": 0.185, + "step": 8080 + }, + { + "epoch": 0.18943136599916782, + "grad_norm": 0.2721780240535736, + "learning_rate": 0.0001828415992128469, + "loss": 0.0504, + "step": 8081 + }, + { + "epoch": 0.18945480757397282, + "grad_norm": 0.5798007845878601, + "learning_rate": 0.00018283747310489772, + "loss": 0.1225, + "step": 8082 + }, + { + "epoch": 0.1894782491487778, + "grad_norm": 0.48741617798805237, + "learning_rate": 0.00018283334654747008, + "loss": 0.0822, + "step": 8083 + }, + { + "epoch": 0.1895016907235828, + "grad_norm": 0.3466722369194031, + "learning_rate": 0.0001828292195405864, + "loss": 0.0966, + "step": 8084 + }, + { + "epoch": 0.1895251322983878, + "grad_norm": 1.05417799949646, + "learning_rate": 0.000182825092084269, + "loss": 0.1695, + "step": 8085 + }, + { + "epoch": 0.1895485738731928, + "grad_norm": 0.5247150659561157, + "learning_rate": 0.00018282096417854033, + "loss": 0.1195, + "step": 8086 + }, + { + "epoch": 0.1895720154479978, + "grad_norm": 0.20892253518104553, + "learning_rate": 0.0001828168358234228, + "loss": 0.0411, + "step": 8087 + }, + { + "epoch": 0.18959545702280278, + "grad_norm": 0.5072241425514221, + "learning_rate": 0.00018281270701893875, + "loss": 0.1271, + "step": 8088 + }, + { + "epoch": 0.18961889859760778, + "grad_norm": 0.7243735194206238, + "learning_rate": 0.00018280857776511065, + "loss": 0.096, + "step": 8089 + }, + { + "epoch": 0.18964234017241277, + "grad_norm": 0.33705469965934753, + "learning_rate": 0.00018280444806196089, + "loss": 0.1039, + "step": 8090 + }, + { + "epoch": 0.18966578174721777, + "grad_norm": 0.6160234808921814, + "learning_rate": 0.00018280031790951186, + "loss": 0.0731, + "step": 8091 + }, + { + "epoch": 0.18968922332202276, + "grad_norm": 0.6683740615844727, + "learning_rate": 0.00018279618730778597, + "loss": 0.7891, + "step": 8092 + }, + { + "epoch": 0.18971266489682778, + "grad_norm": 0.5350403785705566, + "learning_rate": 0.00018279205625680565, + "loss": 0.0938, + "step": 8093 + }, + { + "epoch": 0.18973610647163278, + "grad_norm": 0.5347799062728882, + "learning_rate": 0.0001827879247565933, + "loss": 0.1191, + "step": 8094 + }, + { + "epoch": 0.18975954804643777, + "grad_norm": 0.5981982946395874, + "learning_rate": 0.00018278379280717134, + "loss": 0.146, + "step": 8095 + }, + { + "epoch": 0.18978298962124276, + "grad_norm": 0.7508258819580078, + "learning_rate": 0.0001827796604085622, + "loss": 0.2132, + "step": 8096 + }, + { + "epoch": 0.18980643119604776, + "grad_norm": 0.7460482716560364, + "learning_rate": 0.00018277552756078825, + "loss": 0.2359, + "step": 8097 + }, + { + "epoch": 0.18982987277085275, + "grad_norm": 0.5595836639404297, + "learning_rate": 0.000182771394263872, + "loss": 0.1169, + "step": 8098 + }, + { + "epoch": 0.18985331434565775, + "grad_norm": 0.5986171364784241, + "learning_rate": 0.00018276726051783587, + "loss": 0.1808, + "step": 8099 + }, + { + "epoch": 0.18987675592046274, + "grad_norm": 0.39920902252197266, + "learning_rate": 0.00018276312632270223, + "loss": 0.095, + "step": 8100 + }, + { + "epoch": 0.18990019749526774, + "grad_norm": 0.4050222635269165, + "learning_rate": 0.00018275899167849354, + "loss": 0.1359, + "step": 8101 + }, + { + "epoch": 0.18992363907007273, + "grad_norm": 0.21316124498844147, + "learning_rate": 0.00018275485658523222, + "loss": 0.0512, + "step": 8102 + }, + { + "epoch": 0.18994708064487772, + "grad_norm": 0.20612914860248566, + "learning_rate": 0.0001827507210429407, + "loss": 0.0323, + "step": 8103 + }, + { + "epoch": 0.18997052221968272, + "grad_norm": 0.48168325424194336, + "learning_rate": 0.00018274658505164148, + "loss": 0.0484, + "step": 8104 + }, + { + "epoch": 0.1899939637944877, + "grad_norm": 0.5978466868400574, + "learning_rate": 0.00018274244861135697, + "loss": 0.1591, + "step": 8105 + }, + { + "epoch": 0.1900174053692927, + "grad_norm": 0.6051807999610901, + "learning_rate": 0.00018273831172210958, + "loss": 0.1253, + "step": 8106 + }, + { + "epoch": 0.1900408469440977, + "grad_norm": 0.25316327810287476, + "learning_rate": 0.0001827341743839218, + "loss": 0.0449, + "step": 8107 + }, + { + "epoch": 0.1900642885189027, + "grad_norm": 0.5464124083518982, + "learning_rate": 0.00018273003659681605, + "loss": 0.129, + "step": 8108 + }, + { + "epoch": 0.1900877300937077, + "grad_norm": 0.6904551386833191, + "learning_rate": 0.00018272589836081478, + "loss": 0.6925, + "step": 8109 + }, + { + "epoch": 0.19011117166851269, + "grad_norm": 0.5552025437355042, + "learning_rate": 0.00018272175967594048, + "loss": 0.0974, + "step": 8110 + }, + { + "epoch": 0.19013461324331768, + "grad_norm": 0.5517478585243225, + "learning_rate": 0.0001827176205422156, + "loss": 0.1119, + "step": 8111 + }, + { + "epoch": 0.19015805481812267, + "grad_norm": 0.5052191019058228, + "learning_rate": 0.00018271348095966257, + "loss": 0.1553, + "step": 8112 + }, + { + "epoch": 0.19018149639292767, + "grad_norm": 0.5593851208686829, + "learning_rate": 0.00018270934092830387, + "loss": 0.156, + "step": 8113 + }, + { + "epoch": 0.19020493796773266, + "grad_norm": 0.5992788672447205, + "learning_rate": 0.00018270520044816197, + "loss": 0.2451, + "step": 8114 + }, + { + "epoch": 0.19022837954253766, + "grad_norm": 0.13546402752399445, + "learning_rate": 0.0001827010595192593, + "loss": 0.034, + "step": 8115 + }, + { + "epoch": 0.19025182111734265, + "grad_norm": 0.5349294543266296, + "learning_rate": 0.00018269691814161838, + "loss": 0.5521, + "step": 8116 + }, + { + "epoch": 0.19027526269214765, + "grad_norm": 0.33947592973709106, + "learning_rate": 0.00018269277631526165, + "loss": 0.1105, + "step": 8117 + }, + { + "epoch": 0.19029870426695267, + "grad_norm": 0.3994632363319397, + "learning_rate": 0.0001826886340402116, + "loss": 0.0684, + "step": 8118 + }, + { + "epoch": 0.19032214584175766, + "grad_norm": 0.30481892824172974, + "learning_rate": 0.00018268449131649068, + "loss": 0.0701, + "step": 8119 + }, + { + "epoch": 0.19034558741656266, + "grad_norm": 0.23674902319908142, + "learning_rate": 0.0001826803481441214, + "loss": 0.0275, + "step": 8120 + }, + { + "epoch": 0.19036902899136765, + "grad_norm": 0.5418123602867126, + "learning_rate": 0.0001826762045231262, + "loss": 0.0446, + "step": 8121 + }, + { + "epoch": 0.19039247056617264, + "grad_norm": 0.5933468341827393, + "learning_rate": 0.00018267206045352762, + "loss": 0.1566, + "step": 8122 + }, + { + "epoch": 0.19041591214097764, + "grad_norm": 0.1837623566389084, + "learning_rate": 0.00018266791593534812, + "loss": 0.0462, + "step": 8123 + }, + { + "epoch": 0.19043935371578263, + "grad_norm": 0.5802790522575378, + "learning_rate": 0.00018266377096861014, + "loss": 0.1579, + "step": 8124 + }, + { + "epoch": 0.19046279529058763, + "grad_norm": 0.5483971238136292, + "learning_rate": 0.00018265962555333624, + "loss": 0.1573, + "step": 8125 + }, + { + "epoch": 0.19048623686539262, + "grad_norm": 0.758655846118927, + "learning_rate": 0.0001826554796895489, + "loss": 0.1956, + "step": 8126 + }, + { + "epoch": 0.19050967844019762, + "grad_norm": 1.0636143684387207, + "learning_rate": 0.0001826513333772706, + "loss": 0.1526, + "step": 8127 + }, + { + "epoch": 0.1905331200150026, + "grad_norm": 0.23086102306842804, + "learning_rate": 0.0001826471866165238, + "loss": 0.0526, + "step": 8128 + }, + { + "epoch": 0.1905565615898076, + "grad_norm": 0.6348758935928345, + "learning_rate": 0.00018264303940733107, + "loss": 0.1418, + "step": 8129 + }, + { + "epoch": 0.1905800031646126, + "grad_norm": 1.1823760271072388, + "learning_rate": 0.00018263889174971492, + "loss": 0.2604, + "step": 8130 + }, + { + "epoch": 0.1906034447394176, + "grad_norm": 0.4818638563156128, + "learning_rate": 0.00018263474364369777, + "loss": 0.085, + "step": 8131 + }, + { + "epoch": 0.1906268863142226, + "grad_norm": 0.4388597905635834, + "learning_rate": 0.0001826305950893022, + "loss": 0.1006, + "step": 8132 + }, + { + "epoch": 0.19065032788902758, + "grad_norm": 0.28837475180625916, + "learning_rate": 0.0001826264460865507, + "loss": 0.0825, + "step": 8133 + }, + { + "epoch": 0.19067376946383258, + "grad_norm": 0.6392115354537964, + "learning_rate": 0.00018262229663546577, + "loss": 0.2204, + "step": 8134 + }, + { + "epoch": 0.19069721103863757, + "grad_norm": 0.5132818818092346, + "learning_rate": 0.00018261814673606994, + "loss": 0.1085, + "step": 8135 + }, + { + "epoch": 0.19072065261344257, + "grad_norm": 0.21487446129322052, + "learning_rate": 0.00018261399638838573, + "loss": 0.0611, + "step": 8136 + }, + { + "epoch": 0.19074409418824756, + "grad_norm": 0.40039217472076416, + "learning_rate": 0.00018260984559243564, + "loss": 0.1048, + "step": 8137 + }, + { + "epoch": 0.19076753576305255, + "grad_norm": 0.6506914496421814, + "learning_rate": 0.0001826056943482422, + "loss": 0.1542, + "step": 8138 + }, + { + "epoch": 0.19079097733785755, + "grad_norm": 0.8722531795501709, + "learning_rate": 0.00018260154265582794, + "loss": 0.1548, + "step": 8139 + }, + { + "epoch": 0.19081441891266254, + "grad_norm": 0.1470659375190735, + "learning_rate": 0.0001825973905152154, + "loss": 0.0477, + "step": 8140 + }, + { + "epoch": 0.19083786048746754, + "grad_norm": 0.5193430185317993, + "learning_rate": 0.00018259323792642711, + "loss": 0.1123, + "step": 8141 + }, + { + "epoch": 0.19086130206227253, + "grad_norm": 0.8548406958580017, + "learning_rate": 0.00018258908488948556, + "loss": 0.1807, + "step": 8142 + }, + { + "epoch": 0.19088474363707753, + "grad_norm": 0.269748330116272, + "learning_rate": 0.00018258493140441332, + "loss": 0.0537, + "step": 8143 + }, + { + "epoch": 0.19090818521188255, + "grad_norm": 0.32375749945640564, + "learning_rate": 0.00018258077747123293, + "loss": 0.086, + "step": 8144 + }, + { + "epoch": 0.19093162678668754, + "grad_norm": 1.0835992097854614, + "learning_rate": 0.00018257662308996693, + "loss": 0.1097, + "step": 8145 + }, + { + "epoch": 0.19095506836149254, + "grad_norm": 0.2165665626525879, + "learning_rate": 0.00018257246826063782, + "loss": 0.0249, + "step": 8146 + }, + { + "epoch": 0.19097850993629753, + "grad_norm": 0.10156141221523285, + "learning_rate": 0.0001825683129832682, + "loss": 0.0196, + "step": 8147 + }, + { + "epoch": 0.19100195151110252, + "grad_norm": 0.40947604179382324, + "learning_rate": 0.0001825641572578806, + "loss": 0.0805, + "step": 8148 + }, + { + "epoch": 0.19102539308590752, + "grad_norm": 0.5009132623672485, + "learning_rate": 0.00018256000108449752, + "loss": 0.1434, + "step": 8149 + }, + { + "epoch": 0.1910488346607125, + "grad_norm": 0.9145996570587158, + "learning_rate": 0.00018255584446314157, + "loss": 0.2047, + "step": 8150 + }, + { + "epoch": 0.1910722762355175, + "grad_norm": 0.18171733617782593, + "learning_rate": 0.00018255168739383533, + "loss": 0.0511, + "step": 8151 + }, + { + "epoch": 0.1910957178103225, + "grad_norm": 0.5816588997840881, + "learning_rate": 0.00018254752987660125, + "loss": 0.1405, + "step": 8152 + }, + { + "epoch": 0.1911191593851275, + "grad_norm": 0.6590563654899597, + "learning_rate": 0.00018254337191146199, + "loss": 0.1169, + "step": 8153 + }, + { + "epoch": 0.1911426009599325, + "grad_norm": 0.4460527300834656, + "learning_rate": 0.00018253921349844005, + "loss": 0.1246, + "step": 8154 + }, + { + "epoch": 0.19116604253473748, + "grad_norm": 0.5074474811553955, + "learning_rate": 0.00018253505463755804, + "loss": 0.0911, + "step": 8155 + }, + { + "epoch": 0.19118948410954248, + "grad_norm": 1.4319807291030884, + "learning_rate": 0.0001825308953288385, + "loss": 0.1987, + "step": 8156 + }, + { + "epoch": 0.19121292568434747, + "grad_norm": 0.8236105442047119, + "learning_rate": 0.000182526735572304, + "loss": 0.1018, + "step": 8157 + }, + { + "epoch": 0.19123636725915247, + "grad_norm": 0.2815825045108795, + "learning_rate": 0.00018252257536797707, + "loss": 0.0642, + "step": 8158 + }, + { + "epoch": 0.19125980883395746, + "grad_norm": 0.5244436860084534, + "learning_rate": 0.00018251841471588037, + "loss": 0.1458, + "step": 8159 + }, + { + "epoch": 0.19128325040876246, + "grad_norm": 0.4326091408729553, + "learning_rate": 0.00018251425361603645, + "loss": 0.0859, + "step": 8160 + }, + { + "epoch": 0.19130669198356745, + "grad_norm": 0.6013628840446472, + "learning_rate": 0.00018251009206846783, + "loss": 0.1409, + "step": 8161 + }, + { + "epoch": 0.19133013355837244, + "grad_norm": 0.17877842485904694, + "learning_rate": 0.00018250593007319713, + "loss": 0.0444, + "step": 8162 + }, + { + "epoch": 0.19135357513317744, + "grad_norm": 0.6399587392807007, + "learning_rate": 0.00018250176763024694, + "loss": 0.0984, + "step": 8163 + }, + { + "epoch": 0.19137701670798243, + "grad_norm": 0.874876081943512, + "learning_rate": 0.00018249760473963987, + "loss": 0.1863, + "step": 8164 + }, + { + "epoch": 0.19140045828278743, + "grad_norm": 0.7995060682296753, + "learning_rate": 0.00018249344140139847, + "loss": 0.728, + "step": 8165 + }, + { + "epoch": 0.19142389985759242, + "grad_norm": 0.6952192187309265, + "learning_rate": 0.0001824892776155453, + "loss": 0.5693, + "step": 8166 + }, + { + "epoch": 0.19144734143239742, + "grad_norm": 0.7091705203056335, + "learning_rate": 0.00018248511338210302, + "loss": 0.1345, + "step": 8167 + }, + { + "epoch": 0.1914707830072024, + "grad_norm": 0.664176881313324, + "learning_rate": 0.00018248094870109417, + "loss": 0.1469, + "step": 8168 + }, + { + "epoch": 0.19149422458200743, + "grad_norm": 0.48316386342048645, + "learning_rate": 0.0001824767835725414, + "loss": 0.0986, + "step": 8169 + }, + { + "epoch": 0.19151766615681243, + "grad_norm": 0.9024325609207153, + "learning_rate": 0.00018247261799646727, + "loss": 0.1127, + "step": 8170 + }, + { + "epoch": 0.19154110773161742, + "grad_norm": 0.3868875205516815, + "learning_rate": 0.0001824684519728944, + "loss": 0.0483, + "step": 8171 + }, + { + "epoch": 0.19156454930642242, + "grad_norm": 0.5138843655586243, + "learning_rate": 0.00018246428550184538, + "loss": 0.1432, + "step": 8172 + }, + { + "epoch": 0.1915879908812274, + "grad_norm": 0.44383692741394043, + "learning_rate": 0.00018246011858334284, + "loss": 0.0533, + "step": 8173 + }, + { + "epoch": 0.1916114324560324, + "grad_norm": 0.7401746511459351, + "learning_rate": 0.00018245595121740938, + "loss": 0.1502, + "step": 8174 + }, + { + "epoch": 0.1916348740308374, + "grad_norm": 0.2693489193916321, + "learning_rate": 0.00018245178340406762, + "loss": 0.0878, + "step": 8175 + }, + { + "epoch": 0.1916583156056424, + "grad_norm": 0.7748499512672424, + "learning_rate": 0.00018244761514334013, + "loss": 0.1586, + "step": 8176 + }, + { + "epoch": 0.1916817571804474, + "grad_norm": 0.5853844881057739, + "learning_rate": 0.00018244344643524957, + "loss": 0.1042, + "step": 8177 + }, + { + "epoch": 0.19170519875525238, + "grad_norm": 0.5252129435539246, + "learning_rate": 0.0001824392772798186, + "loss": 0.1015, + "step": 8178 + }, + { + "epoch": 0.19172864033005738, + "grad_norm": 0.6511596441268921, + "learning_rate": 0.00018243510767706976, + "loss": 0.1249, + "step": 8179 + }, + { + "epoch": 0.19175208190486237, + "grad_norm": 0.5415927767753601, + "learning_rate": 0.0001824309376270257, + "loss": 0.1359, + "step": 8180 + }, + { + "epoch": 0.19177552347966736, + "grad_norm": 0.4083194434642792, + "learning_rate": 0.00018242676712970903, + "loss": 0.5611, + "step": 8181 + }, + { + "epoch": 0.19179896505447236, + "grad_norm": 0.11064600199460983, + "learning_rate": 0.00018242259618514247, + "loss": 0.0227, + "step": 8182 + }, + { + "epoch": 0.19182240662927735, + "grad_norm": 0.6009275913238525, + "learning_rate": 0.00018241842479334853, + "loss": 0.0908, + "step": 8183 + }, + { + "epoch": 0.19184584820408235, + "grad_norm": 0.4656963348388672, + "learning_rate": 0.00018241425295434992, + "loss": 0.0692, + "step": 8184 + }, + { + "epoch": 0.19186928977888734, + "grad_norm": 0.7356980443000793, + "learning_rate": 0.00018241008066816926, + "loss": 0.1853, + "step": 8185 + }, + { + "epoch": 0.19189273135369234, + "grad_norm": 0.5092816352844238, + "learning_rate": 0.00018240590793482917, + "loss": 0.1083, + "step": 8186 + }, + { + "epoch": 0.19191617292849733, + "grad_norm": 0.37550872564315796, + "learning_rate": 0.00018240173475435234, + "loss": 0.0552, + "step": 8187 + }, + { + "epoch": 0.19193961450330232, + "grad_norm": 0.32253310084342957, + "learning_rate": 0.00018239756112676134, + "loss": 0.045, + "step": 8188 + }, + { + "epoch": 0.19196305607810732, + "grad_norm": 0.7831167578697205, + "learning_rate": 0.00018239338705207888, + "loss": 0.1114, + "step": 8189 + }, + { + "epoch": 0.1919864976529123, + "grad_norm": 0.1787293553352356, + "learning_rate": 0.00018238921253032757, + "loss": 0.0439, + "step": 8190 + }, + { + "epoch": 0.1920099392277173, + "grad_norm": 0.5113260746002197, + "learning_rate": 0.00018238503756153008, + "loss": 0.0874, + "step": 8191 + }, + { + "epoch": 0.1920333808025223, + "grad_norm": 0.5563989281654358, + "learning_rate": 0.0001823808621457091, + "loss": 0.1753, + "step": 8192 + }, + { + "epoch": 0.1920568223773273, + "grad_norm": 0.695446789264679, + "learning_rate": 0.0001823766862828872, + "loss": 0.8011, + "step": 8193 + }, + { + "epoch": 0.1920802639521323, + "grad_norm": 0.35089758038520813, + "learning_rate": 0.00018237250997308707, + "loss": 0.0875, + "step": 8194 + }, + { + "epoch": 0.1921037055269373, + "grad_norm": 0.47805917263031006, + "learning_rate": 0.00018236833321633143, + "loss": 0.0864, + "step": 8195 + }, + { + "epoch": 0.1921271471017423, + "grad_norm": 0.5820274353027344, + "learning_rate": 0.00018236415601264287, + "loss": 0.122, + "step": 8196 + }, + { + "epoch": 0.1921505886765473, + "grad_norm": 0.7293996214866638, + "learning_rate": 0.00018235997836204407, + "loss": 0.6885, + "step": 8197 + }, + { + "epoch": 0.1921740302513523, + "grad_norm": 0.2779168486595154, + "learning_rate": 0.00018235580026455773, + "loss": 0.0551, + "step": 8198 + }, + { + "epoch": 0.1921974718261573, + "grad_norm": 0.6242194771766663, + "learning_rate": 0.00018235162172020652, + "loss": 0.1185, + "step": 8199 + }, + { + "epoch": 0.19222091340096228, + "grad_norm": 0.5174821615219116, + "learning_rate": 0.00018234744272901309, + "loss": 0.1481, + "step": 8200 + }, + { + "epoch": 0.19224435497576728, + "grad_norm": 0.2975151240825653, + "learning_rate": 0.0001823432632910001, + "loss": 0.0675, + "step": 8201 + }, + { + "epoch": 0.19226779655057227, + "grad_norm": 0.7049132585525513, + "learning_rate": 0.00018233908340619027, + "loss": 0.1036, + "step": 8202 + }, + { + "epoch": 0.19229123812537727, + "grad_norm": 0.6268844604492188, + "learning_rate": 0.00018233490307460626, + "loss": 0.1303, + "step": 8203 + }, + { + "epoch": 0.19231467970018226, + "grad_norm": 0.7038108706474304, + "learning_rate": 0.00018233072229627072, + "loss": 0.1167, + "step": 8204 + }, + { + "epoch": 0.19233812127498726, + "grad_norm": 0.4763893187046051, + "learning_rate": 0.00018232654107120638, + "loss": 0.1107, + "step": 8205 + }, + { + "epoch": 0.19236156284979225, + "grad_norm": 0.6500981450080872, + "learning_rate": 0.00018232235939943592, + "loss": 0.6483, + "step": 8206 + }, + { + "epoch": 0.19238500442459724, + "grad_norm": 0.7767022848129272, + "learning_rate": 0.00018231817728098202, + "loss": 0.74, + "step": 8207 + }, + { + "epoch": 0.19240844599940224, + "grad_norm": 0.9962478876113892, + "learning_rate": 0.00018231399471586737, + "loss": 0.1957, + "step": 8208 + }, + { + "epoch": 0.19243188757420723, + "grad_norm": 0.8276905417442322, + "learning_rate": 0.00018230981170411467, + "loss": 0.1404, + "step": 8209 + }, + { + "epoch": 0.19245532914901223, + "grad_norm": 0.5208324790000916, + "learning_rate": 0.00018230562824574661, + "loss": 0.1374, + "step": 8210 + }, + { + "epoch": 0.19247877072381722, + "grad_norm": 0.6988608837127686, + "learning_rate": 0.00018230144434078593, + "loss": 0.1347, + "step": 8211 + }, + { + "epoch": 0.19250221229862222, + "grad_norm": 0.5481346845626831, + "learning_rate": 0.00018229725998925527, + "loss": 0.5088, + "step": 8212 + }, + { + "epoch": 0.1925256538734272, + "grad_norm": 0.211139976978302, + "learning_rate": 0.00018229307519117738, + "loss": 0.0706, + "step": 8213 + }, + { + "epoch": 0.1925490954482322, + "grad_norm": 0.3553377389907837, + "learning_rate": 0.00018228888994657494, + "loss": 0.3557, + "step": 8214 + }, + { + "epoch": 0.1925725370230372, + "grad_norm": 0.19765451550483704, + "learning_rate": 0.00018228470425547068, + "loss": 0.0496, + "step": 8215 + }, + { + "epoch": 0.1925959785978422, + "grad_norm": 0.79488605260849, + "learning_rate": 0.0001822805181178873, + "loss": 0.1927, + "step": 8216 + }, + { + "epoch": 0.1926194201726472, + "grad_norm": 0.6889388561248779, + "learning_rate": 0.0001822763315338475, + "loss": 0.0656, + "step": 8217 + }, + { + "epoch": 0.19264286174745218, + "grad_norm": 0.5948553681373596, + "learning_rate": 0.000182272144503374, + "loss": 0.5463, + "step": 8218 + }, + { + "epoch": 0.19266630332225718, + "grad_norm": 0.7923163771629333, + "learning_rate": 0.00018226795702648959, + "loss": 0.155, + "step": 8219 + }, + { + "epoch": 0.1926897448970622, + "grad_norm": 0.8244780898094177, + "learning_rate": 0.00018226376910321688, + "loss": 0.4126, + "step": 8220 + }, + { + "epoch": 0.1927131864718672, + "grad_norm": 0.24706239998340607, + "learning_rate": 0.00018225958073357865, + "loss": 0.0596, + "step": 8221 + }, + { + "epoch": 0.1927366280466722, + "grad_norm": 0.2752763628959656, + "learning_rate": 0.00018225539191759764, + "loss": 0.0726, + "step": 8222 + }, + { + "epoch": 0.19276006962147718, + "grad_norm": 0.7010564804077148, + "learning_rate": 0.00018225120265529654, + "loss": 0.1632, + "step": 8223 + }, + { + "epoch": 0.19278351119628218, + "grad_norm": 0.25884997844696045, + "learning_rate": 0.00018224701294669812, + "loss": 0.0387, + "step": 8224 + }, + { + "epoch": 0.19280695277108717, + "grad_norm": 0.17796078324317932, + "learning_rate": 0.00018224282279182506, + "loss": 0.0484, + "step": 8225 + }, + { + "epoch": 0.19283039434589216, + "grad_norm": 0.4867846369743347, + "learning_rate": 0.0001822386321907002, + "loss": 0.1093, + "step": 8226 + }, + { + "epoch": 0.19285383592069716, + "grad_norm": 0.579906702041626, + "learning_rate": 0.00018223444114334614, + "loss": 0.1159, + "step": 8227 + }, + { + "epoch": 0.19287727749550215, + "grad_norm": 0.4313749372959137, + "learning_rate": 0.00018223024964978573, + "loss": 0.0809, + "step": 8228 + }, + { + "epoch": 0.19290071907030715, + "grad_norm": 0.7039403319358826, + "learning_rate": 0.00018222605771004165, + "loss": 0.1036, + "step": 8229 + }, + { + "epoch": 0.19292416064511214, + "grad_norm": 0.3251975476741791, + "learning_rate": 0.0001822218653241367, + "loss": 0.5904, + "step": 8230 + }, + { + "epoch": 0.19294760221991714, + "grad_norm": 0.2507440745830536, + "learning_rate": 0.00018221767249209357, + "loss": 0.043, + "step": 8231 + }, + { + "epoch": 0.19297104379472213, + "grad_norm": 0.29956725239753723, + "learning_rate": 0.00018221347921393503, + "loss": 0.0146, + "step": 8232 + }, + { + "epoch": 0.19299448536952712, + "grad_norm": 0.7673323750495911, + "learning_rate": 0.00018220928548968388, + "loss": 0.0987, + "step": 8233 + }, + { + "epoch": 0.19301792694433212, + "grad_norm": 0.5709177255630493, + "learning_rate": 0.00018220509131936278, + "loss": 0.0792, + "step": 8234 + }, + { + "epoch": 0.1930413685191371, + "grad_norm": 0.7380501627922058, + "learning_rate": 0.00018220089670299458, + "loss": 0.1398, + "step": 8235 + }, + { + "epoch": 0.1930648100939421, + "grad_norm": 0.3590962886810303, + "learning_rate": 0.000182196701640602, + "loss": 0.0684, + "step": 8236 + }, + { + "epoch": 0.1930882516687471, + "grad_norm": 0.405613511800766, + "learning_rate": 0.0001821925061322078, + "loss": 0.0568, + "step": 8237 + }, + { + "epoch": 0.1931116932435521, + "grad_norm": 0.4216843843460083, + "learning_rate": 0.00018218831017783476, + "loss": 0.1085, + "step": 8238 + }, + { + "epoch": 0.1931351348183571, + "grad_norm": 0.4737604558467865, + "learning_rate": 0.00018218411377750565, + "loss": 0.117, + "step": 8239 + }, + { + "epoch": 0.19315857639316208, + "grad_norm": 0.820476770401001, + "learning_rate": 0.00018217991693124318, + "loss": 0.131, + "step": 8240 + }, + { + "epoch": 0.19318201796796708, + "grad_norm": 0.3952380120754242, + "learning_rate": 0.00018217571963907017, + "loss": 0.4782, + "step": 8241 + }, + { + "epoch": 0.19320545954277207, + "grad_norm": 0.26548418402671814, + "learning_rate": 0.00018217152190100946, + "loss": 0.0477, + "step": 8242 + }, + { + "epoch": 0.19322890111757707, + "grad_norm": 0.14799880981445312, + "learning_rate": 0.00018216732371708372, + "loss": 0.0529, + "step": 8243 + }, + { + "epoch": 0.19325234269238206, + "grad_norm": 0.2758166790008545, + "learning_rate": 0.0001821631250873158, + "loss": 0.0551, + "step": 8244 + }, + { + "epoch": 0.19327578426718706, + "grad_norm": 0.2818252444267273, + "learning_rate": 0.00018215892601172842, + "loss": 0.04, + "step": 8245 + }, + { + "epoch": 0.19329922584199208, + "grad_norm": 0.5655741691589355, + "learning_rate": 0.0001821547264903444, + "loss": 0.5678, + "step": 8246 + }, + { + "epoch": 0.19332266741679707, + "grad_norm": 0.4251398742198944, + "learning_rate": 0.00018215052652318652, + "loss": 0.5169, + "step": 8247 + }, + { + "epoch": 0.19334610899160207, + "grad_norm": 1.2709320783615112, + "learning_rate": 0.0001821463261102776, + "loss": 0.1462, + "step": 8248 + }, + { + "epoch": 0.19336955056640706, + "grad_norm": 0.3641168177127838, + "learning_rate": 0.00018214212525164038, + "loss": 0.0855, + "step": 8249 + }, + { + "epoch": 0.19339299214121206, + "grad_norm": 0.6760038137435913, + "learning_rate": 0.0001821379239472977, + "loss": 0.1839, + "step": 8250 + }, + { + "epoch": 0.19341643371601705, + "grad_norm": 0.23751705884933472, + "learning_rate": 0.00018213372219727232, + "loss": 0.0382, + "step": 8251 + }, + { + "epoch": 0.19343987529082204, + "grad_norm": 0.0905093252658844, + "learning_rate": 0.00018212952000158704, + "loss": 0.0141, + "step": 8252 + }, + { + "epoch": 0.19346331686562704, + "grad_norm": 0.5764830112457275, + "learning_rate": 0.0001821253173602647, + "loss": 0.8111, + "step": 8253 + }, + { + "epoch": 0.19348675844043203, + "grad_norm": 0.35134539008140564, + "learning_rate": 0.00018212111427332806, + "loss": 0.0726, + "step": 8254 + }, + { + "epoch": 0.19351020001523703, + "grad_norm": 0.4917551875114441, + "learning_rate": 0.00018211691074079996, + "loss": 0.0963, + "step": 8255 + }, + { + "epoch": 0.19353364159004202, + "grad_norm": 0.1367403119802475, + "learning_rate": 0.00018211270676270317, + "loss": 0.0225, + "step": 8256 + }, + { + "epoch": 0.19355708316484702, + "grad_norm": 0.4761177599430084, + "learning_rate": 0.00018210850233906057, + "loss": 0.0835, + "step": 8257 + }, + { + "epoch": 0.193580524739652, + "grad_norm": 0.4341386556625366, + "learning_rate": 0.0001821042974698949, + "loss": 0.0876, + "step": 8258 + }, + { + "epoch": 0.193603966314457, + "grad_norm": 0.37416887283325195, + "learning_rate": 0.00018210009215522899, + "loss": 0.0731, + "step": 8259 + }, + { + "epoch": 0.193627407889262, + "grad_norm": 0.13796155154705048, + "learning_rate": 0.00018209588639508568, + "loss": 0.0263, + "step": 8260 + }, + { + "epoch": 0.193650849464067, + "grad_norm": 0.15616151690483093, + "learning_rate": 0.0001820916801894878, + "loss": 0.0389, + "step": 8261 + }, + { + "epoch": 0.193674291038872, + "grad_norm": 0.29609882831573486, + "learning_rate": 0.00018208747353845814, + "loss": 0.0676, + "step": 8262 + }, + { + "epoch": 0.19369773261367698, + "grad_norm": 1.169240951538086, + "learning_rate": 0.00018208326644201953, + "loss": 0.1259, + "step": 8263 + }, + { + "epoch": 0.19372117418848198, + "grad_norm": 0.5159767270088196, + "learning_rate": 0.00018207905890019484, + "loss": 0.0689, + "step": 8264 + }, + { + "epoch": 0.19374461576328697, + "grad_norm": 0.4926445782184601, + "learning_rate": 0.00018207485091300684, + "loss": 0.1451, + "step": 8265 + }, + { + "epoch": 0.19376805733809196, + "grad_norm": 0.3673040270805359, + "learning_rate": 0.0001820706424804784, + "loss": 0.0717, + "step": 8266 + }, + { + "epoch": 0.19379149891289696, + "grad_norm": 0.47527140378952026, + "learning_rate": 0.00018206643360263236, + "loss": 0.1435, + "step": 8267 + }, + { + "epoch": 0.19381494048770195, + "grad_norm": 0.5338281393051147, + "learning_rate": 0.00018206222427949152, + "loss": 0.113, + "step": 8268 + }, + { + "epoch": 0.19383838206250695, + "grad_norm": 0.6780043244361877, + "learning_rate": 0.00018205801451107875, + "loss": 0.1371, + "step": 8269 + }, + { + "epoch": 0.19386182363731194, + "grad_norm": 0.4254222810268402, + "learning_rate": 0.0001820538042974169, + "loss": 0.1124, + "step": 8270 + }, + { + "epoch": 0.19388526521211694, + "grad_norm": 0.788031280040741, + "learning_rate": 0.0001820495936385288, + "loss": 0.0795, + "step": 8271 + }, + { + "epoch": 0.19390870678692196, + "grad_norm": 0.49286535382270813, + "learning_rate": 0.00018204538253443727, + "loss": 0.1448, + "step": 8272 + }, + { + "epoch": 0.19393214836172695, + "grad_norm": 0.3105529546737671, + "learning_rate": 0.0001820411709851652, + "loss": 0.0684, + "step": 8273 + }, + { + "epoch": 0.19395558993653195, + "grad_norm": 0.6445198059082031, + "learning_rate": 0.00018203695899073546, + "loss": 0.0831, + "step": 8274 + }, + { + "epoch": 0.19397903151133694, + "grad_norm": 0.8080342411994934, + "learning_rate": 0.00018203274655117084, + "loss": 0.2187, + "step": 8275 + }, + { + "epoch": 0.19400247308614194, + "grad_norm": 0.5850443840026855, + "learning_rate": 0.00018202853366649423, + "loss": 0.6297, + "step": 8276 + }, + { + "epoch": 0.19402591466094693, + "grad_norm": 0.4286741018295288, + "learning_rate": 0.00018202432033672853, + "loss": 0.1156, + "step": 8277 + }, + { + "epoch": 0.19404935623575192, + "grad_norm": 0.5837370157241821, + "learning_rate": 0.0001820201065618965, + "loss": 0.1639, + "step": 8278 + }, + { + "epoch": 0.19407279781055692, + "grad_norm": 0.6505921483039856, + "learning_rate": 0.00018201589234202113, + "loss": 0.1775, + "step": 8279 + }, + { + "epoch": 0.1940962393853619, + "grad_norm": 0.2390703409910202, + "learning_rate": 0.00018201167767712518, + "loss": 0.0775, + "step": 8280 + }, + { + "epoch": 0.1941196809601669, + "grad_norm": 0.35667508840560913, + "learning_rate": 0.00018200746256723157, + "loss": 0.1194, + "step": 8281 + }, + { + "epoch": 0.1941431225349719, + "grad_norm": 0.3124512732028961, + "learning_rate": 0.00018200324701236314, + "loss": 0.092, + "step": 8282 + }, + { + "epoch": 0.1941665641097769, + "grad_norm": 0.5052964687347412, + "learning_rate": 0.00018199903101254284, + "loss": 0.1381, + "step": 8283 + }, + { + "epoch": 0.1941900056845819, + "grad_norm": 0.19781756401062012, + "learning_rate": 0.00018199481456779347, + "loss": 0.0896, + "step": 8284 + }, + { + "epoch": 0.19421344725938688, + "grad_norm": 0.5814071893692017, + "learning_rate": 0.0001819905976781379, + "loss": 0.1579, + "step": 8285 + }, + { + "epoch": 0.19423688883419188, + "grad_norm": 0.5026907324790955, + "learning_rate": 0.00018198638034359907, + "loss": 0.0897, + "step": 8286 + }, + { + "epoch": 0.19426033040899687, + "grad_norm": 0.564132571220398, + "learning_rate": 0.00018198216256419983, + "loss": 0.2283, + "step": 8287 + }, + { + "epoch": 0.19428377198380187, + "grad_norm": 0.5003170967102051, + "learning_rate": 0.00018197794433996305, + "loss": 0.0994, + "step": 8288 + }, + { + "epoch": 0.19430721355860686, + "grad_norm": 0.5105032920837402, + "learning_rate": 0.00018197372567091167, + "loss": 0.1391, + "step": 8289 + }, + { + "epoch": 0.19433065513341186, + "grad_norm": 0.515654981136322, + "learning_rate": 0.00018196950655706854, + "loss": 0.1377, + "step": 8290 + }, + { + "epoch": 0.19435409670821685, + "grad_norm": 0.6607921719551086, + "learning_rate": 0.00018196528699845655, + "loss": 0.2057, + "step": 8291 + }, + { + "epoch": 0.19437753828302184, + "grad_norm": 0.6812745332717896, + "learning_rate": 0.00018196106699509864, + "loss": 0.4137, + "step": 8292 + }, + { + "epoch": 0.19440097985782684, + "grad_norm": 0.30327898263931274, + "learning_rate": 0.00018195684654701764, + "loss": 0.0681, + "step": 8293 + }, + { + "epoch": 0.19442442143263183, + "grad_norm": 0.40495795011520386, + "learning_rate": 0.0001819526256542365, + "loss": 0.0684, + "step": 8294 + }, + { + "epoch": 0.19444786300743683, + "grad_norm": 0.4391038119792938, + "learning_rate": 0.0001819484043167781, + "loss": 0.0914, + "step": 8295 + }, + { + "epoch": 0.19447130458224182, + "grad_norm": 0.5046908259391785, + "learning_rate": 0.00018194418253466538, + "loss": 0.1859, + "step": 8296 + }, + { + "epoch": 0.19449474615704684, + "grad_norm": 0.4563748240470886, + "learning_rate": 0.00018193996030792123, + "loss": 0.165, + "step": 8297 + }, + { + "epoch": 0.19451818773185184, + "grad_norm": 0.4584156572818756, + "learning_rate": 0.0001819357376365685, + "loss": 0.1373, + "step": 8298 + }, + { + "epoch": 0.19454162930665683, + "grad_norm": 0.5452895164489746, + "learning_rate": 0.00018193151452063018, + "loss": 0.0816, + "step": 8299 + }, + { + "epoch": 0.19456507088146183, + "grad_norm": 0.3619781732559204, + "learning_rate": 0.00018192729096012914, + "loss": 0.0644, + "step": 8300 + }, + { + "epoch": 0.19458851245626682, + "grad_norm": 0.46335211396217346, + "learning_rate": 0.00018192306695508835, + "loss": 0.0958, + "step": 8301 + }, + { + "epoch": 0.19461195403107182, + "grad_norm": 0.327405720949173, + "learning_rate": 0.0001819188425055307, + "loss": 0.1311, + "step": 8302 + }, + { + "epoch": 0.1946353956058768, + "grad_norm": 0.611291766166687, + "learning_rate": 0.00018191461761147908, + "loss": 0.1339, + "step": 8303 + }, + { + "epoch": 0.1946588371806818, + "grad_norm": 0.5628547668457031, + "learning_rate": 0.00018191039227295647, + "loss": 0.1144, + "step": 8304 + }, + { + "epoch": 0.1946822787554868, + "grad_norm": 0.7026132941246033, + "learning_rate": 0.00018190616648998571, + "loss": 0.178, + "step": 8305 + }, + { + "epoch": 0.1947057203302918, + "grad_norm": 0.41143834590911865, + "learning_rate": 0.00018190194026258984, + "loss": 0.0997, + "step": 8306 + }, + { + "epoch": 0.1947291619050968, + "grad_norm": 0.6273947358131409, + "learning_rate": 0.00018189771359079172, + "loss": 0.7174, + "step": 8307 + }, + { + "epoch": 0.19475260347990178, + "grad_norm": 0.4971364140510559, + "learning_rate": 0.0001818934864746143, + "loss": 0.1508, + "step": 8308 + }, + { + "epoch": 0.19477604505470678, + "grad_norm": 0.31750521063804626, + "learning_rate": 0.00018188925891408054, + "loss": 0.0378, + "step": 8309 + }, + { + "epoch": 0.19479948662951177, + "grad_norm": 0.36047521233558655, + "learning_rate": 0.00018188503090921335, + "loss": 0.0786, + "step": 8310 + }, + { + "epoch": 0.19482292820431676, + "grad_norm": 0.1743033230304718, + "learning_rate": 0.00018188080246003566, + "loss": 0.0383, + "step": 8311 + }, + { + "epoch": 0.19484636977912176, + "grad_norm": 0.41951096057891846, + "learning_rate": 0.00018187657356657044, + "loss": 0.1489, + "step": 8312 + }, + { + "epoch": 0.19486981135392675, + "grad_norm": 0.5278777480125427, + "learning_rate": 0.00018187234422884066, + "loss": 0.1016, + "step": 8313 + }, + { + "epoch": 0.19489325292873175, + "grad_norm": 0.33838707208633423, + "learning_rate": 0.0001818681144468692, + "loss": 0.1031, + "step": 8314 + }, + { + "epoch": 0.19491669450353674, + "grad_norm": 0.513846218585968, + "learning_rate": 0.0001818638842206791, + "loss": 0.1715, + "step": 8315 + }, + { + "epoch": 0.19494013607834174, + "grad_norm": 0.43515849113464355, + "learning_rate": 0.0001818596535502932, + "loss": 0.1321, + "step": 8316 + }, + { + "epoch": 0.19496357765314673, + "grad_norm": 0.40312033891677856, + "learning_rate": 0.00018185542243573456, + "loss": 0.0976, + "step": 8317 + }, + { + "epoch": 0.19498701922795172, + "grad_norm": 0.7631856203079224, + "learning_rate": 0.0001818511908770261, + "loss": 0.1902, + "step": 8318 + }, + { + "epoch": 0.19501046080275672, + "grad_norm": 0.669377863407135, + "learning_rate": 0.0001818469588741907, + "loss": 0.1252, + "step": 8319 + }, + { + "epoch": 0.1950339023775617, + "grad_norm": 0.4334631562232971, + "learning_rate": 0.00018184272642725148, + "loss": 0.1456, + "step": 8320 + }, + { + "epoch": 0.1950573439523667, + "grad_norm": 0.6340333223342896, + "learning_rate": 0.0001818384935362313, + "loss": 0.0807, + "step": 8321 + }, + { + "epoch": 0.1950807855271717, + "grad_norm": 0.5243640542030334, + "learning_rate": 0.00018183426020115318, + "loss": 0.1402, + "step": 8322 + }, + { + "epoch": 0.19510422710197672, + "grad_norm": 0.040109504014253616, + "learning_rate": 0.00018183002642204002, + "loss": 0.0072, + "step": 8323 + }, + { + "epoch": 0.19512766867678172, + "grad_norm": 0.5931088924407959, + "learning_rate": 0.00018182579219891488, + "loss": 0.1588, + "step": 8324 + }, + { + "epoch": 0.1951511102515867, + "grad_norm": 0.5549536347389221, + "learning_rate": 0.00018182155753180065, + "loss": 0.1367, + "step": 8325 + }, + { + "epoch": 0.1951745518263917, + "grad_norm": 0.21490690112113953, + "learning_rate": 0.00018181732242072038, + "loss": 0.0288, + "step": 8326 + }, + { + "epoch": 0.1951979934011967, + "grad_norm": 0.5525678992271423, + "learning_rate": 0.000181813086865697, + "loss": 0.1187, + "step": 8327 + }, + { + "epoch": 0.1952214349760017, + "grad_norm": 0.2725343704223633, + "learning_rate": 0.0001818088508667535, + "loss": 0.086, + "step": 8328 + }, + { + "epoch": 0.1952448765508067, + "grad_norm": 0.44431257247924805, + "learning_rate": 0.0001818046144239129, + "loss": 0.1403, + "step": 8329 + }, + { + "epoch": 0.19526831812561168, + "grad_norm": 0.5843494534492493, + "learning_rate": 0.00018180037753719815, + "loss": 0.1636, + "step": 8330 + }, + { + "epoch": 0.19529175970041668, + "grad_norm": 0.28792041540145874, + "learning_rate": 0.00018179614020663224, + "loss": 0.0692, + "step": 8331 + }, + { + "epoch": 0.19531520127522167, + "grad_norm": 0.53193199634552, + "learning_rate": 0.0001817919024322382, + "loss": 0.1416, + "step": 8332 + }, + { + "epoch": 0.19533864285002667, + "grad_norm": 0.5574225187301636, + "learning_rate": 0.00018178766421403898, + "loss": 0.7746, + "step": 8333 + }, + { + "epoch": 0.19536208442483166, + "grad_norm": 0.49322429299354553, + "learning_rate": 0.00018178342555205757, + "loss": 0.6519, + "step": 8334 + }, + { + "epoch": 0.19538552599963666, + "grad_norm": 0.2295655906200409, + "learning_rate": 0.00018177918644631707, + "loss": 0.0294, + "step": 8335 + }, + { + "epoch": 0.19540896757444165, + "grad_norm": 0.7531381845474243, + "learning_rate": 0.00018177494689684034, + "loss": 0.2523, + "step": 8336 + }, + { + "epoch": 0.19543240914924664, + "grad_norm": 0.5150001049041748, + "learning_rate": 0.00018177070690365046, + "loss": 0.1094, + "step": 8337 + }, + { + "epoch": 0.19545585072405164, + "grad_norm": 0.504675567150116, + "learning_rate": 0.00018176646646677044, + "loss": 0.151, + "step": 8338 + }, + { + "epoch": 0.19547929229885663, + "grad_norm": 0.12805397808551788, + "learning_rate": 0.00018176222558622326, + "loss": 0.0281, + "step": 8339 + }, + { + "epoch": 0.19550273387366163, + "grad_norm": 0.22368714213371277, + "learning_rate": 0.00018175798426203193, + "loss": 0.0431, + "step": 8340 + }, + { + "epoch": 0.19552617544846662, + "grad_norm": 0.16200348734855652, + "learning_rate": 0.0001817537424942195, + "loss": 0.0218, + "step": 8341 + }, + { + "epoch": 0.19554961702327162, + "grad_norm": 0.4245620667934418, + "learning_rate": 0.00018174950028280898, + "loss": 0.1199, + "step": 8342 + }, + { + "epoch": 0.1955730585980766, + "grad_norm": 0.25777682662010193, + "learning_rate": 0.00018174525762782335, + "loss": 0.0347, + "step": 8343 + }, + { + "epoch": 0.1955965001728816, + "grad_norm": 0.47340577840805054, + "learning_rate": 0.00018174101452928567, + "loss": 0.113, + "step": 8344 + }, + { + "epoch": 0.1956199417476866, + "grad_norm": 0.5480863451957703, + "learning_rate": 0.00018173677098721892, + "loss": 0.1013, + "step": 8345 + }, + { + "epoch": 0.1956433833224916, + "grad_norm": 0.43617942929267883, + "learning_rate": 0.00018173252700164617, + "loss": 0.0806, + "step": 8346 + }, + { + "epoch": 0.1956668248972966, + "grad_norm": 0.2849056124687195, + "learning_rate": 0.0001817282825725904, + "loss": 0.0396, + "step": 8347 + }, + { + "epoch": 0.1956902664721016, + "grad_norm": 0.5518757700920105, + "learning_rate": 0.00018172403770007472, + "loss": 0.0863, + "step": 8348 + }, + { + "epoch": 0.1957137080469066, + "grad_norm": 0.11191239953041077, + "learning_rate": 0.0001817197923841221, + "loss": 0.0179, + "step": 8349 + }, + { + "epoch": 0.1957371496217116, + "grad_norm": 0.3686104118824005, + "learning_rate": 0.0001817155466247556, + "loss": 0.061, + "step": 8350 + }, + { + "epoch": 0.1957605911965166, + "grad_norm": 0.36774948239326477, + "learning_rate": 0.00018171130042199821, + "loss": 0.5504, + "step": 8351 + }, + { + "epoch": 0.1957840327713216, + "grad_norm": 0.3111710250377655, + "learning_rate": 0.00018170705377587302, + "loss": 0.0778, + "step": 8352 + }, + { + "epoch": 0.19580747434612658, + "grad_norm": 0.5020843148231506, + "learning_rate": 0.00018170280668640304, + "loss": 0.1201, + "step": 8353 + }, + { + "epoch": 0.19583091592093158, + "grad_norm": 0.5449276566505432, + "learning_rate": 0.00018169855915361137, + "loss": 0.191, + "step": 8354 + }, + { + "epoch": 0.19585435749573657, + "grad_norm": 0.3001125156879425, + "learning_rate": 0.00018169431117752102, + "loss": 0.08, + "step": 8355 + }, + { + "epoch": 0.19587779907054156, + "grad_norm": 0.781292200088501, + "learning_rate": 0.00018169006275815504, + "loss": 0.1156, + "step": 8356 + }, + { + "epoch": 0.19590124064534656, + "grad_norm": 0.5543990731239319, + "learning_rate": 0.00018168581389553645, + "loss": 0.7883, + "step": 8357 + }, + { + "epoch": 0.19592468222015155, + "grad_norm": 0.3475675880908966, + "learning_rate": 0.00018168156458968837, + "loss": 0.0877, + "step": 8358 + }, + { + "epoch": 0.19594812379495655, + "grad_norm": 0.5609899759292603, + "learning_rate": 0.0001816773148406338, + "loss": 0.1028, + "step": 8359 + }, + { + "epoch": 0.19597156536976154, + "grad_norm": 0.12040575593709946, + "learning_rate": 0.00018167306464839583, + "loss": 0.0203, + "step": 8360 + }, + { + "epoch": 0.19599500694456654, + "grad_norm": 0.4277724623680115, + "learning_rate": 0.00018166881401299753, + "loss": 0.0492, + "step": 8361 + }, + { + "epoch": 0.19601844851937153, + "grad_norm": 1.00889253616333, + "learning_rate": 0.00018166456293446196, + "loss": 0.1538, + "step": 8362 + }, + { + "epoch": 0.19604189009417652, + "grad_norm": 0.4544396996498108, + "learning_rate": 0.00018166031141281214, + "loss": 0.1238, + "step": 8363 + }, + { + "epoch": 0.19606533166898152, + "grad_norm": 0.21297599375247955, + "learning_rate": 0.00018165605944807118, + "loss": 0.0513, + "step": 8364 + }, + { + "epoch": 0.1960887732437865, + "grad_norm": 0.5649656653404236, + "learning_rate": 0.00018165180704026213, + "loss": 0.0526, + "step": 8365 + }, + { + "epoch": 0.1961122148185915, + "grad_norm": 0.7209000587463379, + "learning_rate": 0.00018164755418940813, + "loss": 0.2067, + "step": 8366 + }, + { + "epoch": 0.1961356563933965, + "grad_norm": 0.5674104690551758, + "learning_rate": 0.00018164330089553217, + "loss": 0.122, + "step": 8367 + }, + { + "epoch": 0.1961590979682015, + "grad_norm": 0.30145546793937683, + "learning_rate": 0.00018163904715865736, + "loss": 0.0405, + "step": 8368 + }, + { + "epoch": 0.1961825395430065, + "grad_norm": 0.4803714156150818, + "learning_rate": 0.0001816347929788068, + "loss": 0.1147, + "step": 8369 + }, + { + "epoch": 0.19620598111781148, + "grad_norm": 0.4605691134929657, + "learning_rate": 0.00018163053835600356, + "loss": 0.1199, + "step": 8370 + }, + { + "epoch": 0.19622942269261648, + "grad_norm": 0.29481440782546997, + "learning_rate": 0.00018162628329027072, + "loss": 0.0561, + "step": 8371 + }, + { + "epoch": 0.19625286426742147, + "grad_norm": 0.6385668516159058, + "learning_rate": 0.00018162202778163134, + "loss": 0.098, + "step": 8372 + }, + { + "epoch": 0.19627630584222647, + "grad_norm": 0.9552730917930603, + "learning_rate": 0.0001816177718301086, + "loss": 0.179, + "step": 8373 + }, + { + "epoch": 0.1962997474170315, + "grad_norm": 0.34520167112350464, + "learning_rate": 0.00018161351543572552, + "loss": 0.5058, + "step": 8374 + }, + { + "epoch": 0.19632318899183648, + "grad_norm": 0.3784148395061493, + "learning_rate": 0.00018160925859850516, + "loss": 0.0463, + "step": 8375 + }, + { + "epoch": 0.19634663056664148, + "grad_norm": 0.24077960848808289, + "learning_rate": 0.00018160500131847074, + "loss": 0.0402, + "step": 8376 + }, + { + "epoch": 0.19637007214144647, + "grad_norm": 0.49630871415138245, + "learning_rate": 0.00018160074359564523, + "loss": 0.0903, + "step": 8377 + }, + { + "epoch": 0.19639351371625147, + "grad_norm": 0.24958766996860504, + "learning_rate": 0.0001815964854300518, + "loss": 0.0495, + "step": 8378 + }, + { + "epoch": 0.19641695529105646, + "grad_norm": 0.2429221272468567, + "learning_rate": 0.00018159222682171357, + "loss": 0.0441, + "step": 8379 + }, + { + "epoch": 0.19644039686586146, + "grad_norm": 0.3736417591571808, + "learning_rate": 0.0001815879677706536, + "loss": 0.0944, + "step": 8380 + }, + { + "epoch": 0.19646383844066645, + "grad_norm": 0.38496601581573486, + "learning_rate": 0.00018158370827689506, + "loss": 0.1152, + "step": 8381 + }, + { + "epoch": 0.19648728001547144, + "grad_norm": 0.5870498418807983, + "learning_rate": 0.00018157944834046102, + "loss": 0.091, + "step": 8382 + }, + { + "epoch": 0.19651072159027644, + "grad_norm": 0.6165640950202942, + "learning_rate": 0.00018157518796137456, + "loss": 0.1495, + "step": 8383 + }, + { + "epoch": 0.19653416316508143, + "grad_norm": 0.539188802242279, + "learning_rate": 0.00018157092713965887, + "loss": 0.1494, + "step": 8384 + }, + { + "epoch": 0.19655760473988643, + "grad_norm": 0.1985633224248886, + "learning_rate": 0.00018156666587533702, + "loss": 0.037, + "step": 8385 + }, + { + "epoch": 0.19658104631469142, + "grad_norm": 0.43071916699409485, + "learning_rate": 0.00018156240416843214, + "loss": 0.1067, + "step": 8386 + }, + { + "epoch": 0.19660448788949642, + "grad_norm": 0.7623417973518372, + "learning_rate": 0.00018155814201896737, + "loss": 0.128, + "step": 8387 + }, + { + "epoch": 0.1966279294643014, + "grad_norm": 0.8203821182250977, + "learning_rate": 0.00018155387942696583, + "loss": 0.1532, + "step": 8388 + }, + { + "epoch": 0.1966513710391064, + "grad_norm": 0.17002378404140472, + "learning_rate": 0.00018154961639245065, + "loss": 0.0306, + "step": 8389 + }, + { + "epoch": 0.1966748126139114, + "grad_norm": 0.9478054046630859, + "learning_rate": 0.00018154535291544494, + "loss": 0.2232, + "step": 8390 + }, + { + "epoch": 0.1966982541887164, + "grad_norm": 0.551575243473053, + "learning_rate": 0.00018154108899597188, + "loss": 0.0663, + "step": 8391 + }, + { + "epoch": 0.1967216957635214, + "grad_norm": 0.4587637782096863, + "learning_rate": 0.00018153682463405458, + "loss": 0.0571, + "step": 8392 + }, + { + "epoch": 0.19674513733832638, + "grad_norm": 0.5624250769615173, + "learning_rate": 0.00018153255982971615, + "loss": 0.0858, + "step": 8393 + }, + { + "epoch": 0.19676857891313138, + "grad_norm": 0.5188046097755432, + "learning_rate": 0.00018152829458297978, + "loss": 0.1283, + "step": 8394 + }, + { + "epoch": 0.19679202048793637, + "grad_norm": 0.46285924315452576, + "learning_rate": 0.00018152402889386857, + "loss": 0.0608, + "step": 8395 + }, + { + "epoch": 0.19681546206274136, + "grad_norm": 0.21364136040210724, + "learning_rate": 0.0001815197627624057, + "loss": 0.0375, + "step": 8396 + }, + { + "epoch": 0.19683890363754636, + "grad_norm": 0.4520265460014343, + "learning_rate": 0.0001815154961886143, + "loss": 0.0838, + "step": 8397 + }, + { + "epoch": 0.19686234521235135, + "grad_norm": 0.4183449447154999, + "learning_rate": 0.00018151122917251753, + "loss": 0.1259, + "step": 8398 + }, + { + "epoch": 0.19688578678715637, + "grad_norm": 0.4961702823638916, + "learning_rate": 0.00018150696171413854, + "loss": 0.1093, + "step": 8399 + }, + { + "epoch": 0.19690922836196137, + "grad_norm": 0.6854214668273926, + "learning_rate": 0.0001815026938135005, + "loss": 0.1349, + "step": 8400 + }, + { + "epoch": 0.19693266993676636, + "grad_norm": 0.2152298092842102, + "learning_rate": 0.00018149842547062652, + "loss": 0.0456, + "step": 8401 + }, + { + "epoch": 0.19695611151157136, + "grad_norm": 0.3536198139190674, + "learning_rate": 0.0001814941566855398, + "loss": 0.1055, + "step": 8402 + }, + { + "epoch": 0.19697955308637635, + "grad_norm": 0.6270764470100403, + "learning_rate": 0.00018148988745826352, + "loss": 0.1213, + "step": 8403 + }, + { + "epoch": 0.19700299466118135, + "grad_norm": 0.28011444211006165, + "learning_rate": 0.00018148561778882078, + "loss": 0.0434, + "step": 8404 + }, + { + "epoch": 0.19702643623598634, + "grad_norm": 0.8719365000724792, + "learning_rate": 0.00018148134767723481, + "loss": 0.1514, + "step": 8405 + }, + { + "epoch": 0.19704987781079134, + "grad_norm": 0.841310441493988, + "learning_rate": 0.00018147707712352874, + "loss": 0.2337, + "step": 8406 + }, + { + "epoch": 0.19707331938559633, + "grad_norm": 0.8816095590591431, + "learning_rate": 0.00018147280612772576, + "loss": 0.1924, + "step": 8407 + }, + { + "epoch": 0.19709676096040132, + "grad_norm": 0.23224563896656036, + "learning_rate": 0.00018146853468984907, + "loss": 0.0423, + "step": 8408 + }, + { + "epoch": 0.19712020253520632, + "grad_norm": 0.6234045028686523, + "learning_rate": 0.00018146426280992182, + "loss": 0.1398, + "step": 8409 + }, + { + "epoch": 0.1971436441100113, + "grad_norm": 0.6077029705047607, + "learning_rate": 0.00018145999048796717, + "loss": 0.1136, + "step": 8410 + }, + { + "epoch": 0.1971670856848163, + "grad_norm": 0.7703057527542114, + "learning_rate": 0.00018145571772400833, + "loss": 0.1213, + "step": 8411 + }, + { + "epoch": 0.1971905272596213, + "grad_norm": 0.18179906904697418, + "learning_rate": 0.0001814514445180685, + "loss": 0.0393, + "step": 8412 + }, + { + "epoch": 0.1972139688344263, + "grad_norm": 0.6648778915405273, + "learning_rate": 0.00018144717087017077, + "loss": 0.6194, + "step": 8413 + }, + { + "epoch": 0.1972374104092313, + "grad_norm": 0.22399009764194489, + "learning_rate": 0.00018144289678033847, + "loss": 0.0639, + "step": 8414 + }, + { + "epoch": 0.19726085198403628, + "grad_norm": 0.09205186367034912, + "learning_rate": 0.00018143862224859468, + "loss": 0.0131, + "step": 8415 + }, + { + "epoch": 0.19728429355884128, + "grad_norm": 0.3755621314048767, + "learning_rate": 0.00018143434727496265, + "loss": 0.0678, + "step": 8416 + }, + { + "epoch": 0.19730773513364627, + "grad_norm": 0.6144974231719971, + "learning_rate": 0.00018143007185946556, + "loss": 0.1246, + "step": 8417 + }, + { + "epoch": 0.19733117670845127, + "grad_norm": 0.37855827808380127, + "learning_rate": 0.0001814257960021266, + "loss": 0.0546, + "step": 8418 + }, + { + "epoch": 0.19735461828325626, + "grad_norm": 0.5546435117721558, + "learning_rate": 0.00018142151970296902, + "loss": 0.646, + "step": 8419 + }, + { + "epoch": 0.19737805985806126, + "grad_norm": 0.3532821834087372, + "learning_rate": 0.00018141724296201595, + "loss": 0.0515, + "step": 8420 + }, + { + "epoch": 0.19740150143286625, + "grad_norm": 0.3198471963405609, + "learning_rate": 0.00018141296577929062, + "loss": 0.0591, + "step": 8421 + }, + { + "epoch": 0.19742494300767124, + "grad_norm": 0.8459601998329163, + "learning_rate": 0.0001814086881548163, + "loss": 0.1112, + "step": 8422 + }, + { + "epoch": 0.19744838458247624, + "grad_norm": 0.5639168620109558, + "learning_rate": 0.00018140441008861611, + "loss": 0.6085, + "step": 8423 + }, + { + "epoch": 0.19747182615728123, + "grad_norm": 0.4275533854961395, + "learning_rate": 0.00018140013158071333, + "loss": 0.0793, + "step": 8424 + }, + { + "epoch": 0.19749526773208625, + "grad_norm": 0.6942738890647888, + "learning_rate": 0.0001813958526311311, + "loss": 0.5186, + "step": 8425 + }, + { + "epoch": 0.19751870930689125, + "grad_norm": 0.34293410181999207, + "learning_rate": 0.0001813915732398927, + "loss": 0.067, + "step": 8426 + }, + { + "epoch": 0.19754215088169624, + "grad_norm": 0.303539514541626, + "learning_rate": 0.00018138729340702137, + "loss": 0.0535, + "step": 8427 + }, + { + "epoch": 0.19756559245650124, + "grad_norm": 0.34954774379730225, + "learning_rate": 0.00018138301313254024, + "loss": 0.0816, + "step": 8428 + }, + { + "epoch": 0.19758903403130623, + "grad_norm": 0.6433753371238708, + "learning_rate": 0.00018137873241647266, + "loss": 0.1759, + "step": 8429 + }, + { + "epoch": 0.19761247560611123, + "grad_norm": 0.30090001225471497, + "learning_rate": 0.00018137445125884172, + "loss": 0.0768, + "step": 8430 + }, + { + "epoch": 0.19763591718091622, + "grad_norm": 0.5876416563987732, + "learning_rate": 0.00018137016965967079, + "loss": 0.1195, + "step": 8431 + }, + { + "epoch": 0.19765935875572122, + "grad_norm": 0.5659796595573425, + "learning_rate": 0.000181365887618983, + "loss": 0.4776, + "step": 8432 + }, + { + "epoch": 0.1976828003305262, + "grad_norm": 0.4440523684024811, + "learning_rate": 0.00018136160513680162, + "loss": 0.1135, + "step": 8433 + }, + { + "epoch": 0.1977062419053312, + "grad_norm": 0.3880212903022766, + "learning_rate": 0.00018135732221314984, + "loss": 0.0596, + "step": 8434 + }, + { + "epoch": 0.1977296834801362, + "grad_norm": 0.5704978704452515, + "learning_rate": 0.000181353038848051, + "loss": 0.1148, + "step": 8435 + }, + { + "epoch": 0.1977531250549412, + "grad_norm": 0.6539776921272278, + "learning_rate": 0.00018134875504152825, + "loss": 0.0528, + "step": 8436 + }, + { + "epoch": 0.1977765666297462, + "grad_norm": 0.4852358102798462, + "learning_rate": 0.0001813444707936049, + "loss": 0.3106, + "step": 8437 + }, + { + "epoch": 0.19780000820455118, + "grad_norm": 0.8087131977081299, + "learning_rate": 0.00018134018610430414, + "loss": 0.1275, + "step": 8438 + }, + { + "epoch": 0.19782344977935618, + "grad_norm": 0.4278604984283447, + "learning_rate": 0.00018133590097364923, + "loss": 0.1056, + "step": 8439 + }, + { + "epoch": 0.19784689135416117, + "grad_norm": 0.9008005857467651, + "learning_rate": 0.00018133161540166346, + "loss": 0.142, + "step": 8440 + }, + { + "epoch": 0.19787033292896616, + "grad_norm": 0.8373420834541321, + "learning_rate": 0.00018132732938837008, + "loss": 0.1238, + "step": 8441 + }, + { + "epoch": 0.19789377450377116, + "grad_norm": 0.4959527850151062, + "learning_rate": 0.00018132304293379226, + "loss": 0.0785, + "step": 8442 + }, + { + "epoch": 0.19791721607857615, + "grad_norm": 1.512812852859497, + "learning_rate": 0.00018131875603795335, + "loss": 0.1215, + "step": 8443 + }, + { + "epoch": 0.19794065765338115, + "grad_norm": 0.37834692001342773, + "learning_rate": 0.0001813144687008766, + "loss": 0.0364, + "step": 8444 + }, + { + "epoch": 0.19796409922818614, + "grad_norm": 0.6017890572547913, + "learning_rate": 0.00018131018092258525, + "loss": 0.0739, + "step": 8445 + }, + { + "epoch": 0.19798754080299114, + "grad_norm": 0.7016398906707764, + "learning_rate": 0.00018130589270310256, + "loss": 0.7909, + "step": 8446 + }, + { + "epoch": 0.19801098237779613, + "grad_norm": 0.20946544408798218, + "learning_rate": 0.00018130160404245183, + "loss": 0.0222, + "step": 8447 + }, + { + "epoch": 0.19803442395260112, + "grad_norm": 0.5617585778236389, + "learning_rate": 0.0001812973149406563, + "loss": 0.113, + "step": 8448 + }, + { + "epoch": 0.19805786552740612, + "grad_norm": 0.44350960850715637, + "learning_rate": 0.00018129302539773924, + "loss": 0.1089, + "step": 8449 + }, + { + "epoch": 0.1980813071022111, + "grad_norm": 0.2570452392101288, + "learning_rate": 0.00018128873541372397, + "loss": 0.0547, + "step": 8450 + }, + { + "epoch": 0.19810474867701613, + "grad_norm": 1.4527335166931152, + "learning_rate": 0.0001812844449886337, + "loss": 0.2603, + "step": 8451 + }, + { + "epoch": 0.19812819025182113, + "grad_norm": 0.5504583120346069, + "learning_rate": 0.00018128015412249179, + "loss": 0.0783, + "step": 8452 + }, + { + "epoch": 0.19815163182662612, + "grad_norm": 0.36258405447006226, + "learning_rate": 0.00018127586281532142, + "loss": 0.0314, + "step": 8453 + }, + { + "epoch": 0.19817507340143112, + "grad_norm": 0.515236496925354, + "learning_rate": 0.00018127157106714598, + "loss": 0.4309, + "step": 8454 + }, + { + "epoch": 0.1981985149762361, + "grad_norm": 0.553528904914856, + "learning_rate": 0.0001812672788779887, + "loss": 0.0813, + "step": 8455 + }, + { + "epoch": 0.1982219565510411, + "grad_norm": 0.8283093571662903, + "learning_rate": 0.00018126298624787288, + "loss": 0.1456, + "step": 8456 + }, + { + "epoch": 0.1982453981258461, + "grad_norm": 0.5583999156951904, + "learning_rate": 0.00018125869317682177, + "loss": 0.1705, + "step": 8457 + }, + { + "epoch": 0.1982688397006511, + "grad_norm": 0.7637509107589722, + "learning_rate": 0.00018125439966485877, + "loss": 0.1404, + "step": 8458 + }, + { + "epoch": 0.1982922812754561, + "grad_norm": 0.9133081436157227, + "learning_rate": 0.00018125010571200706, + "loss": 0.1761, + "step": 8459 + }, + { + "epoch": 0.19831572285026108, + "grad_norm": 0.987614631652832, + "learning_rate": 0.00018124581131829005, + "loss": 0.2057, + "step": 8460 + }, + { + "epoch": 0.19833916442506608, + "grad_norm": 0.5019797086715698, + "learning_rate": 0.00018124151648373093, + "loss": 0.6708, + "step": 8461 + }, + { + "epoch": 0.19836260599987107, + "grad_norm": 0.5200096964836121, + "learning_rate": 0.00018123722120835307, + "loss": 0.1318, + "step": 8462 + }, + { + "epoch": 0.19838604757467607, + "grad_norm": 0.6147804856300354, + "learning_rate": 0.00018123292549217975, + "loss": 0.1964, + "step": 8463 + }, + { + "epoch": 0.19840948914948106, + "grad_norm": 0.33091992139816284, + "learning_rate": 0.00018122862933523434, + "loss": 0.0578, + "step": 8464 + }, + { + "epoch": 0.19843293072428606, + "grad_norm": 0.9133055806159973, + "learning_rate": 0.00018122433273754006, + "loss": 0.1469, + "step": 8465 + }, + { + "epoch": 0.19845637229909105, + "grad_norm": 0.2421533167362213, + "learning_rate": 0.00018122003569912029, + "loss": 0.0503, + "step": 8466 + }, + { + "epoch": 0.19847981387389604, + "grad_norm": 0.5807479023933411, + "learning_rate": 0.00018121573821999832, + "loss": 0.1382, + "step": 8467 + }, + { + "epoch": 0.19850325544870104, + "grad_norm": 0.5746414661407471, + "learning_rate": 0.00018121144030019748, + "loss": 0.1413, + "step": 8468 + }, + { + "epoch": 0.19852669702350603, + "grad_norm": 0.5604755282402039, + "learning_rate": 0.00018120714193974106, + "loss": 0.136, + "step": 8469 + }, + { + "epoch": 0.19855013859831103, + "grad_norm": 0.15655109286308289, + "learning_rate": 0.00018120284313865242, + "loss": 0.0332, + "step": 8470 + }, + { + "epoch": 0.19857358017311602, + "grad_norm": 0.6162576079368591, + "learning_rate": 0.00018119854389695483, + "loss": 0.1972, + "step": 8471 + }, + { + "epoch": 0.19859702174792102, + "grad_norm": 0.5127005577087402, + "learning_rate": 0.0001811942442146717, + "loss": 0.1354, + "step": 8472 + }, + { + "epoch": 0.198620463322726, + "grad_norm": 0.6040886044502258, + "learning_rate": 0.0001811899440918263, + "loss": 0.8155, + "step": 8473 + }, + { + "epoch": 0.198643904897531, + "grad_norm": 0.5734472870826721, + "learning_rate": 0.00018118564352844198, + "loss": 0.6897, + "step": 8474 + }, + { + "epoch": 0.198667346472336, + "grad_norm": 0.28061479330062866, + "learning_rate": 0.0001811813425245421, + "loss": 0.0834, + "step": 8475 + }, + { + "epoch": 0.19869078804714102, + "grad_norm": 0.28284090757369995, + "learning_rate": 0.00018117704108014997, + "loss": 0.0687, + "step": 8476 + }, + { + "epoch": 0.19871422962194601, + "grad_norm": 0.6300418972969055, + "learning_rate": 0.00018117273919528892, + "loss": 0.7457, + "step": 8477 + }, + { + "epoch": 0.198737671196751, + "grad_norm": 0.7397487163543701, + "learning_rate": 0.00018116843686998227, + "loss": 0.169, + "step": 8478 + }, + { + "epoch": 0.198761112771556, + "grad_norm": 0.4921896457672119, + "learning_rate": 0.00018116413410425346, + "loss": 0.1203, + "step": 8479 + }, + { + "epoch": 0.198784554346361, + "grad_norm": 0.5028294920921326, + "learning_rate": 0.00018115983089812574, + "loss": 0.1372, + "step": 8480 + }, + { + "epoch": 0.198807995921166, + "grad_norm": 0.5161466598510742, + "learning_rate": 0.0001811555272516225, + "loss": 0.1226, + "step": 8481 + }, + { + "epoch": 0.198831437495971, + "grad_norm": 0.42682212591171265, + "learning_rate": 0.0001811512231647671, + "loss": 0.0875, + "step": 8482 + }, + { + "epoch": 0.19885487907077598, + "grad_norm": 0.5055699944496155, + "learning_rate": 0.0001811469186375829, + "loss": 0.1, + "step": 8483 + }, + { + "epoch": 0.19887832064558097, + "grad_norm": 0.2018814980983734, + "learning_rate": 0.0001811426136700932, + "loss": 0.0463, + "step": 8484 + }, + { + "epoch": 0.19890176222038597, + "grad_norm": 0.4616326093673706, + "learning_rate": 0.0001811383082623214, + "loss": 0.1091, + "step": 8485 + }, + { + "epoch": 0.19892520379519096, + "grad_norm": 0.5065670609474182, + "learning_rate": 0.00018113400241429086, + "loss": 0.1246, + "step": 8486 + }, + { + "epoch": 0.19894864536999596, + "grad_norm": 0.5638619065284729, + "learning_rate": 0.00018112969612602494, + "loss": 0.1271, + "step": 8487 + }, + { + "epoch": 0.19897208694480095, + "grad_norm": 0.4377663731575012, + "learning_rate": 0.00018112538939754703, + "loss": 0.1058, + "step": 8488 + }, + { + "epoch": 0.19899552851960595, + "grad_norm": 0.3718479871749878, + "learning_rate": 0.00018112108222888046, + "loss": 0.0736, + "step": 8489 + }, + { + "epoch": 0.19901897009441094, + "grad_norm": 0.276547908782959, + "learning_rate": 0.00018111677462004862, + "loss": 0.0714, + "step": 8490 + }, + { + "epoch": 0.19904241166921594, + "grad_norm": 0.6470116376876831, + "learning_rate": 0.0001811124665710749, + "loss": 0.1703, + "step": 8491 + }, + { + "epoch": 0.19906585324402093, + "grad_norm": 0.16642402112483978, + "learning_rate": 0.0001811081580819826, + "loss": 0.0276, + "step": 8492 + }, + { + "epoch": 0.19908929481882592, + "grad_norm": 0.4676654636859894, + "learning_rate": 0.0001811038491527952, + "loss": 0.1139, + "step": 8493 + }, + { + "epoch": 0.19911273639363092, + "grad_norm": 0.48846301436424255, + "learning_rate": 0.00018109953978353603, + "loss": 0.1172, + "step": 8494 + }, + { + "epoch": 0.1991361779684359, + "grad_norm": 0.14514021575450897, + "learning_rate": 0.00018109522997422842, + "loss": 0.0429, + "step": 8495 + }, + { + "epoch": 0.1991596195432409, + "grad_norm": 0.42219269275665283, + "learning_rate": 0.0001810909197248959, + "loss": 0.1211, + "step": 8496 + }, + { + "epoch": 0.1991830611180459, + "grad_norm": 0.7330620884895325, + "learning_rate": 0.0001810866090355617, + "loss": 0.0746, + "step": 8497 + }, + { + "epoch": 0.1992065026928509, + "grad_norm": 0.8572062849998474, + "learning_rate": 0.00018108229790624934, + "loss": 0.1942, + "step": 8498 + }, + { + "epoch": 0.1992299442676559, + "grad_norm": 0.3556850254535675, + "learning_rate": 0.00018107798633698212, + "loss": 0.0943, + "step": 8499 + }, + { + "epoch": 0.19925338584246088, + "grad_norm": 0.49129652976989746, + "learning_rate": 0.00018107367432778346, + "loss": 0.0937, + "step": 8500 + }, + { + "epoch": 0.19927682741726588, + "grad_norm": 0.4202575385570526, + "learning_rate": 0.00018106936187867677, + "loss": 0.1427, + "step": 8501 + }, + { + "epoch": 0.1993002689920709, + "grad_norm": 0.6553333401679993, + "learning_rate": 0.00018106504898968547, + "loss": 0.1661, + "step": 8502 + }, + { + "epoch": 0.1993237105668759, + "grad_norm": 0.38922497630119324, + "learning_rate": 0.00018106073566083287, + "loss": 0.11, + "step": 8503 + }, + { + "epoch": 0.1993471521416809, + "grad_norm": 0.5213153958320618, + "learning_rate": 0.0001810564218921425, + "loss": 0.1073, + "step": 8504 + }, + { + "epoch": 0.19937059371648588, + "grad_norm": 0.5226986408233643, + "learning_rate": 0.00018105210768363768, + "loss": 0.7981, + "step": 8505 + }, + { + "epoch": 0.19939403529129088, + "grad_norm": 0.6041629910469055, + "learning_rate": 0.00018104779303534184, + "loss": 0.1151, + "step": 8506 + }, + { + "epoch": 0.19941747686609587, + "grad_norm": 0.28046610951423645, + "learning_rate": 0.00018104347794727841, + "loss": 0.0573, + "step": 8507 + }, + { + "epoch": 0.19944091844090087, + "grad_norm": 0.30590638518333435, + "learning_rate": 0.00018103916241947078, + "loss": 0.0678, + "step": 8508 + }, + { + "epoch": 0.19946436001570586, + "grad_norm": 0.6212975978851318, + "learning_rate": 0.00018103484645194237, + "loss": 0.15, + "step": 8509 + }, + { + "epoch": 0.19948780159051085, + "grad_norm": 0.501365602016449, + "learning_rate": 0.0001810305300447166, + "loss": 0.1582, + "step": 8510 + }, + { + "epoch": 0.19951124316531585, + "grad_norm": 0.44901537895202637, + "learning_rate": 0.00018102621319781692, + "loss": 0.1459, + "step": 8511 + }, + { + "epoch": 0.19953468474012084, + "grad_norm": 0.6220707893371582, + "learning_rate": 0.00018102189591126674, + "loss": 0.166, + "step": 8512 + }, + { + "epoch": 0.19955812631492584, + "grad_norm": 0.5201054215431213, + "learning_rate": 0.00018101757818508944, + "loss": 0.1532, + "step": 8513 + }, + { + "epoch": 0.19958156788973083, + "grad_norm": 0.16963103413581848, + "learning_rate": 0.0001810132600193085, + "loss": 0.0309, + "step": 8514 + }, + { + "epoch": 0.19960500946453583, + "grad_norm": 0.48468083143234253, + "learning_rate": 0.00018100894141394734, + "loss": 0.0857, + "step": 8515 + }, + { + "epoch": 0.19962845103934082, + "grad_norm": 0.6443337202072144, + "learning_rate": 0.00018100462236902938, + "loss": 0.7638, + "step": 8516 + }, + { + "epoch": 0.19965189261414582, + "grad_norm": 0.7298932671546936, + "learning_rate": 0.00018100030288457805, + "loss": 0.0845, + "step": 8517 + }, + { + "epoch": 0.1996753341889508, + "grad_norm": 0.3203544318675995, + "learning_rate": 0.0001809959829606168, + "loss": 0.0787, + "step": 8518 + }, + { + "epoch": 0.1996987757637558, + "grad_norm": 0.6929791569709778, + "learning_rate": 0.00018099166259716908, + "loss": 0.1431, + "step": 8519 + }, + { + "epoch": 0.1997222173385608, + "grad_norm": 0.30805766582489014, + "learning_rate": 0.0001809873417942583, + "loss": 0.0646, + "step": 8520 + }, + { + "epoch": 0.1997456589133658, + "grad_norm": 0.43272122740745544, + "learning_rate": 0.00018098302055190798, + "loss": 0.1057, + "step": 8521 + }, + { + "epoch": 0.1997691004881708, + "grad_norm": 0.09567991644144058, + "learning_rate": 0.00018097869887014146, + "loss": 0.0141, + "step": 8522 + }, + { + "epoch": 0.19979254206297578, + "grad_norm": 0.08182520419359207, + "learning_rate": 0.00018097437674898225, + "loss": 0.0152, + "step": 8523 + }, + { + "epoch": 0.19981598363778078, + "grad_norm": 0.6219624876976013, + "learning_rate": 0.0001809700541884538, + "loss": 0.0938, + "step": 8524 + }, + { + "epoch": 0.19983942521258577, + "grad_norm": 0.4275258183479309, + "learning_rate": 0.00018096573118857955, + "loss": 0.0808, + "step": 8525 + }, + { + "epoch": 0.19986286678739076, + "grad_norm": 0.46567854285240173, + "learning_rate": 0.000180961407749383, + "loss": 0.1237, + "step": 8526 + }, + { + "epoch": 0.19988630836219579, + "grad_norm": 0.4827771484851837, + "learning_rate": 0.00018095708387088757, + "loss": 0.103, + "step": 8527 + }, + { + "epoch": 0.19990974993700078, + "grad_norm": 0.8490985631942749, + "learning_rate": 0.0001809527595531167, + "loss": 0.2068, + "step": 8528 + }, + { + "epoch": 0.19993319151180577, + "grad_norm": 0.2563524544239044, + "learning_rate": 0.0001809484347960939, + "loss": 0.0523, + "step": 8529 + }, + { + "epoch": 0.19995663308661077, + "grad_norm": 0.16090741753578186, + "learning_rate": 0.0001809441095998426, + "loss": 0.0287, + "step": 8530 + }, + { + "epoch": 0.19998007466141576, + "grad_norm": 0.5117418766021729, + "learning_rate": 0.0001809397839643863, + "loss": 0.1478, + "step": 8531 + }, + { + "epoch": 0.20000351623622076, + "grad_norm": 0.5637942552566528, + "learning_rate": 0.00018093545788974846, + "loss": 0.1579, + "step": 8532 + }, + { + "epoch": 0.20002695781102575, + "grad_norm": 0.5242329239845276, + "learning_rate": 0.00018093113137595252, + "loss": 0.1388, + "step": 8533 + }, + { + "epoch": 0.20005039938583075, + "grad_norm": 0.2534771263599396, + "learning_rate": 0.00018092680442302202, + "loss": 0.0513, + "step": 8534 + }, + { + "epoch": 0.20007384096063574, + "grad_norm": 0.2755577862262726, + "learning_rate": 0.0001809224770309804, + "loss": 0.0616, + "step": 8535 + }, + { + "epoch": 0.20009728253544073, + "grad_norm": 0.4032500088214874, + "learning_rate": 0.00018091814919985111, + "loss": 0.1295, + "step": 8536 + }, + { + "epoch": 0.20012072411024573, + "grad_norm": 0.36649519205093384, + "learning_rate": 0.0001809138209296577, + "loss": 0.115, + "step": 8537 + }, + { + "epoch": 0.20014416568505072, + "grad_norm": 0.5806264877319336, + "learning_rate": 0.0001809094922204236, + "loss": 0.2138, + "step": 8538 + }, + { + "epoch": 0.20016760725985572, + "grad_norm": 0.520412027835846, + "learning_rate": 0.00018090516307217236, + "loss": 0.1222, + "step": 8539 + }, + { + "epoch": 0.2001910488346607, + "grad_norm": 0.08080889284610748, + "learning_rate": 0.0001809008334849274, + "loss": 0.0138, + "step": 8540 + }, + { + "epoch": 0.2002144904094657, + "grad_norm": 0.3028675317764282, + "learning_rate": 0.00018089650345871222, + "loss": 0.0641, + "step": 8541 + }, + { + "epoch": 0.2002379319842707, + "grad_norm": 0.8251224756240845, + "learning_rate": 0.00018089217299355038, + "loss": 0.1958, + "step": 8542 + }, + { + "epoch": 0.2002613735590757, + "grad_norm": 0.14559166133403778, + "learning_rate": 0.00018088784208946534, + "loss": 0.0275, + "step": 8543 + }, + { + "epoch": 0.2002848151338807, + "grad_norm": 0.4986998736858368, + "learning_rate": 0.00018088351074648055, + "loss": 0.0673, + "step": 8544 + }, + { + "epoch": 0.20030825670868568, + "grad_norm": 0.412029504776001, + "learning_rate": 0.0001808791789646196, + "loss": 0.0632, + "step": 8545 + }, + { + "epoch": 0.20033169828349068, + "grad_norm": 0.16407258808612823, + "learning_rate": 0.00018087484674390592, + "loss": 0.0374, + "step": 8546 + }, + { + "epoch": 0.20035513985829567, + "grad_norm": 0.7571607232093811, + "learning_rate": 0.00018087051408436307, + "loss": 0.1276, + "step": 8547 + }, + { + "epoch": 0.20037858143310067, + "grad_norm": 0.3800141513347626, + "learning_rate": 0.00018086618098601452, + "loss": 0.4867, + "step": 8548 + }, + { + "epoch": 0.20040202300790566, + "grad_norm": 0.5507010817527771, + "learning_rate": 0.00018086184744888381, + "loss": 0.149, + "step": 8549 + }, + { + "epoch": 0.20042546458271066, + "grad_norm": 0.6151976585388184, + "learning_rate": 0.0001808575134729944, + "loss": 0.1421, + "step": 8550 + }, + { + "epoch": 0.20044890615751565, + "grad_norm": 0.40184450149536133, + "learning_rate": 0.0001808531790583699, + "loss": 0.1299, + "step": 8551 + }, + { + "epoch": 0.20047234773232064, + "grad_norm": 0.2305312156677246, + "learning_rate": 0.0001808488442050338, + "loss": 0.072, + "step": 8552 + }, + { + "epoch": 0.20049578930712567, + "grad_norm": 0.5930656790733337, + "learning_rate": 0.00018084450891300952, + "loss": 0.1435, + "step": 8553 + }, + { + "epoch": 0.20051923088193066, + "grad_norm": 0.3873889446258545, + "learning_rate": 0.0001808401731823207, + "loss": 0.0881, + "step": 8554 + }, + { + "epoch": 0.20054267245673565, + "grad_norm": 0.4684235155582428, + "learning_rate": 0.00018083583701299083, + "loss": 0.0673, + "step": 8555 + }, + { + "epoch": 0.20056611403154065, + "grad_norm": 0.6214765310287476, + "learning_rate": 0.00018083150040504344, + "loss": 0.1234, + "step": 8556 + }, + { + "epoch": 0.20058955560634564, + "grad_norm": 0.6748062372207642, + "learning_rate": 0.00018082716335850204, + "loss": 0.5534, + "step": 8557 + }, + { + "epoch": 0.20061299718115064, + "grad_norm": 0.5171861052513123, + "learning_rate": 0.0001808228258733902, + "loss": 0.4729, + "step": 8558 + }, + { + "epoch": 0.20063643875595563, + "grad_norm": 0.4812859892845154, + "learning_rate": 0.00018081848794973144, + "loss": 0.121, + "step": 8559 + }, + { + "epoch": 0.20065988033076063, + "grad_norm": 0.4968555271625519, + "learning_rate": 0.00018081414958754928, + "loss": 0.1293, + "step": 8560 + }, + { + "epoch": 0.20068332190556562, + "grad_norm": 0.4331395924091339, + "learning_rate": 0.00018080981078686726, + "loss": 0.1245, + "step": 8561 + }, + { + "epoch": 0.20070676348037061, + "grad_norm": 0.5860663652420044, + "learning_rate": 0.00018080547154770892, + "loss": 0.135, + "step": 8562 + }, + { + "epoch": 0.2007302050551756, + "grad_norm": 0.29969629645347595, + "learning_rate": 0.00018080113187009787, + "loss": 0.0713, + "step": 8563 + }, + { + "epoch": 0.2007536466299806, + "grad_norm": 0.4114657938480377, + "learning_rate": 0.00018079679175405757, + "loss": 0.077, + "step": 8564 + }, + { + "epoch": 0.2007770882047856, + "grad_norm": 0.5164487957954407, + "learning_rate": 0.00018079245119961162, + "loss": 0.0851, + "step": 8565 + }, + { + "epoch": 0.2008005297795906, + "grad_norm": 0.39338934421539307, + "learning_rate": 0.00018078811020678354, + "loss": 0.0958, + "step": 8566 + }, + { + "epoch": 0.2008239713543956, + "grad_norm": 0.911985456943512, + "learning_rate": 0.00018078376877559693, + "loss": 0.2077, + "step": 8567 + }, + { + "epoch": 0.20084741292920058, + "grad_norm": 0.44346973299980164, + "learning_rate": 0.00018077942690607527, + "loss": 0.1404, + "step": 8568 + }, + { + "epoch": 0.20087085450400557, + "grad_norm": 0.4092410206794739, + "learning_rate": 0.0001807750845982422, + "loss": 0.112, + "step": 8569 + }, + { + "epoch": 0.20089429607881057, + "grad_norm": 0.580609142780304, + "learning_rate": 0.00018077074185212126, + "loss": 0.2025, + "step": 8570 + }, + { + "epoch": 0.20091773765361556, + "grad_norm": 0.18408317863941193, + "learning_rate": 0.000180766398667736, + "loss": 0.0404, + "step": 8571 + }, + { + "epoch": 0.20094117922842056, + "grad_norm": 0.44724684953689575, + "learning_rate": 0.00018076205504511, + "loss": 0.1296, + "step": 8572 + }, + { + "epoch": 0.20096462080322555, + "grad_norm": 0.8008028864860535, + "learning_rate": 0.0001807577109842668, + "loss": 0.1887, + "step": 8573 + }, + { + "epoch": 0.20098806237803055, + "grad_norm": 0.3648020923137665, + "learning_rate": 0.00018075336648523, + "loss": 0.0934, + "step": 8574 + }, + { + "epoch": 0.20101150395283554, + "grad_norm": 0.31737220287323, + "learning_rate": 0.00018074902154802315, + "loss": 0.0795, + "step": 8575 + }, + { + "epoch": 0.20103494552764054, + "grad_norm": 0.6559204459190369, + "learning_rate": 0.00018074467617266985, + "loss": 0.2184, + "step": 8576 + }, + { + "epoch": 0.20105838710244553, + "grad_norm": 0.1668802946805954, + "learning_rate": 0.00018074033035919365, + "loss": 0.0323, + "step": 8577 + }, + { + "epoch": 0.20108182867725055, + "grad_norm": 0.31511080265045166, + "learning_rate": 0.00018073598410761816, + "loss": 0.0581, + "step": 8578 + }, + { + "epoch": 0.20110527025205555, + "grad_norm": 0.5246996283531189, + "learning_rate": 0.00018073163741796693, + "loss": 0.1259, + "step": 8579 + }, + { + "epoch": 0.20112871182686054, + "grad_norm": 0.3417578637599945, + "learning_rate": 0.0001807272902902636, + "loss": 0.0661, + "step": 8580 + }, + { + "epoch": 0.20115215340166553, + "grad_norm": 0.21083462238311768, + "learning_rate": 0.0001807229427245317, + "loss": 0.0412, + "step": 8581 + }, + { + "epoch": 0.20117559497647053, + "grad_norm": 1.0348528623580933, + "learning_rate": 0.00018071859472079487, + "loss": 0.3376, + "step": 8582 + }, + { + "epoch": 0.20119903655127552, + "grad_norm": 0.49961623549461365, + "learning_rate": 0.00018071424627907665, + "loss": 0.107, + "step": 8583 + }, + { + "epoch": 0.20122247812608052, + "grad_norm": 0.21893180906772614, + "learning_rate": 0.00018070989739940065, + "loss": 0.0407, + "step": 8584 + }, + { + "epoch": 0.2012459197008855, + "grad_norm": 0.8455453515052795, + "learning_rate": 0.00018070554808179048, + "loss": 0.2625, + "step": 8585 + }, + { + "epoch": 0.2012693612756905, + "grad_norm": 0.2538416087627411, + "learning_rate": 0.00018070119832626978, + "loss": 0.0504, + "step": 8586 + }, + { + "epoch": 0.2012928028504955, + "grad_norm": 0.7211131453514099, + "learning_rate": 0.00018069684813286206, + "loss": 0.178, + "step": 8587 + }, + { + "epoch": 0.2013162444253005, + "grad_norm": 0.22537554800510406, + "learning_rate": 0.000180692497501591, + "loss": 0.0413, + "step": 8588 + }, + { + "epoch": 0.2013396860001055, + "grad_norm": 0.44573986530303955, + "learning_rate": 0.00018068814643248016, + "loss": 0.0619, + "step": 8589 + }, + { + "epoch": 0.20136312757491048, + "grad_norm": 0.6217576861381531, + "learning_rate": 0.00018068379492555318, + "loss": 0.681, + "step": 8590 + }, + { + "epoch": 0.20138656914971548, + "grad_norm": 0.6600984930992126, + "learning_rate": 0.00018067944298083365, + "loss": 0.1594, + "step": 8591 + }, + { + "epoch": 0.20141001072452047, + "grad_norm": 0.620028555393219, + "learning_rate": 0.00018067509059834522, + "loss": 0.1551, + "step": 8592 + }, + { + "epoch": 0.20143345229932547, + "grad_norm": 0.74853515625, + "learning_rate": 0.0001806707377781115, + "loss": 0.1408, + "step": 8593 + }, + { + "epoch": 0.20145689387413046, + "grad_norm": 0.6208356022834778, + "learning_rate": 0.00018066638452015604, + "loss": 0.1257, + "step": 8594 + }, + { + "epoch": 0.20148033544893545, + "grad_norm": 0.4955652356147766, + "learning_rate": 0.00018066203082450248, + "loss": 0.5802, + "step": 8595 + }, + { + "epoch": 0.20150377702374045, + "grad_norm": 0.7860310673713684, + "learning_rate": 0.00018065767669117453, + "loss": 0.198, + "step": 8596 + }, + { + "epoch": 0.20152721859854544, + "grad_norm": 0.6618441343307495, + "learning_rate": 0.00018065332212019574, + "loss": 0.1661, + "step": 8597 + }, + { + "epoch": 0.20155066017335044, + "grad_norm": 0.4094386100769043, + "learning_rate": 0.00018064896711158975, + "loss": 0.1003, + "step": 8598 + }, + { + "epoch": 0.20157410174815543, + "grad_norm": 0.5612719655036926, + "learning_rate": 0.00018064461166538022, + "loss": 0.1994, + "step": 8599 + }, + { + "epoch": 0.20159754332296043, + "grad_norm": 0.24641098082065582, + "learning_rate": 0.00018064025578159076, + "loss": 0.0571, + "step": 8600 + }, + { + "epoch": 0.20162098489776542, + "grad_norm": 0.33319395780563354, + "learning_rate": 0.00018063589946024498, + "loss": 0.0651, + "step": 8601 + }, + { + "epoch": 0.20164442647257041, + "grad_norm": 0.28761979937553406, + "learning_rate": 0.00018063154270136657, + "loss": 0.0673, + "step": 8602 + }, + { + "epoch": 0.2016678680473754, + "grad_norm": 0.6472399830818176, + "learning_rate": 0.00018062718550497913, + "loss": 0.2582, + "step": 8603 + }, + { + "epoch": 0.20169130962218043, + "grad_norm": 0.7300455570220947, + "learning_rate": 0.0001806228278711063, + "loss": 0.2238, + "step": 8604 + }, + { + "epoch": 0.20171475119698543, + "grad_norm": 0.7252531051635742, + "learning_rate": 0.00018061846979977174, + "loss": 0.2038, + "step": 8605 + }, + { + "epoch": 0.20173819277179042, + "grad_norm": 0.09297958016395569, + "learning_rate": 0.0001806141112909991, + "loss": 0.0235, + "step": 8606 + }, + { + "epoch": 0.20176163434659541, + "grad_norm": 0.46465983986854553, + "learning_rate": 0.00018060975234481202, + "loss": 0.1457, + "step": 8607 + }, + { + "epoch": 0.2017850759214004, + "grad_norm": 0.5742157101631165, + "learning_rate": 0.0001806053929612342, + "loss": 0.1528, + "step": 8608 + }, + { + "epoch": 0.2018085174962054, + "grad_norm": 0.7428197860717773, + "learning_rate": 0.0001806010331402892, + "loss": 0.1272, + "step": 8609 + }, + { + "epoch": 0.2018319590710104, + "grad_norm": 0.6866782307624817, + "learning_rate": 0.00018059667288200076, + "loss": 0.8313, + "step": 8610 + }, + { + "epoch": 0.2018554006458154, + "grad_norm": 0.21750208735466003, + "learning_rate": 0.00018059231218639247, + "loss": 0.0482, + "step": 8611 + }, + { + "epoch": 0.20187884222062039, + "grad_norm": 0.15468671917915344, + "learning_rate": 0.00018058795105348806, + "loss": 0.0407, + "step": 8612 + }, + { + "epoch": 0.20190228379542538, + "grad_norm": 0.19418483972549438, + "learning_rate": 0.00018058358948331115, + "loss": 0.0496, + "step": 8613 + }, + { + "epoch": 0.20192572537023037, + "grad_norm": 0.6855537295341492, + "learning_rate": 0.00018057922747588543, + "loss": 0.172, + "step": 8614 + }, + { + "epoch": 0.20194916694503537, + "grad_norm": 0.5402629971504211, + "learning_rate": 0.00018057486503123452, + "loss": 0.2133, + "step": 8615 + }, + { + "epoch": 0.20197260851984036, + "grad_norm": 0.39298132061958313, + "learning_rate": 0.00018057050214938215, + "loss": 0.1016, + "step": 8616 + }, + { + "epoch": 0.20199605009464536, + "grad_norm": 0.5922165513038635, + "learning_rate": 0.00018056613883035198, + "loss": 0.1923, + "step": 8617 + }, + { + "epoch": 0.20201949166945035, + "grad_norm": 0.4680246114730835, + "learning_rate": 0.00018056177507416767, + "loss": 0.6161, + "step": 8618 + }, + { + "epoch": 0.20204293324425535, + "grad_norm": 0.5436411499977112, + "learning_rate": 0.00018055741088085291, + "loss": 0.1253, + "step": 8619 + }, + { + "epoch": 0.20206637481906034, + "grad_norm": 0.7532602548599243, + "learning_rate": 0.00018055304625043137, + "loss": 0.2059, + "step": 8620 + }, + { + "epoch": 0.20208981639386533, + "grad_norm": 0.41715312004089355, + "learning_rate": 0.00018054868118292674, + "loss": 0.0752, + "step": 8621 + }, + { + "epoch": 0.20211325796867033, + "grad_norm": 0.1998397409915924, + "learning_rate": 0.0001805443156783627, + "loss": 0.0365, + "step": 8622 + }, + { + "epoch": 0.20213669954347532, + "grad_norm": 0.7986249327659607, + "learning_rate": 0.0001805399497367629, + "loss": 0.1553, + "step": 8623 + }, + { + "epoch": 0.20216014111828032, + "grad_norm": 0.745429277420044, + "learning_rate": 0.0001805355833581511, + "loss": 0.1845, + "step": 8624 + }, + { + "epoch": 0.2021835826930853, + "grad_norm": 0.6598851084709167, + "learning_rate": 0.00018053121654255095, + "loss": 0.1575, + "step": 8625 + }, + { + "epoch": 0.2022070242678903, + "grad_norm": 0.42436254024505615, + "learning_rate": 0.00018052684928998618, + "loss": 0.1157, + "step": 8626 + }, + { + "epoch": 0.2022304658426953, + "grad_norm": 0.5402257442474365, + "learning_rate": 0.00018052248160048046, + "loss": 0.1278, + "step": 8627 + }, + { + "epoch": 0.2022539074175003, + "grad_norm": 0.4107092022895813, + "learning_rate": 0.00018051811347405747, + "loss": 0.1244, + "step": 8628 + }, + { + "epoch": 0.2022773489923053, + "grad_norm": 0.27401477098464966, + "learning_rate": 0.00018051374491074093, + "loss": 0.3542, + "step": 8629 + }, + { + "epoch": 0.2023007905671103, + "grad_norm": 0.5203835964202881, + "learning_rate": 0.00018050937591055456, + "loss": 0.1267, + "step": 8630 + }, + { + "epoch": 0.2023242321419153, + "grad_norm": 0.22134923934936523, + "learning_rate": 0.00018050500647352204, + "loss": 0.0574, + "step": 8631 + }, + { + "epoch": 0.2023476737167203, + "grad_norm": 0.38208743929862976, + "learning_rate": 0.0001805006365996671, + "loss": 0.5065, + "step": 8632 + }, + { + "epoch": 0.2023711152915253, + "grad_norm": 0.5471987128257751, + "learning_rate": 0.00018049626628901345, + "loss": 0.1049, + "step": 8633 + }, + { + "epoch": 0.2023945568663303, + "grad_norm": 0.23511499166488647, + "learning_rate": 0.00018049189554158478, + "loss": 0.0524, + "step": 8634 + }, + { + "epoch": 0.20241799844113528, + "grad_norm": 0.7164549231529236, + "learning_rate": 0.00018048752435740484, + "loss": 0.1836, + "step": 8635 + }, + { + "epoch": 0.20244144001594028, + "grad_norm": 0.42564404010772705, + "learning_rate": 0.00018048315273649733, + "loss": 0.0986, + "step": 8636 + }, + { + "epoch": 0.20246488159074527, + "grad_norm": 0.8629196882247925, + "learning_rate": 0.00018047878067888596, + "loss": 0.1611, + "step": 8637 + }, + { + "epoch": 0.20248832316555027, + "grad_norm": 0.5192515254020691, + "learning_rate": 0.00018047440818459444, + "loss": 0.0713, + "step": 8638 + }, + { + "epoch": 0.20251176474035526, + "grad_norm": 0.1638125479221344, + "learning_rate": 0.00018047003525364653, + "loss": 0.0339, + "step": 8639 + }, + { + "epoch": 0.20253520631516025, + "grad_norm": 0.479766845703125, + "learning_rate": 0.00018046566188606596, + "loss": 0.1217, + "step": 8640 + }, + { + "epoch": 0.20255864788996525, + "grad_norm": 0.18593743443489075, + "learning_rate": 0.00018046128808187645, + "loss": 0.0425, + "step": 8641 + }, + { + "epoch": 0.20258208946477024, + "grad_norm": 0.5519899129867554, + "learning_rate": 0.00018045691384110173, + "loss": 0.1831, + "step": 8642 + }, + { + "epoch": 0.20260553103957524, + "grad_norm": 0.44976285099983215, + "learning_rate": 0.00018045253916376553, + "loss": 0.4971, + "step": 8643 + }, + { + "epoch": 0.20262897261438023, + "grad_norm": 0.7716842293739319, + "learning_rate": 0.00018044816404989158, + "loss": 0.1244, + "step": 8644 + }, + { + "epoch": 0.20265241418918523, + "grad_norm": 0.2174830138683319, + "learning_rate": 0.00018044378849950364, + "loss": 0.0398, + "step": 8645 + }, + { + "epoch": 0.20267585576399022, + "grad_norm": 0.8582330346107483, + "learning_rate": 0.00018043941251262542, + "loss": 0.218, + "step": 8646 + }, + { + "epoch": 0.20269929733879521, + "grad_norm": 1.0377960205078125, + "learning_rate": 0.0001804350360892807, + "loss": 0.2515, + "step": 8647 + }, + { + "epoch": 0.2027227389136002, + "grad_norm": 0.36170390248298645, + "learning_rate": 0.00018043065922949324, + "loss": 0.0991, + "step": 8648 + }, + { + "epoch": 0.2027461804884052, + "grad_norm": 0.44305601716041565, + "learning_rate": 0.00018042628193328672, + "loss": 0.0766, + "step": 8649 + }, + { + "epoch": 0.2027696220632102, + "grad_norm": 0.4952988028526306, + "learning_rate": 0.00018042190420068495, + "loss": 0.1867, + "step": 8650 + }, + { + "epoch": 0.2027930636380152, + "grad_norm": 0.5620678067207336, + "learning_rate": 0.00018041752603171167, + "loss": 0.143, + "step": 8651 + }, + { + "epoch": 0.2028165052128202, + "grad_norm": 0.5808992385864258, + "learning_rate": 0.0001804131474263906, + "loss": 0.131, + "step": 8652 + }, + { + "epoch": 0.20283994678762518, + "grad_norm": 0.2687598168849945, + "learning_rate": 0.00018040876838474557, + "loss": 0.0646, + "step": 8653 + }, + { + "epoch": 0.20286338836243017, + "grad_norm": 0.4238126873970032, + "learning_rate": 0.0001804043889068003, + "loss": 0.1238, + "step": 8654 + }, + { + "epoch": 0.2028868299372352, + "grad_norm": 0.40605807304382324, + "learning_rate": 0.00018040000899257854, + "loss": 0.0606, + "step": 8655 + }, + { + "epoch": 0.2029102715120402, + "grad_norm": 0.26570644974708557, + "learning_rate": 0.00018039562864210408, + "loss": 0.0996, + "step": 8656 + }, + { + "epoch": 0.20293371308684519, + "grad_norm": 0.4981594979763031, + "learning_rate": 0.00018039124785540066, + "loss": 0.1878, + "step": 8657 + }, + { + "epoch": 0.20295715466165018, + "grad_norm": 0.3953964412212372, + "learning_rate": 0.0001803868666324921, + "loss": 0.1197, + "step": 8658 + }, + { + "epoch": 0.20298059623645517, + "grad_norm": 0.2895045876502991, + "learning_rate": 0.0001803824849734021, + "loss": 0.0368, + "step": 8659 + }, + { + "epoch": 0.20300403781126017, + "grad_norm": 0.7215723991394043, + "learning_rate": 0.00018037810287815454, + "loss": 0.1078, + "step": 8660 + }, + { + "epoch": 0.20302747938606516, + "grad_norm": 0.236420139670372, + "learning_rate": 0.00018037372034677306, + "loss": 0.0337, + "step": 8661 + }, + { + "epoch": 0.20305092096087016, + "grad_norm": 0.1889704018831253, + "learning_rate": 0.00018036933737928157, + "loss": 0.0329, + "step": 8662 + }, + { + "epoch": 0.20307436253567515, + "grad_norm": 0.740744948387146, + "learning_rate": 0.0001803649539757038, + "loss": 0.2255, + "step": 8663 + }, + { + "epoch": 0.20309780411048015, + "grad_norm": 0.422523558139801, + "learning_rate": 0.00018036057013606348, + "loss": 0.0955, + "step": 8664 + }, + { + "epoch": 0.20312124568528514, + "grad_norm": 0.6262863278388977, + "learning_rate": 0.00018035618586038446, + "loss": 0.1405, + "step": 8665 + }, + { + "epoch": 0.20314468726009013, + "grad_norm": 0.23849093914031982, + "learning_rate": 0.00018035180114869053, + "loss": 0.0242, + "step": 8666 + }, + { + "epoch": 0.20316812883489513, + "grad_norm": 0.26272016763687134, + "learning_rate": 0.00018034741600100545, + "loss": 0.0472, + "step": 8667 + }, + { + "epoch": 0.20319157040970012, + "grad_norm": 0.6620708703994751, + "learning_rate": 0.00018034303041735306, + "loss": 0.09, + "step": 8668 + }, + { + "epoch": 0.20321501198450512, + "grad_norm": 1.0354483127593994, + "learning_rate": 0.00018033864439775713, + "loss": 0.2016, + "step": 8669 + }, + { + "epoch": 0.2032384535593101, + "grad_norm": 0.42867064476013184, + "learning_rate": 0.00018033425794224144, + "loss": 0.1101, + "step": 8670 + }, + { + "epoch": 0.2032618951341151, + "grad_norm": 0.3420883119106293, + "learning_rate": 0.0001803298710508298, + "loss": 0.0476, + "step": 8671 + }, + { + "epoch": 0.2032853367089201, + "grad_norm": 0.6048081517219543, + "learning_rate": 0.00018032548372354606, + "loss": 0.1174, + "step": 8672 + }, + { + "epoch": 0.2033087782837251, + "grad_norm": 0.42552539706230164, + "learning_rate": 0.00018032109596041393, + "loss": 0.1053, + "step": 8673 + }, + { + "epoch": 0.2033322198585301, + "grad_norm": 0.22042511403560638, + "learning_rate": 0.00018031670776145727, + "loss": 0.0489, + "step": 8674 + }, + { + "epoch": 0.20335566143333508, + "grad_norm": 0.5184459686279297, + "learning_rate": 0.00018031231912669994, + "loss": 0.6997, + "step": 8675 + }, + { + "epoch": 0.20337910300814008, + "grad_norm": 0.47888270020484924, + "learning_rate": 0.00018030793005616573, + "loss": 0.6863, + "step": 8676 + }, + { + "epoch": 0.20340254458294507, + "grad_norm": 0.8007267117500305, + "learning_rate": 0.0001803035405498784, + "loss": 0.2709, + "step": 8677 + }, + { + "epoch": 0.20342598615775007, + "grad_norm": 0.25830012559890747, + "learning_rate": 0.00018029915060786177, + "loss": 0.0771, + "step": 8678 + }, + { + "epoch": 0.20344942773255506, + "grad_norm": 0.6834151744842529, + "learning_rate": 0.00018029476023013972, + "loss": 0.1457, + "step": 8679 + }, + { + "epoch": 0.20347286930736005, + "grad_norm": 0.5555875301361084, + "learning_rate": 0.00018029036941673608, + "loss": 0.0819, + "step": 8680 + }, + { + "epoch": 0.20349631088216508, + "grad_norm": 0.32753995060920715, + "learning_rate": 0.0001802859781676746, + "loss": 0.0972, + "step": 8681 + }, + { + "epoch": 0.20351975245697007, + "grad_norm": 0.6606007814407349, + "learning_rate": 0.00018028158648297912, + "loss": 0.1542, + "step": 8682 + }, + { + "epoch": 0.20354319403177507, + "grad_norm": 1.1149041652679443, + "learning_rate": 0.00018027719436267353, + "loss": 0.2721, + "step": 8683 + }, + { + "epoch": 0.20356663560658006, + "grad_norm": 0.3981739282608032, + "learning_rate": 0.0001802728018067816, + "loss": 0.1045, + "step": 8684 + }, + { + "epoch": 0.20359007718138505, + "grad_norm": 0.6056767106056213, + "learning_rate": 0.00018026840881532722, + "loss": 0.0977, + "step": 8685 + }, + { + "epoch": 0.20361351875619005, + "grad_norm": 0.5099706053733826, + "learning_rate": 0.00018026401538833418, + "loss": 0.1223, + "step": 8686 + }, + { + "epoch": 0.20363696033099504, + "grad_norm": 0.4656122028827667, + "learning_rate": 0.00018025962152582633, + "loss": 0.1454, + "step": 8687 + }, + { + "epoch": 0.20366040190580004, + "grad_norm": 0.2825790345668793, + "learning_rate": 0.00018025522722782755, + "loss": 0.052, + "step": 8688 + }, + { + "epoch": 0.20368384348060503, + "grad_norm": 0.5171502232551575, + "learning_rate": 0.0001802508324943616, + "loss": 0.1395, + "step": 8689 + }, + { + "epoch": 0.20370728505541003, + "grad_norm": 0.4816376566886902, + "learning_rate": 0.00018024643732545244, + "loss": 0.0523, + "step": 8690 + }, + { + "epoch": 0.20373072663021502, + "grad_norm": 0.20843352377414703, + "learning_rate": 0.00018024204172112382, + "loss": 0.0574, + "step": 8691 + }, + { + "epoch": 0.20375416820502001, + "grad_norm": 0.7252232432365417, + "learning_rate": 0.00018023764568139963, + "loss": 0.1776, + "step": 8692 + }, + { + "epoch": 0.203777609779825, + "grad_norm": 0.5551912784576416, + "learning_rate": 0.0001802332492063037, + "loss": 0.1416, + "step": 8693 + }, + { + "epoch": 0.20380105135463, + "grad_norm": 0.7471585273742676, + "learning_rate": 0.00018022885229585991, + "loss": 0.1662, + "step": 8694 + }, + { + "epoch": 0.203824492929435, + "grad_norm": 0.7022367715835571, + "learning_rate": 0.00018022445495009213, + "loss": 0.137, + "step": 8695 + }, + { + "epoch": 0.20384793450424, + "grad_norm": 0.557439923286438, + "learning_rate": 0.00018022005716902419, + "loss": 0.1253, + "step": 8696 + }, + { + "epoch": 0.20387137607904499, + "grad_norm": 0.5747105479240417, + "learning_rate": 0.00018021565895267995, + "loss": 0.1486, + "step": 8697 + }, + { + "epoch": 0.20389481765384998, + "grad_norm": 0.22454750537872314, + "learning_rate": 0.0001802112603010833, + "loss": 0.0601, + "step": 8698 + }, + { + "epoch": 0.20391825922865497, + "grad_norm": 0.47565919160842896, + "learning_rate": 0.00018020686121425809, + "loss": 0.1129, + "step": 8699 + }, + { + "epoch": 0.20394170080345997, + "grad_norm": 0.14543519914150238, + "learning_rate": 0.0001802024616922282, + "loss": 0.0316, + "step": 8700 + }, + { + "epoch": 0.20396514237826496, + "grad_norm": 0.582962155342102, + "learning_rate": 0.00018019806173501753, + "loss": 0.162, + "step": 8701 + }, + { + "epoch": 0.20398858395306996, + "grad_norm": 0.6432943344116211, + "learning_rate": 0.0001801936613426499, + "loss": 0.1481, + "step": 8702 + }, + { + "epoch": 0.20401202552787495, + "grad_norm": 0.4699691832065582, + "learning_rate": 0.0001801892605151492, + "loss": 0.1009, + "step": 8703 + }, + { + "epoch": 0.20403546710267995, + "grad_norm": 0.8564994931221008, + "learning_rate": 0.0001801848592525393, + "loss": 0.2146, + "step": 8704 + }, + { + "epoch": 0.20405890867748494, + "grad_norm": 0.6081379652023315, + "learning_rate": 0.0001801804575548441, + "loss": 0.1335, + "step": 8705 + }, + { + "epoch": 0.20408235025228996, + "grad_norm": 0.7465633749961853, + "learning_rate": 0.00018017605542208754, + "loss": 0.6368, + "step": 8706 + }, + { + "epoch": 0.20410579182709496, + "grad_norm": 0.5114008784294128, + "learning_rate": 0.00018017165285429343, + "loss": 0.0769, + "step": 8707 + }, + { + "epoch": 0.20412923340189995, + "grad_norm": 0.4404764473438263, + "learning_rate": 0.00018016724985148562, + "loss": 0.0868, + "step": 8708 + }, + { + "epoch": 0.20415267497670495, + "grad_norm": 0.19603344798088074, + "learning_rate": 0.0001801628464136881, + "loss": 0.0465, + "step": 8709 + }, + { + "epoch": 0.20417611655150994, + "grad_norm": 0.22991310060024261, + "learning_rate": 0.00018015844254092476, + "loss": 0.065, + "step": 8710 + }, + { + "epoch": 0.20419955812631493, + "grad_norm": 0.5823919773101807, + "learning_rate": 0.0001801540382332194, + "loss": 0.1402, + "step": 8711 + }, + { + "epoch": 0.20422299970111993, + "grad_norm": 0.8368767499923706, + "learning_rate": 0.00018014963349059597, + "loss": 0.1842, + "step": 8712 + }, + { + "epoch": 0.20424644127592492, + "grad_norm": 0.4265610873699188, + "learning_rate": 0.0001801452283130784, + "loss": 0.0763, + "step": 8713 + }, + { + "epoch": 0.20426988285072992, + "grad_norm": 0.4142268896102905, + "learning_rate": 0.00018014082270069058, + "loss": 0.1039, + "step": 8714 + }, + { + "epoch": 0.2042933244255349, + "grad_norm": 0.5878127217292786, + "learning_rate": 0.0001801364166534564, + "loss": 0.1379, + "step": 8715 + }, + { + "epoch": 0.2043167660003399, + "grad_norm": 0.19334742426872253, + "learning_rate": 0.00018013201017139972, + "loss": 0.0438, + "step": 8716 + }, + { + "epoch": 0.2043402075751449, + "grad_norm": 0.729448139667511, + "learning_rate": 0.00018012760325454455, + "loss": 0.181, + "step": 8717 + }, + { + "epoch": 0.2043636491499499, + "grad_norm": 0.5924577116966248, + "learning_rate": 0.00018012319590291474, + "loss": 0.1639, + "step": 8718 + }, + { + "epoch": 0.2043870907247549, + "grad_norm": 0.44697749614715576, + "learning_rate": 0.00018011878811653422, + "loss": 0.0988, + "step": 8719 + }, + { + "epoch": 0.20441053229955988, + "grad_norm": 0.7993769645690918, + "learning_rate": 0.00018011437989542687, + "loss": 0.1558, + "step": 8720 + }, + { + "epoch": 0.20443397387436488, + "grad_norm": 0.6208627820014954, + "learning_rate": 0.0001801099712396167, + "loss": 0.1991, + "step": 8721 + }, + { + "epoch": 0.20445741544916987, + "grad_norm": 0.10853661596775055, + "learning_rate": 0.00018010556214912752, + "loss": 0.0169, + "step": 8722 + }, + { + "epoch": 0.20448085702397487, + "grad_norm": 0.3980400264263153, + "learning_rate": 0.00018010115262398336, + "loss": 0.0547, + "step": 8723 + }, + { + "epoch": 0.20450429859877986, + "grad_norm": 0.16605046391487122, + "learning_rate": 0.00018009674266420805, + "loss": 0.0406, + "step": 8724 + }, + { + "epoch": 0.20452774017358485, + "grad_norm": 0.47693371772766113, + "learning_rate": 0.00018009233226982558, + "loss": 0.1063, + "step": 8725 + }, + { + "epoch": 0.20455118174838985, + "grad_norm": 0.7200890183448792, + "learning_rate": 0.00018008792144085986, + "loss": 0.6986, + "step": 8726 + }, + { + "epoch": 0.20457462332319484, + "grad_norm": 0.46174514293670654, + "learning_rate": 0.00018008351017733483, + "loss": 0.0635, + "step": 8727 + }, + { + "epoch": 0.20459806489799984, + "grad_norm": 0.4796469509601593, + "learning_rate": 0.0001800790984792744, + "loss": 0.1208, + "step": 8728 + }, + { + "epoch": 0.20462150647280483, + "grad_norm": 0.43386778235435486, + "learning_rate": 0.00018007468634670255, + "loss": 0.1047, + "step": 8729 + }, + { + "epoch": 0.20464494804760983, + "grad_norm": 0.16844189167022705, + "learning_rate": 0.0001800702737796432, + "loss": 0.0458, + "step": 8730 + }, + { + "epoch": 0.20466838962241482, + "grad_norm": 0.5297343730926514, + "learning_rate": 0.00018006586077812028, + "loss": 0.0959, + "step": 8731 + }, + { + "epoch": 0.20469183119721984, + "grad_norm": 0.6011572480201721, + "learning_rate": 0.00018006144734215775, + "loss": 0.1259, + "step": 8732 + }, + { + "epoch": 0.20471527277202484, + "grad_norm": 0.7411909699440002, + "learning_rate": 0.0001800570334717796, + "loss": 0.2041, + "step": 8733 + }, + { + "epoch": 0.20473871434682983, + "grad_norm": 0.6071872115135193, + "learning_rate": 0.00018005261916700967, + "loss": 0.0959, + "step": 8734 + }, + { + "epoch": 0.20476215592163483, + "grad_norm": 0.4332064986228943, + "learning_rate": 0.000180048204427872, + "loss": 0.1398, + "step": 8735 + }, + { + "epoch": 0.20478559749643982, + "grad_norm": 0.4329114854335785, + "learning_rate": 0.00018004378925439056, + "loss": 0.2194, + "step": 8736 + }, + { + "epoch": 0.20480903907124481, + "grad_norm": 0.5629383325576782, + "learning_rate": 0.00018003937364658923, + "loss": 0.1104, + "step": 8737 + }, + { + "epoch": 0.2048324806460498, + "grad_norm": 0.3680645823478699, + "learning_rate": 0.00018003495760449202, + "loss": 0.0905, + "step": 8738 + }, + { + "epoch": 0.2048559222208548, + "grad_norm": 0.6005721092224121, + "learning_rate": 0.00018003054112812286, + "loss": 0.1764, + "step": 8739 + }, + { + "epoch": 0.2048793637956598, + "grad_norm": 0.5793299078941345, + "learning_rate": 0.0001800261242175058, + "loss": 0.1429, + "step": 8740 + }, + { + "epoch": 0.2049028053704648, + "grad_norm": 0.3909156918525696, + "learning_rate": 0.00018002170687266464, + "loss": 0.0273, + "step": 8741 + }, + { + "epoch": 0.20492624694526979, + "grad_norm": 0.6063154935836792, + "learning_rate": 0.0001800172890936235, + "loss": 0.1153, + "step": 8742 + }, + { + "epoch": 0.20494968852007478, + "grad_norm": 0.2948063313961029, + "learning_rate": 0.0001800128708804063, + "loss": 0.0658, + "step": 8743 + }, + { + "epoch": 0.20497313009487977, + "grad_norm": 0.32340019941329956, + "learning_rate": 0.00018000845223303702, + "loss": 0.0564, + "step": 8744 + }, + { + "epoch": 0.20499657166968477, + "grad_norm": 0.3178112804889679, + "learning_rate": 0.0001800040331515396, + "loss": 0.0438, + "step": 8745 + }, + { + "epoch": 0.20502001324448976, + "grad_norm": 0.40012064576148987, + "learning_rate": 0.0001799996136359381, + "loss": 0.0909, + "step": 8746 + }, + { + "epoch": 0.20504345481929476, + "grad_norm": 0.49787023663520813, + "learning_rate": 0.0001799951936862564, + "loss": 0.1665, + "step": 8747 + }, + { + "epoch": 0.20506689639409975, + "grad_norm": 0.6398888230323792, + "learning_rate": 0.00017999077330251853, + "loss": 0.1579, + "step": 8748 + }, + { + "epoch": 0.20509033796890475, + "grad_norm": 0.8612897992134094, + "learning_rate": 0.00017998635248474848, + "loss": 0.1557, + "step": 8749 + }, + { + "epoch": 0.20511377954370974, + "grad_norm": 0.40145742893218994, + "learning_rate": 0.00017998193123297025, + "loss": 0.0992, + "step": 8750 + }, + { + "epoch": 0.20513722111851473, + "grad_norm": 0.3730372488498688, + "learning_rate": 0.0001799775095472078, + "loss": 0.0746, + "step": 8751 + }, + { + "epoch": 0.20516066269331973, + "grad_norm": 0.5121384859085083, + "learning_rate": 0.00017997308742748517, + "loss": 0.1241, + "step": 8752 + }, + { + "epoch": 0.20518410426812472, + "grad_norm": 0.6574753522872925, + "learning_rate": 0.00017996866487382626, + "loss": 0.115, + "step": 8753 + }, + { + "epoch": 0.20520754584292972, + "grad_norm": 0.8830963373184204, + "learning_rate": 0.00017996424188625513, + "loss": 0.8148, + "step": 8754 + }, + { + "epoch": 0.2052309874177347, + "grad_norm": 0.3705771267414093, + "learning_rate": 0.00017995981846479582, + "loss": 0.0876, + "step": 8755 + }, + { + "epoch": 0.2052544289925397, + "grad_norm": 0.7238439917564392, + "learning_rate": 0.00017995539460947225, + "loss": 0.155, + "step": 8756 + }, + { + "epoch": 0.20527787056734473, + "grad_norm": 0.40438178181648254, + "learning_rate": 0.0001799509703203085, + "loss": 0.078, + "step": 8757 + }, + { + "epoch": 0.20530131214214972, + "grad_norm": 0.5177406668663025, + "learning_rate": 0.0001799465455973285, + "loss": 0.1435, + "step": 8758 + }, + { + "epoch": 0.20532475371695472, + "grad_norm": 0.0909590870141983, + "learning_rate": 0.00017994212044055628, + "loss": 0.0199, + "step": 8759 + }, + { + "epoch": 0.2053481952917597, + "grad_norm": 0.3523876965045929, + "learning_rate": 0.0001799376948500159, + "loss": 0.0717, + "step": 8760 + }, + { + "epoch": 0.2053716368665647, + "grad_norm": 0.5329910516738892, + "learning_rate": 0.00017993326882573136, + "loss": 0.1064, + "step": 8761 + }, + { + "epoch": 0.2053950784413697, + "grad_norm": 0.48684048652648926, + "learning_rate": 0.00017992884236772661, + "loss": 0.6787, + "step": 8762 + }, + { + "epoch": 0.2054185200161747, + "grad_norm": 0.35137027502059937, + "learning_rate": 0.00017992441547602574, + "loss": 0.0948, + "step": 8763 + }, + { + "epoch": 0.2054419615909797, + "grad_norm": 0.4117944538593292, + "learning_rate": 0.00017991998815065273, + "loss": 0.0573, + "step": 8764 + }, + { + "epoch": 0.20546540316578468, + "grad_norm": 0.26857155561447144, + "learning_rate": 0.00017991556039163165, + "loss": 0.0486, + "step": 8765 + }, + { + "epoch": 0.20548884474058968, + "grad_norm": 0.3489537537097931, + "learning_rate": 0.00017991113219898645, + "loss": 0.0708, + "step": 8766 + }, + { + "epoch": 0.20551228631539467, + "grad_norm": 0.7128647565841675, + "learning_rate": 0.00017990670357274121, + "loss": 0.1678, + "step": 8767 + }, + { + "epoch": 0.20553572789019967, + "grad_norm": 0.5228469967842102, + "learning_rate": 0.00017990227451291998, + "loss": 0.1251, + "step": 8768 + }, + { + "epoch": 0.20555916946500466, + "grad_norm": 0.8272022604942322, + "learning_rate": 0.00017989784501954673, + "loss": 0.1983, + "step": 8769 + }, + { + "epoch": 0.20558261103980965, + "grad_norm": 0.6474776864051819, + "learning_rate": 0.0001798934150926455, + "loss": 0.1406, + "step": 8770 + }, + { + "epoch": 0.20560605261461465, + "grad_norm": 0.7674769759178162, + "learning_rate": 0.0001798889847322404, + "loss": 0.2046, + "step": 8771 + }, + { + "epoch": 0.20562949418941964, + "grad_norm": 0.8059005737304688, + "learning_rate": 0.00017988455393835542, + "loss": 0.2158, + "step": 8772 + }, + { + "epoch": 0.20565293576422464, + "grad_norm": 0.4490198791027069, + "learning_rate": 0.0001798801227110146, + "loss": 0.1082, + "step": 8773 + }, + { + "epoch": 0.20567637733902963, + "grad_norm": 0.9478737115859985, + "learning_rate": 0.00017987569105024198, + "loss": 0.1289, + "step": 8774 + }, + { + "epoch": 0.20569981891383463, + "grad_norm": 0.807776689529419, + "learning_rate": 0.0001798712589560616, + "loss": 0.2165, + "step": 8775 + }, + { + "epoch": 0.20572326048863962, + "grad_norm": 0.7804641723632812, + "learning_rate": 0.00017986682642849754, + "loss": 0.1709, + "step": 8776 + }, + { + "epoch": 0.20574670206344461, + "grad_norm": 0.4424547553062439, + "learning_rate": 0.00017986239346757382, + "loss": 0.068, + "step": 8777 + }, + { + "epoch": 0.2057701436382496, + "grad_norm": 0.679220974445343, + "learning_rate": 0.00017985796007331456, + "loss": 0.1426, + "step": 8778 + }, + { + "epoch": 0.2057935852130546, + "grad_norm": 0.24035850167274475, + "learning_rate": 0.0001798535262457437, + "loss": 0.0303, + "step": 8779 + }, + { + "epoch": 0.2058170267878596, + "grad_norm": 0.3215738534927368, + "learning_rate": 0.0001798490919848854, + "loss": 0.0712, + "step": 8780 + }, + { + "epoch": 0.2058404683626646, + "grad_norm": 0.8829333186149597, + "learning_rate": 0.00017984465729076367, + "loss": 0.1686, + "step": 8781 + }, + { + "epoch": 0.20586390993746959, + "grad_norm": 0.5001850724220276, + "learning_rate": 0.00017984022216340257, + "loss": 0.4197, + "step": 8782 + }, + { + "epoch": 0.2058873515122746, + "grad_norm": 0.8385021090507507, + "learning_rate": 0.0001798357866028262, + "loss": 0.2188, + "step": 8783 + }, + { + "epoch": 0.2059107930870796, + "grad_norm": 0.7253649234771729, + "learning_rate": 0.0001798313506090586, + "loss": 0.121, + "step": 8784 + }, + { + "epoch": 0.2059342346618846, + "grad_norm": 0.6304307579994202, + "learning_rate": 0.00017982691418212384, + "loss": 0.1233, + "step": 8785 + }, + { + "epoch": 0.2059576762366896, + "grad_norm": 0.3540430963039398, + "learning_rate": 0.000179822477322046, + "loss": 0.0605, + "step": 8786 + }, + { + "epoch": 0.20598111781149459, + "grad_norm": 0.7514124512672424, + "learning_rate": 0.00017981804002884918, + "loss": 0.1175, + "step": 8787 + }, + { + "epoch": 0.20600455938629958, + "grad_norm": 0.5750205516815186, + "learning_rate": 0.0001798136023025574, + "loss": 0.1567, + "step": 8788 + }, + { + "epoch": 0.20602800096110457, + "grad_norm": 0.5154857635498047, + "learning_rate": 0.0001798091641431948, + "loss": 0.1019, + "step": 8789 + }, + { + "epoch": 0.20605144253590957, + "grad_norm": 0.8332803845405579, + "learning_rate": 0.0001798047255507854, + "loss": 0.1255, + "step": 8790 + }, + { + "epoch": 0.20607488411071456, + "grad_norm": 0.4989287257194519, + "learning_rate": 0.00017980028652535334, + "loss": 0.1277, + "step": 8791 + }, + { + "epoch": 0.20609832568551956, + "grad_norm": 0.37931591272354126, + "learning_rate": 0.00017979584706692266, + "loss": 0.0883, + "step": 8792 + }, + { + "epoch": 0.20612176726032455, + "grad_norm": 0.32819023728370667, + "learning_rate": 0.00017979140717551748, + "loss": 0.1145, + "step": 8793 + }, + { + "epoch": 0.20614520883512955, + "grad_norm": 0.5356243252754211, + "learning_rate": 0.00017978696685116185, + "loss": 0.7676, + "step": 8794 + }, + { + "epoch": 0.20616865040993454, + "grad_norm": 0.5669243931770325, + "learning_rate": 0.00017978252609387993, + "loss": 0.1589, + "step": 8795 + }, + { + "epoch": 0.20619209198473953, + "grad_norm": 0.5530673265457153, + "learning_rate": 0.00017977808490369577, + "loss": 0.1565, + "step": 8796 + }, + { + "epoch": 0.20621553355954453, + "grad_norm": 0.14583145081996918, + "learning_rate": 0.00017977364328063347, + "loss": 0.0296, + "step": 8797 + }, + { + "epoch": 0.20623897513434952, + "grad_norm": 0.5660419464111328, + "learning_rate": 0.00017976920122471715, + "loss": 0.1521, + "step": 8798 + }, + { + "epoch": 0.20626241670915452, + "grad_norm": 0.782599687576294, + "learning_rate": 0.00017976475873597092, + "loss": 0.1638, + "step": 8799 + }, + { + "epoch": 0.2062858582839595, + "grad_norm": 0.6012731194496155, + "learning_rate": 0.00017976031581441886, + "loss": 0.1017, + "step": 8800 + }, + { + "epoch": 0.2063092998587645, + "grad_norm": 0.11913345009088516, + "learning_rate": 0.00017975587246008506, + "loss": 0.0324, + "step": 8801 + }, + { + "epoch": 0.2063327414335695, + "grad_norm": 0.2479640543460846, + "learning_rate": 0.00017975142867299366, + "loss": 0.0785, + "step": 8802 + }, + { + "epoch": 0.2063561830083745, + "grad_norm": 0.41693007946014404, + "learning_rate": 0.00017974698445316877, + "loss": 0.1765, + "step": 8803 + }, + { + "epoch": 0.2063796245831795, + "grad_norm": 0.1761176884174347, + "learning_rate": 0.0001797425398006345, + "loss": 0.037, + "step": 8804 + }, + { + "epoch": 0.20640306615798448, + "grad_norm": 0.3100127577781677, + "learning_rate": 0.000179738094715415, + "loss": 0.1163, + "step": 8805 + }, + { + "epoch": 0.20642650773278948, + "grad_norm": 0.6594531536102295, + "learning_rate": 0.00017973364919753428, + "loss": 0.2305, + "step": 8806 + }, + { + "epoch": 0.20644994930759447, + "grad_norm": 0.7067521214485168, + "learning_rate": 0.00017972920324701657, + "loss": 0.1046, + "step": 8807 + }, + { + "epoch": 0.2064733908823995, + "grad_norm": 0.16620683670043945, + "learning_rate": 0.00017972475686388602, + "loss": 0.0359, + "step": 8808 + }, + { + "epoch": 0.2064968324572045, + "grad_norm": 0.3831525444984436, + "learning_rate": 0.00017972031004816663, + "loss": 0.0918, + "step": 8809 + }, + { + "epoch": 0.20652027403200948, + "grad_norm": 0.3628937304019928, + "learning_rate": 0.00017971586279988262, + "loss": 0.0385, + "step": 8810 + }, + { + "epoch": 0.20654371560681448, + "grad_norm": 0.38020560145378113, + "learning_rate": 0.00017971141511905807, + "loss": 0.1011, + "step": 8811 + }, + { + "epoch": 0.20656715718161947, + "grad_norm": 0.14585831761360168, + "learning_rate": 0.0001797069670057172, + "loss": 0.046, + "step": 8812 + }, + { + "epoch": 0.20659059875642447, + "grad_norm": 0.6623074412345886, + "learning_rate": 0.00017970251845988403, + "loss": 0.1442, + "step": 8813 + }, + { + "epoch": 0.20661404033122946, + "grad_norm": 0.22679558396339417, + "learning_rate": 0.00017969806948158276, + "loss": 0.0511, + "step": 8814 + }, + { + "epoch": 0.20663748190603445, + "grad_norm": 0.5281689167022705, + "learning_rate": 0.00017969362007083753, + "loss": 0.1304, + "step": 8815 + }, + { + "epoch": 0.20666092348083945, + "grad_norm": 0.37024542689323425, + "learning_rate": 0.0001796891702276725, + "loss": 0.1051, + "step": 8816 + }, + { + "epoch": 0.20668436505564444, + "grad_norm": 0.3297891914844513, + "learning_rate": 0.00017968471995211179, + "loss": 0.0558, + "step": 8817 + }, + { + "epoch": 0.20670780663044944, + "grad_norm": 0.6210941076278687, + "learning_rate": 0.0001796802692441795, + "loss": 0.1467, + "step": 8818 + }, + { + "epoch": 0.20673124820525443, + "grad_norm": 0.5744023323059082, + "learning_rate": 0.00017967581810389988, + "loss": 0.1019, + "step": 8819 + }, + { + "epoch": 0.20675468978005943, + "grad_norm": 0.46427375078201294, + "learning_rate": 0.00017967136653129702, + "loss": 0.1279, + "step": 8820 + }, + { + "epoch": 0.20677813135486442, + "grad_norm": 0.5789769291877747, + "learning_rate": 0.00017966691452639508, + "loss": 0.1029, + "step": 8821 + }, + { + "epoch": 0.2068015729296694, + "grad_norm": 0.7329282760620117, + "learning_rate": 0.0001796624620892182, + "loss": 0.1709, + "step": 8822 + }, + { + "epoch": 0.2068250145044744, + "grad_norm": 0.5634263753890991, + "learning_rate": 0.0001796580092197906, + "loss": 0.6778, + "step": 8823 + }, + { + "epoch": 0.2068484560792794, + "grad_norm": 0.6201493144035339, + "learning_rate": 0.0001796535559181364, + "loss": 0.1219, + "step": 8824 + }, + { + "epoch": 0.2068718976540844, + "grad_norm": 0.1388527899980545, + "learning_rate": 0.00017964910218427974, + "loss": 0.0277, + "step": 8825 + }, + { + "epoch": 0.2068953392288894, + "grad_norm": 0.21100805699825287, + "learning_rate": 0.00017964464801824482, + "loss": 0.03, + "step": 8826 + }, + { + "epoch": 0.20691878080369439, + "grad_norm": 0.17224526405334473, + "learning_rate": 0.0001796401934200558, + "loss": 0.0414, + "step": 8827 + }, + { + "epoch": 0.20694222237849938, + "grad_norm": 0.6120057702064514, + "learning_rate": 0.00017963573838973687, + "loss": 0.0845, + "step": 8828 + }, + { + "epoch": 0.20696566395330437, + "grad_norm": 0.1860004961490631, + "learning_rate": 0.00017963128292731214, + "loss": 0.0453, + "step": 8829 + }, + { + "epoch": 0.20698910552810937, + "grad_norm": 0.46929100155830383, + "learning_rate": 0.00017962682703280588, + "loss": 0.5369, + "step": 8830 + }, + { + "epoch": 0.20701254710291436, + "grad_norm": 0.6629065275192261, + "learning_rate": 0.0001796223707062422, + "loss": 0.1335, + "step": 8831 + }, + { + "epoch": 0.20703598867771936, + "grad_norm": 0.23452672362327576, + "learning_rate": 0.00017961791394764526, + "loss": 0.0633, + "step": 8832 + }, + { + "epoch": 0.20705943025252435, + "grad_norm": 0.573896586894989, + "learning_rate": 0.00017961345675703933, + "loss": 0.1508, + "step": 8833 + }, + { + "epoch": 0.20708287182732937, + "grad_norm": 1.021958351135254, + "learning_rate": 0.00017960899913444854, + "loss": 0.0891, + "step": 8834 + }, + { + "epoch": 0.20710631340213437, + "grad_norm": 0.41412582993507385, + "learning_rate": 0.00017960454107989704, + "loss": 0.1096, + "step": 8835 + }, + { + "epoch": 0.20712975497693936, + "grad_norm": 0.6319136619567871, + "learning_rate": 0.0001796000825934091, + "loss": 0.1489, + "step": 8836 + }, + { + "epoch": 0.20715319655174436, + "grad_norm": 0.35350245237350464, + "learning_rate": 0.00017959562367500889, + "loss": 0.0903, + "step": 8837 + }, + { + "epoch": 0.20717663812654935, + "grad_norm": 0.5557618141174316, + "learning_rate": 0.00017959116432472055, + "loss": 0.1192, + "step": 8838 + }, + { + "epoch": 0.20720007970135434, + "grad_norm": 0.28787270188331604, + "learning_rate": 0.00017958670454256833, + "loss": 0.0673, + "step": 8839 + }, + { + "epoch": 0.20722352127615934, + "grad_norm": 0.5969753861427307, + "learning_rate": 0.0001795822443285764, + "loss": 0.1216, + "step": 8840 + }, + { + "epoch": 0.20724696285096433, + "grad_norm": 0.49771907925605774, + "learning_rate": 0.000179577783682769, + "loss": 0.116, + "step": 8841 + }, + { + "epoch": 0.20727040442576933, + "grad_norm": 0.47583910822868347, + "learning_rate": 0.0001795733226051703, + "loss": 0.4636, + "step": 8842 + }, + { + "epoch": 0.20729384600057432, + "grad_norm": 0.5619797110557556, + "learning_rate": 0.00017956886109580448, + "loss": 0.1408, + "step": 8843 + }, + { + "epoch": 0.20731728757537932, + "grad_norm": 0.32854121923446655, + "learning_rate": 0.00017956439915469582, + "loss": 0.0762, + "step": 8844 + }, + { + "epoch": 0.2073407291501843, + "grad_norm": 0.5044355988502502, + "learning_rate": 0.0001795599367818685, + "loss": 0.155, + "step": 8845 + }, + { + "epoch": 0.2073641707249893, + "grad_norm": 0.2397080361843109, + "learning_rate": 0.0001795554739773467, + "loss": 0.0679, + "step": 8846 + }, + { + "epoch": 0.2073876122997943, + "grad_norm": 0.20154230296611786, + "learning_rate": 0.00017955101074115465, + "loss": 0.0308, + "step": 8847 + }, + { + "epoch": 0.2074110538745993, + "grad_norm": 0.8119200468063354, + "learning_rate": 0.00017954654707331665, + "loss": 0.747, + "step": 8848 + }, + { + "epoch": 0.2074344954494043, + "grad_norm": 0.3931174874305725, + "learning_rate": 0.00017954208297385677, + "loss": 0.1123, + "step": 8849 + }, + { + "epoch": 0.20745793702420928, + "grad_norm": 0.584611177444458, + "learning_rate": 0.00017953761844279934, + "loss": 0.1551, + "step": 8850 + }, + { + "epoch": 0.20748137859901428, + "grad_norm": 0.5247711539268494, + "learning_rate": 0.00017953315348016857, + "loss": 0.075, + "step": 8851 + }, + { + "epoch": 0.20750482017381927, + "grad_norm": 0.512065052986145, + "learning_rate": 0.00017952868808598864, + "loss": 0.1277, + "step": 8852 + }, + { + "epoch": 0.20752826174862427, + "grad_norm": 0.2839510142803192, + "learning_rate": 0.00017952422226028385, + "loss": 0.0552, + "step": 8853 + }, + { + "epoch": 0.20755170332342926, + "grad_norm": 0.2535901665687561, + "learning_rate": 0.00017951975600307836, + "loss": 0.0658, + "step": 8854 + }, + { + "epoch": 0.20757514489823425, + "grad_norm": 0.11044342070817947, + "learning_rate": 0.00017951528931439646, + "loss": 0.0211, + "step": 8855 + }, + { + "epoch": 0.20759858647303925, + "grad_norm": 0.6655513048171997, + "learning_rate": 0.00017951082219426237, + "loss": 0.1886, + "step": 8856 + }, + { + "epoch": 0.20762202804784424, + "grad_norm": 0.5788015127182007, + "learning_rate": 0.00017950635464270028, + "loss": 0.1912, + "step": 8857 + }, + { + "epoch": 0.20764546962264924, + "grad_norm": 0.25033286213874817, + "learning_rate": 0.00017950188665973453, + "loss": 0.0391, + "step": 8858 + }, + { + "epoch": 0.20766891119745423, + "grad_norm": 0.26019078493118286, + "learning_rate": 0.00017949741824538928, + "loss": 0.0741, + "step": 8859 + }, + { + "epoch": 0.20769235277225925, + "grad_norm": 0.534540057182312, + "learning_rate": 0.0001794929493996888, + "loss": 0.1239, + "step": 8860 + }, + { + "epoch": 0.20771579434706425, + "grad_norm": 0.46189621090888977, + "learning_rate": 0.00017948848012265734, + "loss": 0.1359, + "step": 8861 + }, + { + "epoch": 0.20773923592186924, + "grad_norm": 0.4927632212638855, + "learning_rate": 0.00017948401041431916, + "loss": 0.083, + "step": 8862 + }, + { + "epoch": 0.20776267749667424, + "grad_norm": 0.5494357943534851, + "learning_rate": 0.0001794795402746985, + "loss": 0.0723, + "step": 8863 + }, + { + "epoch": 0.20778611907147923, + "grad_norm": 0.9197568297386169, + "learning_rate": 0.00017947506970381963, + "loss": 0.1787, + "step": 8864 + }, + { + "epoch": 0.20780956064628422, + "grad_norm": 0.5818542838096619, + "learning_rate": 0.00017947059870170682, + "loss": 0.1975, + "step": 8865 + }, + { + "epoch": 0.20783300222108922, + "grad_norm": 0.1797255575656891, + "learning_rate": 0.00017946612726838426, + "loss": 0.029, + "step": 8866 + }, + { + "epoch": 0.2078564437958942, + "grad_norm": 0.27048972249031067, + "learning_rate": 0.00017946165540387628, + "loss": 0.0612, + "step": 8867 + }, + { + "epoch": 0.2078798853706992, + "grad_norm": 0.20524582266807556, + "learning_rate": 0.00017945718310820712, + "loss": 0.0651, + "step": 8868 + }, + { + "epoch": 0.2079033269455042, + "grad_norm": 0.691024124622345, + "learning_rate": 0.00017945271038140103, + "loss": 0.8164, + "step": 8869 + }, + { + "epoch": 0.2079267685203092, + "grad_norm": 0.38782861828804016, + "learning_rate": 0.00017944823722348233, + "loss": 0.0582, + "step": 8870 + }, + { + "epoch": 0.2079502100951142, + "grad_norm": 0.6903435587882996, + "learning_rate": 0.00017944376363447526, + "loss": 0.1697, + "step": 8871 + }, + { + "epoch": 0.20797365166991919, + "grad_norm": 0.1064407229423523, + "learning_rate": 0.0001794392896144041, + "loss": 0.0257, + "step": 8872 + }, + { + "epoch": 0.20799709324472418, + "grad_norm": 0.11313275992870331, + "learning_rate": 0.0001794348151632931, + "loss": 0.0219, + "step": 8873 + }, + { + "epoch": 0.20802053481952917, + "grad_norm": 0.6404618620872498, + "learning_rate": 0.00017943034028116656, + "loss": 0.1186, + "step": 8874 + }, + { + "epoch": 0.20804397639433417, + "grad_norm": 0.4658365547657013, + "learning_rate": 0.00017942586496804878, + "loss": 0.1099, + "step": 8875 + }, + { + "epoch": 0.20806741796913916, + "grad_norm": 0.8616634011268616, + "learning_rate": 0.000179421389223964, + "loss": 0.7218, + "step": 8876 + }, + { + "epoch": 0.20809085954394416, + "grad_norm": 0.7671387195587158, + "learning_rate": 0.00017941691304893654, + "loss": 0.2236, + "step": 8877 + }, + { + "epoch": 0.20811430111874915, + "grad_norm": 0.5593684315681458, + "learning_rate": 0.00017941243644299066, + "loss": 0.1488, + "step": 8878 + }, + { + "epoch": 0.20813774269355415, + "grad_norm": 0.5432280898094177, + "learning_rate": 0.00017940795940615066, + "loss": 0.0847, + "step": 8879 + }, + { + "epoch": 0.20816118426835914, + "grad_norm": 0.5395185947418213, + "learning_rate": 0.00017940348193844089, + "loss": 0.8443, + "step": 8880 + }, + { + "epoch": 0.20818462584316413, + "grad_norm": 0.5693603157997131, + "learning_rate": 0.00017939900403988557, + "loss": 0.0922, + "step": 8881 + }, + { + "epoch": 0.20820806741796913, + "grad_norm": 0.7311559915542603, + "learning_rate": 0.00017939452571050902, + "loss": 0.206, + "step": 8882 + }, + { + "epoch": 0.20823150899277412, + "grad_norm": 0.5601897835731506, + "learning_rate": 0.00017939004695033555, + "loss": 0.1085, + "step": 8883 + }, + { + "epoch": 0.20825495056757912, + "grad_norm": 0.5704688429832458, + "learning_rate": 0.00017938556775938942, + "loss": 0.1658, + "step": 8884 + }, + { + "epoch": 0.20827839214238414, + "grad_norm": 0.34039366245269775, + "learning_rate": 0.00017938108813769497, + "loss": 0.1014, + "step": 8885 + }, + { + "epoch": 0.20830183371718913, + "grad_norm": 0.13416698575019836, + "learning_rate": 0.00017937660808527657, + "loss": 0.0207, + "step": 8886 + }, + { + "epoch": 0.20832527529199413, + "grad_norm": 0.3676689863204956, + "learning_rate": 0.0001793721276021584, + "loss": 0.0818, + "step": 8887 + }, + { + "epoch": 0.20834871686679912, + "grad_norm": 0.35427772998809814, + "learning_rate": 0.00017936764668836485, + "loss": 0.1011, + "step": 8888 + }, + { + "epoch": 0.20837215844160412, + "grad_norm": 0.10671483725309372, + "learning_rate": 0.0001793631653439202, + "loss": 0.02, + "step": 8889 + }, + { + "epoch": 0.2083956000164091, + "grad_norm": 0.5702257752418518, + "learning_rate": 0.0001793586835688488, + "loss": 0.1328, + "step": 8890 + }, + { + "epoch": 0.2084190415912141, + "grad_norm": 0.4059233069419861, + "learning_rate": 0.00017935420136317494, + "loss": 0.1272, + "step": 8891 + }, + { + "epoch": 0.2084424831660191, + "grad_norm": 0.5470440983772278, + "learning_rate": 0.00017934971872692298, + "loss": 0.1432, + "step": 8892 + }, + { + "epoch": 0.2084659247408241, + "grad_norm": 0.5439013242721558, + "learning_rate": 0.0001793452356601172, + "loss": 0.1121, + "step": 8893 + }, + { + "epoch": 0.2084893663156291, + "grad_norm": 0.4824609160423279, + "learning_rate": 0.00017934075216278196, + "loss": 0.0874, + "step": 8894 + }, + { + "epoch": 0.20851280789043408, + "grad_norm": 0.35666903853416443, + "learning_rate": 0.00017933626823494155, + "loss": 0.0573, + "step": 8895 + }, + { + "epoch": 0.20853624946523908, + "grad_norm": 0.16620957851409912, + "learning_rate": 0.00017933178387662032, + "loss": 0.048, + "step": 8896 + }, + { + "epoch": 0.20855969104004407, + "grad_norm": 0.2462289184331894, + "learning_rate": 0.00017932729908784259, + "loss": 0.0453, + "step": 8897 + }, + { + "epoch": 0.20858313261484906, + "grad_norm": 0.44849538803100586, + "learning_rate": 0.00017932281386863269, + "loss": 0.0915, + "step": 8898 + }, + { + "epoch": 0.20860657418965406, + "grad_norm": 0.5241284966468811, + "learning_rate": 0.000179318328219015, + "loss": 0.0998, + "step": 8899 + }, + { + "epoch": 0.20863001576445905, + "grad_norm": 0.5269938111305237, + "learning_rate": 0.00017931384213901383, + "loss": 0.0912, + "step": 8900 + }, + { + "epoch": 0.20865345733926405, + "grad_norm": 0.11258967220783234, + "learning_rate": 0.0001793093556286535, + "loss": 0.0188, + "step": 8901 + }, + { + "epoch": 0.20867689891406904, + "grad_norm": 0.2091676890850067, + "learning_rate": 0.0001793048686879584, + "loss": 0.0515, + "step": 8902 + }, + { + "epoch": 0.20870034048887404, + "grad_norm": 0.2748352289199829, + "learning_rate": 0.00017930038131695283, + "loss": 0.0632, + "step": 8903 + }, + { + "epoch": 0.20872378206367903, + "grad_norm": 0.5254852771759033, + "learning_rate": 0.0001792958935156612, + "loss": 0.0989, + "step": 8904 + }, + { + "epoch": 0.20874722363848403, + "grad_norm": 0.2239304780960083, + "learning_rate": 0.0001792914052841078, + "loss": 0.0457, + "step": 8905 + }, + { + "epoch": 0.20877066521328902, + "grad_norm": 1.1136926412582397, + "learning_rate": 0.000179286916622317, + "loss": 0.366, + "step": 8906 + }, + { + "epoch": 0.208794106788094, + "grad_norm": 0.5572555065155029, + "learning_rate": 0.00017928242753031314, + "loss": 0.1226, + "step": 8907 + }, + { + "epoch": 0.208817548362899, + "grad_norm": 0.5185630321502686, + "learning_rate": 0.00017927793800812062, + "loss": 0.1118, + "step": 8908 + }, + { + "epoch": 0.208840989937704, + "grad_norm": 0.44913920760154724, + "learning_rate": 0.0001792734480557638, + "loss": 0.0358, + "step": 8909 + }, + { + "epoch": 0.208864431512509, + "grad_norm": 0.31286606192588806, + "learning_rate": 0.000179268957673267, + "loss": 0.0793, + "step": 8910 + }, + { + "epoch": 0.20888787308731402, + "grad_norm": 0.5058275461196899, + "learning_rate": 0.00017926446686065462, + "loss": 0.1144, + "step": 8911 + }, + { + "epoch": 0.208911314662119, + "grad_norm": 0.4102218449115753, + "learning_rate": 0.000179259975617951, + "loss": 0.0915, + "step": 8912 + }, + { + "epoch": 0.208934756236924, + "grad_norm": 0.40409937500953674, + "learning_rate": 0.00017925548394518052, + "loss": 0.0947, + "step": 8913 + }, + { + "epoch": 0.208958197811729, + "grad_norm": 0.304043173789978, + "learning_rate": 0.0001792509918423676, + "loss": 0.0475, + "step": 8914 + }, + { + "epoch": 0.208981639386534, + "grad_norm": 0.4116867482662201, + "learning_rate": 0.0001792464993095365, + "loss": 0.0593, + "step": 8915 + }, + { + "epoch": 0.209005080961339, + "grad_norm": 0.41045165061950684, + "learning_rate": 0.0001792420063467117, + "loss": 0.1221, + "step": 8916 + }, + { + "epoch": 0.20902852253614398, + "grad_norm": 0.5131996273994446, + "learning_rate": 0.00017923751295391755, + "loss": 0.0618, + "step": 8917 + }, + { + "epoch": 0.20905196411094898, + "grad_norm": 0.5946773290634155, + "learning_rate": 0.00017923301913117844, + "loss": 0.567, + "step": 8918 + }, + { + "epoch": 0.20907540568575397, + "grad_norm": 0.2547484040260315, + "learning_rate": 0.0001792285248785187, + "loss": 0.0529, + "step": 8919 + }, + { + "epoch": 0.20909884726055897, + "grad_norm": 0.7245180010795593, + "learning_rate": 0.0001792240301959628, + "loss": 0.184, + "step": 8920 + }, + { + "epoch": 0.20912228883536396, + "grad_norm": 0.5394506454467773, + "learning_rate": 0.00017921953508353506, + "loss": 0.2008, + "step": 8921 + }, + { + "epoch": 0.20914573041016896, + "grad_norm": 0.6831104159355164, + "learning_rate": 0.0001792150395412599, + "loss": 0.1242, + "step": 8922 + }, + { + "epoch": 0.20916917198497395, + "grad_norm": 0.46336790919303894, + "learning_rate": 0.0001792105435691617, + "loss": 0.147, + "step": 8923 + }, + { + "epoch": 0.20919261355977894, + "grad_norm": 0.24483822286128998, + "learning_rate": 0.0001792060471672649, + "loss": 0.0603, + "step": 8924 + }, + { + "epoch": 0.20921605513458394, + "grad_norm": 0.2496897280216217, + "learning_rate": 0.0001792015503355938, + "loss": 0.0689, + "step": 8925 + }, + { + "epoch": 0.20923949670938893, + "grad_norm": 0.4146885573863983, + "learning_rate": 0.0001791970530741729, + "loss": 0.1187, + "step": 8926 + }, + { + "epoch": 0.20926293828419393, + "grad_norm": 0.7615490555763245, + "learning_rate": 0.00017919255538302654, + "loss": 0.1744, + "step": 8927 + }, + { + "epoch": 0.20928637985899892, + "grad_norm": 0.06466952711343765, + "learning_rate": 0.00017918805726217918, + "loss": 0.0075, + "step": 8928 + }, + { + "epoch": 0.20930982143380392, + "grad_norm": 0.32179850339889526, + "learning_rate": 0.00017918355871165517, + "loss": 0.0769, + "step": 8929 + }, + { + "epoch": 0.2093332630086089, + "grad_norm": 0.5511574149131775, + "learning_rate": 0.00017917905973147895, + "loss": 0.1396, + "step": 8930 + }, + { + "epoch": 0.2093567045834139, + "grad_norm": 0.27986347675323486, + "learning_rate": 0.00017917456032167493, + "loss": 0.0753, + "step": 8931 + }, + { + "epoch": 0.2093801461582189, + "grad_norm": 0.6504880785942078, + "learning_rate": 0.00017917006048226753, + "loss": 0.1405, + "step": 8932 + }, + { + "epoch": 0.2094035877330239, + "grad_norm": 0.5968223214149475, + "learning_rate": 0.00017916556021328113, + "loss": 0.0995, + "step": 8933 + }, + { + "epoch": 0.2094270293078289, + "grad_norm": 0.6661072373390198, + "learning_rate": 0.0001791610595147402, + "loss": 0.9673, + "step": 8934 + }, + { + "epoch": 0.20945047088263388, + "grad_norm": 0.18278932571411133, + "learning_rate": 0.00017915655838666913, + "loss": 0.0232, + "step": 8935 + }, + { + "epoch": 0.2094739124574389, + "grad_norm": 0.3499493896961212, + "learning_rate": 0.00017915205682909232, + "loss": 0.0741, + "step": 8936 + }, + { + "epoch": 0.2094973540322439, + "grad_norm": 0.2704675495624542, + "learning_rate": 0.00017914755484203427, + "loss": 0.0697, + "step": 8937 + }, + { + "epoch": 0.2095207956070489, + "grad_norm": 0.8123922348022461, + "learning_rate": 0.00017914305242551932, + "loss": 0.2091, + "step": 8938 + }, + { + "epoch": 0.2095442371818539, + "grad_norm": 0.5749266743659973, + "learning_rate": 0.00017913854957957197, + "loss": 0.6725, + "step": 8939 + }, + { + "epoch": 0.20956767875665888, + "grad_norm": 0.16132879257202148, + "learning_rate": 0.0001791340463042166, + "loss": 0.032, + "step": 8940 + }, + { + "epoch": 0.20959112033146388, + "grad_norm": 0.4863854646682739, + "learning_rate": 0.0001791295425994777, + "loss": 0.1433, + "step": 8941 + }, + { + "epoch": 0.20961456190626887, + "grad_norm": 0.7418196201324463, + "learning_rate": 0.00017912503846537968, + "loss": 0.4789, + "step": 8942 + }, + { + "epoch": 0.20963800348107386, + "grad_norm": 0.16494165360927582, + "learning_rate": 0.0001791205339019469, + "loss": 0.0216, + "step": 8943 + }, + { + "epoch": 0.20966144505587886, + "grad_norm": 0.4724571704864502, + "learning_rate": 0.000179116028909204, + "loss": 0.1077, + "step": 8944 + }, + { + "epoch": 0.20968488663068385, + "grad_norm": 0.2805664539337158, + "learning_rate": 0.00017911152348717518, + "loss": 0.248, + "step": 8945 + }, + { + "epoch": 0.20970832820548885, + "grad_norm": 0.1874750256538391, + "learning_rate": 0.0001791070176358851, + "loss": 0.0363, + "step": 8946 + }, + { + "epoch": 0.20973176978029384, + "grad_norm": 0.3616485893726349, + "learning_rate": 0.00017910251135535806, + "loss": 0.1061, + "step": 8947 + }, + { + "epoch": 0.20975521135509884, + "grad_norm": 0.47502753138542175, + "learning_rate": 0.00017909800464561862, + "loss": 0.1294, + "step": 8948 + }, + { + "epoch": 0.20977865292990383, + "grad_norm": 0.48841214179992676, + "learning_rate": 0.00017909349750669116, + "loss": 0.1456, + "step": 8949 + }, + { + "epoch": 0.20980209450470882, + "grad_norm": 0.46009695529937744, + "learning_rate": 0.00017908898993860016, + "loss": 0.0624, + "step": 8950 + }, + { + "epoch": 0.20982553607951382, + "grad_norm": 0.7788807153701782, + "learning_rate": 0.00017908448194137007, + "loss": 0.114, + "step": 8951 + }, + { + "epoch": 0.2098489776543188, + "grad_norm": 0.133303701877594, + "learning_rate": 0.00017907997351502536, + "loss": 0.0304, + "step": 8952 + }, + { + "epoch": 0.2098724192291238, + "grad_norm": 0.5120387077331543, + "learning_rate": 0.00017907546465959048, + "loss": 0.1671, + "step": 8953 + }, + { + "epoch": 0.2098958608039288, + "grad_norm": 0.4778299927711487, + "learning_rate": 0.00017907095537508992, + "loss": 0.0953, + "step": 8954 + }, + { + "epoch": 0.2099193023787338, + "grad_norm": 0.4908263683319092, + "learning_rate": 0.00017906644566154813, + "loss": 0.1272, + "step": 8955 + }, + { + "epoch": 0.2099427439535388, + "grad_norm": 0.21922358870506287, + "learning_rate": 0.00017906193551898959, + "loss": 0.0351, + "step": 8956 + }, + { + "epoch": 0.20996618552834379, + "grad_norm": 0.4885677695274353, + "learning_rate": 0.00017905742494743875, + "loss": 0.1228, + "step": 8957 + }, + { + "epoch": 0.20998962710314878, + "grad_norm": 0.5228813290596008, + "learning_rate": 0.0001790529139469201, + "loss": 0.1282, + "step": 8958 + }, + { + "epoch": 0.21001306867795377, + "grad_norm": 0.4977901577949524, + "learning_rate": 0.00017904840251745812, + "loss": 0.2, + "step": 8959 + }, + { + "epoch": 0.21003651025275877, + "grad_norm": 0.6170253753662109, + "learning_rate": 0.00017904389065907727, + "loss": 0.1463, + "step": 8960 + }, + { + "epoch": 0.21005995182756376, + "grad_norm": 0.2668013572692871, + "learning_rate": 0.0001790393783718021, + "loss": 0.0461, + "step": 8961 + }, + { + "epoch": 0.21008339340236878, + "grad_norm": 0.19054648280143738, + "learning_rate": 0.000179034865655657, + "loss": 0.0305, + "step": 8962 + }, + { + "epoch": 0.21010683497717378, + "grad_norm": 0.8179089426994324, + "learning_rate": 0.0001790303525106665, + "loss": 0.1159, + "step": 8963 + }, + { + "epoch": 0.21013027655197877, + "grad_norm": 0.5773341655731201, + "learning_rate": 0.0001790258389368551, + "loss": 0.0931, + "step": 8964 + }, + { + "epoch": 0.21015371812678377, + "grad_norm": 0.5306981801986694, + "learning_rate": 0.00017902132493424723, + "loss": 0.1291, + "step": 8965 + }, + { + "epoch": 0.21017715970158876, + "grad_norm": 0.8891834616661072, + "learning_rate": 0.00017901681050286748, + "loss": 0.1309, + "step": 8966 + }, + { + "epoch": 0.21020060127639376, + "grad_norm": 0.6944369673728943, + "learning_rate": 0.00017901229564274026, + "loss": 0.1898, + "step": 8967 + }, + { + "epoch": 0.21022404285119875, + "grad_norm": 0.3309853672981262, + "learning_rate": 0.0001790077803538901, + "loss": 0.0723, + "step": 8968 + }, + { + "epoch": 0.21024748442600374, + "grad_norm": 0.5530970692634583, + "learning_rate": 0.00017900326463634155, + "loss": 0.1808, + "step": 8969 + }, + { + "epoch": 0.21027092600080874, + "grad_norm": 0.38931936025619507, + "learning_rate": 0.00017899874849011904, + "loss": 0.0666, + "step": 8970 + }, + { + "epoch": 0.21029436757561373, + "grad_norm": 0.6419046521186829, + "learning_rate": 0.00017899423191524707, + "loss": 0.1565, + "step": 8971 + }, + { + "epoch": 0.21031780915041873, + "grad_norm": 0.30420053005218506, + "learning_rate": 0.00017898971491175022, + "loss": 0.0539, + "step": 8972 + }, + { + "epoch": 0.21034125072522372, + "grad_norm": 0.7895926833152771, + "learning_rate": 0.00017898519747965293, + "loss": 0.2134, + "step": 8973 + }, + { + "epoch": 0.21036469230002872, + "grad_norm": 0.8012529611587524, + "learning_rate": 0.00017898067961897974, + "loss": 0.1657, + "step": 8974 + }, + { + "epoch": 0.2103881338748337, + "grad_norm": 0.2685394287109375, + "learning_rate": 0.00017897616132975518, + "loss": 0.0436, + "step": 8975 + }, + { + "epoch": 0.2104115754496387, + "grad_norm": 0.5417284369468689, + "learning_rate": 0.00017897164261200372, + "loss": 0.1577, + "step": 8976 + }, + { + "epoch": 0.2104350170244437, + "grad_norm": 0.8423824310302734, + "learning_rate": 0.00017896712346574992, + "loss": 0.1173, + "step": 8977 + }, + { + "epoch": 0.2104584585992487, + "grad_norm": 0.6397481560707092, + "learning_rate": 0.00017896260389101829, + "loss": 0.112, + "step": 8978 + }, + { + "epoch": 0.2104819001740537, + "grad_norm": 0.17022481560707092, + "learning_rate": 0.00017895808388783334, + "loss": 0.0299, + "step": 8979 + }, + { + "epoch": 0.21050534174885868, + "grad_norm": 0.5379864573478699, + "learning_rate": 0.00017895356345621965, + "loss": 0.1049, + "step": 8980 + }, + { + "epoch": 0.21052878332366368, + "grad_norm": 0.6442903876304626, + "learning_rate": 0.00017894904259620166, + "loss": 0.7276, + "step": 8981 + }, + { + "epoch": 0.21055222489846867, + "grad_norm": 0.3131876289844513, + "learning_rate": 0.00017894452130780394, + "loss": 0.0985, + "step": 8982 + }, + { + "epoch": 0.21057566647327366, + "grad_norm": 0.49791479110717773, + "learning_rate": 0.00017893999959105105, + "loss": 0.1618, + "step": 8983 + }, + { + "epoch": 0.21059910804807866, + "grad_norm": 0.6003378033638, + "learning_rate": 0.0001789354774459675, + "loss": 0.0918, + "step": 8984 + }, + { + "epoch": 0.21062254962288365, + "grad_norm": 0.6937844753265381, + "learning_rate": 0.00017893095487257784, + "loss": 0.7499, + "step": 8985 + }, + { + "epoch": 0.21064599119768865, + "grad_norm": 0.5843356847763062, + "learning_rate": 0.00017892643187090656, + "loss": 0.1548, + "step": 8986 + }, + { + "epoch": 0.21066943277249367, + "grad_norm": 0.2603113651275635, + "learning_rate": 0.0001789219084409783, + "loss": 0.0684, + "step": 8987 + }, + { + "epoch": 0.21069287434729866, + "grad_norm": 0.641190767288208, + "learning_rate": 0.0001789173845828175, + "loss": 0.1087, + "step": 8988 + }, + { + "epoch": 0.21071631592210366, + "grad_norm": 0.4526830315589905, + "learning_rate": 0.00017891286029644875, + "loss": 0.1119, + "step": 8989 + }, + { + "epoch": 0.21073975749690865, + "grad_norm": 0.6005496382713318, + "learning_rate": 0.0001789083355818966, + "loss": 0.1287, + "step": 8990 + }, + { + "epoch": 0.21076319907171365, + "grad_norm": 0.8683573007583618, + "learning_rate": 0.00017890381043918566, + "loss": 0.1476, + "step": 8991 + }, + { + "epoch": 0.21078664064651864, + "grad_norm": 0.45440325140953064, + "learning_rate": 0.00017889928486834037, + "loss": 0.1403, + "step": 8992 + }, + { + "epoch": 0.21081008222132364, + "grad_norm": 0.528097927570343, + "learning_rate": 0.00017889475886938535, + "loss": 0.1965, + "step": 8993 + }, + { + "epoch": 0.21083352379612863, + "grad_norm": 0.5570186376571655, + "learning_rate": 0.00017889023244234515, + "loss": 0.0911, + "step": 8994 + }, + { + "epoch": 0.21085696537093362, + "grad_norm": 0.51927250623703, + "learning_rate": 0.00017888570558724433, + "loss": 0.1078, + "step": 8995 + }, + { + "epoch": 0.21088040694573862, + "grad_norm": 0.4800107181072235, + "learning_rate": 0.0001788811783041075, + "loss": 0.0977, + "step": 8996 + }, + { + "epoch": 0.2109038485205436, + "grad_norm": 0.3379712402820587, + "learning_rate": 0.0001788766505929591, + "loss": 0.0633, + "step": 8997 + }, + { + "epoch": 0.2109272900953486, + "grad_norm": 0.5705545544624329, + "learning_rate": 0.00017887212245382383, + "loss": 0.1576, + "step": 8998 + }, + { + "epoch": 0.2109507316701536, + "grad_norm": 0.3310067355632782, + "learning_rate": 0.0001788675938867262, + "loss": 0.1059, + "step": 8999 + }, + { + "epoch": 0.2109741732449586, + "grad_norm": 0.48429074883461, + "learning_rate": 0.0001788630648916908, + "loss": 0.1475, + "step": 9000 + }, + { + "epoch": 0.2109976148197636, + "grad_norm": 1.0104094743728638, + "learning_rate": 0.00017885853546874216, + "loss": 0.2211, + "step": 9001 + }, + { + "epoch": 0.21102105639456858, + "grad_norm": 0.5613245964050293, + "learning_rate": 0.0001788540056179049, + "loss": 0.1357, + "step": 9002 + }, + { + "epoch": 0.21104449796937358, + "grad_norm": 0.4849991500377655, + "learning_rate": 0.00017884947533920357, + "loss": 0.0786, + "step": 9003 + }, + { + "epoch": 0.21106793954417857, + "grad_norm": 0.5183973908424377, + "learning_rate": 0.0001788449446326628, + "loss": 0.1334, + "step": 9004 + }, + { + "epoch": 0.21109138111898357, + "grad_norm": 0.5675663352012634, + "learning_rate": 0.00017884041349830714, + "loss": 0.0701, + "step": 9005 + }, + { + "epoch": 0.21111482269378856, + "grad_norm": 0.6690622568130493, + "learning_rate": 0.00017883588193616116, + "loss": 0.099, + "step": 9006 + }, + { + "epoch": 0.21113826426859356, + "grad_norm": 0.6792713403701782, + "learning_rate": 0.00017883134994624948, + "loss": 0.1337, + "step": 9007 + }, + { + "epoch": 0.21116170584339855, + "grad_norm": 0.19065500795841217, + "learning_rate": 0.00017882681752859667, + "loss": 0.0399, + "step": 9008 + }, + { + "epoch": 0.21118514741820354, + "grad_norm": 0.2732146084308624, + "learning_rate": 0.00017882228468322735, + "loss": 0.075, + "step": 9009 + }, + { + "epoch": 0.21120858899300854, + "grad_norm": 0.6771293878555298, + "learning_rate": 0.0001788177514101661, + "loss": 0.1296, + "step": 9010 + }, + { + "epoch": 0.21123203056781353, + "grad_norm": 0.49362388253211975, + "learning_rate": 0.0001788132177094375, + "loss": 0.0942, + "step": 9011 + }, + { + "epoch": 0.21125547214261853, + "grad_norm": 0.48486483097076416, + "learning_rate": 0.00017880868358106615, + "loss": 0.6325, + "step": 9012 + }, + { + "epoch": 0.21127891371742355, + "grad_norm": 0.5649060010910034, + "learning_rate": 0.0001788041490250767, + "loss": 0.1584, + "step": 9013 + }, + { + "epoch": 0.21130235529222854, + "grad_norm": 0.5059124827384949, + "learning_rate": 0.0001787996140414937, + "loss": 0.1534, + "step": 9014 + }, + { + "epoch": 0.21132579686703354, + "grad_norm": 1.0516419410705566, + "learning_rate": 0.0001787950786303418, + "loss": 0.2018, + "step": 9015 + }, + { + "epoch": 0.21134923844183853, + "grad_norm": 0.5639469623565674, + "learning_rate": 0.00017879054279164555, + "loss": 0.1401, + "step": 9016 + }, + { + "epoch": 0.21137268001664353, + "grad_norm": 0.7801406979560852, + "learning_rate": 0.00017878600652542962, + "loss": 1.119, + "step": 9017 + }, + { + "epoch": 0.21139612159144852, + "grad_norm": 0.39432692527770996, + "learning_rate": 0.0001787814698317186, + "loss": 0.0668, + "step": 9018 + }, + { + "epoch": 0.21141956316625352, + "grad_norm": 0.49655410647392273, + "learning_rate": 0.00017877693271053713, + "loss": 0.1306, + "step": 9019 + }, + { + "epoch": 0.2114430047410585, + "grad_norm": 0.7034127116203308, + "learning_rate": 0.0001787723951619098, + "loss": 0.1076, + "step": 9020 + }, + { + "epoch": 0.2114664463158635, + "grad_norm": 0.22100163996219635, + "learning_rate": 0.00017876785718586122, + "loss": 0.0651, + "step": 9021 + }, + { + "epoch": 0.2114898878906685, + "grad_norm": 0.8167222738265991, + "learning_rate": 0.00017876331878241604, + "loss": 0.1793, + "step": 9022 + }, + { + "epoch": 0.2115133294654735, + "grad_norm": 1.063579797744751, + "learning_rate": 0.00017875877995159888, + "loss": 0.2619, + "step": 9023 + }, + { + "epoch": 0.2115367710402785, + "grad_norm": 0.6653071641921997, + "learning_rate": 0.00017875424069343436, + "loss": 0.1294, + "step": 9024 + }, + { + "epoch": 0.21156021261508348, + "grad_norm": 0.5311400294303894, + "learning_rate": 0.00017874970100794713, + "loss": 0.0985, + "step": 9025 + }, + { + "epoch": 0.21158365418988848, + "grad_norm": 0.431695818901062, + "learning_rate": 0.0001787451608951618, + "loss": 0.1024, + "step": 9026 + }, + { + "epoch": 0.21160709576469347, + "grad_norm": 0.5809057950973511, + "learning_rate": 0.00017874062035510302, + "loss": 0.1258, + "step": 9027 + }, + { + "epoch": 0.21163053733949846, + "grad_norm": 0.5171260237693787, + "learning_rate": 0.00017873607938779541, + "loss": 0.1011, + "step": 9028 + }, + { + "epoch": 0.21165397891430346, + "grad_norm": 0.2174760103225708, + "learning_rate": 0.0001787315379932636, + "loss": 0.033, + "step": 9029 + }, + { + "epoch": 0.21167742048910845, + "grad_norm": 0.3253825604915619, + "learning_rate": 0.0001787269961715323, + "loss": 0.0712, + "step": 9030 + }, + { + "epoch": 0.21170086206391345, + "grad_norm": 0.9687361717224121, + "learning_rate": 0.00017872245392262608, + "loss": 0.0698, + "step": 9031 + }, + { + "epoch": 0.21172430363871844, + "grad_norm": 0.38126882910728455, + "learning_rate": 0.0001787179112465696, + "loss": 0.1074, + "step": 9032 + }, + { + "epoch": 0.21174774521352344, + "grad_norm": 0.5909558534622192, + "learning_rate": 0.00017871336814338751, + "loss": 0.1494, + "step": 9033 + }, + { + "epoch": 0.21177118678832843, + "grad_norm": 0.5123745799064636, + "learning_rate": 0.00017870882461310452, + "loss": 0.087, + "step": 9034 + }, + { + "epoch": 0.21179462836313342, + "grad_norm": 0.2304714322090149, + "learning_rate": 0.00017870428065574521, + "loss": 0.0463, + "step": 9035 + }, + { + "epoch": 0.21181806993793842, + "grad_norm": 0.42230382561683655, + "learning_rate": 0.00017869973627133426, + "loss": 0.1133, + "step": 9036 + }, + { + "epoch": 0.2118415115127434, + "grad_norm": 0.5043180584907532, + "learning_rate": 0.0001786951914598963, + "loss": 0.1144, + "step": 9037 + }, + { + "epoch": 0.2118649530875484, + "grad_norm": 0.4970242977142334, + "learning_rate": 0.00017869064622145605, + "loss": 0.1187, + "step": 9038 + }, + { + "epoch": 0.21188839466235343, + "grad_norm": 0.39992257952690125, + "learning_rate": 0.00017868610055603814, + "loss": 0.1049, + "step": 9039 + }, + { + "epoch": 0.21191183623715842, + "grad_norm": 0.5753785371780396, + "learning_rate": 0.0001786815544636672, + "loss": 0.1406, + "step": 9040 + }, + { + "epoch": 0.21193527781196342, + "grad_norm": 0.5189141631126404, + "learning_rate": 0.00017867700794436798, + "loss": 0.1244, + "step": 9041 + }, + { + "epoch": 0.2119587193867684, + "grad_norm": 0.37230515480041504, + "learning_rate": 0.0001786724609981651, + "loss": 0.0599, + "step": 9042 + }, + { + "epoch": 0.2119821609615734, + "grad_norm": 0.40537142753601074, + "learning_rate": 0.00017866791362508322, + "loss": 0.1001, + "step": 9043 + }, + { + "epoch": 0.2120056025363784, + "grad_norm": 0.309272438287735, + "learning_rate": 0.00017866336582514702, + "loss": 0.067, + "step": 9044 + }, + { + "epoch": 0.2120290441111834, + "grad_norm": 0.4707411527633667, + "learning_rate": 0.00017865881759838122, + "loss": 0.1138, + "step": 9045 + }, + { + "epoch": 0.2120524856859884, + "grad_norm": 0.6122572422027588, + "learning_rate": 0.00017865426894481042, + "loss": 0.1153, + "step": 9046 + }, + { + "epoch": 0.21207592726079338, + "grad_norm": 0.5262412428855896, + "learning_rate": 0.00017864971986445935, + "loss": 0.1557, + "step": 9047 + }, + { + "epoch": 0.21209936883559838, + "grad_norm": 0.698156476020813, + "learning_rate": 0.00017864517035735272, + "loss": 0.7126, + "step": 9048 + }, + { + "epoch": 0.21212281041040337, + "grad_norm": 0.2269383668899536, + "learning_rate": 0.00017864062042351515, + "loss": 0.0372, + "step": 9049 + }, + { + "epoch": 0.21214625198520837, + "grad_norm": 0.582976758480072, + "learning_rate": 0.0001786360700629714, + "loss": 0.1118, + "step": 9050 + }, + { + "epoch": 0.21216969356001336, + "grad_norm": 0.60455721616745, + "learning_rate": 0.00017863151927574608, + "loss": 0.1847, + "step": 9051 + }, + { + "epoch": 0.21219313513481836, + "grad_norm": 0.9063741564750671, + "learning_rate": 0.00017862696806186397, + "loss": 0.3101, + "step": 9052 + }, + { + "epoch": 0.21221657670962335, + "grad_norm": 0.388729065656662, + "learning_rate": 0.00017862241642134969, + "loss": 0.0574, + "step": 9053 + }, + { + "epoch": 0.21224001828442834, + "grad_norm": 0.24130307137966156, + "learning_rate": 0.00017861786435422795, + "loss": 0.0227, + "step": 9054 + }, + { + "epoch": 0.21226345985923334, + "grad_norm": 0.4851987659931183, + "learning_rate": 0.0001786133118605235, + "loss": 0.1162, + "step": 9055 + }, + { + "epoch": 0.21228690143403833, + "grad_norm": 0.5732377767562866, + "learning_rate": 0.000178608758940261, + "loss": 0.1458, + "step": 9056 + }, + { + "epoch": 0.21231034300884333, + "grad_norm": 0.17309673130512238, + "learning_rate": 0.00017860420559346517, + "loss": 0.0563, + "step": 9057 + }, + { + "epoch": 0.21233378458364832, + "grad_norm": 0.7822044491767883, + "learning_rate": 0.00017859965182016071, + "loss": 0.3098, + "step": 9058 + }, + { + "epoch": 0.21235722615845332, + "grad_norm": 0.15998093783855438, + "learning_rate": 0.00017859509762037235, + "loss": 0.049, + "step": 9059 + }, + { + "epoch": 0.2123806677332583, + "grad_norm": 0.4390997886657715, + "learning_rate": 0.00017859054299412472, + "loss": 0.0609, + "step": 9060 + }, + { + "epoch": 0.2124041093080633, + "grad_norm": 0.7180513143539429, + "learning_rate": 0.00017858598794144267, + "loss": 0.2132, + "step": 9061 + }, + { + "epoch": 0.2124275508828683, + "grad_norm": 0.5709378123283386, + "learning_rate": 0.00017858143246235076, + "loss": 0.1835, + "step": 9062 + }, + { + "epoch": 0.2124509924576733, + "grad_norm": 0.6708611249923706, + "learning_rate": 0.00017857687655687383, + "loss": 0.6984, + "step": 9063 + }, + { + "epoch": 0.21247443403247832, + "grad_norm": 0.5559596419334412, + "learning_rate": 0.0001785723202250366, + "loss": 0.1352, + "step": 9064 + }, + { + "epoch": 0.2124978756072833, + "grad_norm": 0.5111309289932251, + "learning_rate": 0.00017856776346686368, + "loss": 0.2115, + "step": 9065 + }, + { + "epoch": 0.2125213171820883, + "grad_norm": 0.41765961050987244, + "learning_rate": 0.00017856320628237987, + "loss": 0.1175, + "step": 9066 + }, + { + "epoch": 0.2125447587568933, + "grad_norm": 1.172131061553955, + "learning_rate": 0.00017855864867160994, + "loss": 0.1469, + "step": 9067 + }, + { + "epoch": 0.2125682003316983, + "grad_norm": 0.4147167503833771, + "learning_rate": 0.00017855409063457852, + "loss": 0.0834, + "step": 9068 + }, + { + "epoch": 0.2125916419065033, + "grad_norm": 0.5439594984054565, + "learning_rate": 0.00017854953217131045, + "loss": 0.2087, + "step": 9069 + }, + { + "epoch": 0.21261508348130828, + "grad_norm": 0.6685320138931274, + "learning_rate": 0.00017854497328183036, + "loss": 0.1634, + "step": 9070 + }, + { + "epoch": 0.21263852505611328, + "grad_norm": 0.6089337468147278, + "learning_rate": 0.0001785404139661631, + "loss": 0.1349, + "step": 9071 + }, + { + "epoch": 0.21266196663091827, + "grad_norm": 0.17848992347717285, + "learning_rate": 0.0001785358542243333, + "loss": 0.0324, + "step": 9072 + }, + { + "epoch": 0.21268540820572326, + "grad_norm": 0.4884072542190552, + "learning_rate": 0.00017853129405636574, + "loss": 0.7701, + "step": 9073 + }, + { + "epoch": 0.21270884978052826, + "grad_norm": 0.49944397807121277, + "learning_rate": 0.0001785267334622852, + "loss": 0.0827, + "step": 9074 + }, + { + "epoch": 0.21273229135533325, + "grad_norm": 0.4543008804321289, + "learning_rate": 0.0001785221724421164, + "loss": 0.0797, + "step": 9075 + }, + { + "epoch": 0.21275573293013825, + "grad_norm": 0.299651175737381, + "learning_rate": 0.00017851761099588404, + "loss": 0.0651, + "step": 9076 + }, + { + "epoch": 0.21277917450494324, + "grad_norm": 0.22216889262199402, + "learning_rate": 0.00017851304912361296, + "loss": 0.0361, + "step": 9077 + }, + { + "epoch": 0.21280261607974824, + "grad_norm": 0.1682184636592865, + "learning_rate": 0.00017850848682532785, + "loss": 0.0251, + "step": 9078 + }, + { + "epoch": 0.21282605765455323, + "grad_norm": 0.8718963265419006, + "learning_rate": 0.00017850392410105346, + "loss": 0.2443, + "step": 9079 + }, + { + "epoch": 0.21284949922935822, + "grad_norm": 0.2639635503292084, + "learning_rate": 0.00017849936095081458, + "loss": 0.0481, + "step": 9080 + }, + { + "epoch": 0.21287294080416322, + "grad_norm": 0.5481154322624207, + "learning_rate": 0.00017849479737463598, + "loss": 0.1593, + "step": 9081 + }, + { + "epoch": 0.2128963823789682, + "grad_norm": 0.8553960919380188, + "learning_rate": 0.0001784902333725424, + "loss": 0.1465, + "step": 9082 + }, + { + "epoch": 0.2129198239537732, + "grad_norm": 0.6736152768135071, + "learning_rate": 0.0001784856689445586, + "loss": 0.2539, + "step": 9083 + }, + { + "epoch": 0.2129432655285782, + "grad_norm": 0.4887605607509613, + "learning_rate": 0.00017848110409070936, + "loss": 0.1589, + "step": 9084 + }, + { + "epoch": 0.2129667071033832, + "grad_norm": 0.6684427261352539, + "learning_rate": 0.00017847653881101943, + "loss": 0.6901, + "step": 9085 + }, + { + "epoch": 0.2129901486781882, + "grad_norm": 0.4184075891971588, + "learning_rate": 0.0001784719731055136, + "loss": 0.084, + "step": 9086 + }, + { + "epoch": 0.21301359025299318, + "grad_norm": 0.31855833530426025, + "learning_rate": 0.00017846740697421664, + "loss": 0.0679, + "step": 9087 + }, + { + "epoch": 0.21303703182779818, + "grad_norm": 0.6275452971458435, + "learning_rate": 0.00017846284041715332, + "loss": 0.1836, + "step": 9088 + }, + { + "epoch": 0.21306047340260317, + "grad_norm": 0.5681195259094238, + "learning_rate": 0.00017845827343434842, + "loss": 0.7724, + "step": 9089 + }, + { + "epoch": 0.2130839149774082, + "grad_norm": 0.1483229398727417, + "learning_rate": 0.00017845370602582677, + "loss": 0.0302, + "step": 9090 + }, + { + "epoch": 0.2131073565522132, + "grad_norm": 0.4666050672531128, + "learning_rate": 0.00017844913819161303, + "loss": 0.1111, + "step": 9091 + }, + { + "epoch": 0.21313079812701818, + "grad_norm": 0.24507687985897064, + "learning_rate": 0.00017844456993173209, + "loss": 0.0328, + "step": 9092 + }, + { + "epoch": 0.21315423970182318, + "grad_norm": 0.13438265025615692, + "learning_rate": 0.0001784400012462087, + "loss": 0.0279, + "step": 9093 + }, + { + "epoch": 0.21317768127662817, + "grad_norm": 0.34775310754776, + "learning_rate": 0.00017843543213506767, + "loss": 0.027, + "step": 9094 + }, + { + "epoch": 0.21320112285143317, + "grad_norm": 0.9870791435241699, + "learning_rate": 0.00017843086259833378, + "loss": 0.1391, + "step": 9095 + }, + { + "epoch": 0.21322456442623816, + "grad_norm": 0.6866376996040344, + "learning_rate": 0.00017842629263603184, + "loss": 0.1341, + "step": 9096 + }, + { + "epoch": 0.21324800600104316, + "grad_norm": 0.6748077273368835, + "learning_rate": 0.00017842172224818658, + "loss": 0.0628, + "step": 9097 + }, + { + "epoch": 0.21327144757584815, + "grad_norm": 0.33803483843803406, + "learning_rate": 0.0001784171514348229, + "loss": 0.0839, + "step": 9098 + }, + { + "epoch": 0.21329488915065314, + "grad_norm": 0.5298961400985718, + "learning_rate": 0.00017841258019596554, + "loss": 0.1542, + "step": 9099 + }, + { + "epoch": 0.21331833072545814, + "grad_norm": 0.6111764311790466, + "learning_rate": 0.0001784080085316393, + "loss": 0.1688, + "step": 9100 + }, + { + "epoch": 0.21334177230026313, + "grad_norm": 0.6206469535827637, + "learning_rate": 0.00017840343644186902, + "loss": 0.1076, + "step": 9101 + }, + { + "epoch": 0.21336521387506813, + "grad_norm": 0.7481014728546143, + "learning_rate": 0.00017839886392667949, + "loss": 0.1569, + "step": 9102 + }, + { + "epoch": 0.21338865544987312, + "grad_norm": 0.503788948059082, + "learning_rate": 0.00017839429098609548, + "loss": 0.0989, + "step": 9103 + }, + { + "epoch": 0.21341209702467812, + "grad_norm": 0.43106383085250854, + "learning_rate": 0.0001783897176201419, + "loss": 0.0783, + "step": 9104 + }, + { + "epoch": 0.2134355385994831, + "grad_norm": 0.6517298221588135, + "learning_rate": 0.00017838514382884348, + "loss": 0.0837, + "step": 9105 + }, + { + "epoch": 0.2134589801742881, + "grad_norm": 0.6661410927772522, + "learning_rate": 0.00017838056961222507, + "loss": 0.1738, + "step": 9106 + }, + { + "epoch": 0.2134824217490931, + "grad_norm": 0.29566633701324463, + "learning_rate": 0.00017837599497031148, + "loss": 0.0466, + "step": 9107 + }, + { + "epoch": 0.2135058633238981, + "grad_norm": 0.5811222791671753, + "learning_rate": 0.00017837141990312754, + "loss": 0.1191, + "step": 9108 + }, + { + "epoch": 0.2135293048987031, + "grad_norm": 0.6828871369361877, + "learning_rate": 0.00017836684441069805, + "loss": 0.1045, + "step": 9109 + }, + { + "epoch": 0.21355274647350808, + "grad_norm": 0.4290938675403595, + "learning_rate": 0.0001783622684930479, + "loss": 0.0982, + "step": 9110 + }, + { + "epoch": 0.21357618804831308, + "grad_norm": 0.6101336479187012, + "learning_rate": 0.00017835769215020185, + "loss": 0.1404, + "step": 9111 + }, + { + "epoch": 0.21359962962311807, + "grad_norm": 0.5291793346405029, + "learning_rate": 0.00017835311538218475, + "loss": 0.0941, + "step": 9112 + }, + { + "epoch": 0.21362307119792306, + "grad_norm": 0.599776566028595, + "learning_rate": 0.00017834853818902147, + "loss": 0.1193, + "step": 9113 + }, + { + "epoch": 0.21364651277272806, + "grad_norm": 0.44390547275543213, + "learning_rate": 0.0001783439605707368, + "loss": 0.1562, + "step": 9114 + }, + { + "epoch": 0.21366995434753308, + "grad_norm": 0.15110257267951965, + "learning_rate": 0.0001783393825273556, + "loss": 0.0259, + "step": 9115 + }, + { + "epoch": 0.21369339592233808, + "grad_norm": 0.15490232408046722, + "learning_rate": 0.00017833480405890273, + "loss": 0.0373, + "step": 9116 + }, + { + "epoch": 0.21371683749714307, + "grad_norm": 0.7137459516525269, + "learning_rate": 0.00017833022516540301, + "loss": 0.2046, + "step": 9117 + }, + { + "epoch": 0.21374027907194806, + "grad_norm": 0.2956506907939911, + "learning_rate": 0.00017832564584688125, + "loss": 0.08, + "step": 9118 + }, + { + "epoch": 0.21376372064675306, + "grad_norm": 0.3258841931819916, + "learning_rate": 0.00017832106610336234, + "loss": 0.0598, + "step": 9119 + }, + { + "epoch": 0.21378716222155805, + "grad_norm": 0.4447345733642578, + "learning_rate": 0.00017831648593487115, + "loss": 0.117, + "step": 9120 + }, + { + "epoch": 0.21381060379636305, + "grad_norm": 0.7992703914642334, + "learning_rate": 0.00017831190534143247, + "loss": 0.1641, + "step": 9121 + }, + { + "epoch": 0.21383404537116804, + "grad_norm": 0.160650372505188, + "learning_rate": 0.00017830732432307122, + "loss": 0.0412, + "step": 9122 + }, + { + "epoch": 0.21385748694597304, + "grad_norm": 0.383186012506485, + "learning_rate": 0.00017830274287981225, + "loss": 0.1173, + "step": 9123 + }, + { + "epoch": 0.21388092852077803, + "grad_norm": 0.33689212799072266, + "learning_rate": 0.00017829816101168033, + "loss": 0.0769, + "step": 9124 + }, + { + "epoch": 0.21390437009558302, + "grad_norm": 0.2608790993690491, + "learning_rate": 0.00017829357871870043, + "loss": 0.0452, + "step": 9125 + }, + { + "epoch": 0.21392781167038802, + "grad_norm": 0.6297788023948669, + "learning_rate": 0.00017828899600089737, + "loss": 0.0859, + "step": 9126 + }, + { + "epoch": 0.213951253245193, + "grad_norm": 0.6116383671760559, + "learning_rate": 0.000178284412858296, + "loss": 0.0855, + "step": 9127 + }, + { + "epoch": 0.213974694819998, + "grad_norm": 0.2223491221666336, + "learning_rate": 0.00017827982929092122, + "loss": 0.0577, + "step": 9128 + }, + { + "epoch": 0.213998136394803, + "grad_norm": 0.6941012740135193, + "learning_rate": 0.00017827524529879786, + "loss": 0.1537, + "step": 9129 + }, + { + "epoch": 0.214021577969608, + "grad_norm": 0.2042597234249115, + "learning_rate": 0.00017827066088195085, + "loss": 0.0457, + "step": 9130 + }, + { + "epoch": 0.214045019544413, + "grad_norm": 0.6362863779067993, + "learning_rate": 0.000178266076040405, + "loss": 0.1527, + "step": 9131 + }, + { + "epoch": 0.21406846111921798, + "grad_norm": 0.830958366394043, + "learning_rate": 0.00017826149077418527, + "loss": 0.2062, + "step": 9132 + }, + { + "epoch": 0.21409190269402298, + "grad_norm": 0.2278081327676773, + "learning_rate": 0.00017825690508331644, + "loss": 0.0417, + "step": 9133 + }, + { + "epoch": 0.21411534426882797, + "grad_norm": 0.6239277124404907, + "learning_rate": 0.00017825231896782343, + "loss": 0.1746, + "step": 9134 + }, + { + "epoch": 0.21413878584363297, + "grad_norm": 0.7837158441543579, + "learning_rate": 0.00017824773242773118, + "loss": 0.1115, + "step": 9135 + }, + { + "epoch": 0.21416222741843796, + "grad_norm": 0.4224492907524109, + "learning_rate": 0.00017824314546306452, + "loss": 0.1251, + "step": 9136 + }, + { + "epoch": 0.21418566899324296, + "grad_norm": 0.5976373553276062, + "learning_rate": 0.00017823855807384836, + "loss": 0.212, + "step": 9137 + }, + { + "epoch": 0.21420911056804795, + "grad_norm": 0.2749262750148773, + "learning_rate": 0.0001782339702601076, + "loss": 0.0647, + "step": 9138 + }, + { + "epoch": 0.21423255214285294, + "grad_norm": 0.7072374820709229, + "learning_rate": 0.0001782293820218671, + "loss": 0.2042, + "step": 9139 + }, + { + "epoch": 0.21425599371765794, + "grad_norm": 0.6351097226142883, + "learning_rate": 0.00017822479335915175, + "loss": 0.2088, + "step": 9140 + }, + { + "epoch": 0.21427943529246296, + "grad_norm": 1.0056241750717163, + "learning_rate": 0.0001782202042719865, + "loss": 0.1478, + "step": 9141 + }, + { + "epoch": 0.21430287686726796, + "grad_norm": 0.5127723813056946, + "learning_rate": 0.00017821561476039622, + "loss": 0.146, + "step": 9142 + }, + { + "epoch": 0.21432631844207295, + "grad_norm": 0.17142029106616974, + "learning_rate": 0.0001782110248244058, + "loss": 0.055, + "step": 9143 + }, + { + "epoch": 0.21434976001687794, + "grad_norm": 0.30604150891304016, + "learning_rate": 0.00017820643446404016, + "loss": 0.0914, + "step": 9144 + }, + { + "epoch": 0.21437320159168294, + "grad_norm": 0.26963841915130615, + "learning_rate": 0.00017820184367932422, + "loss": 0.059, + "step": 9145 + }, + { + "epoch": 0.21439664316648793, + "grad_norm": 0.18112987279891968, + "learning_rate": 0.00017819725247028288, + "loss": 0.0379, + "step": 9146 + }, + { + "epoch": 0.21442008474129293, + "grad_norm": 0.5723358988761902, + "learning_rate": 0.00017819266083694103, + "loss": 0.1016, + "step": 9147 + }, + { + "epoch": 0.21444352631609792, + "grad_norm": 0.5520337224006653, + "learning_rate": 0.00017818806877932365, + "loss": 0.1986, + "step": 9148 + }, + { + "epoch": 0.21446696789090292, + "grad_norm": 0.5592267513275146, + "learning_rate": 0.00017818347629745555, + "loss": 0.1734, + "step": 9149 + }, + { + "epoch": 0.2144904094657079, + "grad_norm": 0.3239561915397644, + "learning_rate": 0.00017817888339136174, + "loss": 0.0575, + "step": 9150 + }, + { + "epoch": 0.2145138510405129, + "grad_norm": 0.1393336057662964, + "learning_rate": 0.00017817429006106712, + "loss": 0.0505, + "step": 9151 + }, + { + "epoch": 0.2145372926153179, + "grad_norm": 0.3919665217399597, + "learning_rate": 0.00017816969630659658, + "loss": 0.1402, + "step": 9152 + }, + { + "epoch": 0.2145607341901229, + "grad_norm": 0.5993861556053162, + "learning_rate": 0.00017816510212797512, + "loss": 0.8066, + "step": 9153 + }, + { + "epoch": 0.2145841757649279, + "grad_norm": 0.23582400381565094, + "learning_rate": 0.00017816050752522756, + "loss": 0.0506, + "step": 9154 + }, + { + "epoch": 0.21460761733973288, + "grad_norm": 0.8581938743591309, + "learning_rate": 0.00017815591249837888, + "loss": 0.1461, + "step": 9155 + }, + { + "epoch": 0.21463105891453788, + "grad_norm": 0.6587904095649719, + "learning_rate": 0.00017815131704745408, + "loss": 0.1033, + "step": 9156 + }, + { + "epoch": 0.21465450048934287, + "grad_norm": 0.5793978571891785, + "learning_rate": 0.000178146721172478, + "loss": 0.1381, + "step": 9157 + }, + { + "epoch": 0.21467794206414786, + "grad_norm": 0.4783765971660614, + "learning_rate": 0.0001781421248734756, + "loss": 0.1138, + "step": 9158 + }, + { + "epoch": 0.21470138363895286, + "grad_norm": 0.3805655539035797, + "learning_rate": 0.00017813752815047188, + "loss": 0.0848, + "step": 9159 + }, + { + "epoch": 0.21472482521375785, + "grad_norm": 0.5045800805091858, + "learning_rate": 0.00017813293100349167, + "loss": 0.1387, + "step": 9160 + }, + { + "epoch": 0.21474826678856285, + "grad_norm": 0.39452436566352844, + "learning_rate": 0.00017812833343256004, + "loss": 0.0635, + "step": 9161 + }, + { + "epoch": 0.21477170836336784, + "grad_norm": 0.6784062385559082, + "learning_rate": 0.00017812373543770184, + "loss": 0.1502, + "step": 9162 + }, + { + "epoch": 0.21479514993817284, + "grad_norm": 0.47182679176330566, + "learning_rate": 0.00017811913701894208, + "loss": 0.0948, + "step": 9163 + }, + { + "epoch": 0.21481859151297783, + "grad_norm": 0.7966069579124451, + "learning_rate": 0.00017811453817630568, + "loss": 0.1276, + "step": 9164 + }, + { + "epoch": 0.21484203308778282, + "grad_norm": 0.3504107892513275, + "learning_rate": 0.00017810993890981757, + "loss": 0.0603, + "step": 9165 + }, + { + "epoch": 0.21486547466258785, + "grad_norm": 0.308706670999527, + "learning_rate": 0.0001781053392195028, + "loss": 0.0638, + "step": 9166 + }, + { + "epoch": 0.21488891623739284, + "grad_norm": 0.4680926501750946, + "learning_rate": 0.0001781007391053862, + "loss": 0.1246, + "step": 9167 + }, + { + "epoch": 0.21491235781219784, + "grad_norm": 0.6066713929176331, + "learning_rate": 0.0001780961385674928, + "loss": 0.1337, + "step": 9168 + }, + { + "epoch": 0.21493579938700283, + "grad_norm": 0.495093435049057, + "learning_rate": 0.00017809153760584757, + "loss": 0.7362, + "step": 9169 + }, + { + "epoch": 0.21495924096180782, + "grad_norm": 0.6644469499588013, + "learning_rate": 0.00017808693622047548, + "loss": 0.1867, + "step": 9170 + }, + { + "epoch": 0.21498268253661282, + "grad_norm": 0.38951802253723145, + "learning_rate": 0.00017808233441140146, + "loss": 0.0551, + "step": 9171 + }, + { + "epoch": 0.2150061241114178, + "grad_norm": 0.4398663341999054, + "learning_rate": 0.0001780777321786505, + "loss": 0.0705, + "step": 9172 + }, + { + "epoch": 0.2150295656862228, + "grad_norm": 0.43554240465164185, + "learning_rate": 0.00017807312952224758, + "loss": 0.0902, + "step": 9173 + }, + { + "epoch": 0.2150530072610278, + "grad_norm": 0.4508645534515381, + "learning_rate": 0.00017806852644221765, + "loss": 0.1146, + "step": 9174 + }, + { + "epoch": 0.2150764488358328, + "grad_norm": 0.6440017819404602, + "learning_rate": 0.0001780639229385857, + "loss": 0.151, + "step": 9175 + }, + { + "epoch": 0.2150998904106378, + "grad_norm": 0.5501804947853088, + "learning_rate": 0.00017805931901137672, + "loss": 0.1443, + "step": 9176 + }, + { + "epoch": 0.21512333198544278, + "grad_norm": 0.1642502397298813, + "learning_rate": 0.00017805471466061568, + "loss": 0.0324, + "step": 9177 + }, + { + "epoch": 0.21514677356024778, + "grad_norm": 0.3863295018672943, + "learning_rate": 0.00017805010988632757, + "loss": 0.0636, + "step": 9178 + }, + { + "epoch": 0.21517021513505277, + "grad_norm": 0.3076837360858917, + "learning_rate": 0.00017804550468853735, + "loss": 0.0835, + "step": 9179 + }, + { + "epoch": 0.21519365670985777, + "grad_norm": 0.7204993367195129, + "learning_rate": 0.00017804089906727, + "loss": 0.1641, + "step": 9180 + }, + { + "epoch": 0.21521709828466276, + "grad_norm": 0.12320862710475922, + "learning_rate": 0.00017803629302255057, + "loss": 0.0199, + "step": 9181 + }, + { + "epoch": 0.21524053985946776, + "grad_norm": 0.23650628328323364, + "learning_rate": 0.00017803168655440403, + "loss": 0.0264, + "step": 9182 + }, + { + "epoch": 0.21526398143427275, + "grad_norm": 0.20116287469863892, + "learning_rate": 0.00017802707966285537, + "loss": 0.0329, + "step": 9183 + }, + { + "epoch": 0.21528742300907774, + "grad_norm": 0.7137426137924194, + "learning_rate": 0.00017802247234792958, + "loss": 0.1261, + "step": 9184 + }, + { + "epoch": 0.21531086458388274, + "grad_norm": 0.7032276391983032, + "learning_rate": 0.00017801786460965163, + "loss": 0.5712, + "step": 9185 + }, + { + "epoch": 0.21533430615868773, + "grad_norm": 0.6268220543861389, + "learning_rate": 0.00017801325644804656, + "loss": 0.1105, + "step": 9186 + }, + { + "epoch": 0.21535774773349273, + "grad_norm": 0.5126611590385437, + "learning_rate": 0.00017800864786313936, + "loss": 0.0992, + "step": 9187 + }, + { + "epoch": 0.21538118930829772, + "grad_norm": 0.4742869734764099, + "learning_rate": 0.00017800403885495509, + "loss": 0.1259, + "step": 9188 + }, + { + "epoch": 0.21540463088310272, + "grad_norm": 1.115169644355774, + "learning_rate": 0.0001779994294235187, + "loss": 0.2193, + "step": 9189 + }, + { + "epoch": 0.2154280724579077, + "grad_norm": 0.6577510833740234, + "learning_rate": 0.00017799481956885517, + "loss": 0.1991, + "step": 9190 + }, + { + "epoch": 0.2154515140327127, + "grad_norm": 0.5955620408058167, + "learning_rate": 0.00017799020929098956, + "loss": 0.696, + "step": 9191 + }, + { + "epoch": 0.21547495560751773, + "grad_norm": 0.46764567494392395, + "learning_rate": 0.00017798559858994692, + "loss": 0.124, + "step": 9192 + }, + { + "epoch": 0.21549839718232272, + "grad_norm": 0.49791112542152405, + "learning_rate": 0.00017798098746575223, + "loss": 0.0837, + "step": 9193 + }, + { + "epoch": 0.21552183875712771, + "grad_norm": 0.8451802730560303, + "learning_rate": 0.0001779763759184305, + "loss": 0.1643, + "step": 9194 + }, + { + "epoch": 0.2155452803319327, + "grad_norm": 0.47412949800491333, + "learning_rate": 0.00017797176394800674, + "loss": 0.1084, + "step": 9195 + }, + { + "epoch": 0.2155687219067377, + "grad_norm": 0.42174145579338074, + "learning_rate": 0.00017796715155450602, + "loss": 0.1409, + "step": 9196 + }, + { + "epoch": 0.2155921634815427, + "grad_norm": 0.2329004853963852, + "learning_rate": 0.00017796253873795332, + "loss": 0.0634, + "step": 9197 + }, + { + "epoch": 0.2156156050563477, + "grad_norm": 0.2331550568342209, + "learning_rate": 0.0001779579254983737, + "loss": 0.0392, + "step": 9198 + }, + { + "epoch": 0.2156390466311527, + "grad_norm": 0.20381228625774384, + "learning_rate": 0.0001779533118357922, + "loss": 0.0567, + "step": 9199 + }, + { + "epoch": 0.21566248820595768, + "grad_norm": 0.6749442219734192, + "learning_rate": 0.00017794869775023382, + "loss": 0.1973, + "step": 9200 + }, + { + "epoch": 0.21568592978076268, + "grad_norm": 0.6076815724372864, + "learning_rate": 0.00017794408324172363, + "loss": 0.2104, + "step": 9201 + }, + { + "epoch": 0.21570937135556767, + "grad_norm": 1.0667072534561157, + "learning_rate": 0.00017793946831028666, + "loss": 0.0441, + "step": 9202 + }, + { + "epoch": 0.21573281293037266, + "grad_norm": 0.49468883872032166, + "learning_rate": 0.0001779348529559479, + "loss": 0.143, + "step": 9203 + }, + { + "epoch": 0.21575625450517766, + "grad_norm": 0.5759472846984863, + "learning_rate": 0.0001779302371787325, + "loss": 0.1861, + "step": 9204 + }, + { + "epoch": 0.21577969607998265, + "grad_norm": 0.915363073348999, + "learning_rate": 0.0001779256209786654, + "loss": 0.3408, + "step": 9205 + }, + { + "epoch": 0.21580313765478765, + "grad_norm": 0.5399860739707947, + "learning_rate": 0.0001779210043557717, + "loss": 0.1221, + "step": 9206 + }, + { + "epoch": 0.21582657922959264, + "grad_norm": 0.48389461636543274, + "learning_rate": 0.00017791638731007642, + "loss": 0.0964, + "step": 9207 + }, + { + "epoch": 0.21585002080439764, + "grad_norm": 0.2633078992366791, + "learning_rate": 0.00017791176984160466, + "loss": 0.0565, + "step": 9208 + }, + { + "epoch": 0.21587346237920263, + "grad_norm": 0.3838725984096527, + "learning_rate": 0.00017790715195038144, + "loss": 0.0481, + "step": 9209 + }, + { + "epoch": 0.21589690395400762, + "grad_norm": 0.4487365782260895, + "learning_rate": 0.0001779025336364318, + "loss": 0.1105, + "step": 9210 + }, + { + "epoch": 0.21592034552881262, + "grad_norm": 0.5455501079559326, + "learning_rate": 0.00017789791489978085, + "loss": 0.1404, + "step": 9211 + }, + { + "epoch": 0.2159437871036176, + "grad_norm": 0.364915668964386, + "learning_rate": 0.00017789329574045362, + "loss": 0.132, + "step": 9212 + }, + { + "epoch": 0.2159672286784226, + "grad_norm": 0.32485467195510864, + "learning_rate": 0.00017788867615847515, + "loss": 0.1016, + "step": 9213 + }, + { + "epoch": 0.2159906702532276, + "grad_norm": 0.3674456775188446, + "learning_rate": 0.00017788405615387057, + "loss": 0.0805, + "step": 9214 + }, + { + "epoch": 0.2160141118280326, + "grad_norm": 0.43305861949920654, + "learning_rate": 0.00017787943572666488, + "loss": 0.143, + "step": 9215 + }, + { + "epoch": 0.2160375534028376, + "grad_norm": 0.3606945276260376, + "learning_rate": 0.0001778748148768832, + "loss": 0.0739, + "step": 9216 + }, + { + "epoch": 0.21606099497764258, + "grad_norm": 0.8950464725494385, + "learning_rate": 0.0001778701936045506, + "loss": 0.1795, + "step": 9217 + }, + { + "epoch": 0.2160844365524476, + "grad_norm": 0.5325526595115662, + "learning_rate": 0.0001778655719096921, + "loss": 0.7263, + "step": 9218 + }, + { + "epoch": 0.2161078781272526, + "grad_norm": 0.5433641076087952, + "learning_rate": 0.00017786094979233284, + "loss": 0.1115, + "step": 9219 + }, + { + "epoch": 0.2161313197020576, + "grad_norm": 0.6593860387802124, + "learning_rate": 0.00017785632725249789, + "loss": 0.1317, + "step": 9220 + }, + { + "epoch": 0.2161547612768626, + "grad_norm": 0.435111939907074, + "learning_rate": 0.0001778517042902123, + "loss": 0.0788, + "step": 9221 + }, + { + "epoch": 0.21617820285166758, + "grad_norm": 0.7629940509796143, + "learning_rate": 0.0001778470809055012, + "loss": 0.2057, + "step": 9222 + }, + { + "epoch": 0.21620164442647258, + "grad_norm": 0.9477643370628357, + "learning_rate": 0.0001778424570983896, + "loss": 0.1323, + "step": 9223 + }, + { + "epoch": 0.21622508600127757, + "grad_norm": 0.6780590415000916, + "learning_rate": 0.00017783783286890267, + "loss": 0.1198, + "step": 9224 + }, + { + "epoch": 0.21624852757608257, + "grad_norm": 0.34811103343963623, + "learning_rate": 0.00017783320821706546, + "loss": 0.0854, + "step": 9225 + }, + { + "epoch": 0.21627196915088756, + "grad_norm": 0.5029367208480835, + "learning_rate": 0.0001778285831429031, + "loss": 0.6512, + "step": 9226 + }, + { + "epoch": 0.21629541072569256, + "grad_norm": 0.18026387691497803, + "learning_rate": 0.00017782395764644063, + "loss": 0.0315, + "step": 9227 + }, + { + "epoch": 0.21631885230049755, + "grad_norm": 0.3394416570663452, + "learning_rate": 0.0001778193317277032, + "loss": 0.0573, + "step": 9228 + }, + { + "epoch": 0.21634229387530254, + "grad_norm": 0.486545205116272, + "learning_rate": 0.00017781470538671588, + "loss": 0.102, + "step": 9229 + }, + { + "epoch": 0.21636573545010754, + "grad_norm": 1.0655509233474731, + "learning_rate": 0.0001778100786235038, + "loss": 0.2125, + "step": 9230 + }, + { + "epoch": 0.21638917702491253, + "grad_norm": 0.5848194360733032, + "learning_rate": 0.000177805451438092, + "loss": 0.0909, + "step": 9231 + }, + { + "epoch": 0.21641261859971753, + "grad_norm": 0.212505504488945, + "learning_rate": 0.00017780082383050564, + "loss": 0.0243, + "step": 9232 + }, + { + "epoch": 0.21643606017452252, + "grad_norm": 0.5611519813537598, + "learning_rate": 0.00017779619580076986, + "loss": 0.1188, + "step": 9233 + }, + { + "epoch": 0.21645950174932752, + "grad_norm": 0.8316395282745361, + "learning_rate": 0.00017779156734890972, + "loss": 0.1123, + "step": 9234 + }, + { + "epoch": 0.2164829433241325, + "grad_norm": 0.7622222304344177, + "learning_rate": 0.00017778693847495032, + "loss": 0.1707, + "step": 9235 + }, + { + "epoch": 0.2165063848989375, + "grad_norm": 0.4052189886569977, + "learning_rate": 0.00017778230917891682, + "loss": 0.0571, + "step": 9236 + }, + { + "epoch": 0.2165298264737425, + "grad_norm": 0.6876880526542664, + "learning_rate": 0.00017777767946083433, + "loss": 0.1712, + "step": 9237 + }, + { + "epoch": 0.2165532680485475, + "grad_norm": 0.37457525730133057, + "learning_rate": 0.00017777304932072797, + "loss": 0.0974, + "step": 9238 + }, + { + "epoch": 0.2165767096233525, + "grad_norm": 0.35032108426094055, + "learning_rate": 0.00017776841875862288, + "loss": 0.0732, + "step": 9239 + }, + { + "epoch": 0.21660015119815748, + "grad_norm": 0.21496334671974182, + "learning_rate": 0.0001777637877745441, + "loss": 0.0567, + "step": 9240 + }, + { + "epoch": 0.21662359277296248, + "grad_norm": 0.8540540337562561, + "learning_rate": 0.00017775915636851687, + "loss": 0.5188, + "step": 9241 + }, + { + "epoch": 0.21664703434776747, + "grad_norm": 0.8917461633682251, + "learning_rate": 0.00017775452454056625, + "loss": 0.2064, + "step": 9242 + }, + { + "epoch": 0.2166704759225725, + "grad_norm": 0.46920540928840637, + "learning_rate": 0.00017774989229071738, + "loss": 0.1235, + "step": 9243 + }, + { + "epoch": 0.2166939174973775, + "grad_norm": 0.41059231758117676, + "learning_rate": 0.00017774525961899545, + "loss": 0.4001, + "step": 9244 + }, + { + "epoch": 0.21671735907218248, + "grad_norm": 0.5432890057563782, + "learning_rate": 0.00017774062652542552, + "loss": 0.1281, + "step": 9245 + }, + { + "epoch": 0.21674080064698747, + "grad_norm": 0.22151587903499603, + "learning_rate": 0.00017773599301003275, + "loss": 0.0554, + "step": 9246 + }, + { + "epoch": 0.21676424222179247, + "grad_norm": 0.472746342420578, + "learning_rate": 0.00017773135907284235, + "loss": 0.1319, + "step": 9247 + }, + { + "epoch": 0.21678768379659746, + "grad_norm": 0.38564401865005493, + "learning_rate": 0.0001777267247138794, + "loss": 0.0578, + "step": 9248 + }, + { + "epoch": 0.21681112537140246, + "grad_norm": 0.8213809132575989, + "learning_rate": 0.00017772208993316904, + "loss": 0.258, + "step": 9249 + }, + { + "epoch": 0.21683456694620745, + "grad_norm": 0.5505796670913696, + "learning_rate": 0.00017771745473073644, + "loss": 0.1114, + "step": 9250 + }, + { + "epoch": 0.21685800852101245, + "grad_norm": 0.7314544320106506, + "learning_rate": 0.00017771281910660674, + "loss": 0.1745, + "step": 9251 + }, + { + "epoch": 0.21688145009581744, + "grad_norm": 0.6077522039413452, + "learning_rate": 0.00017770818306080507, + "loss": 0.1875, + "step": 9252 + }, + { + "epoch": 0.21690489167062244, + "grad_norm": 0.2184506356716156, + "learning_rate": 0.00017770354659335665, + "loss": 0.0457, + "step": 9253 + }, + { + "epoch": 0.21692833324542743, + "grad_norm": 0.8259330987930298, + "learning_rate": 0.0001776989097042866, + "loss": 0.1743, + "step": 9254 + }, + { + "epoch": 0.21695177482023242, + "grad_norm": 0.6130579113960266, + "learning_rate": 0.00017769427239362008, + "loss": 0.1739, + "step": 9255 + }, + { + "epoch": 0.21697521639503742, + "grad_norm": 0.24152608215808868, + "learning_rate": 0.00017768963466138223, + "loss": 0.0808, + "step": 9256 + }, + { + "epoch": 0.2169986579698424, + "grad_norm": 0.6470065116882324, + "learning_rate": 0.00017768499650759826, + "loss": 0.1254, + "step": 9257 + }, + { + "epoch": 0.2170220995446474, + "grad_norm": 0.48302745819091797, + "learning_rate": 0.00017768035793229334, + "loss": 0.154, + "step": 9258 + }, + { + "epoch": 0.2170455411194524, + "grad_norm": 0.6271643042564392, + "learning_rate": 0.00017767571893549254, + "loss": 0.1472, + "step": 9259 + }, + { + "epoch": 0.2170689826942574, + "grad_norm": 0.5298148989677429, + "learning_rate": 0.00017767107951722116, + "loss": 0.1521, + "step": 9260 + }, + { + "epoch": 0.2170924242690624, + "grad_norm": 0.10857109725475311, + "learning_rate": 0.00017766643967750433, + "loss": 0.0243, + "step": 9261 + }, + { + "epoch": 0.21711586584386738, + "grad_norm": 0.8559702038764954, + "learning_rate": 0.0001776617994163672, + "loss": 0.2057, + "step": 9262 + }, + { + "epoch": 0.21713930741867238, + "grad_norm": 0.40915775299072266, + "learning_rate": 0.00017765715873383495, + "loss": 0.0819, + "step": 9263 + }, + { + "epoch": 0.21716274899347737, + "grad_norm": 0.7077546715736389, + "learning_rate": 0.00017765251762993274, + "loss": 0.1611, + "step": 9264 + }, + { + "epoch": 0.21718619056828237, + "grad_norm": 0.239286869764328, + "learning_rate": 0.00017764787610468586, + "loss": 0.0564, + "step": 9265 + }, + { + "epoch": 0.21720963214308736, + "grad_norm": 0.9113888740539551, + "learning_rate": 0.0001776432341581194, + "loss": 0.2407, + "step": 9266 + }, + { + "epoch": 0.21723307371789236, + "grad_norm": 0.1348278820514679, + "learning_rate": 0.00017763859179025854, + "loss": 0.02, + "step": 9267 + }, + { + "epoch": 0.21725651529269735, + "grad_norm": 0.45727118849754333, + "learning_rate": 0.0001776339490011285, + "loss": 0.1061, + "step": 9268 + }, + { + "epoch": 0.21727995686750237, + "grad_norm": 0.6140025854110718, + "learning_rate": 0.00017762930579075448, + "loss": 0.1659, + "step": 9269 + }, + { + "epoch": 0.21730339844230737, + "grad_norm": 0.30279111862182617, + "learning_rate": 0.00017762466215916168, + "loss": 0.0479, + "step": 9270 + }, + { + "epoch": 0.21732684001711236, + "grad_norm": 0.769779622554779, + "learning_rate": 0.00017762001810637528, + "loss": 0.2244, + "step": 9271 + }, + { + "epoch": 0.21735028159191735, + "grad_norm": 0.5659551620483398, + "learning_rate": 0.00017761537363242047, + "loss": 0.1225, + "step": 9272 + }, + { + "epoch": 0.21737372316672235, + "grad_norm": 0.40873467922210693, + "learning_rate": 0.00017761072873732244, + "loss": 0.092, + "step": 9273 + }, + { + "epoch": 0.21739716474152734, + "grad_norm": 0.45317912101745605, + "learning_rate": 0.00017760608342110644, + "loss": 0.0581, + "step": 9274 + }, + { + "epoch": 0.21742060631633234, + "grad_norm": 0.3955463767051697, + "learning_rate": 0.00017760143768379762, + "loss": 0.0556, + "step": 9275 + }, + { + "epoch": 0.21744404789113733, + "grad_norm": 0.5811648964881897, + "learning_rate": 0.00017759679152542127, + "loss": 0.0565, + "step": 9276 + }, + { + "epoch": 0.21746748946594233, + "grad_norm": 0.18300476670265198, + "learning_rate": 0.00017759214494600248, + "loss": 0.0433, + "step": 9277 + }, + { + "epoch": 0.21749093104074732, + "grad_norm": 0.5249030590057373, + "learning_rate": 0.00017758749794556658, + "loss": 0.0639, + "step": 9278 + }, + { + "epoch": 0.21751437261555231, + "grad_norm": 0.48977214097976685, + "learning_rate": 0.00017758285052413872, + "loss": 0.1336, + "step": 9279 + }, + { + "epoch": 0.2175378141903573, + "grad_norm": 0.4171541631221771, + "learning_rate": 0.00017757820268174412, + "loss": 0.5375, + "step": 9280 + }, + { + "epoch": 0.2175612557651623, + "grad_norm": 0.28784406185150146, + "learning_rate": 0.000177573554418408, + "loss": 0.0759, + "step": 9281 + }, + { + "epoch": 0.2175846973399673, + "grad_norm": 0.6088464856147766, + "learning_rate": 0.00017756890573415564, + "loss": 0.565, + "step": 9282 + }, + { + "epoch": 0.2176081389147723, + "grad_norm": 0.5960171818733215, + "learning_rate": 0.00017756425662901219, + "loss": 0.0786, + "step": 9283 + }, + { + "epoch": 0.2176315804895773, + "grad_norm": 0.21600256860256195, + "learning_rate": 0.0001775596071030029, + "loss": 0.0443, + "step": 9284 + }, + { + "epoch": 0.21765502206438228, + "grad_norm": 0.8887595534324646, + "learning_rate": 0.000177554957156153, + "loss": 0.1505, + "step": 9285 + }, + { + "epoch": 0.21767846363918728, + "grad_norm": 0.24637767672538757, + "learning_rate": 0.0001775503067884877, + "loss": 0.0449, + "step": 9286 + }, + { + "epoch": 0.21770190521399227, + "grad_norm": 0.631088376045227, + "learning_rate": 0.0001775456560000323, + "loss": 0.1224, + "step": 9287 + }, + { + "epoch": 0.21772534678879726, + "grad_norm": 0.19075095653533936, + "learning_rate": 0.00017754100479081198, + "loss": 0.0331, + "step": 9288 + }, + { + "epoch": 0.21774878836360226, + "grad_norm": 0.44926467537879944, + "learning_rate": 0.00017753635316085196, + "loss": 0.0987, + "step": 9289 + }, + { + "epoch": 0.21777222993840725, + "grad_norm": 0.4650879502296448, + "learning_rate": 0.00017753170111017754, + "loss": 0.1391, + "step": 9290 + }, + { + "epoch": 0.21779567151321225, + "grad_norm": 0.3669434189796448, + "learning_rate": 0.00017752704863881392, + "loss": 0.0668, + "step": 9291 + }, + { + "epoch": 0.21781911308801724, + "grad_norm": 0.4603593349456787, + "learning_rate": 0.00017752239574678633, + "loss": 0.0699, + "step": 9292 + }, + { + "epoch": 0.21784255466282224, + "grad_norm": 0.9117306470870972, + "learning_rate": 0.00017751774243412007, + "loss": 0.2573, + "step": 9293 + }, + { + "epoch": 0.21786599623762726, + "grad_norm": 0.5865415334701538, + "learning_rate": 0.00017751308870084037, + "loss": 0.1228, + "step": 9294 + }, + { + "epoch": 0.21788943781243225, + "grad_norm": 0.6596537232398987, + "learning_rate": 0.00017750843454697243, + "loss": 0.7205, + "step": 9295 + }, + { + "epoch": 0.21791287938723725, + "grad_norm": 0.5053126811981201, + "learning_rate": 0.00017750377997254155, + "loss": 0.1322, + "step": 9296 + }, + { + "epoch": 0.21793632096204224, + "grad_norm": 0.4113384485244751, + "learning_rate": 0.00017749912497757298, + "loss": 0.0689, + "step": 9297 + }, + { + "epoch": 0.21795976253684723, + "grad_norm": 0.7100510597229004, + "learning_rate": 0.00017749446956209203, + "loss": 0.1491, + "step": 9298 + }, + { + "epoch": 0.21798320411165223, + "grad_norm": 0.30498865246772766, + "learning_rate": 0.00017748981372612388, + "loss": 0.088, + "step": 9299 + }, + { + "epoch": 0.21800664568645722, + "grad_norm": 0.9515901207923889, + "learning_rate": 0.00017748515746969382, + "loss": 0.2119, + "step": 9300 + }, + { + "epoch": 0.21803008726126222, + "grad_norm": 0.5217675566673279, + "learning_rate": 0.00017748050079282712, + "loss": 0.5198, + "step": 9301 + }, + { + "epoch": 0.2180535288360672, + "grad_norm": 0.4102204740047455, + "learning_rate": 0.00017747584369554902, + "loss": 0.4735, + "step": 9302 + }, + { + "epoch": 0.2180769704108722, + "grad_norm": 0.2734188139438629, + "learning_rate": 0.00017747118617788483, + "loss": 0.0745, + "step": 9303 + }, + { + "epoch": 0.2181004119856772, + "grad_norm": 0.7810125350952148, + "learning_rate": 0.0001774665282398598, + "loss": 0.6931, + "step": 9304 + }, + { + "epoch": 0.2181238535604822, + "grad_norm": 0.5799183249473572, + "learning_rate": 0.00017746186988149923, + "loss": 0.6149, + "step": 9305 + }, + { + "epoch": 0.2181472951352872, + "grad_norm": 0.27607887983322144, + "learning_rate": 0.00017745721110282837, + "loss": 0.0675, + "step": 9306 + }, + { + "epoch": 0.21817073671009218, + "grad_norm": 0.12406173348426819, + "learning_rate": 0.0001774525519038725, + "loss": 0.0235, + "step": 9307 + }, + { + "epoch": 0.21819417828489718, + "grad_norm": 0.4518776834011078, + "learning_rate": 0.00017744789228465687, + "loss": 0.1216, + "step": 9308 + }, + { + "epoch": 0.21821761985970217, + "grad_norm": 0.6214583516120911, + "learning_rate": 0.00017744323224520687, + "loss": 0.1211, + "step": 9309 + }, + { + "epoch": 0.21824106143450717, + "grad_norm": 0.7888907790184021, + "learning_rate": 0.00017743857178554765, + "loss": 0.1675, + "step": 9310 + }, + { + "epoch": 0.21826450300931216, + "grad_norm": 0.20234090089797974, + "learning_rate": 0.0001774339109057046, + "loss": 0.0363, + "step": 9311 + }, + { + "epoch": 0.21828794458411716, + "grad_norm": 0.46773412823677063, + "learning_rate": 0.00017742924960570297, + "loss": 0.1294, + "step": 9312 + }, + { + "epoch": 0.21831138615892215, + "grad_norm": 0.44595518708229065, + "learning_rate": 0.00017742458788556803, + "loss": 0.0828, + "step": 9313 + }, + { + "epoch": 0.21833482773372714, + "grad_norm": 0.4179916977882385, + "learning_rate": 0.0001774199257453251, + "loss": 0.102, + "step": 9314 + }, + { + "epoch": 0.21835826930853214, + "grad_norm": 0.36803242564201355, + "learning_rate": 0.0001774152631849995, + "loss": 0.0742, + "step": 9315 + }, + { + "epoch": 0.21838171088333713, + "grad_norm": 0.6742575764656067, + "learning_rate": 0.00017741060020461647, + "loss": 0.1592, + "step": 9316 + }, + { + "epoch": 0.21840515245814213, + "grad_norm": 0.4228438138961792, + "learning_rate": 0.0001774059368042014, + "loss": 0.0712, + "step": 9317 + }, + { + "epoch": 0.21842859403294712, + "grad_norm": 0.6697778105735779, + "learning_rate": 0.0001774012729837795, + "loss": 0.1616, + "step": 9318 + }, + { + "epoch": 0.21845203560775212, + "grad_norm": 0.1423385739326477, + "learning_rate": 0.00017739660874337615, + "loss": 0.037, + "step": 9319 + }, + { + "epoch": 0.21847547718255714, + "grad_norm": 0.3448365330696106, + "learning_rate": 0.0001773919440830166, + "loss": 0.0654, + "step": 9320 + }, + { + "epoch": 0.21849891875736213, + "grad_norm": 0.577443540096283, + "learning_rate": 0.00017738727900272618, + "loss": 0.1464, + "step": 9321 + }, + { + "epoch": 0.21852236033216713, + "grad_norm": 0.09141240268945694, + "learning_rate": 0.00017738261350253023, + "loss": 0.0241, + "step": 9322 + }, + { + "epoch": 0.21854580190697212, + "grad_norm": 0.5767099261283875, + "learning_rate": 0.000177377947582454, + "loss": 0.1091, + "step": 9323 + }, + { + "epoch": 0.21856924348177711, + "grad_norm": 0.20817995071411133, + "learning_rate": 0.0001773732812425229, + "loss": 0.0296, + "step": 9324 + }, + { + "epoch": 0.2185926850565821, + "grad_norm": 0.26841866970062256, + "learning_rate": 0.00017736861448276219, + "loss": 0.0561, + "step": 9325 + }, + { + "epoch": 0.2186161266313871, + "grad_norm": 0.4509830176830292, + "learning_rate": 0.0001773639473031972, + "loss": 0.0844, + "step": 9326 + }, + { + "epoch": 0.2186395682061921, + "grad_norm": 0.3924781382083893, + "learning_rate": 0.00017735927970385324, + "loss": 0.0637, + "step": 9327 + }, + { + "epoch": 0.2186630097809971, + "grad_norm": 0.7132023572921753, + "learning_rate": 0.00017735461168475567, + "loss": 0.1592, + "step": 9328 + }, + { + "epoch": 0.2186864513558021, + "grad_norm": 0.44327545166015625, + "learning_rate": 0.0001773499432459298, + "loss": 0.559, + "step": 9329 + }, + { + "epoch": 0.21870989293060708, + "grad_norm": 0.3280065953731537, + "learning_rate": 0.00017734527438740093, + "loss": 0.0507, + "step": 9330 + }, + { + "epoch": 0.21873333450541207, + "grad_norm": 0.3080361783504486, + "learning_rate": 0.00017734060510919445, + "loss": 0.0937, + "step": 9331 + }, + { + "epoch": 0.21875677608021707, + "grad_norm": 0.4043236970901489, + "learning_rate": 0.00017733593541133572, + "loss": 0.0645, + "step": 9332 + }, + { + "epoch": 0.21878021765502206, + "grad_norm": 0.963289737701416, + "learning_rate": 0.00017733126529384996, + "loss": 0.2301, + "step": 9333 + }, + { + "epoch": 0.21880365922982706, + "grad_norm": 0.4800208806991577, + "learning_rate": 0.00017732659475676261, + "loss": 0.1437, + "step": 9334 + }, + { + "epoch": 0.21882710080463205, + "grad_norm": 0.4980350434780121, + "learning_rate": 0.00017732192380009898, + "loss": 0.0978, + "step": 9335 + }, + { + "epoch": 0.21885054237943705, + "grad_norm": 1.3474876880645752, + "learning_rate": 0.00017731725242388442, + "loss": 0.2086, + "step": 9336 + }, + { + "epoch": 0.21887398395424204, + "grad_norm": 0.555706799030304, + "learning_rate": 0.00017731258062814428, + "loss": 0.1249, + "step": 9337 + }, + { + "epoch": 0.21889742552904703, + "grad_norm": 0.6589659452438354, + "learning_rate": 0.0001773079084129039, + "loss": 0.1601, + "step": 9338 + }, + { + "epoch": 0.21892086710385203, + "grad_norm": 0.49732792377471924, + "learning_rate": 0.0001773032357781886, + "loss": 0.0886, + "step": 9339 + }, + { + "epoch": 0.21894430867865702, + "grad_norm": 0.40424844622612, + "learning_rate": 0.00017729856272402384, + "loss": 0.4489, + "step": 9340 + }, + { + "epoch": 0.21896775025346202, + "grad_norm": 0.8248707056045532, + "learning_rate": 0.00017729388925043487, + "loss": 0.1164, + "step": 9341 + }, + { + "epoch": 0.218991191828267, + "grad_norm": 0.44915875792503357, + "learning_rate": 0.00017728921535744706, + "loss": 0.0657, + "step": 9342 + }, + { + "epoch": 0.219014633403072, + "grad_norm": 0.6252944469451904, + "learning_rate": 0.00017728454104508581, + "loss": 0.1559, + "step": 9343 + }, + { + "epoch": 0.219038074977877, + "grad_norm": 0.6840646266937256, + "learning_rate": 0.00017727986631337647, + "loss": 0.1899, + "step": 9344 + }, + { + "epoch": 0.21906151655268202, + "grad_norm": 0.5474720001220703, + "learning_rate": 0.00017727519116234438, + "loss": 0.1363, + "step": 9345 + }, + { + "epoch": 0.21908495812748702, + "grad_norm": 0.13154935836791992, + "learning_rate": 0.00017727051559201498, + "loss": 0.0154, + "step": 9346 + }, + { + "epoch": 0.219108399702292, + "grad_norm": 0.44995859265327454, + "learning_rate": 0.00017726583960241357, + "loss": 0.074, + "step": 9347 + }, + { + "epoch": 0.219131841277097, + "grad_norm": 0.65727299451828, + "learning_rate": 0.00017726116319356552, + "loss": 0.189, + "step": 9348 + }, + { + "epoch": 0.219155282851902, + "grad_norm": 0.500123918056488, + "learning_rate": 0.00017725648636549623, + "loss": 0.0968, + "step": 9349 + }, + { + "epoch": 0.219178724426707, + "grad_norm": 0.301680326461792, + "learning_rate": 0.0001772518091182311, + "loss": 0.0505, + "step": 9350 + }, + { + "epoch": 0.219202166001512, + "grad_norm": 0.36933594942092896, + "learning_rate": 0.00017724713145179545, + "loss": 0.0744, + "step": 9351 + }, + { + "epoch": 0.21922560757631698, + "grad_norm": 0.3642148971557617, + "learning_rate": 0.0001772424533662147, + "loss": 0.0591, + "step": 9352 + }, + { + "epoch": 0.21924904915112198, + "grad_norm": 0.6354667544364929, + "learning_rate": 0.00017723777486151424, + "loss": 0.0992, + "step": 9353 + }, + { + "epoch": 0.21927249072592697, + "grad_norm": 0.6233209371566772, + "learning_rate": 0.00017723309593771944, + "loss": 0.1418, + "step": 9354 + }, + { + "epoch": 0.21929593230073197, + "grad_norm": 0.28639551997184753, + "learning_rate": 0.00017722841659485568, + "loss": 0.0695, + "step": 9355 + }, + { + "epoch": 0.21931937387553696, + "grad_norm": 0.13196174800395966, + "learning_rate": 0.00017722373683294837, + "loss": 0.0349, + "step": 9356 + }, + { + "epoch": 0.21934281545034195, + "grad_norm": 0.1992545872926712, + "learning_rate": 0.0001772190566520229, + "loss": 0.0342, + "step": 9357 + }, + { + "epoch": 0.21936625702514695, + "grad_norm": 0.46618548035621643, + "learning_rate": 0.00017721437605210466, + "loss": 0.1076, + "step": 9358 + }, + { + "epoch": 0.21938969859995194, + "grad_norm": 0.43494459986686707, + "learning_rate": 0.000177209695033219, + "loss": 0.1504, + "step": 9359 + }, + { + "epoch": 0.21941314017475694, + "grad_norm": 0.22909314930438995, + "learning_rate": 0.0001772050135953914, + "loss": 0.0527, + "step": 9360 + }, + { + "epoch": 0.21943658174956193, + "grad_norm": 0.42824044823646545, + "learning_rate": 0.0001772003317386472, + "loss": 0.0893, + "step": 9361 + }, + { + "epoch": 0.21946002332436693, + "grad_norm": 0.45455577969551086, + "learning_rate": 0.00017719564946301185, + "loss": 0.0877, + "step": 9362 + }, + { + "epoch": 0.21948346489917192, + "grad_norm": 0.32091712951660156, + "learning_rate": 0.00017719096676851076, + "loss": 0.093, + "step": 9363 + }, + { + "epoch": 0.21950690647397691, + "grad_norm": 0.664953351020813, + "learning_rate": 0.00017718628365516928, + "loss": 0.185, + "step": 9364 + }, + { + "epoch": 0.2195303480487819, + "grad_norm": 0.10503192245960236, + "learning_rate": 0.00017718160012301286, + "loss": 0.0236, + "step": 9365 + }, + { + "epoch": 0.2195537896235869, + "grad_norm": 0.5509812235832214, + "learning_rate": 0.00017717691617206692, + "loss": 0.0575, + "step": 9366 + }, + { + "epoch": 0.2195772311983919, + "grad_norm": 0.14677050709724426, + "learning_rate": 0.00017717223180235684, + "loss": 0.057, + "step": 9367 + }, + { + "epoch": 0.2196006727731969, + "grad_norm": 0.5354702472686768, + "learning_rate": 0.00017716754701390807, + "loss": 0.104, + "step": 9368 + }, + { + "epoch": 0.2196241143480019, + "grad_norm": 0.6109145879745483, + "learning_rate": 0.00017716286180674602, + "loss": 0.1147, + "step": 9369 + }, + { + "epoch": 0.21964755592280688, + "grad_norm": 1.1027675867080688, + "learning_rate": 0.0001771581761808961, + "loss": 0.3706, + "step": 9370 + }, + { + "epoch": 0.2196709974976119, + "grad_norm": 0.6397554278373718, + "learning_rate": 0.00017715349013638378, + "loss": 0.7098, + "step": 9371 + }, + { + "epoch": 0.2196944390724169, + "grad_norm": 0.5532655715942383, + "learning_rate": 0.00017714880367323445, + "loss": 0.0708, + "step": 9372 + }, + { + "epoch": 0.2197178806472219, + "grad_norm": 0.6766468286514282, + "learning_rate": 0.0001771441167914735, + "loss": 0.2324, + "step": 9373 + }, + { + "epoch": 0.21974132222202689, + "grad_norm": 0.4473623037338257, + "learning_rate": 0.00017713942949112646, + "loss": 0.1028, + "step": 9374 + }, + { + "epoch": 0.21976476379683188, + "grad_norm": 0.4434335231781006, + "learning_rate": 0.00017713474177221867, + "loss": 0.0525, + "step": 9375 + }, + { + "epoch": 0.21978820537163687, + "grad_norm": 0.29984065890312195, + "learning_rate": 0.0001771300536347756, + "loss": 0.1128, + "step": 9376 + }, + { + "epoch": 0.21981164694644187, + "grad_norm": 0.5505114197731018, + "learning_rate": 0.0001771253650788227, + "loss": 0.1561, + "step": 9377 + }, + { + "epoch": 0.21983508852124686, + "grad_norm": 0.22036167979240417, + "learning_rate": 0.0001771206761043854, + "loss": 0.039, + "step": 9378 + }, + { + "epoch": 0.21985853009605186, + "grad_norm": 0.2049759030342102, + "learning_rate": 0.00017711598671148914, + "loss": 0.0403, + "step": 9379 + }, + { + "epoch": 0.21988197167085685, + "grad_norm": 0.39640533924102783, + "learning_rate": 0.00017711129690015938, + "loss": 0.1013, + "step": 9380 + }, + { + "epoch": 0.21990541324566185, + "grad_norm": 0.8468468189239502, + "learning_rate": 0.00017710660667042154, + "loss": 0.1199, + "step": 9381 + }, + { + "epoch": 0.21992885482046684, + "grad_norm": 1.1626948118209839, + "learning_rate": 0.00017710191602230107, + "loss": 0.2021, + "step": 9382 + }, + { + "epoch": 0.21995229639527183, + "grad_norm": 0.3951576352119446, + "learning_rate": 0.00017709722495582345, + "loss": 0.1123, + "step": 9383 + }, + { + "epoch": 0.21997573797007683, + "grad_norm": 0.28967389464378357, + "learning_rate": 0.00017709253347101417, + "loss": 0.0827, + "step": 9384 + }, + { + "epoch": 0.21999917954488182, + "grad_norm": 0.23540878295898438, + "learning_rate": 0.00017708784156789857, + "loss": 0.0459, + "step": 9385 + }, + { + "epoch": 0.22002262111968682, + "grad_norm": 0.789093554019928, + "learning_rate": 0.0001770831492465022, + "loss": 0.122, + "step": 9386 + }, + { + "epoch": 0.2200460626944918, + "grad_norm": 0.1820783019065857, + "learning_rate": 0.0001770784565068505, + "loss": 0.0545, + "step": 9387 + }, + { + "epoch": 0.2200695042692968, + "grad_norm": 0.5012380480766296, + "learning_rate": 0.00017707376334896892, + "loss": 0.1439, + "step": 9388 + }, + { + "epoch": 0.2200929458441018, + "grad_norm": 0.5011993646621704, + "learning_rate": 0.00017706906977288291, + "loss": 0.1195, + "step": 9389 + }, + { + "epoch": 0.2201163874189068, + "grad_norm": 0.7236985564231873, + "learning_rate": 0.000177064375778618, + "loss": 0.1704, + "step": 9390 + }, + { + "epoch": 0.2201398289937118, + "grad_norm": 0.8162919878959656, + "learning_rate": 0.0001770596813661996, + "loss": 0.1507, + "step": 9391 + }, + { + "epoch": 0.22016327056851678, + "grad_norm": 0.3989415466785431, + "learning_rate": 0.00017705498653565317, + "loss": 0.1355, + "step": 9392 + }, + { + "epoch": 0.22018671214332178, + "grad_norm": 0.4011996388435364, + "learning_rate": 0.00017705029128700428, + "loss": 0.1, + "step": 9393 + }, + { + "epoch": 0.22021015371812677, + "grad_norm": 0.45488470792770386, + "learning_rate": 0.0001770455956202783, + "loss": 0.1327, + "step": 9394 + }, + { + "epoch": 0.22023359529293177, + "grad_norm": 0.5297275185585022, + "learning_rate": 0.00017704089953550075, + "loss": 0.6072, + "step": 9395 + }, + { + "epoch": 0.22025703686773676, + "grad_norm": 0.5271209478378296, + "learning_rate": 0.00017703620303269714, + "loss": 0.1382, + "step": 9396 + }, + { + "epoch": 0.22028047844254178, + "grad_norm": 0.24925078451633453, + "learning_rate": 0.00017703150611189292, + "loss": 0.053, + "step": 9397 + }, + { + "epoch": 0.22030392001734678, + "grad_norm": 0.5217573642730713, + "learning_rate": 0.00017702680877311358, + "loss": 0.1227, + "step": 9398 + }, + { + "epoch": 0.22032736159215177, + "grad_norm": 0.5022732019424438, + "learning_rate": 0.0001770221110163846, + "loss": 0.1442, + "step": 9399 + }, + { + "epoch": 0.22035080316695677, + "grad_norm": 0.3136340081691742, + "learning_rate": 0.0001770174128417315, + "loss": 0.0474, + "step": 9400 + }, + { + "epoch": 0.22037424474176176, + "grad_norm": 0.21808871626853943, + "learning_rate": 0.0001770127142491797, + "loss": 0.0165, + "step": 9401 + }, + { + "epoch": 0.22039768631656675, + "grad_norm": 0.4473646283149719, + "learning_rate": 0.00017700801523875482, + "loss": 0.0976, + "step": 9402 + }, + { + "epoch": 0.22042112789137175, + "grad_norm": 0.6239801049232483, + "learning_rate": 0.00017700331581048224, + "loss": 0.1479, + "step": 9403 + }, + { + "epoch": 0.22044456946617674, + "grad_norm": 0.46982160210609436, + "learning_rate": 0.0001769986159643875, + "loss": 0.1288, + "step": 9404 + }, + { + "epoch": 0.22046801104098174, + "grad_norm": 0.7721350789070129, + "learning_rate": 0.00017699391570049613, + "loss": 0.2145, + "step": 9405 + }, + { + "epoch": 0.22049145261578673, + "grad_norm": 0.2517414689064026, + "learning_rate": 0.0001769892150188336, + "loss": 0.0423, + "step": 9406 + }, + { + "epoch": 0.22051489419059173, + "grad_norm": 0.5697616934776306, + "learning_rate": 0.00017698451391942542, + "loss": 0.1144, + "step": 9407 + }, + { + "epoch": 0.22053833576539672, + "grad_norm": 0.5219829678535461, + "learning_rate": 0.0001769798124022971, + "loss": 0.1245, + "step": 9408 + }, + { + "epoch": 0.22056177734020171, + "grad_norm": 1.1754258871078491, + "learning_rate": 0.00017697511046747418, + "loss": 0.3162, + "step": 9409 + }, + { + "epoch": 0.2205852189150067, + "grad_norm": 0.7483277916908264, + "learning_rate": 0.00017697040811498208, + "loss": 0.6069, + "step": 9410 + }, + { + "epoch": 0.2206086604898117, + "grad_norm": 0.8081483244895935, + "learning_rate": 0.00017696570534484645, + "loss": 0.2156, + "step": 9411 + }, + { + "epoch": 0.2206321020646167, + "grad_norm": 1.035145878791809, + "learning_rate": 0.0001769610021570927, + "loss": 0.1792, + "step": 9412 + }, + { + "epoch": 0.2206555436394217, + "grad_norm": 0.3407277464866638, + "learning_rate": 0.0001769562985517464, + "loss": 0.0648, + "step": 9413 + }, + { + "epoch": 0.22067898521422669, + "grad_norm": 0.7417298555374146, + "learning_rate": 0.00017695159452883305, + "loss": 0.7069, + "step": 9414 + }, + { + "epoch": 0.22070242678903168, + "grad_norm": 0.5152727365493774, + "learning_rate": 0.0001769468900883782, + "loss": 0.1054, + "step": 9415 + }, + { + "epoch": 0.22072586836383667, + "grad_norm": 0.5846399664878845, + "learning_rate": 0.00017694218523040735, + "loss": 0.1296, + "step": 9416 + }, + { + "epoch": 0.22074930993864167, + "grad_norm": 0.5311324596405029, + "learning_rate": 0.00017693747995494604, + "loss": 0.1506, + "step": 9417 + }, + { + "epoch": 0.22077275151344666, + "grad_norm": 0.23249903321266174, + "learning_rate": 0.00017693277426201985, + "loss": 0.0678, + "step": 9418 + }, + { + "epoch": 0.22079619308825166, + "grad_norm": 0.40767696499824524, + "learning_rate": 0.0001769280681516542, + "loss": 0.0662, + "step": 9419 + }, + { + "epoch": 0.22081963466305665, + "grad_norm": 0.5002607703208923, + "learning_rate": 0.0001769233616238747, + "loss": 0.152, + "step": 9420 + }, + { + "epoch": 0.22084307623786165, + "grad_norm": 0.4888341724872589, + "learning_rate": 0.0001769186546787069, + "loss": 0.139, + "step": 9421 + }, + { + "epoch": 0.22086651781266667, + "grad_norm": 0.8250799775123596, + "learning_rate": 0.0001769139473161763, + "loss": 0.2607, + "step": 9422 + }, + { + "epoch": 0.22088995938747166, + "grad_norm": 0.5216615200042725, + "learning_rate": 0.00017690923953630846, + "loss": 0.8052, + "step": 9423 + }, + { + "epoch": 0.22091340096227666, + "grad_norm": 0.1523924022912979, + "learning_rate": 0.00017690453133912892, + "loss": 0.0262, + "step": 9424 + }, + { + "epoch": 0.22093684253708165, + "grad_norm": 0.5326470732688904, + "learning_rate": 0.00017689982272466325, + "loss": 0.1435, + "step": 9425 + }, + { + "epoch": 0.22096028411188665, + "grad_norm": 0.5742672681808472, + "learning_rate": 0.00017689511369293698, + "loss": 0.1747, + "step": 9426 + }, + { + "epoch": 0.22098372568669164, + "grad_norm": 0.2956394851207733, + "learning_rate": 0.00017689040424397564, + "loss": 0.057, + "step": 9427 + }, + { + "epoch": 0.22100716726149663, + "grad_norm": 0.7995259761810303, + "learning_rate": 0.0001768856943778048, + "loss": 0.2273, + "step": 9428 + }, + { + "epoch": 0.22103060883630163, + "grad_norm": 0.3758296072483063, + "learning_rate": 0.00017688098409445005, + "loss": 0.0809, + "step": 9429 + }, + { + "epoch": 0.22105405041110662, + "grad_norm": 0.2916949689388275, + "learning_rate": 0.00017687627339393688, + "loss": 0.0662, + "step": 9430 + }, + { + "epoch": 0.22107749198591162, + "grad_norm": 0.42913416028022766, + "learning_rate": 0.00017687156227629092, + "loss": 0.0909, + "step": 9431 + }, + { + "epoch": 0.2211009335607166, + "grad_norm": 0.2211947739124298, + "learning_rate": 0.00017686685074153768, + "loss": 0.0582, + "step": 9432 + }, + { + "epoch": 0.2211243751355216, + "grad_norm": 0.7545976042747498, + "learning_rate": 0.00017686213878970278, + "loss": 0.3201, + "step": 9433 + }, + { + "epoch": 0.2211478167103266, + "grad_norm": 0.5688100457191467, + "learning_rate": 0.0001768574264208117, + "loss": 0.1552, + "step": 9434 + }, + { + "epoch": 0.2211712582851316, + "grad_norm": 0.5394184589385986, + "learning_rate": 0.0001768527136348901, + "loss": 0.1317, + "step": 9435 + }, + { + "epoch": 0.2211946998599366, + "grad_norm": 0.6893871426582336, + "learning_rate": 0.00017684800043196353, + "loss": 0.7475, + "step": 9436 + }, + { + "epoch": 0.22121814143474158, + "grad_norm": 0.8407506942749023, + "learning_rate": 0.00017684328681205752, + "loss": 0.1826, + "step": 9437 + }, + { + "epoch": 0.22124158300954658, + "grad_norm": 0.774921715259552, + "learning_rate": 0.00017683857277519768, + "loss": 0.8325, + "step": 9438 + }, + { + "epoch": 0.22126502458435157, + "grad_norm": 0.3191255033016205, + "learning_rate": 0.0001768338583214096, + "loss": 0.0776, + "step": 9439 + }, + { + "epoch": 0.22128846615915657, + "grad_norm": 0.3897872269153595, + "learning_rate": 0.00017682914345071882, + "loss": 0.0609, + "step": 9440 + }, + { + "epoch": 0.22131190773396156, + "grad_norm": 0.41318264603614807, + "learning_rate": 0.00017682442816315097, + "loss": 0.0777, + "step": 9441 + }, + { + "epoch": 0.22133534930876655, + "grad_norm": 0.36317381262779236, + "learning_rate": 0.0001768197124587316, + "loss": 0.0734, + "step": 9442 + }, + { + "epoch": 0.22135879088357155, + "grad_norm": 0.38515451550483704, + "learning_rate": 0.0001768149963374863, + "loss": 0.1248, + "step": 9443 + }, + { + "epoch": 0.22138223245837654, + "grad_norm": 0.3960838317871094, + "learning_rate": 0.00017681027979944068, + "loss": 0.0858, + "step": 9444 + }, + { + "epoch": 0.22140567403318154, + "grad_norm": 0.24181433022022247, + "learning_rate": 0.00017680556284462034, + "loss": 0.0563, + "step": 9445 + }, + { + "epoch": 0.22142911560798653, + "grad_norm": 0.4459158480167389, + "learning_rate": 0.00017680084547305085, + "loss": 0.1004, + "step": 9446 + }, + { + "epoch": 0.22145255718279153, + "grad_norm": 0.5399752855300903, + "learning_rate": 0.0001767961276847578, + "loss": 0.133, + "step": 9447 + }, + { + "epoch": 0.22147599875759655, + "grad_norm": 0.5118715167045593, + "learning_rate": 0.00017679140947976678, + "loss": 0.1396, + "step": 9448 + }, + { + "epoch": 0.22149944033240154, + "grad_norm": 0.524748682975769, + "learning_rate": 0.00017678669085810346, + "loss": 0.1542, + "step": 9449 + }, + { + "epoch": 0.22152288190720654, + "grad_norm": 0.19310441613197327, + "learning_rate": 0.00017678197181979338, + "loss": 0.0333, + "step": 9450 + }, + { + "epoch": 0.22154632348201153, + "grad_norm": 0.47186478972435, + "learning_rate": 0.00017677725236486214, + "loss": 0.4462, + "step": 9451 + }, + { + "epoch": 0.22156976505681653, + "grad_norm": 0.48683544993400574, + "learning_rate": 0.00017677253249333538, + "loss": 0.1083, + "step": 9452 + }, + { + "epoch": 0.22159320663162152, + "grad_norm": 0.58660888671875, + "learning_rate": 0.0001767678122052387, + "loss": 0.1096, + "step": 9453 + }, + { + "epoch": 0.22161664820642651, + "grad_norm": 0.3456367254257202, + "learning_rate": 0.00017676309150059774, + "loss": 0.0968, + "step": 9454 + }, + { + "epoch": 0.2216400897812315, + "grad_norm": 0.349166601896286, + "learning_rate": 0.0001767583703794381, + "loss": 0.073, + "step": 9455 + }, + { + "epoch": 0.2216635313560365, + "grad_norm": 0.7682031989097595, + "learning_rate": 0.00017675364884178532, + "loss": 0.0703, + "step": 9456 + }, + { + "epoch": 0.2216869729308415, + "grad_norm": 0.4993293285369873, + "learning_rate": 0.0001767489268876651, + "loss": 0.0804, + "step": 9457 + }, + { + "epoch": 0.2217104145056465, + "grad_norm": 0.32245099544525146, + "learning_rate": 0.00017674420451710307, + "loss": 0.0553, + "step": 9458 + }, + { + "epoch": 0.22173385608045149, + "grad_norm": 0.1573197841644287, + "learning_rate": 0.0001767394817301248, + "loss": 0.0314, + "step": 9459 + }, + { + "epoch": 0.22175729765525648, + "grad_norm": 0.2524549067020416, + "learning_rate": 0.00017673475852675598, + "loss": 0.0707, + "step": 9460 + }, + { + "epoch": 0.22178073923006147, + "grad_norm": 0.44788676500320435, + "learning_rate": 0.00017673003490702218, + "loss": 0.1277, + "step": 9461 + }, + { + "epoch": 0.22180418080486647, + "grad_norm": 0.9346414804458618, + "learning_rate": 0.00017672531087094905, + "loss": 0.1714, + "step": 9462 + }, + { + "epoch": 0.22182762237967146, + "grad_norm": 0.6474072933197021, + "learning_rate": 0.00017672058641856223, + "loss": 0.1127, + "step": 9463 + }, + { + "epoch": 0.22185106395447646, + "grad_norm": 0.9745696783065796, + "learning_rate": 0.00017671586154988737, + "loss": 0.2074, + "step": 9464 + }, + { + "epoch": 0.22187450552928145, + "grad_norm": 0.43857038021087646, + "learning_rate": 0.00017671113626495006, + "loss": 0.0651, + "step": 9465 + }, + { + "epoch": 0.22189794710408645, + "grad_norm": 0.6436682343482971, + "learning_rate": 0.00017670641056377598, + "loss": 0.1098, + "step": 9466 + }, + { + "epoch": 0.22192138867889144, + "grad_norm": 0.29465076327323914, + "learning_rate": 0.00017670168444639076, + "loss": 0.0494, + "step": 9467 + }, + { + "epoch": 0.22194483025369643, + "grad_norm": 0.6199662685394287, + "learning_rate": 0.00017669695791282006, + "loss": 0.2103, + "step": 9468 + }, + { + "epoch": 0.22196827182850143, + "grad_norm": 0.6977688074111938, + "learning_rate": 0.0001766922309630895, + "loss": 0.102, + "step": 9469 + }, + { + "epoch": 0.22199171340330642, + "grad_norm": 0.4895387887954712, + "learning_rate": 0.00017668750359722474, + "loss": 0.071, + "step": 9470 + }, + { + "epoch": 0.22201515497811142, + "grad_norm": 0.8791958689689636, + "learning_rate": 0.0001766827758152514, + "loss": 0.0776, + "step": 9471 + }, + { + "epoch": 0.2220385965529164, + "grad_norm": 0.3550080358982086, + "learning_rate": 0.00017667804761719517, + "loss": 0.083, + "step": 9472 + }, + { + "epoch": 0.22206203812772143, + "grad_norm": 0.37750986218452454, + "learning_rate": 0.00017667331900308173, + "loss": 0.4443, + "step": 9473 + }, + { + "epoch": 0.22208547970252643, + "grad_norm": 0.5066747069358826, + "learning_rate": 0.00017666858997293668, + "loss": 0.1202, + "step": 9474 + }, + { + "epoch": 0.22210892127733142, + "grad_norm": 0.7157706022262573, + "learning_rate": 0.0001766638605267857, + "loss": 0.1472, + "step": 9475 + }, + { + "epoch": 0.22213236285213642, + "grad_norm": 0.1132802739739418, + "learning_rate": 0.00017665913066465446, + "loss": 0.0187, + "step": 9476 + }, + { + "epoch": 0.2221558044269414, + "grad_norm": 0.4143143892288208, + "learning_rate": 0.00017665440038656865, + "loss": 0.0752, + "step": 9477 + }, + { + "epoch": 0.2221792460017464, + "grad_norm": 0.695498526096344, + "learning_rate": 0.00017664966969255388, + "loss": 0.0732, + "step": 9478 + }, + { + "epoch": 0.2222026875765514, + "grad_norm": 0.6704642176628113, + "learning_rate": 0.00017664493858263587, + "loss": 0.2059, + "step": 9479 + }, + { + "epoch": 0.2222261291513564, + "grad_norm": 0.9855329394340515, + "learning_rate": 0.00017664020705684022, + "loss": 0.1502, + "step": 9480 + }, + { + "epoch": 0.2222495707261614, + "grad_norm": 0.31712606549263, + "learning_rate": 0.0001766354751151927, + "loss": 0.0519, + "step": 9481 + }, + { + "epoch": 0.22227301230096638, + "grad_norm": 0.5879018306732178, + "learning_rate": 0.0001766307427577189, + "loss": 0.1077, + "step": 9482 + }, + { + "epoch": 0.22229645387577138, + "grad_norm": 0.504437267780304, + "learning_rate": 0.00017662600998444457, + "loss": 0.1139, + "step": 9483 + }, + { + "epoch": 0.22231989545057637, + "grad_norm": 0.3724059760570526, + "learning_rate": 0.00017662127679539538, + "loss": 0.0729, + "step": 9484 + }, + { + "epoch": 0.22234333702538137, + "grad_norm": 0.2157432585954666, + "learning_rate": 0.00017661654319059693, + "loss": 0.0541, + "step": 9485 + }, + { + "epoch": 0.22236677860018636, + "grad_norm": 0.5667347311973572, + "learning_rate": 0.000176611809170075, + "loss": 0.1948, + "step": 9486 + }, + { + "epoch": 0.22239022017499135, + "grad_norm": 0.08473974466323853, + "learning_rate": 0.0001766070747338552, + "loss": 0.0174, + "step": 9487 + }, + { + "epoch": 0.22241366174979635, + "grad_norm": 0.28259244561195374, + "learning_rate": 0.00017660233988196327, + "loss": 0.0588, + "step": 9488 + }, + { + "epoch": 0.22243710332460134, + "grad_norm": 0.5138009786605835, + "learning_rate": 0.00017659760461442492, + "loss": 0.101, + "step": 9489 + }, + { + "epoch": 0.22246054489940634, + "grad_norm": 0.07973767817020416, + "learning_rate": 0.0001765928689312658, + "loss": 0.0126, + "step": 9490 + }, + { + "epoch": 0.22248398647421133, + "grad_norm": 0.2927822172641754, + "learning_rate": 0.00017658813283251158, + "loss": 0.0462, + "step": 9491 + }, + { + "epoch": 0.22250742804901633, + "grad_norm": 0.4530125558376312, + "learning_rate": 0.00017658339631818802, + "loss": 0.0882, + "step": 9492 + }, + { + "epoch": 0.22253086962382132, + "grad_norm": 0.5587961077690125, + "learning_rate": 0.00017657865938832084, + "loss": 0.0882, + "step": 9493 + }, + { + "epoch": 0.22255431119862631, + "grad_norm": 0.32872703671455383, + "learning_rate": 0.00017657392204293565, + "loss": 0.0699, + "step": 9494 + }, + { + "epoch": 0.2225777527734313, + "grad_norm": 0.47849440574645996, + "learning_rate": 0.00017656918428205824, + "loss": 0.1058, + "step": 9495 + }, + { + "epoch": 0.2226011943482363, + "grad_norm": 0.7489807605743408, + "learning_rate": 0.00017656444610571428, + "loss": 0.2108, + "step": 9496 + }, + { + "epoch": 0.2226246359230413, + "grad_norm": 0.5263688564300537, + "learning_rate": 0.00017655970751392946, + "loss": 0.1046, + "step": 9497 + }, + { + "epoch": 0.2226480774978463, + "grad_norm": 0.6529802680015564, + "learning_rate": 0.00017655496850672953, + "loss": 0.1617, + "step": 9498 + }, + { + "epoch": 0.2226715190726513, + "grad_norm": 0.21287420392036438, + "learning_rate": 0.00017655022908414018, + "loss": 0.0334, + "step": 9499 + }, + { + "epoch": 0.2226949606474563, + "grad_norm": 0.5784285664558411, + "learning_rate": 0.00017654548924618714, + "loss": 0.1798, + "step": 9500 + }, + { + "epoch": 0.2227184022222613, + "grad_norm": 1.0283507108688354, + "learning_rate": 0.00017654074899289612, + "loss": 0.1973, + "step": 9501 + }, + { + "epoch": 0.2227418437970663, + "grad_norm": 0.4510066509246826, + "learning_rate": 0.00017653600832429284, + "loss": 0.0789, + "step": 9502 + }, + { + "epoch": 0.2227652853718713, + "grad_norm": 0.6151838898658752, + "learning_rate": 0.00017653126724040305, + "loss": 0.2171, + "step": 9503 + }, + { + "epoch": 0.22278872694667629, + "grad_norm": 0.43425658345222473, + "learning_rate": 0.0001765265257412524, + "loss": 0.1106, + "step": 9504 + }, + { + "epoch": 0.22281216852148128, + "grad_norm": 0.223083034157753, + "learning_rate": 0.00017652178382686673, + "loss": 0.0364, + "step": 9505 + }, + { + "epoch": 0.22283561009628627, + "grad_norm": 0.6821253299713135, + "learning_rate": 0.00017651704149727171, + "loss": 0.1229, + "step": 9506 + }, + { + "epoch": 0.22285905167109127, + "grad_norm": 0.48645132780075073, + "learning_rate": 0.00017651229875249302, + "loss": 0.0905, + "step": 9507 + }, + { + "epoch": 0.22288249324589626, + "grad_norm": 0.18818219006061554, + "learning_rate": 0.00017650755559255648, + "loss": 0.0422, + "step": 9508 + }, + { + "epoch": 0.22290593482070126, + "grad_norm": 0.25407883524894714, + "learning_rate": 0.00017650281201748777, + "loss": 0.0537, + "step": 9509 + }, + { + "epoch": 0.22292937639550625, + "grad_norm": 0.6671158671379089, + "learning_rate": 0.00017649806802731268, + "loss": 0.6235, + "step": 9510 + }, + { + "epoch": 0.22295281797031125, + "grad_norm": 0.6563150882720947, + "learning_rate": 0.0001764933236220569, + "loss": 0.157, + "step": 9511 + }, + { + "epoch": 0.22297625954511624, + "grad_norm": 0.09821595251560211, + "learning_rate": 0.0001764885788017462, + "loss": 0.0272, + "step": 9512 + }, + { + "epoch": 0.22299970111992123, + "grad_norm": 0.3028154969215393, + "learning_rate": 0.0001764838335664063, + "loss": 0.0661, + "step": 9513 + }, + { + "epoch": 0.22302314269472623, + "grad_norm": 0.5915244221687317, + "learning_rate": 0.00017647908791606302, + "loss": 0.1647, + "step": 9514 + }, + { + "epoch": 0.22304658426953122, + "grad_norm": 0.3283836245536804, + "learning_rate": 0.000176474341850742, + "loss": 0.0483, + "step": 9515 + }, + { + "epoch": 0.22307002584433622, + "grad_norm": 0.6568018198013306, + "learning_rate": 0.0001764695953704691, + "loss": 0.1546, + "step": 9516 + }, + { + "epoch": 0.2230934674191412, + "grad_norm": 0.14550863206386566, + "learning_rate": 0.00017646484847527, + "loss": 0.0553, + "step": 9517 + }, + { + "epoch": 0.2231169089939462, + "grad_norm": 0.14202357828617096, + "learning_rate": 0.00017646010116517048, + "loss": 0.0245, + "step": 9518 + }, + { + "epoch": 0.2231403505687512, + "grad_norm": 0.4234565794467926, + "learning_rate": 0.0001764553534401963, + "loss": 0.0672, + "step": 9519 + }, + { + "epoch": 0.2231637921435562, + "grad_norm": 0.168277308344841, + "learning_rate": 0.00017645060530037325, + "loss": 0.0383, + "step": 9520 + }, + { + "epoch": 0.2231872337183612, + "grad_norm": 0.3834403455257416, + "learning_rate": 0.00017644585674572705, + "loss": 0.1145, + "step": 9521 + }, + { + "epoch": 0.22321067529316618, + "grad_norm": 0.41878917813301086, + "learning_rate": 0.00017644110777628348, + "loss": 0.0748, + "step": 9522 + }, + { + "epoch": 0.22323411686797118, + "grad_norm": 0.23097418248653412, + "learning_rate": 0.0001764363583920683, + "loss": 0.0383, + "step": 9523 + }, + { + "epoch": 0.2232575584427762, + "grad_norm": 0.647487223148346, + "learning_rate": 0.0001764316085931073, + "loss": 0.1419, + "step": 9524 + }, + { + "epoch": 0.2232810000175812, + "grad_norm": 0.5489042401313782, + "learning_rate": 0.00017642685837942622, + "loss": 0.0875, + "step": 9525 + }, + { + "epoch": 0.2233044415923862, + "grad_norm": 0.5389214158058167, + "learning_rate": 0.0001764221077510509, + "loss": 0.4652, + "step": 9526 + }, + { + "epoch": 0.22332788316719118, + "grad_norm": 0.7285433411598206, + "learning_rate": 0.00017641735670800705, + "loss": 0.2327, + "step": 9527 + }, + { + "epoch": 0.22335132474199618, + "grad_norm": 0.40583863854408264, + "learning_rate": 0.0001764126052503205, + "loss": 0.1162, + "step": 9528 + }, + { + "epoch": 0.22337476631680117, + "grad_norm": 0.5922127366065979, + "learning_rate": 0.00017640785337801694, + "loss": 0.1186, + "step": 9529 + }, + { + "epoch": 0.22339820789160617, + "grad_norm": 0.36060842871665955, + "learning_rate": 0.0001764031010911223, + "loss": 0.4774, + "step": 9530 + }, + { + "epoch": 0.22342164946641116, + "grad_norm": 0.6485096216201782, + "learning_rate": 0.00017639834838966225, + "loss": 0.1639, + "step": 9531 + }, + { + "epoch": 0.22344509104121615, + "grad_norm": 0.39616692066192627, + "learning_rate": 0.00017639359527366261, + "loss": 0.0602, + "step": 9532 + }, + { + "epoch": 0.22346853261602115, + "grad_norm": 0.6150946617126465, + "learning_rate": 0.00017638884174314917, + "loss": 0.6211, + "step": 9533 + }, + { + "epoch": 0.22349197419082614, + "grad_norm": 0.16097617149353027, + "learning_rate": 0.00017638408779814774, + "loss": 0.0287, + "step": 9534 + }, + { + "epoch": 0.22351541576563114, + "grad_norm": 0.37284842133522034, + "learning_rate": 0.00017637933343868408, + "loss": 0.1194, + "step": 9535 + }, + { + "epoch": 0.22353885734043613, + "grad_norm": 0.23791730403900146, + "learning_rate": 0.00017637457866478403, + "loss": 0.0652, + "step": 9536 + }, + { + "epoch": 0.22356229891524113, + "grad_norm": 0.7690894603729248, + "learning_rate": 0.00017636982347647338, + "loss": 0.1309, + "step": 9537 + }, + { + "epoch": 0.22358574049004612, + "grad_norm": 0.3267422914505005, + "learning_rate": 0.00017636506787377786, + "loss": 0.1058, + "step": 9538 + }, + { + "epoch": 0.22360918206485111, + "grad_norm": 0.3854861855506897, + "learning_rate": 0.0001763603118567234, + "loss": 0.1012, + "step": 9539 + }, + { + "epoch": 0.2236326236396561, + "grad_norm": 0.5835559964179993, + "learning_rate": 0.00017635555542533572, + "loss": 0.1715, + "step": 9540 + }, + { + "epoch": 0.2236560652144611, + "grad_norm": 0.4264010787010193, + "learning_rate": 0.00017635079857964065, + "loss": 0.0935, + "step": 9541 + }, + { + "epoch": 0.2236795067892661, + "grad_norm": 0.7111018300056458, + "learning_rate": 0.000176346041319664, + "loss": 0.1675, + "step": 9542 + }, + { + "epoch": 0.2237029483640711, + "grad_norm": 0.4438384175300598, + "learning_rate": 0.00017634128364543158, + "loss": 0.1052, + "step": 9543 + }, + { + "epoch": 0.22372638993887609, + "grad_norm": 0.6033090353012085, + "learning_rate": 0.0001763365255569692, + "loss": 0.1137, + "step": 9544 + }, + { + "epoch": 0.22374983151368108, + "grad_norm": 0.5339131951332092, + "learning_rate": 0.00017633176705430267, + "loss": 0.5926, + "step": 9545 + }, + { + "epoch": 0.22377327308848607, + "grad_norm": 0.5706179141998291, + "learning_rate": 0.00017632700813745785, + "loss": 0.1246, + "step": 9546 + }, + { + "epoch": 0.22379671466329107, + "grad_norm": 0.38176050782203674, + "learning_rate": 0.00017632224880646057, + "loss": 0.0797, + "step": 9547 + }, + { + "epoch": 0.22382015623809606, + "grad_norm": 0.30984634160995483, + "learning_rate": 0.00017631748906133657, + "loss": 0.0823, + "step": 9548 + }, + { + "epoch": 0.22384359781290106, + "grad_norm": 0.395106703042984, + "learning_rate": 0.00017631272890211177, + "loss": 0.0657, + "step": 9549 + }, + { + "epoch": 0.22386703938770608, + "grad_norm": 0.2895471155643463, + "learning_rate": 0.00017630796832881191, + "loss": 0.0906, + "step": 9550 + }, + { + "epoch": 0.22389048096251107, + "grad_norm": 0.8976348042488098, + "learning_rate": 0.0001763032073414629, + "loss": 0.2057, + "step": 9551 + }, + { + "epoch": 0.22391392253731607, + "grad_norm": 0.4037836194038391, + "learning_rate": 0.00017629844594009052, + "loss": 0.1133, + "step": 9552 + }, + { + "epoch": 0.22393736411212106, + "grad_norm": 0.40994730591773987, + "learning_rate": 0.00017629368412472066, + "loss": 0.0516, + "step": 9553 + }, + { + "epoch": 0.22396080568692606, + "grad_norm": 0.7935268878936768, + "learning_rate": 0.0001762889218953791, + "loss": 0.1579, + "step": 9554 + }, + { + "epoch": 0.22398424726173105, + "grad_norm": 0.41938984394073486, + "learning_rate": 0.0001762841592520917, + "loss": 0.0814, + "step": 9555 + }, + { + "epoch": 0.22400768883653605, + "grad_norm": 0.5210335850715637, + "learning_rate": 0.00017627939619488432, + "loss": 0.2129, + "step": 9556 + }, + { + "epoch": 0.22403113041134104, + "grad_norm": 0.29691559076309204, + "learning_rate": 0.0001762746327237828, + "loss": 0.0625, + "step": 9557 + }, + { + "epoch": 0.22405457198614603, + "grad_norm": 0.538474440574646, + "learning_rate": 0.000176269868838813, + "loss": 0.0908, + "step": 9558 + }, + { + "epoch": 0.22407801356095103, + "grad_norm": 0.1986548751592636, + "learning_rate": 0.0001762651045400007, + "loss": 0.0354, + "step": 9559 + }, + { + "epoch": 0.22410145513575602, + "grad_norm": 0.5902890563011169, + "learning_rate": 0.0001762603398273718, + "loss": 0.2084, + "step": 9560 + }, + { + "epoch": 0.22412489671056102, + "grad_norm": 0.623262882232666, + "learning_rate": 0.00017625557470095216, + "loss": 0.2334, + "step": 9561 + }, + { + "epoch": 0.224148338285366, + "grad_norm": 0.5782968997955322, + "learning_rate": 0.00017625080916076763, + "loss": 0.0706, + "step": 9562 + }, + { + "epoch": 0.224171779860171, + "grad_norm": 0.5622221231460571, + "learning_rate": 0.00017624604320684407, + "loss": 0.1734, + "step": 9563 + }, + { + "epoch": 0.224195221434976, + "grad_norm": 0.4875849187374115, + "learning_rate": 0.00017624127683920734, + "loss": 0.1007, + "step": 9564 + }, + { + "epoch": 0.224218663009781, + "grad_norm": 0.567709743976593, + "learning_rate": 0.0001762365100578833, + "loss": 0.1203, + "step": 9565 + }, + { + "epoch": 0.224242104584586, + "grad_norm": 0.3541492521762848, + "learning_rate": 0.0001762317428628978, + "loss": 0.3845, + "step": 9566 + }, + { + "epoch": 0.22426554615939098, + "grad_norm": 0.6537215709686279, + "learning_rate": 0.00017622697525427673, + "loss": 0.2215, + "step": 9567 + }, + { + "epoch": 0.22428898773419598, + "grad_norm": 0.19679901003837585, + "learning_rate": 0.00017622220723204594, + "loss": 0.0518, + "step": 9568 + }, + { + "epoch": 0.22431242930900097, + "grad_norm": 0.9015858173370361, + "learning_rate": 0.0001762174387962313, + "loss": 0.1905, + "step": 9569 + }, + { + "epoch": 0.22433587088380597, + "grad_norm": 0.5066947937011719, + "learning_rate": 0.00017621266994685873, + "loss": 0.1032, + "step": 9570 + }, + { + "epoch": 0.22435931245861096, + "grad_norm": 0.3930661976337433, + "learning_rate": 0.00017620790068395404, + "loss": 0.0623, + "step": 9571 + }, + { + "epoch": 0.22438275403341595, + "grad_norm": 0.2681390345096588, + "learning_rate": 0.00017620313100754316, + "loss": 0.0533, + "step": 9572 + }, + { + "epoch": 0.22440619560822095, + "grad_norm": 0.5397644639015198, + "learning_rate": 0.00017619836091765193, + "loss": 0.1036, + "step": 9573 + }, + { + "epoch": 0.22442963718302594, + "grad_norm": 0.4434335231781006, + "learning_rate": 0.00017619359041430626, + "loss": 0.0792, + "step": 9574 + }, + { + "epoch": 0.22445307875783094, + "grad_norm": 0.39053642749786377, + "learning_rate": 0.000176188819497532, + "loss": 0.1022, + "step": 9575 + }, + { + "epoch": 0.22447652033263596, + "grad_norm": 0.6749141216278076, + "learning_rate": 0.00017618404816735513, + "loss": 0.1076, + "step": 9576 + }, + { + "epoch": 0.22449996190744095, + "grad_norm": 0.3122097849845886, + "learning_rate": 0.0001761792764238014, + "loss": 0.061, + "step": 9577 + }, + { + "epoch": 0.22452340348224595, + "grad_norm": 0.4983031749725342, + "learning_rate": 0.0001761745042668968, + "loss": 0.4823, + "step": 9578 + }, + { + "epoch": 0.22454684505705094, + "grad_norm": 0.18623775243759155, + "learning_rate": 0.0001761697316966672, + "loss": 0.0482, + "step": 9579 + }, + { + "epoch": 0.22457028663185594, + "grad_norm": 0.5370126366615295, + "learning_rate": 0.0001761649587131385, + "loss": 0.1184, + "step": 9580 + }, + { + "epoch": 0.22459372820666093, + "grad_norm": 0.5170368552207947, + "learning_rate": 0.0001761601853163366, + "loss": 0.1094, + "step": 9581 + }, + { + "epoch": 0.22461716978146593, + "grad_norm": 0.26673758029937744, + "learning_rate": 0.00017615541150628735, + "loss": 0.0486, + "step": 9582 + }, + { + "epoch": 0.22464061135627092, + "grad_norm": 0.44574451446533203, + "learning_rate": 0.00017615063728301672, + "loss": 0.1667, + "step": 9583 + }, + { + "epoch": 0.2246640529310759, + "grad_norm": 0.19354841113090515, + "learning_rate": 0.00017614586264655063, + "loss": 0.0441, + "step": 9584 + }, + { + "epoch": 0.2246874945058809, + "grad_norm": 0.23089343309402466, + "learning_rate": 0.0001761410875969149, + "loss": 0.0513, + "step": 9585 + }, + { + "epoch": 0.2247109360806859, + "grad_norm": 0.17129428684711456, + "learning_rate": 0.0001761363121341355, + "loss": 0.031, + "step": 9586 + }, + { + "epoch": 0.2247343776554909, + "grad_norm": 0.7163568139076233, + "learning_rate": 0.00017613153625823836, + "loss": 0.2109, + "step": 9587 + }, + { + "epoch": 0.2247578192302959, + "grad_norm": 0.47879067063331604, + "learning_rate": 0.00017612675996924932, + "loss": 0.0916, + "step": 9588 + }, + { + "epoch": 0.22478126080510089, + "grad_norm": 0.519028902053833, + "learning_rate": 0.00017612198326719437, + "loss": 0.706, + "step": 9589 + }, + { + "epoch": 0.22480470237990588, + "grad_norm": 0.58564293384552, + "learning_rate": 0.00017611720615209935, + "loss": 0.1181, + "step": 9590 + }, + { + "epoch": 0.22482814395471087, + "grad_norm": 0.5725695490837097, + "learning_rate": 0.00017611242862399027, + "loss": 0.1474, + "step": 9591 + }, + { + "epoch": 0.22485158552951587, + "grad_norm": 0.1423484981060028, + "learning_rate": 0.000176107650682893, + "loss": 0.0206, + "step": 9592 + }, + { + "epoch": 0.22487502710432086, + "grad_norm": 0.21584196388721466, + "learning_rate": 0.0001761028723288335, + "loss": 0.0482, + "step": 9593 + }, + { + "epoch": 0.22489846867912586, + "grad_norm": 0.2839513123035431, + "learning_rate": 0.00017609809356183766, + "loss": 0.067, + "step": 9594 + }, + { + "epoch": 0.22492191025393085, + "grad_norm": 0.7775712013244629, + "learning_rate": 0.0001760933143819314, + "loss": 0.1185, + "step": 9595 + }, + { + "epoch": 0.22494535182873585, + "grad_norm": 0.6358787417411804, + "learning_rate": 0.0001760885347891407, + "loss": 0.1372, + "step": 9596 + }, + { + "epoch": 0.22496879340354084, + "grad_norm": 0.6314387321472168, + "learning_rate": 0.0001760837547834915, + "loss": 0.1485, + "step": 9597 + }, + { + "epoch": 0.22499223497834583, + "grad_norm": 0.5388137698173523, + "learning_rate": 0.00017607897436500966, + "loss": 0.0999, + "step": 9598 + }, + { + "epoch": 0.22501567655315083, + "grad_norm": 0.4730847179889679, + "learning_rate": 0.00017607419353372116, + "loss": 0.089, + "step": 9599 + }, + { + "epoch": 0.22503911812795582, + "grad_norm": 0.4759657680988312, + "learning_rate": 0.000176069412289652, + "loss": 0.0967, + "step": 9600 + }, + { + "epoch": 0.22506255970276084, + "grad_norm": 0.08146951347589493, + "learning_rate": 0.00017606463063282802, + "loss": 0.0145, + "step": 9601 + }, + { + "epoch": 0.22508600127756584, + "grad_norm": 0.6840486526489258, + "learning_rate": 0.00017605984856327523, + "loss": 0.8024, + "step": 9602 + }, + { + "epoch": 0.22510944285237083, + "grad_norm": 0.3900815546512604, + "learning_rate": 0.00017605506608101958, + "loss": 0.0958, + "step": 9603 + }, + { + "epoch": 0.22513288442717583, + "grad_norm": 0.7590513229370117, + "learning_rate": 0.000176050283186087, + "loss": 0.1462, + "step": 9604 + }, + { + "epoch": 0.22515632600198082, + "grad_norm": 0.8408234119415283, + "learning_rate": 0.00017604549987850344, + "loss": 0.8716, + "step": 9605 + }, + { + "epoch": 0.22517976757678582, + "grad_norm": 0.5175727009773254, + "learning_rate": 0.00017604071615829486, + "loss": 0.1093, + "step": 9606 + }, + { + "epoch": 0.2252032091515908, + "grad_norm": 0.5956825017929077, + "learning_rate": 0.00017603593202548717, + "loss": 0.0835, + "step": 9607 + }, + { + "epoch": 0.2252266507263958, + "grad_norm": 0.4794912338256836, + "learning_rate": 0.00017603114748010644, + "loss": 0.1131, + "step": 9608 + }, + { + "epoch": 0.2252500923012008, + "grad_norm": 0.5871694684028625, + "learning_rate": 0.00017602636252217856, + "loss": 0.7317, + "step": 9609 + }, + { + "epoch": 0.2252735338760058, + "grad_norm": 0.3713077902793884, + "learning_rate": 0.00017602157715172946, + "loss": 0.0629, + "step": 9610 + }, + { + "epoch": 0.2252969754508108, + "grad_norm": 0.7280458211898804, + "learning_rate": 0.00017601679136878518, + "loss": 0.1603, + "step": 9611 + }, + { + "epoch": 0.22532041702561578, + "grad_norm": 0.3941148817539215, + "learning_rate": 0.00017601200517337165, + "loss": 0.1304, + "step": 9612 + }, + { + "epoch": 0.22534385860042078, + "grad_norm": 0.975128710269928, + "learning_rate": 0.00017600721856551483, + "loss": 0.1185, + "step": 9613 + }, + { + "epoch": 0.22536730017522577, + "grad_norm": 0.157391756772995, + "learning_rate": 0.0001760024315452407, + "loss": 0.0378, + "step": 9614 + }, + { + "epoch": 0.22539074175003077, + "grad_norm": 0.7527853846549988, + "learning_rate": 0.00017599764411257524, + "loss": 0.1738, + "step": 9615 + }, + { + "epoch": 0.22541418332483576, + "grad_norm": 0.8664771914482117, + "learning_rate": 0.00017599285626754445, + "loss": 0.2018, + "step": 9616 + }, + { + "epoch": 0.22543762489964075, + "grad_norm": 0.4084315001964569, + "learning_rate": 0.00017598806801017426, + "loss": 0.0842, + "step": 9617 + }, + { + "epoch": 0.22546106647444575, + "grad_norm": 0.33664435148239136, + "learning_rate": 0.00017598327934049068, + "loss": 0.0756, + "step": 9618 + }, + { + "epoch": 0.22548450804925074, + "grad_norm": 0.8600096106529236, + "learning_rate": 0.0001759784902585197, + "loss": 0.087, + "step": 9619 + }, + { + "epoch": 0.22550794962405574, + "grad_norm": 0.40073901414871216, + "learning_rate": 0.0001759737007642873, + "loss": 0.0925, + "step": 9620 + }, + { + "epoch": 0.22553139119886073, + "grad_norm": 0.5653279423713684, + "learning_rate": 0.00017596891085781943, + "loss": 0.1043, + "step": 9621 + }, + { + "epoch": 0.22555483277366573, + "grad_norm": 0.7136891484260559, + "learning_rate": 0.00017596412053914214, + "loss": 0.1634, + "step": 9622 + }, + { + "epoch": 0.22557827434847072, + "grad_norm": 1.1641905307769775, + "learning_rate": 0.0001759593298082814, + "loss": 0.1256, + "step": 9623 + }, + { + "epoch": 0.22560171592327571, + "grad_norm": 0.5101980566978455, + "learning_rate": 0.0001759545386652632, + "loss": 0.1046, + "step": 9624 + }, + { + "epoch": 0.2256251574980807, + "grad_norm": 0.16809606552124023, + "learning_rate": 0.00017594974711011353, + "loss": 0.0368, + "step": 9625 + }, + { + "epoch": 0.2256485990728857, + "grad_norm": 0.32331156730651855, + "learning_rate": 0.00017594495514285837, + "loss": 0.0906, + "step": 9626 + }, + { + "epoch": 0.22567204064769072, + "grad_norm": 0.7602250576019287, + "learning_rate": 0.0001759401627635238, + "loss": 0.1806, + "step": 9627 + }, + { + "epoch": 0.22569548222249572, + "grad_norm": 0.6386082172393799, + "learning_rate": 0.00017593536997213575, + "loss": 0.0947, + "step": 9628 + }, + { + "epoch": 0.2257189237973007, + "grad_norm": 0.41444045305252075, + "learning_rate": 0.00017593057676872025, + "loss": 0.532, + "step": 9629 + }, + { + "epoch": 0.2257423653721057, + "grad_norm": 0.3617234528064728, + "learning_rate": 0.0001759257831533033, + "loss": 0.1401, + "step": 9630 + }, + { + "epoch": 0.2257658069469107, + "grad_norm": 0.5572789907455444, + "learning_rate": 0.0001759209891259109, + "loss": 0.1875, + "step": 9631 + }, + { + "epoch": 0.2257892485217157, + "grad_norm": 0.16574354469776154, + "learning_rate": 0.0001759161946865691, + "loss": 0.0485, + "step": 9632 + }, + { + "epoch": 0.2258126900965207, + "grad_norm": 0.256206750869751, + "learning_rate": 0.0001759113998353039, + "loss": 0.0382, + "step": 9633 + }, + { + "epoch": 0.22583613167132568, + "grad_norm": 0.47009778022766113, + "learning_rate": 0.0001759066045721413, + "loss": 0.093, + "step": 9634 + }, + { + "epoch": 0.22585957324613068, + "grad_norm": 0.23074723780155182, + "learning_rate": 0.00017590180889710732, + "loss": 0.0358, + "step": 9635 + }, + { + "epoch": 0.22588301482093567, + "grad_norm": 0.509398877620697, + "learning_rate": 0.000175897012810228, + "loss": 0.1619, + "step": 9636 + }, + { + "epoch": 0.22590645639574067, + "grad_norm": 0.3548509478569031, + "learning_rate": 0.00017589221631152935, + "loss": 0.0567, + "step": 9637 + }, + { + "epoch": 0.22592989797054566, + "grad_norm": 0.8275394439697266, + "learning_rate": 0.00017588741940103742, + "loss": 0.1679, + "step": 9638 + }, + { + "epoch": 0.22595333954535066, + "grad_norm": 0.39927592873573303, + "learning_rate": 0.0001758826220787782, + "loss": 0.0775, + "step": 9639 + }, + { + "epoch": 0.22597678112015565, + "grad_norm": 0.909253716468811, + "learning_rate": 0.0001758778243447777, + "loss": 0.183, + "step": 9640 + }, + { + "epoch": 0.22600022269496065, + "grad_norm": 0.14408168196678162, + "learning_rate": 0.00017587302619906205, + "loss": 0.0358, + "step": 9641 + }, + { + "epoch": 0.22602366426976564, + "grad_norm": 0.4577227234840393, + "learning_rate": 0.0001758682276416572, + "loss": 0.0588, + "step": 9642 + }, + { + "epoch": 0.22604710584457063, + "grad_norm": 0.42167237401008606, + "learning_rate": 0.0001758634286725892, + "loss": 0.0596, + "step": 9643 + }, + { + "epoch": 0.22607054741937563, + "grad_norm": 0.26493537425994873, + "learning_rate": 0.00017585862929188416, + "loss": 0.0369, + "step": 9644 + }, + { + "epoch": 0.22609398899418062, + "grad_norm": 0.7121134996414185, + "learning_rate": 0.000175853829499568, + "loss": 0.2009, + "step": 9645 + }, + { + "epoch": 0.22611743056898562, + "grad_norm": 0.4005148112773895, + "learning_rate": 0.0001758490292956669, + "loss": 0.0783, + "step": 9646 + }, + { + "epoch": 0.2261408721437906, + "grad_norm": 0.7892176508903503, + "learning_rate": 0.00017584422868020678, + "loss": 0.1272, + "step": 9647 + }, + { + "epoch": 0.2261643137185956, + "grad_norm": 0.20026659965515137, + "learning_rate": 0.00017583942765321374, + "loss": 0.0471, + "step": 9648 + }, + { + "epoch": 0.2261877552934006, + "grad_norm": 0.2918206453323364, + "learning_rate": 0.00017583462621471388, + "loss": 0.0368, + "step": 9649 + }, + { + "epoch": 0.2262111968682056, + "grad_norm": 0.30136799812316895, + "learning_rate": 0.00017582982436473317, + "loss": 0.0714, + "step": 9650 + }, + { + "epoch": 0.2262346384430106, + "grad_norm": 0.3681640326976776, + "learning_rate": 0.00017582502210329769, + "loss": 0.0689, + "step": 9651 + }, + { + "epoch": 0.2262580800178156, + "grad_norm": 0.7284814119338989, + "learning_rate": 0.00017582021943043352, + "loss": 0.1711, + "step": 9652 + }, + { + "epoch": 0.2262815215926206, + "grad_norm": 0.5907042026519775, + "learning_rate": 0.00017581541634616673, + "loss": 0.6128, + "step": 9653 + }, + { + "epoch": 0.2263049631674256, + "grad_norm": 0.6526538729667664, + "learning_rate": 0.00017581061285052333, + "loss": 0.1504, + "step": 9654 + }, + { + "epoch": 0.2263284047422306, + "grad_norm": 0.8438609838485718, + "learning_rate": 0.0001758058089435294, + "loss": 0.0706, + "step": 9655 + }, + { + "epoch": 0.2263518463170356, + "grad_norm": 0.8331573605537415, + "learning_rate": 0.00017580100462521105, + "loss": 0.1163, + "step": 9656 + }, + { + "epoch": 0.22637528789184058, + "grad_norm": 0.49606797099113464, + "learning_rate": 0.00017579619989559434, + "loss": 0.1729, + "step": 9657 + }, + { + "epoch": 0.22639872946664558, + "grad_norm": 0.21218253672122955, + "learning_rate": 0.00017579139475470528, + "loss": 0.0239, + "step": 9658 + }, + { + "epoch": 0.22642217104145057, + "grad_norm": 0.5096657872200012, + "learning_rate": 0.00017578658920256997, + "loss": 0.1246, + "step": 9659 + }, + { + "epoch": 0.22644561261625556, + "grad_norm": 0.6034578680992126, + "learning_rate": 0.00017578178323921452, + "loss": 0.221, + "step": 9660 + }, + { + "epoch": 0.22646905419106056, + "grad_norm": 0.15860003232955933, + "learning_rate": 0.00017577697686466497, + "loss": 0.0411, + "step": 9661 + }, + { + "epoch": 0.22649249576586555, + "grad_norm": 0.14653296768665314, + "learning_rate": 0.0001757721700789474, + "loss": 0.033, + "step": 9662 + }, + { + "epoch": 0.22651593734067055, + "grad_norm": 0.4583609402179718, + "learning_rate": 0.00017576736288208794, + "loss": 0.1258, + "step": 9663 + }, + { + "epoch": 0.22653937891547554, + "grad_norm": 0.7069479823112488, + "learning_rate": 0.00017576255527411262, + "loss": 0.1981, + "step": 9664 + }, + { + "epoch": 0.22656282049028054, + "grad_norm": 0.8390872478485107, + "learning_rate": 0.00017575774725504757, + "loss": 0.7326, + "step": 9665 + }, + { + "epoch": 0.22658626206508553, + "grad_norm": 0.5956442356109619, + "learning_rate": 0.00017575293882491881, + "loss": 0.1224, + "step": 9666 + }, + { + "epoch": 0.22660970363989053, + "grad_norm": 0.6977460384368896, + "learning_rate": 0.0001757481299837525, + "loss": 0.2354, + "step": 9667 + }, + { + "epoch": 0.22663314521469552, + "grad_norm": 0.3688708245754242, + "learning_rate": 0.00017574332073157472, + "loss": 0.0926, + "step": 9668 + }, + { + "epoch": 0.2266565867895005, + "grad_norm": 0.36992207169532776, + "learning_rate": 0.0001757385110684115, + "loss": 0.0798, + "step": 9669 + }, + { + "epoch": 0.2266800283643055, + "grad_norm": 0.46035298705101013, + "learning_rate": 0.00017573370099428906, + "loss": 0.0974, + "step": 9670 + }, + { + "epoch": 0.2267034699391105, + "grad_norm": 0.40192320942878723, + "learning_rate": 0.00017572889050923335, + "loss": 0.0784, + "step": 9671 + }, + { + "epoch": 0.2267269115139155, + "grad_norm": 0.18580210208892822, + "learning_rate": 0.0001757240796132706, + "loss": 0.0442, + "step": 9672 + }, + { + "epoch": 0.2267503530887205, + "grad_norm": 0.2842116951942444, + "learning_rate": 0.00017571926830642687, + "loss": 0.0751, + "step": 9673 + }, + { + "epoch": 0.22677379466352549, + "grad_norm": 0.3066639006137848, + "learning_rate": 0.00017571445658872824, + "loss": 0.0672, + "step": 9674 + }, + { + "epoch": 0.22679723623833048, + "grad_norm": 0.6220678687095642, + "learning_rate": 0.00017570964446020088, + "loss": 0.1182, + "step": 9675 + }, + { + "epoch": 0.22682067781313547, + "grad_norm": 0.49002084136009216, + "learning_rate": 0.0001757048319208708, + "loss": 0.0954, + "step": 9676 + }, + { + "epoch": 0.22684411938794047, + "grad_norm": 0.5594515800476074, + "learning_rate": 0.0001757000189707642, + "loss": 0.0611, + "step": 9677 + }, + { + "epoch": 0.2268675609627455, + "grad_norm": 0.5643510818481445, + "learning_rate": 0.00017569520560990718, + "loss": 0.1924, + "step": 9678 + }, + { + "epoch": 0.22689100253755048, + "grad_norm": 0.22122301161289215, + "learning_rate": 0.00017569039183832586, + "loss": 0.0363, + "step": 9679 + }, + { + "epoch": 0.22691444411235548, + "grad_norm": 0.5507302284240723, + "learning_rate": 0.00017568557765604628, + "loss": 0.3535, + "step": 9680 + }, + { + "epoch": 0.22693788568716047, + "grad_norm": 0.4306029975414276, + "learning_rate": 0.0001756807630630947, + "loss": 0.0913, + "step": 9681 + }, + { + "epoch": 0.22696132726196547, + "grad_norm": 0.4920825660228729, + "learning_rate": 0.00017567594805949715, + "loss": 0.1164, + "step": 9682 + }, + { + "epoch": 0.22698476883677046, + "grad_norm": 1.0393046140670776, + "learning_rate": 0.00017567113264527976, + "loss": 0.194, + "step": 9683 + }, + { + "epoch": 0.22700821041157546, + "grad_norm": 0.7136772871017456, + "learning_rate": 0.00017566631682046867, + "loss": 0.1021, + "step": 9684 + }, + { + "epoch": 0.22703165198638045, + "grad_norm": 0.24457195401191711, + "learning_rate": 0.00017566150058509006, + "loss": 0.0621, + "step": 9685 + }, + { + "epoch": 0.22705509356118544, + "grad_norm": 0.19889838993549347, + "learning_rate": 0.00017565668393916997, + "loss": 0.0681, + "step": 9686 + }, + { + "epoch": 0.22707853513599044, + "grad_norm": 0.27010419964790344, + "learning_rate": 0.00017565186688273463, + "loss": 0.049, + "step": 9687 + }, + { + "epoch": 0.22710197671079543, + "grad_norm": 0.6010445356369019, + "learning_rate": 0.00017564704941581012, + "loss": 0.1023, + "step": 9688 + }, + { + "epoch": 0.22712541828560043, + "grad_norm": 0.33880123496055603, + "learning_rate": 0.00017564223153842257, + "loss": 0.0613, + "step": 9689 + }, + { + "epoch": 0.22714885986040542, + "grad_norm": 0.4579766094684601, + "learning_rate": 0.00017563741325059818, + "loss": 0.1163, + "step": 9690 + }, + { + "epoch": 0.22717230143521042, + "grad_norm": 0.40687233209609985, + "learning_rate": 0.00017563259455236302, + "loss": 0.0776, + "step": 9691 + }, + { + "epoch": 0.2271957430100154, + "grad_norm": 1.0326560735702515, + "learning_rate": 0.00017562777544374332, + "loss": 0.1882, + "step": 9692 + }, + { + "epoch": 0.2272191845848204, + "grad_norm": 0.5659423470497131, + "learning_rate": 0.00017562295592476513, + "loss": 0.0899, + "step": 9693 + }, + { + "epoch": 0.2272426261596254, + "grad_norm": 0.35467082262039185, + "learning_rate": 0.0001756181359954547, + "loss": 0.0656, + "step": 9694 + }, + { + "epoch": 0.2272660677344304, + "grad_norm": 0.7923446893692017, + "learning_rate": 0.0001756133156558381, + "loss": 0.1356, + "step": 9695 + }, + { + "epoch": 0.2272895093092354, + "grad_norm": 0.25195086002349854, + "learning_rate": 0.00017560849490594156, + "loss": 0.0646, + "step": 9696 + }, + { + "epoch": 0.22731295088404038, + "grad_norm": 0.5514370799064636, + "learning_rate": 0.0001756036737457912, + "loss": 0.035, + "step": 9697 + }, + { + "epoch": 0.22733639245884538, + "grad_norm": 0.14460690319538116, + "learning_rate": 0.00017559885217541317, + "loss": 0.0217, + "step": 9698 + }, + { + "epoch": 0.22735983403365037, + "grad_norm": 0.666865348815918, + "learning_rate": 0.0001755940301948336, + "loss": 0.1358, + "step": 9699 + }, + { + "epoch": 0.22738327560845537, + "grad_norm": 0.9055926203727722, + "learning_rate": 0.00017558920780407877, + "loss": 0.1281, + "step": 9700 + }, + { + "epoch": 0.22740671718326036, + "grad_norm": 1.019667387008667, + "learning_rate": 0.00017558438500317475, + "loss": 0.2381, + "step": 9701 + }, + { + "epoch": 0.22743015875806535, + "grad_norm": 0.5240825414657593, + "learning_rate": 0.00017557956179214772, + "loss": 0.0837, + "step": 9702 + }, + { + "epoch": 0.22745360033287038, + "grad_norm": 0.43513187766075134, + "learning_rate": 0.00017557473817102386, + "loss": 0.1211, + "step": 9703 + }, + { + "epoch": 0.22747704190767537, + "grad_norm": 0.4540485441684723, + "learning_rate": 0.00017556991413982936, + "loss": 0.106, + "step": 9704 + }, + { + "epoch": 0.22750048348248036, + "grad_norm": 0.3856518268585205, + "learning_rate": 0.00017556508969859036, + "loss": 0.0626, + "step": 9705 + }, + { + "epoch": 0.22752392505728536, + "grad_norm": 0.5272149443626404, + "learning_rate": 0.00017556026484733306, + "loss": 0.1384, + "step": 9706 + }, + { + "epoch": 0.22754736663209035, + "grad_norm": 0.5742329955101013, + "learning_rate": 0.00017555543958608363, + "loss": 0.6995, + "step": 9707 + }, + { + "epoch": 0.22757080820689535, + "grad_norm": 0.18079589307308197, + "learning_rate": 0.00017555061391486828, + "loss": 0.0257, + "step": 9708 + }, + { + "epoch": 0.22759424978170034, + "grad_norm": 0.17280367016792297, + "learning_rate": 0.00017554578783371318, + "loss": 0.0276, + "step": 9709 + }, + { + "epoch": 0.22761769135650534, + "grad_norm": 1.0267850160598755, + "learning_rate": 0.0001755409613426445, + "loss": 0.0843, + "step": 9710 + }, + { + "epoch": 0.22764113293131033, + "grad_norm": 0.3366166055202484, + "learning_rate": 0.00017553613444168845, + "loss": 0.0794, + "step": 9711 + }, + { + "epoch": 0.22766457450611532, + "grad_norm": 0.5435163974761963, + "learning_rate": 0.0001755313071308712, + "loss": 0.1214, + "step": 9712 + }, + { + "epoch": 0.22768801608092032, + "grad_norm": 0.3606480658054352, + "learning_rate": 0.00017552647941021897, + "loss": 0.0737, + "step": 9713 + }, + { + "epoch": 0.2277114576557253, + "grad_norm": 0.6948080062866211, + "learning_rate": 0.0001755216512797579, + "loss": 0.5178, + "step": 9714 + }, + { + "epoch": 0.2277348992305303, + "grad_norm": 0.39001739025115967, + "learning_rate": 0.00017551682273951426, + "loss": 0.1114, + "step": 9715 + }, + { + "epoch": 0.2277583408053353, + "grad_norm": 0.6780891418457031, + "learning_rate": 0.00017551199378951422, + "loss": 0.6519, + "step": 9716 + }, + { + "epoch": 0.2277817823801403, + "grad_norm": 0.6664759516716003, + "learning_rate": 0.00017550716442978396, + "loss": 0.1548, + "step": 9717 + }, + { + "epoch": 0.2278052239549453, + "grad_norm": 0.318346232175827, + "learning_rate": 0.0001755023346603497, + "loss": 0.0373, + "step": 9718 + }, + { + "epoch": 0.22782866552975028, + "grad_norm": 0.9576539397239685, + "learning_rate": 0.0001754975044812377, + "loss": 0.1394, + "step": 9719 + }, + { + "epoch": 0.22785210710455528, + "grad_norm": 0.6999953389167786, + "learning_rate": 0.00017549267389247404, + "loss": 0.1722, + "step": 9720 + }, + { + "epoch": 0.22787554867936027, + "grad_norm": 0.1680929809808731, + "learning_rate": 0.00017548784289408506, + "loss": 0.0383, + "step": 9721 + }, + { + "epoch": 0.22789899025416527, + "grad_norm": 0.5651943683624268, + "learning_rate": 0.00017548301148609692, + "loss": 0.1213, + "step": 9722 + }, + { + "epoch": 0.22792243182897026, + "grad_norm": 0.6131528615951538, + "learning_rate": 0.00017547817966853582, + "loss": 0.07, + "step": 9723 + }, + { + "epoch": 0.22794587340377526, + "grad_norm": 0.6261026263237, + "learning_rate": 0.00017547334744142797, + "loss": 0.1117, + "step": 9724 + }, + { + "epoch": 0.22796931497858025, + "grad_norm": 0.17736677825450897, + "learning_rate": 0.00017546851480479967, + "loss": 0.0502, + "step": 9725 + }, + { + "epoch": 0.22799275655338525, + "grad_norm": 0.5325525999069214, + "learning_rate": 0.00017546368175867704, + "loss": 0.1429, + "step": 9726 + }, + { + "epoch": 0.22801619812819024, + "grad_norm": 0.6921891570091248, + "learning_rate": 0.00017545884830308639, + "loss": 0.1224, + "step": 9727 + }, + { + "epoch": 0.22803963970299523, + "grad_norm": 0.38081902265548706, + "learning_rate": 0.0001754540144380539, + "loss": 0.0673, + "step": 9728 + }, + { + "epoch": 0.22806308127780026, + "grad_norm": 0.37402835488319397, + "learning_rate": 0.0001754491801636058, + "loss": 0.0687, + "step": 9729 + }, + { + "epoch": 0.22808652285260525, + "grad_norm": 0.3978549838066101, + "learning_rate": 0.00017544434547976832, + "loss": 0.0698, + "step": 9730 + }, + { + "epoch": 0.22810996442741024, + "grad_norm": 0.3258509039878845, + "learning_rate": 0.0001754395103865677, + "loss": 0.0733, + "step": 9731 + }, + { + "epoch": 0.22813340600221524, + "grad_norm": 0.893958568572998, + "learning_rate": 0.00017543467488403017, + "loss": 0.2489, + "step": 9732 + }, + { + "epoch": 0.22815684757702023, + "grad_norm": 0.5735569596290588, + "learning_rate": 0.00017542983897218196, + "loss": 0.171, + "step": 9733 + }, + { + "epoch": 0.22818028915182523, + "grad_norm": 0.44390666484832764, + "learning_rate": 0.00017542500265104938, + "loss": 0.1299, + "step": 9734 + }, + { + "epoch": 0.22820373072663022, + "grad_norm": 0.2945023775100708, + "learning_rate": 0.00017542016592065858, + "loss": 0.0839, + "step": 9735 + }, + { + "epoch": 0.22822717230143522, + "grad_norm": 0.16915114223957062, + "learning_rate": 0.00017541532878103584, + "loss": 0.0351, + "step": 9736 + }, + { + "epoch": 0.2282506138762402, + "grad_norm": 0.9426842331886292, + "learning_rate": 0.0001754104912322074, + "loss": 0.1641, + "step": 9737 + }, + { + "epoch": 0.2282740554510452, + "grad_norm": 0.41849416494369507, + "learning_rate": 0.0001754056532741995, + "loss": 0.0757, + "step": 9738 + }, + { + "epoch": 0.2282974970258502, + "grad_norm": 1.1891090869903564, + "learning_rate": 0.00017540081490703842, + "loss": 0.3949, + "step": 9739 + }, + { + "epoch": 0.2283209386006552, + "grad_norm": 0.5299246311187744, + "learning_rate": 0.00017539597613075043, + "loss": 0.1233, + "step": 9740 + }, + { + "epoch": 0.2283443801754602, + "grad_norm": 0.6183013319969177, + "learning_rate": 0.0001753911369453617, + "loss": 0.1278, + "step": 9741 + }, + { + "epoch": 0.22836782175026518, + "grad_norm": 0.6720483303070068, + "learning_rate": 0.00017538629735089857, + "loss": 0.103, + "step": 9742 + }, + { + "epoch": 0.22839126332507018, + "grad_norm": 0.36184871196746826, + "learning_rate": 0.00017538145734738726, + "loss": 0.0689, + "step": 9743 + }, + { + "epoch": 0.22841470489987517, + "grad_norm": 0.47910749912261963, + "learning_rate": 0.00017537661693485407, + "loss": 0.1439, + "step": 9744 + }, + { + "epoch": 0.22843814647468016, + "grad_norm": 0.38731321692466736, + "learning_rate": 0.0001753717761133252, + "loss": 0.0841, + "step": 9745 + }, + { + "epoch": 0.22846158804948516, + "grad_norm": 0.5440340638160706, + "learning_rate": 0.00017536693488282698, + "loss": 0.1192, + "step": 9746 + }, + { + "epoch": 0.22848502962429015, + "grad_norm": 0.5945267677307129, + "learning_rate": 0.00017536209324338565, + "loss": 0.6861, + "step": 9747 + }, + { + "epoch": 0.22850847119909515, + "grad_norm": 0.7981118559837341, + "learning_rate": 0.0001753572511950275, + "loss": 0.1636, + "step": 9748 + }, + { + "epoch": 0.22853191277390014, + "grad_norm": 0.6974382400512695, + "learning_rate": 0.00017535240873777874, + "loss": 0.1167, + "step": 9749 + }, + { + "epoch": 0.22855535434870514, + "grad_norm": 0.6362550854682922, + "learning_rate": 0.0001753475658716657, + "loss": 0.114, + "step": 9750 + }, + { + "epoch": 0.22857879592351013, + "grad_norm": 0.4911833703517914, + "learning_rate": 0.00017534272259671466, + "loss": 0.1107, + "step": 9751 + }, + { + "epoch": 0.22860223749831513, + "grad_norm": 0.7125426530838013, + "learning_rate": 0.0001753378789129519, + "loss": 0.1166, + "step": 9752 + }, + { + "epoch": 0.22862567907312012, + "grad_norm": 0.21129737794399261, + "learning_rate": 0.00017533303482040368, + "loss": 0.0442, + "step": 9753 + }, + { + "epoch": 0.22864912064792514, + "grad_norm": 0.8589712381362915, + "learning_rate": 0.0001753281903190963, + "loss": 0.1353, + "step": 9754 + }, + { + "epoch": 0.22867256222273014, + "grad_norm": 0.56416255235672, + "learning_rate": 0.00017532334540905603, + "loss": 0.1345, + "step": 9755 + }, + { + "epoch": 0.22869600379753513, + "grad_norm": 0.4962238669395447, + "learning_rate": 0.0001753185000903092, + "loss": 0.1147, + "step": 9756 + }, + { + "epoch": 0.22871944537234012, + "grad_norm": 0.5958768725395203, + "learning_rate": 0.00017531365436288204, + "loss": 0.5448, + "step": 9757 + }, + { + "epoch": 0.22874288694714512, + "grad_norm": 0.47737205028533936, + "learning_rate": 0.00017530880822680088, + "loss": 0.1032, + "step": 9758 + }, + { + "epoch": 0.2287663285219501, + "grad_norm": 0.44610559940338135, + "learning_rate": 0.000175303961682092, + "loss": 0.1139, + "step": 9759 + }, + { + "epoch": 0.2287897700967551, + "grad_norm": 0.6675467491149902, + "learning_rate": 0.00017529911472878172, + "loss": 0.0942, + "step": 9760 + }, + { + "epoch": 0.2288132116715601, + "grad_norm": 0.3792840838432312, + "learning_rate": 0.00017529426736689633, + "loss": 0.041, + "step": 9761 + }, + { + "epoch": 0.2288366532463651, + "grad_norm": 0.49361997842788696, + "learning_rate": 0.00017528941959646212, + "loss": 0.0914, + "step": 9762 + }, + { + "epoch": 0.2288600948211701, + "grad_norm": 0.5075053572654724, + "learning_rate": 0.0001752845714175054, + "loss": 0.1234, + "step": 9763 + }, + { + "epoch": 0.22888353639597508, + "grad_norm": 0.5294249057769775, + "learning_rate": 0.00017527972283005249, + "loss": 0.1258, + "step": 9764 + }, + { + "epoch": 0.22890697797078008, + "grad_norm": 0.24137823283672333, + "learning_rate": 0.00017527487383412968, + "loss": 0.0523, + "step": 9765 + }, + { + "epoch": 0.22893041954558507, + "grad_norm": 0.6800529956817627, + "learning_rate": 0.00017527002442976327, + "loss": 0.1533, + "step": 9766 + }, + { + "epoch": 0.22895386112039007, + "grad_norm": 0.7076426148414612, + "learning_rate": 0.00017526517461697965, + "loss": 0.2098, + "step": 9767 + }, + { + "epoch": 0.22897730269519506, + "grad_norm": 0.37514010071754456, + "learning_rate": 0.00017526032439580503, + "loss": 0.131, + "step": 9768 + }, + { + "epoch": 0.22900074427000006, + "grad_norm": 0.7343341112136841, + "learning_rate": 0.00017525547376626578, + "loss": 0.6797, + "step": 9769 + }, + { + "epoch": 0.22902418584480505, + "grad_norm": 0.37200355529785156, + "learning_rate": 0.0001752506227283882, + "loss": 0.0886, + "step": 9770 + }, + { + "epoch": 0.22904762741961004, + "grad_norm": 0.946801483631134, + "learning_rate": 0.00017524577128219865, + "loss": 0.1032, + "step": 9771 + }, + { + "epoch": 0.22907106899441504, + "grad_norm": 0.7025704383850098, + "learning_rate": 0.0001752409194277234, + "loss": 0.1638, + "step": 9772 + }, + { + "epoch": 0.22909451056922003, + "grad_norm": 1.022057056427002, + "learning_rate": 0.00017523606716498885, + "loss": 0.2789, + "step": 9773 + }, + { + "epoch": 0.22911795214402503, + "grad_norm": 0.330609530210495, + "learning_rate": 0.00017523121449402123, + "loss": 0.0662, + "step": 9774 + }, + { + "epoch": 0.22914139371883002, + "grad_norm": 0.7719431519508362, + "learning_rate": 0.00017522636141484695, + "loss": 0.8216, + "step": 9775 + }, + { + "epoch": 0.22916483529363502, + "grad_norm": 0.5847081542015076, + "learning_rate": 0.0001752215079274923, + "loss": 0.0902, + "step": 9776 + }, + { + "epoch": 0.22918827686844, + "grad_norm": 0.30836358666419983, + "learning_rate": 0.0001752166540319837, + "loss": 0.0737, + "step": 9777 + }, + { + "epoch": 0.229211718443245, + "grad_norm": 0.475717693567276, + "learning_rate": 0.00017521179972834736, + "loss": 0.1215, + "step": 9778 + }, + { + "epoch": 0.22923516001805, + "grad_norm": 0.27801772952079773, + "learning_rate": 0.00017520694501660971, + "loss": 0.0732, + "step": 9779 + }, + { + "epoch": 0.22925860159285502, + "grad_norm": 0.7200757265090942, + "learning_rate": 0.00017520208989679702, + "loss": 0.1053, + "step": 9780 + }, + { + "epoch": 0.22928204316766002, + "grad_norm": 0.5579872131347656, + "learning_rate": 0.00017519723436893573, + "loss": 0.13, + "step": 9781 + }, + { + "epoch": 0.229305484742465, + "grad_norm": 0.42698773741722107, + "learning_rate": 0.00017519237843305208, + "loss": 0.0915, + "step": 9782 + }, + { + "epoch": 0.22932892631727, + "grad_norm": 0.5919968485832214, + "learning_rate": 0.0001751875220891725, + "loss": 0.748, + "step": 9783 + }, + { + "epoch": 0.229352367892075, + "grad_norm": 0.5103535056114197, + "learning_rate": 0.0001751826653373233, + "loss": 0.1285, + "step": 9784 + }, + { + "epoch": 0.22937580946688, + "grad_norm": 0.471973717212677, + "learning_rate": 0.00017517780817753083, + "loss": 0.091, + "step": 9785 + }, + { + "epoch": 0.229399251041685, + "grad_norm": 0.40043342113494873, + "learning_rate": 0.00017517295060982151, + "loss": 0.1142, + "step": 9786 + }, + { + "epoch": 0.22942269261648998, + "grad_norm": 1.0445165634155273, + "learning_rate": 0.0001751680926342216, + "loss": 0.3031, + "step": 9787 + }, + { + "epoch": 0.22944613419129498, + "grad_norm": 0.18012568354606628, + "learning_rate": 0.0001751632342507575, + "loss": 0.048, + "step": 9788 + }, + { + "epoch": 0.22946957576609997, + "grad_norm": 0.9087527394294739, + "learning_rate": 0.00017515837545945565, + "loss": 0.2154, + "step": 9789 + }, + { + "epoch": 0.22949301734090496, + "grad_norm": 0.1220538318157196, + "learning_rate": 0.00017515351626034225, + "loss": 0.026, + "step": 9790 + }, + { + "epoch": 0.22951645891570996, + "grad_norm": 0.30732256174087524, + "learning_rate": 0.0001751486566534438, + "loss": 0.0668, + "step": 9791 + }, + { + "epoch": 0.22953990049051495, + "grad_norm": 0.5899932980537415, + "learning_rate": 0.00017514379663878665, + "loss": 0.104, + "step": 9792 + }, + { + "epoch": 0.22956334206531995, + "grad_norm": 0.6055413484573364, + "learning_rate": 0.0001751389362163971, + "loss": 0.1482, + "step": 9793 + }, + { + "epoch": 0.22958678364012494, + "grad_norm": 0.13381381332874298, + "learning_rate": 0.0001751340753863016, + "loss": 0.0163, + "step": 9794 + }, + { + "epoch": 0.22961022521492994, + "grad_norm": 0.48676377534866333, + "learning_rate": 0.00017512921414852646, + "loss": 0.1584, + "step": 9795 + }, + { + "epoch": 0.22963366678973493, + "grad_norm": 0.2820839583873749, + "learning_rate": 0.0001751243525030981, + "loss": 0.0824, + "step": 9796 + }, + { + "epoch": 0.22965710836453992, + "grad_norm": 0.43090683221817017, + "learning_rate": 0.0001751194904500429, + "loss": 0.0801, + "step": 9797 + }, + { + "epoch": 0.22968054993934492, + "grad_norm": 0.5340858101844788, + "learning_rate": 0.00017511462798938725, + "loss": 0.5479, + "step": 9798 + }, + { + "epoch": 0.2297039915141499, + "grad_norm": 0.36648908257484436, + "learning_rate": 0.00017510976512115752, + "loss": 0.0933, + "step": 9799 + }, + { + "epoch": 0.2297274330889549, + "grad_norm": 0.3721422851085663, + "learning_rate": 0.00017510490184538007, + "loss": 0.103, + "step": 9800 + }, + { + "epoch": 0.2297508746637599, + "grad_norm": 0.47195374965667725, + "learning_rate": 0.00017510003816208132, + "loss": 0.1492, + "step": 9801 + }, + { + "epoch": 0.2297743162385649, + "grad_norm": 0.30489474534988403, + "learning_rate": 0.00017509517407128766, + "loss": 0.0651, + "step": 9802 + }, + { + "epoch": 0.2297977578133699, + "grad_norm": 0.3462965190410614, + "learning_rate": 0.00017509030957302545, + "loss": 0.0601, + "step": 9803 + }, + { + "epoch": 0.22982119938817488, + "grad_norm": 0.5851407647132874, + "learning_rate": 0.00017508544466732112, + "loss": 0.1223, + "step": 9804 + }, + { + "epoch": 0.22984464096297988, + "grad_norm": 0.27984505891799927, + "learning_rate": 0.00017508057935420104, + "loss": 0.0631, + "step": 9805 + }, + { + "epoch": 0.2298680825377849, + "grad_norm": 0.5259984731674194, + "learning_rate": 0.00017507571363369165, + "loss": 0.6776, + "step": 9806 + }, + { + "epoch": 0.2298915241125899, + "grad_norm": 0.5006417036056519, + "learning_rate": 0.00017507084750581934, + "loss": 0.1075, + "step": 9807 + }, + { + "epoch": 0.2299149656873949, + "grad_norm": 0.6448625326156616, + "learning_rate": 0.00017506598097061045, + "loss": 0.6613, + "step": 9808 + }, + { + "epoch": 0.22993840726219988, + "grad_norm": 0.732010006904602, + "learning_rate": 0.00017506111402809148, + "loss": 0.118, + "step": 9809 + }, + { + "epoch": 0.22996184883700488, + "grad_norm": 0.21162743866443634, + "learning_rate": 0.00017505624667828878, + "loss": 0.0364, + "step": 9810 + }, + { + "epoch": 0.22998529041180987, + "grad_norm": 0.38666969537734985, + "learning_rate": 0.00017505137892122877, + "loss": 0.0977, + "step": 9811 + }, + { + "epoch": 0.23000873198661487, + "grad_norm": 0.4796125590801239, + "learning_rate": 0.0001750465107569379, + "loss": 0.1599, + "step": 9812 + }, + { + "epoch": 0.23003217356141986, + "grad_norm": 0.47867387533187866, + "learning_rate": 0.0001750416421854425, + "loss": 0.1606, + "step": 9813 + }, + { + "epoch": 0.23005561513622486, + "grad_norm": 0.5916839241981506, + "learning_rate": 0.0001750367732067691, + "loss": 0.2368, + "step": 9814 + }, + { + "epoch": 0.23007905671102985, + "grad_norm": 0.43050557374954224, + "learning_rate": 0.000175031903820944, + "loss": 0.1597, + "step": 9815 + }, + { + "epoch": 0.23010249828583484, + "grad_norm": 0.2609682083129883, + "learning_rate": 0.0001750270340279937, + "loss": 0.0456, + "step": 9816 + }, + { + "epoch": 0.23012593986063984, + "grad_norm": 0.7600001692771912, + "learning_rate": 0.0001750221638279446, + "loss": 0.2012, + "step": 9817 + }, + { + "epoch": 0.23014938143544483, + "grad_norm": 0.3902962803840637, + "learning_rate": 0.00017501729322082316, + "loss": 0.0692, + "step": 9818 + }, + { + "epoch": 0.23017282301024983, + "grad_norm": 0.5847262740135193, + "learning_rate": 0.00017501242220665576, + "loss": 0.1424, + "step": 9819 + }, + { + "epoch": 0.23019626458505482, + "grad_norm": 0.20167535543441772, + "learning_rate": 0.00017500755078546886, + "loss": 0.0412, + "step": 9820 + }, + { + "epoch": 0.23021970615985982, + "grad_norm": 0.6504358649253845, + "learning_rate": 0.00017500267895728886, + "loss": 0.1769, + "step": 9821 + }, + { + "epoch": 0.2302431477346648, + "grad_norm": 0.8800047636032104, + "learning_rate": 0.00017499780672214222, + "loss": 0.1445, + "step": 9822 + }, + { + "epoch": 0.2302665893094698, + "grad_norm": 0.2560732066631317, + "learning_rate": 0.00017499293408005534, + "loss": 0.0667, + "step": 9823 + }, + { + "epoch": 0.2302900308842748, + "grad_norm": 0.5576605796813965, + "learning_rate": 0.00017498806103105477, + "loss": 0.1148, + "step": 9824 + }, + { + "epoch": 0.2303134724590798, + "grad_norm": 0.5347234010696411, + "learning_rate": 0.0001749831875751668, + "loss": 0.1263, + "step": 9825 + }, + { + "epoch": 0.2303369140338848, + "grad_norm": 0.7622226476669312, + "learning_rate": 0.00017497831371241797, + "loss": 0.7375, + "step": 9826 + }, + { + "epoch": 0.23036035560868978, + "grad_norm": 0.3676221966743469, + "learning_rate": 0.00017497343944283472, + "loss": 0.0702, + "step": 9827 + }, + { + "epoch": 0.23038379718349478, + "grad_norm": 0.3668425679206848, + "learning_rate": 0.00017496856476644345, + "loss": 0.0532, + "step": 9828 + }, + { + "epoch": 0.23040723875829977, + "grad_norm": 0.26714247465133667, + "learning_rate": 0.00017496368968327066, + "loss": 0.0346, + "step": 9829 + }, + { + "epoch": 0.23043068033310476, + "grad_norm": 0.47808173298835754, + "learning_rate": 0.00017495881419334278, + "loss": 0.1474, + "step": 9830 + }, + { + "epoch": 0.2304541219079098, + "grad_norm": 0.6498208045959473, + "learning_rate": 0.00017495393829668624, + "loss": 0.0897, + "step": 9831 + }, + { + "epoch": 0.23047756348271478, + "grad_norm": 0.7452696561813354, + "learning_rate": 0.00017494906199332755, + "loss": 0.6544, + "step": 9832 + }, + { + "epoch": 0.23050100505751978, + "grad_norm": 0.37157705426216125, + "learning_rate": 0.00017494418528329314, + "loss": 0.0736, + "step": 9833 + }, + { + "epoch": 0.23052444663232477, + "grad_norm": 0.6506941914558411, + "learning_rate": 0.00017493930816660946, + "loss": 0.5502, + "step": 9834 + }, + { + "epoch": 0.23054788820712976, + "grad_norm": 0.35321277379989624, + "learning_rate": 0.000174934430643303, + "loss": 0.1092, + "step": 9835 + }, + { + "epoch": 0.23057132978193476, + "grad_norm": 0.6789724230766296, + "learning_rate": 0.0001749295527134002, + "loss": 0.1395, + "step": 9836 + }, + { + "epoch": 0.23059477135673975, + "grad_norm": 0.39725691080093384, + "learning_rate": 0.00017492467437692756, + "loss": 0.0916, + "step": 9837 + }, + { + "epoch": 0.23061821293154475, + "grad_norm": 0.3730846345424652, + "learning_rate": 0.0001749197956339115, + "loss": 0.0558, + "step": 9838 + }, + { + "epoch": 0.23064165450634974, + "grad_norm": 0.6384469270706177, + "learning_rate": 0.00017491491648437854, + "loss": 0.176, + "step": 9839 + }, + { + "epoch": 0.23066509608115474, + "grad_norm": 0.31619027256965637, + "learning_rate": 0.00017491003692835517, + "loss": 0.0743, + "step": 9840 + }, + { + "epoch": 0.23068853765595973, + "grad_norm": 0.4176599979400635, + "learning_rate": 0.0001749051569658678, + "loss": 0.094, + "step": 9841 + }, + { + "epoch": 0.23071197923076472, + "grad_norm": 0.6423042416572571, + "learning_rate": 0.00017490027659694294, + "loss": 0.7313, + "step": 9842 + }, + { + "epoch": 0.23073542080556972, + "grad_norm": 0.5971770286560059, + "learning_rate": 0.00017489539582160707, + "loss": 0.1423, + "step": 9843 + }, + { + "epoch": 0.2307588623803747, + "grad_norm": 0.599929928779602, + "learning_rate": 0.00017489051463988667, + "loss": 0.1708, + "step": 9844 + }, + { + "epoch": 0.2307823039551797, + "grad_norm": 0.25532886385917664, + "learning_rate": 0.00017488563305180827, + "loss": 0.052, + "step": 9845 + }, + { + "epoch": 0.2308057455299847, + "grad_norm": 0.3639717400074005, + "learning_rate": 0.0001748807510573983, + "loss": 0.0751, + "step": 9846 + }, + { + "epoch": 0.2308291871047897, + "grad_norm": 0.749863862991333, + "learning_rate": 0.00017487586865668324, + "loss": 0.1377, + "step": 9847 + }, + { + "epoch": 0.2308526286795947, + "grad_norm": 0.7583582401275635, + "learning_rate": 0.00017487098584968963, + "loss": 0.1429, + "step": 9848 + }, + { + "epoch": 0.23087607025439968, + "grad_norm": 0.19735203683376312, + "learning_rate": 0.00017486610263644397, + "loss": 0.0612, + "step": 9849 + }, + { + "epoch": 0.23089951182920468, + "grad_norm": 0.35509002208709717, + "learning_rate": 0.00017486121901697272, + "loss": 0.0841, + "step": 9850 + }, + { + "epoch": 0.23092295340400967, + "grad_norm": 0.42357319593429565, + "learning_rate": 0.0001748563349913024, + "loss": 0.1421, + "step": 9851 + }, + { + "epoch": 0.23094639497881467, + "grad_norm": 0.40122419595718384, + "learning_rate": 0.00017485145055945944, + "loss": 0.0631, + "step": 9852 + }, + { + "epoch": 0.23096983655361966, + "grad_norm": 0.4783705770969391, + "learning_rate": 0.00017484656572147047, + "loss": 0.0969, + "step": 9853 + }, + { + "epoch": 0.23099327812842466, + "grad_norm": 0.5752087235450745, + "learning_rate": 0.00017484168047736192, + "loss": 0.1664, + "step": 9854 + }, + { + "epoch": 0.23101671970322965, + "grad_norm": 0.8634099960327148, + "learning_rate": 0.00017483679482716027, + "loss": 0.1098, + "step": 9855 + }, + { + "epoch": 0.23104016127803464, + "grad_norm": 0.2008368968963623, + "learning_rate": 0.0001748319087708921, + "loss": 0.0241, + "step": 9856 + }, + { + "epoch": 0.23106360285283967, + "grad_norm": 0.7911242246627808, + "learning_rate": 0.0001748270223085839, + "loss": 0.1577, + "step": 9857 + }, + { + "epoch": 0.23108704442764466, + "grad_norm": 0.14765292406082153, + "learning_rate": 0.00017482213544026216, + "loss": 0.0261, + "step": 9858 + }, + { + "epoch": 0.23111048600244966, + "grad_norm": 0.29254668951034546, + "learning_rate": 0.0001748172481659534, + "loss": 0.0401, + "step": 9859 + }, + { + "epoch": 0.23113392757725465, + "grad_norm": 0.8062102794647217, + "learning_rate": 0.00017481236048568413, + "loss": 0.1107, + "step": 9860 + }, + { + "epoch": 0.23115736915205964, + "grad_norm": 0.35180652141571045, + "learning_rate": 0.0001748074723994809, + "loss": 0.114, + "step": 9861 + }, + { + "epoch": 0.23118081072686464, + "grad_norm": 0.6152297258377075, + "learning_rate": 0.0001748025839073703, + "loss": 0.1088, + "step": 9862 + }, + { + "epoch": 0.23120425230166963, + "grad_norm": 0.2752155065536499, + "learning_rate": 0.00017479769500937866, + "loss": 0.0249, + "step": 9863 + }, + { + "epoch": 0.23122769387647463, + "grad_norm": 0.4573328495025635, + "learning_rate": 0.00017479280570553268, + "loss": 0.0769, + "step": 9864 + }, + { + "epoch": 0.23125113545127962, + "grad_norm": 0.7412363290786743, + "learning_rate": 0.00017478791599585884, + "loss": 0.1538, + "step": 9865 + }, + { + "epoch": 0.23127457702608462, + "grad_norm": 0.1566327065229416, + "learning_rate": 0.00017478302588038363, + "loss": 0.0308, + "step": 9866 + }, + { + "epoch": 0.2312980186008896, + "grad_norm": 0.41631197929382324, + "learning_rate": 0.00017477813535913365, + "loss": 0.0864, + "step": 9867 + }, + { + "epoch": 0.2313214601756946, + "grad_norm": 0.332247257232666, + "learning_rate": 0.0001747732444321354, + "loss": 0.0921, + "step": 9868 + }, + { + "epoch": 0.2313449017504996, + "grad_norm": 0.3195398449897766, + "learning_rate": 0.0001747683530994154, + "loss": 0.0418, + "step": 9869 + }, + { + "epoch": 0.2313683433253046, + "grad_norm": 0.19244763255119324, + "learning_rate": 0.00017476346136100025, + "loss": 0.0359, + "step": 9870 + }, + { + "epoch": 0.2313917849001096, + "grad_norm": 0.5766055583953857, + "learning_rate": 0.00017475856921691642, + "loss": 0.1247, + "step": 9871 + }, + { + "epoch": 0.23141522647491458, + "grad_norm": 0.45784297585487366, + "learning_rate": 0.00017475367666719053, + "loss": 0.1391, + "step": 9872 + }, + { + "epoch": 0.23143866804971958, + "grad_norm": 0.32424405217170715, + "learning_rate": 0.00017474878371184907, + "loss": 0.0981, + "step": 9873 + }, + { + "epoch": 0.23146210962452457, + "grad_norm": 0.45578598976135254, + "learning_rate": 0.0001747438903509186, + "loss": 0.1253, + "step": 9874 + }, + { + "epoch": 0.23148555119932956, + "grad_norm": 0.4389733672142029, + "learning_rate": 0.00017473899658442568, + "loss": 0.0773, + "step": 9875 + }, + { + "epoch": 0.23150899277413456, + "grad_norm": 0.5937752723693848, + "learning_rate": 0.00017473410241239688, + "loss": 0.8187, + "step": 9876 + }, + { + "epoch": 0.23153243434893955, + "grad_norm": 0.8557516932487488, + "learning_rate": 0.0001747292078348587, + "loss": 0.1491, + "step": 9877 + }, + { + "epoch": 0.23155587592374455, + "grad_norm": 0.5423198342323303, + "learning_rate": 0.0001747243128518378, + "loss": 0.0901, + "step": 9878 + }, + { + "epoch": 0.23157931749854954, + "grad_norm": 0.4685702919960022, + "learning_rate": 0.0001747194174633606, + "loss": 0.1628, + "step": 9879 + }, + { + "epoch": 0.23160275907335454, + "grad_norm": 0.1033565029501915, + "learning_rate": 0.0001747145216694538, + "loss": 0.0113, + "step": 9880 + }, + { + "epoch": 0.23162620064815953, + "grad_norm": 0.954095184803009, + "learning_rate": 0.00017470962547014388, + "loss": 0.2048, + "step": 9881 + }, + { + "epoch": 0.23164964222296455, + "grad_norm": 0.2622700333595276, + "learning_rate": 0.00017470472886545744, + "loss": 0.0489, + "step": 9882 + }, + { + "epoch": 0.23167308379776955, + "grad_norm": 0.4592219591140747, + "learning_rate": 0.00017469983185542108, + "loss": 0.0667, + "step": 9883 + }, + { + "epoch": 0.23169652537257454, + "grad_norm": 0.6742193102836609, + "learning_rate": 0.00017469493444006125, + "loss": 0.1658, + "step": 9884 + }, + { + "epoch": 0.23171996694737954, + "grad_norm": 0.517116129398346, + "learning_rate": 0.00017469003661940463, + "loss": 0.04, + "step": 9885 + }, + { + "epoch": 0.23174340852218453, + "grad_norm": 0.518973708152771, + "learning_rate": 0.00017468513839347783, + "loss": 0.131, + "step": 9886 + }, + { + "epoch": 0.23176685009698952, + "grad_norm": 0.9123970866203308, + "learning_rate": 0.00017468023976230729, + "loss": 0.1721, + "step": 9887 + }, + { + "epoch": 0.23179029167179452, + "grad_norm": 0.33586961030960083, + "learning_rate": 0.0001746753407259197, + "loss": 0.0558, + "step": 9888 + }, + { + "epoch": 0.2318137332465995, + "grad_norm": 0.12756885588169098, + "learning_rate": 0.00017467044128434165, + "loss": 0.0393, + "step": 9889 + }, + { + "epoch": 0.2318371748214045, + "grad_norm": 0.3688812255859375, + "learning_rate": 0.00017466554143759963, + "loss": 0.0815, + "step": 9890 + }, + { + "epoch": 0.2318606163962095, + "grad_norm": 0.877680778503418, + "learning_rate": 0.00017466064118572033, + "loss": 0.3031, + "step": 9891 + }, + { + "epoch": 0.2318840579710145, + "grad_norm": 0.770320475101471, + "learning_rate": 0.00017465574052873026, + "loss": 0.1671, + "step": 9892 + }, + { + "epoch": 0.2319074995458195, + "grad_norm": 0.3350052237510681, + "learning_rate": 0.00017465083946665604, + "loss": 0.0603, + "step": 9893 + }, + { + "epoch": 0.23193094112062448, + "grad_norm": 0.6190816760063171, + "learning_rate": 0.0001746459379995243, + "loss": 0.1024, + "step": 9894 + }, + { + "epoch": 0.23195438269542948, + "grad_norm": 0.635679304599762, + "learning_rate": 0.00017464103612736156, + "loss": 0.1158, + "step": 9895 + }, + { + "epoch": 0.23197782427023447, + "grad_norm": 0.11799385398626328, + "learning_rate": 0.00017463613385019448, + "loss": 0.0244, + "step": 9896 + }, + { + "epoch": 0.23200126584503947, + "grad_norm": 0.664746880531311, + "learning_rate": 0.00017463123116804964, + "loss": 0.1826, + "step": 9897 + }, + { + "epoch": 0.23202470741984446, + "grad_norm": 0.33889949321746826, + "learning_rate": 0.00017462632808095364, + "loss": 0.1075, + "step": 9898 + }, + { + "epoch": 0.23204814899464946, + "grad_norm": 0.324491947889328, + "learning_rate": 0.0001746214245889331, + "loss": 0.0772, + "step": 9899 + }, + { + "epoch": 0.23207159056945445, + "grad_norm": 0.6742551922798157, + "learning_rate": 0.0001746165206920146, + "loss": 0.1756, + "step": 9900 + }, + { + "epoch": 0.23209503214425944, + "grad_norm": 0.5571437478065491, + "learning_rate": 0.00017461161639022475, + "loss": 0.1287, + "step": 9901 + }, + { + "epoch": 0.23211847371906444, + "grad_norm": 0.4235038757324219, + "learning_rate": 0.00017460671168359015, + "loss": 0.0866, + "step": 9902 + }, + { + "epoch": 0.23214191529386943, + "grad_norm": 0.16620220243930817, + "learning_rate": 0.00017460180657213748, + "loss": 0.0224, + "step": 9903 + }, + { + "epoch": 0.23216535686867443, + "grad_norm": 0.7823532819747925, + "learning_rate": 0.0001745969010558933, + "loss": 0.1242, + "step": 9904 + }, + { + "epoch": 0.23218879844347942, + "grad_norm": 0.7900081872940063, + "learning_rate": 0.00017459199513488423, + "loss": 0.1473, + "step": 9905 + }, + { + "epoch": 0.23221224001828442, + "grad_norm": 0.1747620403766632, + "learning_rate": 0.00017458708880913688, + "loss": 0.0322, + "step": 9906 + }, + { + "epoch": 0.2322356815930894, + "grad_norm": 0.1997542530298233, + "learning_rate": 0.0001745821820786779, + "loss": 0.0265, + "step": 9907 + }, + { + "epoch": 0.23225912316789443, + "grad_norm": 0.8206465244293213, + "learning_rate": 0.00017457727494353393, + "loss": 0.1285, + "step": 9908 + }, + { + "epoch": 0.23228256474269943, + "grad_norm": 0.5455538630485535, + "learning_rate": 0.00017457236740373152, + "loss": 0.1432, + "step": 9909 + }, + { + "epoch": 0.23230600631750442, + "grad_norm": 0.8133078217506409, + "learning_rate": 0.0001745674594592974, + "loss": 0.1151, + "step": 9910 + }, + { + "epoch": 0.23232944789230942, + "grad_norm": 0.7910841703414917, + "learning_rate": 0.0001745625511102581, + "loss": 0.1215, + "step": 9911 + }, + { + "epoch": 0.2323528894671144, + "grad_norm": 0.5211089253425598, + "learning_rate": 0.00017455764235664033, + "loss": 0.1487, + "step": 9912 + }, + { + "epoch": 0.2323763310419194, + "grad_norm": 0.8027048110961914, + "learning_rate": 0.0001745527331984707, + "loss": 0.5956, + "step": 9913 + }, + { + "epoch": 0.2323997726167244, + "grad_norm": 0.859510600566864, + "learning_rate": 0.00017454782363577581, + "loss": 0.1117, + "step": 9914 + }, + { + "epoch": 0.2324232141915294, + "grad_norm": 0.9182872772216797, + "learning_rate": 0.00017454291366858238, + "loss": 0.1621, + "step": 9915 + }, + { + "epoch": 0.2324466557663344, + "grad_norm": 0.7439157366752625, + "learning_rate": 0.00017453800329691698, + "loss": 0.1719, + "step": 9916 + }, + { + "epoch": 0.23247009734113938, + "grad_norm": 0.5824978351593018, + "learning_rate": 0.00017453309252080627, + "loss": 0.1564, + "step": 9917 + }, + { + "epoch": 0.23249353891594438, + "grad_norm": 0.3055647909641266, + "learning_rate": 0.00017452818134027694, + "loss": 0.074, + "step": 9918 + }, + { + "epoch": 0.23251698049074937, + "grad_norm": 0.47838929295539856, + "learning_rate": 0.00017452326975535555, + "loss": 0.0672, + "step": 9919 + }, + { + "epoch": 0.23254042206555436, + "grad_norm": 0.653824508190155, + "learning_rate": 0.00017451835776606883, + "loss": 0.2164, + "step": 9920 + }, + { + "epoch": 0.23256386364035936, + "grad_norm": 0.8688134551048279, + "learning_rate": 0.00017451344537244343, + "loss": 0.1294, + "step": 9921 + }, + { + "epoch": 0.23258730521516435, + "grad_norm": 0.44116896390914917, + "learning_rate": 0.00017450853257450596, + "loss": 0.1053, + "step": 9922 + }, + { + "epoch": 0.23261074678996935, + "grad_norm": 0.2639327347278595, + "learning_rate": 0.0001745036193722831, + "loss": 0.0678, + "step": 9923 + }, + { + "epoch": 0.23263418836477434, + "grad_norm": 0.38529762625694275, + "learning_rate": 0.0001744987057658015, + "loss": 0.0815, + "step": 9924 + }, + { + "epoch": 0.23265762993957934, + "grad_norm": 0.36066219210624695, + "learning_rate": 0.00017449379175508783, + "loss": 0.0516, + "step": 9925 + }, + { + "epoch": 0.23268107151438433, + "grad_norm": 0.32539790868759155, + "learning_rate": 0.0001744888773401687, + "loss": 0.0643, + "step": 9926 + }, + { + "epoch": 0.23270451308918932, + "grad_norm": 0.6494882702827454, + "learning_rate": 0.00017448396252107092, + "loss": 0.1101, + "step": 9927 + }, + { + "epoch": 0.23272795466399432, + "grad_norm": 0.3911370038986206, + "learning_rate": 0.000174479047297821, + "loss": 0.1155, + "step": 9928 + }, + { + "epoch": 0.2327513962387993, + "grad_norm": 0.4276430010795593, + "learning_rate": 0.00017447413167044572, + "loss": 0.0809, + "step": 9929 + }, + { + "epoch": 0.2327748378136043, + "grad_norm": 0.2648073434829712, + "learning_rate": 0.00017446921563897164, + "loss": 0.0542, + "step": 9930 + }, + { + "epoch": 0.2327982793884093, + "grad_norm": 0.6795045137405396, + "learning_rate": 0.0001744642992034256, + "loss": 0.1478, + "step": 9931 + }, + { + "epoch": 0.2328217209632143, + "grad_norm": 0.7153986692428589, + "learning_rate": 0.0001744593823638341, + "loss": 0.1699, + "step": 9932 + }, + { + "epoch": 0.23284516253801932, + "grad_norm": 0.1718771606683731, + "learning_rate": 0.00017445446512022393, + "loss": 0.0388, + "step": 9933 + }, + { + "epoch": 0.2328686041128243, + "grad_norm": 0.8925146460533142, + "learning_rate": 0.00017444954747262173, + "loss": 0.1427, + "step": 9934 + }, + { + "epoch": 0.2328920456876293, + "grad_norm": 0.7147164344787598, + "learning_rate": 0.00017444462942105418, + "loss": 0.8914, + "step": 9935 + }, + { + "epoch": 0.2329154872624343, + "grad_norm": 0.5033244490623474, + "learning_rate": 0.000174439710965548, + "loss": 0.1234, + "step": 9936 + }, + { + "epoch": 0.2329389288372393, + "grad_norm": 0.7474193572998047, + "learning_rate": 0.00017443479210612982, + "loss": 0.1332, + "step": 9937 + }, + { + "epoch": 0.2329623704120443, + "grad_norm": 0.6125195622444153, + "learning_rate": 0.00017442987284282642, + "loss": 0.6106, + "step": 9938 + }, + { + "epoch": 0.23298581198684928, + "grad_norm": 0.26031193137168884, + "learning_rate": 0.0001744249531756644, + "loss": 0.0779, + "step": 9939 + }, + { + "epoch": 0.23300925356165428, + "grad_norm": 0.7913921475410461, + "learning_rate": 0.0001744200331046705, + "loss": 0.1452, + "step": 9940 + }, + { + "epoch": 0.23303269513645927, + "grad_norm": 0.607687771320343, + "learning_rate": 0.00017441511262987142, + "loss": 0.7721, + "step": 9941 + }, + { + "epoch": 0.23305613671126427, + "grad_norm": 0.8287720680236816, + "learning_rate": 0.00017441019175129383, + "loss": 0.1755, + "step": 9942 + }, + { + "epoch": 0.23307957828606926, + "grad_norm": 0.357349693775177, + "learning_rate": 0.00017440527046896445, + "loss": 0.0778, + "step": 9943 + }, + { + "epoch": 0.23310301986087426, + "grad_norm": 0.18451261520385742, + "learning_rate": 0.00017440034878290995, + "loss": 0.0349, + "step": 9944 + }, + { + "epoch": 0.23312646143567925, + "grad_norm": 0.28723764419555664, + "learning_rate": 0.0001743954266931571, + "loss": 0.0692, + "step": 9945 + }, + { + "epoch": 0.23314990301048424, + "grad_norm": 0.5956594944000244, + "learning_rate": 0.0001743905041997326, + "loss": 0.1289, + "step": 9946 + }, + { + "epoch": 0.23317334458528924, + "grad_norm": 0.8244083523750305, + "learning_rate": 0.0001743855813026631, + "loss": 0.1378, + "step": 9947 + }, + { + "epoch": 0.23319678616009423, + "grad_norm": 0.6383642554283142, + "learning_rate": 0.00017438065800197533, + "loss": 0.1169, + "step": 9948 + }, + { + "epoch": 0.23322022773489923, + "grad_norm": 0.48731130361557007, + "learning_rate": 0.00017437573429769602, + "loss": 0.0739, + "step": 9949 + }, + { + "epoch": 0.23324366930970422, + "grad_norm": 0.3911266624927521, + "learning_rate": 0.00017437081018985192, + "loss": 0.6987, + "step": 9950 + }, + { + "epoch": 0.23326711088450922, + "grad_norm": 0.9485647678375244, + "learning_rate": 0.0001743658856784697, + "loss": 0.1686, + "step": 9951 + }, + { + "epoch": 0.2332905524593142, + "grad_norm": 0.4620005786418915, + "learning_rate": 0.00017436096076357607, + "loss": 0.1026, + "step": 9952 + }, + { + "epoch": 0.2333139940341192, + "grad_norm": 0.8147702813148499, + "learning_rate": 0.0001743560354451978, + "loss": 0.1445, + "step": 9953 + }, + { + "epoch": 0.2333374356089242, + "grad_norm": 0.5581479668617249, + "learning_rate": 0.00017435110972336158, + "loss": 0.115, + "step": 9954 + }, + { + "epoch": 0.2333608771837292, + "grad_norm": 0.5915439128875732, + "learning_rate": 0.00017434618359809414, + "loss": 0.1255, + "step": 9955 + }, + { + "epoch": 0.2333843187585342, + "grad_norm": 0.6439775228500366, + "learning_rate": 0.00017434125706942223, + "loss": 0.5279, + "step": 9956 + }, + { + "epoch": 0.23340776033333918, + "grad_norm": 0.33209800720214844, + "learning_rate": 0.00017433633013737255, + "loss": 0.0636, + "step": 9957 + }, + { + "epoch": 0.23343120190814418, + "grad_norm": 0.7902156114578247, + "learning_rate": 0.00017433140280197188, + "loss": 0.1691, + "step": 9958 + }, + { + "epoch": 0.2334546434829492, + "grad_norm": 0.5790836215019226, + "learning_rate": 0.00017432647506324694, + "loss": 0.1157, + "step": 9959 + }, + { + "epoch": 0.2334780850577542, + "grad_norm": 0.4148259162902832, + "learning_rate": 0.00017432154692122441, + "loss": 0.0796, + "step": 9960 + }, + { + "epoch": 0.2335015266325592, + "grad_norm": 0.6583589315414429, + "learning_rate": 0.00017431661837593112, + "loss": 0.6286, + "step": 9961 + }, + { + "epoch": 0.23352496820736418, + "grad_norm": 0.40855613350868225, + "learning_rate": 0.00017431168942739375, + "loss": 0.0687, + "step": 9962 + }, + { + "epoch": 0.23354840978216918, + "grad_norm": 0.5386926531791687, + "learning_rate": 0.00017430676007563906, + "loss": 0.0618, + "step": 9963 + }, + { + "epoch": 0.23357185135697417, + "grad_norm": 0.5073639750480652, + "learning_rate": 0.00017430183032069382, + "loss": 0.1272, + "step": 9964 + }, + { + "epoch": 0.23359529293177916, + "grad_norm": 0.6195089817047119, + "learning_rate": 0.00017429690016258474, + "loss": 0.6962, + "step": 9965 + }, + { + "epoch": 0.23361873450658416, + "grad_norm": 0.24739079177379608, + "learning_rate": 0.00017429196960133857, + "loss": 0.0588, + "step": 9966 + }, + { + "epoch": 0.23364217608138915, + "grad_norm": 0.4510112702846527, + "learning_rate": 0.00017428703863698212, + "loss": 0.1283, + "step": 9967 + }, + { + "epoch": 0.23366561765619415, + "grad_norm": 0.20108221471309662, + "learning_rate": 0.0001742821072695421, + "loss": 0.0298, + "step": 9968 + }, + { + "epoch": 0.23368905923099914, + "grad_norm": 0.2500898838043213, + "learning_rate": 0.00017427717549904528, + "loss": 0.0232, + "step": 9969 + }, + { + "epoch": 0.23371250080580414, + "grad_norm": 0.49574220180511475, + "learning_rate": 0.00017427224332551843, + "loss": 0.1117, + "step": 9970 + }, + { + "epoch": 0.23373594238060913, + "grad_norm": 0.7291527986526489, + "learning_rate": 0.00017426731074898829, + "loss": 0.1554, + "step": 9971 + }, + { + "epoch": 0.23375938395541412, + "grad_norm": 0.6787426471710205, + "learning_rate": 0.0001742623777694816, + "loss": 0.1825, + "step": 9972 + }, + { + "epoch": 0.23378282553021912, + "grad_norm": 0.24935109913349152, + "learning_rate": 0.00017425744438702518, + "loss": 0.0402, + "step": 9973 + }, + { + "epoch": 0.2338062671050241, + "grad_norm": 0.6732781529426575, + "learning_rate": 0.0001742525106016458, + "loss": 0.1787, + "step": 9974 + }, + { + "epoch": 0.2338297086798291, + "grad_norm": 0.7260283827781677, + "learning_rate": 0.0001742475764133702, + "loss": 0.1073, + "step": 9975 + }, + { + "epoch": 0.2338531502546341, + "grad_norm": 0.8440982103347778, + "learning_rate": 0.00017424264182222513, + "loss": 0.1225, + "step": 9976 + }, + { + "epoch": 0.2338765918294391, + "grad_norm": 0.3387334942817688, + "learning_rate": 0.00017423770682823743, + "loss": 0.089, + "step": 9977 + }, + { + "epoch": 0.2339000334042441, + "grad_norm": 0.6856550574302673, + "learning_rate": 0.00017423277143143384, + "loss": 0.0745, + "step": 9978 + }, + { + "epoch": 0.23392347497904908, + "grad_norm": 0.2599451541900635, + "learning_rate": 0.00017422783563184115, + "loss": 0.0693, + "step": 9979 + }, + { + "epoch": 0.23394691655385408, + "grad_norm": 0.6015354990959167, + "learning_rate": 0.00017422289942948616, + "loss": 0.1141, + "step": 9980 + }, + { + "epoch": 0.23397035812865907, + "grad_norm": 0.17835596203804016, + "learning_rate": 0.00017421796282439557, + "loss": 0.033, + "step": 9981 + }, + { + "epoch": 0.23399379970346407, + "grad_norm": 0.5880522131919861, + "learning_rate": 0.00017421302581659628, + "loss": 0.1311, + "step": 9982 + }, + { + "epoch": 0.23401724127826906, + "grad_norm": 0.936578631401062, + "learning_rate": 0.00017420808840611497, + "loss": 0.3086, + "step": 9983 + }, + { + "epoch": 0.23404068285307406, + "grad_norm": 0.5981115698814392, + "learning_rate": 0.00017420315059297854, + "loss": 0.1433, + "step": 9984 + }, + { + "epoch": 0.23406412442787908, + "grad_norm": 0.4425642192363739, + "learning_rate": 0.00017419821237721372, + "loss": 0.0936, + "step": 9985 + }, + { + "epoch": 0.23408756600268407, + "grad_norm": 0.380551278591156, + "learning_rate": 0.00017419327375884727, + "loss": 0.0872, + "step": 9986 + }, + { + "epoch": 0.23411100757748907, + "grad_norm": 0.4085692763328552, + "learning_rate": 0.00017418833473790607, + "loss": 0.0911, + "step": 9987 + }, + { + "epoch": 0.23413444915229406, + "grad_norm": 0.5555424094200134, + "learning_rate": 0.00017418339531441687, + "loss": 0.112, + "step": 9988 + }, + { + "epoch": 0.23415789072709906, + "grad_norm": 0.726901113986969, + "learning_rate": 0.0001741784554884065, + "loss": 0.0938, + "step": 9989 + }, + { + "epoch": 0.23418133230190405, + "grad_norm": 0.8051747679710388, + "learning_rate": 0.00017417351525990174, + "loss": 0.0773, + "step": 9990 + }, + { + "epoch": 0.23420477387670904, + "grad_norm": 0.5897985100746155, + "learning_rate": 0.0001741685746289294, + "loss": 0.1575, + "step": 9991 + }, + { + "epoch": 0.23422821545151404, + "grad_norm": 0.4683057963848114, + "learning_rate": 0.00017416363359551627, + "loss": 0.4585, + "step": 9992 + }, + { + "epoch": 0.23425165702631903, + "grad_norm": 0.5292109251022339, + "learning_rate": 0.00017415869215968915, + "loss": 0.7592, + "step": 9993 + }, + { + "epoch": 0.23427509860112403, + "grad_norm": 0.9906411170959473, + "learning_rate": 0.00017415375032147495, + "loss": 0.2207, + "step": 9994 + }, + { + "epoch": 0.23429854017592902, + "grad_norm": 0.6125018000602722, + "learning_rate": 0.00017414880808090037, + "loss": 0.1333, + "step": 9995 + }, + { + "epoch": 0.23432198175073402, + "grad_norm": 0.2135237604379654, + "learning_rate": 0.00017414386543799227, + "loss": 0.0452, + "step": 9996 + }, + { + "epoch": 0.234345423325539, + "grad_norm": 0.4313287138938904, + "learning_rate": 0.00017413892239277748, + "loss": 0.1521, + "step": 9997 + }, + { + "epoch": 0.234368864900344, + "grad_norm": 0.430908739566803, + "learning_rate": 0.00017413397894528284, + "loss": 0.0868, + "step": 9998 + }, + { + "epoch": 0.234392306475149, + "grad_norm": 0.6231163144111633, + "learning_rate": 0.00017412903509553512, + "loss": 0.0622, + "step": 9999 + }, + { + "epoch": 0.234415748049954, + "grad_norm": 0.5437332391738892, + "learning_rate": 0.00017412409084356122, + "loss": 0.107, + "step": 10000 + }, + { + "epoch": 0.234439189624759, + "grad_norm": 0.1402120590209961, + "learning_rate": 0.00017411914618938786, + "loss": 0.04, + "step": 10001 + }, + { + "epoch": 0.23446263119956398, + "grad_norm": 0.6037971377372742, + "learning_rate": 0.00017411420113304192, + "loss": 0.1758, + "step": 10002 + }, + { + "epoch": 0.23448607277436898, + "grad_norm": 0.4461596608161926, + "learning_rate": 0.00017410925567455028, + "loss": 0.0668, + "step": 10003 + }, + { + "epoch": 0.23450951434917397, + "grad_norm": 0.3912249505519867, + "learning_rate": 0.0001741043098139397, + "loss": 0.2308, + "step": 10004 + }, + { + "epoch": 0.23453295592397896, + "grad_norm": 0.5924336314201355, + "learning_rate": 0.0001740993635512371, + "loss": 0.1654, + "step": 10005 + }, + { + "epoch": 0.23455639749878396, + "grad_norm": 0.5121138095855713, + "learning_rate": 0.00017409441688646924, + "loss": 0.1969, + "step": 10006 + }, + { + "epoch": 0.23457983907358895, + "grad_norm": 0.4407414495944977, + "learning_rate": 0.00017408946981966298, + "loss": 0.1476, + "step": 10007 + }, + { + "epoch": 0.23460328064839395, + "grad_norm": 0.5748093724250793, + "learning_rate": 0.0001740845223508452, + "loss": 0.5751, + "step": 10008 + }, + { + "epoch": 0.23462672222319894, + "grad_norm": 0.16962620615959167, + "learning_rate": 0.0001740795744800427, + "loss": 0.0314, + "step": 10009 + }, + { + "epoch": 0.23465016379800396, + "grad_norm": 0.324923038482666, + "learning_rate": 0.00017407462620728235, + "loss": 0.0876, + "step": 10010 + }, + { + "epoch": 0.23467360537280896, + "grad_norm": 0.5283911824226379, + "learning_rate": 0.00017406967753259102, + "loss": 0.0959, + "step": 10011 + }, + { + "epoch": 0.23469704694761395, + "grad_norm": 0.4384599030017853, + "learning_rate": 0.00017406472845599548, + "loss": 0.114, + "step": 10012 + }, + { + "epoch": 0.23472048852241895, + "grad_norm": 0.46463742852211, + "learning_rate": 0.00017405977897752269, + "loss": 0.0828, + "step": 10013 + }, + { + "epoch": 0.23474393009722394, + "grad_norm": 0.2645675837993622, + "learning_rate": 0.0001740548290971994, + "loss": 0.0643, + "step": 10014 + }, + { + "epoch": 0.23476737167202893, + "grad_norm": 0.5662010312080383, + "learning_rate": 0.0001740498788150526, + "loss": 0.0699, + "step": 10015 + }, + { + "epoch": 0.23479081324683393, + "grad_norm": 0.301310658454895, + "learning_rate": 0.00017404492813110904, + "loss": 0.0477, + "step": 10016 + }, + { + "epoch": 0.23481425482163892, + "grad_norm": 0.6835817694664001, + "learning_rate": 0.00017403997704539562, + "loss": 0.1296, + "step": 10017 + }, + { + "epoch": 0.23483769639644392, + "grad_norm": 0.7079688310623169, + "learning_rate": 0.00017403502555793918, + "loss": 0.1252, + "step": 10018 + }, + { + "epoch": 0.2348611379712489, + "grad_norm": 0.6718003749847412, + "learning_rate": 0.00017403007366876666, + "loss": 0.1309, + "step": 10019 + }, + { + "epoch": 0.2348845795460539, + "grad_norm": 1.0552210807800293, + "learning_rate": 0.00017402512137790484, + "loss": 0.1249, + "step": 10020 + }, + { + "epoch": 0.2349080211208589, + "grad_norm": 0.2636600434780121, + "learning_rate": 0.00017402016868538062, + "loss": 0.0809, + "step": 10021 + }, + { + "epoch": 0.2349314626956639, + "grad_norm": 0.5317531824111938, + "learning_rate": 0.00017401521559122092, + "loss": 0.1317, + "step": 10022 + }, + { + "epoch": 0.2349549042704689, + "grad_norm": 0.5732740759849548, + "learning_rate": 0.00017401026209545254, + "loss": 0.6805, + "step": 10023 + }, + { + "epoch": 0.23497834584527388, + "grad_norm": 0.1776454895734787, + "learning_rate": 0.00017400530819810243, + "loss": 0.0469, + "step": 10024 + }, + { + "epoch": 0.23500178742007888, + "grad_norm": 0.8106613755226135, + "learning_rate": 0.00017400035389919744, + "loss": 0.0942, + "step": 10025 + }, + { + "epoch": 0.23502522899488387, + "grad_norm": 0.18944691121578217, + "learning_rate": 0.00017399539919876443, + "loss": 0.0291, + "step": 10026 + }, + { + "epoch": 0.23504867056968887, + "grad_norm": 0.2891485095024109, + "learning_rate": 0.0001739904440968303, + "loss": 0.0569, + "step": 10027 + }, + { + "epoch": 0.23507211214449386, + "grad_norm": 0.18392544984817505, + "learning_rate": 0.00017398548859342197, + "loss": 0.0392, + "step": 10028 + }, + { + "epoch": 0.23509555371929886, + "grad_norm": 0.5221707224845886, + "learning_rate": 0.00017398053268856626, + "loss": 0.1671, + "step": 10029 + }, + { + "epoch": 0.23511899529410385, + "grad_norm": 0.6847456693649292, + "learning_rate": 0.0001739755763822901, + "loss": 0.1712, + "step": 10030 + }, + { + "epoch": 0.23514243686890884, + "grad_norm": 0.30433914065361023, + "learning_rate": 0.00017397061967462042, + "loss": 0.0618, + "step": 10031 + }, + { + "epoch": 0.23516587844371384, + "grad_norm": 0.6198539733886719, + "learning_rate": 0.00017396566256558407, + "loss": 0.0964, + "step": 10032 + }, + { + "epoch": 0.23518932001851883, + "grad_norm": 0.3214361071586609, + "learning_rate": 0.00017396070505520795, + "loss": 0.0513, + "step": 10033 + }, + { + "epoch": 0.23521276159332383, + "grad_norm": 0.4690834879875183, + "learning_rate": 0.00017395574714351895, + "loss": 0.1204, + "step": 10034 + }, + { + "epoch": 0.23523620316812882, + "grad_norm": 0.14818356931209564, + "learning_rate": 0.000173950788830544, + "loss": 0.0464, + "step": 10035 + }, + { + "epoch": 0.23525964474293384, + "grad_norm": 0.6123438477516174, + "learning_rate": 0.00017394583011630995, + "loss": 0.1654, + "step": 10036 + }, + { + "epoch": 0.23528308631773884, + "grad_norm": 0.5757198333740234, + "learning_rate": 0.0001739408710008438, + "loss": 0.1674, + "step": 10037 + }, + { + "epoch": 0.23530652789254383, + "grad_norm": 0.14659175276756287, + "learning_rate": 0.00017393591148417238, + "loss": 0.0217, + "step": 10038 + }, + { + "epoch": 0.23532996946734883, + "grad_norm": 1.3370903730392456, + "learning_rate": 0.00017393095156632263, + "loss": 0.1064, + "step": 10039 + }, + { + "epoch": 0.23535341104215382, + "grad_norm": 0.6645917296409607, + "learning_rate": 0.00017392599124732146, + "loss": 0.1407, + "step": 10040 + }, + { + "epoch": 0.23537685261695881, + "grad_norm": 0.5296242237091064, + "learning_rate": 0.0001739210305271958, + "loss": 0.1061, + "step": 10041 + }, + { + "epoch": 0.2354002941917638, + "grad_norm": 0.5809794068336487, + "learning_rate": 0.00017391606940597254, + "loss": 0.0926, + "step": 10042 + }, + { + "epoch": 0.2354237357665688, + "grad_norm": 0.8242725729942322, + "learning_rate": 0.0001739111078836786, + "loss": 0.1334, + "step": 10043 + }, + { + "epoch": 0.2354471773413738, + "grad_norm": 0.5517762303352356, + "learning_rate": 0.00017390614596034088, + "loss": 0.13, + "step": 10044 + }, + { + "epoch": 0.2354706189161788, + "grad_norm": 0.6413998007774353, + "learning_rate": 0.00017390118363598637, + "loss": 0.1197, + "step": 10045 + }, + { + "epoch": 0.2354940604909838, + "grad_norm": 0.13005876541137695, + "learning_rate": 0.00017389622091064198, + "loss": 0.0461, + "step": 10046 + }, + { + "epoch": 0.23551750206578878, + "grad_norm": 0.506151556968689, + "learning_rate": 0.00017389125778433456, + "loss": 0.112, + "step": 10047 + }, + { + "epoch": 0.23554094364059378, + "grad_norm": 0.8565847873687744, + "learning_rate": 0.00017388629425709113, + "loss": 0.1891, + "step": 10048 + }, + { + "epoch": 0.23556438521539877, + "grad_norm": 0.15677808225154877, + "learning_rate": 0.0001738813303289386, + "loss": 0.0287, + "step": 10049 + }, + { + "epoch": 0.23558782679020376, + "grad_norm": 0.26808542013168335, + "learning_rate": 0.0001738763659999039, + "loss": 0.0405, + "step": 10050 + }, + { + "epoch": 0.23561126836500876, + "grad_norm": 0.5985084176063538, + "learning_rate": 0.00017387140127001394, + "loss": 0.1503, + "step": 10051 + }, + { + "epoch": 0.23563470993981375, + "grad_norm": 0.7959159016609192, + "learning_rate": 0.00017386643613929566, + "loss": 0.1355, + "step": 10052 + }, + { + "epoch": 0.23565815151461875, + "grad_norm": 0.52142333984375, + "learning_rate": 0.00017386147060777606, + "loss": 0.1077, + "step": 10053 + }, + { + "epoch": 0.23568159308942374, + "grad_norm": 0.9380134344100952, + "learning_rate": 0.00017385650467548205, + "loss": 0.2572, + "step": 10054 + }, + { + "epoch": 0.23570503466422874, + "grad_norm": 0.16394869983196259, + "learning_rate": 0.00017385153834244052, + "loss": 0.0216, + "step": 10055 + }, + { + "epoch": 0.23572847623903373, + "grad_norm": 0.5677983164787292, + "learning_rate": 0.0001738465716086785, + "loss": 0.1158, + "step": 10056 + }, + { + "epoch": 0.23575191781383872, + "grad_norm": 0.3649616837501526, + "learning_rate": 0.00017384160447422292, + "loss": 0.0538, + "step": 10057 + }, + { + "epoch": 0.23577535938864372, + "grad_norm": 0.4789288640022278, + "learning_rate": 0.00017383663693910069, + "loss": 0.1213, + "step": 10058 + }, + { + "epoch": 0.2357988009634487, + "grad_norm": 0.6092037558555603, + "learning_rate": 0.0001738316690033388, + "loss": 0.1081, + "step": 10059 + }, + { + "epoch": 0.2358222425382537, + "grad_norm": 0.6458043456077576, + "learning_rate": 0.00017382670066696417, + "loss": 0.2074, + "step": 10060 + }, + { + "epoch": 0.23584568411305873, + "grad_norm": 0.5501415133476257, + "learning_rate": 0.00017382173193000383, + "loss": 0.0872, + "step": 10061 + }, + { + "epoch": 0.23586912568786372, + "grad_norm": 0.4572150409221649, + "learning_rate": 0.0001738167627924847, + "loss": 0.0626, + "step": 10062 + }, + { + "epoch": 0.23589256726266872, + "grad_norm": 0.5019553899765015, + "learning_rate": 0.00017381179325443372, + "loss": 0.1896, + "step": 10063 + }, + { + "epoch": 0.2359160088374737, + "grad_norm": 0.3338049054145813, + "learning_rate": 0.00017380682331587789, + "loss": 0.0531, + "step": 10064 + }, + { + "epoch": 0.2359394504122787, + "grad_norm": 0.6971047520637512, + "learning_rate": 0.00017380185297684412, + "loss": 0.7215, + "step": 10065 + }, + { + "epoch": 0.2359628919870837, + "grad_norm": 0.5513423681259155, + "learning_rate": 0.00017379688223735946, + "loss": 0.122, + "step": 10066 + }, + { + "epoch": 0.2359863335618887, + "grad_norm": 0.6905301213264465, + "learning_rate": 0.0001737919110974508, + "loss": 0.1668, + "step": 10067 + }, + { + "epoch": 0.2360097751366937, + "grad_norm": 0.7024086713790894, + "learning_rate": 0.0001737869395571452, + "loss": 0.625, + "step": 10068 + }, + { + "epoch": 0.23603321671149868, + "grad_norm": 0.7843540906906128, + "learning_rate": 0.00017378196761646958, + "loss": 0.0654, + "step": 10069 + }, + { + "epoch": 0.23605665828630368, + "grad_norm": 0.7573236227035522, + "learning_rate": 0.00017377699527545092, + "loss": 0.1894, + "step": 10070 + }, + { + "epoch": 0.23608009986110867, + "grad_norm": 0.1892940253019333, + "learning_rate": 0.00017377202253411622, + "loss": 0.042, + "step": 10071 + }, + { + "epoch": 0.23610354143591367, + "grad_norm": 0.5872131586074829, + "learning_rate": 0.00017376704939249247, + "loss": 0.7829, + "step": 10072 + }, + { + "epoch": 0.23612698301071866, + "grad_norm": 0.5175689458847046, + "learning_rate": 0.00017376207585060658, + "loss": 0.1356, + "step": 10073 + }, + { + "epoch": 0.23615042458552366, + "grad_norm": 0.6594922542572021, + "learning_rate": 0.00017375710190848563, + "loss": 0.0735, + "step": 10074 + }, + { + "epoch": 0.23617386616032865, + "grad_norm": 0.37005189061164856, + "learning_rate": 0.00017375212756615658, + "loss": 0.093, + "step": 10075 + }, + { + "epoch": 0.23619730773513364, + "grad_norm": 0.46448054909706116, + "learning_rate": 0.00017374715282364637, + "loss": 0.0589, + "step": 10076 + }, + { + "epoch": 0.23622074930993864, + "grad_norm": 0.4806431829929352, + "learning_rate": 0.0001737421776809821, + "loss": 0.0636, + "step": 10077 + }, + { + "epoch": 0.23624419088474363, + "grad_norm": 0.38152313232421875, + "learning_rate": 0.00017373720213819066, + "loss": 0.0483, + "step": 10078 + }, + { + "epoch": 0.23626763245954863, + "grad_norm": 0.7191515564918518, + "learning_rate": 0.00017373222619529907, + "loss": 0.1785, + "step": 10079 + }, + { + "epoch": 0.23629107403435362, + "grad_norm": 0.7384956479072571, + "learning_rate": 0.0001737272498523344, + "loss": 0.7496, + "step": 10080 + }, + { + "epoch": 0.23631451560915862, + "grad_norm": 0.3936556875705719, + "learning_rate": 0.00017372227310932353, + "loss": 0.0678, + "step": 10081 + }, + { + "epoch": 0.2363379571839636, + "grad_norm": 0.675161600112915, + "learning_rate": 0.0001737172959662936, + "loss": 0.1931, + "step": 10082 + }, + { + "epoch": 0.2363613987587686, + "grad_norm": 0.5440999269485474, + "learning_rate": 0.00017371231842327152, + "loss": 0.1073, + "step": 10083 + }, + { + "epoch": 0.2363848403335736, + "grad_norm": 0.6762790083885193, + "learning_rate": 0.00017370734048028432, + "loss": 0.1867, + "step": 10084 + }, + { + "epoch": 0.2364082819083786, + "grad_norm": 0.47453728318214417, + "learning_rate": 0.00017370236213735903, + "loss": 0.0845, + "step": 10085 + }, + { + "epoch": 0.2364317234831836, + "grad_norm": 0.3349112272262573, + "learning_rate": 0.00017369738339452265, + "loss": 0.0679, + "step": 10086 + }, + { + "epoch": 0.2364551650579886, + "grad_norm": 0.2488349974155426, + "learning_rate": 0.00017369240425180218, + "loss": 0.0409, + "step": 10087 + }, + { + "epoch": 0.2364786066327936, + "grad_norm": 0.21234886348247528, + "learning_rate": 0.00017368742470922466, + "loss": 0.032, + "step": 10088 + }, + { + "epoch": 0.2365020482075986, + "grad_norm": 0.22639472782611847, + "learning_rate": 0.0001736824447668171, + "loss": 0.0424, + "step": 10089 + }, + { + "epoch": 0.2365254897824036, + "grad_norm": 0.41283106803894043, + "learning_rate": 0.00017367746442460656, + "loss": 0.0552, + "step": 10090 + }, + { + "epoch": 0.23654893135720859, + "grad_norm": 0.2784019708633423, + "learning_rate": 0.00017367248368261996, + "loss": 0.0335, + "step": 10091 + }, + { + "epoch": 0.23657237293201358, + "grad_norm": 0.3160547614097595, + "learning_rate": 0.00017366750254088444, + "loss": 0.1189, + "step": 10092 + }, + { + "epoch": 0.23659581450681857, + "grad_norm": 0.9978629350662231, + "learning_rate": 0.00017366252099942697, + "loss": 0.1882, + "step": 10093 + }, + { + "epoch": 0.23661925608162357, + "grad_norm": 0.35904088616371155, + "learning_rate": 0.0001736575390582746, + "loss": 0.7071, + "step": 10094 + }, + { + "epoch": 0.23664269765642856, + "grad_norm": 0.5836021304130554, + "learning_rate": 0.00017365255671745433, + "loss": 0.7013, + "step": 10095 + }, + { + "epoch": 0.23666613923123356, + "grad_norm": 0.29849281907081604, + "learning_rate": 0.00017364757397699324, + "loss": 0.0347, + "step": 10096 + }, + { + "epoch": 0.23668958080603855, + "grad_norm": 0.4212730824947357, + "learning_rate": 0.00017364259083691832, + "loss": 0.0799, + "step": 10097 + }, + { + "epoch": 0.23671302238084355, + "grad_norm": 0.6366367340087891, + "learning_rate": 0.00017363760729725662, + "loss": 0.1393, + "step": 10098 + }, + { + "epoch": 0.23673646395564854, + "grad_norm": 0.2789618968963623, + "learning_rate": 0.00017363262335803523, + "loss": 0.2084, + "step": 10099 + }, + { + "epoch": 0.23675990553045353, + "grad_norm": 0.2599222958087921, + "learning_rate": 0.00017362763901928113, + "loss": 0.0718, + "step": 10100 + }, + { + "epoch": 0.23678334710525853, + "grad_norm": 0.4409581124782562, + "learning_rate": 0.0001736226542810214, + "loss": 0.0899, + "step": 10101 + }, + { + "epoch": 0.23680678868006352, + "grad_norm": 0.2925170660018921, + "learning_rate": 0.0001736176691432831, + "loss": 0.0868, + "step": 10102 + }, + { + "epoch": 0.23683023025486852, + "grad_norm": 0.41358616948127747, + "learning_rate": 0.00017361268360609322, + "loss": 0.104, + "step": 10103 + }, + { + "epoch": 0.2368536718296735, + "grad_norm": 0.642440140247345, + "learning_rate": 0.00017360769766947888, + "loss": 0.154, + "step": 10104 + }, + { + "epoch": 0.2368771134044785, + "grad_norm": 0.8154634237289429, + "learning_rate": 0.00017360271133346706, + "loss": 0.2012, + "step": 10105 + }, + { + "epoch": 0.2369005549792835, + "grad_norm": 0.35103029012680054, + "learning_rate": 0.00017359772459808487, + "loss": 0.0785, + "step": 10106 + }, + { + "epoch": 0.2369239965540885, + "grad_norm": 0.45356181263923645, + "learning_rate": 0.00017359273746335937, + "loss": 0.096, + "step": 10107 + }, + { + "epoch": 0.2369474381288935, + "grad_norm": 0.5278180241584778, + "learning_rate": 0.00017358774992931762, + "loss": 0.0692, + "step": 10108 + }, + { + "epoch": 0.23697087970369848, + "grad_norm": 0.17772595584392548, + "learning_rate": 0.00017358276199598666, + "loss": 0.0349, + "step": 10109 + }, + { + "epoch": 0.23699432127850348, + "grad_norm": 0.5230815410614014, + "learning_rate": 0.00017357777366339356, + "loss": 0.0983, + "step": 10110 + }, + { + "epoch": 0.23701776285330847, + "grad_norm": 0.15178847312927246, + "learning_rate": 0.00017357278493156542, + "loss": 0.027, + "step": 10111 + }, + { + "epoch": 0.2370412044281135, + "grad_norm": 0.22361022233963013, + "learning_rate": 0.00017356779580052925, + "loss": 0.0507, + "step": 10112 + }, + { + "epoch": 0.2370646460029185, + "grad_norm": 0.4602797329425812, + "learning_rate": 0.00017356280627031214, + "loss": 0.1251, + "step": 10113 + }, + { + "epoch": 0.23708808757772348, + "grad_norm": 0.5025404095649719, + "learning_rate": 0.00017355781634094118, + "loss": 0.0816, + "step": 10114 + }, + { + "epoch": 0.23711152915252848, + "grad_norm": 0.3676910102367401, + "learning_rate": 0.00017355282601244344, + "loss": 0.0777, + "step": 10115 + }, + { + "epoch": 0.23713497072733347, + "grad_norm": 0.07802839577198029, + "learning_rate": 0.00017354783528484603, + "loss": 0.0183, + "step": 10116 + }, + { + "epoch": 0.23715841230213847, + "grad_norm": 0.8145523071289062, + "learning_rate": 0.00017354284415817598, + "loss": 0.1371, + "step": 10117 + }, + { + "epoch": 0.23718185387694346, + "grad_norm": 0.3685770332813263, + "learning_rate": 0.00017353785263246038, + "loss": 0.1089, + "step": 10118 + }, + { + "epoch": 0.23720529545174845, + "grad_norm": 0.9134306311607361, + "learning_rate": 0.00017353286070772631, + "loss": 0.1903, + "step": 10119 + }, + { + "epoch": 0.23722873702655345, + "grad_norm": 0.5268301963806152, + "learning_rate": 0.0001735278683840009, + "loss": 0.1444, + "step": 10120 + }, + { + "epoch": 0.23725217860135844, + "grad_norm": 0.29681822657585144, + "learning_rate": 0.0001735228756613112, + "loss": 0.0468, + "step": 10121 + }, + { + "epoch": 0.23727562017616344, + "grad_norm": 0.37664303183555603, + "learning_rate": 0.00017351788253968433, + "loss": 0.1392, + "step": 10122 + }, + { + "epoch": 0.23729906175096843, + "grad_norm": 0.5661227107048035, + "learning_rate": 0.00017351288901914732, + "loss": 0.135, + "step": 10123 + }, + { + "epoch": 0.23732250332577343, + "grad_norm": 0.4538155496120453, + "learning_rate": 0.00017350789509972735, + "loss": 0.1366, + "step": 10124 + }, + { + "epoch": 0.23734594490057842, + "grad_norm": 0.45101064443588257, + "learning_rate": 0.00017350290078145143, + "loss": 0.4178, + "step": 10125 + }, + { + "epoch": 0.23736938647538341, + "grad_norm": 0.5643322467803955, + "learning_rate": 0.00017349790606434673, + "loss": 0.5171, + "step": 10126 + }, + { + "epoch": 0.2373928280501884, + "grad_norm": 0.6485903263092041, + "learning_rate": 0.00017349291094844032, + "loss": 0.1022, + "step": 10127 + }, + { + "epoch": 0.2374162696249934, + "grad_norm": 0.6306473016738892, + "learning_rate": 0.00017348791543375932, + "loss": 0.1237, + "step": 10128 + }, + { + "epoch": 0.2374397111997984, + "grad_norm": 0.720855176448822, + "learning_rate": 0.0001734829195203308, + "loss": 0.2041, + "step": 10129 + }, + { + "epoch": 0.2374631527746034, + "grad_norm": 0.567216157913208, + "learning_rate": 0.0001734779232081819, + "loss": 0.1343, + "step": 10130 + }, + { + "epoch": 0.2374865943494084, + "grad_norm": 0.5229189991950989, + "learning_rate": 0.00017347292649733974, + "loss": 0.1397, + "step": 10131 + }, + { + "epoch": 0.23751003592421338, + "grad_norm": 0.41629040241241455, + "learning_rate": 0.0001734679293878314, + "loss": 0.1404, + "step": 10132 + }, + { + "epoch": 0.23753347749901838, + "grad_norm": 0.38129231333732605, + "learning_rate": 0.000173462931879684, + "loss": 0.1029, + "step": 10133 + }, + { + "epoch": 0.23755691907382337, + "grad_norm": 0.46337008476257324, + "learning_rate": 0.0001734579339729247, + "loss": 0.1141, + "step": 10134 + }, + { + "epoch": 0.23758036064862836, + "grad_norm": 0.612399160861969, + "learning_rate": 0.00017345293566758057, + "loss": 0.8181, + "step": 10135 + }, + { + "epoch": 0.23760380222343336, + "grad_norm": 0.37817680835723877, + "learning_rate": 0.00017344793696367873, + "loss": 0.107, + "step": 10136 + }, + { + "epoch": 0.23762724379823835, + "grad_norm": 0.4683060944080353, + "learning_rate": 0.00017344293786124632, + "loss": 0.1164, + "step": 10137 + }, + { + "epoch": 0.23765068537304337, + "grad_norm": 0.5029197931289673, + "learning_rate": 0.0001734379383603105, + "loss": 0.0888, + "step": 10138 + }, + { + "epoch": 0.23767412694784837, + "grad_norm": 0.35529470443725586, + "learning_rate": 0.00017343293846089832, + "loss": 0.0819, + "step": 10139 + }, + { + "epoch": 0.23769756852265336, + "grad_norm": 0.5909577012062073, + "learning_rate": 0.00017342793816303695, + "loss": 0.0967, + "step": 10140 + }, + { + "epoch": 0.23772101009745836, + "grad_norm": 0.5591868758201599, + "learning_rate": 0.00017342293746675357, + "loss": 0.1182, + "step": 10141 + }, + { + "epoch": 0.23774445167226335, + "grad_norm": 0.26334285736083984, + "learning_rate": 0.0001734179363720752, + "loss": 0.0665, + "step": 10142 + }, + { + "epoch": 0.23776789324706835, + "grad_norm": 0.5135504603385925, + "learning_rate": 0.0001734129348790291, + "loss": 0.1824, + "step": 10143 + }, + { + "epoch": 0.23779133482187334, + "grad_norm": 0.24132439494132996, + "learning_rate": 0.00017340793298764232, + "loss": 0.0845, + "step": 10144 + }, + { + "epoch": 0.23781477639667833, + "grad_norm": 0.4754857122898102, + "learning_rate": 0.00017340293069794204, + "loss": 0.0885, + "step": 10145 + }, + { + "epoch": 0.23783821797148333, + "grad_norm": 0.4988524913787842, + "learning_rate": 0.00017339792800995541, + "loss": 0.1044, + "step": 10146 + }, + { + "epoch": 0.23786165954628832, + "grad_norm": 0.18401630222797394, + "learning_rate": 0.00017339292492370953, + "loss": 0.0274, + "step": 10147 + }, + { + "epoch": 0.23788510112109332, + "grad_norm": 0.7363924980163574, + "learning_rate": 0.00017338792143923158, + "loss": 0.1002, + "step": 10148 + }, + { + "epoch": 0.2379085426958983, + "grad_norm": 0.619953453540802, + "learning_rate": 0.00017338291755654873, + "loss": 0.1406, + "step": 10149 + }, + { + "epoch": 0.2379319842707033, + "grad_norm": 0.48465755581855774, + "learning_rate": 0.0001733779132756881, + "loss": 0.0866, + "step": 10150 + }, + { + "epoch": 0.2379554258455083, + "grad_norm": 0.7981391549110413, + "learning_rate": 0.00017337290859667683, + "loss": 0.1384, + "step": 10151 + }, + { + "epoch": 0.2379788674203133, + "grad_norm": 0.49396878480911255, + "learning_rate": 0.0001733679035195421, + "loss": 0.1178, + "step": 10152 + }, + { + "epoch": 0.2380023089951183, + "grad_norm": 0.3897983431816101, + "learning_rate": 0.00017336289804431107, + "loss": 0.0588, + "step": 10153 + }, + { + "epoch": 0.23802575056992328, + "grad_norm": 0.6785329580307007, + "learning_rate": 0.00017335789217101088, + "loss": 0.0992, + "step": 10154 + }, + { + "epoch": 0.23804919214472828, + "grad_norm": 0.3642464280128479, + "learning_rate": 0.00017335288589966874, + "loss": 0.0728, + "step": 10155 + }, + { + "epoch": 0.23807263371953327, + "grad_norm": 0.4491799771785736, + "learning_rate": 0.00017334787923031176, + "loss": 0.094, + "step": 10156 + }, + { + "epoch": 0.23809607529433827, + "grad_norm": 0.5760008692741394, + "learning_rate": 0.00017334287216296709, + "loss": 0.1288, + "step": 10157 + }, + { + "epoch": 0.23811951686914326, + "grad_norm": 0.1990918666124344, + "learning_rate": 0.00017333786469766197, + "loss": 0.0664, + "step": 10158 + }, + { + "epoch": 0.23814295844394825, + "grad_norm": 0.22834907472133636, + "learning_rate": 0.00017333285683442354, + "loss": 0.0456, + "step": 10159 + }, + { + "epoch": 0.23816640001875325, + "grad_norm": 0.4761970639228821, + "learning_rate": 0.00017332784857327895, + "loss": 0.0866, + "step": 10160 + }, + { + "epoch": 0.23818984159355824, + "grad_norm": 0.6673595905303955, + "learning_rate": 0.00017332283991425537, + "loss": 0.1517, + "step": 10161 + }, + { + "epoch": 0.23821328316836324, + "grad_norm": 0.7355926632881165, + "learning_rate": 0.00017331783085738006, + "loss": 0.222, + "step": 10162 + }, + { + "epoch": 0.23823672474316823, + "grad_norm": 0.5996819734573364, + "learning_rate": 0.00017331282140268007, + "loss": 0.7723, + "step": 10163 + }, + { + "epoch": 0.23826016631797325, + "grad_norm": 0.5924426913261414, + "learning_rate": 0.00017330781155018267, + "loss": 0.4686, + "step": 10164 + }, + { + "epoch": 0.23828360789277825, + "grad_norm": 0.3987446129322052, + "learning_rate": 0.00017330280129991505, + "loss": 0.1092, + "step": 10165 + }, + { + "epoch": 0.23830704946758324, + "grad_norm": 0.4978594481945038, + "learning_rate": 0.00017329779065190434, + "loss": 0.1187, + "step": 10166 + }, + { + "epoch": 0.23833049104238824, + "grad_norm": 0.6765627264976501, + "learning_rate": 0.00017329277960617778, + "loss": 0.7628, + "step": 10167 + }, + { + "epoch": 0.23835393261719323, + "grad_norm": 0.1709720492362976, + "learning_rate": 0.00017328776816276253, + "loss": 0.0334, + "step": 10168 + }, + { + "epoch": 0.23837737419199823, + "grad_norm": 0.47327110171318054, + "learning_rate": 0.0001732827563216858, + "loss": 0.1082, + "step": 10169 + }, + { + "epoch": 0.23840081576680322, + "grad_norm": 0.2070228010416031, + "learning_rate": 0.00017327774408297473, + "loss": 0.0288, + "step": 10170 + }, + { + "epoch": 0.23842425734160821, + "grad_norm": 0.683274507522583, + "learning_rate": 0.0001732727314466566, + "loss": 0.092, + "step": 10171 + }, + { + "epoch": 0.2384476989164132, + "grad_norm": 0.28210747241973877, + "learning_rate": 0.00017326771841275857, + "loss": 0.0538, + "step": 10172 + }, + { + "epoch": 0.2384711404912182, + "grad_norm": 0.6713701486587524, + "learning_rate": 0.00017326270498130782, + "loss": 0.1317, + "step": 10173 + }, + { + "epoch": 0.2384945820660232, + "grad_norm": 0.4793860912322998, + "learning_rate": 0.00017325769115233159, + "loss": 0.1078, + "step": 10174 + }, + { + "epoch": 0.2385180236408282, + "grad_norm": 0.15206967294216156, + "learning_rate": 0.00017325267692585707, + "loss": 0.0229, + "step": 10175 + }, + { + "epoch": 0.23854146521563319, + "grad_norm": 0.13444924354553223, + "learning_rate": 0.00017324766230191144, + "loss": 0.0308, + "step": 10176 + }, + { + "epoch": 0.23856490679043818, + "grad_norm": 0.42907091975212097, + "learning_rate": 0.00017324264728052198, + "loss": 0.0349, + "step": 10177 + }, + { + "epoch": 0.23858834836524317, + "grad_norm": 0.20186582207679749, + "learning_rate": 0.0001732376318617158, + "loss": 0.0313, + "step": 10178 + }, + { + "epoch": 0.23861178994004817, + "grad_norm": 1.0719506740570068, + "learning_rate": 0.0001732326160455202, + "loss": 0.1798, + "step": 10179 + }, + { + "epoch": 0.23863523151485316, + "grad_norm": 0.8420286178588867, + "learning_rate": 0.00017322759983196234, + "loss": 0.1798, + "step": 10180 + }, + { + "epoch": 0.23865867308965816, + "grad_norm": 0.6509289741516113, + "learning_rate": 0.00017322258322106953, + "loss": 0.8087, + "step": 10181 + }, + { + "epoch": 0.23868211466446315, + "grad_norm": 0.29509779810905457, + "learning_rate": 0.00017321756621286886, + "loss": 0.0598, + "step": 10182 + }, + { + "epoch": 0.23870555623926815, + "grad_norm": 0.4973064661026001, + "learning_rate": 0.00017321254880738764, + "loss": 0.0787, + "step": 10183 + }, + { + "epoch": 0.23872899781407314, + "grad_norm": 0.14782263338565826, + "learning_rate": 0.00017320753100465307, + "loss": 0.0247, + "step": 10184 + }, + { + "epoch": 0.23875243938887813, + "grad_norm": 0.1349996030330658, + "learning_rate": 0.00017320251280469239, + "loss": 0.0354, + "step": 10185 + }, + { + "epoch": 0.23877588096368313, + "grad_norm": 0.4457310736179352, + "learning_rate": 0.0001731974942075328, + "loss": 0.3688, + "step": 10186 + }, + { + "epoch": 0.23879932253848812, + "grad_norm": 0.6162009835243225, + "learning_rate": 0.00017319247521320157, + "loss": 0.1448, + "step": 10187 + }, + { + "epoch": 0.23882276411329312, + "grad_norm": 0.3062056005001068, + "learning_rate": 0.00017318745582172587, + "loss": 0.0818, + "step": 10188 + }, + { + "epoch": 0.23884620568809814, + "grad_norm": 0.624832034111023, + "learning_rate": 0.00017318243603313302, + "loss": 0.7787, + "step": 10189 + }, + { + "epoch": 0.23886964726290313, + "grad_norm": 0.4158954918384552, + "learning_rate": 0.00017317741584745017, + "loss": 0.0573, + "step": 10190 + }, + { + "epoch": 0.23889308883770813, + "grad_norm": 0.46912142634391785, + "learning_rate": 0.00017317239526470462, + "loss": 0.3753, + "step": 10191 + }, + { + "epoch": 0.23891653041251312, + "grad_norm": 0.7865447402000427, + "learning_rate": 0.00017316737428492362, + "loss": 0.183, + "step": 10192 + }, + { + "epoch": 0.23893997198731812, + "grad_norm": 0.3288024961948395, + "learning_rate": 0.00017316235290813442, + "loss": 0.0597, + "step": 10193 + }, + { + "epoch": 0.2389634135621231, + "grad_norm": 0.2518462836742401, + "learning_rate": 0.00017315733113436416, + "loss": 0.0702, + "step": 10194 + }, + { + "epoch": 0.2389868551369281, + "grad_norm": 0.4958207905292511, + "learning_rate": 0.00017315230896364022, + "loss": 0.0887, + "step": 10195 + }, + { + "epoch": 0.2390102967117331, + "grad_norm": 0.510270357131958, + "learning_rate": 0.00017314728639598976, + "loss": 0.1105, + "step": 10196 + }, + { + "epoch": 0.2390337382865381, + "grad_norm": 0.33386704325675964, + "learning_rate": 0.0001731422634314401, + "loss": 0.0812, + "step": 10197 + }, + { + "epoch": 0.2390571798613431, + "grad_norm": 0.16915850341320038, + "learning_rate": 0.0001731372400700184, + "loss": 0.029, + "step": 10198 + }, + { + "epoch": 0.23908062143614808, + "grad_norm": 0.15201009809970856, + "learning_rate": 0.00017313221631175206, + "loss": 0.0392, + "step": 10199 + }, + { + "epoch": 0.23910406301095308, + "grad_norm": 0.4528534412384033, + "learning_rate": 0.0001731271921566682, + "loss": 0.6682, + "step": 10200 + }, + { + "epoch": 0.23912750458575807, + "grad_norm": 0.6325749158859253, + "learning_rate": 0.00017312216760479417, + "loss": 0.8705, + "step": 10201 + }, + { + "epoch": 0.23915094616056307, + "grad_norm": 0.5721643567085266, + "learning_rate": 0.00017311714265615718, + "loss": 0.0989, + "step": 10202 + }, + { + "epoch": 0.23917438773536806, + "grad_norm": 0.36006832122802734, + "learning_rate": 0.00017311211731078452, + "loss": 0.0916, + "step": 10203 + }, + { + "epoch": 0.23919782931017305, + "grad_norm": 0.6362485885620117, + "learning_rate": 0.00017310709156870347, + "loss": 0.1361, + "step": 10204 + }, + { + "epoch": 0.23922127088497805, + "grad_norm": 0.32819393277168274, + "learning_rate": 0.00017310206542994126, + "loss": 0.0544, + "step": 10205 + }, + { + "epoch": 0.23924471245978304, + "grad_norm": 0.718072235584259, + "learning_rate": 0.00017309703889452523, + "loss": 0.0991, + "step": 10206 + }, + { + "epoch": 0.23926815403458804, + "grad_norm": 0.5539748668670654, + "learning_rate": 0.0001730920119624826, + "loss": 0.0965, + "step": 10207 + }, + { + "epoch": 0.23929159560939303, + "grad_norm": 0.4569169282913208, + "learning_rate": 0.00017308698463384064, + "loss": 0.1131, + "step": 10208 + }, + { + "epoch": 0.23931503718419803, + "grad_norm": 0.2958424687385559, + "learning_rate": 0.00017308195690862664, + "loss": 0.0782, + "step": 10209 + }, + { + "epoch": 0.23933847875900302, + "grad_norm": 0.5702525973320007, + "learning_rate": 0.0001730769287868679, + "loss": 0.0856, + "step": 10210 + }, + { + "epoch": 0.23936192033380801, + "grad_norm": 0.300466388463974, + "learning_rate": 0.0001730719002685917, + "loss": 0.1098, + "step": 10211 + }, + { + "epoch": 0.239385361908613, + "grad_norm": 0.3087124526500702, + "learning_rate": 0.0001730668713538253, + "loss": 0.0629, + "step": 10212 + }, + { + "epoch": 0.239408803483418, + "grad_norm": 0.10565980523824692, + "learning_rate": 0.000173061842042596, + "loss": 0.0219, + "step": 10213 + }, + { + "epoch": 0.239432245058223, + "grad_norm": 0.5305407643318176, + "learning_rate": 0.00017305681233493108, + "loss": 0.1167, + "step": 10214 + }, + { + "epoch": 0.23945568663302802, + "grad_norm": 0.5331280827522278, + "learning_rate": 0.00017305178223085788, + "loss": 0.0933, + "step": 10215 + }, + { + "epoch": 0.23947912820783301, + "grad_norm": 0.5834969282150269, + "learning_rate": 0.00017304675173040362, + "loss": 0.1249, + "step": 10216 + }, + { + "epoch": 0.239502569782638, + "grad_norm": 0.19445136189460754, + "learning_rate": 0.00017304172083359567, + "loss": 0.0723, + "step": 10217 + }, + { + "epoch": 0.239526011357443, + "grad_norm": 0.5170983672142029, + "learning_rate": 0.00017303668954046124, + "loss": 0.1818, + "step": 10218 + }, + { + "epoch": 0.239549452932248, + "grad_norm": 0.6235284209251404, + "learning_rate": 0.0001730316578510277, + "loss": 0.1481, + "step": 10219 + }, + { + "epoch": 0.239572894507053, + "grad_norm": 0.4539189636707306, + "learning_rate": 0.00017302662576532235, + "loss": 0.0669, + "step": 10220 + }, + { + "epoch": 0.23959633608185799, + "grad_norm": 0.8477569222450256, + "learning_rate": 0.00017302159328337244, + "loss": 0.1994, + "step": 10221 + }, + { + "epoch": 0.23961977765666298, + "grad_norm": 0.04963661730289459, + "learning_rate": 0.00017301656040520533, + "loss": 0.0101, + "step": 10222 + }, + { + "epoch": 0.23964321923146797, + "grad_norm": 0.7951521277427673, + "learning_rate": 0.00017301152713084834, + "loss": 0.1202, + "step": 10223 + }, + { + "epoch": 0.23966666080627297, + "grad_norm": 0.4508524239063263, + "learning_rate": 0.0001730064934603287, + "loss": 0.1094, + "step": 10224 + }, + { + "epoch": 0.23969010238107796, + "grad_norm": 0.20807787775993347, + "learning_rate": 0.0001730014593936738, + "loss": 0.0305, + "step": 10225 + }, + { + "epoch": 0.23971354395588296, + "grad_norm": 0.598412275314331, + "learning_rate": 0.00017299642493091093, + "loss": 0.1089, + "step": 10226 + }, + { + "epoch": 0.23973698553068795, + "grad_norm": 0.42372336983680725, + "learning_rate": 0.0001729913900720674, + "loss": 0.0924, + "step": 10227 + }, + { + "epoch": 0.23976042710549295, + "grad_norm": 0.1999523639678955, + "learning_rate": 0.00017298635481717055, + "loss": 0.0606, + "step": 10228 + }, + { + "epoch": 0.23978386868029794, + "grad_norm": 0.26691123843193054, + "learning_rate": 0.00017298131916624766, + "loss": 0.0496, + "step": 10229 + }, + { + "epoch": 0.23980731025510293, + "grad_norm": 0.5536004304885864, + "learning_rate": 0.0001729762831193261, + "loss": 0.1207, + "step": 10230 + }, + { + "epoch": 0.23983075182990793, + "grad_norm": 0.25270596146583557, + "learning_rate": 0.00017297124667643317, + "loss": 0.043, + "step": 10231 + }, + { + "epoch": 0.23985419340471292, + "grad_norm": 0.35086581110954285, + "learning_rate": 0.0001729662098375962, + "loss": 0.0718, + "step": 10232 + }, + { + "epoch": 0.23987763497951792, + "grad_norm": 0.7039461135864258, + "learning_rate": 0.00017296117260284251, + "loss": 0.1227, + "step": 10233 + }, + { + "epoch": 0.2399010765543229, + "grad_norm": 0.6678536534309387, + "learning_rate": 0.00017295613497219948, + "loss": 0.1676, + "step": 10234 + }, + { + "epoch": 0.2399245181291279, + "grad_norm": 0.6935195326805115, + "learning_rate": 0.00017295109694569438, + "loss": 0.0839, + "step": 10235 + }, + { + "epoch": 0.2399479597039329, + "grad_norm": 0.4253934919834137, + "learning_rate": 0.00017294605852335459, + "loss": 0.1244, + "step": 10236 + }, + { + "epoch": 0.2399714012787379, + "grad_norm": 0.7092583179473877, + "learning_rate": 0.00017294101970520743, + "loss": 0.2194, + "step": 10237 + }, + { + "epoch": 0.2399948428535429, + "grad_norm": 0.5399492383003235, + "learning_rate": 0.00017293598049128025, + "loss": 0.1187, + "step": 10238 + }, + { + "epoch": 0.24001828442834788, + "grad_norm": 0.4055677652359009, + "learning_rate": 0.0001729309408816004, + "loss": 0.0679, + "step": 10239 + }, + { + "epoch": 0.2400417260031529, + "grad_norm": 0.8215765357017517, + "learning_rate": 0.0001729259008761952, + "loss": 0.0517, + "step": 10240 + }, + { + "epoch": 0.2400651675779579, + "grad_norm": 0.5744391083717346, + "learning_rate": 0.000172920860475092, + "loss": 0.1224, + "step": 10241 + }, + { + "epoch": 0.2400886091527629, + "grad_norm": 0.28055503964424133, + "learning_rate": 0.0001729158196783182, + "loss": 0.0334, + "step": 10242 + }, + { + "epoch": 0.2401120507275679, + "grad_norm": 0.4027986228466034, + "learning_rate": 0.00017291077848590106, + "loss": 0.0788, + "step": 10243 + }, + { + "epoch": 0.24013549230237288, + "grad_norm": 0.6850246787071228, + "learning_rate": 0.00017290573689786803, + "loss": 0.793, + "step": 10244 + }, + { + "epoch": 0.24015893387717788, + "grad_norm": 0.268706351518631, + "learning_rate": 0.0001729006949142464, + "loss": 0.079, + "step": 10245 + }, + { + "epoch": 0.24018237545198287, + "grad_norm": 0.49181273579597473, + "learning_rate": 0.00017289565253506356, + "loss": 0.1243, + "step": 10246 + }, + { + "epoch": 0.24020581702678787, + "grad_norm": 0.5036120414733887, + "learning_rate": 0.00017289060976034685, + "loss": 0.1417, + "step": 10247 + }, + { + "epoch": 0.24022925860159286, + "grad_norm": 0.5484995245933533, + "learning_rate": 0.00017288556659012368, + "loss": 0.1439, + "step": 10248 + }, + { + "epoch": 0.24025270017639785, + "grad_norm": 0.17666004598140717, + "learning_rate": 0.0001728805230244213, + "loss": 0.0477, + "step": 10249 + }, + { + "epoch": 0.24027614175120285, + "grad_norm": 0.14341989159584045, + "learning_rate": 0.00017287547906326723, + "loss": 0.0268, + "step": 10250 + }, + { + "epoch": 0.24029958332600784, + "grad_norm": 0.4132755398750305, + "learning_rate": 0.0001728704347066887, + "loss": 0.1352, + "step": 10251 + }, + { + "epoch": 0.24032302490081284, + "grad_norm": 0.5288094878196716, + "learning_rate": 0.00017286538995471318, + "loss": 0.6626, + "step": 10252 + }, + { + "epoch": 0.24034646647561783, + "grad_norm": 0.5904994010925293, + "learning_rate": 0.00017286034480736798, + "loss": 0.6649, + "step": 10253 + }, + { + "epoch": 0.24036990805042283, + "grad_norm": 0.873271644115448, + "learning_rate": 0.00017285529926468053, + "loss": 0.371, + "step": 10254 + }, + { + "epoch": 0.24039334962522782, + "grad_norm": 0.6005449891090393, + "learning_rate": 0.00017285025332667816, + "loss": 0.0727, + "step": 10255 + }, + { + "epoch": 0.24041679120003281, + "grad_norm": 0.3916987478733063, + "learning_rate": 0.00017284520699338825, + "loss": 0.1118, + "step": 10256 + }, + { + "epoch": 0.2404402327748378, + "grad_norm": 0.5709614753723145, + "learning_rate": 0.00017284016026483822, + "loss": 0.0874, + "step": 10257 + }, + { + "epoch": 0.2404636743496428, + "grad_norm": 0.4584828019142151, + "learning_rate": 0.00017283511314105545, + "loss": 0.6587, + "step": 10258 + }, + { + "epoch": 0.2404871159244478, + "grad_norm": 0.15676306188106537, + "learning_rate": 0.0001728300656220673, + "loss": 0.024, + "step": 10259 + }, + { + "epoch": 0.2405105574992528, + "grad_norm": 0.6036669611930847, + "learning_rate": 0.00017282501770790116, + "loss": 0.1087, + "step": 10260 + }, + { + "epoch": 0.24053399907405779, + "grad_norm": 0.8227988481521606, + "learning_rate": 0.0001728199693985844, + "loss": 0.1085, + "step": 10261 + }, + { + "epoch": 0.24055744064886278, + "grad_norm": 0.4206632375717163, + "learning_rate": 0.00017281492069414446, + "loss": 0.0671, + "step": 10262 + }, + { + "epoch": 0.24058088222366777, + "grad_norm": 0.5786264538764954, + "learning_rate": 0.00017280987159460872, + "loss": 0.1218, + "step": 10263 + }, + { + "epoch": 0.24060432379847277, + "grad_norm": 0.6699943542480469, + "learning_rate": 0.00017280482210000456, + "loss": 0.1609, + "step": 10264 + }, + { + "epoch": 0.24062776537327776, + "grad_norm": 0.1484641432762146, + "learning_rate": 0.0001727997722103594, + "loss": 0.0431, + "step": 10265 + }, + { + "epoch": 0.24065120694808279, + "grad_norm": 0.5215839147567749, + "learning_rate": 0.00017279472192570063, + "loss": 0.1189, + "step": 10266 + }, + { + "epoch": 0.24067464852288778, + "grad_norm": 0.47917255759239197, + "learning_rate": 0.00017278967124605563, + "loss": 0.5525, + "step": 10267 + }, + { + "epoch": 0.24069809009769277, + "grad_norm": 0.1984848976135254, + "learning_rate": 0.00017278462017145184, + "loss": 0.0596, + "step": 10268 + }, + { + "epoch": 0.24072153167249777, + "grad_norm": 0.2720576226711273, + "learning_rate": 0.00017277956870191666, + "loss": 0.0321, + "step": 10269 + }, + { + "epoch": 0.24074497324730276, + "grad_norm": 0.3758627474308014, + "learning_rate": 0.00017277451683747746, + "loss": 0.1272, + "step": 10270 + }, + { + "epoch": 0.24076841482210776, + "grad_norm": 0.64534592628479, + "learning_rate": 0.00017276946457816172, + "loss": 0.1267, + "step": 10271 + }, + { + "epoch": 0.24079185639691275, + "grad_norm": 0.4478974938392639, + "learning_rate": 0.00017276441192399684, + "loss": 0.1202, + "step": 10272 + }, + { + "epoch": 0.24081529797171775, + "grad_norm": 0.5207543969154358, + "learning_rate": 0.00017275935887501017, + "loss": 0.1137, + "step": 10273 + }, + { + "epoch": 0.24083873954652274, + "grad_norm": 0.5904563665390015, + "learning_rate": 0.0001727543054312292, + "loss": 0.1539, + "step": 10274 + }, + { + "epoch": 0.24086218112132773, + "grad_norm": 0.6702696681022644, + "learning_rate": 0.00017274925159268132, + "loss": 0.1439, + "step": 10275 + }, + { + "epoch": 0.24088562269613273, + "grad_norm": 0.35393524169921875, + "learning_rate": 0.00017274419735939392, + "loss": 0.0456, + "step": 10276 + }, + { + "epoch": 0.24090906427093772, + "grad_norm": 0.3167506456375122, + "learning_rate": 0.0001727391427313945, + "loss": 0.0772, + "step": 10277 + }, + { + "epoch": 0.24093250584574272, + "grad_norm": 0.6661062240600586, + "learning_rate": 0.00017273408770871043, + "loss": 0.1136, + "step": 10278 + }, + { + "epoch": 0.2409559474205477, + "grad_norm": 0.43029695749282837, + "learning_rate": 0.00017272903229136917, + "loss": 0.1154, + "step": 10279 + }, + { + "epoch": 0.2409793889953527, + "grad_norm": 0.6709954738616943, + "learning_rate": 0.0001727239764793981, + "loss": 0.1269, + "step": 10280 + }, + { + "epoch": 0.2410028305701577, + "grad_norm": 0.466411828994751, + "learning_rate": 0.00017271892027282472, + "loss": 0.124, + "step": 10281 + }, + { + "epoch": 0.2410262721449627, + "grad_norm": 0.4040454030036926, + "learning_rate": 0.00017271386367167642, + "loss": 0.0763, + "step": 10282 + }, + { + "epoch": 0.2410497137197677, + "grad_norm": 0.32627540826797485, + "learning_rate": 0.0001727088066759807, + "loss": 0.0745, + "step": 10283 + }, + { + "epoch": 0.24107315529457268, + "grad_norm": 0.5666033625602722, + "learning_rate": 0.00017270374928576488, + "loss": 0.1104, + "step": 10284 + }, + { + "epoch": 0.24109659686937768, + "grad_norm": 0.1845308244228363, + "learning_rate": 0.00017269869150105652, + "loss": 0.0361, + "step": 10285 + }, + { + "epoch": 0.24112003844418267, + "grad_norm": 0.5736370086669922, + "learning_rate": 0.00017269363332188297, + "loss": 0.7893, + "step": 10286 + }, + { + "epoch": 0.24114348001898767, + "grad_norm": 0.4101828336715698, + "learning_rate": 0.00017268857474827176, + "loss": 0.0968, + "step": 10287 + }, + { + "epoch": 0.24116692159379266, + "grad_norm": 0.2873794138431549, + "learning_rate": 0.0001726835157802503, + "loss": 0.0519, + "step": 10288 + }, + { + "epoch": 0.24119036316859765, + "grad_norm": 0.7001160979270935, + "learning_rate": 0.000172678456417846, + "loss": 0.1214, + "step": 10289 + }, + { + "epoch": 0.24121380474340265, + "grad_norm": 0.6031533479690552, + "learning_rate": 0.0001726733966610864, + "loss": 0.1256, + "step": 10290 + }, + { + "epoch": 0.24123724631820767, + "grad_norm": 0.874365508556366, + "learning_rate": 0.00017266833650999888, + "loss": 0.2041, + "step": 10291 + }, + { + "epoch": 0.24126068789301267, + "grad_norm": 0.7012614607810974, + "learning_rate": 0.00017266327596461093, + "loss": 0.2072, + "step": 10292 + }, + { + "epoch": 0.24128412946781766, + "grad_norm": 0.6429399251937866, + "learning_rate": 0.00017265821502495, + "loss": 0.1439, + "step": 10293 + }, + { + "epoch": 0.24130757104262265, + "grad_norm": 0.5377576351165771, + "learning_rate": 0.00017265315369104355, + "loss": 0.1032, + "step": 10294 + }, + { + "epoch": 0.24133101261742765, + "grad_norm": 0.2107095718383789, + "learning_rate": 0.00017264809196291907, + "loss": 0.0645, + "step": 10295 + }, + { + "epoch": 0.24135445419223264, + "grad_norm": 0.32248562574386597, + "learning_rate": 0.00017264302984060397, + "loss": 0.0262, + "step": 10296 + }, + { + "epoch": 0.24137789576703764, + "grad_norm": 0.49949127435684204, + "learning_rate": 0.00017263796732412576, + "loss": 0.6857, + "step": 10297 + }, + { + "epoch": 0.24140133734184263, + "grad_norm": 0.30738118290901184, + "learning_rate": 0.0001726329044135119, + "loss": 0.2077, + "step": 10298 + }, + { + "epoch": 0.24142477891664763, + "grad_norm": 0.21435129642486572, + "learning_rate": 0.00017262784110878983, + "loss": 0.0602, + "step": 10299 + }, + { + "epoch": 0.24144822049145262, + "grad_norm": 0.14432229101657867, + "learning_rate": 0.00017262277740998707, + "loss": 0.0294, + "step": 10300 + }, + { + "epoch": 0.24147166206625761, + "grad_norm": 0.5167664289474487, + "learning_rate": 0.00017261771331713107, + "loss": 0.7732, + "step": 10301 + }, + { + "epoch": 0.2414951036410626, + "grad_norm": 0.506373941898346, + "learning_rate": 0.00017261264883024934, + "loss": 0.1121, + "step": 10302 + }, + { + "epoch": 0.2415185452158676, + "grad_norm": 0.6095906496047974, + "learning_rate": 0.00017260758394936932, + "loss": 0.3864, + "step": 10303 + }, + { + "epoch": 0.2415419867906726, + "grad_norm": 0.18648463487625122, + "learning_rate": 0.0001726025186745185, + "loss": 0.0386, + "step": 10304 + }, + { + "epoch": 0.2415654283654776, + "grad_norm": 0.6509620547294617, + "learning_rate": 0.0001725974530057244, + "loss": 0.1017, + "step": 10305 + }, + { + "epoch": 0.24158886994028259, + "grad_norm": 0.19001786410808563, + "learning_rate": 0.00017259238694301447, + "loss": 0.0256, + "step": 10306 + }, + { + "epoch": 0.24161231151508758, + "grad_norm": 0.31319183111190796, + "learning_rate": 0.0001725873204864162, + "loss": 0.0404, + "step": 10307 + }, + { + "epoch": 0.24163575308989257, + "grad_norm": 0.20520424842834473, + "learning_rate": 0.00017258225363595708, + "loss": 0.0256, + "step": 10308 + }, + { + "epoch": 0.24165919466469757, + "grad_norm": 0.46334511041641235, + "learning_rate": 0.00017257718639166462, + "loss": 0.416, + "step": 10309 + }, + { + "epoch": 0.24168263623950256, + "grad_norm": 0.47211965918540955, + "learning_rate": 0.00017257211875356632, + "loss": 0.1184, + "step": 10310 + }, + { + "epoch": 0.24170607781430756, + "grad_norm": 0.8292112350463867, + "learning_rate": 0.00017256705072168964, + "loss": 0.2416, + "step": 10311 + }, + { + "epoch": 0.24172951938911255, + "grad_norm": 0.42899951338768005, + "learning_rate": 0.00017256198229606212, + "loss": 0.1254, + "step": 10312 + }, + { + "epoch": 0.24175296096391755, + "grad_norm": 0.5795680284500122, + "learning_rate": 0.00017255691347671122, + "loss": 0.1435, + "step": 10313 + }, + { + "epoch": 0.24177640253872254, + "grad_norm": 0.5677819848060608, + "learning_rate": 0.0001725518442636645, + "loss": 0.1458, + "step": 10314 + }, + { + "epoch": 0.24179984411352753, + "grad_norm": 0.7136471271514893, + "learning_rate": 0.00017254677465694938, + "loss": 0.2455, + "step": 10315 + }, + { + "epoch": 0.24182328568833253, + "grad_norm": 0.3838413655757904, + "learning_rate": 0.00017254170465659345, + "loss": 0.1048, + "step": 10316 + }, + { + "epoch": 0.24184672726313755, + "grad_norm": 0.09876592457294464, + "learning_rate": 0.00017253663426262424, + "loss": 0.0251, + "step": 10317 + }, + { + "epoch": 0.24187016883794255, + "grad_norm": 0.2620088756084442, + "learning_rate": 0.00017253156347506914, + "loss": 0.0305, + "step": 10318 + }, + { + "epoch": 0.24189361041274754, + "grad_norm": 0.43909013271331787, + "learning_rate": 0.00017252649229395576, + "loss": 0.1181, + "step": 10319 + }, + { + "epoch": 0.24191705198755253, + "grad_norm": 0.36715149879455566, + "learning_rate": 0.0001725214207193116, + "loss": 0.0458, + "step": 10320 + }, + { + "epoch": 0.24194049356235753, + "grad_norm": 0.5613242983818054, + "learning_rate": 0.00017251634875116418, + "loss": 0.1557, + "step": 10321 + }, + { + "epoch": 0.24196393513716252, + "grad_norm": 0.4170677363872528, + "learning_rate": 0.00017251127638954098, + "loss": 0.1266, + "step": 10322 + }, + { + "epoch": 0.24198737671196752, + "grad_norm": 0.7584241628646851, + "learning_rate": 0.00017250620363446962, + "loss": 0.1972, + "step": 10323 + }, + { + "epoch": 0.2420108182867725, + "grad_norm": 0.6618577837944031, + "learning_rate": 0.00017250113048597753, + "loss": 0.6981, + "step": 10324 + }, + { + "epoch": 0.2420342598615775, + "grad_norm": 0.50611412525177, + "learning_rate": 0.00017249605694409222, + "loss": 0.1519, + "step": 10325 + }, + { + "epoch": 0.2420577014363825, + "grad_norm": 0.42848217487335205, + "learning_rate": 0.0001724909830088413, + "loss": 0.1178, + "step": 10326 + }, + { + "epoch": 0.2420811430111875, + "grad_norm": 0.08866824954748154, + "learning_rate": 0.0001724859086802523, + "loss": 0.0181, + "step": 10327 + }, + { + "epoch": 0.2421045845859925, + "grad_norm": 0.816166341304779, + "learning_rate": 0.00017248083395835265, + "loss": 0.627, + "step": 10328 + }, + { + "epoch": 0.24212802616079748, + "grad_norm": 0.5892238020896912, + "learning_rate": 0.00017247575884317003, + "loss": 0.0986, + "step": 10329 + }, + { + "epoch": 0.24215146773560248, + "grad_norm": 0.7560095191001892, + "learning_rate": 0.00017247068333473185, + "loss": 0.1314, + "step": 10330 + }, + { + "epoch": 0.24217490931040747, + "grad_norm": 0.1872389167547226, + "learning_rate": 0.00017246560743306574, + "loss": 0.0354, + "step": 10331 + }, + { + "epoch": 0.24219835088521247, + "grad_norm": 0.6592507362365723, + "learning_rate": 0.00017246053113819922, + "loss": 0.1269, + "step": 10332 + }, + { + "epoch": 0.24222179246001746, + "grad_norm": 0.44185328483581543, + "learning_rate": 0.00017245545445015978, + "loss": 0.088, + "step": 10333 + }, + { + "epoch": 0.24224523403482245, + "grad_norm": 0.10461698472499847, + "learning_rate": 0.00017245037736897503, + "loss": 0.022, + "step": 10334 + }, + { + "epoch": 0.24226867560962745, + "grad_norm": 0.44285908341407776, + "learning_rate": 0.00017244529989467248, + "loss": 0.1081, + "step": 10335 + }, + { + "epoch": 0.24229211718443244, + "grad_norm": 0.29318225383758545, + "learning_rate": 0.0001724402220272797, + "loss": 0.0189, + "step": 10336 + }, + { + "epoch": 0.24231555875923744, + "grad_norm": 0.2785991132259369, + "learning_rate": 0.00017243514376682426, + "loss": 0.0249, + "step": 10337 + }, + { + "epoch": 0.24233900033404243, + "grad_norm": 0.272067129611969, + "learning_rate": 0.00017243006511333367, + "loss": 0.0737, + "step": 10338 + }, + { + "epoch": 0.24236244190884743, + "grad_norm": 0.765201985836029, + "learning_rate": 0.0001724249860668355, + "loss": 0.1394, + "step": 10339 + }, + { + "epoch": 0.24238588348365242, + "grad_norm": 0.7992691397666931, + "learning_rate": 0.00017241990662735735, + "loss": 0.138, + "step": 10340 + }, + { + "epoch": 0.24240932505845741, + "grad_norm": 0.2100738286972046, + "learning_rate": 0.00017241482679492673, + "loss": 0.0538, + "step": 10341 + }, + { + "epoch": 0.2424327666332624, + "grad_norm": 0.8874862194061279, + "learning_rate": 0.00017240974656957124, + "loss": 0.1206, + "step": 10342 + }, + { + "epoch": 0.24245620820806743, + "grad_norm": 0.23479004204273224, + "learning_rate": 0.00017240466595131842, + "loss": 0.0476, + "step": 10343 + }, + { + "epoch": 0.24247964978287243, + "grad_norm": 0.6109858751296997, + "learning_rate": 0.00017239958494019586, + "loss": 0.1112, + "step": 10344 + }, + { + "epoch": 0.24250309135767742, + "grad_norm": 0.35591721534729004, + "learning_rate": 0.0001723945035362311, + "loss": 0.0351, + "step": 10345 + }, + { + "epoch": 0.2425265329324824, + "grad_norm": 0.6098877191543579, + "learning_rate": 0.00017238942173945172, + "loss": 0.1143, + "step": 10346 + }, + { + "epoch": 0.2425499745072874, + "grad_norm": 0.6159080862998962, + "learning_rate": 0.00017238433954988531, + "loss": 0.6126, + "step": 10347 + }, + { + "epoch": 0.2425734160820924, + "grad_norm": 0.2635570168495178, + "learning_rate": 0.00017237925696755944, + "loss": 0.0585, + "step": 10348 + }, + { + "epoch": 0.2425968576568974, + "grad_norm": 0.2562665045261383, + "learning_rate": 0.00017237417399250168, + "loss": 0.0466, + "step": 10349 + }, + { + "epoch": 0.2426202992317024, + "grad_norm": 0.5607483386993408, + "learning_rate": 0.00017236909062473964, + "loss": 0.7409, + "step": 10350 + }, + { + "epoch": 0.24264374080650739, + "grad_norm": 0.5389678478240967, + "learning_rate": 0.00017236400686430084, + "loss": 0.0762, + "step": 10351 + }, + { + "epoch": 0.24266718238131238, + "grad_norm": 0.13542819023132324, + "learning_rate": 0.00017235892271121292, + "loss": 0.0224, + "step": 10352 + }, + { + "epoch": 0.24269062395611737, + "grad_norm": 0.7566735148429871, + "learning_rate": 0.00017235383816550348, + "loss": 0.1754, + "step": 10353 + }, + { + "epoch": 0.24271406553092237, + "grad_norm": 0.2960110306739807, + "learning_rate": 0.00017234875322720005, + "loss": 0.0868, + "step": 10354 + }, + { + "epoch": 0.24273750710572736, + "grad_norm": 0.555684506893158, + "learning_rate": 0.00017234366789633024, + "loss": 0.4598, + "step": 10355 + }, + { + "epoch": 0.24276094868053236, + "grad_norm": 0.5227928757667542, + "learning_rate": 0.0001723385821729217, + "loss": 0.082, + "step": 10356 + }, + { + "epoch": 0.24278439025533735, + "grad_norm": 0.321305513381958, + "learning_rate": 0.00017233349605700194, + "loss": 0.1007, + "step": 10357 + }, + { + "epoch": 0.24280783183014235, + "grad_norm": 0.5930278301239014, + "learning_rate": 0.00017232840954859861, + "loss": 0.1183, + "step": 10358 + }, + { + "epoch": 0.24283127340494734, + "grad_norm": 0.3836132287979126, + "learning_rate": 0.0001723233226477393, + "loss": 0.1191, + "step": 10359 + }, + { + "epoch": 0.24285471497975233, + "grad_norm": 0.5483798980712891, + "learning_rate": 0.00017231823535445158, + "loss": 0.148, + "step": 10360 + }, + { + "epoch": 0.24287815655455733, + "grad_norm": 0.7069822549819946, + "learning_rate": 0.00017231314766876312, + "loss": 0.1731, + "step": 10361 + }, + { + "epoch": 0.24290159812936232, + "grad_norm": 0.6196601986885071, + "learning_rate": 0.0001723080595907015, + "loss": 0.2015, + "step": 10362 + }, + { + "epoch": 0.24292503970416732, + "grad_norm": 0.5908275842666626, + "learning_rate": 0.00017230297112029428, + "loss": 0.1312, + "step": 10363 + }, + { + "epoch": 0.2429484812789723, + "grad_norm": 0.3871102035045624, + "learning_rate": 0.00017229788225756908, + "loss": 0.1208, + "step": 10364 + }, + { + "epoch": 0.2429719228537773, + "grad_norm": 0.7628756761550903, + "learning_rate": 0.0001722927930025536, + "loss": 0.2732, + "step": 10365 + }, + { + "epoch": 0.2429953644285823, + "grad_norm": 0.3509989082813263, + "learning_rate": 0.00017228770335527536, + "loss": 0.1035, + "step": 10366 + }, + { + "epoch": 0.2430188060033873, + "grad_norm": 0.388164222240448, + "learning_rate": 0.000172282613315762, + "loss": 0.0575, + "step": 10367 + }, + { + "epoch": 0.24304224757819232, + "grad_norm": 0.5004040002822876, + "learning_rate": 0.00017227752288404116, + "loss": 0.0843, + "step": 10368 + }, + { + "epoch": 0.2430656891529973, + "grad_norm": 0.8413380980491638, + "learning_rate": 0.00017227243206014047, + "loss": 0.1982, + "step": 10369 + }, + { + "epoch": 0.2430891307278023, + "grad_norm": 0.38575324416160583, + "learning_rate": 0.0001722673408440875, + "loss": 0.0879, + "step": 10370 + }, + { + "epoch": 0.2431125723026073, + "grad_norm": 0.4438917338848114, + "learning_rate": 0.00017226224923590992, + "loss": 0.1269, + "step": 10371 + }, + { + "epoch": 0.2431360138774123, + "grad_norm": 0.3063514232635498, + "learning_rate": 0.00017225715723563533, + "loss": 0.091, + "step": 10372 + }, + { + "epoch": 0.2431594554522173, + "grad_norm": 0.312131404876709, + "learning_rate": 0.00017225206484329137, + "loss": 0.0666, + "step": 10373 + }, + { + "epoch": 0.24318289702702228, + "grad_norm": 0.3050001263618469, + "learning_rate": 0.0001722469720589057, + "loss": 0.0576, + "step": 10374 + }, + { + "epoch": 0.24320633860182728, + "grad_norm": 0.18032269179821014, + "learning_rate": 0.00017224187888250592, + "loss": 0.0351, + "step": 10375 + }, + { + "epoch": 0.24322978017663227, + "grad_norm": 0.7541733384132385, + "learning_rate": 0.00017223678531411965, + "loss": 0.2077, + "step": 10376 + }, + { + "epoch": 0.24325322175143727, + "grad_norm": 0.4977695345878601, + "learning_rate": 0.00017223169135377456, + "loss": 0.111, + "step": 10377 + }, + { + "epoch": 0.24327666332624226, + "grad_norm": 0.5011966228485107, + "learning_rate": 0.00017222659700149833, + "loss": 0.1205, + "step": 10378 + }, + { + "epoch": 0.24330010490104725, + "grad_norm": 0.5191999077796936, + "learning_rate": 0.0001722215022573185, + "loss": 0.1266, + "step": 10379 + }, + { + "epoch": 0.24332354647585225, + "grad_norm": 0.4498714804649353, + "learning_rate": 0.0001722164071212628, + "loss": 0.0782, + "step": 10380 + }, + { + "epoch": 0.24334698805065724, + "grad_norm": 0.9417638182640076, + "learning_rate": 0.00017221131159335883, + "loss": 0.0915, + "step": 10381 + }, + { + "epoch": 0.24337042962546224, + "grad_norm": 0.34587034583091736, + "learning_rate": 0.00017220621567363425, + "loss": 0.2173, + "step": 10382 + }, + { + "epoch": 0.24339387120026723, + "grad_norm": 0.7577885389328003, + "learning_rate": 0.0001722011193621167, + "loss": 0.1679, + "step": 10383 + }, + { + "epoch": 0.24341731277507223, + "grad_norm": 0.5049103498458862, + "learning_rate": 0.00017219602265883388, + "loss": 0.1246, + "step": 10384 + }, + { + "epoch": 0.24344075434987722, + "grad_norm": 0.9142827391624451, + "learning_rate": 0.0001721909255638134, + "loss": 0.2321, + "step": 10385 + }, + { + "epoch": 0.24346419592468221, + "grad_norm": 0.6313765645027161, + "learning_rate": 0.0001721858280770829, + "loss": 0.4595, + "step": 10386 + }, + { + "epoch": 0.2434876374994872, + "grad_norm": 0.25233909487724304, + "learning_rate": 0.0001721807301986701, + "loss": 0.0667, + "step": 10387 + }, + { + "epoch": 0.2435110790742922, + "grad_norm": 0.3216259479522705, + "learning_rate": 0.00017217563192860262, + "loss": 0.0339, + "step": 10388 + }, + { + "epoch": 0.2435345206490972, + "grad_norm": 2.084895610809326, + "learning_rate": 0.00017217053326690812, + "loss": 0.1525, + "step": 10389 + }, + { + "epoch": 0.2435579622239022, + "grad_norm": 0.4741005599498749, + "learning_rate": 0.0001721654342136143, + "loss": 0.0696, + "step": 10390 + }, + { + "epoch": 0.24358140379870719, + "grad_norm": 0.18609045445919037, + "learning_rate": 0.00017216033476874878, + "loss": 0.0379, + "step": 10391 + }, + { + "epoch": 0.24360484537351218, + "grad_norm": 0.16310325264930725, + "learning_rate": 0.00017215523493233924, + "loss": 0.0453, + "step": 10392 + }, + { + "epoch": 0.24362828694831717, + "grad_norm": 1.287947654724121, + "learning_rate": 0.00017215013470441343, + "loss": 0.2512, + "step": 10393 + }, + { + "epoch": 0.2436517285231222, + "grad_norm": 0.36103343963623047, + "learning_rate": 0.00017214503408499891, + "loss": 0.0711, + "step": 10394 + }, + { + "epoch": 0.2436751700979272, + "grad_norm": 0.5622011423110962, + "learning_rate": 0.00017213993307412345, + "loss": 0.152, + "step": 10395 + }, + { + "epoch": 0.24369861167273218, + "grad_norm": 0.5947824716567993, + "learning_rate": 0.00017213483167181464, + "loss": 0.0973, + "step": 10396 + }, + { + "epoch": 0.24372205324753718, + "grad_norm": 0.46176019310951233, + "learning_rate": 0.0001721297298781002, + "loss": 0.1095, + "step": 10397 + }, + { + "epoch": 0.24374549482234217, + "grad_norm": 0.6107857823371887, + "learning_rate": 0.00017212462769300782, + "loss": 0.1103, + "step": 10398 + }, + { + "epoch": 0.24376893639714717, + "grad_norm": 0.2891556918621063, + "learning_rate": 0.0001721195251165652, + "loss": 0.0696, + "step": 10399 + }, + { + "epoch": 0.24379237797195216, + "grad_norm": 0.1787806898355484, + "learning_rate": 0.00017211442214880003, + "loss": 0.0371, + "step": 10400 + }, + { + "epoch": 0.24381581954675716, + "grad_norm": 0.3846605122089386, + "learning_rate": 0.00017210931878973998, + "loss": 0.1192, + "step": 10401 + }, + { + "epoch": 0.24383926112156215, + "grad_norm": 0.7776139974594116, + "learning_rate": 0.00017210421503941267, + "loss": 0.1418, + "step": 10402 + }, + { + "epoch": 0.24386270269636715, + "grad_norm": 0.23162022233009338, + "learning_rate": 0.00017209911089784594, + "loss": 0.0227, + "step": 10403 + }, + { + "epoch": 0.24388614427117214, + "grad_norm": 0.4118684232234955, + "learning_rate": 0.00017209400636506735, + "loss": 0.067, + "step": 10404 + }, + { + "epoch": 0.24390958584597713, + "grad_norm": 0.49955835938453674, + "learning_rate": 0.0001720889014411047, + "loss": 0.0939, + "step": 10405 + }, + { + "epoch": 0.24393302742078213, + "grad_norm": 0.42964521050453186, + "learning_rate": 0.00017208379612598562, + "loss": 0.0934, + "step": 10406 + }, + { + "epoch": 0.24395646899558712, + "grad_norm": 0.9770145416259766, + "learning_rate": 0.00017207869041973783, + "loss": 0.1163, + "step": 10407 + }, + { + "epoch": 0.24397991057039212, + "grad_norm": 0.8450762629508972, + "learning_rate": 0.00017207358432238902, + "loss": 0.1855, + "step": 10408 + }, + { + "epoch": 0.2440033521451971, + "grad_norm": 1.0001012086868286, + "learning_rate": 0.00017206847783396697, + "loss": 0.1748, + "step": 10409 + }, + { + "epoch": 0.2440267937200021, + "grad_norm": 0.31916895508766174, + "learning_rate": 0.0001720633709544993, + "loss": 0.0548, + "step": 10410 + }, + { + "epoch": 0.2440502352948071, + "grad_norm": 0.6633515954017639, + "learning_rate": 0.00017205826368401375, + "loss": 0.1428, + "step": 10411 + }, + { + "epoch": 0.2440736768696121, + "grad_norm": 0.20704680681228638, + "learning_rate": 0.00017205315602253804, + "loss": 0.0334, + "step": 10412 + }, + { + "epoch": 0.2440971184444171, + "grad_norm": 0.32116201519966125, + "learning_rate": 0.00017204804797009987, + "loss": 0.0575, + "step": 10413 + }, + { + "epoch": 0.24412056001922208, + "grad_norm": 0.46272918581962585, + "learning_rate": 0.000172042939526727, + "loss": 0.1295, + "step": 10414 + }, + { + "epoch": 0.24414400159402708, + "grad_norm": 0.4957813024520874, + "learning_rate": 0.0001720378306924471, + "loss": 0.0898, + "step": 10415 + }, + { + "epoch": 0.24416744316883207, + "grad_norm": 0.5111292004585266, + "learning_rate": 0.00017203272146728787, + "loss": 0.1323, + "step": 10416 + }, + { + "epoch": 0.24419088474363707, + "grad_norm": 0.2264803647994995, + "learning_rate": 0.00017202761185127708, + "loss": 0.0198, + "step": 10417 + }, + { + "epoch": 0.24421432631844206, + "grad_norm": 0.3911817669868469, + "learning_rate": 0.0001720225018444425, + "loss": 0.0331, + "step": 10418 + }, + { + "epoch": 0.24423776789324708, + "grad_norm": 0.7907423377037048, + "learning_rate": 0.00017201739144681174, + "loss": 0.1276, + "step": 10419 + }, + { + "epoch": 0.24426120946805208, + "grad_norm": 0.5385693907737732, + "learning_rate": 0.0001720122806584126, + "loss": 0.1017, + "step": 10420 + }, + { + "epoch": 0.24428465104285707, + "grad_norm": 0.8338881731033325, + "learning_rate": 0.00017200716947927283, + "loss": 0.5411, + "step": 10421 + }, + { + "epoch": 0.24430809261766206, + "grad_norm": 0.6565908789634705, + "learning_rate": 0.0001720020579094201, + "loss": 0.1011, + "step": 10422 + }, + { + "epoch": 0.24433153419246706, + "grad_norm": 0.5935671925544739, + "learning_rate": 0.00017199694594888218, + "loss": 0.1837, + "step": 10423 + }, + { + "epoch": 0.24435497576727205, + "grad_norm": 0.5631996393203735, + "learning_rate": 0.0001719918335976868, + "loss": 0.1921, + "step": 10424 + }, + { + "epoch": 0.24437841734207705, + "grad_norm": 0.12973126769065857, + "learning_rate": 0.00017198672085586174, + "loss": 0.032, + "step": 10425 + }, + { + "epoch": 0.24440185891688204, + "grad_norm": 0.7108723521232605, + "learning_rate": 0.00017198160772343468, + "loss": 0.1195, + "step": 10426 + }, + { + "epoch": 0.24442530049168704, + "grad_norm": 0.4214595556259155, + "learning_rate": 0.00017197649420043343, + "loss": 0.0938, + "step": 10427 + }, + { + "epoch": 0.24444874206649203, + "grad_norm": 0.31724104285240173, + "learning_rate": 0.00017197138028688565, + "loss": 0.0819, + "step": 10428 + }, + { + "epoch": 0.24447218364129703, + "grad_norm": 0.4724537134170532, + "learning_rate": 0.00017196626598281917, + "loss": 0.0801, + "step": 10429 + }, + { + "epoch": 0.24449562521610202, + "grad_norm": 0.33619430661201477, + "learning_rate": 0.0001719611512882617, + "loss": 0.0887, + "step": 10430 + }, + { + "epoch": 0.244519066790907, + "grad_norm": 0.7613144516944885, + "learning_rate": 0.000171956036203241, + "loss": 0.1369, + "step": 10431 + }, + { + "epoch": 0.244542508365712, + "grad_norm": 0.40647485852241516, + "learning_rate": 0.0001719509207277848, + "loss": 0.0761, + "step": 10432 + }, + { + "epoch": 0.244565949940517, + "grad_norm": 0.2418343871831894, + "learning_rate": 0.0001719458048619209, + "loss": 0.0515, + "step": 10433 + }, + { + "epoch": 0.244589391515322, + "grad_norm": 0.5828571319580078, + "learning_rate": 0.00017194068860567705, + "loss": 0.1058, + "step": 10434 + }, + { + "epoch": 0.244612833090127, + "grad_norm": 0.9039673209190369, + "learning_rate": 0.000171935571959081, + "loss": 0.2479, + "step": 10435 + }, + { + "epoch": 0.24463627466493199, + "grad_norm": 0.5404134392738342, + "learning_rate": 0.0001719304549221605, + "loss": 0.1251, + "step": 10436 + }, + { + "epoch": 0.24465971623973698, + "grad_norm": 0.4520339071750641, + "learning_rate": 0.00017192533749494336, + "loss": 0.0896, + "step": 10437 + }, + { + "epoch": 0.24468315781454197, + "grad_norm": 0.558517575263977, + "learning_rate": 0.0001719202196774573, + "loss": 0.1767, + "step": 10438 + }, + { + "epoch": 0.24470659938934697, + "grad_norm": 0.24477525055408478, + "learning_rate": 0.00017191510146973008, + "loss": 0.0658, + "step": 10439 + }, + { + "epoch": 0.24473004096415196, + "grad_norm": 0.1561620533466339, + "learning_rate": 0.0001719099828717895, + "loss": 0.0377, + "step": 10440 + }, + { + "epoch": 0.24475348253895696, + "grad_norm": 0.49943864345550537, + "learning_rate": 0.00017190486388366335, + "loss": 0.0882, + "step": 10441 + }, + { + "epoch": 0.24477692411376195, + "grad_norm": 0.3384992182254791, + "learning_rate": 0.00017189974450537937, + "loss": 0.0801, + "step": 10442 + }, + { + "epoch": 0.24480036568856695, + "grad_norm": 0.4875170886516571, + "learning_rate": 0.00017189462473696538, + "loss": 0.0868, + "step": 10443 + }, + { + "epoch": 0.24482380726337194, + "grad_norm": 0.22303986549377441, + "learning_rate": 0.0001718895045784491, + "loss": 0.0486, + "step": 10444 + }, + { + "epoch": 0.24484724883817696, + "grad_norm": 0.5308917760848999, + "learning_rate": 0.0001718843840298584, + "loss": 0.0901, + "step": 10445 + }, + { + "epoch": 0.24487069041298196, + "grad_norm": 0.8845254778862, + "learning_rate": 0.00017187926309122098, + "loss": 0.1821, + "step": 10446 + }, + { + "epoch": 0.24489413198778695, + "grad_norm": 0.6032607555389404, + "learning_rate": 0.00017187414176256466, + "loss": 0.1339, + "step": 10447 + }, + { + "epoch": 0.24491757356259194, + "grad_norm": 0.6862090826034546, + "learning_rate": 0.00017186902004391721, + "loss": 0.0989, + "step": 10448 + }, + { + "epoch": 0.24494101513739694, + "grad_norm": 0.5800846219062805, + "learning_rate": 0.00017186389793530645, + "loss": 0.1751, + "step": 10449 + }, + { + "epoch": 0.24496445671220193, + "grad_norm": 0.10214214771986008, + "learning_rate": 0.00017185877543676018, + "loss": 0.0126, + "step": 10450 + }, + { + "epoch": 0.24498789828700693, + "grad_norm": 1.2777515649795532, + "learning_rate": 0.00017185365254830613, + "loss": 0.1283, + "step": 10451 + }, + { + "epoch": 0.24501133986181192, + "grad_norm": 0.5428986549377441, + "learning_rate": 0.00017184852926997217, + "loss": 0.1127, + "step": 10452 + }, + { + "epoch": 0.24503478143661692, + "grad_norm": 0.6180917024612427, + "learning_rate": 0.00017184340560178607, + "loss": 0.2023, + "step": 10453 + }, + { + "epoch": 0.2450582230114219, + "grad_norm": 0.31354305148124695, + "learning_rate": 0.0001718382815437756, + "loss": 0.0724, + "step": 10454 + }, + { + "epoch": 0.2450816645862269, + "grad_norm": 0.7066413164138794, + "learning_rate": 0.00017183315709596865, + "loss": 0.1585, + "step": 10455 + }, + { + "epoch": 0.2451051061610319, + "grad_norm": 0.7280223369598389, + "learning_rate": 0.00017182803225839293, + "loss": 0.149, + "step": 10456 + }, + { + "epoch": 0.2451285477358369, + "grad_norm": 0.6842271089553833, + "learning_rate": 0.0001718229070310763, + "loss": 0.1351, + "step": 10457 + }, + { + "epoch": 0.2451519893106419, + "grad_norm": 0.6726391911506653, + "learning_rate": 0.00017181778141404656, + "loss": 0.6623, + "step": 10458 + }, + { + "epoch": 0.24517543088544688, + "grad_norm": 0.543451189994812, + "learning_rate": 0.00017181265540733152, + "loss": 0.0899, + "step": 10459 + }, + { + "epoch": 0.24519887246025188, + "grad_norm": 0.7656391263008118, + "learning_rate": 0.00017180752901095898, + "loss": 0.1628, + "step": 10460 + }, + { + "epoch": 0.24522231403505687, + "grad_norm": 0.7499356865882874, + "learning_rate": 0.00017180240222495677, + "loss": 0.5809, + "step": 10461 + }, + { + "epoch": 0.24524575560986187, + "grad_norm": 0.595203161239624, + "learning_rate": 0.00017179727504935271, + "loss": 0.0725, + "step": 10462 + }, + { + "epoch": 0.24526919718466686, + "grad_norm": 0.3484635055065155, + "learning_rate": 0.0001717921474841746, + "loss": 0.1154, + "step": 10463 + }, + { + "epoch": 0.24529263875947185, + "grad_norm": 0.7447439432144165, + "learning_rate": 0.0001717870195294503, + "loss": 0.1414, + "step": 10464 + }, + { + "epoch": 0.24531608033427685, + "grad_norm": 0.4863469898700714, + "learning_rate": 0.0001717818911852076, + "loss": 0.0798, + "step": 10465 + }, + { + "epoch": 0.24533952190908184, + "grad_norm": 0.2205309420824051, + "learning_rate": 0.00017177676245147438, + "loss": 0.0274, + "step": 10466 + }, + { + "epoch": 0.24536296348388684, + "grad_norm": 0.43361392617225647, + "learning_rate": 0.0001717716333282784, + "loss": 0.5886, + "step": 10467 + }, + { + "epoch": 0.24538640505869183, + "grad_norm": 0.16013722121715546, + "learning_rate": 0.00017176650381564752, + "loss": 0.0405, + "step": 10468 + }, + { + "epoch": 0.24540984663349683, + "grad_norm": 0.7453486323356628, + "learning_rate": 0.00017176137391360956, + "loss": 0.1676, + "step": 10469 + }, + { + "epoch": 0.24543328820830185, + "grad_norm": 0.6131359934806824, + "learning_rate": 0.0001717562436221924, + "loss": 0.1455, + "step": 10470 + }, + { + "epoch": 0.24545672978310684, + "grad_norm": 0.3699627220630646, + "learning_rate": 0.0001717511129414238, + "loss": 0.0925, + "step": 10471 + }, + { + "epoch": 0.24548017135791184, + "grad_norm": 0.5826449394226074, + "learning_rate": 0.00017174598187133167, + "loss": 0.6822, + "step": 10472 + }, + { + "epoch": 0.24550361293271683, + "grad_norm": 0.5156713724136353, + "learning_rate": 0.0001717408504119438, + "loss": 0.0994, + "step": 10473 + }, + { + "epoch": 0.24552705450752182, + "grad_norm": 0.7317230701446533, + "learning_rate": 0.0001717357185632881, + "loss": 0.1513, + "step": 10474 + }, + { + "epoch": 0.24555049608232682, + "grad_norm": 0.24020077288150787, + "learning_rate": 0.00017173058632539236, + "loss": 0.0661, + "step": 10475 + }, + { + "epoch": 0.2455739376571318, + "grad_norm": 0.6385550498962402, + "learning_rate": 0.0001717254536982844, + "loss": 0.1173, + "step": 10476 + }, + { + "epoch": 0.2455973792319368, + "grad_norm": 0.7971703410148621, + "learning_rate": 0.00017172032068199218, + "loss": 0.1012, + "step": 10477 + }, + { + "epoch": 0.2456208208067418, + "grad_norm": 0.470656156539917, + "learning_rate": 0.00017171518727654347, + "loss": 0.1551, + "step": 10478 + }, + { + "epoch": 0.2456442623815468, + "grad_norm": 0.30251839756965637, + "learning_rate": 0.00017171005348196608, + "loss": 0.0664, + "step": 10479 + }, + { + "epoch": 0.2456677039563518, + "grad_norm": 0.46908849477767944, + "learning_rate": 0.00017170491929828797, + "loss": 0.1106, + "step": 10480 + }, + { + "epoch": 0.24569114553115678, + "grad_norm": 0.39165493845939636, + "learning_rate": 0.00017169978472553695, + "loss": 0.0499, + "step": 10481 + }, + { + "epoch": 0.24571458710596178, + "grad_norm": 0.6180287599563599, + "learning_rate": 0.00017169464976374087, + "loss": 0.2072, + "step": 10482 + }, + { + "epoch": 0.24573802868076677, + "grad_norm": 0.5323134064674377, + "learning_rate": 0.0001716895144129276, + "loss": 0.0986, + "step": 10483 + }, + { + "epoch": 0.24576147025557177, + "grad_norm": 0.3620815873146057, + "learning_rate": 0.00017168437867312505, + "loss": 0.0637, + "step": 10484 + }, + { + "epoch": 0.24578491183037676, + "grad_norm": 0.16888366639614105, + "learning_rate": 0.000171679242544361, + "loss": 0.0284, + "step": 10485 + }, + { + "epoch": 0.24580835340518176, + "grad_norm": 0.510675311088562, + "learning_rate": 0.00017167410602666336, + "loss": 0.1519, + "step": 10486 + }, + { + "epoch": 0.24583179497998675, + "grad_norm": 0.4788723289966583, + "learning_rate": 0.00017166896912006006, + "loss": 0.0975, + "step": 10487 + }, + { + "epoch": 0.24585523655479175, + "grad_norm": 0.0988776683807373, + "learning_rate": 0.00017166383182457885, + "loss": 0.0348, + "step": 10488 + }, + { + "epoch": 0.24587867812959674, + "grad_norm": 0.47403016686439514, + "learning_rate": 0.0001716586941402477, + "loss": 0.1164, + "step": 10489 + }, + { + "epoch": 0.24590211970440173, + "grad_norm": 0.660778284072876, + "learning_rate": 0.0001716535560670945, + "loss": 0.1793, + "step": 10490 + }, + { + "epoch": 0.24592556127920673, + "grad_norm": 0.7865579128265381, + "learning_rate": 0.00017164841760514702, + "loss": 0.1924, + "step": 10491 + }, + { + "epoch": 0.24594900285401172, + "grad_norm": 0.6828781366348267, + "learning_rate": 0.00017164327875443327, + "loss": 0.9683, + "step": 10492 + }, + { + "epoch": 0.24597244442881672, + "grad_norm": 0.5930101275444031, + "learning_rate": 0.00017163813951498106, + "loss": 0.1249, + "step": 10493 + }, + { + "epoch": 0.2459958860036217, + "grad_norm": 0.6004011034965515, + "learning_rate": 0.0001716329998868183, + "loss": 0.0928, + "step": 10494 + }, + { + "epoch": 0.2460193275784267, + "grad_norm": 0.6438028812408447, + "learning_rate": 0.00017162785986997282, + "loss": 0.1102, + "step": 10495 + }, + { + "epoch": 0.24604276915323173, + "grad_norm": 0.5042461156845093, + "learning_rate": 0.0001716227194644726, + "loss": 0.5057, + "step": 10496 + }, + { + "epoch": 0.24606621072803672, + "grad_norm": 0.657270610332489, + "learning_rate": 0.00017161757867034547, + "loss": 0.6121, + "step": 10497 + }, + { + "epoch": 0.24608965230284172, + "grad_norm": 0.7524641156196594, + "learning_rate": 0.00017161243748761937, + "loss": 0.2151, + "step": 10498 + }, + { + "epoch": 0.2461130938776467, + "grad_norm": 0.6761435270309448, + "learning_rate": 0.00017160729591632215, + "loss": 0.0555, + "step": 10499 + }, + { + "epoch": 0.2461365354524517, + "grad_norm": 0.7976429462432861, + "learning_rate": 0.00017160215395648173, + "loss": 0.1822, + "step": 10500 + }, + { + "epoch": 0.2461599770272567, + "grad_norm": 0.6852023601531982, + "learning_rate": 0.00017159701160812603, + "loss": 0.7269, + "step": 10501 + }, + { + "epoch": 0.2461834186020617, + "grad_norm": 0.4652552902698517, + "learning_rate": 0.0001715918688712829, + "loss": 0.1352, + "step": 10502 + }, + { + "epoch": 0.2462068601768667, + "grad_norm": 0.2858380079269409, + "learning_rate": 0.00017158672574598031, + "loss": 0.0275, + "step": 10503 + }, + { + "epoch": 0.24623030175167168, + "grad_norm": 0.2639264166355133, + "learning_rate": 0.0001715815822322461, + "loss": 0.077, + "step": 10504 + }, + { + "epoch": 0.24625374332647668, + "grad_norm": 0.25378191471099854, + "learning_rate": 0.00017157643833010824, + "loss": 0.0619, + "step": 10505 + }, + { + "epoch": 0.24627718490128167, + "grad_norm": 0.43709900975227356, + "learning_rate": 0.0001715712940395946, + "loss": 0.1026, + "step": 10506 + }, + { + "epoch": 0.24630062647608666, + "grad_norm": 0.21394671499729156, + "learning_rate": 0.0001715661493607331, + "loss": 0.0376, + "step": 10507 + }, + { + "epoch": 0.24632406805089166, + "grad_norm": 0.6523244976997375, + "learning_rate": 0.00017156100429355169, + "loss": 0.1185, + "step": 10508 + }, + { + "epoch": 0.24634750962569665, + "grad_norm": 0.151919886469841, + "learning_rate": 0.00017155585883807823, + "loss": 0.0464, + "step": 10509 + }, + { + "epoch": 0.24637095120050165, + "grad_norm": 0.6359995007514954, + "learning_rate": 0.00017155071299434067, + "loss": 0.1008, + "step": 10510 + }, + { + "epoch": 0.24639439277530664, + "grad_norm": 0.6571186184883118, + "learning_rate": 0.00017154556676236694, + "loss": 0.0964, + "step": 10511 + }, + { + "epoch": 0.24641783435011164, + "grad_norm": 0.5761663317680359, + "learning_rate": 0.00017154042014218493, + "loss": 0.1441, + "step": 10512 + }, + { + "epoch": 0.24644127592491663, + "grad_norm": 0.204716295003891, + "learning_rate": 0.00017153527313382264, + "loss": 0.0148, + "step": 10513 + }, + { + "epoch": 0.24646471749972163, + "grad_norm": 0.5703563094139099, + "learning_rate": 0.0001715301257373079, + "loss": 0.1262, + "step": 10514 + }, + { + "epoch": 0.24648815907452662, + "grad_norm": 0.37475666403770447, + "learning_rate": 0.00017152497795266868, + "loss": 0.4235, + "step": 10515 + }, + { + "epoch": 0.2465116006493316, + "grad_norm": 0.25294217467308044, + "learning_rate": 0.00017151982977993294, + "loss": 0.0601, + "step": 10516 + }, + { + "epoch": 0.2465350422241366, + "grad_norm": 0.392120361328125, + "learning_rate": 0.0001715146812191286, + "loss": 0.1162, + "step": 10517 + }, + { + "epoch": 0.2465584837989416, + "grad_norm": 0.38905543088912964, + "learning_rate": 0.0001715095322702836, + "loss": 0.0915, + "step": 10518 + }, + { + "epoch": 0.2465819253737466, + "grad_norm": 0.33888357877731323, + "learning_rate": 0.00017150438293342583, + "loss": 0.3145, + "step": 10519 + }, + { + "epoch": 0.2466053669485516, + "grad_norm": 0.20312117040157318, + "learning_rate": 0.00017149923320858328, + "loss": 0.0383, + "step": 10520 + }, + { + "epoch": 0.2466288085233566, + "grad_norm": 0.5814870595932007, + "learning_rate": 0.0001714940830957839, + "loss": 0.096, + "step": 10521 + }, + { + "epoch": 0.2466522500981616, + "grad_norm": 0.49274903535842896, + "learning_rate": 0.0001714889325950556, + "loss": 0.116, + "step": 10522 + }, + { + "epoch": 0.2466756916729666, + "grad_norm": 0.56732177734375, + "learning_rate": 0.0001714837817064263, + "loss": 0.1133, + "step": 10523 + }, + { + "epoch": 0.2466991332477716, + "grad_norm": 0.602167546749115, + "learning_rate": 0.00017147863042992406, + "loss": 0.1293, + "step": 10524 + }, + { + "epoch": 0.2467225748225766, + "grad_norm": 0.912916898727417, + "learning_rate": 0.00017147347876557675, + "loss": 0.1233, + "step": 10525 + }, + { + "epoch": 0.24674601639738158, + "grad_norm": 0.5241827964782715, + "learning_rate": 0.0001714683267134123, + "loss": 0.8681, + "step": 10526 + }, + { + "epoch": 0.24676945797218658, + "grad_norm": 0.6528499126434326, + "learning_rate": 0.00017146317427345874, + "loss": 0.1148, + "step": 10527 + }, + { + "epoch": 0.24679289954699157, + "grad_norm": 0.48694056272506714, + "learning_rate": 0.00017145802144574395, + "loss": 0.1509, + "step": 10528 + }, + { + "epoch": 0.24681634112179657, + "grad_norm": 0.6997070908546448, + "learning_rate": 0.00017145286823029592, + "loss": 0.1084, + "step": 10529 + }, + { + "epoch": 0.24683978269660156, + "grad_norm": 0.3461887240409851, + "learning_rate": 0.00017144771462714267, + "loss": 0.0776, + "step": 10530 + }, + { + "epoch": 0.24686322427140656, + "grad_norm": 0.43348968029022217, + "learning_rate": 0.00017144256063631206, + "loss": 0.0946, + "step": 10531 + }, + { + "epoch": 0.24688666584621155, + "grad_norm": 1.0059682130813599, + "learning_rate": 0.00017143740625783212, + "loss": 0.31, + "step": 10532 + }, + { + "epoch": 0.24691010742101654, + "grad_norm": 2.3712732791900635, + "learning_rate": 0.00017143225149173082, + "loss": 0.1734, + "step": 10533 + }, + { + "epoch": 0.24693354899582154, + "grad_norm": 0.4301984906196594, + "learning_rate": 0.00017142709633803614, + "loss": 0.1036, + "step": 10534 + }, + { + "epoch": 0.24695699057062653, + "grad_norm": 0.6078599095344543, + "learning_rate": 0.00017142194079677596, + "loss": 0.0686, + "step": 10535 + }, + { + "epoch": 0.24698043214543153, + "grad_norm": 0.3375696539878845, + "learning_rate": 0.00017141678486797836, + "loss": 0.077, + "step": 10536 + }, + { + "epoch": 0.24700387372023652, + "grad_norm": 0.14864733815193176, + "learning_rate": 0.00017141162855167127, + "loss": 0.0348, + "step": 10537 + }, + { + "epoch": 0.24702731529504152, + "grad_norm": 0.16918663680553436, + "learning_rate": 0.0001714064718478827, + "loss": 0.0236, + "step": 10538 + }, + { + "epoch": 0.2470507568698465, + "grad_norm": 0.4522308111190796, + "learning_rate": 0.00017140131475664057, + "loss": 0.0804, + "step": 10539 + }, + { + "epoch": 0.2470741984446515, + "grad_norm": 0.7145341634750366, + "learning_rate": 0.00017139615727797294, + "loss": 0.1822, + "step": 10540 + }, + { + "epoch": 0.2470976400194565, + "grad_norm": 0.46810880303382874, + "learning_rate": 0.0001713909994119077, + "loss": 0.0788, + "step": 10541 + }, + { + "epoch": 0.2471210815942615, + "grad_norm": 0.5743310451507568, + "learning_rate": 0.00017138584115847293, + "loss": 0.0963, + "step": 10542 + }, + { + "epoch": 0.2471445231690665, + "grad_norm": 0.6419326066970825, + "learning_rate": 0.00017138068251769655, + "loss": 0.1312, + "step": 10543 + }, + { + "epoch": 0.24716796474387148, + "grad_norm": 0.29663339257240295, + "learning_rate": 0.00017137552348960662, + "loss": 0.0913, + "step": 10544 + }, + { + "epoch": 0.24719140631867648, + "grad_norm": 0.4346647560596466, + "learning_rate": 0.00017137036407423106, + "loss": 0.0854, + "step": 10545 + }, + { + "epoch": 0.24721484789348147, + "grad_norm": 0.6105912327766418, + "learning_rate": 0.0001713652042715979, + "loss": 0.1698, + "step": 10546 + }, + { + "epoch": 0.2472382894682865, + "grad_norm": 0.5425085425376892, + "learning_rate": 0.00017136004408173513, + "loss": 0.1495, + "step": 10547 + }, + { + "epoch": 0.2472617310430915, + "grad_norm": 0.672114908695221, + "learning_rate": 0.00017135488350467076, + "loss": 0.1331, + "step": 10548 + }, + { + "epoch": 0.24728517261789648, + "grad_norm": 0.32048940658569336, + "learning_rate": 0.0001713497225404328, + "loss": 0.0672, + "step": 10549 + }, + { + "epoch": 0.24730861419270148, + "grad_norm": 0.25433585047721863, + "learning_rate": 0.00017134456118904923, + "loss": 0.0654, + "step": 10550 + }, + { + "epoch": 0.24733205576750647, + "grad_norm": 0.30254682898521423, + "learning_rate": 0.00017133939945054806, + "loss": 0.0547, + "step": 10551 + }, + { + "epoch": 0.24735549734231146, + "grad_norm": 0.2350485622882843, + "learning_rate": 0.0001713342373249573, + "loss": 0.0414, + "step": 10552 + }, + { + "epoch": 0.24737893891711646, + "grad_norm": 0.7469107508659363, + "learning_rate": 0.000171329074812305, + "loss": 0.1635, + "step": 10553 + }, + { + "epoch": 0.24740238049192145, + "grad_norm": 0.4970915913581848, + "learning_rate": 0.0001713239119126191, + "loss": 0.0996, + "step": 10554 + }, + { + "epoch": 0.24742582206672645, + "grad_norm": 0.5180898904800415, + "learning_rate": 0.00017131874862592766, + "loss": 0.1672, + "step": 10555 + }, + { + "epoch": 0.24744926364153144, + "grad_norm": 0.37500831484794617, + "learning_rate": 0.00017131358495225865, + "loss": 0.0852, + "step": 10556 + }, + { + "epoch": 0.24747270521633644, + "grad_norm": 0.09571556746959686, + "learning_rate": 0.00017130842089164015, + "loss": 0.0267, + "step": 10557 + }, + { + "epoch": 0.24749614679114143, + "grad_norm": 0.5624181628227234, + "learning_rate": 0.00017130325644410018, + "loss": 0.1868, + "step": 10558 + }, + { + "epoch": 0.24751958836594642, + "grad_norm": 0.22833454608917236, + "learning_rate": 0.00017129809160966667, + "loss": 0.0537, + "step": 10559 + }, + { + "epoch": 0.24754302994075142, + "grad_norm": 0.46978214383125305, + "learning_rate": 0.00017129292638836776, + "loss": 0.0853, + "step": 10560 + }, + { + "epoch": 0.2475664715155564, + "grad_norm": 1.2081446647644043, + "learning_rate": 0.0001712877607802314, + "loss": 0.2459, + "step": 10561 + }, + { + "epoch": 0.2475899130903614, + "grad_norm": 0.5574521422386169, + "learning_rate": 0.00017128259478528567, + "loss": 0.0903, + "step": 10562 + }, + { + "epoch": 0.2476133546651664, + "grad_norm": 0.489193856716156, + "learning_rate": 0.0001712774284035586, + "loss": 0.1127, + "step": 10563 + }, + { + "epoch": 0.2476367962399714, + "grad_norm": 0.5847039222717285, + "learning_rate": 0.00017127226163507814, + "loss": 0.1098, + "step": 10564 + }, + { + "epoch": 0.2476602378147764, + "grad_norm": 0.41153037548065186, + "learning_rate": 0.00017126709447987242, + "loss": 0.1535, + "step": 10565 + }, + { + "epoch": 0.24768367938958138, + "grad_norm": 0.6630839109420776, + "learning_rate": 0.00017126192693796944, + "loss": 0.1514, + "step": 10566 + }, + { + "epoch": 0.24770712096438638, + "grad_norm": 0.4277642071247101, + "learning_rate": 0.00017125675900939725, + "loss": 0.0906, + "step": 10567 + }, + { + "epoch": 0.24773056253919137, + "grad_norm": 0.5569185018539429, + "learning_rate": 0.00017125159069418384, + "loss": 0.1581, + "step": 10568 + }, + { + "epoch": 0.24775400411399637, + "grad_norm": 0.5997418165206909, + "learning_rate": 0.00017124642199235732, + "loss": 0.1221, + "step": 10569 + }, + { + "epoch": 0.24777744568880136, + "grad_norm": 0.5799440145492554, + "learning_rate": 0.00017124125290394573, + "loss": 0.1271, + "step": 10570 + }, + { + "epoch": 0.24780088726360636, + "grad_norm": 0.13740438222885132, + "learning_rate": 0.00017123608342897706, + "loss": 0.0267, + "step": 10571 + }, + { + "epoch": 0.24782432883841135, + "grad_norm": 0.4515792727470398, + "learning_rate": 0.00017123091356747942, + "loss": 0.1041, + "step": 10572 + }, + { + "epoch": 0.24784777041321637, + "grad_norm": 0.5553436279296875, + "learning_rate": 0.00017122574331948084, + "loss": 0.1767, + "step": 10573 + }, + { + "epoch": 0.24787121198802137, + "grad_norm": 0.9329277873039246, + "learning_rate": 0.00017122057268500936, + "loss": 0.2405, + "step": 10574 + }, + { + "epoch": 0.24789465356282636, + "grad_norm": 0.4940938651561737, + "learning_rate": 0.00017121540166409308, + "loss": 0.1115, + "step": 10575 + }, + { + "epoch": 0.24791809513763136, + "grad_norm": 0.27256038784980774, + "learning_rate": 0.00017121023025676, + "loss": 0.0377, + "step": 10576 + }, + { + "epoch": 0.24794153671243635, + "grad_norm": 0.48588117957115173, + "learning_rate": 0.00017120505846303824, + "loss": 0.0833, + "step": 10577 + }, + { + "epoch": 0.24796497828724134, + "grad_norm": 0.7030973434448242, + "learning_rate": 0.0001711998862829558, + "loss": 0.9221, + "step": 10578 + }, + { + "epoch": 0.24798841986204634, + "grad_norm": 0.3802284598350525, + "learning_rate": 0.00017119471371654078, + "loss": 0.0675, + "step": 10579 + }, + { + "epoch": 0.24801186143685133, + "grad_norm": 0.6283760666847229, + "learning_rate": 0.00017118954076382126, + "loss": 0.1234, + "step": 10580 + }, + { + "epoch": 0.24803530301165633, + "grad_norm": 0.13610920310020447, + "learning_rate": 0.0001711843674248253, + "loss": 0.0314, + "step": 10581 + }, + { + "epoch": 0.24805874458646132, + "grad_norm": 0.4483596682548523, + "learning_rate": 0.00017117919369958093, + "loss": 0.0533, + "step": 10582 + }, + { + "epoch": 0.24808218616126632, + "grad_norm": 0.9054425954818726, + "learning_rate": 0.00017117401958811624, + "loss": 0.1252, + "step": 10583 + }, + { + "epoch": 0.2481056277360713, + "grad_norm": 0.6941733360290527, + "learning_rate": 0.00017116884509045933, + "loss": 0.0873, + "step": 10584 + }, + { + "epoch": 0.2481290693108763, + "grad_norm": 0.8911787271499634, + "learning_rate": 0.0001711636702066383, + "loss": 0.1183, + "step": 10585 + }, + { + "epoch": 0.2481525108856813, + "grad_norm": 0.4402931034564972, + "learning_rate": 0.00017115849493668116, + "loss": 0.1153, + "step": 10586 + }, + { + "epoch": 0.2481759524604863, + "grad_norm": 0.5686041712760925, + "learning_rate": 0.00017115331928061603, + "loss": 0.1251, + "step": 10587 + }, + { + "epoch": 0.2481993940352913, + "grad_norm": 0.5252244472503662, + "learning_rate": 0.000171148143238471, + "loss": 0.1274, + "step": 10588 + }, + { + "epoch": 0.24822283561009628, + "grad_norm": 0.4284338355064392, + "learning_rate": 0.00017114296681027414, + "loss": 0.0755, + "step": 10589 + }, + { + "epoch": 0.24824627718490128, + "grad_norm": 1.158592700958252, + "learning_rate": 0.00017113778999605354, + "loss": 0.1697, + "step": 10590 + }, + { + "epoch": 0.24826971875970627, + "grad_norm": 0.28266435861587524, + "learning_rate": 0.0001711326127958373, + "loss": 0.0622, + "step": 10591 + }, + { + "epoch": 0.24829316033451126, + "grad_norm": 0.41069889068603516, + "learning_rate": 0.0001711274352096535, + "loss": 0.3965, + "step": 10592 + }, + { + "epoch": 0.24831660190931626, + "grad_norm": 0.6347628831863403, + "learning_rate": 0.00017112225723753022, + "loss": 0.1654, + "step": 10593 + }, + { + "epoch": 0.24834004348412125, + "grad_norm": 0.4711586833000183, + "learning_rate": 0.00017111707887949558, + "loss": 0.0582, + "step": 10594 + }, + { + "epoch": 0.24836348505892625, + "grad_norm": 0.5025475025177002, + "learning_rate": 0.0001711119001355777, + "loss": 0.1127, + "step": 10595 + }, + { + "epoch": 0.24838692663373124, + "grad_norm": 0.48622429370880127, + "learning_rate": 0.0001711067210058046, + "loss": 0.1206, + "step": 10596 + }, + { + "epoch": 0.24841036820853624, + "grad_norm": 0.215740367770195, + "learning_rate": 0.00017110154149020446, + "loss": 0.0472, + "step": 10597 + }, + { + "epoch": 0.24843380978334126, + "grad_norm": 0.8246989846229553, + "learning_rate": 0.00017109636158880534, + "loss": 0.2, + "step": 10598 + }, + { + "epoch": 0.24845725135814625, + "grad_norm": 0.6235707998275757, + "learning_rate": 0.00017109118130163539, + "loss": 0.1993, + "step": 10599 + }, + { + "epoch": 0.24848069293295125, + "grad_norm": 0.8760494589805603, + "learning_rate": 0.0001710860006287227, + "loss": 0.1479, + "step": 10600 + }, + { + "epoch": 0.24850413450775624, + "grad_norm": 0.49107927083969116, + "learning_rate": 0.00017108081957009535, + "loss": 0.1605, + "step": 10601 + }, + { + "epoch": 0.24852757608256124, + "grad_norm": 0.842990517616272, + "learning_rate": 0.00017107563812578146, + "loss": 0.1953, + "step": 10602 + }, + { + "epoch": 0.24855101765736623, + "grad_norm": 0.5392521023750305, + "learning_rate": 0.0001710704562958092, + "loss": 0.1046, + "step": 10603 + }, + { + "epoch": 0.24857445923217122, + "grad_norm": 0.13455446064472198, + "learning_rate": 0.0001710652740802066, + "loss": 0.0366, + "step": 10604 + }, + { + "epoch": 0.24859790080697622, + "grad_norm": 0.6409425735473633, + "learning_rate": 0.00017106009147900184, + "loss": 0.1269, + "step": 10605 + }, + { + "epoch": 0.2486213423817812, + "grad_norm": 0.9782203435897827, + "learning_rate": 0.00017105490849222304, + "loss": 0.2051, + "step": 10606 + }, + { + "epoch": 0.2486447839565862, + "grad_norm": 0.6962677240371704, + "learning_rate": 0.00017104972511989828, + "loss": 0.1186, + "step": 10607 + }, + { + "epoch": 0.2486682255313912, + "grad_norm": 0.10424994677305222, + "learning_rate": 0.00017104454136205574, + "loss": 0.0336, + "step": 10608 + }, + { + "epoch": 0.2486916671061962, + "grad_norm": 0.46413207054138184, + "learning_rate": 0.00017103935721872354, + "loss": 0.129, + "step": 10609 + }, + { + "epoch": 0.2487151086810012, + "grad_norm": 0.7216601371765137, + "learning_rate": 0.00017103417268992974, + "loss": 0.1891, + "step": 10610 + }, + { + "epoch": 0.24873855025580618, + "grad_norm": 0.6672570705413818, + "learning_rate": 0.00017102898777570253, + "loss": 0.1355, + "step": 10611 + }, + { + "epoch": 0.24876199183061118, + "grad_norm": 0.7188601493835449, + "learning_rate": 0.00017102380247607007, + "loss": 0.1569, + "step": 10612 + }, + { + "epoch": 0.24878543340541617, + "grad_norm": 0.7458955645561218, + "learning_rate": 0.00017101861679106044, + "loss": 0.1068, + "step": 10613 + }, + { + "epoch": 0.24880887498022117, + "grad_norm": 0.4926430881023407, + "learning_rate": 0.0001710134307207018, + "loss": 0.0841, + "step": 10614 + }, + { + "epoch": 0.24883231655502616, + "grad_norm": 0.46742215752601624, + "learning_rate": 0.0001710082442650223, + "loss": 0.101, + "step": 10615 + }, + { + "epoch": 0.24885575812983116, + "grad_norm": 0.4634242355823517, + "learning_rate": 0.00017100305742405003, + "loss": 0.0848, + "step": 10616 + }, + { + "epoch": 0.24887919970463615, + "grad_norm": 0.31974154710769653, + "learning_rate": 0.0001709978701978132, + "loss": 0.0498, + "step": 10617 + }, + { + "epoch": 0.24890264127944114, + "grad_norm": 0.4560358226299286, + "learning_rate": 0.00017099268258633993, + "loss": 0.1096, + "step": 10618 + }, + { + "epoch": 0.24892608285424614, + "grad_norm": 0.6056327223777771, + "learning_rate": 0.0001709874945896584, + "loss": 0.1501, + "step": 10619 + }, + { + "epoch": 0.24894952442905113, + "grad_norm": 0.3982192277908325, + "learning_rate": 0.0001709823062077967, + "loss": 0.0725, + "step": 10620 + }, + { + "epoch": 0.24897296600385613, + "grad_norm": 0.4251832664012909, + "learning_rate": 0.000170977117440783, + "loss": 0.1093, + "step": 10621 + }, + { + "epoch": 0.24899640757866112, + "grad_norm": 0.6377979516983032, + "learning_rate": 0.00017097192828864546, + "loss": 0.721, + "step": 10622 + }, + { + "epoch": 0.24901984915346612, + "grad_norm": 1.080267310142517, + "learning_rate": 0.00017096673875141227, + "loss": 0.1743, + "step": 10623 + }, + { + "epoch": 0.24904329072827114, + "grad_norm": 0.2893345355987549, + "learning_rate": 0.00017096154882911152, + "loss": 0.035, + "step": 10624 + }, + { + "epoch": 0.24906673230307613, + "grad_norm": 0.42911940813064575, + "learning_rate": 0.00017095635852177146, + "loss": 0.0935, + "step": 10625 + }, + { + "epoch": 0.24909017387788113, + "grad_norm": 0.5791428089141846, + "learning_rate": 0.00017095116782942017, + "loss": 0.1652, + "step": 10626 + }, + { + "epoch": 0.24911361545268612, + "grad_norm": 0.6614278554916382, + "learning_rate": 0.00017094597675208585, + "loss": 0.1566, + "step": 10627 + }, + { + "epoch": 0.24913705702749112, + "grad_norm": 0.35152697563171387, + "learning_rate": 0.0001709407852897967, + "loss": 0.0772, + "step": 10628 + }, + { + "epoch": 0.2491604986022961, + "grad_norm": 0.3639163076877594, + "learning_rate": 0.00017093559344258082, + "loss": 0.0976, + "step": 10629 + }, + { + "epoch": 0.2491839401771011, + "grad_norm": 0.6690696477890015, + "learning_rate": 0.00017093040121046643, + "loss": 0.1353, + "step": 10630 + }, + { + "epoch": 0.2492073817519061, + "grad_norm": 0.5434431433677673, + "learning_rate": 0.00017092520859348167, + "loss": 0.7432, + "step": 10631 + }, + { + "epoch": 0.2492308233267111, + "grad_norm": 0.4994957447052002, + "learning_rate": 0.00017092001559165473, + "loss": 0.0836, + "step": 10632 + }, + { + "epoch": 0.2492542649015161, + "grad_norm": 0.49937403202056885, + "learning_rate": 0.00017091482220501383, + "loss": 0.0681, + "step": 10633 + }, + { + "epoch": 0.24927770647632108, + "grad_norm": 0.5845094919204712, + "learning_rate": 0.00017090962843358708, + "loss": 0.1423, + "step": 10634 + }, + { + "epoch": 0.24930114805112608, + "grad_norm": 0.501728892326355, + "learning_rate": 0.0001709044342774027, + "loss": 0.0571, + "step": 10635 + }, + { + "epoch": 0.24932458962593107, + "grad_norm": 0.8236398100852966, + "learning_rate": 0.00017089923973648887, + "loss": 0.2129, + "step": 10636 + }, + { + "epoch": 0.24934803120073606, + "grad_norm": 0.4349720776081085, + "learning_rate": 0.00017089404481087376, + "loss": 0.1046, + "step": 10637 + }, + { + "epoch": 0.24937147277554106, + "grad_norm": 0.18869327008724213, + "learning_rate": 0.00017088884950058558, + "loss": 0.0429, + "step": 10638 + }, + { + "epoch": 0.24939491435034605, + "grad_norm": 0.3984345495700836, + "learning_rate": 0.0001708836538056525, + "loss": 0.1163, + "step": 10639 + }, + { + "epoch": 0.24941835592515105, + "grad_norm": 0.4336862862110138, + "learning_rate": 0.00017087845772610275, + "loss": 0.066, + "step": 10640 + }, + { + "epoch": 0.24944179749995604, + "grad_norm": 0.2983720898628235, + "learning_rate": 0.00017087326126196447, + "loss": 0.053, + "step": 10641 + }, + { + "epoch": 0.24946523907476104, + "grad_norm": 0.5712759494781494, + "learning_rate": 0.0001708680644132659, + "loss": 0.5147, + "step": 10642 + }, + { + "epoch": 0.24948868064956603, + "grad_norm": 0.786989688873291, + "learning_rate": 0.0001708628671800352, + "loss": 0.1891, + "step": 10643 + }, + { + "epoch": 0.24951212222437102, + "grad_norm": 0.5459268689155579, + "learning_rate": 0.00017085766956230058, + "loss": 0.5931, + "step": 10644 + }, + { + "epoch": 0.24953556379917602, + "grad_norm": 0.7017382383346558, + "learning_rate": 0.00017085247156009028, + "loss": 0.1193, + "step": 10645 + }, + { + "epoch": 0.249559005373981, + "grad_norm": 0.4462102949619293, + "learning_rate": 0.00017084727317343246, + "loss": 0.1182, + "step": 10646 + }, + { + "epoch": 0.249582446948786, + "grad_norm": 0.24757793545722961, + "learning_rate": 0.00017084207440235533, + "loss": 0.0493, + "step": 10647 + }, + { + "epoch": 0.249605888523591, + "grad_norm": 0.5710713267326355, + "learning_rate": 0.00017083687524688715, + "loss": 0.0886, + "step": 10648 + }, + { + "epoch": 0.24962933009839602, + "grad_norm": 0.48243504762649536, + "learning_rate": 0.00017083167570705607, + "loss": 0.1186, + "step": 10649 + }, + { + "epoch": 0.24965277167320102, + "grad_norm": 0.1467808187007904, + "learning_rate": 0.00017082647578289033, + "loss": 0.0216, + "step": 10650 + }, + { + "epoch": 0.249676213248006, + "grad_norm": 0.5795167684555054, + "learning_rate": 0.00017082127547441816, + "loss": 0.1216, + "step": 10651 + }, + { + "epoch": 0.249699654822811, + "grad_norm": 0.5106654167175293, + "learning_rate": 0.00017081607478166771, + "loss": 0.0685, + "step": 10652 + }, + { + "epoch": 0.249723096397616, + "grad_norm": 0.4980693459510803, + "learning_rate": 0.0001708108737046673, + "loss": 0.1176, + "step": 10653 + }, + { + "epoch": 0.249746537972421, + "grad_norm": 0.5598974227905273, + "learning_rate": 0.00017080567224344505, + "loss": 0.171, + "step": 10654 + }, + { + "epoch": 0.249769979547226, + "grad_norm": 0.5298475623130798, + "learning_rate": 0.00017080047039802926, + "loss": 0.6328, + "step": 10655 + }, + { + "epoch": 0.24979342112203098, + "grad_norm": 0.9007928371429443, + "learning_rate": 0.00017079526816844812, + "loss": 0.2215, + "step": 10656 + }, + { + "epoch": 0.24981686269683598, + "grad_norm": 0.6880490183830261, + "learning_rate": 0.00017079006555472987, + "loss": 0.174, + "step": 10657 + }, + { + "epoch": 0.24984030427164097, + "grad_norm": 0.34272557497024536, + "learning_rate": 0.00017078486255690274, + "loss": 0.0767, + "step": 10658 + }, + { + "epoch": 0.24986374584644597, + "grad_norm": 0.3685002028942108, + "learning_rate": 0.00017077965917499494, + "loss": 0.102, + "step": 10659 + }, + { + "epoch": 0.24988718742125096, + "grad_norm": 0.4206659197807312, + "learning_rate": 0.00017077445540903473, + "loss": 0.0951, + "step": 10660 + }, + { + "epoch": 0.24991062899605596, + "grad_norm": 0.568590521812439, + "learning_rate": 0.00017076925125905033, + "loss": 0.0979, + "step": 10661 + }, + { + "epoch": 0.24993407057086095, + "grad_norm": 0.5585440993309021, + "learning_rate": 0.00017076404672506997, + "loss": 0.19, + "step": 10662 + }, + { + "epoch": 0.24995751214566594, + "grad_norm": 0.5310125946998596, + "learning_rate": 0.00017075884180712194, + "loss": 0.1235, + "step": 10663 + }, + { + "epoch": 0.24998095372047094, + "grad_norm": 0.4317227900028229, + "learning_rate": 0.00017075363650523442, + "loss": 0.4722, + "step": 10664 + }, + { + "epoch": 0.25000439529527596, + "grad_norm": 0.669491171836853, + "learning_rate": 0.00017074843081943568, + "loss": 0.1242, + "step": 10665 + }, + { + "epoch": 0.25002783687008096, + "grad_norm": 0.31975439190864563, + "learning_rate": 0.00017074322474975397, + "loss": 0.0885, + "step": 10666 + }, + { + "epoch": 0.25005127844488595, + "grad_norm": 0.18351401388645172, + "learning_rate": 0.00017073801829621752, + "loss": 0.0305, + "step": 10667 + }, + { + "epoch": 0.25007472001969094, + "grad_norm": 0.16031630337238312, + "learning_rate": 0.0001707328114588546, + "loss": 0.0503, + "step": 10668 + }, + { + "epoch": 0.25009816159449594, + "grad_norm": 0.15214857459068298, + "learning_rate": 0.00017072760423769348, + "loss": 0.0241, + "step": 10669 + }, + { + "epoch": 0.25012160316930093, + "grad_norm": 0.4343738853931427, + "learning_rate": 0.00017072239663276236, + "loss": 0.0631, + "step": 10670 + }, + { + "epoch": 0.2501450447441059, + "grad_norm": 0.5515048503875732, + "learning_rate": 0.00017071718864408952, + "loss": 0.1443, + "step": 10671 + }, + { + "epoch": 0.2501684863189109, + "grad_norm": 0.44112536311149597, + "learning_rate": 0.00017071198027170326, + "loss": 0.099, + "step": 10672 + }, + { + "epoch": 0.2501919278937159, + "grad_norm": 0.45185941457748413, + "learning_rate": 0.00017070677151563178, + "loss": 0.1322, + "step": 10673 + }, + { + "epoch": 0.2502153694685209, + "grad_norm": 0.9840788841247559, + "learning_rate": 0.00017070156237590338, + "loss": 0.2479, + "step": 10674 + }, + { + "epoch": 0.2502388110433259, + "grad_norm": 0.7577517628669739, + "learning_rate": 0.0001706963528525463, + "loss": 0.274, + "step": 10675 + }, + { + "epoch": 0.2502622526181309, + "grad_norm": 0.2833043336868286, + "learning_rate": 0.00017069114294558885, + "loss": 0.0345, + "step": 10676 + }, + { + "epoch": 0.2502856941929359, + "grad_norm": 0.5482519268989563, + "learning_rate": 0.00017068593265505926, + "loss": 0.0749, + "step": 10677 + }, + { + "epoch": 0.2503091357677409, + "grad_norm": 0.516278862953186, + "learning_rate": 0.0001706807219809858, + "loss": 0.1384, + "step": 10678 + }, + { + "epoch": 0.2503325773425459, + "grad_norm": 0.6015802621841431, + "learning_rate": 0.00017067551092339676, + "loss": 0.181, + "step": 10679 + }, + { + "epoch": 0.2503560189173509, + "grad_norm": 0.14253903925418854, + "learning_rate": 0.00017067029948232042, + "loss": 0.0347, + "step": 10680 + }, + { + "epoch": 0.25037946049215587, + "grad_norm": 0.7271518111228943, + "learning_rate": 0.00017066508765778505, + "loss": 0.141, + "step": 10681 + }, + { + "epoch": 0.25040290206696086, + "grad_norm": 0.13006529211997986, + "learning_rate": 0.0001706598754498189, + "loss": 0.0316, + "step": 10682 + }, + { + "epoch": 0.25042634364176586, + "grad_norm": 0.6455012559890747, + "learning_rate": 0.00017065466285845032, + "loss": 0.1466, + "step": 10683 + }, + { + "epoch": 0.25044978521657085, + "grad_norm": 0.45180419087409973, + "learning_rate": 0.0001706494498837075, + "loss": 0.1123, + "step": 10684 + }, + { + "epoch": 0.25047322679137585, + "grad_norm": 0.23791635036468506, + "learning_rate": 0.00017064423652561882, + "loss": 0.0483, + "step": 10685 + }, + { + "epoch": 0.25049666836618084, + "grad_norm": 0.5474951863288879, + "learning_rate": 0.00017063902278421252, + "loss": 0.1485, + "step": 10686 + }, + { + "epoch": 0.25052010994098584, + "grad_norm": 0.19580291211605072, + "learning_rate": 0.0001706338086595169, + "loss": 0.0616, + "step": 10687 + }, + { + "epoch": 0.25054355151579083, + "grad_norm": 0.5272239446640015, + "learning_rate": 0.00017062859415156023, + "loss": 0.1139, + "step": 10688 + }, + { + "epoch": 0.2505669930905958, + "grad_norm": 0.24011985957622528, + "learning_rate": 0.0001706233792603708, + "loss": 0.0406, + "step": 10689 + }, + { + "epoch": 0.2505904346654008, + "grad_norm": 0.3089829981327057, + "learning_rate": 0.00017061816398597702, + "loss": 0.0466, + "step": 10690 + }, + { + "epoch": 0.2506138762402058, + "grad_norm": 0.6340471506118774, + "learning_rate": 0.00017061294832840698, + "loss": 0.1557, + "step": 10691 + }, + { + "epoch": 0.2506373178150108, + "grad_norm": 0.34946468472480774, + "learning_rate": 0.00017060773228768917, + "loss": 0.0365, + "step": 10692 + }, + { + "epoch": 0.2506607593898158, + "grad_norm": 1.1691447496414185, + "learning_rate": 0.00017060251586385183, + "loss": 0.3941, + "step": 10693 + }, + { + "epoch": 0.2506842009646208, + "grad_norm": 0.6095716953277588, + "learning_rate": 0.00017059729905692322, + "loss": 0.5415, + "step": 10694 + }, + { + "epoch": 0.2507076425394258, + "grad_norm": 0.1384119838476181, + "learning_rate": 0.00017059208186693167, + "loss": 0.0229, + "step": 10695 + }, + { + "epoch": 0.2507310841142308, + "grad_norm": 0.6138153672218323, + "learning_rate": 0.0001705868642939055, + "loss": 0.0868, + "step": 10696 + }, + { + "epoch": 0.2507545256890358, + "grad_norm": 0.40646278858184814, + "learning_rate": 0.00017058164633787306, + "loss": 0.3204, + "step": 10697 + }, + { + "epoch": 0.2507779672638408, + "grad_norm": 0.4262591600418091, + "learning_rate": 0.0001705764279988626, + "loss": 0.1314, + "step": 10698 + }, + { + "epoch": 0.25080140883864577, + "grad_norm": 0.35958585143089294, + "learning_rate": 0.00017057120927690248, + "loss": 0.0997, + "step": 10699 + }, + { + "epoch": 0.25082485041345076, + "grad_norm": 0.26901987195014954, + "learning_rate": 0.00017056599017202095, + "loss": 0.0752, + "step": 10700 + }, + { + "epoch": 0.25084829198825576, + "grad_norm": 1.5863429307937622, + "learning_rate": 0.0001705607706842464, + "loss": 0.1123, + "step": 10701 + }, + { + "epoch": 0.25087173356306075, + "grad_norm": 0.4479251503944397, + "learning_rate": 0.0001705555508136071, + "loss": 0.1013, + "step": 10702 + }, + { + "epoch": 0.25089517513786574, + "grad_norm": 0.1318780481815338, + "learning_rate": 0.00017055033056013143, + "loss": 0.0406, + "step": 10703 + }, + { + "epoch": 0.25091861671267074, + "grad_norm": 0.12682323157787323, + "learning_rate": 0.00017054510992384767, + "loss": 0.0303, + "step": 10704 + }, + { + "epoch": 0.25094205828747573, + "grad_norm": 0.4386073052883148, + "learning_rate": 0.00017053988890478413, + "loss": 0.4768, + "step": 10705 + }, + { + "epoch": 0.2509654998622807, + "grad_norm": 0.49995288252830505, + "learning_rate": 0.00017053466750296923, + "loss": 0.0284, + "step": 10706 + }, + { + "epoch": 0.2509889414370857, + "grad_norm": 0.34583958983421326, + "learning_rate": 0.00017052944571843118, + "loss": 0.0828, + "step": 10707 + }, + { + "epoch": 0.2510123830118907, + "grad_norm": 1.5490190982818604, + "learning_rate": 0.0001705242235511984, + "loss": 0.2931, + "step": 10708 + }, + { + "epoch": 0.2510358245866957, + "grad_norm": 0.21718615293502808, + "learning_rate": 0.00017051900100129922, + "loss": 0.0619, + "step": 10709 + }, + { + "epoch": 0.2510592661615007, + "grad_norm": 0.6281791925430298, + "learning_rate": 0.0001705137780687619, + "loss": 0.1822, + "step": 10710 + }, + { + "epoch": 0.2510827077363057, + "grad_norm": 0.36641979217529297, + "learning_rate": 0.0001705085547536149, + "loss": 0.0693, + "step": 10711 + }, + { + "epoch": 0.2511061493111107, + "grad_norm": 0.12615320086479187, + "learning_rate": 0.00017050333105588647, + "loss": 0.0257, + "step": 10712 + }, + { + "epoch": 0.25112959088591574, + "grad_norm": 0.5425575375556946, + "learning_rate": 0.000170498106975605, + "loss": 0.1048, + "step": 10713 + }, + { + "epoch": 0.25115303246072074, + "grad_norm": 0.9752046465873718, + "learning_rate": 0.0001704928825127988, + "loss": 0.2542, + "step": 10714 + }, + { + "epoch": 0.25117647403552573, + "grad_norm": 0.5276920795440674, + "learning_rate": 0.00017048765766749626, + "loss": 0.1372, + "step": 10715 + }, + { + "epoch": 0.2511999156103307, + "grad_norm": 0.26078668236732483, + "learning_rate": 0.0001704824324397257, + "loss": 0.0485, + "step": 10716 + }, + { + "epoch": 0.2512233571851357, + "grad_norm": 0.38182884454727173, + "learning_rate": 0.00017047720682951547, + "loss": 0.0911, + "step": 10717 + }, + { + "epoch": 0.2512467987599407, + "grad_norm": 0.421579509973526, + "learning_rate": 0.00017047198083689394, + "loss": 0.1036, + "step": 10718 + }, + { + "epoch": 0.2512702403347457, + "grad_norm": 0.16791272163391113, + "learning_rate": 0.00017046675446188948, + "loss": 0.0449, + "step": 10719 + }, + { + "epoch": 0.2512936819095507, + "grad_norm": 0.6259785890579224, + "learning_rate": 0.0001704615277045304, + "loss": 0.1361, + "step": 10720 + }, + { + "epoch": 0.2513171234843557, + "grad_norm": 0.3691766560077667, + "learning_rate": 0.0001704563005648451, + "loss": 0.0486, + "step": 10721 + }, + { + "epoch": 0.2513405650591607, + "grad_norm": 0.5143055319786072, + "learning_rate": 0.00017045107304286192, + "loss": 0.1134, + "step": 10722 + }, + { + "epoch": 0.2513640066339657, + "grad_norm": 0.5800526738166809, + "learning_rate": 0.00017044584513860924, + "loss": 0.7234, + "step": 10723 + }, + { + "epoch": 0.2513874482087707, + "grad_norm": 0.7014310956001282, + "learning_rate": 0.00017044061685211542, + "loss": 0.598, + "step": 10724 + }, + { + "epoch": 0.2514108897835757, + "grad_norm": 0.40805840492248535, + "learning_rate": 0.00017043538818340884, + "loss": 0.0797, + "step": 10725 + }, + { + "epoch": 0.25143433135838067, + "grad_norm": 0.27736425399780273, + "learning_rate": 0.00017043015913251788, + "loss": 0.0646, + "step": 10726 + }, + { + "epoch": 0.25145777293318566, + "grad_norm": 0.6511843204498291, + "learning_rate": 0.00017042492969947088, + "loss": 0.1518, + "step": 10727 + }, + { + "epoch": 0.25148121450799066, + "grad_norm": 0.3574136197566986, + "learning_rate": 0.0001704196998842962, + "loss": 0.0627, + "step": 10728 + }, + { + "epoch": 0.25150465608279565, + "grad_norm": 0.4758891463279724, + "learning_rate": 0.00017041446968702232, + "loss": 0.104, + "step": 10729 + }, + { + "epoch": 0.25152809765760065, + "grad_norm": 0.8733404278755188, + "learning_rate": 0.00017040923910767748, + "loss": 0.0644, + "step": 10730 + }, + { + "epoch": 0.25155153923240564, + "grad_norm": 0.6787099838256836, + "learning_rate": 0.00017040400814629018, + "loss": 0.202, + "step": 10731 + }, + { + "epoch": 0.25157498080721064, + "grad_norm": 0.3646237850189209, + "learning_rate": 0.00017039877680288871, + "loss": 0.1006, + "step": 10732 + }, + { + "epoch": 0.25159842238201563, + "grad_norm": 0.6053150296211243, + "learning_rate": 0.00017039354507750155, + "loss": 0.14, + "step": 10733 + }, + { + "epoch": 0.2516218639568206, + "grad_norm": 0.2821817398071289, + "learning_rate": 0.00017038831297015697, + "loss": 0.0687, + "step": 10734 + }, + { + "epoch": 0.2516453055316256, + "grad_norm": 0.7038359045982361, + "learning_rate": 0.00017038308048088348, + "loss": 0.7277, + "step": 10735 + }, + { + "epoch": 0.2516687471064306, + "grad_norm": 0.3544371724128723, + "learning_rate": 0.0001703778476097094, + "loss": 0.1229, + "step": 10736 + }, + { + "epoch": 0.2516921886812356, + "grad_norm": 0.6090645790100098, + "learning_rate": 0.0001703726143566631, + "loss": 0.1104, + "step": 10737 + }, + { + "epoch": 0.2517156302560406, + "grad_norm": 0.6385695934295654, + "learning_rate": 0.00017036738072177308, + "loss": 0.5954, + "step": 10738 + }, + { + "epoch": 0.2517390718308456, + "grad_norm": 0.6686074137687683, + "learning_rate": 0.00017036214670506763, + "loss": 0.0908, + "step": 10739 + }, + { + "epoch": 0.2517625134056506, + "grad_norm": 0.44266417622566223, + "learning_rate": 0.0001703569123065752, + "loss": 0.5311, + "step": 10740 + }, + { + "epoch": 0.2517859549804556, + "grad_norm": 1.4697191715240479, + "learning_rate": 0.00017035167752632422, + "loss": 0.1901, + "step": 10741 + }, + { + "epoch": 0.2518093965552606, + "grad_norm": 0.942637026309967, + "learning_rate": 0.000170346442364343, + "loss": 0.2839, + "step": 10742 + }, + { + "epoch": 0.2518328381300656, + "grad_norm": 0.4396228492259979, + "learning_rate": 0.00017034120682066006, + "loss": 0.0752, + "step": 10743 + }, + { + "epoch": 0.25185627970487057, + "grad_norm": 0.1955181062221527, + "learning_rate": 0.00017033597089530372, + "loss": 0.0604, + "step": 10744 + }, + { + "epoch": 0.25187972127967556, + "grad_norm": 0.8648996949195862, + "learning_rate": 0.0001703307345883024, + "loss": 0.2742, + "step": 10745 + }, + { + "epoch": 0.25190316285448056, + "grad_norm": 0.33375078439712524, + "learning_rate": 0.00017032549789968462, + "loss": 0.0782, + "step": 10746 + }, + { + "epoch": 0.25192660442928555, + "grad_norm": 0.13982060551643372, + "learning_rate": 0.00017032026082947866, + "loss": 0.031, + "step": 10747 + }, + { + "epoch": 0.25195004600409054, + "grad_norm": 0.5543082356452942, + "learning_rate": 0.00017031502337771297, + "loss": 0.1173, + "step": 10748 + }, + { + "epoch": 0.25197348757889554, + "grad_norm": 0.1604231297969818, + "learning_rate": 0.000170309785544416, + "loss": 0.0418, + "step": 10749 + }, + { + "epoch": 0.25199692915370053, + "grad_norm": 0.5387109518051147, + "learning_rate": 0.00017030454732961617, + "loss": 0.1557, + "step": 10750 + }, + { + "epoch": 0.2520203707285055, + "grad_norm": 0.1648717224597931, + "learning_rate": 0.0001702993087333419, + "loss": 0.0312, + "step": 10751 + }, + { + "epoch": 0.2520438123033105, + "grad_norm": 0.9868321418762207, + "learning_rate": 0.0001702940697556216, + "loss": 0.1336, + "step": 10752 + }, + { + "epoch": 0.2520672538781155, + "grad_norm": 0.15392856299877167, + "learning_rate": 0.00017028883039648366, + "loss": 0.0398, + "step": 10753 + }, + { + "epoch": 0.2520906954529205, + "grad_norm": 0.23485824465751648, + "learning_rate": 0.0001702835906559566, + "loss": 0.0376, + "step": 10754 + }, + { + "epoch": 0.2521141370277255, + "grad_norm": 0.6086127758026123, + "learning_rate": 0.00017027835053406876, + "loss": 0.1382, + "step": 10755 + }, + { + "epoch": 0.2521375786025305, + "grad_norm": 0.7465330362319946, + "learning_rate": 0.00017027311003084864, + "loss": 0.0519, + "step": 10756 + }, + { + "epoch": 0.2521610201773355, + "grad_norm": 0.5147605538368225, + "learning_rate": 0.00017026786914632464, + "loss": 0.1348, + "step": 10757 + }, + { + "epoch": 0.2521844617521405, + "grad_norm": 0.4777024984359741, + "learning_rate": 0.0001702626278805252, + "loss": 0.1264, + "step": 10758 + }, + { + "epoch": 0.2522079033269455, + "grad_norm": 0.4738243520259857, + "learning_rate": 0.00017025738623347879, + "loss": 0.453, + "step": 10759 + }, + { + "epoch": 0.2522313449017505, + "grad_norm": 0.5944973826408386, + "learning_rate": 0.0001702521442052138, + "loss": 0.1718, + "step": 10760 + }, + { + "epoch": 0.25225478647655547, + "grad_norm": 1.0043946504592896, + "learning_rate": 0.0001702469017957587, + "loss": 0.3628, + "step": 10761 + }, + { + "epoch": 0.25227822805136046, + "grad_norm": 0.3509388267993927, + "learning_rate": 0.00017024165900514195, + "loss": 0.0677, + "step": 10762 + }, + { + "epoch": 0.25230166962616546, + "grad_norm": 0.4740431010723114, + "learning_rate": 0.00017023641583339199, + "loss": 0.08, + "step": 10763 + }, + { + "epoch": 0.2523251112009705, + "grad_norm": 0.36652520298957825, + "learning_rate": 0.00017023117228053724, + "loss": 0.0702, + "step": 10764 + }, + { + "epoch": 0.2523485527757755, + "grad_norm": 0.41480299830436707, + "learning_rate": 0.0001702259283466062, + "loss": 0.0914, + "step": 10765 + }, + { + "epoch": 0.2523719943505805, + "grad_norm": 0.44897255301475525, + "learning_rate": 0.0001702206840316273, + "loss": 0.1087, + "step": 10766 + }, + { + "epoch": 0.2523954359253855, + "grad_norm": 0.2293419986963272, + "learning_rate": 0.00017021543933562895, + "loss": 0.0404, + "step": 10767 + }, + { + "epoch": 0.2524188775001905, + "grad_norm": 0.5508582592010498, + "learning_rate": 0.00017021019425863968, + "loss": 0.1069, + "step": 10768 + }, + { + "epoch": 0.2524423190749955, + "grad_norm": 0.46842843294143677, + "learning_rate": 0.00017020494880068792, + "loss": 0.0949, + "step": 10769 + }, + { + "epoch": 0.2524657606498005, + "grad_norm": 0.755721926689148, + "learning_rate": 0.00017019970296180213, + "loss": 0.1442, + "step": 10770 + }, + { + "epoch": 0.25248920222460547, + "grad_norm": 0.18383066356182098, + "learning_rate": 0.0001701944567420108, + "loss": 0.0289, + "step": 10771 + }, + { + "epoch": 0.25251264379941046, + "grad_norm": 0.5512399077415466, + "learning_rate": 0.00017018921014134238, + "loss": 0.6512, + "step": 10772 + }, + { + "epoch": 0.25253608537421546, + "grad_norm": 0.20186150074005127, + "learning_rate": 0.0001701839631598253, + "loss": 0.0331, + "step": 10773 + }, + { + "epoch": 0.25255952694902045, + "grad_norm": 0.4546481668949127, + "learning_rate": 0.00017017871579748806, + "loss": 0.0727, + "step": 10774 + }, + { + "epoch": 0.25258296852382545, + "grad_norm": 0.1849886029958725, + "learning_rate": 0.00017017346805435913, + "loss": 0.0414, + "step": 10775 + }, + { + "epoch": 0.25260641009863044, + "grad_norm": 0.4716874957084656, + "learning_rate": 0.00017016821993046704, + "loss": 0.103, + "step": 10776 + }, + { + "epoch": 0.25262985167343543, + "grad_norm": 0.43852517008781433, + "learning_rate": 0.0001701629714258402, + "loss": 0.0958, + "step": 10777 + }, + { + "epoch": 0.25265329324824043, + "grad_norm": 0.42055565118789673, + "learning_rate": 0.00017015772254050707, + "loss": 0.0748, + "step": 10778 + }, + { + "epoch": 0.2526767348230454, + "grad_norm": 0.5909906625747681, + "learning_rate": 0.00017015247327449617, + "loss": 0.1342, + "step": 10779 + }, + { + "epoch": 0.2527001763978504, + "grad_norm": 0.6791220307350159, + "learning_rate": 0.000170147223627836, + "loss": 0.1961, + "step": 10780 + }, + { + "epoch": 0.2527236179726554, + "grad_norm": 0.2099103182554245, + "learning_rate": 0.000170141973600555, + "loss": 0.04, + "step": 10781 + }, + { + "epoch": 0.2527470595474604, + "grad_norm": 0.4243866801261902, + "learning_rate": 0.00017013672319268167, + "loss": 0.0853, + "step": 10782 + }, + { + "epoch": 0.2527705011222654, + "grad_norm": 0.3000257909297943, + "learning_rate": 0.00017013147240424456, + "loss": 0.0635, + "step": 10783 + }, + { + "epoch": 0.2527939426970704, + "grad_norm": 0.6770594120025635, + "learning_rate": 0.00017012622123527206, + "loss": 0.2059, + "step": 10784 + }, + { + "epoch": 0.2528173842718754, + "grad_norm": 0.34519702196121216, + "learning_rate": 0.00017012096968579274, + "loss": 0.0719, + "step": 10785 + }, + { + "epoch": 0.2528408258466804, + "grad_norm": 0.5898627042770386, + "learning_rate": 0.00017011571775583505, + "loss": 0.1071, + "step": 10786 + }, + { + "epoch": 0.2528642674214854, + "grad_norm": 0.14636388421058655, + "learning_rate": 0.00017011046544542752, + "loss": 0.0305, + "step": 10787 + }, + { + "epoch": 0.2528877089962904, + "grad_norm": 0.6617730855941772, + "learning_rate": 0.00017010521275459862, + "loss": 0.1262, + "step": 10788 + }, + { + "epoch": 0.25291115057109537, + "grad_norm": 0.9244785308837891, + "learning_rate": 0.00017009995968337687, + "loss": 0.1462, + "step": 10789 + }, + { + "epoch": 0.25293459214590036, + "grad_norm": 0.6526480317115784, + "learning_rate": 0.00017009470623179077, + "loss": 0.1618, + "step": 10790 + }, + { + "epoch": 0.25295803372070536, + "grad_norm": 0.9917035698890686, + "learning_rate": 0.0001700894523998688, + "loss": 0.2418, + "step": 10791 + }, + { + "epoch": 0.25298147529551035, + "grad_norm": 1.0856308937072754, + "learning_rate": 0.0001700841981876395, + "loss": 0.2039, + "step": 10792 + }, + { + "epoch": 0.25300491687031534, + "grad_norm": 1.1061698198318481, + "learning_rate": 0.0001700789435951314, + "loss": 0.0991, + "step": 10793 + }, + { + "epoch": 0.25302835844512034, + "grad_norm": 0.6766337156295776, + "learning_rate": 0.00017007368862237296, + "loss": 0.0795, + "step": 10794 + }, + { + "epoch": 0.25305180001992533, + "grad_norm": 0.1279618889093399, + "learning_rate": 0.00017006843326939272, + "loss": 0.0198, + "step": 10795 + }, + { + "epoch": 0.2530752415947303, + "grad_norm": 0.5076586008071899, + "learning_rate": 0.00017006317753621918, + "loss": 0.0861, + "step": 10796 + }, + { + "epoch": 0.2530986831695353, + "grad_norm": 0.6749037504196167, + "learning_rate": 0.00017005792142288088, + "loss": 0.5622, + "step": 10797 + }, + { + "epoch": 0.2531221247443403, + "grad_norm": 0.19386139512062073, + "learning_rate": 0.0001700526649294063, + "loss": 0.0407, + "step": 10798 + }, + { + "epoch": 0.2531455663191453, + "grad_norm": 0.4415215849876404, + "learning_rate": 0.00017004740805582404, + "loss": 0.1052, + "step": 10799 + }, + { + "epoch": 0.2531690078939503, + "grad_norm": 0.38773927092552185, + "learning_rate": 0.00017004215080216256, + "loss": 0.0849, + "step": 10800 + }, + { + "epoch": 0.2531924494687553, + "grad_norm": 0.7041864991188049, + "learning_rate": 0.00017003689316845037, + "loss": 0.743, + "step": 10801 + }, + { + "epoch": 0.2532158910435603, + "grad_norm": 0.129940465092659, + "learning_rate": 0.00017003163515471605, + "loss": 0.0245, + "step": 10802 + }, + { + "epoch": 0.2532393326183653, + "grad_norm": 0.6043221354484558, + "learning_rate": 0.00017002637676098812, + "loss": 0.0928, + "step": 10803 + }, + { + "epoch": 0.2532627741931703, + "grad_norm": 0.20916225016117096, + "learning_rate": 0.00017002111798729508, + "loss": 0.0557, + "step": 10804 + }, + { + "epoch": 0.2532862157679753, + "grad_norm": 0.4833003580570221, + "learning_rate": 0.00017001585883366548, + "loss": 0.1079, + "step": 10805 + }, + { + "epoch": 0.25330965734278027, + "grad_norm": 0.6594139337539673, + "learning_rate": 0.00017001059930012787, + "loss": 0.1091, + "step": 10806 + }, + { + "epoch": 0.25333309891758526, + "grad_norm": 0.48159554600715637, + "learning_rate": 0.0001700053393867108, + "loss": 0.1601, + "step": 10807 + }, + { + "epoch": 0.25335654049239026, + "grad_norm": 0.5486742258071899, + "learning_rate": 0.00017000007909344274, + "loss": 0.0844, + "step": 10808 + }, + { + "epoch": 0.25337998206719525, + "grad_norm": 0.7064542174339294, + "learning_rate": 0.00016999481842035238, + "loss": 0.1989, + "step": 10809 + }, + { + "epoch": 0.25340342364200025, + "grad_norm": 0.4134048819541931, + "learning_rate": 0.00016998955736746804, + "loss": 0.075, + "step": 10810 + }, + { + "epoch": 0.25342686521680524, + "grad_norm": 0.08845259994268417, + "learning_rate": 0.00016998429593481848, + "loss": 0.0185, + "step": 10811 + }, + { + "epoch": 0.25345030679161024, + "grad_norm": 0.4897208511829376, + "learning_rate": 0.00016997903412243215, + "loss": 0.0841, + "step": 10812 + }, + { + "epoch": 0.25347374836641523, + "grad_norm": 0.35026004910469055, + "learning_rate": 0.0001699737719303376, + "loss": 0.0558, + "step": 10813 + }, + { + "epoch": 0.2534971899412202, + "grad_norm": 0.3289344906806946, + "learning_rate": 0.0001699685093585634, + "loss": 0.1006, + "step": 10814 + }, + { + "epoch": 0.2535206315160252, + "grad_norm": 0.07457751035690308, + "learning_rate": 0.00016996324640713807, + "loss": 0.0068, + "step": 10815 + }, + { + "epoch": 0.25354407309083027, + "grad_norm": 0.5287301540374756, + "learning_rate": 0.00016995798307609025, + "loss": 0.127, + "step": 10816 + }, + { + "epoch": 0.25356751466563526, + "grad_norm": 0.4465089440345764, + "learning_rate": 0.0001699527193654484, + "loss": 0.1422, + "step": 10817 + }, + { + "epoch": 0.25359095624044026, + "grad_norm": 0.4926787316799164, + "learning_rate": 0.00016994745527524116, + "loss": 0.1287, + "step": 10818 + }, + { + "epoch": 0.25361439781524525, + "grad_norm": 0.33587098121643066, + "learning_rate": 0.00016994219080549707, + "loss": 0.133, + "step": 10819 + }, + { + "epoch": 0.25363783939005025, + "grad_norm": 0.7113810181617737, + "learning_rate": 0.00016993692595624466, + "loss": 0.15, + "step": 10820 + }, + { + "epoch": 0.25366128096485524, + "grad_norm": 0.32994547486305237, + "learning_rate": 0.00016993166072751253, + "loss": 0.0457, + "step": 10821 + }, + { + "epoch": 0.25368472253966023, + "grad_norm": 0.4280984401702881, + "learning_rate": 0.00016992639511932925, + "loss": 0.0767, + "step": 10822 + }, + { + "epoch": 0.25370816411446523, + "grad_norm": 0.392612099647522, + "learning_rate": 0.00016992112913172337, + "loss": 0.1337, + "step": 10823 + }, + { + "epoch": 0.2537316056892702, + "grad_norm": 0.7353304624557495, + "learning_rate": 0.0001699158627647235, + "loss": 0.1701, + "step": 10824 + }, + { + "epoch": 0.2537550472640752, + "grad_norm": 0.33832573890686035, + "learning_rate": 0.00016991059601835818, + "loss": 0.0843, + "step": 10825 + }, + { + "epoch": 0.2537784888388802, + "grad_norm": 0.44659844040870667, + "learning_rate": 0.00016990532889265598, + "loss": 0.1101, + "step": 10826 + }, + { + "epoch": 0.2538019304136852, + "grad_norm": 0.6642935276031494, + "learning_rate": 0.00016990006138764555, + "loss": 0.1407, + "step": 10827 + }, + { + "epoch": 0.2538253719884902, + "grad_norm": 0.512456476688385, + "learning_rate": 0.00016989479350335536, + "loss": 0.1261, + "step": 10828 + }, + { + "epoch": 0.2538488135632952, + "grad_norm": 0.674625813961029, + "learning_rate": 0.0001698895252398141, + "loss": 0.1797, + "step": 10829 + }, + { + "epoch": 0.2538722551381002, + "grad_norm": 0.23840954899787903, + "learning_rate": 0.0001698842565970503, + "loss": 0.0408, + "step": 10830 + }, + { + "epoch": 0.2538956967129052, + "grad_norm": 0.3779658377170563, + "learning_rate": 0.00016987898757509252, + "loss": 0.4342, + "step": 10831 + }, + { + "epoch": 0.2539191382877102, + "grad_norm": 0.6861881613731384, + "learning_rate": 0.00016987371817396945, + "loss": 0.1619, + "step": 10832 + }, + { + "epoch": 0.25394257986251517, + "grad_norm": 0.5887826085090637, + "learning_rate": 0.00016986844839370958, + "loss": 0.1924, + "step": 10833 + }, + { + "epoch": 0.25396602143732017, + "grad_norm": 0.6329207420349121, + "learning_rate": 0.00016986317823434156, + "loss": 0.1445, + "step": 10834 + }, + { + "epoch": 0.25398946301212516, + "grad_norm": 0.9351248741149902, + "learning_rate": 0.000169857907695894, + "loss": 0.1326, + "step": 10835 + }, + { + "epoch": 0.25401290458693015, + "grad_norm": 0.3991246819496155, + "learning_rate": 0.00016985263677839543, + "loss": 0.0749, + "step": 10836 + }, + { + "epoch": 0.25403634616173515, + "grad_norm": 0.22621022164821625, + "learning_rate": 0.0001698473654818745, + "loss": 0.0563, + "step": 10837 + }, + { + "epoch": 0.25405978773654014, + "grad_norm": 0.3985416889190674, + "learning_rate": 0.00016984209380635978, + "loss": 0.0537, + "step": 10838 + }, + { + "epoch": 0.25408322931134514, + "grad_norm": 0.24973812699317932, + "learning_rate": 0.00016983682175187993, + "loss": 0.0784, + "step": 10839 + }, + { + "epoch": 0.25410667088615013, + "grad_norm": 0.3026748299598694, + "learning_rate": 0.00016983154931846347, + "loss": 0.0619, + "step": 10840 + }, + { + "epoch": 0.2541301124609551, + "grad_norm": 0.4377194941043854, + "learning_rate": 0.0001698262765061391, + "loss": 0.1161, + "step": 10841 + }, + { + "epoch": 0.2541535540357601, + "grad_norm": 0.45535656809806824, + "learning_rate": 0.0001698210033149354, + "loss": 0.0665, + "step": 10842 + }, + { + "epoch": 0.2541769956105651, + "grad_norm": 0.6035432815551758, + "learning_rate": 0.00016981572974488092, + "loss": 0.129, + "step": 10843 + }, + { + "epoch": 0.2542004371853701, + "grad_norm": 0.6195052862167358, + "learning_rate": 0.00016981045579600438, + "loss": 0.8056, + "step": 10844 + }, + { + "epoch": 0.2542238787601751, + "grad_norm": 0.1760999858379364, + "learning_rate": 0.00016980518146833434, + "loss": 0.022, + "step": 10845 + }, + { + "epoch": 0.2542473203349801, + "grad_norm": 0.20951835811138153, + "learning_rate": 0.00016979990676189942, + "loss": 0.033, + "step": 10846 + }, + { + "epoch": 0.2542707619097851, + "grad_norm": 0.38403815031051636, + "learning_rate": 0.00016979463167672823, + "loss": 0.0742, + "step": 10847 + }, + { + "epoch": 0.2542942034845901, + "grad_norm": 0.34929317235946655, + "learning_rate": 0.00016978935621284942, + "loss": 0.117, + "step": 10848 + }, + { + "epoch": 0.2543176450593951, + "grad_norm": 0.39825841784477234, + "learning_rate": 0.00016978408037029156, + "loss": 0.0695, + "step": 10849 + }, + { + "epoch": 0.2543410866342001, + "grad_norm": 1.007974624633789, + "learning_rate": 0.00016977880414908335, + "loss": 0.2148, + "step": 10850 + }, + { + "epoch": 0.25436452820900507, + "grad_norm": 0.6473968029022217, + "learning_rate": 0.0001697735275492534, + "loss": 0.0946, + "step": 10851 + }, + { + "epoch": 0.25438796978381006, + "grad_norm": 0.6920773386955261, + "learning_rate": 0.00016976825057083032, + "loss": 0.0991, + "step": 10852 + }, + { + "epoch": 0.25441141135861506, + "grad_norm": 0.32781293988227844, + "learning_rate": 0.00016976297321384274, + "loss": 0.0704, + "step": 10853 + }, + { + "epoch": 0.25443485293342005, + "grad_norm": 0.18031007051467896, + "learning_rate": 0.0001697576954783193, + "loss": 0.0248, + "step": 10854 + }, + { + "epoch": 0.25445829450822505, + "grad_norm": 0.3812943994998932, + "learning_rate": 0.00016975241736428868, + "loss": 0.064, + "step": 10855 + }, + { + "epoch": 0.25448173608303004, + "grad_norm": 0.5090324878692627, + "learning_rate": 0.00016974713887177943, + "loss": 0.044, + "step": 10856 + }, + { + "epoch": 0.25450517765783504, + "grad_norm": 0.4194522500038147, + "learning_rate": 0.00016974186000082029, + "loss": 0.1157, + "step": 10857 + }, + { + "epoch": 0.25452861923264003, + "grad_norm": 0.1743794083595276, + "learning_rate": 0.0001697365807514398, + "loss": 0.0286, + "step": 10858 + }, + { + "epoch": 0.254552060807445, + "grad_norm": 0.2044985294342041, + "learning_rate": 0.0001697313011236667, + "loss": 0.0596, + "step": 10859 + }, + { + "epoch": 0.25457550238225, + "grad_norm": 0.36838507652282715, + "learning_rate": 0.0001697260211175296, + "loss": 0.0654, + "step": 10860 + }, + { + "epoch": 0.254598943957055, + "grad_norm": 0.25147905945777893, + "learning_rate": 0.00016972074073305716, + "loss": 0.0413, + "step": 10861 + }, + { + "epoch": 0.25462238553186, + "grad_norm": 0.3446400463581085, + "learning_rate": 0.00016971545997027804, + "loss": 0.0714, + "step": 10862 + }, + { + "epoch": 0.254645827106665, + "grad_norm": 0.46219196915626526, + "learning_rate": 0.0001697101788292208, + "loss": 0.1068, + "step": 10863 + }, + { + "epoch": 0.25466926868147, + "grad_norm": 0.8302043080329895, + "learning_rate": 0.0001697048973099142, + "loss": 0.1317, + "step": 10864 + }, + { + "epoch": 0.254692710256275, + "grad_norm": 1.7137349843978882, + "learning_rate": 0.0001696996154123869, + "loss": 0.1588, + "step": 10865 + }, + { + "epoch": 0.25471615183108, + "grad_norm": 0.5802058577537537, + "learning_rate": 0.0001696943331366675, + "loss": 0.5046, + "step": 10866 + }, + { + "epoch": 0.25473959340588503, + "grad_norm": 0.614695131778717, + "learning_rate": 0.0001696890504827847, + "loss": 0.0826, + "step": 10867 + }, + { + "epoch": 0.25476303498069003, + "grad_norm": 0.9264271855354309, + "learning_rate": 0.00016968376745076712, + "loss": 0.1353, + "step": 10868 + }, + { + "epoch": 0.254786476555495, + "grad_norm": 0.846402645111084, + "learning_rate": 0.0001696784840406435, + "loss": 0.2375, + "step": 10869 + }, + { + "epoch": 0.2548099181303, + "grad_norm": 0.20108488202095032, + "learning_rate": 0.00016967320025244244, + "loss": 0.0169, + "step": 10870 + }, + { + "epoch": 0.254833359705105, + "grad_norm": 0.7494037747383118, + "learning_rate": 0.00016966791608619266, + "loss": 0.0975, + "step": 10871 + }, + { + "epoch": 0.25485680127991, + "grad_norm": 0.5905589461326599, + "learning_rate": 0.0001696626315419228, + "loss": 0.1086, + "step": 10872 + }, + { + "epoch": 0.254880242854715, + "grad_norm": 0.2706553637981415, + "learning_rate": 0.00016965734661966153, + "loss": 0.0765, + "step": 10873 + }, + { + "epoch": 0.25490368442952, + "grad_norm": 0.4182336628437042, + "learning_rate": 0.00016965206131943755, + "loss": 0.0704, + "step": 10874 + }, + { + "epoch": 0.254927126004325, + "grad_norm": 0.38037627935409546, + "learning_rate": 0.0001696467756412795, + "loss": 0.0534, + "step": 10875 + }, + { + "epoch": 0.25495056757913, + "grad_norm": 0.43059638142585754, + "learning_rate": 0.00016964148958521612, + "loss": 0.0896, + "step": 10876 + }, + { + "epoch": 0.254974009153935, + "grad_norm": 0.41634872555732727, + "learning_rate": 0.00016963620315127605, + "loss": 0.0799, + "step": 10877 + }, + { + "epoch": 0.25499745072873997, + "grad_norm": 0.6033285856246948, + "learning_rate": 0.00016963091633948797, + "loss": 0.1109, + "step": 10878 + }, + { + "epoch": 0.25502089230354497, + "grad_norm": 0.4802660644054413, + "learning_rate": 0.0001696256291498806, + "loss": 0.0987, + "step": 10879 + }, + { + "epoch": 0.25504433387834996, + "grad_norm": 0.2575869560241699, + "learning_rate": 0.0001696203415824826, + "loss": 0.0519, + "step": 10880 + }, + { + "epoch": 0.25506777545315495, + "grad_norm": 0.12474773079156876, + "learning_rate": 0.0001696150536373227, + "loss": 0.0296, + "step": 10881 + }, + { + "epoch": 0.25509121702795995, + "grad_norm": 0.555296003818512, + "learning_rate": 0.00016960976531442951, + "loss": 0.4063, + "step": 10882 + }, + { + "epoch": 0.25511465860276494, + "grad_norm": 0.5506271719932556, + "learning_rate": 0.00016960447661383177, + "loss": 0.1318, + "step": 10883 + }, + { + "epoch": 0.25513810017756994, + "grad_norm": 0.40723946690559387, + "learning_rate": 0.00016959918753555822, + "loss": 0.1101, + "step": 10884 + }, + { + "epoch": 0.25516154175237493, + "grad_norm": 0.21971996128559113, + "learning_rate": 0.0001695938980796375, + "loss": 0.0402, + "step": 10885 + }, + { + "epoch": 0.2551849833271799, + "grad_norm": 0.5122407078742981, + "learning_rate": 0.00016958860824609838, + "loss": 0.1234, + "step": 10886 + }, + { + "epoch": 0.2552084249019849, + "grad_norm": 0.43648555874824524, + "learning_rate": 0.0001695833180349695, + "loss": 0.1377, + "step": 10887 + }, + { + "epoch": 0.2552318664767899, + "grad_norm": 0.38132917881011963, + "learning_rate": 0.00016957802744627952, + "loss": 0.0611, + "step": 10888 + }, + { + "epoch": 0.2552553080515949, + "grad_norm": 0.7640248537063599, + "learning_rate": 0.00016957273648005727, + "loss": 0.1519, + "step": 10889 + }, + { + "epoch": 0.2552787496263999, + "grad_norm": 0.6134025454521179, + "learning_rate": 0.00016956744513633136, + "loss": 0.1059, + "step": 10890 + }, + { + "epoch": 0.2553021912012049, + "grad_norm": 0.7808334827423096, + "learning_rate": 0.00016956215341513056, + "loss": 0.3314, + "step": 10891 + }, + { + "epoch": 0.2553256327760099, + "grad_norm": 0.530225396156311, + "learning_rate": 0.00016955686131648354, + "loss": 0.0921, + "step": 10892 + }, + { + "epoch": 0.2553490743508149, + "grad_norm": 0.838852047920227, + "learning_rate": 0.00016955156884041904, + "loss": 0.188, + "step": 10893 + }, + { + "epoch": 0.2553725159256199, + "grad_norm": 0.5917839407920837, + "learning_rate": 0.00016954627598696578, + "loss": 0.5329, + "step": 10894 + }, + { + "epoch": 0.2553959575004249, + "grad_norm": 0.545798659324646, + "learning_rate": 0.00016954098275615245, + "loss": 0.0631, + "step": 10895 + }, + { + "epoch": 0.25541939907522987, + "grad_norm": 0.28213661909103394, + "learning_rate": 0.00016953568914800782, + "loss": 0.0607, + "step": 10896 + }, + { + "epoch": 0.25544284065003486, + "grad_norm": 0.7677700519561768, + "learning_rate": 0.00016953039516256056, + "loss": 0.1639, + "step": 10897 + }, + { + "epoch": 0.25546628222483986, + "grad_norm": 0.7905531525611877, + "learning_rate": 0.00016952510079983944, + "loss": 0.0877, + "step": 10898 + }, + { + "epoch": 0.25548972379964485, + "grad_norm": 0.09886692464351654, + "learning_rate": 0.00016951980605987317, + "loss": 0.023, + "step": 10899 + }, + { + "epoch": 0.25551316537444985, + "grad_norm": 0.18446624279022217, + "learning_rate": 0.00016951451094269045, + "loss": 0.0354, + "step": 10900 + }, + { + "epoch": 0.25553660694925484, + "grad_norm": 0.736690104007721, + "learning_rate": 0.00016950921544832008, + "loss": 0.1144, + "step": 10901 + }, + { + "epoch": 0.25556004852405984, + "grad_norm": 0.33129218220710754, + "learning_rate": 0.00016950391957679074, + "loss": 0.0655, + "step": 10902 + }, + { + "epoch": 0.25558349009886483, + "grad_norm": 0.07764976471662521, + "learning_rate": 0.00016949862332813116, + "loss": 0.0201, + "step": 10903 + }, + { + "epoch": 0.2556069316736698, + "grad_norm": 0.3156512975692749, + "learning_rate": 0.0001694933267023701, + "loss": 0.0558, + "step": 10904 + }, + { + "epoch": 0.2556303732484748, + "grad_norm": 0.7024522423744202, + "learning_rate": 0.0001694880296995363, + "loss": 0.1277, + "step": 10905 + }, + { + "epoch": 0.2556538148232798, + "grad_norm": 0.48472756147384644, + "learning_rate": 0.00016948273231965852, + "loss": 0.0692, + "step": 10906 + }, + { + "epoch": 0.2556772563980848, + "grad_norm": 0.3258594572544098, + "learning_rate": 0.00016947743456276543, + "loss": 0.0525, + "step": 10907 + }, + { + "epoch": 0.2557006979728898, + "grad_norm": 0.15714235603809357, + "learning_rate": 0.00016947213642888587, + "loss": 0.0141, + "step": 10908 + }, + { + "epoch": 0.2557241395476948, + "grad_norm": 0.5020485520362854, + "learning_rate": 0.0001694668379180485, + "loss": 0.0783, + "step": 10909 + }, + { + "epoch": 0.2557475811224998, + "grad_norm": 0.23653282225131989, + "learning_rate": 0.00016946153903028217, + "loss": 0.0497, + "step": 10910 + }, + { + "epoch": 0.2557710226973048, + "grad_norm": 0.5567378997802734, + "learning_rate": 0.00016945623976561555, + "loss": 0.1454, + "step": 10911 + }, + { + "epoch": 0.2557944642721098, + "grad_norm": 0.3939257860183716, + "learning_rate": 0.0001694509401240774, + "loss": 0.1353, + "step": 10912 + }, + { + "epoch": 0.2558179058469148, + "grad_norm": 0.9195933938026428, + "learning_rate": 0.00016944564010569653, + "loss": 0.14, + "step": 10913 + }, + { + "epoch": 0.25584134742171977, + "grad_norm": 0.5231193900108337, + "learning_rate": 0.00016944033971050165, + "loss": 0.1342, + "step": 10914 + }, + { + "epoch": 0.25586478899652476, + "grad_norm": 0.3950524628162384, + "learning_rate": 0.00016943503893852152, + "loss": 0.0969, + "step": 10915 + }, + { + "epoch": 0.25588823057132976, + "grad_norm": 0.2587001323699951, + "learning_rate": 0.00016942973778978496, + "loss": 0.0483, + "step": 10916 + }, + { + "epoch": 0.25591167214613475, + "grad_norm": 0.7186914682388306, + "learning_rate": 0.00016942443626432064, + "loss": 0.6228, + "step": 10917 + }, + { + "epoch": 0.2559351137209398, + "grad_norm": 0.6745159029960632, + "learning_rate": 0.00016941913436215739, + "loss": 0.1496, + "step": 10918 + }, + { + "epoch": 0.2559585552957448, + "grad_norm": 0.3515014350414276, + "learning_rate": 0.00016941383208332397, + "loss": 0.0781, + "step": 10919 + }, + { + "epoch": 0.2559819968705498, + "grad_norm": 0.2906809151172638, + "learning_rate": 0.00016940852942784918, + "loss": 0.075, + "step": 10920 + }, + { + "epoch": 0.2560054384453548, + "grad_norm": 0.5860452055931091, + "learning_rate": 0.00016940322639576172, + "loss": 0.1442, + "step": 10921 + }, + { + "epoch": 0.2560288800201598, + "grad_norm": 0.2792896032333374, + "learning_rate": 0.0001693979229870904, + "loss": 0.0716, + "step": 10922 + }, + { + "epoch": 0.25605232159496477, + "grad_norm": 0.5294543504714966, + "learning_rate": 0.000169392619201864, + "loss": 0.1169, + "step": 10923 + }, + { + "epoch": 0.25607576316976977, + "grad_norm": 0.4045732915401459, + "learning_rate": 0.0001693873150401113, + "loss": 0.1065, + "step": 10924 + }, + { + "epoch": 0.25609920474457476, + "grad_norm": 0.516793966293335, + "learning_rate": 0.00016938201050186108, + "loss": 0.8648, + "step": 10925 + }, + { + "epoch": 0.25612264631937975, + "grad_norm": 0.49758410453796387, + "learning_rate": 0.00016937670558714213, + "loss": 0.0883, + "step": 10926 + }, + { + "epoch": 0.25614608789418475, + "grad_norm": 0.20989464223384857, + "learning_rate": 0.0001693714002959832, + "loss": 0.0389, + "step": 10927 + }, + { + "epoch": 0.25616952946898974, + "grad_norm": 0.6069733500480652, + "learning_rate": 0.00016936609462841316, + "loss": 0.8059, + "step": 10928 + }, + { + "epoch": 0.25619297104379474, + "grad_norm": 0.45255398750305176, + "learning_rate": 0.00016936078858446066, + "loss": 0.1018, + "step": 10929 + }, + { + "epoch": 0.25621641261859973, + "grad_norm": 0.5151517987251282, + "learning_rate": 0.00016935548216415463, + "loss": 0.5364, + "step": 10930 + }, + { + "epoch": 0.2562398541934047, + "grad_norm": 0.15880221128463745, + "learning_rate": 0.00016935017536752378, + "loss": 0.0477, + "step": 10931 + }, + { + "epoch": 0.2562632957682097, + "grad_norm": 0.40112847089767456, + "learning_rate": 0.00016934486819459693, + "loss": 0.0733, + "step": 10932 + }, + { + "epoch": 0.2562867373430147, + "grad_norm": 0.39139994978904724, + "learning_rate": 0.00016933956064540286, + "loss": 0.1118, + "step": 10933 + }, + { + "epoch": 0.2563101789178197, + "grad_norm": 0.23452413082122803, + "learning_rate": 0.00016933425271997043, + "loss": 0.0466, + "step": 10934 + }, + { + "epoch": 0.2563336204926247, + "grad_norm": 0.34876927733421326, + "learning_rate": 0.00016932894441832837, + "loss": 0.0753, + "step": 10935 + }, + { + "epoch": 0.2563570620674297, + "grad_norm": 0.380651593208313, + "learning_rate": 0.0001693236357405055, + "loss": 0.0735, + "step": 10936 + }, + { + "epoch": 0.2563805036422347, + "grad_norm": 0.47576940059661865, + "learning_rate": 0.00016931832668653062, + "loss": 0.0648, + "step": 10937 + }, + { + "epoch": 0.2564039452170397, + "grad_norm": 0.15675702691078186, + "learning_rate": 0.00016931301725643257, + "loss": 0.0212, + "step": 10938 + }, + { + "epoch": 0.2564273867918447, + "grad_norm": 0.5325210094451904, + "learning_rate": 0.00016930770745024015, + "loss": 0.1652, + "step": 10939 + }, + { + "epoch": 0.2564508283666497, + "grad_norm": 0.2024589627981186, + "learning_rate": 0.00016930239726798213, + "loss": 0.0346, + "step": 10940 + }, + { + "epoch": 0.25647426994145467, + "grad_norm": 0.1909940242767334, + "learning_rate": 0.0001692970867096874, + "loss": 0.0364, + "step": 10941 + }, + { + "epoch": 0.25649771151625966, + "grad_norm": 0.28132060170173645, + "learning_rate": 0.00016929177577538468, + "loss": 0.0407, + "step": 10942 + }, + { + "epoch": 0.25652115309106466, + "grad_norm": 0.29350897669792175, + "learning_rate": 0.00016928646446510284, + "loss": 0.0756, + "step": 10943 + }, + { + "epoch": 0.25654459466586965, + "grad_norm": 0.13009357452392578, + "learning_rate": 0.00016928115277887074, + "loss": 0.0473, + "step": 10944 + }, + { + "epoch": 0.25656803624067465, + "grad_norm": 0.43856579065322876, + "learning_rate": 0.00016927584071671712, + "loss": 0.1644, + "step": 10945 + }, + { + "epoch": 0.25659147781547964, + "grad_norm": 0.11460557579994202, + "learning_rate": 0.00016927052827867085, + "loss": 0.0147, + "step": 10946 + }, + { + "epoch": 0.25661491939028463, + "grad_norm": 0.5178530812263489, + "learning_rate": 0.00016926521546476074, + "loss": 0.1068, + "step": 10947 + }, + { + "epoch": 0.25663836096508963, + "grad_norm": 0.7135165333747864, + "learning_rate": 0.00016925990227501563, + "loss": 0.5928, + "step": 10948 + }, + { + "epoch": 0.2566618025398946, + "grad_norm": 0.23805639147758484, + "learning_rate": 0.00016925458870946435, + "loss": 0.0178, + "step": 10949 + }, + { + "epoch": 0.2566852441146996, + "grad_norm": 1.061281681060791, + "learning_rate": 0.0001692492747681357, + "loss": 0.1548, + "step": 10950 + }, + { + "epoch": 0.2567086856895046, + "grad_norm": 0.507284939289093, + "learning_rate": 0.00016924396045105857, + "loss": 0.1142, + "step": 10951 + }, + { + "epoch": 0.2567321272643096, + "grad_norm": 0.3545617461204529, + "learning_rate": 0.00016923864575826177, + "loss": 0.0296, + "step": 10952 + }, + { + "epoch": 0.2567555688391146, + "grad_norm": 0.16553227603435516, + "learning_rate": 0.00016923333068977406, + "loss": 0.0332, + "step": 10953 + }, + { + "epoch": 0.2567790104139196, + "grad_norm": 0.18883949518203735, + "learning_rate": 0.00016922801524562445, + "loss": 0.0485, + "step": 10954 + }, + { + "epoch": 0.2568024519887246, + "grad_norm": 0.4936503767967224, + "learning_rate": 0.00016922269942584158, + "loss": 0.1151, + "step": 10955 + }, + { + "epoch": 0.2568258935635296, + "grad_norm": 0.4709455072879791, + "learning_rate": 0.0001692173832304545, + "loss": 0.0535, + "step": 10956 + }, + { + "epoch": 0.2568493351383346, + "grad_norm": 0.2682858109474182, + "learning_rate": 0.0001692120666594919, + "loss": 0.0437, + "step": 10957 + }, + { + "epoch": 0.2568727767131396, + "grad_norm": 0.6436682939529419, + "learning_rate": 0.00016920674971298265, + "loss": 0.7534, + "step": 10958 + }, + { + "epoch": 0.25689621828794457, + "grad_norm": 0.6028342843055725, + "learning_rate": 0.00016920143239095568, + "loss": 0.1222, + "step": 10959 + }, + { + "epoch": 0.25691965986274956, + "grad_norm": 0.5691989660263062, + "learning_rate": 0.00016919611469343976, + "loss": 0.0863, + "step": 10960 + }, + { + "epoch": 0.25694310143755456, + "grad_norm": 0.27467936277389526, + "learning_rate": 0.0001691907966204638, + "loss": 0.0879, + "step": 10961 + }, + { + "epoch": 0.25696654301235955, + "grad_norm": 0.9245221614837646, + "learning_rate": 0.00016918547817205664, + "loss": 0.2093, + "step": 10962 + }, + { + "epoch": 0.25698998458716454, + "grad_norm": 0.6537737250328064, + "learning_rate": 0.00016918015934824708, + "loss": 0.1049, + "step": 10963 + }, + { + "epoch": 0.25701342616196954, + "grad_norm": 0.34864699840545654, + "learning_rate": 0.00016917484014906408, + "loss": 0.3994, + "step": 10964 + }, + { + "epoch": 0.25703686773677453, + "grad_norm": 0.18671341240406036, + "learning_rate": 0.00016916952057453646, + "loss": 0.0383, + "step": 10965 + }, + { + "epoch": 0.2570603093115795, + "grad_norm": 0.5112318396568298, + "learning_rate": 0.00016916420062469304, + "loss": 0.1049, + "step": 10966 + }, + { + "epoch": 0.2570837508863845, + "grad_norm": 0.4858752489089966, + "learning_rate": 0.00016915888029956272, + "loss": 0.6762, + "step": 10967 + }, + { + "epoch": 0.2571071924611895, + "grad_norm": 0.6022621393203735, + "learning_rate": 0.00016915355959917438, + "loss": 0.0553, + "step": 10968 + }, + { + "epoch": 0.25713063403599457, + "grad_norm": 0.12557430565357208, + "learning_rate": 0.0001691482385235569, + "loss": 0.0186, + "step": 10969 + }, + { + "epoch": 0.25715407561079956, + "grad_norm": 0.7195578217506409, + "learning_rate": 0.00016914291707273912, + "loss": 0.6429, + "step": 10970 + }, + { + "epoch": 0.25717751718560455, + "grad_norm": 0.1317610740661621, + "learning_rate": 0.00016913759524674992, + "loss": 0.0271, + "step": 10971 + }, + { + "epoch": 0.25720095876040955, + "grad_norm": 0.5701876282691956, + "learning_rate": 0.00016913227304561819, + "loss": 0.1279, + "step": 10972 + }, + { + "epoch": 0.25722440033521454, + "grad_norm": 0.24626867473125458, + "learning_rate": 0.00016912695046937282, + "loss": 0.0235, + "step": 10973 + }, + { + "epoch": 0.25724784191001954, + "grad_norm": 0.40903225541114807, + "learning_rate": 0.00016912162751804262, + "loss": 0.0717, + "step": 10974 + }, + { + "epoch": 0.25727128348482453, + "grad_norm": 0.3940884470939636, + "learning_rate": 0.0001691163041916566, + "loss": 0.0506, + "step": 10975 + }, + { + "epoch": 0.2572947250596295, + "grad_norm": 0.3985580801963806, + "learning_rate": 0.00016911098049024353, + "loss": 0.0894, + "step": 10976 + }, + { + "epoch": 0.2573181666344345, + "grad_norm": 0.5815785527229309, + "learning_rate": 0.00016910565641383235, + "loss": 0.1363, + "step": 10977 + }, + { + "epoch": 0.2573416082092395, + "grad_norm": 0.2654179036617279, + "learning_rate": 0.00016910033196245193, + "loss": 0.0742, + "step": 10978 + }, + { + "epoch": 0.2573650497840445, + "grad_norm": 0.2422369420528412, + "learning_rate": 0.00016909500713613114, + "loss": 0.0343, + "step": 10979 + }, + { + "epoch": 0.2573884913588495, + "grad_norm": 0.3180494010448456, + "learning_rate": 0.0001690896819348989, + "loss": 0.0512, + "step": 10980 + }, + { + "epoch": 0.2574119329336545, + "grad_norm": 0.32332584261894226, + "learning_rate": 0.00016908435635878415, + "loss": 0.0856, + "step": 10981 + }, + { + "epoch": 0.2574353745084595, + "grad_norm": 0.5510647892951965, + "learning_rate": 0.0001690790304078157, + "loss": 0.1149, + "step": 10982 + }, + { + "epoch": 0.2574588160832645, + "grad_norm": 0.41202589869499207, + "learning_rate": 0.0001690737040820225, + "loss": 0.0755, + "step": 10983 + }, + { + "epoch": 0.2574822576580695, + "grad_norm": 0.15801747143268585, + "learning_rate": 0.00016906837738143345, + "loss": 0.0102, + "step": 10984 + }, + { + "epoch": 0.2575056992328745, + "grad_norm": 0.6083974838256836, + "learning_rate": 0.00016906305030607745, + "loss": 0.1528, + "step": 10985 + }, + { + "epoch": 0.25752914080767947, + "grad_norm": 0.18684732913970947, + "learning_rate": 0.00016905772285598337, + "loss": 0.0343, + "step": 10986 + }, + { + "epoch": 0.25755258238248446, + "grad_norm": 0.9105584025382996, + "learning_rate": 0.00016905239503118015, + "loss": 0.6487, + "step": 10987 + }, + { + "epoch": 0.25757602395728946, + "grad_norm": 0.3702869415283203, + "learning_rate": 0.0001690470668316967, + "loss": 0.0828, + "step": 10988 + }, + { + "epoch": 0.25759946553209445, + "grad_norm": 0.3804193139076233, + "learning_rate": 0.00016904173825756195, + "loss": 0.0402, + "step": 10989 + }, + { + "epoch": 0.25762290710689945, + "grad_norm": 0.552461564540863, + "learning_rate": 0.00016903640930880474, + "loss": 0.1154, + "step": 10990 + }, + { + "epoch": 0.25764634868170444, + "grad_norm": 0.5587748289108276, + "learning_rate": 0.00016903107998545404, + "loss": 0.1143, + "step": 10991 + }, + { + "epoch": 0.25766979025650943, + "grad_norm": 0.5526704788208008, + "learning_rate": 0.00016902575028753875, + "loss": 0.656, + "step": 10992 + }, + { + "epoch": 0.25769323183131443, + "grad_norm": 0.16138722002506256, + "learning_rate": 0.00016902042021508783, + "loss": 0.0307, + "step": 10993 + }, + { + "epoch": 0.2577166734061194, + "grad_norm": 0.5664640665054321, + "learning_rate": 0.00016901508976813017, + "loss": 0.1282, + "step": 10994 + }, + { + "epoch": 0.2577401149809244, + "grad_norm": 0.4392664134502411, + "learning_rate": 0.00016900975894669466, + "loss": 0.1299, + "step": 10995 + }, + { + "epoch": 0.2577635565557294, + "grad_norm": 0.5791068077087402, + "learning_rate": 0.0001690044277508103, + "loss": 0.1222, + "step": 10996 + }, + { + "epoch": 0.2577869981305344, + "grad_norm": 0.5985374450683594, + "learning_rate": 0.00016899909618050594, + "loss": 0.9162, + "step": 10997 + }, + { + "epoch": 0.2578104397053394, + "grad_norm": 0.07800135761499405, + "learning_rate": 0.00016899376423581055, + "loss": 0.0196, + "step": 10998 + }, + { + "epoch": 0.2578338812801444, + "grad_norm": 0.6958053708076477, + "learning_rate": 0.00016898843191675306, + "loss": 0.1489, + "step": 10999 + }, + { + "epoch": 0.2578573228549494, + "grad_norm": 0.43805092573165894, + "learning_rate": 0.0001689830992233624, + "loss": 0.0979, + "step": 11000 + }, + { + "epoch": 0.2578807644297544, + "grad_norm": 0.3779895007610321, + "learning_rate": 0.0001689777661556675, + "loss": 0.0693, + "step": 11001 + }, + { + "epoch": 0.2579042060045594, + "grad_norm": 0.14167314767837524, + "learning_rate": 0.00016897243271369732, + "loss": 0.0268, + "step": 11002 + }, + { + "epoch": 0.25792764757936437, + "grad_norm": 1.0153909921646118, + "learning_rate": 0.00016896709889748074, + "loss": 0.7813, + "step": 11003 + }, + { + "epoch": 0.25795108915416937, + "grad_norm": 0.07616525888442993, + "learning_rate": 0.00016896176470704676, + "loss": 0.0134, + "step": 11004 + }, + { + "epoch": 0.25797453072897436, + "grad_norm": 0.6316256523132324, + "learning_rate": 0.00016895643014242434, + "loss": 0.1378, + "step": 11005 + }, + { + "epoch": 0.25799797230377935, + "grad_norm": 0.3431378901004791, + "learning_rate": 0.00016895109520364236, + "loss": 0.1002, + "step": 11006 + }, + { + "epoch": 0.25802141387858435, + "grad_norm": 0.5785248279571533, + "learning_rate": 0.0001689457598907298, + "loss": 0.1106, + "step": 11007 + }, + { + "epoch": 0.25804485545338934, + "grad_norm": 0.5691595077514648, + "learning_rate": 0.00016894042420371563, + "loss": 0.1164, + "step": 11008 + }, + { + "epoch": 0.25806829702819434, + "grad_norm": 0.8948394656181335, + "learning_rate": 0.00016893508814262874, + "loss": 0.2002, + "step": 11009 + }, + { + "epoch": 0.25809173860299933, + "grad_norm": 0.4075659513473511, + "learning_rate": 0.00016892975170749815, + "loss": 0.0972, + "step": 11010 + }, + { + "epoch": 0.2581151801778043, + "grad_norm": 0.557311475276947, + "learning_rate": 0.00016892441489835277, + "loss": 0.1591, + "step": 11011 + }, + { + "epoch": 0.2581386217526093, + "grad_norm": 0.3240739703178406, + "learning_rate": 0.0001689190777152216, + "loss": 0.0568, + "step": 11012 + }, + { + "epoch": 0.2581620633274143, + "grad_norm": 0.23063348233699799, + "learning_rate": 0.0001689137401581336, + "loss": 0.0404, + "step": 11013 + }, + { + "epoch": 0.2581855049022193, + "grad_norm": 0.1348811239004135, + "learning_rate": 0.00016890840222711766, + "loss": 0.0335, + "step": 11014 + }, + { + "epoch": 0.2582089464770243, + "grad_norm": 0.5054349303245544, + "learning_rate": 0.0001689030639222028, + "loss": 0.0632, + "step": 11015 + }, + { + "epoch": 0.2582323880518293, + "grad_norm": 0.5558384656906128, + "learning_rate": 0.000168897725243418, + "loss": 0.0857, + "step": 11016 + }, + { + "epoch": 0.2582558296266343, + "grad_norm": 0.17300774157047272, + "learning_rate": 0.00016889238619079215, + "loss": 0.0326, + "step": 11017 + }, + { + "epoch": 0.2582792712014393, + "grad_norm": 0.527838945388794, + "learning_rate": 0.00016888704676435432, + "loss": 0.5584, + "step": 11018 + }, + { + "epoch": 0.2583027127762443, + "grad_norm": 0.3223264813423157, + "learning_rate": 0.00016888170696413346, + "loss": 0.0466, + "step": 11019 + }, + { + "epoch": 0.25832615435104933, + "grad_norm": 0.5047590136528015, + "learning_rate": 0.0001688763667901585, + "loss": 0.1351, + "step": 11020 + }, + { + "epoch": 0.2583495959258543, + "grad_norm": 0.2067560851573944, + "learning_rate": 0.00016887102624245842, + "loss": 0.068, + "step": 11021 + }, + { + "epoch": 0.2583730375006593, + "grad_norm": 0.5045052170753479, + "learning_rate": 0.00016886568532106222, + "loss": 0.1221, + "step": 11022 + }, + { + "epoch": 0.2583964790754643, + "grad_norm": 0.5079661011695862, + "learning_rate": 0.00016886034402599888, + "loss": 0.123, + "step": 11023 + }, + { + "epoch": 0.2584199206502693, + "grad_norm": 0.17873570322990417, + "learning_rate": 0.0001688550023572974, + "loss": 0.0286, + "step": 11024 + }, + { + "epoch": 0.2584433622250743, + "grad_norm": 0.5964670181274414, + "learning_rate": 0.0001688496603149867, + "loss": 0.066, + "step": 11025 + }, + { + "epoch": 0.2584668037998793, + "grad_norm": 0.47104746103286743, + "learning_rate": 0.00016884431789909585, + "loss": 0.1047, + "step": 11026 + }, + { + "epoch": 0.2584902453746843, + "grad_norm": 0.08318386971950531, + "learning_rate": 0.00016883897510965376, + "loss": 0.0274, + "step": 11027 + }, + { + "epoch": 0.2585136869494893, + "grad_norm": 0.6755596995353699, + "learning_rate": 0.00016883363194668944, + "loss": 0.1517, + "step": 11028 + }, + { + "epoch": 0.2585371285242943, + "grad_norm": 0.6104267239570618, + "learning_rate": 0.00016882828841023194, + "loss": 0.1022, + "step": 11029 + }, + { + "epoch": 0.2585605700990993, + "grad_norm": 0.5957956314086914, + "learning_rate": 0.00016882294450031022, + "loss": 0.1426, + "step": 11030 + }, + { + "epoch": 0.25858401167390427, + "grad_norm": 0.5165791511535645, + "learning_rate": 0.00016881760021695324, + "loss": 0.1457, + "step": 11031 + }, + { + "epoch": 0.25860745324870926, + "grad_norm": 0.45881006121635437, + "learning_rate": 0.00016881225556019004, + "loss": 0.0927, + "step": 11032 + }, + { + "epoch": 0.25863089482351426, + "grad_norm": 0.41522714495658875, + "learning_rate": 0.0001688069105300496, + "loss": 0.1244, + "step": 11033 + }, + { + "epoch": 0.25865433639831925, + "grad_norm": 0.5765830278396606, + "learning_rate": 0.00016880156512656092, + "loss": 0.1684, + "step": 11034 + }, + { + "epoch": 0.25867777797312425, + "grad_norm": 0.7064173817634583, + "learning_rate": 0.00016879621934975302, + "loss": 0.2204, + "step": 11035 + }, + { + "epoch": 0.25870121954792924, + "grad_norm": 0.34881147742271423, + "learning_rate": 0.0001687908731996549, + "loss": 0.1066, + "step": 11036 + }, + { + "epoch": 0.25872466112273423, + "grad_norm": 0.5015280246734619, + "learning_rate": 0.00016878552667629556, + "loss": 0.1185, + "step": 11037 + }, + { + "epoch": 0.25874810269753923, + "grad_norm": 0.42915040254592896, + "learning_rate": 0.00016878017977970403, + "loss": 0.1217, + "step": 11038 + }, + { + "epoch": 0.2587715442723442, + "grad_norm": 1.028767704963684, + "learning_rate": 0.0001687748325099093, + "loss": 0.2425, + "step": 11039 + }, + { + "epoch": 0.2587949858471492, + "grad_norm": 0.19869518280029297, + "learning_rate": 0.00016876948486694042, + "loss": 0.0303, + "step": 11040 + }, + { + "epoch": 0.2588184274219542, + "grad_norm": 0.4366282522678375, + "learning_rate": 0.00016876413685082634, + "loss": 0.0768, + "step": 11041 + }, + { + "epoch": 0.2588418689967592, + "grad_norm": 0.3239380717277527, + "learning_rate": 0.00016875878846159617, + "loss": 0.062, + "step": 11042 + }, + { + "epoch": 0.2588653105715642, + "grad_norm": 0.4038643538951874, + "learning_rate": 0.00016875343969927883, + "loss": 0.1159, + "step": 11043 + }, + { + "epoch": 0.2588887521463692, + "grad_norm": 0.2203063666820526, + "learning_rate": 0.00016874809056390344, + "loss": 0.0463, + "step": 11044 + }, + { + "epoch": 0.2589121937211742, + "grad_norm": 0.15099208056926727, + "learning_rate": 0.0001687427410554989, + "loss": 0.0494, + "step": 11045 + }, + { + "epoch": 0.2589356352959792, + "grad_norm": 0.7327306270599365, + "learning_rate": 0.0001687373911740944, + "loss": 0.2325, + "step": 11046 + }, + { + "epoch": 0.2589590768707842, + "grad_norm": 0.16355860233306885, + "learning_rate": 0.00016873204091971884, + "loss": 0.019, + "step": 11047 + }, + { + "epoch": 0.25898251844558917, + "grad_norm": 0.46995997428894043, + "learning_rate": 0.00016872669029240132, + "loss": 0.0835, + "step": 11048 + }, + { + "epoch": 0.25900596002039417, + "grad_norm": 0.43643859028816223, + "learning_rate": 0.00016872133929217084, + "loss": 0.1106, + "step": 11049 + }, + { + "epoch": 0.25902940159519916, + "grad_norm": 0.49992987513542175, + "learning_rate": 0.00016871598791905643, + "loss": 0.1015, + "step": 11050 + }, + { + "epoch": 0.25905284317000415, + "grad_norm": 0.16862891614437103, + "learning_rate": 0.00016871063617308714, + "loss": 0.0279, + "step": 11051 + }, + { + "epoch": 0.25907628474480915, + "grad_norm": 0.5490517616271973, + "learning_rate": 0.000168705284054292, + "loss": 0.2147, + "step": 11052 + }, + { + "epoch": 0.25909972631961414, + "grad_norm": 0.47988685965538025, + "learning_rate": 0.00016869993156270005, + "loss": 0.1267, + "step": 11053 + }, + { + "epoch": 0.25912316789441914, + "grad_norm": 0.47011861205101013, + "learning_rate": 0.00016869457869834035, + "loss": 0.0777, + "step": 11054 + }, + { + "epoch": 0.25914660946922413, + "grad_norm": 0.8983186483383179, + "learning_rate": 0.00016868922546124196, + "loss": 0.2683, + "step": 11055 + }, + { + "epoch": 0.2591700510440291, + "grad_norm": 0.6142249703407288, + "learning_rate": 0.00016868387185143386, + "loss": 0.1225, + "step": 11056 + }, + { + "epoch": 0.2591934926188341, + "grad_norm": 0.2549769878387451, + "learning_rate": 0.00016867851786894517, + "loss": 0.0402, + "step": 11057 + }, + { + "epoch": 0.2592169341936391, + "grad_norm": 0.3848269581794739, + "learning_rate": 0.0001686731635138049, + "loss": 0.4199, + "step": 11058 + }, + { + "epoch": 0.2592403757684441, + "grad_norm": 1.1240845918655396, + "learning_rate": 0.0001686678087860421, + "loss": 0.3609, + "step": 11059 + }, + { + "epoch": 0.2592638173432491, + "grad_norm": 0.21910235285758972, + "learning_rate": 0.00016866245368568584, + "loss": 0.0445, + "step": 11060 + }, + { + "epoch": 0.2592872589180541, + "grad_norm": 0.5810951590538025, + "learning_rate": 0.00016865709821276519, + "loss": 0.1623, + "step": 11061 + }, + { + "epoch": 0.2593107004928591, + "grad_norm": 0.22440467774868011, + "learning_rate": 0.0001686517423673092, + "loss": 0.0473, + "step": 11062 + }, + { + "epoch": 0.2593341420676641, + "grad_norm": 0.6049763560295105, + "learning_rate": 0.00016864638614934693, + "loss": 0.3355, + "step": 11063 + }, + { + "epoch": 0.2593575836424691, + "grad_norm": 0.6815288662910461, + "learning_rate": 0.0001686410295589074, + "loss": 0.1501, + "step": 11064 + }, + { + "epoch": 0.2593810252172741, + "grad_norm": 0.5126118659973145, + "learning_rate": 0.00016863567259601975, + "loss": 0.6137, + "step": 11065 + }, + { + "epoch": 0.25940446679207907, + "grad_norm": 0.5505920648574829, + "learning_rate": 0.00016863031526071298, + "loss": 0.6393, + "step": 11066 + }, + { + "epoch": 0.25942790836688406, + "grad_norm": 0.45048338174819946, + "learning_rate": 0.00016862495755301624, + "loss": 0.0575, + "step": 11067 + }, + { + "epoch": 0.25945134994168906, + "grad_norm": 0.6221013069152832, + "learning_rate": 0.0001686195994729585, + "loss": 0.1048, + "step": 11068 + }, + { + "epoch": 0.25947479151649405, + "grad_norm": 0.3768232762813568, + "learning_rate": 0.0001686142410205689, + "loss": 0.4994, + "step": 11069 + }, + { + "epoch": 0.25949823309129905, + "grad_norm": 0.5222107768058777, + "learning_rate": 0.0001686088821958765, + "loss": 0.1001, + "step": 11070 + }, + { + "epoch": 0.2595216746661041, + "grad_norm": 0.0970631018280983, + "learning_rate": 0.00016860352299891037, + "loss": 0.0108, + "step": 11071 + }, + { + "epoch": 0.2595451162409091, + "grad_norm": 0.7304882407188416, + "learning_rate": 0.00016859816342969962, + "loss": 0.1964, + "step": 11072 + }, + { + "epoch": 0.2595685578157141, + "grad_norm": 0.22375202178955078, + "learning_rate": 0.0001685928034882733, + "loss": 0.0444, + "step": 11073 + }, + { + "epoch": 0.2595919993905191, + "grad_norm": 0.27498695254325867, + "learning_rate": 0.00016858744317466048, + "loss": 0.042, + "step": 11074 + }, + { + "epoch": 0.2596154409653241, + "grad_norm": 0.27443215250968933, + "learning_rate": 0.00016858208248889028, + "loss": 0.0483, + "step": 11075 + }, + { + "epoch": 0.25963888254012907, + "grad_norm": 0.1315544843673706, + "learning_rate": 0.00016857672143099178, + "loss": 0.0094, + "step": 11076 + }, + { + "epoch": 0.25966232411493406, + "grad_norm": 0.8558392524719238, + "learning_rate": 0.00016857136000099403, + "loss": 0.2274, + "step": 11077 + }, + { + "epoch": 0.25968576568973906, + "grad_norm": 0.38732773065567017, + "learning_rate": 0.00016856599819892618, + "loss": 0.1162, + "step": 11078 + }, + { + "epoch": 0.25970920726454405, + "grad_norm": 0.6104490160942078, + "learning_rate": 0.00016856063602481726, + "loss": 0.7752, + "step": 11079 + }, + { + "epoch": 0.25973264883934905, + "grad_norm": 0.3625788986682892, + "learning_rate": 0.00016855527347869645, + "loss": 0.0831, + "step": 11080 + }, + { + "epoch": 0.25975609041415404, + "grad_norm": 0.6218829154968262, + "learning_rate": 0.00016854991056059276, + "loss": 0.0535, + "step": 11081 + }, + { + "epoch": 0.25977953198895903, + "grad_norm": 0.8388184309005737, + "learning_rate": 0.00016854454727053536, + "loss": 0.2362, + "step": 11082 + }, + { + "epoch": 0.25980297356376403, + "grad_norm": 0.6049408912658691, + "learning_rate": 0.0001685391836085533, + "loss": 0.09, + "step": 11083 + }, + { + "epoch": 0.259826415138569, + "grad_norm": 0.4923383295536041, + "learning_rate": 0.0001685338195746757, + "loss": 0.0723, + "step": 11084 + }, + { + "epoch": 0.259849856713374, + "grad_norm": 0.619515597820282, + "learning_rate": 0.00016852845516893166, + "loss": 0.1579, + "step": 11085 + }, + { + "epoch": 0.259873298288179, + "grad_norm": 0.38932928442955017, + "learning_rate": 0.00016852309039135034, + "loss": 0.1336, + "step": 11086 + }, + { + "epoch": 0.259896739862984, + "grad_norm": 0.2500521242618561, + "learning_rate": 0.00016851772524196078, + "loss": 0.0693, + "step": 11087 + }, + { + "epoch": 0.259920181437789, + "grad_norm": 0.5418628454208374, + "learning_rate": 0.00016851235972079208, + "loss": 0.1001, + "step": 11088 + }, + { + "epoch": 0.259943623012594, + "grad_norm": 0.41494324803352356, + "learning_rate": 0.00016850699382787342, + "loss": 0.108, + "step": 11089 + }, + { + "epoch": 0.259967064587399, + "grad_norm": 0.5155040621757507, + "learning_rate": 0.0001685016275632339, + "loss": 0.1292, + "step": 11090 + }, + { + "epoch": 0.259990506162204, + "grad_norm": 0.5911110639572144, + "learning_rate": 0.0001684962609269026, + "loss": 0.1229, + "step": 11091 + }, + { + "epoch": 0.260013947737009, + "grad_norm": 0.4845779538154602, + "learning_rate": 0.00016849089391890868, + "loss": 0.1116, + "step": 11092 + }, + { + "epoch": 0.26003738931181397, + "grad_norm": 0.5506205558776855, + "learning_rate": 0.00016848552653928123, + "loss": 0.6554, + "step": 11093 + }, + { + "epoch": 0.26006083088661897, + "grad_norm": 0.44284704327583313, + "learning_rate": 0.0001684801587880494, + "loss": 0.1118, + "step": 11094 + }, + { + "epoch": 0.26008427246142396, + "grad_norm": 0.3329842984676361, + "learning_rate": 0.0001684747906652423, + "loss": 0.0554, + "step": 11095 + }, + { + "epoch": 0.26010771403622895, + "grad_norm": 0.21024325489997864, + "learning_rate": 0.00016846942217088904, + "loss": 0.053, + "step": 11096 + }, + { + "epoch": 0.26013115561103395, + "grad_norm": 0.6391198039054871, + "learning_rate": 0.0001684640533050188, + "loss": 0.1615, + "step": 11097 + }, + { + "epoch": 0.26015459718583894, + "grad_norm": 0.40132832527160645, + "learning_rate": 0.00016845868406766066, + "loss": 0.0878, + "step": 11098 + }, + { + "epoch": 0.26017803876064394, + "grad_norm": 0.5438514351844788, + "learning_rate": 0.00016845331445884378, + "loss": 0.152, + "step": 11099 + }, + { + "epoch": 0.26020148033544893, + "grad_norm": 0.21987435221672058, + "learning_rate": 0.0001684479444785973, + "loss": 0.0303, + "step": 11100 + }, + { + "epoch": 0.2602249219102539, + "grad_norm": 0.6004465222358704, + "learning_rate": 0.00016844257412695034, + "loss": 0.2043, + "step": 11101 + }, + { + "epoch": 0.2602483634850589, + "grad_norm": 0.34499701857566833, + "learning_rate": 0.00016843720340393206, + "loss": 0.0866, + "step": 11102 + }, + { + "epoch": 0.2602718050598639, + "grad_norm": 0.46529918909072876, + "learning_rate": 0.00016843183230957158, + "loss": 0.1539, + "step": 11103 + }, + { + "epoch": 0.2602952466346689, + "grad_norm": 0.38113105297088623, + "learning_rate": 0.00016842646084389804, + "loss": 0.0252, + "step": 11104 + }, + { + "epoch": 0.2603186882094739, + "grad_norm": 0.939130425453186, + "learning_rate": 0.00016842108900694063, + "loss": 0.1073, + "step": 11105 + }, + { + "epoch": 0.2603421297842789, + "grad_norm": 0.48973530530929565, + "learning_rate": 0.00016841571679872844, + "loss": 0.1185, + "step": 11106 + }, + { + "epoch": 0.2603655713590839, + "grad_norm": 0.12376939505338669, + "learning_rate": 0.00016841034421929065, + "loss": 0.0095, + "step": 11107 + }, + { + "epoch": 0.2603890129338889, + "grad_norm": 0.5205577611923218, + "learning_rate": 0.0001684049712686564, + "loss": 0.6729, + "step": 11108 + }, + { + "epoch": 0.2604124545086939, + "grad_norm": 0.8744951486587524, + "learning_rate": 0.00016839959794685488, + "loss": 0.2209, + "step": 11109 + }, + { + "epoch": 0.2604358960834989, + "grad_norm": 0.3206281363964081, + "learning_rate": 0.00016839422425391523, + "loss": 0.0572, + "step": 11110 + }, + { + "epoch": 0.26045933765830387, + "grad_norm": 0.15200510621070862, + "learning_rate": 0.00016838885018986656, + "loss": 0.0353, + "step": 11111 + }, + { + "epoch": 0.26048277923310886, + "grad_norm": 0.5716545581817627, + "learning_rate": 0.0001683834757547381, + "loss": 0.1362, + "step": 11112 + }, + { + "epoch": 0.26050622080791386, + "grad_norm": 0.6585162281990051, + "learning_rate": 0.00016837810094855895, + "loss": 0.1454, + "step": 11113 + }, + { + "epoch": 0.26052966238271885, + "grad_norm": 0.8075419664382935, + "learning_rate": 0.0001683727257713583, + "loss": 0.1046, + "step": 11114 + }, + { + "epoch": 0.26055310395752385, + "grad_norm": 0.19698354601860046, + "learning_rate": 0.00016836735022316532, + "loss": 0.0327, + "step": 11115 + }, + { + "epoch": 0.26057654553232884, + "grad_norm": 0.36965885758399963, + "learning_rate": 0.0001683619743040092, + "loss": 0.0297, + "step": 11116 + }, + { + "epoch": 0.26059998710713383, + "grad_norm": 0.485565185546875, + "learning_rate": 0.00016835659801391905, + "loss": 0.0644, + "step": 11117 + }, + { + "epoch": 0.26062342868193883, + "grad_norm": 0.27051275968551636, + "learning_rate": 0.0001683512213529241, + "loss": 0.0409, + "step": 11118 + }, + { + "epoch": 0.2606468702567438, + "grad_norm": 0.09640590846538544, + "learning_rate": 0.0001683458443210535, + "loss": 0.0271, + "step": 11119 + }, + { + "epoch": 0.2606703118315488, + "grad_norm": 0.23828516900539398, + "learning_rate": 0.0001683404669183364, + "loss": 0.0372, + "step": 11120 + }, + { + "epoch": 0.2606937534063538, + "grad_norm": 0.5361570715904236, + "learning_rate": 0.00016833508914480203, + "loss": 0.0883, + "step": 11121 + }, + { + "epoch": 0.26071719498115886, + "grad_norm": 0.5691536664962769, + "learning_rate": 0.00016832971100047957, + "loss": 0.632, + "step": 11122 + }, + { + "epoch": 0.26074063655596386, + "grad_norm": 0.8155473470687866, + "learning_rate": 0.00016832433248539812, + "loss": 0.1599, + "step": 11123 + }, + { + "epoch": 0.26076407813076885, + "grad_norm": 0.531785249710083, + "learning_rate": 0.00016831895359958695, + "loss": 0.7824, + "step": 11124 + }, + { + "epoch": 0.26078751970557384, + "grad_norm": 0.7267643809318542, + "learning_rate": 0.00016831357434307526, + "loss": 0.1857, + "step": 11125 + }, + { + "epoch": 0.26081096128037884, + "grad_norm": 0.22953593730926514, + "learning_rate": 0.0001683081947158921, + "loss": 0.0238, + "step": 11126 + }, + { + "epoch": 0.26083440285518383, + "grad_norm": 0.799075186252594, + "learning_rate": 0.00016830281471806683, + "loss": 0.1326, + "step": 11127 + }, + { + "epoch": 0.2608578444299888, + "grad_norm": 0.14818087220191956, + "learning_rate": 0.00016829743434962853, + "loss": 0.0337, + "step": 11128 + }, + { + "epoch": 0.2608812860047938, + "grad_norm": 0.7031774520874023, + "learning_rate": 0.00016829205361060647, + "loss": 0.1565, + "step": 11129 + }, + { + "epoch": 0.2609047275795988, + "grad_norm": 0.9151565432548523, + "learning_rate": 0.0001682866725010298, + "loss": 0.1802, + "step": 11130 + }, + { + "epoch": 0.2609281691544038, + "grad_norm": 0.16460710763931274, + "learning_rate": 0.00016828129102092768, + "loss": 0.0394, + "step": 11131 + }, + { + "epoch": 0.2609516107292088, + "grad_norm": 0.20115883648395538, + "learning_rate": 0.0001682759091703294, + "loss": 0.0356, + "step": 11132 + }, + { + "epoch": 0.2609750523040138, + "grad_norm": 0.5195479989051819, + "learning_rate": 0.00016827052694926405, + "loss": 0.151, + "step": 11133 + }, + { + "epoch": 0.2609984938788188, + "grad_norm": 0.46234941482543945, + "learning_rate": 0.00016826514435776098, + "loss": 0.1134, + "step": 11134 + }, + { + "epoch": 0.2610219354536238, + "grad_norm": 0.4598771929740906, + "learning_rate": 0.00016825976139584929, + "loss": 0.0933, + "step": 11135 + }, + { + "epoch": 0.2610453770284288, + "grad_norm": 0.1943250298500061, + "learning_rate": 0.0001682543780635582, + "loss": 0.0394, + "step": 11136 + }, + { + "epoch": 0.2610688186032338, + "grad_norm": 0.6535816192626953, + "learning_rate": 0.00016824899436091695, + "loss": 0.0894, + "step": 11137 + }, + { + "epoch": 0.26109226017803877, + "grad_norm": 0.9795627593994141, + "learning_rate": 0.00016824361028795472, + "loss": 0.16, + "step": 11138 + }, + { + "epoch": 0.26111570175284377, + "grad_norm": 0.6699813008308411, + "learning_rate": 0.00016823822584470074, + "loss": 0.1855, + "step": 11139 + }, + { + "epoch": 0.26113914332764876, + "grad_norm": 0.1428484171628952, + "learning_rate": 0.0001682328410311843, + "loss": 0.0337, + "step": 11140 + }, + { + "epoch": 0.26116258490245375, + "grad_norm": 0.8263479471206665, + "learning_rate": 0.00016822745584743447, + "loss": 0.192, + "step": 11141 + }, + { + "epoch": 0.26118602647725875, + "grad_norm": 0.6898497343063354, + "learning_rate": 0.00016822207029348055, + "loss": 0.0783, + "step": 11142 + }, + { + "epoch": 0.26120946805206374, + "grad_norm": 0.10310444235801697, + "learning_rate": 0.00016821668436935177, + "loss": 0.016, + "step": 11143 + }, + { + "epoch": 0.26123290962686874, + "grad_norm": 0.6546668410301208, + "learning_rate": 0.00016821129807507735, + "loss": 0.1456, + "step": 11144 + }, + { + "epoch": 0.26125635120167373, + "grad_norm": 0.2738611400127411, + "learning_rate": 0.00016820591141068652, + "loss": 0.0564, + "step": 11145 + }, + { + "epoch": 0.2612797927764787, + "grad_norm": 0.5077426433563232, + "learning_rate": 0.00016820052437620847, + "loss": 0.1179, + "step": 11146 + }, + { + "epoch": 0.2613032343512837, + "grad_norm": 0.4886651039123535, + "learning_rate": 0.00016819513697167247, + "loss": 0.08, + "step": 11147 + }, + { + "epoch": 0.2613266759260887, + "grad_norm": 0.3766149580478668, + "learning_rate": 0.00016818974919710774, + "loss": 0.1416, + "step": 11148 + }, + { + "epoch": 0.2613501175008937, + "grad_norm": 0.6481250524520874, + "learning_rate": 0.00016818436105254354, + "loss": 0.1538, + "step": 11149 + }, + { + "epoch": 0.2613735590756987, + "grad_norm": 0.5955440998077393, + "learning_rate": 0.00016817897253800903, + "loss": 0.1061, + "step": 11150 + }, + { + "epoch": 0.2613970006505037, + "grad_norm": 0.5004754662513733, + "learning_rate": 0.0001681735836535335, + "loss": 0.1276, + "step": 11151 + }, + { + "epoch": 0.2614204422253087, + "grad_norm": 0.30247992277145386, + "learning_rate": 0.00016816819439914623, + "loss": 0.0706, + "step": 11152 + }, + { + "epoch": 0.2614438838001137, + "grad_norm": 0.8259243965148926, + "learning_rate": 0.0001681628047748764, + "loss": 0.1475, + "step": 11153 + }, + { + "epoch": 0.2614673253749187, + "grad_norm": 0.4747508764266968, + "learning_rate": 0.00016815741478075328, + "loss": 0.1566, + "step": 11154 + }, + { + "epoch": 0.2614907669497237, + "grad_norm": 0.5393422842025757, + "learning_rate": 0.00016815202441680611, + "loss": 0.1242, + "step": 11155 + }, + { + "epoch": 0.26151420852452867, + "grad_norm": 0.43559205532073975, + "learning_rate": 0.00016814663368306412, + "loss": 0.0746, + "step": 11156 + }, + { + "epoch": 0.26153765009933366, + "grad_norm": 1.3944655656814575, + "learning_rate": 0.00016814124257955658, + "loss": 0.1074, + "step": 11157 + }, + { + "epoch": 0.26156109167413866, + "grad_norm": 0.5505359768867493, + "learning_rate": 0.0001681358511063128, + "loss": 0.6401, + "step": 11158 + }, + { + "epoch": 0.26158453324894365, + "grad_norm": 0.8458204865455627, + "learning_rate": 0.0001681304592633619, + "loss": 0.0868, + "step": 11159 + }, + { + "epoch": 0.26160797482374865, + "grad_norm": 0.4049227833747864, + "learning_rate": 0.00016812506705073325, + "loss": 0.0901, + "step": 11160 + }, + { + "epoch": 0.26163141639855364, + "grad_norm": 0.23875217139720917, + "learning_rate": 0.00016811967446845607, + "loss": 0.0574, + "step": 11161 + }, + { + "epoch": 0.26165485797335863, + "grad_norm": 0.22147461771965027, + "learning_rate": 0.00016811428151655963, + "loss": 0.0625, + "step": 11162 + }, + { + "epoch": 0.26167829954816363, + "grad_norm": 0.2688615024089813, + "learning_rate": 0.00016810888819507316, + "loss": 0.0544, + "step": 11163 + }, + { + "epoch": 0.2617017411229686, + "grad_norm": 0.6187421083450317, + "learning_rate": 0.00016810349450402593, + "loss": 0.0701, + "step": 11164 + }, + { + "epoch": 0.2617251826977736, + "grad_norm": 0.5279889702796936, + "learning_rate": 0.00016809810044344728, + "loss": 0.1452, + "step": 11165 + }, + { + "epoch": 0.2617486242725786, + "grad_norm": 0.7410956025123596, + "learning_rate": 0.0001680927060133664, + "loss": 0.1244, + "step": 11166 + }, + { + "epoch": 0.2617720658473836, + "grad_norm": 0.35169458389282227, + "learning_rate": 0.00016808731121381257, + "loss": 0.0573, + "step": 11167 + }, + { + "epoch": 0.2617955074221886, + "grad_norm": 0.4692017734050751, + "learning_rate": 0.00016808191604481505, + "loss": 0.3691, + "step": 11168 + }, + { + "epoch": 0.2618189489969936, + "grad_norm": 0.166940376162529, + "learning_rate": 0.0001680765205064032, + "loss": 0.0156, + "step": 11169 + }, + { + "epoch": 0.2618423905717986, + "grad_norm": 0.514896035194397, + "learning_rate": 0.0001680711245986062, + "loss": 0.146, + "step": 11170 + }, + { + "epoch": 0.2618658321466036, + "grad_norm": 0.22816503047943115, + "learning_rate": 0.00016806572832145337, + "loss": 0.044, + "step": 11171 + }, + { + "epoch": 0.2618892737214086, + "grad_norm": 0.6170011162757874, + "learning_rate": 0.00016806033167497396, + "loss": 0.1469, + "step": 11172 + }, + { + "epoch": 0.2619127152962136, + "grad_norm": 0.13953301310539246, + "learning_rate": 0.00016805493465919732, + "loss": 0.0414, + "step": 11173 + }, + { + "epoch": 0.2619361568710186, + "grad_norm": 0.7229751944541931, + "learning_rate": 0.00016804953727415266, + "loss": 0.1485, + "step": 11174 + }, + { + "epoch": 0.2619595984458236, + "grad_norm": 0.585942804813385, + "learning_rate": 0.0001680441395198693, + "loss": 0.1336, + "step": 11175 + }, + { + "epoch": 0.2619830400206286, + "grad_norm": 0.49201974272727966, + "learning_rate": 0.00016803874139637658, + "loss": 0.0573, + "step": 11176 + }, + { + "epoch": 0.2620064815954336, + "grad_norm": 0.8137434124946594, + "learning_rate": 0.00016803334290370366, + "loss": 0.662, + "step": 11177 + }, + { + "epoch": 0.2620299231702386, + "grad_norm": 0.5962467789649963, + "learning_rate": 0.00016802794404187994, + "loss": 0.1534, + "step": 11178 + }, + { + "epoch": 0.2620533647450436, + "grad_norm": 0.7056680917739868, + "learning_rate": 0.00016802254481093469, + "loss": 0.2135, + "step": 11179 + }, + { + "epoch": 0.2620768063198486, + "grad_norm": 0.38553211092948914, + "learning_rate": 0.00016801714521089717, + "loss": 0.0506, + "step": 11180 + }, + { + "epoch": 0.2621002478946536, + "grad_norm": 0.20161111652851105, + "learning_rate": 0.00016801174524179671, + "loss": 0.0475, + "step": 11181 + }, + { + "epoch": 0.2621236894694586, + "grad_norm": 0.8780447244644165, + "learning_rate": 0.00016800634490366265, + "loss": 1.2523, + "step": 11182 + }, + { + "epoch": 0.26214713104426357, + "grad_norm": 0.8616023659706116, + "learning_rate": 0.00016800094419652423, + "loss": 0.158, + "step": 11183 + }, + { + "epoch": 0.26217057261906856, + "grad_norm": 0.5883771777153015, + "learning_rate": 0.00016799554312041078, + "loss": 0.1264, + "step": 11184 + }, + { + "epoch": 0.26219401419387356, + "grad_norm": 0.4673301875591278, + "learning_rate": 0.0001679901416753516, + "loss": 0.0731, + "step": 11185 + }, + { + "epoch": 0.26221745576867855, + "grad_norm": 0.5644683241844177, + "learning_rate": 0.00016798473986137598, + "loss": 0.1276, + "step": 11186 + }, + { + "epoch": 0.26224089734348355, + "grad_norm": 0.4733586311340332, + "learning_rate": 0.00016797933767851326, + "loss": 0.4925, + "step": 11187 + }, + { + "epoch": 0.26226433891828854, + "grad_norm": 0.44368985295295715, + "learning_rate": 0.00016797393512679273, + "loss": 0.0594, + "step": 11188 + }, + { + "epoch": 0.26228778049309354, + "grad_norm": 0.6797733902931213, + "learning_rate": 0.00016796853220624373, + "loss": 0.1351, + "step": 11189 + }, + { + "epoch": 0.26231122206789853, + "grad_norm": 0.7358852624893188, + "learning_rate": 0.00016796312891689558, + "loss": 0.1366, + "step": 11190 + }, + { + "epoch": 0.2623346636427035, + "grad_norm": 0.43392622470855713, + "learning_rate": 0.00016795772525877759, + "loss": 0.3982, + "step": 11191 + }, + { + "epoch": 0.2623581052175085, + "grad_norm": 0.6735832691192627, + "learning_rate": 0.00016795232123191902, + "loss": 0.1449, + "step": 11192 + }, + { + "epoch": 0.2623815467923135, + "grad_norm": 0.6240580677986145, + "learning_rate": 0.00016794691683634925, + "loss": 0.1273, + "step": 11193 + }, + { + "epoch": 0.2624049883671185, + "grad_norm": 0.4446828067302704, + "learning_rate": 0.00016794151207209765, + "loss": 0.1311, + "step": 11194 + }, + { + "epoch": 0.2624284299419235, + "grad_norm": 0.6762906908988953, + "learning_rate": 0.00016793610693919346, + "loss": 0.1338, + "step": 11195 + }, + { + "epoch": 0.2624518715167285, + "grad_norm": 0.22816473245620728, + "learning_rate": 0.00016793070143766606, + "loss": 0.0537, + "step": 11196 + }, + { + "epoch": 0.2624753130915335, + "grad_norm": 0.205770805478096, + "learning_rate": 0.00016792529556754478, + "loss": 0.0422, + "step": 11197 + }, + { + "epoch": 0.2624987546663385, + "grad_norm": 0.24224723875522614, + "learning_rate": 0.0001679198893288589, + "loss": 0.0296, + "step": 11198 + }, + { + "epoch": 0.2625221962411435, + "grad_norm": 0.527225136756897, + "learning_rate": 0.00016791448272163777, + "loss": 0.0747, + "step": 11199 + }, + { + "epoch": 0.2625456378159485, + "grad_norm": 0.5518425107002258, + "learning_rate": 0.00016790907574591079, + "loss": 0.1884, + "step": 11200 + }, + { + "epoch": 0.26256907939075347, + "grad_norm": 0.5503342747688293, + "learning_rate": 0.00016790366840170723, + "loss": 0.807, + "step": 11201 + }, + { + "epoch": 0.26259252096555846, + "grad_norm": 0.4234360456466675, + "learning_rate": 0.0001678982606890565, + "loss": 0.1091, + "step": 11202 + }, + { + "epoch": 0.26261596254036346, + "grad_norm": 0.8082767724990845, + "learning_rate": 0.00016789285260798787, + "loss": 0.1808, + "step": 11203 + }, + { + "epoch": 0.26263940411516845, + "grad_norm": 0.28477486968040466, + "learning_rate": 0.0001678874441585307, + "loss": 0.0741, + "step": 11204 + }, + { + "epoch": 0.26266284568997345, + "grad_norm": 0.607519805431366, + "learning_rate": 0.00016788203534071435, + "loss": 0.1765, + "step": 11205 + }, + { + "epoch": 0.26268628726477844, + "grad_norm": 0.35519087314605713, + "learning_rate": 0.00016787662615456815, + "loss": 0.3195, + "step": 11206 + }, + { + "epoch": 0.26270972883958343, + "grad_norm": 0.5680370926856995, + "learning_rate": 0.00016787121660012154, + "loss": 0.1158, + "step": 11207 + }, + { + "epoch": 0.26273317041438843, + "grad_norm": 0.8116618990898132, + "learning_rate": 0.00016786580667740376, + "loss": 0.1861, + "step": 11208 + }, + { + "epoch": 0.2627566119891934, + "grad_norm": 0.8347171545028687, + "learning_rate": 0.00016786039638644418, + "loss": 0.1168, + "step": 11209 + }, + { + "epoch": 0.2627800535639984, + "grad_norm": 0.17237110435962677, + "learning_rate": 0.0001678549857272722, + "loss": 0.0337, + "step": 11210 + }, + { + "epoch": 0.2628034951388034, + "grad_norm": 0.550467848777771, + "learning_rate": 0.00016784957469991716, + "loss": 0.1032, + "step": 11211 + }, + { + "epoch": 0.2628269367136084, + "grad_norm": 0.40244200825691223, + "learning_rate": 0.0001678441633044084, + "loss": 0.0821, + "step": 11212 + }, + { + "epoch": 0.2628503782884134, + "grad_norm": 0.40921252965927124, + "learning_rate": 0.00016783875154077535, + "loss": 0.0538, + "step": 11213 + }, + { + "epoch": 0.2628738198632184, + "grad_norm": 0.298898845911026, + "learning_rate": 0.0001678333394090473, + "loss": 0.0615, + "step": 11214 + }, + { + "epoch": 0.2628972614380234, + "grad_norm": 0.92995685338974, + "learning_rate": 0.00016782792690925365, + "loss": 0.1116, + "step": 11215 + }, + { + "epoch": 0.2629207030128284, + "grad_norm": 0.5242421627044678, + "learning_rate": 0.00016782251404142375, + "loss": 0.1197, + "step": 11216 + }, + { + "epoch": 0.2629441445876334, + "grad_norm": 0.3895799517631531, + "learning_rate": 0.00016781710080558698, + "loss": 0.1106, + "step": 11217 + }, + { + "epoch": 0.26296758616243837, + "grad_norm": 0.15961523354053497, + "learning_rate": 0.00016781168720177272, + "loss": 0.0376, + "step": 11218 + }, + { + "epoch": 0.26299102773724337, + "grad_norm": 0.5072868466377258, + "learning_rate": 0.00016780627323001034, + "loss": 0.1059, + "step": 11219 + }, + { + "epoch": 0.26301446931204836, + "grad_norm": 0.6098797917366028, + "learning_rate": 0.0001678008588903292, + "loss": 0.1049, + "step": 11220 + }, + { + "epoch": 0.26303791088685335, + "grad_norm": 0.5586089491844177, + "learning_rate": 0.00016779544418275873, + "loss": 0.1048, + "step": 11221 + }, + { + "epoch": 0.26306135246165835, + "grad_norm": 0.32755813002586365, + "learning_rate": 0.00016779002910732827, + "loss": 0.4891, + "step": 11222 + }, + { + "epoch": 0.26308479403646334, + "grad_norm": 0.5024951100349426, + "learning_rate": 0.00016778461366406716, + "loss": 0.136, + "step": 11223 + }, + { + "epoch": 0.26310823561126834, + "grad_norm": 0.6579483151435852, + "learning_rate": 0.00016777919785300487, + "loss": 0.1486, + "step": 11224 + }, + { + "epoch": 0.2631316771860734, + "grad_norm": 0.3661322295665741, + "learning_rate": 0.00016777378167417073, + "loss": 0.0406, + "step": 11225 + }, + { + "epoch": 0.2631551187608784, + "grad_norm": 0.5277058482170105, + "learning_rate": 0.00016776836512759415, + "loss": 0.08, + "step": 11226 + }, + { + "epoch": 0.2631785603356834, + "grad_norm": 0.3654125928878784, + "learning_rate": 0.00016776294821330452, + "loss": 0.0897, + "step": 11227 + }, + { + "epoch": 0.26320200191048837, + "grad_norm": 0.4706968665122986, + "learning_rate": 0.0001677575309313312, + "loss": 0.0898, + "step": 11228 + }, + { + "epoch": 0.26322544348529336, + "grad_norm": 0.33051714301109314, + "learning_rate": 0.00016775211328170364, + "loss": 0.0505, + "step": 11229 + }, + { + "epoch": 0.26324888506009836, + "grad_norm": 0.5404694676399231, + "learning_rate": 0.0001677466952644512, + "loss": 0.1088, + "step": 11230 + }, + { + "epoch": 0.26327232663490335, + "grad_norm": 0.5540599226951599, + "learning_rate": 0.00016774127687960332, + "loss": 0.0915, + "step": 11231 + }, + { + "epoch": 0.26329576820970835, + "grad_norm": 0.8332512974739075, + "learning_rate": 0.00016773585812718934, + "loss": 0.1815, + "step": 11232 + }, + { + "epoch": 0.26331920978451334, + "grad_norm": 0.49893561005592346, + "learning_rate": 0.00016773043900723865, + "loss": 0.1605, + "step": 11233 + }, + { + "epoch": 0.26334265135931834, + "grad_norm": 0.7822564840316772, + "learning_rate": 0.00016772501951978073, + "loss": 0.1891, + "step": 11234 + }, + { + "epoch": 0.26336609293412333, + "grad_norm": 0.16648255288600922, + "learning_rate": 0.00016771959966484495, + "loss": 0.0476, + "step": 11235 + }, + { + "epoch": 0.2633895345089283, + "grad_norm": 0.7471863627433777, + "learning_rate": 0.00016771417944246071, + "loss": 0.7073, + "step": 11236 + }, + { + "epoch": 0.2634129760837333, + "grad_norm": 1.8614203929901123, + "learning_rate": 0.00016770875885265744, + "loss": 0.3252, + "step": 11237 + }, + { + "epoch": 0.2634364176585383, + "grad_norm": 0.5491059422492981, + "learning_rate": 0.00016770333789546454, + "loss": 0.1377, + "step": 11238 + }, + { + "epoch": 0.2634598592333433, + "grad_norm": 0.6135349869728088, + "learning_rate": 0.00016769791657091142, + "loss": 0.1575, + "step": 11239 + }, + { + "epoch": 0.2634833008081483, + "grad_norm": 0.38134825229644775, + "learning_rate": 0.0001676924948790275, + "loss": 0.0737, + "step": 11240 + }, + { + "epoch": 0.2635067423829533, + "grad_norm": 0.4147808253765106, + "learning_rate": 0.0001676870728198422, + "loss": 0.1348, + "step": 11241 + }, + { + "epoch": 0.2635301839577583, + "grad_norm": 0.4632175862789154, + "learning_rate": 0.00016768165039338494, + "loss": 0.1386, + "step": 11242 + }, + { + "epoch": 0.2635536255325633, + "grad_norm": 0.4878372848033905, + "learning_rate": 0.00016767622759968512, + "loss": 0.0905, + "step": 11243 + }, + { + "epoch": 0.2635770671073683, + "grad_norm": 0.873866081237793, + "learning_rate": 0.00016767080443877224, + "loss": 0.1329, + "step": 11244 + }, + { + "epoch": 0.2636005086821733, + "grad_norm": 0.5206285715103149, + "learning_rate": 0.00016766538091067562, + "loss": 0.1172, + "step": 11245 + }, + { + "epoch": 0.26362395025697827, + "grad_norm": 0.13926610350608826, + "learning_rate": 0.0001676599570154248, + "loss": 0.0375, + "step": 11246 + }, + { + "epoch": 0.26364739183178326, + "grad_norm": 0.8082709908485413, + "learning_rate": 0.00016765453275304908, + "loss": 0.1427, + "step": 11247 + }, + { + "epoch": 0.26367083340658826, + "grad_norm": 0.6804131269454956, + "learning_rate": 0.00016764910812357798, + "loss": 0.1497, + "step": 11248 + }, + { + "epoch": 0.26369427498139325, + "grad_norm": 0.19373397529125214, + "learning_rate": 0.00016764368312704094, + "loss": 0.0423, + "step": 11249 + }, + { + "epoch": 0.26371771655619825, + "grad_norm": 0.6159773468971252, + "learning_rate": 0.00016763825776346737, + "loss": 0.1905, + "step": 11250 + }, + { + "epoch": 0.26374115813100324, + "grad_norm": 0.129631906747818, + "learning_rate": 0.00016763283203288673, + "loss": 0.0384, + "step": 11251 + }, + { + "epoch": 0.26376459970580823, + "grad_norm": 0.35902512073516846, + "learning_rate": 0.00016762740593532843, + "loss": 0.0591, + "step": 11252 + }, + { + "epoch": 0.26378804128061323, + "grad_norm": 0.5978168845176697, + "learning_rate": 0.0001676219794708219, + "loss": 0.1529, + "step": 11253 + }, + { + "epoch": 0.2638114828554182, + "grad_norm": 0.417913019657135, + "learning_rate": 0.0001676165526393966, + "loss": 0.0866, + "step": 11254 + }, + { + "epoch": 0.2638349244302232, + "grad_norm": 0.26062560081481934, + "learning_rate": 0.00016761112544108202, + "loss": 0.1005, + "step": 11255 + }, + { + "epoch": 0.2638583660050282, + "grad_norm": 0.7655705213546753, + "learning_rate": 0.00016760569787590757, + "loss": 0.1158, + "step": 11256 + }, + { + "epoch": 0.2638818075798332, + "grad_norm": 0.1906888782978058, + "learning_rate": 0.0001676002699439027, + "loss": 0.0334, + "step": 11257 + }, + { + "epoch": 0.2639052491546382, + "grad_norm": 0.11024913191795349, + "learning_rate": 0.00016759484164509685, + "loss": 0.0183, + "step": 11258 + }, + { + "epoch": 0.2639286907294432, + "grad_norm": 0.6884051561355591, + "learning_rate": 0.00016758941297951948, + "loss": 0.1231, + "step": 11259 + }, + { + "epoch": 0.2639521323042482, + "grad_norm": 0.6380037069320679, + "learning_rate": 0.00016758398394720007, + "loss": 0.1098, + "step": 11260 + }, + { + "epoch": 0.2639755738790532, + "grad_norm": 0.7616057991981506, + "learning_rate": 0.00016757855454816806, + "loss": 0.134, + "step": 11261 + }, + { + "epoch": 0.2639990154538582, + "grad_norm": 0.5317718982696533, + "learning_rate": 0.00016757312478245291, + "loss": 0.2114, + "step": 11262 + }, + { + "epoch": 0.26402245702866317, + "grad_norm": 0.693315863609314, + "learning_rate": 0.0001675676946500841, + "loss": 0.1479, + "step": 11263 + }, + { + "epoch": 0.26404589860346817, + "grad_norm": 0.4545246660709381, + "learning_rate": 0.00016756226415109106, + "loss": 0.1419, + "step": 11264 + }, + { + "epoch": 0.26406934017827316, + "grad_norm": 0.6034801602363586, + "learning_rate": 0.00016755683328550325, + "loss": 0.7279, + "step": 11265 + }, + { + "epoch": 0.26409278175307815, + "grad_norm": 0.4654789865016937, + "learning_rate": 0.00016755140205335022, + "loss": 0.0695, + "step": 11266 + }, + { + "epoch": 0.26411622332788315, + "grad_norm": 0.264799565076828, + "learning_rate": 0.00016754597045466136, + "loss": 0.0711, + "step": 11267 + }, + { + "epoch": 0.26413966490268814, + "grad_norm": 0.6565135717391968, + "learning_rate": 0.00016754053848946616, + "loss": 0.2025, + "step": 11268 + }, + { + "epoch": 0.26416310647749314, + "grad_norm": 0.7565827965736389, + "learning_rate": 0.0001675351061577941, + "loss": 0.0522, + "step": 11269 + }, + { + "epoch": 0.26418654805229813, + "grad_norm": 0.2783643305301666, + "learning_rate": 0.00016752967345967467, + "loss": 0.0292, + "step": 11270 + }, + { + "epoch": 0.2642099896271031, + "grad_norm": 0.31816187500953674, + "learning_rate": 0.0001675242403951373, + "loss": 0.0443, + "step": 11271 + }, + { + "epoch": 0.2642334312019081, + "grad_norm": 0.18247278034687042, + "learning_rate": 0.00016751880696421154, + "loss": 0.0502, + "step": 11272 + }, + { + "epoch": 0.2642568727767131, + "grad_norm": 0.3772483170032501, + "learning_rate": 0.00016751337316692682, + "loss": 0.0586, + "step": 11273 + }, + { + "epoch": 0.2642803143515181, + "grad_norm": 0.6739150285720825, + "learning_rate": 0.00016750793900331264, + "loss": 0.1528, + "step": 11274 + }, + { + "epoch": 0.2643037559263231, + "grad_norm": 0.2812805473804474, + "learning_rate": 0.00016750250447339848, + "loss": 0.3797, + "step": 11275 + }, + { + "epoch": 0.26432719750112815, + "grad_norm": 0.6665917038917542, + "learning_rate": 0.00016749706957721385, + "loss": 0.1694, + "step": 11276 + }, + { + "epoch": 0.26435063907593315, + "grad_norm": 0.497014582157135, + "learning_rate": 0.0001674916343147882, + "loss": 0.0988, + "step": 11277 + }, + { + "epoch": 0.26437408065073814, + "grad_norm": 0.45713362097740173, + "learning_rate": 0.00016748619868615106, + "loss": 0.1028, + "step": 11278 + }, + { + "epoch": 0.26439752222554314, + "grad_norm": 0.7434092164039612, + "learning_rate": 0.0001674807626913319, + "loss": 0.1538, + "step": 11279 + }, + { + "epoch": 0.26442096380034813, + "grad_norm": 0.4768427610397339, + "learning_rate": 0.00016747532633036021, + "loss": 0.1538, + "step": 11280 + }, + { + "epoch": 0.2644444053751531, + "grad_norm": 0.14716191589832306, + "learning_rate": 0.00016746988960326552, + "loss": 0.037, + "step": 11281 + }, + { + "epoch": 0.2644678469499581, + "grad_norm": 0.15619631111621857, + "learning_rate": 0.00016746445251007732, + "loss": 0.0349, + "step": 11282 + }, + { + "epoch": 0.2644912885247631, + "grad_norm": 0.3315783143043518, + "learning_rate": 0.0001674590150508251, + "loss": 0.1044, + "step": 11283 + }, + { + "epoch": 0.2645147300995681, + "grad_norm": 0.6757332682609558, + "learning_rate": 0.00016745357722553833, + "loss": 0.1144, + "step": 11284 + }, + { + "epoch": 0.2645381716743731, + "grad_norm": 0.770926296710968, + "learning_rate": 0.0001674481390342466, + "loss": 0.247, + "step": 11285 + }, + { + "epoch": 0.2645616132491781, + "grad_norm": 0.6408274173736572, + "learning_rate": 0.00016744270047697936, + "loss": 0.77, + "step": 11286 + }, + { + "epoch": 0.2645850548239831, + "grad_norm": 0.30070292949676514, + "learning_rate": 0.00016743726155376613, + "loss": 0.0463, + "step": 11287 + }, + { + "epoch": 0.2646084963987881, + "grad_norm": 0.44996607303619385, + "learning_rate": 0.0001674318222646364, + "loss": 0.1227, + "step": 11288 + }, + { + "epoch": 0.2646319379735931, + "grad_norm": 0.24290573596954346, + "learning_rate": 0.00016742638260961975, + "loss": 0.0544, + "step": 11289 + }, + { + "epoch": 0.2646553795483981, + "grad_norm": 0.42510247230529785, + "learning_rate": 0.0001674209425887456, + "loss": 0.1266, + "step": 11290 + }, + { + "epoch": 0.26467882112320307, + "grad_norm": 0.43301495909690857, + "learning_rate": 0.00016741550220204356, + "loss": 0.0797, + "step": 11291 + }, + { + "epoch": 0.26470226269800806, + "grad_norm": 0.494600385427475, + "learning_rate": 0.0001674100614495431, + "loss": 0.1113, + "step": 11292 + }, + { + "epoch": 0.26472570427281306, + "grad_norm": 0.6058350205421448, + "learning_rate": 0.00016740462033127375, + "loss": 0.124, + "step": 11293 + }, + { + "epoch": 0.26474914584761805, + "grad_norm": 0.5869075655937195, + "learning_rate": 0.00016739917884726503, + "loss": 0.0618, + "step": 11294 + }, + { + "epoch": 0.26477258742242304, + "grad_norm": 0.9263664484024048, + "learning_rate": 0.0001673937369975465, + "loss": 0.6226, + "step": 11295 + }, + { + "epoch": 0.26479602899722804, + "grad_norm": 0.41941767930984497, + "learning_rate": 0.0001673882947821476, + "loss": 0.0832, + "step": 11296 + }, + { + "epoch": 0.26481947057203303, + "grad_norm": 0.49246111512184143, + "learning_rate": 0.00016738285220109796, + "loss": 0.7369, + "step": 11297 + }, + { + "epoch": 0.264842912146838, + "grad_norm": 0.1307586133480072, + "learning_rate": 0.00016737740925442705, + "loss": 0.0295, + "step": 11298 + }, + { + "epoch": 0.264866353721643, + "grad_norm": 0.45529666543006897, + "learning_rate": 0.00016737196594216445, + "loss": 0.0639, + "step": 11299 + }, + { + "epoch": 0.264889795296448, + "grad_norm": 0.7873699069023132, + "learning_rate": 0.00016736652226433964, + "loss": 0.1624, + "step": 11300 + }, + { + "epoch": 0.264913236871253, + "grad_norm": 0.375659316778183, + "learning_rate": 0.0001673610782209822, + "loss": 0.0952, + "step": 11301 + }, + { + "epoch": 0.264936678446058, + "grad_norm": 0.3668907582759857, + "learning_rate": 0.00016735563381212166, + "loss": 0.0818, + "step": 11302 + }, + { + "epoch": 0.264960120020863, + "grad_norm": 0.6775267720222473, + "learning_rate": 0.00016735018903778751, + "loss": 0.1681, + "step": 11303 + }, + { + "epoch": 0.264983561595668, + "grad_norm": 0.3262353241443634, + "learning_rate": 0.0001673447438980094, + "loss": 0.057, + "step": 11304 + }, + { + "epoch": 0.265007003170473, + "grad_norm": 0.6058533191680908, + "learning_rate": 0.0001673392983928168, + "loss": 0.6243, + "step": 11305 + }, + { + "epoch": 0.265030444745278, + "grad_norm": 0.4884392023086548, + "learning_rate": 0.00016733385252223925, + "loss": 0.1539, + "step": 11306 + }, + { + "epoch": 0.265053886320083, + "grad_norm": 0.3587169349193573, + "learning_rate": 0.00016732840628630634, + "loss": 0.0536, + "step": 11307 + }, + { + "epoch": 0.26507732789488797, + "grad_norm": 0.2525702118873596, + "learning_rate": 0.0001673229596850476, + "loss": 0.0584, + "step": 11308 + }, + { + "epoch": 0.26510076946969297, + "grad_norm": 0.4282258450984955, + "learning_rate": 0.00016731751271849258, + "loss": 0.377, + "step": 11309 + }, + { + "epoch": 0.26512421104449796, + "grad_norm": 0.4932349622249603, + "learning_rate": 0.00016731206538667088, + "loss": 0.5806, + "step": 11310 + }, + { + "epoch": 0.26514765261930295, + "grad_norm": 0.1801937073469162, + "learning_rate": 0.000167306617689612, + "loss": 0.0264, + "step": 11311 + }, + { + "epoch": 0.26517109419410795, + "grad_norm": 0.5015469789505005, + "learning_rate": 0.00016730116962734553, + "loss": 0.1126, + "step": 11312 + }, + { + "epoch": 0.26519453576891294, + "grad_norm": 0.8633090853691101, + "learning_rate": 0.00016729572119990102, + "loss": 0.1042, + "step": 11313 + }, + { + "epoch": 0.26521797734371794, + "grad_norm": 0.5705689191818237, + "learning_rate": 0.00016729027240730798, + "loss": 0.7483, + "step": 11314 + }, + { + "epoch": 0.26524141891852293, + "grad_norm": 0.5924447178840637, + "learning_rate": 0.0001672848232495961, + "loss": 0.1579, + "step": 11315 + }, + { + "epoch": 0.2652648604933279, + "grad_norm": 0.6253990530967712, + "learning_rate": 0.00016727937372679484, + "loss": 0.6823, + "step": 11316 + }, + { + "epoch": 0.2652883020681329, + "grad_norm": 0.7099354863166809, + "learning_rate": 0.0001672739238389338, + "loss": 0.1493, + "step": 11317 + }, + { + "epoch": 0.2653117436429379, + "grad_norm": 0.3350987434387207, + "learning_rate": 0.0001672684735860426, + "loss": 0.0998, + "step": 11318 + }, + { + "epoch": 0.2653351852177429, + "grad_norm": 0.49152979254722595, + "learning_rate": 0.00016726302296815075, + "loss": 0.1298, + "step": 11319 + }, + { + "epoch": 0.2653586267925479, + "grad_norm": 0.7118203043937683, + "learning_rate": 0.00016725757198528783, + "loss": 0.1912, + "step": 11320 + }, + { + "epoch": 0.2653820683673529, + "grad_norm": 0.16757269203662872, + "learning_rate": 0.00016725212063748343, + "loss": 0.0375, + "step": 11321 + }, + { + "epoch": 0.2654055099421579, + "grad_norm": 0.46878451108932495, + "learning_rate": 0.00016724666892476716, + "loss": 0.0777, + "step": 11322 + }, + { + "epoch": 0.2654289515169629, + "grad_norm": 0.45881980657577515, + "learning_rate": 0.00016724121684716855, + "loss": 0.1094, + "step": 11323 + }, + { + "epoch": 0.2654523930917679, + "grad_norm": 0.4936524033546448, + "learning_rate": 0.00016723576440471724, + "loss": 0.11, + "step": 11324 + }, + { + "epoch": 0.2654758346665729, + "grad_norm": 1.165604829788208, + "learning_rate": 0.00016723031159744276, + "loss": 0.1318, + "step": 11325 + }, + { + "epoch": 0.26549927624137787, + "grad_norm": 0.5176200270652771, + "learning_rate": 0.0001672248584253747, + "loss": 0.0803, + "step": 11326 + }, + { + "epoch": 0.2655227178161829, + "grad_norm": 0.520946204662323, + "learning_rate": 0.0001672194048885427, + "loss": 0.0794, + "step": 11327 + }, + { + "epoch": 0.2655461593909879, + "grad_norm": 0.3203471004962921, + "learning_rate": 0.0001672139509869763, + "loss": 0.0988, + "step": 11328 + }, + { + "epoch": 0.2655696009657929, + "grad_norm": 0.7996640205383301, + "learning_rate": 0.00016720849672070514, + "loss": 0.173, + "step": 11329 + }, + { + "epoch": 0.2655930425405979, + "grad_norm": 0.23412957787513733, + "learning_rate": 0.00016720304208975875, + "loss": 0.0631, + "step": 11330 + }, + { + "epoch": 0.2656164841154029, + "grad_norm": 0.5602070689201355, + "learning_rate": 0.00016719758709416674, + "loss": 0.0813, + "step": 11331 + }, + { + "epoch": 0.2656399256902079, + "grad_norm": 0.40366026759147644, + "learning_rate": 0.0001671921317339588, + "loss": 0.0847, + "step": 11332 + }, + { + "epoch": 0.2656633672650129, + "grad_norm": 0.6128084659576416, + "learning_rate": 0.0001671866760091644, + "loss": 0.656, + "step": 11333 + }, + { + "epoch": 0.2656868088398179, + "grad_norm": 0.3524670898914337, + "learning_rate": 0.0001671812199198133, + "loss": 0.1021, + "step": 11334 + }, + { + "epoch": 0.2657102504146229, + "grad_norm": 1.0243451595306396, + "learning_rate": 0.00016717576346593493, + "loss": 0.121, + "step": 11335 + }, + { + "epoch": 0.26573369198942787, + "grad_norm": 0.38082870841026306, + "learning_rate": 0.000167170306647559, + "loss": 0.1095, + "step": 11336 + }, + { + "epoch": 0.26575713356423286, + "grad_norm": 0.13466688990592957, + "learning_rate": 0.00016716484946471512, + "loss": 0.0379, + "step": 11337 + }, + { + "epoch": 0.26578057513903786, + "grad_norm": 0.17306353151798248, + "learning_rate": 0.00016715939191743284, + "loss": 0.0352, + "step": 11338 + }, + { + "epoch": 0.26580401671384285, + "grad_norm": 1.1412584781646729, + "learning_rate": 0.00016715393400574185, + "loss": 0.2647, + "step": 11339 + }, + { + "epoch": 0.26582745828864784, + "grad_norm": 0.5624119639396667, + "learning_rate": 0.00016714847572967169, + "loss": 0.1378, + "step": 11340 + }, + { + "epoch": 0.26585089986345284, + "grad_norm": 0.5982601046562195, + "learning_rate": 0.00016714301708925202, + "loss": 0.215, + "step": 11341 + }, + { + "epoch": 0.26587434143825783, + "grad_norm": 0.49731895327568054, + "learning_rate": 0.00016713755808451245, + "loss": 0.1368, + "step": 11342 + }, + { + "epoch": 0.2658977830130628, + "grad_norm": 0.464184433221817, + "learning_rate": 0.0001671320987154826, + "loss": 0.0803, + "step": 11343 + }, + { + "epoch": 0.2659212245878678, + "grad_norm": 0.207984060049057, + "learning_rate": 0.0001671266389821921, + "loss": 0.0347, + "step": 11344 + }, + { + "epoch": 0.2659446661626728, + "grad_norm": 0.4477972388267517, + "learning_rate": 0.0001671211788846706, + "loss": 0.1583, + "step": 11345 + }, + { + "epoch": 0.2659681077374778, + "grad_norm": 0.24240221083164215, + "learning_rate": 0.00016711571842294767, + "loss": 0.0449, + "step": 11346 + }, + { + "epoch": 0.2659915493122828, + "grad_norm": 0.5884044170379639, + "learning_rate": 0.00016711025759705297, + "loss": 0.1515, + "step": 11347 + }, + { + "epoch": 0.2660149908870878, + "grad_norm": 0.26309967041015625, + "learning_rate": 0.00016710479640701613, + "loss": 0.0578, + "step": 11348 + }, + { + "epoch": 0.2660384324618928, + "grad_norm": 0.6004077792167664, + "learning_rate": 0.00016709933485286674, + "loss": 0.6334, + "step": 11349 + }, + { + "epoch": 0.2660618740366978, + "grad_norm": 0.6268410682678223, + "learning_rate": 0.0001670938729346345, + "loss": 0.5641, + "step": 11350 + }, + { + "epoch": 0.2660853156115028, + "grad_norm": 0.435531884431839, + "learning_rate": 0.00016708841065234904, + "loss": 0.0795, + "step": 11351 + }, + { + "epoch": 0.2661087571863078, + "grad_norm": 0.38763973116874695, + "learning_rate": 0.00016708294800603995, + "loss": 0.0995, + "step": 11352 + }, + { + "epoch": 0.26613219876111277, + "grad_norm": 0.3825923502445221, + "learning_rate": 0.00016707748499573692, + "loss": 0.1073, + "step": 11353 + }, + { + "epoch": 0.26615564033591776, + "grad_norm": 0.6028677225112915, + "learning_rate": 0.00016707202162146955, + "loss": 0.1392, + "step": 11354 + }, + { + "epoch": 0.26617908191072276, + "grad_norm": 0.6666698455810547, + "learning_rate": 0.00016706655788326753, + "loss": 0.1722, + "step": 11355 + }, + { + "epoch": 0.26620252348552775, + "grad_norm": 0.29381871223449707, + "learning_rate": 0.00016706109378116047, + "loss": 0.049, + "step": 11356 + }, + { + "epoch": 0.26622596506033275, + "grad_norm": 0.9398775696754456, + "learning_rate": 0.00016705562931517798, + "loss": 0.1855, + "step": 11357 + }, + { + "epoch": 0.26624940663513774, + "grad_norm": 1.1845823526382446, + "learning_rate": 0.00016705016448534981, + "loss": 0.1406, + "step": 11358 + }, + { + "epoch": 0.26627284820994274, + "grad_norm": 0.4777280390262604, + "learning_rate": 0.00016704469929170558, + "loss": 0.1094, + "step": 11359 + }, + { + "epoch": 0.26629628978474773, + "grad_norm": 0.5582742691040039, + "learning_rate": 0.0001670392337342749, + "loss": 0.1765, + "step": 11360 + }, + { + "epoch": 0.2663197313595527, + "grad_norm": 0.6266387701034546, + "learning_rate": 0.00016703376781308744, + "loss": 0.1342, + "step": 11361 + }, + { + "epoch": 0.2663431729343577, + "grad_norm": 0.5239847898483276, + "learning_rate": 0.00016702830152817287, + "loss": 0.6225, + "step": 11362 + }, + { + "epoch": 0.2663666145091627, + "grad_norm": 1.8774793148040771, + "learning_rate": 0.00016702283487956087, + "loss": 0.0954, + "step": 11363 + }, + { + "epoch": 0.2663900560839677, + "grad_norm": 0.33580198884010315, + "learning_rate": 0.0001670173678672811, + "loss": 0.0533, + "step": 11364 + }, + { + "epoch": 0.2664134976587727, + "grad_norm": 0.47727853059768677, + "learning_rate": 0.00016701190049136317, + "loss": 0.1009, + "step": 11365 + }, + { + "epoch": 0.2664369392335777, + "grad_norm": 0.7464223504066467, + "learning_rate": 0.0001670064327518368, + "loss": 0.6639, + "step": 11366 + }, + { + "epoch": 0.2664603808083827, + "grad_norm": 0.40925106406211853, + "learning_rate": 0.00016700096464873166, + "loss": 0.0898, + "step": 11367 + }, + { + "epoch": 0.2664838223831877, + "grad_norm": 0.4082413911819458, + "learning_rate": 0.00016699549618207738, + "loss": 0.0695, + "step": 11368 + }, + { + "epoch": 0.2665072639579927, + "grad_norm": 0.49343693256378174, + "learning_rate": 0.00016699002735190365, + "loss": 0.1757, + "step": 11369 + }, + { + "epoch": 0.2665307055327977, + "grad_norm": 0.635323166847229, + "learning_rate": 0.00016698455815824015, + "loss": 0.8449, + "step": 11370 + }, + { + "epoch": 0.26655414710760267, + "grad_norm": 0.6639947295188904, + "learning_rate": 0.00016697908860111658, + "loss": 0.0656, + "step": 11371 + }, + { + "epoch": 0.26657758868240766, + "grad_norm": 0.3209404945373535, + "learning_rate": 0.00016697361868056257, + "loss": 0.068, + "step": 11372 + }, + { + "epoch": 0.26660103025721266, + "grad_norm": 0.6598321199417114, + "learning_rate": 0.0001669681483966078, + "loss": 0.1005, + "step": 11373 + }, + { + "epoch": 0.26662447183201765, + "grad_norm": 0.4728342890739441, + "learning_rate": 0.000166962677749282, + "loss": 0.1111, + "step": 11374 + }, + { + "epoch": 0.26664791340682265, + "grad_norm": 0.0836276039481163, + "learning_rate": 0.00016695720673861483, + "loss": 0.0134, + "step": 11375 + }, + { + "epoch": 0.26667135498162764, + "grad_norm": 0.5979195833206177, + "learning_rate": 0.00016695173536463597, + "loss": 0.1022, + "step": 11376 + }, + { + "epoch": 0.26669479655643263, + "grad_norm": 0.40267083048820496, + "learning_rate": 0.00016694626362737512, + "loss": 0.0881, + "step": 11377 + }, + { + "epoch": 0.2667182381312377, + "grad_norm": 0.1669238805770874, + "learning_rate": 0.00016694079152686193, + "loss": 0.013, + "step": 11378 + }, + { + "epoch": 0.2667416797060427, + "grad_norm": 0.1522228866815567, + "learning_rate": 0.00016693531906312615, + "loss": 0.0424, + "step": 11379 + }, + { + "epoch": 0.2667651212808477, + "grad_norm": 0.35170087218284607, + "learning_rate": 0.00016692984623619745, + "loss": 0.0914, + "step": 11380 + }, + { + "epoch": 0.26678856285565267, + "grad_norm": 0.40930670499801636, + "learning_rate": 0.0001669243730461055, + "loss": 0.0935, + "step": 11381 + }, + { + "epoch": 0.26681200443045766, + "grad_norm": 0.4867016077041626, + "learning_rate": 0.00016691889949288003, + "loss": 0.096, + "step": 11382 + }, + { + "epoch": 0.26683544600526266, + "grad_norm": 0.2669858932495117, + "learning_rate": 0.00016691342557655076, + "loss": 0.0577, + "step": 11383 + }, + { + "epoch": 0.26685888758006765, + "grad_norm": 0.5888375639915466, + "learning_rate": 0.00016690795129714733, + "loss": 0.1395, + "step": 11384 + }, + { + "epoch": 0.26688232915487264, + "grad_norm": 0.13023898005485535, + "learning_rate": 0.00016690247665469946, + "loss": 0.0236, + "step": 11385 + }, + { + "epoch": 0.26690577072967764, + "grad_norm": 0.48935192823410034, + "learning_rate": 0.00016689700164923692, + "loss": 0.5738, + "step": 11386 + }, + { + "epoch": 0.26692921230448263, + "grad_norm": 0.5890626907348633, + "learning_rate": 0.0001668915262807893, + "loss": 0.0796, + "step": 11387 + }, + { + "epoch": 0.2669526538792876, + "grad_norm": 0.9378913044929504, + "learning_rate": 0.00016688605054938644, + "loss": 0.0511, + "step": 11388 + }, + { + "epoch": 0.2669760954540926, + "grad_norm": 0.5565419793128967, + "learning_rate": 0.00016688057445505796, + "loss": 0.8064, + "step": 11389 + }, + { + "epoch": 0.2669995370288976, + "grad_norm": 0.4003426730632782, + "learning_rate": 0.00016687509799783362, + "loss": 0.0968, + "step": 11390 + }, + { + "epoch": 0.2670229786037026, + "grad_norm": 0.3976491391658783, + "learning_rate": 0.0001668696211777431, + "loss": 0.0973, + "step": 11391 + }, + { + "epoch": 0.2670464201785076, + "grad_norm": 0.3642350733280182, + "learning_rate": 0.00016686414399481615, + "loss": 0.0722, + "step": 11392 + }, + { + "epoch": 0.2670698617533126, + "grad_norm": 0.6340997815132141, + "learning_rate": 0.00016685866644908245, + "loss": 0.1676, + "step": 11393 + }, + { + "epoch": 0.2670933033281176, + "grad_norm": 0.25172173976898193, + "learning_rate": 0.00016685318854057176, + "loss": 0.0721, + "step": 11394 + }, + { + "epoch": 0.2671167449029226, + "grad_norm": 0.4161185920238495, + "learning_rate": 0.0001668477102693138, + "loss": 0.3864, + "step": 11395 + }, + { + "epoch": 0.2671401864777276, + "grad_norm": 0.6725022792816162, + "learning_rate": 0.00016684223163533827, + "loss": 0.1196, + "step": 11396 + }, + { + "epoch": 0.2671636280525326, + "grad_norm": 0.4740893244743347, + "learning_rate": 0.00016683675263867492, + "loss": 0.1157, + "step": 11397 + }, + { + "epoch": 0.26718706962733757, + "grad_norm": 0.5781293511390686, + "learning_rate": 0.00016683127327935345, + "loss": 0.1212, + "step": 11398 + }, + { + "epoch": 0.26721051120214256, + "grad_norm": 0.10561169683933258, + "learning_rate": 0.00016682579355740363, + "loss": 0.0201, + "step": 11399 + }, + { + "epoch": 0.26723395277694756, + "grad_norm": 0.23362401127815247, + "learning_rate": 0.00016682031347285518, + "loss": 0.0732, + "step": 11400 + }, + { + "epoch": 0.26725739435175255, + "grad_norm": 0.7447993159294128, + "learning_rate": 0.00016681483302573785, + "loss": 0.1948, + "step": 11401 + }, + { + "epoch": 0.26728083592655755, + "grad_norm": 0.5434473156929016, + "learning_rate": 0.00016680935221608133, + "loss": 0.0916, + "step": 11402 + }, + { + "epoch": 0.26730427750136254, + "grad_norm": 0.5375645160675049, + "learning_rate": 0.0001668038710439154, + "loss": 0.0742, + "step": 11403 + }, + { + "epoch": 0.26732771907616754, + "grad_norm": 0.305620402097702, + "learning_rate": 0.00016679838950926977, + "loss": 0.0955, + "step": 11404 + }, + { + "epoch": 0.26735116065097253, + "grad_norm": 0.31612256169319153, + "learning_rate": 0.0001667929076121742, + "loss": 0.036, + "step": 11405 + }, + { + "epoch": 0.2673746022257775, + "grad_norm": 1.2637666463851929, + "learning_rate": 0.00016678742535265842, + "loss": 0.1017, + "step": 11406 + }, + { + "epoch": 0.2673980438005825, + "grad_norm": 0.6055604219436646, + "learning_rate": 0.00016678194273075223, + "loss": 0.6494, + "step": 11407 + }, + { + "epoch": 0.2674214853753875, + "grad_norm": 0.3503898084163666, + "learning_rate": 0.00016677645974648534, + "loss": 0.0408, + "step": 11408 + }, + { + "epoch": 0.2674449269501925, + "grad_norm": 0.326006144285202, + "learning_rate": 0.0001667709763998875, + "loss": 0.0993, + "step": 11409 + }, + { + "epoch": 0.2674683685249975, + "grad_norm": 0.5765440464019775, + "learning_rate": 0.0001667654926909884, + "loss": 0.6633, + "step": 11410 + }, + { + "epoch": 0.2674918100998025, + "grad_norm": 0.5114235281944275, + "learning_rate": 0.00016676000861981792, + "loss": 0.0824, + "step": 11411 + }, + { + "epoch": 0.2675152516746075, + "grad_norm": 0.21100151538848877, + "learning_rate": 0.00016675452418640576, + "loss": 0.0433, + "step": 11412 + }, + { + "epoch": 0.2675386932494125, + "grad_norm": 0.4098925292491913, + "learning_rate": 0.00016674903939078165, + "loss": 0.0717, + "step": 11413 + }, + { + "epoch": 0.2675621348242175, + "grad_norm": 0.5967240333557129, + "learning_rate": 0.0001667435542329754, + "loss": 0.1484, + "step": 11414 + }, + { + "epoch": 0.2675855763990225, + "grad_norm": 0.5545629262924194, + "learning_rate": 0.00016673806871301674, + "loss": 0.131, + "step": 11415 + }, + { + "epoch": 0.26760901797382747, + "grad_norm": 0.3418125808238983, + "learning_rate": 0.00016673258283093545, + "loss": 0.0634, + "step": 11416 + }, + { + "epoch": 0.26763245954863246, + "grad_norm": 1.129927635192871, + "learning_rate": 0.00016672709658676124, + "loss": 0.6903, + "step": 11417 + }, + { + "epoch": 0.26765590112343746, + "grad_norm": 0.24444715678691864, + "learning_rate": 0.00016672160998052398, + "loss": 0.0203, + "step": 11418 + }, + { + "epoch": 0.26767934269824245, + "grad_norm": 0.8176482319831848, + "learning_rate": 0.00016671612301225337, + "loss": 0.1321, + "step": 11419 + }, + { + "epoch": 0.26770278427304744, + "grad_norm": 0.4532478451728821, + "learning_rate": 0.0001667106356819792, + "loss": 0.0924, + "step": 11420 + }, + { + "epoch": 0.26772622584785244, + "grad_norm": 0.25622889399528503, + "learning_rate": 0.00016670514798973124, + "loss": 0.0511, + "step": 11421 + }, + { + "epoch": 0.26774966742265743, + "grad_norm": 0.4794962406158447, + "learning_rate": 0.00016669965993553932, + "loss": 0.0932, + "step": 11422 + }, + { + "epoch": 0.26777310899746243, + "grad_norm": 0.9270980358123779, + "learning_rate": 0.0001666941715194331, + "loss": 0.2067, + "step": 11423 + }, + { + "epoch": 0.2677965505722674, + "grad_norm": 0.5228202939033508, + "learning_rate": 0.00016668868274144244, + "loss": 0.1024, + "step": 11424 + }, + { + "epoch": 0.2678199921470724, + "grad_norm": 0.5743786096572876, + "learning_rate": 0.00016668319360159717, + "loss": 0.1105, + "step": 11425 + }, + { + "epoch": 0.2678434337218774, + "grad_norm": 0.6773889660835266, + "learning_rate": 0.00016667770409992697, + "loss": 0.6422, + "step": 11426 + }, + { + "epoch": 0.2678668752966824, + "grad_norm": 0.5087195038795471, + "learning_rate": 0.00016667221423646167, + "loss": 0.0628, + "step": 11427 + }, + { + "epoch": 0.2678903168714874, + "grad_norm": 0.9465935826301575, + "learning_rate": 0.0001666667240112311, + "loss": 0.0466, + "step": 11428 + }, + { + "epoch": 0.26791375844629245, + "grad_norm": 0.6928030252456665, + "learning_rate": 0.000166661233424265, + "loss": 0.0727, + "step": 11429 + }, + { + "epoch": 0.26793720002109744, + "grad_norm": 0.41665658354759216, + "learning_rate": 0.0001666557424755931, + "loss": 0.0962, + "step": 11430 + }, + { + "epoch": 0.26796064159590244, + "grad_norm": 0.18362438678741455, + "learning_rate": 0.00016665025116524533, + "loss": 0.0522, + "step": 11431 + }, + { + "epoch": 0.26798408317070743, + "grad_norm": 0.10498866438865662, + "learning_rate": 0.0001666447594932514, + "loss": 0.0173, + "step": 11432 + }, + { + "epoch": 0.2680075247455124, + "grad_norm": 0.5157957673072815, + "learning_rate": 0.0001666392674596412, + "loss": 0.1295, + "step": 11433 + }, + { + "epoch": 0.2680309663203174, + "grad_norm": 0.41744735836982727, + "learning_rate": 0.0001666337750644444, + "loss": 0.0646, + "step": 11434 + }, + { + "epoch": 0.2680544078951224, + "grad_norm": 0.8173471093177795, + "learning_rate": 0.0001666282823076909, + "loss": 0.1794, + "step": 11435 + }, + { + "epoch": 0.2680778494699274, + "grad_norm": 0.6964301466941833, + "learning_rate": 0.00016662278918941041, + "loss": 0.1313, + "step": 11436 + }, + { + "epoch": 0.2681012910447324, + "grad_norm": 0.5269711017608643, + "learning_rate": 0.0001666172957096328, + "loss": 0.1253, + "step": 11437 + }, + { + "epoch": 0.2681247326195374, + "grad_norm": 0.14702366292476654, + "learning_rate": 0.0001666118018683879, + "loss": 0.0227, + "step": 11438 + }, + { + "epoch": 0.2681481741943424, + "grad_norm": 0.5562058091163635, + "learning_rate": 0.0001666063076657055, + "loss": 0.123, + "step": 11439 + }, + { + "epoch": 0.2681716157691474, + "grad_norm": 0.21088451147079468, + "learning_rate": 0.00016660081310161537, + "loss": 0.0416, + "step": 11440 + }, + { + "epoch": 0.2681950573439524, + "grad_norm": 0.7991693615913391, + "learning_rate": 0.00016659531817614736, + "loss": 0.1887, + "step": 11441 + }, + { + "epoch": 0.2682184989187574, + "grad_norm": 0.11176402866840363, + "learning_rate": 0.0001665898228893313, + "loss": 0.0256, + "step": 11442 + }, + { + "epoch": 0.26824194049356237, + "grad_norm": 0.21356242895126343, + "learning_rate": 0.00016658432724119698, + "loss": 0.0695, + "step": 11443 + }, + { + "epoch": 0.26826538206836736, + "grad_norm": 0.5741231441497803, + "learning_rate": 0.0001665788312317742, + "loss": 0.1325, + "step": 11444 + }, + { + "epoch": 0.26828882364317236, + "grad_norm": 0.2822437882423401, + "learning_rate": 0.00016657333486109287, + "loss": 0.063, + "step": 11445 + }, + { + "epoch": 0.26831226521797735, + "grad_norm": 0.4792311191558838, + "learning_rate": 0.0001665678381291827, + "loss": 0.0916, + "step": 11446 + }, + { + "epoch": 0.26833570679278235, + "grad_norm": 0.6251731514930725, + "learning_rate": 0.0001665623410360736, + "loss": 0.8328, + "step": 11447 + }, + { + "epoch": 0.26835914836758734, + "grad_norm": 0.5768560171127319, + "learning_rate": 0.00016655684358179534, + "loss": 0.1005, + "step": 11448 + }, + { + "epoch": 0.26838258994239234, + "grad_norm": 0.23539534211158752, + "learning_rate": 0.0001665513457663778, + "loss": 0.05, + "step": 11449 + }, + { + "epoch": 0.26840603151719733, + "grad_norm": 0.30451682209968567, + "learning_rate": 0.0001665458475898508, + "loss": 0.0664, + "step": 11450 + }, + { + "epoch": 0.2684294730920023, + "grad_norm": 0.08019159734249115, + "learning_rate": 0.00016654034905224413, + "loss": 0.0169, + "step": 11451 + }, + { + "epoch": 0.2684529146668073, + "grad_norm": 0.35209405422210693, + "learning_rate": 0.00016653485015358764, + "loss": 0.3192, + "step": 11452 + }, + { + "epoch": 0.2684763562416123, + "grad_norm": 0.4143047332763672, + "learning_rate": 0.0001665293508939112, + "loss": 0.0953, + "step": 11453 + }, + { + "epoch": 0.2684997978164173, + "grad_norm": 0.6983429789543152, + "learning_rate": 0.00016652385127324462, + "loss": 0.1825, + "step": 11454 + }, + { + "epoch": 0.2685232393912223, + "grad_norm": 0.22916807234287262, + "learning_rate": 0.0001665183512916178, + "loss": 0.0568, + "step": 11455 + }, + { + "epoch": 0.2685466809660273, + "grad_norm": 0.37324967980384827, + "learning_rate": 0.00016651285094906052, + "loss": 0.0696, + "step": 11456 + }, + { + "epoch": 0.2685701225408323, + "grad_norm": 0.25604262948036194, + "learning_rate": 0.0001665073502456026, + "loss": 0.0647, + "step": 11457 + }, + { + "epoch": 0.2685935641156373, + "grad_norm": 0.5693807005882263, + "learning_rate": 0.00016650184918127393, + "loss": 0.792, + "step": 11458 + }, + { + "epoch": 0.2686170056904423, + "grad_norm": 0.6699520945549011, + "learning_rate": 0.00016649634775610438, + "loss": 0.1197, + "step": 11459 + }, + { + "epoch": 0.2686404472652473, + "grad_norm": 0.6239383816719055, + "learning_rate": 0.0001664908459701238, + "loss": 0.1627, + "step": 11460 + }, + { + "epoch": 0.26866388884005227, + "grad_norm": 0.12034525722265244, + "learning_rate": 0.000166485343823362, + "loss": 0.0311, + "step": 11461 + }, + { + "epoch": 0.26868733041485726, + "grad_norm": 0.5275018215179443, + "learning_rate": 0.0001664798413158488, + "loss": 0.1625, + "step": 11462 + }, + { + "epoch": 0.26871077198966226, + "grad_norm": 0.290448933839798, + "learning_rate": 0.0001664743384476142, + "loss": 0.0262, + "step": 11463 + }, + { + "epoch": 0.26873421356446725, + "grad_norm": 0.3774709403514862, + "learning_rate": 0.00016646883521868793, + "loss": 0.0434, + "step": 11464 + }, + { + "epoch": 0.26875765513927224, + "grad_norm": 0.4443769156932831, + "learning_rate": 0.0001664633316290999, + "loss": 0.0852, + "step": 11465 + }, + { + "epoch": 0.26878109671407724, + "grad_norm": 0.20739805698394775, + "learning_rate": 0.00016645782767887995, + "loss": 0.0625, + "step": 11466 + }, + { + "epoch": 0.26880453828888223, + "grad_norm": 0.7868697643280029, + "learning_rate": 0.00016645232336805796, + "loss": 0.1784, + "step": 11467 + }, + { + "epoch": 0.2688279798636872, + "grad_norm": 0.40147075057029724, + "learning_rate": 0.00016644681869666382, + "loss": 0.0608, + "step": 11468 + }, + { + "epoch": 0.2688514214384922, + "grad_norm": 0.6604676842689514, + "learning_rate": 0.00016644131366472736, + "loss": 0.1573, + "step": 11469 + }, + { + "epoch": 0.2688748630132972, + "grad_norm": 0.3522365391254425, + "learning_rate": 0.00016643580827227841, + "loss": 0.0496, + "step": 11470 + }, + { + "epoch": 0.2688983045881022, + "grad_norm": 0.3499758243560791, + "learning_rate": 0.000166430302519347, + "loss": 0.069, + "step": 11471 + }, + { + "epoch": 0.2689217461629072, + "grad_norm": 0.6910494565963745, + "learning_rate": 0.00016642479640596283, + "loss": 0.2057, + "step": 11472 + }, + { + "epoch": 0.2689451877377122, + "grad_norm": 0.1422085165977478, + "learning_rate": 0.00016641928993215585, + "loss": 0.0177, + "step": 11473 + }, + { + "epoch": 0.2689686293125172, + "grad_norm": 0.2085602879524231, + "learning_rate": 0.00016641378309795594, + "loss": 0.1679, + "step": 11474 + }, + { + "epoch": 0.2689920708873222, + "grad_norm": 0.8225347399711609, + "learning_rate": 0.000166408275903393, + "loss": 0.1917, + "step": 11475 + }, + { + "epoch": 0.2690155124621272, + "grad_norm": 0.34389498829841614, + "learning_rate": 0.00016640276834849687, + "loss": 0.4233, + "step": 11476 + }, + { + "epoch": 0.2690389540369322, + "grad_norm": 0.4096132218837738, + "learning_rate": 0.00016639726043329744, + "loss": 0.0858, + "step": 11477 + }, + { + "epoch": 0.26906239561173717, + "grad_norm": 0.7376437187194824, + "learning_rate": 0.0001663917521578246, + "loss": 0.1834, + "step": 11478 + }, + { + "epoch": 0.26908583718654216, + "grad_norm": 0.5438116192817688, + "learning_rate": 0.00016638624352210827, + "loss": 0.0619, + "step": 11479 + }, + { + "epoch": 0.2691092787613472, + "grad_norm": 0.36109599471092224, + "learning_rate": 0.00016638073452617834, + "loss": 0.049, + "step": 11480 + }, + { + "epoch": 0.2691327203361522, + "grad_norm": 0.3832874000072479, + "learning_rate": 0.00016637522517006463, + "loss": 0.0475, + "step": 11481 + }, + { + "epoch": 0.2691561619109572, + "grad_norm": 0.3896692395210266, + "learning_rate": 0.00016636971545379712, + "loss": 0.0665, + "step": 11482 + }, + { + "epoch": 0.2691796034857622, + "grad_norm": 0.5354982614517212, + "learning_rate": 0.00016636420537740566, + "loss": 0.1363, + "step": 11483 + }, + { + "epoch": 0.2692030450605672, + "grad_norm": 0.32238662242889404, + "learning_rate": 0.0001663586949409201, + "loss": 0.0893, + "step": 11484 + }, + { + "epoch": 0.2692264866353722, + "grad_norm": 0.5745851397514343, + "learning_rate": 0.00016635318414437046, + "loss": 0.0755, + "step": 11485 + }, + { + "epoch": 0.2692499282101772, + "grad_norm": 0.28168514370918274, + "learning_rate": 0.00016634767298778658, + "loss": 0.0491, + "step": 11486 + }, + { + "epoch": 0.2692733697849822, + "grad_norm": 0.6907855272293091, + "learning_rate": 0.00016634216147119832, + "loss": 0.1886, + "step": 11487 + }, + { + "epoch": 0.26929681135978717, + "grad_norm": 0.5226420164108276, + "learning_rate": 0.00016633664959463565, + "loss": 0.1113, + "step": 11488 + }, + { + "epoch": 0.26932025293459216, + "grad_norm": 0.5351573824882507, + "learning_rate": 0.00016633113735812846, + "loss": 0.138, + "step": 11489 + }, + { + "epoch": 0.26934369450939716, + "grad_norm": 0.46909117698669434, + "learning_rate": 0.00016632562476170663, + "loss": 0.1175, + "step": 11490 + }, + { + "epoch": 0.26936713608420215, + "grad_norm": 0.6052258014678955, + "learning_rate": 0.00016632011180540014, + "loss": 0.0727, + "step": 11491 + }, + { + "epoch": 0.26939057765900715, + "grad_norm": 0.6126825213432312, + "learning_rate": 0.0001663145984892388, + "loss": 0.0637, + "step": 11492 + }, + { + "epoch": 0.26941401923381214, + "grad_norm": 0.5812461972236633, + "learning_rate": 0.0001663090848132526, + "loss": 0.0853, + "step": 11493 + }, + { + "epoch": 0.26943746080861714, + "grad_norm": 0.5126211643218994, + "learning_rate": 0.00016630357077747148, + "loss": 0.0958, + "step": 11494 + }, + { + "epoch": 0.26946090238342213, + "grad_norm": 0.6031871438026428, + "learning_rate": 0.00016629805638192528, + "loss": 0.1627, + "step": 11495 + }, + { + "epoch": 0.2694843439582271, + "grad_norm": 0.527385413646698, + "learning_rate": 0.00016629254162664397, + "loss": 0.1666, + "step": 11496 + }, + { + "epoch": 0.2695077855330321, + "grad_norm": 0.18863500654697418, + "learning_rate": 0.0001662870265116575, + "loss": 0.0562, + "step": 11497 + }, + { + "epoch": 0.2695312271078371, + "grad_norm": 0.3656337857246399, + "learning_rate": 0.0001662815110369957, + "loss": 0.0749, + "step": 11498 + }, + { + "epoch": 0.2695546686826421, + "grad_norm": 0.26177963614463806, + "learning_rate": 0.00016627599520268859, + "loss": 0.0563, + "step": 11499 + }, + { + "epoch": 0.2695781102574471, + "grad_norm": 0.4265854060649872, + "learning_rate": 0.00016627047900876608, + "loss": 0.1439, + "step": 11500 + }, + { + "epoch": 0.2696015518322521, + "grad_norm": 0.5258122682571411, + "learning_rate": 0.00016626496245525808, + "loss": 0.1223, + "step": 11501 + }, + { + "epoch": 0.2696249934070571, + "grad_norm": 0.12685222923755646, + "learning_rate": 0.0001662594455421945, + "loss": 0.0149, + "step": 11502 + }, + { + "epoch": 0.2696484349818621, + "grad_norm": 0.4962525963783264, + "learning_rate": 0.0001662539282696053, + "loss": 0.4699, + "step": 11503 + }, + { + "epoch": 0.2696718765566671, + "grad_norm": 0.41107141971588135, + "learning_rate": 0.00016624841063752049, + "loss": 0.0596, + "step": 11504 + }, + { + "epoch": 0.2696953181314721, + "grad_norm": 0.23437462747097015, + "learning_rate": 0.0001662428926459699, + "loss": 0.0382, + "step": 11505 + }, + { + "epoch": 0.26971875970627707, + "grad_norm": 0.4213629961013794, + "learning_rate": 0.0001662373742949835, + "loss": 0.0582, + "step": 11506 + }, + { + "epoch": 0.26974220128108206, + "grad_norm": 0.2200252264738083, + "learning_rate": 0.00016623185558459127, + "loss": 0.0511, + "step": 11507 + }, + { + "epoch": 0.26976564285588706, + "grad_norm": 0.6012821197509766, + "learning_rate": 0.00016622633651482312, + "loss": 0.1677, + "step": 11508 + }, + { + "epoch": 0.26978908443069205, + "grad_norm": 0.43696069717407227, + "learning_rate": 0.000166220817085709, + "loss": 0.0921, + "step": 11509 + }, + { + "epoch": 0.26981252600549704, + "grad_norm": 0.7654175162315369, + "learning_rate": 0.00016621529729727886, + "loss": 0.168, + "step": 11510 + }, + { + "epoch": 0.26983596758030204, + "grad_norm": 0.6276290416717529, + "learning_rate": 0.00016620977714956267, + "loss": 0.1651, + "step": 11511 + }, + { + "epoch": 0.26985940915510703, + "grad_norm": 0.2046400010585785, + "learning_rate": 0.00016620425664259035, + "loss": 0.0432, + "step": 11512 + }, + { + "epoch": 0.269882850729912, + "grad_norm": 0.6421555876731873, + "learning_rate": 0.00016619873577639188, + "loss": 0.1068, + "step": 11513 + }, + { + "epoch": 0.269906292304717, + "grad_norm": 0.4075636863708496, + "learning_rate": 0.00016619321455099722, + "loss": 0.0695, + "step": 11514 + }, + { + "epoch": 0.269929733879522, + "grad_norm": 0.38409343361854553, + "learning_rate": 0.0001661876929664363, + "loss": 0.0626, + "step": 11515 + }, + { + "epoch": 0.269953175454327, + "grad_norm": 1.0759918689727783, + "learning_rate": 0.0001661821710227391, + "loss": 0.1351, + "step": 11516 + }, + { + "epoch": 0.269976617029132, + "grad_norm": 0.08347996324300766, + "learning_rate": 0.0001661766487199356, + "loss": 0.0096, + "step": 11517 + }, + { + "epoch": 0.270000058603937, + "grad_norm": 0.8048067688941956, + "learning_rate": 0.0001661711260580557, + "loss": 0.1636, + "step": 11518 + }, + { + "epoch": 0.270023500178742, + "grad_norm": 0.5748031735420227, + "learning_rate": 0.00016616560303712947, + "loss": 0.1767, + "step": 11519 + }, + { + "epoch": 0.270046941753547, + "grad_norm": 0.16082628071308136, + "learning_rate": 0.00016616007965718677, + "loss": 0.0347, + "step": 11520 + }, + { + "epoch": 0.270070383328352, + "grad_norm": 0.13746418058872223, + "learning_rate": 0.00016615455591825762, + "loss": 0.0231, + "step": 11521 + }, + { + "epoch": 0.270093824903157, + "grad_norm": 0.7092039585113525, + "learning_rate": 0.000166149031820372, + "loss": 0.1455, + "step": 11522 + }, + { + "epoch": 0.27011726647796197, + "grad_norm": 0.17947256565093994, + "learning_rate": 0.0001661435073635599, + "loss": 0.0514, + "step": 11523 + }, + { + "epoch": 0.27014070805276696, + "grad_norm": 0.1769135743379593, + "learning_rate": 0.00016613798254785125, + "loss": 0.0356, + "step": 11524 + }, + { + "epoch": 0.27016414962757196, + "grad_norm": 0.2774936258792877, + "learning_rate": 0.000166132457373276, + "loss": 0.0617, + "step": 11525 + }, + { + "epoch": 0.27018759120237695, + "grad_norm": 0.8927958011627197, + "learning_rate": 0.00016612693183986425, + "loss": 0.1792, + "step": 11526 + }, + { + "epoch": 0.27021103277718195, + "grad_norm": 0.43136370182037354, + "learning_rate": 0.00016612140594764586, + "loss": 0.098, + "step": 11527 + }, + { + "epoch": 0.27023447435198694, + "grad_norm": 0.46208491921424866, + "learning_rate": 0.00016611587969665088, + "loss": 0.074, + "step": 11528 + }, + { + "epoch": 0.27025791592679194, + "grad_norm": 0.17619143426418304, + "learning_rate": 0.00016611035308690929, + "loss": 0.0352, + "step": 11529 + }, + { + "epoch": 0.27028135750159693, + "grad_norm": 0.18399442732334137, + "learning_rate": 0.00016610482611845103, + "loss": 0.0451, + "step": 11530 + }, + { + "epoch": 0.270304799076402, + "grad_norm": 0.7178704142570496, + "learning_rate": 0.00016609929879130613, + "loss": 0.4769, + "step": 11531 + }, + { + "epoch": 0.270328240651207, + "grad_norm": 0.6804999113082886, + "learning_rate": 0.00016609377110550461, + "loss": 0.1512, + "step": 11532 + }, + { + "epoch": 0.27035168222601197, + "grad_norm": 0.48050057888031006, + "learning_rate": 0.0001660882430610764, + "loss": 0.1385, + "step": 11533 + }, + { + "epoch": 0.27037512380081696, + "grad_norm": 0.7659132480621338, + "learning_rate": 0.00016608271465805155, + "loss": 0.1666, + "step": 11534 + }, + { + "epoch": 0.27039856537562196, + "grad_norm": 0.7890582084655762, + "learning_rate": 0.00016607718589646003, + "loss": 0.0991, + "step": 11535 + }, + { + "epoch": 0.27042200695042695, + "grad_norm": 0.19696682691574097, + "learning_rate": 0.0001660716567763318, + "loss": 0.0359, + "step": 11536 + }, + { + "epoch": 0.27044544852523195, + "grad_norm": 0.464662104845047, + "learning_rate": 0.00016606612729769697, + "loss": 0.1253, + "step": 11537 + }, + { + "epoch": 0.27046889010003694, + "grad_norm": 0.5652012228965759, + "learning_rate": 0.00016606059746058542, + "loss": 0.1332, + "step": 11538 + }, + { + "epoch": 0.27049233167484193, + "grad_norm": 0.38768061995506287, + "learning_rate": 0.00016605506726502723, + "loss": 0.1001, + "step": 11539 + }, + { + "epoch": 0.27051577324964693, + "grad_norm": 0.25172924995422363, + "learning_rate": 0.0001660495367110524, + "loss": 0.0589, + "step": 11540 + }, + { + "epoch": 0.2705392148244519, + "grad_norm": 1.6981955766677856, + "learning_rate": 0.0001660440057986909, + "loss": 0.0901, + "step": 11541 + }, + { + "epoch": 0.2705626563992569, + "grad_norm": 0.739810049533844, + "learning_rate": 0.0001660384745279728, + "loss": 0.1809, + "step": 11542 + }, + { + "epoch": 0.2705860979740619, + "grad_norm": 1.2357450723648071, + "learning_rate": 0.00016603294289892806, + "loss": 0.2, + "step": 11543 + }, + { + "epoch": 0.2706095395488669, + "grad_norm": 0.13677605986595154, + "learning_rate": 0.0001660274109115867, + "loss": 0.0121, + "step": 11544 + }, + { + "epoch": 0.2706329811236719, + "grad_norm": 0.09321747720241547, + "learning_rate": 0.00016602187856597876, + "loss": 0.0216, + "step": 11545 + }, + { + "epoch": 0.2706564226984769, + "grad_norm": 0.33562397956848145, + "learning_rate": 0.00016601634586213427, + "loss": 0.0503, + "step": 11546 + }, + { + "epoch": 0.2706798642732819, + "grad_norm": 0.1680147796869278, + "learning_rate": 0.00016601081280008322, + "loss": 0.0267, + "step": 11547 + }, + { + "epoch": 0.2707033058480869, + "grad_norm": 0.48153814673423767, + "learning_rate": 0.00016600527937985564, + "loss": 0.1283, + "step": 11548 + }, + { + "epoch": 0.2707267474228919, + "grad_norm": 0.3372439742088318, + "learning_rate": 0.00016599974560148155, + "loss": 0.0896, + "step": 11549 + }, + { + "epoch": 0.2707501889976969, + "grad_norm": 0.21078941226005554, + "learning_rate": 0.000165994211464991, + "loss": 0.0373, + "step": 11550 + }, + { + "epoch": 0.27077363057250187, + "grad_norm": 0.5650157928466797, + "learning_rate": 0.00016598867697041397, + "loss": 0.1007, + "step": 11551 + }, + { + "epoch": 0.27079707214730686, + "grad_norm": 0.6277092695236206, + "learning_rate": 0.0001659831421177806, + "loss": 0.1775, + "step": 11552 + }, + { + "epoch": 0.27082051372211186, + "grad_norm": 0.7124592065811157, + "learning_rate": 0.0001659776069071208, + "loss": 0.1484, + "step": 11553 + }, + { + "epoch": 0.27084395529691685, + "grad_norm": 0.5317879915237427, + "learning_rate": 0.0001659720713384646, + "loss": 0.0754, + "step": 11554 + }, + { + "epoch": 0.27086739687172184, + "grad_norm": 0.5815346837043762, + "learning_rate": 0.00016596653541184217, + "loss": 0.0846, + "step": 11555 + }, + { + "epoch": 0.27089083844652684, + "grad_norm": 0.5555625557899475, + "learning_rate": 0.00016596099912728342, + "loss": 0.0806, + "step": 11556 + }, + { + "epoch": 0.27091428002133183, + "grad_norm": 0.14501862227916718, + "learning_rate": 0.00016595546248481845, + "loss": 0.0225, + "step": 11557 + }, + { + "epoch": 0.2709377215961368, + "grad_norm": 0.6520816683769226, + "learning_rate": 0.00016594992548447728, + "loss": 0.1186, + "step": 11558 + }, + { + "epoch": 0.2709611631709418, + "grad_norm": 0.6792289614677429, + "learning_rate": 0.00016594438812628996, + "loss": 0.1002, + "step": 11559 + }, + { + "epoch": 0.2709846047457468, + "grad_norm": 0.08052041381597519, + "learning_rate": 0.0001659388504102865, + "loss": 0.0191, + "step": 11560 + }, + { + "epoch": 0.2710080463205518, + "grad_norm": 0.8337379097938538, + "learning_rate": 0.00016593331233649705, + "loss": 0.1662, + "step": 11561 + }, + { + "epoch": 0.2710314878953568, + "grad_norm": 0.2972855865955353, + "learning_rate": 0.00016592777390495156, + "loss": 0.0567, + "step": 11562 + }, + { + "epoch": 0.2710549294701618, + "grad_norm": 0.6899270415306091, + "learning_rate": 0.00016592223511568011, + "loss": 0.2092, + "step": 11563 + }, + { + "epoch": 0.2710783710449668, + "grad_norm": 0.14576931297779083, + "learning_rate": 0.00016591669596871276, + "loss": 0.0354, + "step": 11564 + }, + { + "epoch": 0.2711018126197718, + "grad_norm": 0.9861063361167908, + "learning_rate": 0.00016591115646407957, + "loss": 0.2704, + "step": 11565 + }, + { + "epoch": 0.2711252541945768, + "grad_norm": 0.5375507473945618, + "learning_rate": 0.00016590561660181062, + "loss": 0.1545, + "step": 11566 + }, + { + "epoch": 0.2711486957693818, + "grad_norm": 0.17820951342582703, + "learning_rate": 0.00016590007638193587, + "loss": 0.0489, + "step": 11567 + }, + { + "epoch": 0.27117213734418677, + "grad_norm": 0.4089537560939789, + "learning_rate": 0.00016589453580448553, + "loss": 0.091, + "step": 11568 + }, + { + "epoch": 0.27119557891899176, + "grad_norm": 0.48280763626098633, + "learning_rate": 0.00016588899486948952, + "loss": 0.1222, + "step": 11569 + }, + { + "epoch": 0.27121902049379676, + "grad_norm": 0.40506041049957275, + "learning_rate": 0.000165883453576978, + "loss": 0.0882, + "step": 11570 + }, + { + "epoch": 0.27124246206860175, + "grad_norm": 0.592022716999054, + "learning_rate": 0.00016587791192698102, + "loss": 0.1373, + "step": 11571 + }, + { + "epoch": 0.27126590364340675, + "grad_norm": 1.138041615486145, + "learning_rate": 0.00016587236991952864, + "loss": 0.1916, + "step": 11572 + }, + { + "epoch": 0.27128934521821174, + "grad_norm": 0.26122385263442993, + "learning_rate": 0.0001658668275546509, + "loss": 0.0632, + "step": 11573 + }, + { + "epoch": 0.27131278679301674, + "grad_norm": 0.7849442958831787, + "learning_rate": 0.00016586128483237793, + "loss": 0.1294, + "step": 11574 + }, + { + "epoch": 0.27133622836782173, + "grad_norm": 0.45704370737075806, + "learning_rate": 0.00016585574175273976, + "loss": 0.123, + "step": 11575 + }, + { + "epoch": 0.2713596699426267, + "grad_norm": 0.16178645193576813, + "learning_rate": 0.0001658501983157665, + "loss": 0.0406, + "step": 11576 + }, + { + "epoch": 0.2713831115174317, + "grad_norm": 0.19383427500724792, + "learning_rate": 0.00016584465452148818, + "loss": 0.0501, + "step": 11577 + }, + { + "epoch": 0.2714065530922367, + "grad_norm": 0.24791711568832397, + "learning_rate": 0.00016583911036993496, + "loss": 0.0388, + "step": 11578 + }, + { + "epoch": 0.2714299946670417, + "grad_norm": 0.4207589328289032, + "learning_rate": 0.00016583356586113683, + "loss": 0.1607, + "step": 11579 + }, + { + "epoch": 0.2714534362418467, + "grad_norm": 0.7041313648223877, + "learning_rate": 0.00016582802099512396, + "loss": 0.7918, + "step": 11580 + }, + { + "epoch": 0.2714768778166517, + "grad_norm": 0.691490113735199, + "learning_rate": 0.00016582247577192634, + "loss": 0.194, + "step": 11581 + }, + { + "epoch": 0.2715003193914567, + "grad_norm": 0.8242175579071045, + "learning_rate": 0.00016581693019157418, + "loss": 0.1362, + "step": 11582 + }, + { + "epoch": 0.27152376096626174, + "grad_norm": 0.5136280059814453, + "learning_rate": 0.00016581138425409746, + "loss": 0.1128, + "step": 11583 + }, + { + "epoch": 0.27154720254106673, + "grad_norm": 0.7165887951850891, + "learning_rate": 0.00016580583795952633, + "loss": 0.1771, + "step": 11584 + }, + { + "epoch": 0.27157064411587173, + "grad_norm": 0.9257740378379822, + "learning_rate": 0.00016580029130789091, + "loss": 0.119, + "step": 11585 + }, + { + "epoch": 0.2715940856906767, + "grad_norm": 0.4516768455505371, + "learning_rate": 0.00016579474429922125, + "loss": 0.1214, + "step": 11586 + }, + { + "epoch": 0.2716175272654817, + "grad_norm": 0.32859209179878235, + "learning_rate": 0.0001657891969335474, + "loss": 0.0445, + "step": 11587 + }, + { + "epoch": 0.2716409688402867, + "grad_norm": 0.9473690986633301, + "learning_rate": 0.00016578364921089955, + "loss": 0.0878, + "step": 11588 + }, + { + "epoch": 0.2716644104150917, + "grad_norm": 0.6493855714797974, + "learning_rate": 0.00016577810113130778, + "loss": 0.1583, + "step": 11589 + }, + { + "epoch": 0.2716878519898967, + "grad_norm": 0.6568968296051025, + "learning_rate": 0.00016577255269480218, + "loss": 0.2249, + "step": 11590 + }, + { + "epoch": 0.2717112935647017, + "grad_norm": 0.5396184921264648, + "learning_rate": 0.00016576700390141287, + "loss": 0.6342, + "step": 11591 + }, + { + "epoch": 0.2717347351395067, + "grad_norm": 0.25990185141563416, + "learning_rate": 0.00016576145475116994, + "loss": 0.0564, + "step": 11592 + }, + { + "epoch": 0.2717581767143117, + "grad_norm": 1.4634987115859985, + "learning_rate": 0.00016575590524410351, + "loss": 0.2215, + "step": 11593 + }, + { + "epoch": 0.2717816182891167, + "grad_norm": 0.3512308895587921, + "learning_rate": 0.00016575035538024367, + "loss": 0.0781, + "step": 11594 + }, + { + "epoch": 0.27180505986392167, + "grad_norm": 0.48173099756240845, + "learning_rate": 0.00016574480515962057, + "loss": 0.1599, + "step": 11595 + }, + { + "epoch": 0.27182850143872667, + "grad_norm": 0.6680041551589966, + "learning_rate": 0.0001657392545822643, + "loss": 0.1238, + "step": 11596 + }, + { + "epoch": 0.27185194301353166, + "grad_norm": 0.4649474024772644, + "learning_rate": 0.000165733703648205, + "loss": 0.0886, + "step": 11597 + }, + { + "epoch": 0.27187538458833665, + "grad_norm": 0.3466035723686218, + "learning_rate": 0.00016572815235747274, + "loss": 0.0583, + "step": 11598 + }, + { + "epoch": 0.27189882616314165, + "grad_norm": 0.6035022139549255, + "learning_rate": 0.00016572260071009772, + "loss": 0.1234, + "step": 11599 + }, + { + "epoch": 0.27192226773794664, + "grad_norm": 0.21955756843090057, + "learning_rate": 0.00016571704870610999, + "loss": 0.0605, + "step": 11600 + }, + { + "epoch": 0.27194570931275164, + "grad_norm": 0.3509382903575897, + "learning_rate": 0.00016571149634553973, + "loss": 0.1278, + "step": 11601 + }, + { + "epoch": 0.27196915088755663, + "grad_norm": 0.45309978723526, + "learning_rate": 0.00016570594362841704, + "loss": 0.1101, + "step": 11602 + }, + { + "epoch": 0.2719925924623616, + "grad_norm": 0.36871927976608276, + "learning_rate": 0.00016570039055477205, + "loss": 0.0548, + "step": 11603 + }, + { + "epoch": 0.2720160340371666, + "grad_norm": 0.5693396925926208, + "learning_rate": 0.00016569483712463486, + "loss": 0.1402, + "step": 11604 + }, + { + "epoch": 0.2720394756119716, + "grad_norm": 0.5559210181236267, + "learning_rate": 0.00016568928333803568, + "loss": 0.2262, + "step": 11605 + }, + { + "epoch": 0.2720629171867766, + "grad_norm": 0.33830809593200684, + "learning_rate": 0.00016568372919500458, + "loss": 0.1114, + "step": 11606 + }, + { + "epoch": 0.2720863587615816, + "grad_norm": 0.4750021994113922, + "learning_rate": 0.00016567817469557175, + "loss": 0.1167, + "step": 11607 + }, + { + "epoch": 0.2721098003363866, + "grad_norm": 0.772788941860199, + "learning_rate": 0.00016567261983976725, + "loss": 0.1441, + "step": 11608 + }, + { + "epoch": 0.2721332419111916, + "grad_norm": 0.20387309789657593, + "learning_rate": 0.00016566706462762133, + "loss": 0.0411, + "step": 11609 + }, + { + "epoch": 0.2721566834859966, + "grad_norm": 0.96195387840271, + "learning_rate": 0.000165661509059164, + "loss": 0.9811, + "step": 11610 + }, + { + "epoch": 0.2721801250608016, + "grad_norm": 0.6513558626174927, + "learning_rate": 0.0001656559531344255, + "loss": 0.1644, + "step": 11611 + }, + { + "epoch": 0.2722035666356066, + "grad_norm": 0.20836114883422852, + "learning_rate": 0.00016565039685343598, + "loss": 0.0316, + "step": 11612 + }, + { + "epoch": 0.27222700821041157, + "grad_norm": 0.5677258968353271, + "learning_rate": 0.00016564484021622554, + "loss": 0.181, + "step": 11613 + }, + { + "epoch": 0.27225044978521656, + "grad_norm": 0.5405450463294983, + "learning_rate": 0.00016563928322282435, + "loss": 0.1115, + "step": 11614 + }, + { + "epoch": 0.27227389136002156, + "grad_norm": 0.1917732208967209, + "learning_rate": 0.00016563372587326258, + "loss": 0.0475, + "step": 11615 + }, + { + "epoch": 0.27229733293482655, + "grad_norm": 0.4050724506378174, + "learning_rate": 0.00016562816816757033, + "loss": 0.0545, + "step": 11616 + }, + { + "epoch": 0.27232077450963155, + "grad_norm": 0.7250462770462036, + "learning_rate": 0.0001656226101057778, + "loss": 0.7482, + "step": 11617 + }, + { + "epoch": 0.27234421608443654, + "grad_norm": 0.12189839035272598, + "learning_rate": 0.00016561705168791516, + "loss": 0.0195, + "step": 11618 + }, + { + "epoch": 0.27236765765924154, + "grad_norm": 0.2581478953361511, + "learning_rate": 0.00016561149291401255, + "loss": 0.0614, + "step": 11619 + }, + { + "epoch": 0.27239109923404653, + "grad_norm": 0.22902311384677887, + "learning_rate": 0.00016560593378410016, + "loss": 0.0653, + "step": 11620 + }, + { + "epoch": 0.2724145408088515, + "grad_norm": 0.6156408786773682, + "learning_rate": 0.00016560037429820808, + "loss": 0.1239, + "step": 11621 + }, + { + "epoch": 0.2724379823836565, + "grad_norm": 0.5141701698303223, + "learning_rate": 0.00016559481445636658, + "loss": 0.0881, + "step": 11622 + }, + { + "epoch": 0.2724614239584615, + "grad_norm": 0.3479766249656677, + "learning_rate": 0.00016558925425860576, + "loss": 0.0866, + "step": 11623 + }, + { + "epoch": 0.2724848655332665, + "grad_norm": 0.6459321975708008, + "learning_rate": 0.00016558369370495575, + "loss": 0.1611, + "step": 11624 + }, + { + "epoch": 0.2725083071080715, + "grad_norm": 0.6061774492263794, + "learning_rate": 0.00016557813279544683, + "loss": 0.1638, + "step": 11625 + }, + { + "epoch": 0.2725317486828765, + "grad_norm": 1.095157504081726, + "learning_rate": 0.00016557257153010908, + "loss": 0.2065, + "step": 11626 + }, + { + "epoch": 0.2725551902576815, + "grad_norm": 0.5448046326637268, + "learning_rate": 0.00016556700990897274, + "loss": 0.0857, + "step": 11627 + }, + { + "epoch": 0.2725786318324865, + "grad_norm": 0.48567941784858704, + "learning_rate": 0.00016556144793206795, + "loss": 0.1152, + "step": 11628 + }, + { + "epoch": 0.2726020734072915, + "grad_norm": 0.3998231291770935, + "learning_rate": 0.00016555588559942489, + "loss": 0.0669, + "step": 11629 + }, + { + "epoch": 0.2726255149820965, + "grad_norm": 0.5448756814002991, + "learning_rate": 0.00016555032291107376, + "loss": 0.615, + "step": 11630 + }, + { + "epoch": 0.27264895655690147, + "grad_norm": 0.7829899191856384, + "learning_rate": 0.00016554475986704477, + "loss": 0.7432, + "step": 11631 + }, + { + "epoch": 0.27267239813170646, + "grad_norm": 0.2642066776752472, + "learning_rate": 0.000165539196467368, + "loss": 0.0344, + "step": 11632 + }, + { + "epoch": 0.27269583970651146, + "grad_norm": 0.5734572410583496, + "learning_rate": 0.00016553363271207375, + "loss": 0.1053, + "step": 11633 + }, + { + "epoch": 0.2727192812813165, + "grad_norm": 0.7619876265525818, + "learning_rate": 0.00016552806860119218, + "loss": 0.1699, + "step": 11634 + }, + { + "epoch": 0.2727427228561215, + "grad_norm": 0.5064459443092346, + "learning_rate": 0.00016552250413475345, + "loss": 0.0859, + "step": 11635 + }, + { + "epoch": 0.2727661644309265, + "grad_norm": 0.37753599882125854, + "learning_rate": 0.00016551693931278775, + "loss": 0.0609, + "step": 11636 + }, + { + "epoch": 0.2727896060057315, + "grad_norm": 0.20955578982830048, + "learning_rate": 0.0001655113741353253, + "loss": 0.055, + "step": 11637 + }, + { + "epoch": 0.2728130475805365, + "grad_norm": 0.4714408218860626, + "learning_rate": 0.0001655058086023963, + "loss": 0.0774, + "step": 11638 + }, + { + "epoch": 0.2728364891553415, + "grad_norm": 0.13976573944091797, + "learning_rate": 0.000165500242714031, + "loss": 0.0239, + "step": 11639 + }, + { + "epoch": 0.27285993073014647, + "grad_norm": 0.36634185910224915, + "learning_rate": 0.00016549467647025945, + "loss": 0.0419, + "step": 11640 + }, + { + "epoch": 0.27288337230495147, + "grad_norm": 0.6831300258636475, + "learning_rate": 0.000165489109871112, + "loss": 0.1141, + "step": 11641 + }, + { + "epoch": 0.27290681387975646, + "grad_norm": 0.41629621386528015, + "learning_rate": 0.00016548354291661877, + "loss": 0.0963, + "step": 11642 + }, + { + "epoch": 0.27293025545456145, + "grad_norm": 0.6768691539764404, + "learning_rate": 0.00016547797560680997, + "loss": 0.1816, + "step": 11643 + }, + { + "epoch": 0.27295369702936645, + "grad_norm": 0.7482909560203552, + "learning_rate": 0.0001654724079417159, + "loss": 0.2051, + "step": 11644 + }, + { + "epoch": 0.27297713860417144, + "grad_norm": 0.7179736495018005, + "learning_rate": 0.00016546683992136666, + "loss": 0.0758, + "step": 11645 + }, + { + "epoch": 0.27300058017897644, + "grad_norm": 1.0236852169036865, + "learning_rate": 0.00016546127154579252, + "loss": 0.1232, + "step": 11646 + }, + { + "epoch": 0.27302402175378143, + "grad_norm": 0.2426048368215561, + "learning_rate": 0.0001654557028150237, + "loss": 0.0529, + "step": 11647 + }, + { + "epoch": 0.2730474633285864, + "grad_norm": 0.13874243199825287, + "learning_rate": 0.00016545013372909034, + "loss": 0.0186, + "step": 11648 + }, + { + "epoch": 0.2730709049033914, + "grad_norm": 0.5833110809326172, + "learning_rate": 0.00016544456428802277, + "loss": 0.1448, + "step": 11649 + }, + { + "epoch": 0.2730943464781964, + "grad_norm": 0.551142692565918, + "learning_rate": 0.00016543899449185113, + "loss": 0.1977, + "step": 11650 + }, + { + "epoch": 0.2731177880530014, + "grad_norm": 0.17293104529380798, + "learning_rate": 0.00016543342434060566, + "loss": 0.0214, + "step": 11651 + }, + { + "epoch": 0.2731412296278064, + "grad_norm": 0.2318708449602127, + "learning_rate": 0.0001654278538343166, + "loss": 0.064, + "step": 11652 + }, + { + "epoch": 0.2731646712026114, + "grad_norm": 0.8359501957893372, + "learning_rate": 0.00016542228297301416, + "loss": 0.1968, + "step": 11653 + }, + { + "epoch": 0.2731881127774164, + "grad_norm": 0.348213255405426, + "learning_rate": 0.00016541671175672855, + "loss": 0.0888, + "step": 11654 + }, + { + "epoch": 0.2732115543522214, + "grad_norm": 0.5820369720458984, + "learning_rate": 0.00016541114018549005, + "loss": 0.1231, + "step": 11655 + }, + { + "epoch": 0.2732349959270264, + "grad_norm": 1.4772197008132935, + "learning_rate": 0.00016540556825932887, + "loss": 0.2272, + "step": 11656 + }, + { + "epoch": 0.2732584375018314, + "grad_norm": 0.31984642148017883, + "learning_rate": 0.00016539999597827522, + "loss": 0.0493, + "step": 11657 + }, + { + "epoch": 0.27328187907663637, + "grad_norm": 0.18829010426998138, + "learning_rate": 0.0001653944233423594, + "loss": 0.0299, + "step": 11658 + }, + { + "epoch": 0.27330532065144136, + "grad_norm": 1.1130297183990479, + "learning_rate": 0.00016538885035161154, + "loss": 0.2007, + "step": 11659 + }, + { + "epoch": 0.27332876222624636, + "grad_norm": 0.5919719338417053, + "learning_rate": 0.00016538327700606198, + "loss": 0.1511, + "step": 11660 + }, + { + "epoch": 0.27335220380105135, + "grad_norm": 0.40413641929626465, + "learning_rate": 0.00016537770330574087, + "loss": 0.0744, + "step": 11661 + }, + { + "epoch": 0.27337564537585635, + "grad_norm": 0.19748133420944214, + "learning_rate": 0.00016537212925067855, + "loss": 0.0574, + "step": 11662 + }, + { + "epoch": 0.27339908695066134, + "grad_norm": 0.6578659415245056, + "learning_rate": 0.00016536655484090522, + "loss": 0.0705, + "step": 11663 + }, + { + "epoch": 0.27342252852546634, + "grad_norm": 0.7751975059509277, + "learning_rate": 0.00016536098007645112, + "loss": 0.2292, + "step": 11664 + }, + { + "epoch": 0.27344597010027133, + "grad_norm": 0.48563188314437866, + "learning_rate": 0.00016535540495734647, + "loss": 0.1203, + "step": 11665 + }, + { + "epoch": 0.2734694116750763, + "grad_norm": 0.5370609164237976, + "learning_rate": 0.0001653498294836216, + "loss": 0.1473, + "step": 11666 + }, + { + "epoch": 0.2734928532498813, + "grad_norm": 0.4402795732021332, + "learning_rate": 0.00016534425365530666, + "loss": 0.1344, + "step": 11667 + }, + { + "epoch": 0.2735162948246863, + "grad_norm": 0.5680058598518372, + "learning_rate": 0.000165338677472432, + "loss": 0.1446, + "step": 11668 + }, + { + "epoch": 0.2735397363994913, + "grad_norm": 0.5722119212150574, + "learning_rate": 0.00016533310093502784, + "loss": 0.158, + "step": 11669 + }, + { + "epoch": 0.2735631779742963, + "grad_norm": 0.2006864696741104, + "learning_rate": 0.00016532752404312446, + "loss": 0.0343, + "step": 11670 + }, + { + "epoch": 0.2735866195491013, + "grad_norm": 0.6514109373092651, + "learning_rate": 0.00016532194679675207, + "loss": 0.1833, + "step": 11671 + }, + { + "epoch": 0.2736100611239063, + "grad_norm": 0.6169114112854004, + "learning_rate": 0.00016531636919594095, + "loss": 0.1775, + "step": 11672 + }, + { + "epoch": 0.2736335026987113, + "grad_norm": 0.5801546573638916, + "learning_rate": 0.00016531079124072138, + "loss": 0.1074, + "step": 11673 + }, + { + "epoch": 0.2736569442735163, + "grad_norm": 0.23890115320682526, + "learning_rate": 0.00016530521293112364, + "loss": 0.0268, + "step": 11674 + }, + { + "epoch": 0.2736803858483213, + "grad_norm": 0.4100390374660492, + "learning_rate": 0.000165299634267178, + "loss": 0.058, + "step": 11675 + }, + { + "epoch": 0.27370382742312627, + "grad_norm": 0.38646814227104187, + "learning_rate": 0.00016529405524891465, + "loss": 0.0718, + "step": 11676 + }, + { + "epoch": 0.27372726899793126, + "grad_norm": 0.28738245368003845, + "learning_rate": 0.00016528847587636393, + "loss": 0.2542, + "step": 11677 + }, + { + "epoch": 0.27375071057273626, + "grad_norm": 0.2013034075498581, + "learning_rate": 0.00016528289614955614, + "loss": 0.039, + "step": 11678 + }, + { + "epoch": 0.27377415214754125, + "grad_norm": 0.24575889110565186, + "learning_rate": 0.00016527731606852148, + "loss": 0.035, + "step": 11679 + }, + { + "epoch": 0.27379759372234624, + "grad_norm": 0.631445050239563, + "learning_rate": 0.00016527173563329027, + "loss": 0.7585, + "step": 11680 + }, + { + "epoch": 0.27382103529715124, + "grad_norm": 0.4637528657913208, + "learning_rate": 0.00016526615484389282, + "loss": 0.1043, + "step": 11681 + }, + { + "epoch": 0.27384447687195623, + "grad_norm": 0.30218565464019775, + "learning_rate": 0.00016526057370035935, + "loss": 0.0632, + "step": 11682 + }, + { + "epoch": 0.2738679184467612, + "grad_norm": 0.6565988659858704, + "learning_rate": 0.00016525499220272023, + "loss": 0.2359, + "step": 11683 + }, + { + "epoch": 0.2738913600215662, + "grad_norm": 0.5322155952453613, + "learning_rate": 0.00016524941035100563, + "loss": 0.1642, + "step": 11684 + }, + { + "epoch": 0.27391480159637127, + "grad_norm": 0.7942844033241272, + "learning_rate": 0.00016524382814524588, + "loss": 0.1512, + "step": 11685 + }, + { + "epoch": 0.27393824317117627, + "grad_norm": 0.5653080344200134, + "learning_rate": 0.00016523824558547132, + "loss": 0.1351, + "step": 11686 + }, + { + "epoch": 0.27396168474598126, + "grad_norm": 0.48454004526138306, + "learning_rate": 0.0001652326626717122, + "loss": 0.1108, + "step": 11687 + }, + { + "epoch": 0.27398512632078625, + "grad_norm": 0.17289449274539948, + "learning_rate": 0.0001652270794039988, + "loss": 0.0543, + "step": 11688 + }, + { + "epoch": 0.27400856789559125, + "grad_norm": 0.4745146930217743, + "learning_rate": 0.00016522149578236146, + "loss": 0.087, + "step": 11689 + }, + { + "epoch": 0.27403200947039624, + "grad_norm": 0.5693787336349487, + "learning_rate": 0.00016521591180683045, + "loss": 0.1576, + "step": 11690 + }, + { + "epoch": 0.27405545104520124, + "grad_norm": 0.3179265260696411, + "learning_rate": 0.00016521032747743606, + "loss": 0.0496, + "step": 11691 + }, + { + "epoch": 0.27407889262000623, + "grad_norm": 0.43515098094940186, + "learning_rate": 0.0001652047427942086, + "loss": 0.0704, + "step": 11692 + }, + { + "epoch": 0.2741023341948112, + "grad_norm": 0.7686652541160583, + "learning_rate": 0.00016519915775717835, + "loss": 0.6244, + "step": 11693 + }, + { + "epoch": 0.2741257757696162, + "grad_norm": 0.5262572765350342, + "learning_rate": 0.00016519357236637565, + "loss": 0.1471, + "step": 11694 + }, + { + "epoch": 0.2741492173444212, + "grad_norm": 0.2771710753440857, + "learning_rate": 0.0001651879866218308, + "loss": 0.073, + "step": 11695 + }, + { + "epoch": 0.2741726589192262, + "grad_norm": 0.16482263803482056, + "learning_rate": 0.00016518240052357412, + "loss": 0.0301, + "step": 11696 + }, + { + "epoch": 0.2741961004940312, + "grad_norm": 1.0246111154556274, + "learning_rate": 0.00016517681407163585, + "loss": 0.1984, + "step": 11697 + }, + { + "epoch": 0.2742195420688362, + "grad_norm": 0.6111933588981628, + "learning_rate": 0.0001651712272660464, + "loss": 0.1926, + "step": 11698 + }, + { + "epoch": 0.2742429836436412, + "grad_norm": 0.6929341554641724, + "learning_rate": 0.00016516564010683605, + "loss": 0.1289, + "step": 11699 + }, + { + "epoch": 0.2742664252184462, + "grad_norm": 0.6240811347961426, + "learning_rate": 0.00016516005259403505, + "loss": 0.1305, + "step": 11700 + }, + { + "epoch": 0.2742898667932512, + "grad_norm": 0.3729364573955536, + "learning_rate": 0.0001651544647276738, + "loss": 0.0989, + "step": 11701 + }, + { + "epoch": 0.2743133083680562, + "grad_norm": 0.5115390419960022, + "learning_rate": 0.0001651488765077826, + "loss": 0.1446, + "step": 11702 + }, + { + "epoch": 0.27433674994286117, + "grad_norm": 0.5852293968200684, + "learning_rate": 0.00016514328793439174, + "loss": 0.1135, + "step": 11703 + }, + { + "epoch": 0.27436019151766616, + "grad_norm": 0.9433050155639648, + "learning_rate": 0.0001651376990075316, + "loss": 0.1433, + "step": 11704 + }, + { + "epoch": 0.27438363309247116, + "grad_norm": 0.2731723487377167, + "learning_rate": 0.00016513210972723247, + "loss": 0.0597, + "step": 11705 + }, + { + "epoch": 0.27440707466727615, + "grad_norm": 0.28900468349456787, + "learning_rate": 0.00016512652009352468, + "loss": 0.0528, + "step": 11706 + }, + { + "epoch": 0.27443051624208115, + "grad_norm": 0.5739296674728394, + "learning_rate": 0.00016512093010643856, + "loss": 0.7839, + "step": 11707 + }, + { + "epoch": 0.27445395781688614, + "grad_norm": 0.5668864250183105, + "learning_rate": 0.00016511533976600441, + "loss": 0.128, + "step": 11708 + }, + { + "epoch": 0.27447739939169113, + "grad_norm": 0.3857114911079407, + "learning_rate": 0.00016510974907225264, + "loss": 0.1188, + "step": 11709 + }, + { + "epoch": 0.27450084096649613, + "grad_norm": 0.4177510738372803, + "learning_rate": 0.0001651041580252135, + "loss": 0.0983, + "step": 11710 + }, + { + "epoch": 0.2745242825413011, + "grad_norm": 0.20004050433635712, + "learning_rate": 0.0001650985666249174, + "loss": 0.0305, + "step": 11711 + }, + { + "epoch": 0.2745477241161061, + "grad_norm": 0.336902916431427, + "learning_rate": 0.0001650929748713946, + "loss": 0.0562, + "step": 11712 + }, + { + "epoch": 0.2745711656909111, + "grad_norm": 0.6657809019088745, + "learning_rate": 0.00016508738276467555, + "loss": 0.5318, + "step": 11713 + }, + { + "epoch": 0.2745946072657161, + "grad_norm": 0.735014796257019, + "learning_rate": 0.0001650817903047905, + "loss": 0.1882, + "step": 11714 + }, + { + "epoch": 0.2746180488405211, + "grad_norm": 0.5020318627357483, + "learning_rate": 0.00016507619749176985, + "loss": 0.1483, + "step": 11715 + }, + { + "epoch": 0.2746414904153261, + "grad_norm": 0.52122962474823, + "learning_rate": 0.00016507060432564388, + "loss": 0.0846, + "step": 11716 + }, + { + "epoch": 0.2746649319901311, + "grad_norm": 0.6174842715263367, + "learning_rate": 0.000165065010806443, + "loss": 0.1179, + "step": 11717 + }, + { + "epoch": 0.2746883735649361, + "grad_norm": 0.9930493831634521, + "learning_rate": 0.00016505941693419754, + "loss": 0.2527, + "step": 11718 + }, + { + "epoch": 0.2747118151397411, + "grad_norm": 0.16524182260036469, + "learning_rate": 0.0001650538227089379, + "loss": 0.03, + "step": 11719 + }, + { + "epoch": 0.27473525671454607, + "grad_norm": 0.2988797128200531, + "learning_rate": 0.00016504822813069434, + "loss": 0.0668, + "step": 11720 + }, + { + "epoch": 0.27475869828935107, + "grad_norm": 0.21054160594940186, + "learning_rate": 0.00016504263319949724, + "loss": 0.0177, + "step": 11721 + }, + { + "epoch": 0.27478213986415606, + "grad_norm": 0.6322923302650452, + "learning_rate": 0.00016503703791537704, + "loss": 0.1491, + "step": 11722 + }, + { + "epoch": 0.27480558143896106, + "grad_norm": 0.36968618631362915, + "learning_rate": 0.00016503144227836402, + "loss": 0.0868, + "step": 11723 + }, + { + "epoch": 0.27482902301376605, + "grad_norm": 0.26279741525650024, + "learning_rate": 0.00016502584628848857, + "loss": 0.0519, + "step": 11724 + }, + { + "epoch": 0.27485246458857104, + "grad_norm": 0.41608211398124695, + "learning_rate": 0.00016502024994578105, + "loss": 0.0831, + "step": 11725 + }, + { + "epoch": 0.27487590616337604, + "grad_norm": 0.575066328048706, + "learning_rate": 0.00016501465325027183, + "loss": 0.085, + "step": 11726 + }, + { + "epoch": 0.27489934773818103, + "grad_norm": 0.36586901545524597, + "learning_rate": 0.00016500905620199125, + "loss": 0.4463, + "step": 11727 + }, + { + "epoch": 0.274922789312986, + "grad_norm": 0.85024094581604, + "learning_rate": 0.00016500345880096973, + "loss": 0.1657, + "step": 11728 + }, + { + "epoch": 0.274946230887791, + "grad_norm": 0.17645740509033203, + "learning_rate": 0.00016499786104723759, + "loss": 0.0383, + "step": 11729 + }, + { + "epoch": 0.274969672462596, + "grad_norm": 0.5216371417045593, + "learning_rate": 0.00016499226294082524, + "loss": 0.119, + "step": 11730 + }, + { + "epoch": 0.274993114037401, + "grad_norm": 0.2393512725830078, + "learning_rate": 0.00016498666448176307, + "loss": 0.0485, + "step": 11731 + }, + { + "epoch": 0.275016555612206, + "grad_norm": 0.6563356518745422, + "learning_rate": 0.00016498106567008138, + "loss": 0.1172, + "step": 11732 + }, + { + "epoch": 0.275039997187011, + "grad_norm": 0.5687165856361389, + "learning_rate": 0.00016497546650581064, + "loss": 0.1022, + "step": 11733 + }, + { + "epoch": 0.275063438761816, + "grad_norm": 0.21765515208244324, + "learning_rate": 0.00016496986698898115, + "loss": 0.0353, + "step": 11734 + }, + { + "epoch": 0.275086880336621, + "grad_norm": 0.5263541340827942, + "learning_rate": 0.00016496426711962338, + "loss": 0.108, + "step": 11735 + }, + { + "epoch": 0.27511032191142604, + "grad_norm": 0.6470202207565308, + "learning_rate": 0.00016495866689776766, + "loss": 0.79, + "step": 11736 + }, + { + "epoch": 0.27513376348623103, + "grad_norm": 0.23891733586788177, + "learning_rate": 0.0001649530663234444, + "loss": 0.0576, + "step": 11737 + }, + { + "epoch": 0.275157205061036, + "grad_norm": 0.2557658553123474, + "learning_rate": 0.00016494746539668395, + "loss": 0.0287, + "step": 11738 + }, + { + "epoch": 0.275180646635841, + "grad_norm": 0.5908753871917725, + "learning_rate": 0.00016494186411751676, + "loss": 0.1485, + "step": 11739 + }, + { + "epoch": 0.275204088210646, + "grad_norm": 0.2679985463619232, + "learning_rate": 0.0001649362624859731, + "loss": 0.0686, + "step": 11740 + }, + { + "epoch": 0.275227529785451, + "grad_norm": 0.49530288577079773, + "learning_rate": 0.00016493066050208354, + "loss": 0.1045, + "step": 11741 + }, + { + "epoch": 0.275250971360256, + "grad_norm": 0.17567898333072662, + "learning_rate": 0.00016492505816587837, + "loss": 0.0517, + "step": 11742 + }, + { + "epoch": 0.275274412935061, + "grad_norm": 0.5368921160697937, + "learning_rate": 0.00016491945547738804, + "loss": 0.1523, + "step": 11743 + }, + { + "epoch": 0.275297854509866, + "grad_norm": 0.5606570839881897, + "learning_rate": 0.00016491385243664285, + "loss": 0.1345, + "step": 11744 + }, + { + "epoch": 0.275321296084671, + "grad_norm": 0.15372881293296814, + "learning_rate": 0.0001649082490436733, + "loss": 0.0311, + "step": 11745 + }, + { + "epoch": 0.275344737659476, + "grad_norm": 0.6515322923660278, + "learning_rate": 0.0001649026452985098, + "loss": 0.2024, + "step": 11746 + }, + { + "epoch": 0.275368179234281, + "grad_norm": 0.805270791053772, + "learning_rate": 0.0001648970412011827, + "loss": 0.0719, + "step": 11747 + }, + { + "epoch": 0.27539162080908597, + "grad_norm": 0.7565256357192993, + "learning_rate": 0.0001648914367517224, + "loss": 0.1513, + "step": 11748 + }, + { + "epoch": 0.27541506238389096, + "grad_norm": 0.41615030169487, + "learning_rate": 0.00016488583195015938, + "loss": 0.071, + "step": 11749 + }, + { + "epoch": 0.27543850395869596, + "grad_norm": 0.11364638805389404, + "learning_rate": 0.000164880226796524, + "loss": 0.0172, + "step": 11750 + }, + { + "epoch": 0.27546194553350095, + "grad_norm": 0.3985515236854553, + "learning_rate": 0.00016487462129084667, + "loss": 0.1315, + "step": 11751 + }, + { + "epoch": 0.27548538710830595, + "grad_norm": 0.6651350259780884, + "learning_rate": 0.0001648690154331578, + "loss": 0.137, + "step": 11752 + }, + { + "epoch": 0.27550882868311094, + "grad_norm": 0.590934693813324, + "learning_rate": 0.0001648634092234879, + "loss": 0.1542, + "step": 11753 + }, + { + "epoch": 0.27553227025791593, + "grad_norm": 0.27846962213516235, + "learning_rate": 0.00016485780266186728, + "loss": 0.0796, + "step": 11754 + }, + { + "epoch": 0.27555571183272093, + "grad_norm": 0.4595099985599518, + "learning_rate": 0.0001648521957483264, + "loss": 0.1118, + "step": 11755 + }, + { + "epoch": 0.2755791534075259, + "grad_norm": 0.3585030734539032, + "learning_rate": 0.00016484658848289567, + "loss": 0.0897, + "step": 11756 + }, + { + "epoch": 0.2756025949823309, + "grad_norm": 0.16725435853004456, + "learning_rate": 0.00016484098086560554, + "loss": 0.0221, + "step": 11757 + }, + { + "epoch": 0.2756260365571359, + "grad_norm": 0.2853669226169586, + "learning_rate": 0.0001648353728964864, + "loss": 0.0991, + "step": 11758 + }, + { + "epoch": 0.2756494781319409, + "grad_norm": 0.46039536595344543, + "learning_rate": 0.0001648297645755687, + "loss": 0.0359, + "step": 11759 + }, + { + "epoch": 0.2756729197067459, + "grad_norm": 0.5659158229827881, + "learning_rate": 0.00016482415590288294, + "loss": 0.131, + "step": 11760 + }, + { + "epoch": 0.2756963612815509, + "grad_norm": 0.345803827047348, + "learning_rate": 0.00016481854687845946, + "loss": 0.0464, + "step": 11761 + }, + { + "epoch": 0.2757198028563559, + "grad_norm": 1.1404218673706055, + "learning_rate": 0.0001648129375023287, + "loss": 0.1642, + "step": 11762 + }, + { + "epoch": 0.2757432444311609, + "grad_norm": 0.6261040568351746, + "learning_rate": 0.0001648073277745211, + "loss": 0.0992, + "step": 11763 + }, + { + "epoch": 0.2757666860059659, + "grad_norm": 0.5129100680351257, + "learning_rate": 0.00016480171769506715, + "loss": 0.083, + "step": 11764 + }, + { + "epoch": 0.27579012758077087, + "grad_norm": 0.6657751798629761, + "learning_rate": 0.00016479610726399726, + "loss": 0.1597, + "step": 11765 + }, + { + "epoch": 0.27581356915557587, + "grad_norm": 0.466615229845047, + "learning_rate": 0.00016479049648134185, + "loss": 0.1374, + "step": 11766 + }, + { + "epoch": 0.27583701073038086, + "grad_norm": 0.5809943675994873, + "learning_rate": 0.00016478488534713142, + "loss": 0.1659, + "step": 11767 + }, + { + "epoch": 0.27586045230518585, + "grad_norm": 0.4783628582954407, + "learning_rate": 0.00016477927386139633, + "loss": 0.0882, + "step": 11768 + }, + { + "epoch": 0.27588389387999085, + "grad_norm": 1.32477867603302, + "learning_rate": 0.00016477366202416712, + "loss": 0.1358, + "step": 11769 + }, + { + "epoch": 0.27590733545479584, + "grad_norm": 0.904763400554657, + "learning_rate": 0.00016476804983547419, + "loss": 0.1026, + "step": 11770 + }, + { + "epoch": 0.27593077702960084, + "grad_norm": 0.34647369384765625, + "learning_rate": 0.00016476243729534798, + "loss": 0.113, + "step": 11771 + }, + { + "epoch": 0.27595421860440583, + "grad_norm": 0.49000784754753113, + "learning_rate": 0.00016475682440381898, + "loss": 0.0711, + "step": 11772 + }, + { + "epoch": 0.2759776601792108, + "grad_norm": 0.57559734582901, + "learning_rate": 0.00016475121116091764, + "loss": 0.1492, + "step": 11773 + }, + { + "epoch": 0.2760011017540158, + "grad_norm": 0.570023238658905, + "learning_rate": 0.00016474559756667442, + "loss": 0.0749, + "step": 11774 + }, + { + "epoch": 0.2760245433288208, + "grad_norm": 0.14655819535255432, + "learning_rate": 0.00016473998362111973, + "loss": 0.033, + "step": 11775 + }, + { + "epoch": 0.2760479849036258, + "grad_norm": 0.8376701474189758, + "learning_rate": 0.00016473436932428409, + "loss": 0.2278, + "step": 11776 + }, + { + "epoch": 0.2760714264784308, + "grad_norm": 0.5854076743125916, + "learning_rate": 0.00016472875467619792, + "loss": 0.1337, + "step": 11777 + }, + { + "epoch": 0.2760948680532358, + "grad_norm": 0.5941572189331055, + "learning_rate": 0.0001647231396768917, + "loss": 0.1401, + "step": 11778 + }, + { + "epoch": 0.2761183096280408, + "grad_norm": 0.26142269372940063, + "learning_rate": 0.00016471752432639596, + "loss": 0.0611, + "step": 11779 + }, + { + "epoch": 0.2761417512028458, + "grad_norm": 0.4540145993232727, + "learning_rate": 0.00016471190862474107, + "loss": 0.0495, + "step": 11780 + }, + { + "epoch": 0.2761651927776508, + "grad_norm": 0.9118262529373169, + "learning_rate": 0.00016470629257195755, + "loss": 0.0982, + "step": 11781 + }, + { + "epoch": 0.2761886343524558, + "grad_norm": 0.5644153356552124, + "learning_rate": 0.00016470067616807587, + "loss": 0.151, + "step": 11782 + }, + { + "epoch": 0.27621207592726077, + "grad_norm": 0.7180019021034241, + "learning_rate": 0.0001646950594131265, + "loss": 0.1362, + "step": 11783 + }, + { + "epoch": 0.27623551750206576, + "grad_norm": 0.44770383834838867, + "learning_rate": 0.00016468944230713992, + "loss": 0.0712, + "step": 11784 + }, + { + "epoch": 0.27625895907687076, + "grad_norm": 0.4484364688396454, + "learning_rate": 0.0001646838248501466, + "loss": 0.1183, + "step": 11785 + }, + { + "epoch": 0.27628240065167575, + "grad_norm": 0.441854864358902, + "learning_rate": 0.00016467820704217705, + "loss": 0.0898, + "step": 11786 + }, + { + "epoch": 0.2763058422264808, + "grad_norm": 0.7696609497070312, + "learning_rate": 0.00016467258888326172, + "loss": 0.2532, + "step": 11787 + }, + { + "epoch": 0.2763292838012858, + "grad_norm": 0.4635888338088989, + "learning_rate": 0.00016466697037343108, + "loss": 0.1029, + "step": 11788 + }, + { + "epoch": 0.2763527253760908, + "grad_norm": 0.8830252885818481, + "learning_rate": 0.00016466135151271566, + "loss": 0.1493, + "step": 11789 + }, + { + "epoch": 0.2763761669508958, + "grad_norm": 0.15953481197357178, + "learning_rate": 0.00016465573230114595, + "loss": 0.0171, + "step": 11790 + }, + { + "epoch": 0.2763996085257008, + "grad_norm": 0.3972851037979126, + "learning_rate": 0.0001646501127387524, + "loss": 0.4107, + "step": 11791 + }, + { + "epoch": 0.2764230501005058, + "grad_norm": 0.3760698735713959, + "learning_rate": 0.0001646444928255655, + "loss": 0.0703, + "step": 11792 + }, + { + "epoch": 0.27644649167531077, + "grad_norm": 0.29662346839904785, + "learning_rate": 0.00016463887256161578, + "loss": 0.067, + "step": 11793 + }, + { + "epoch": 0.27646993325011576, + "grad_norm": 0.5610926151275635, + "learning_rate": 0.00016463325194693371, + "loss": 0.1107, + "step": 11794 + }, + { + "epoch": 0.27649337482492076, + "grad_norm": 0.8028746247291565, + "learning_rate": 0.00016462763098154984, + "loss": 0.6557, + "step": 11795 + }, + { + "epoch": 0.27651681639972575, + "grad_norm": 0.6933854818344116, + "learning_rate": 0.0001646220096654946, + "loss": 0.1846, + "step": 11796 + }, + { + "epoch": 0.27654025797453075, + "grad_norm": 0.12520408630371094, + "learning_rate": 0.00016461638799879848, + "loss": 0.0361, + "step": 11797 + }, + { + "epoch": 0.27656369954933574, + "grad_norm": 0.42248305678367615, + "learning_rate": 0.00016461076598149209, + "loss": 0.0984, + "step": 11798 + }, + { + "epoch": 0.27658714112414073, + "grad_norm": 0.0624038502573967, + "learning_rate": 0.00016460514361360583, + "loss": 0.0155, + "step": 11799 + }, + { + "epoch": 0.27661058269894573, + "grad_norm": 0.6091800332069397, + "learning_rate": 0.00016459952089517023, + "loss": 0.1663, + "step": 11800 + }, + { + "epoch": 0.2766340242737507, + "grad_norm": 0.637221097946167, + "learning_rate": 0.00016459389782621584, + "loss": 0.1362, + "step": 11801 + }, + { + "epoch": 0.2766574658485557, + "grad_norm": 0.4520901143550873, + "learning_rate": 0.00016458827440677315, + "loss": 0.12, + "step": 11802 + }, + { + "epoch": 0.2766809074233607, + "grad_norm": 0.5095171928405762, + "learning_rate": 0.00016458265063687264, + "loss": 0.1331, + "step": 11803 + }, + { + "epoch": 0.2767043489981657, + "grad_norm": 0.3193233609199524, + "learning_rate": 0.00016457702651654488, + "loss": 0.1182, + "step": 11804 + }, + { + "epoch": 0.2767277905729707, + "grad_norm": 0.5599116086959839, + "learning_rate": 0.00016457140204582032, + "loss": 0.0606, + "step": 11805 + }, + { + "epoch": 0.2767512321477757, + "grad_norm": 0.5281667709350586, + "learning_rate": 0.00016456577722472954, + "loss": 0.1356, + "step": 11806 + }, + { + "epoch": 0.2767746737225807, + "grad_norm": 0.9914854764938354, + "learning_rate": 0.00016456015205330304, + "loss": 0.5273, + "step": 11807 + }, + { + "epoch": 0.2767981152973857, + "grad_norm": 0.12298418581485748, + "learning_rate": 0.00016455452653157132, + "loss": 0.0236, + "step": 11808 + }, + { + "epoch": 0.2768215568721907, + "grad_norm": 0.39184069633483887, + "learning_rate": 0.00016454890065956495, + "loss": 0.1057, + "step": 11809 + }, + { + "epoch": 0.27684499844699567, + "grad_norm": 0.9415633678436279, + "learning_rate": 0.00016454327443731447, + "loss": 0.1485, + "step": 11810 + }, + { + "epoch": 0.27686844002180067, + "grad_norm": 0.6223465204238892, + "learning_rate": 0.00016453764786485028, + "loss": 0.1185, + "step": 11811 + }, + { + "epoch": 0.27689188159660566, + "grad_norm": 0.4334876239299774, + "learning_rate": 0.00016453202094220305, + "loss": 0.1395, + "step": 11812 + }, + { + "epoch": 0.27691532317141065, + "grad_norm": 0.634368896484375, + "learning_rate": 0.00016452639366940325, + "loss": 0.2042, + "step": 11813 + }, + { + "epoch": 0.27693876474621565, + "grad_norm": 0.2377336174249649, + "learning_rate": 0.0001645207660464814, + "loss": 0.0764, + "step": 11814 + }, + { + "epoch": 0.27696220632102064, + "grad_norm": 0.5398734211921692, + "learning_rate": 0.0001645151380734681, + "loss": 0.0795, + "step": 11815 + }, + { + "epoch": 0.27698564789582564, + "grad_norm": 0.2053510695695877, + "learning_rate": 0.00016450950975039383, + "loss": 0.0462, + "step": 11816 + }, + { + "epoch": 0.27700908947063063, + "grad_norm": 0.3385700285434723, + "learning_rate": 0.00016450388107728912, + "loss": 0.0684, + "step": 11817 + }, + { + "epoch": 0.2770325310454356, + "grad_norm": 0.41106027364730835, + "learning_rate": 0.00016449825205418458, + "loss": 0.1573, + "step": 11818 + }, + { + "epoch": 0.2770559726202406, + "grad_norm": 0.34449416399002075, + "learning_rate": 0.00016449262268111066, + "loss": 0.0765, + "step": 11819 + }, + { + "epoch": 0.2770794141950456, + "grad_norm": 0.5653596520423889, + "learning_rate": 0.00016448699295809803, + "loss": 0.1377, + "step": 11820 + }, + { + "epoch": 0.2771028557698506, + "grad_norm": 0.47109073400497437, + "learning_rate": 0.0001644813628851771, + "loss": 0.0929, + "step": 11821 + }, + { + "epoch": 0.2771262973446556, + "grad_norm": 0.6422905325889587, + "learning_rate": 0.0001644757324623785, + "loss": 0.8825, + "step": 11822 + }, + { + "epoch": 0.2771497389194606, + "grad_norm": 0.43961581587791443, + "learning_rate": 0.00016447010168973275, + "loss": 0.0617, + "step": 11823 + }, + { + "epoch": 0.2771731804942656, + "grad_norm": 0.5735809206962585, + "learning_rate": 0.00016446447056727047, + "loss": 0.1351, + "step": 11824 + }, + { + "epoch": 0.2771966220690706, + "grad_norm": 0.8577000498771667, + "learning_rate": 0.0001644588390950221, + "loss": 0.0958, + "step": 11825 + }, + { + "epoch": 0.2772200636438756, + "grad_norm": 0.25278186798095703, + "learning_rate": 0.00016445320727301828, + "loss": 0.0263, + "step": 11826 + }, + { + "epoch": 0.2772435052186806, + "grad_norm": 0.6516640782356262, + "learning_rate": 0.00016444757510128955, + "loss": 0.1305, + "step": 11827 + }, + { + "epoch": 0.27726694679348557, + "grad_norm": 0.5132516622543335, + "learning_rate": 0.00016444194257986646, + "loss": 0.6788, + "step": 11828 + }, + { + "epoch": 0.27729038836829056, + "grad_norm": 0.3462480902671814, + "learning_rate": 0.00016443630970877957, + "loss": 0.0842, + "step": 11829 + }, + { + "epoch": 0.27731382994309556, + "grad_norm": 0.19156497716903687, + "learning_rate": 0.00016443067648805944, + "loss": 0.0396, + "step": 11830 + }, + { + "epoch": 0.27733727151790055, + "grad_norm": 0.5153577327728271, + "learning_rate": 0.00016442504291773666, + "loss": 0.0923, + "step": 11831 + }, + { + "epoch": 0.27736071309270555, + "grad_norm": 0.5579428672790527, + "learning_rate": 0.00016441940899784182, + "loss": 0.2109, + "step": 11832 + }, + { + "epoch": 0.27738415466751054, + "grad_norm": 0.4187687039375305, + "learning_rate": 0.0001644137747284054, + "loss": 0.082, + "step": 11833 + }, + { + "epoch": 0.27740759624231554, + "grad_norm": 0.49472612142562866, + "learning_rate": 0.00016440814010945805, + "loss": 0.1309, + "step": 11834 + }, + { + "epoch": 0.27743103781712053, + "grad_norm": 0.6256398558616638, + "learning_rate": 0.00016440250514103034, + "loss": 0.1481, + "step": 11835 + }, + { + "epoch": 0.2774544793919255, + "grad_norm": 0.8767985701560974, + "learning_rate": 0.00016439686982315283, + "loss": 0.1816, + "step": 11836 + }, + { + "epoch": 0.2774779209667305, + "grad_norm": 0.7656236886978149, + "learning_rate": 0.00016439123415585603, + "loss": 0.1347, + "step": 11837 + }, + { + "epoch": 0.27750136254153557, + "grad_norm": 0.6721825003623962, + "learning_rate": 0.00016438559813917065, + "loss": 0.1723, + "step": 11838 + }, + { + "epoch": 0.27752480411634056, + "grad_norm": 0.7250674962997437, + "learning_rate": 0.00016437996177312718, + "loss": 0.1568, + "step": 11839 + }, + { + "epoch": 0.27754824569114556, + "grad_norm": 0.8675394058227539, + "learning_rate": 0.00016437432505775622, + "loss": 0.2661, + "step": 11840 + }, + { + "epoch": 0.27757168726595055, + "grad_norm": 0.5829012393951416, + "learning_rate": 0.00016436868799308839, + "loss": 0.0563, + "step": 11841 + }, + { + "epoch": 0.27759512884075555, + "grad_norm": 0.9460395574569702, + "learning_rate": 0.00016436305057915425, + "loss": 0.2724, + "step": 11842 + }, + { + "epoch": 0.27761857041556054, + "grad_norm": 0.5330770015716553, + "learning_rate": 0.00016435741281598434, + "loss": 0.1739, + "step": 11843 + }, + { + "epoch": 0.27764201199036553, + "grad_norm": 0.14742669463157654, + "learning_rate": 0.0001643517747036093, + "loss": 0.023, + "step": 11844 + }, + { + "epoch": 0.27766545356517053, + "grad_norm": 0.5330443382263184, + "learning_rate": 0.00016434613624205976, + "loss": 0.1182, + "step": 11845 + }, + { + "epoch": 0.2776888951399755, + "grad_norm": 0.6165685057640076, + "learning_rate": 0.00016434049743136627, + "loss": 0.2246, + "step": 11846 + }, + { + "epoch": 0.2777123367147805, + "grad_norm": 0.19895437359809875, + "learning_rate": 0.00016433485827155943, + "loss": 0.0525, + "step": 11847 + }, + { + "epoch": 0.2777357782895855, + "grad_norm": 0.4914509057998657, + "learning_rate": 0.00016432921876266985, + "loss": 0.0909, + "step": 11848 + }, + { + "epoch": 0.2777592198643905, + "grad_norm": 0.1852705478668213, + "learning_rate": 0.00016432357890472807, + "loss": 0.0496, + "step": 11849 + }, + { + "epoch": 0.2777826614391955, + "grad_norm": 0.42464151978492737, + "learning_rate": 0.0001643179386977648, + "loss": 0.1195, + "step": 11850 + }, + { + "epoch": 0.2778061030140005, + "grad_norm": 0.5287490487098694, + "learning_rate": 0.00016431229814181054, + "loss": 0.1083, + "step": 11851 + }, + { + "epoch": 0.2778295445888055, + "grad_norm": 0.6504444479942322, + "learning_rate": 0.000164306657236896, + "loss": 0.2754, + "step": 11852 + }, + { + "epoch": 0.2778529861636105, + "grad_norm": 0.15412446856498718, + "learning_rate": 0.00016430101598305169, + "loss": 0.0325, + "step": 11853 + }, + { + "epoch": 0.2778764277384155, + "grad_norm": 1.1380234956741333, + "learning_rate": 0.00016429537438030826, + "loss": 0.4706, + "step": 11854 + }, + { + "epoch": 0.27789986931322047, + "grad_norm": 0.5461359620094299, + "learning_rate": 0.00016428973242869633, + "loss": 0.1291, + "step": 11855 + }, + { + "epoch": 0.27792331088802547, + "grad_norm": 0.5290210843086243, + "learning_rate": 0.00016428409012824647, + "loss": 0.0992, + "step": 11856 + }, + { + "epoch": 0.27794675246283046, + "grad_norm": 0.7046567797660828, + "learning_rate": 0.00016427844747898937, + "loss": 0.1113, + "step": 11857 + }, + { + "epoch": 0.27797019403763545, + "grad_norm": 0.7392265796661377, + "learning_rate": 0.00016427280448095562, + "loss": 0.1507, + "step": 11858 + }, + { + "epoch": 0.27799363561244045, + "grad_norm": 0.2318187803030014, + "learning_rate": 0.00016426716113417578, + "loss": 0.0504, + "step": 11859 + }, + { + "epoch": 0.27801707718724544, + "grad_norm": 0.4750198423862457, + "learning_rate": 0.00016426151743868052, + "loss": 0.088, + "step": 11860 + }, + { + "epoch": 0.27804051876205044, + "grad_norm": 0.36068880558013916, + "learning_rate": 0.00016425587339450047, + "loss": 0.0857, + "step": 11861 + }, + { + "epoch": 0.27806396033685543, + "grad_norm": 0.48064208030700684, + "learning_rate": 0.00016425022900166627, + "loss": 0.0816, + "step": 11862 + }, + { + "epoch": 0.2780874019116604, + "grad_norm": 0.505423903465271, + "learning_rate": 0.00016424458426020851, + "loss": 0.1179, + "step": 11863 + }, + { + "epoch": 0.2781108434864654, + "grad_norm": 0.11937631666660309, + "learning_rate": 0.0001642389391701578, + "loss": 0.0367, + "step": 11864 + }, + { + "epoch": 0.2781342850612704, + "grad_norm": 0.6110888123512268, + "learning_rate": 0.00016423329373154482, + "loss": 0.7149, + "step": 11865 + }, + { + "epoch": 0.2781577266360754, + "grad_norm": 0.4317600429058075, + "learning_rate": 0.00016422764794440023, + "loss": 0.0889, + "step": 11866 + }, + { + "epoch": 0.2781811682108804, + "grad_norm": 0.32257601618766785, + "learning_rate": 0.00016422200180875454, + "loss": 0.0773, + "step": 11867 + }, + { + "epoch": 0.2782046097856854, + "grad_norm": 0.5702155232429504, + "learning_rate": 0.0001642163553246385, + "loss": 0.1057, + "step": 11868 + }, + { + "epoch": 0.2782280513604904, + "grad_norm": 0.5487154126167297, + "learning_rate": 0.00016421070849208268, + "loss": 0.0583, + "step": 11869 + }, + { + "epoch": 0.2782514929352954, + "grad_norm": 0.4233360290527344, + "learning_rate": 0.00016420506131111778, + "loss": 0.1175, + "step": 11870 + }, + { + "epoch": 0.2782749345101004, + "grad_norm": 0.12285589426755905, + "learning_rate": 0.0001641994137817744, + "loss": 0.0309, + "step": 11871 + }, + { + "epoch": 0.2782983760849054, + "grad_norm": 0.9037995338439941, + "learning_rate": 0.00016419376590408322, + "loss": 0.1911, + "step": 11872 + }, + { + "epoch": 0.27832181765971037, + "grad_norm": 0.1910996437072754, + "learning_rate": 0.00016418811767807484, + "loss": 0.0378, + "step": 11873 + }, + { + "epoch": 0.27834525923451536, + "grad_norm": 0.4440511465072632, + "learning_rate": 0.00016418246910377993, + "loss": 0.0767, + "step": 11874 + }, + { + "epoch": 0.27836870080932036, + "grad_norm": 0.19605150818824768, + "learning_rate": 0.00016417682018122912, + "loss": 0.048, + "step": 11875 + }, + { + "epoch": 0.27839214238412535, + "grad_norm": 0.8970569372177124, + "learning_rate": 0.00016417117091045312, + "loss": 0.1501, + "step": 11876 + }, + { + "epoch": 0.27841558395893035, + "grad_norm": 0.1720980405807495, + "learning_rate": 0.0001641655212914825, + "loss": 0.032, + "step": 11877 + }, + { + "epoch": 0.27843902553373534, + "grad_norm": 0.3761272430419922, + "learning_rate": 0.000164159871324348, + "loss": 0.0516, + "step": 11878 + }, + { + "epoch": 0.27846246710854033, + "grad_norm": 1.1273998022079468, + "learning_rate": 0.0001641542210090802, + "loss": 0.278, + "step": 11879 + }, + { + "epoch": 0.27848590868334533, + "grad_norm": 0.16277778148651123, + "learning_rate": 0.0001641485703457098, + "loss": 0.0272, + "step": 11880 + }, + { + "epoch": 0.2785093502581503, + "grad_norm": 0.3489777147769928, + "learning_rate": 0.00016414291933426745, + "loss": 0.0383, + "step": 11881 + }, + { + "epoch": 0.2785327918329553, + "grad_norm": 0.30963537096977234, + "learning_rate": 0.00016413726797478385, + "loss": 0.0781, + "step": 11882 + }, + { + "epoch": 0.2785562334077603, + "grad_norm": 0.4492591619491577, + "learning_rate": 0.0001641316162672896, + "loss": 0.081, + "step": 11883 + }, + { + "epoch": 0.2785796749825653, + "grad_norm": 0.5044496059417725, + "learning_rate": 0.0001641259642118154, + "loss": 0.1678, + "step": 11884 + }, + { + "epoch": 0.2786031165573703, + "grad_norm": 1.0495636463165283, + "learning_rate": 0.00016412031180839193, + "loss": 0.109, + "step": 11885 + }, + { + "epoch": 0.2786265581321753, + "grad_norm": 0.38958320021629333, + "learning_rate": 0.0001641146590570498, + "loss": 0.0745, + "step": 11886 + }, + { + "epoch": 0.2786499997069803, + "grad_norm": 0.34056803584098816, + "learning_rate": 0.00016410900595781978, + "loss": 0.1042, + "step": 11887 + }, + { + "epoch": 0.2786734412817853, + "grad_norm": 0.4944712519645691, + "learning_rate": 0.00016410335251073246, + "loss": 0.1234, + "step": 11888 + }, + { + "epoch": 0.27869688285659033, + "grad_norm": 0.5222362279891968, + "learning_rate": 0.00016409769871581856, + "loss": 0.123, + "step": 11889 + }, + { + "epoch": 0.2787203244313953, + "grad_norm": 0.12848074734210968, + "learning_rate": 0.00016409204457310875, + "loss": 0.0173, + "step": 11890 + }, + { + "epoch": 0.2787437660062003, + "grad_norm": 0.31643345952033997, + "learning_rate": 0.0001640863900826337, + "loss": 0.0401, + "step": 11891 + }, + { + "epoch": 0.2787672075810053, + "grad_norm": 0.2824484705924988, + "learning_rate": 0.00016408073524442406, + "loss": 0.0469, + "step": 11892 + }, + { + "epoch": 0.2787906491558103, + "grad_norm": 0.6443544626235962, + "learning_rate": 0.0001640750800585106, + "loss": 0.1249, + "step": 11893 + }, + { + "epoch": 0.2788140907306153, + "grad_norm": 0.16617687046527863, + "learning_rate": 0.0001640694245249239, + "loss": 0.0352, + "step": 11894 + }, + { + "epoch": 0.2788375323054203, + "grad_norm": 0.10517755895853043, + "learning_rate": 0.00016406376864369477, + "loss": 0.026, + "step": 11895 + }, + { + "epoch": 0.2788609738802253, + "grad_norm": 0.7045244574546814, + "learning_rate": 0.00016405811241485379, + "loss": 0.1841, + "step": 11896 + }, + { + "epoch": 0.2788844154550303, + "grad_norm": 0.2249855250120163, + "learning_rate": 0.0001640524558384317, + "loss": 0.0443, + "step": 11897 + }, + { + "epoch": 0.2789078570298353, + "grad_norm": 0.6540083289146423, + "learning_rate": 0.00016404679891445919, + "loss": 0.1658, + "step": 11898 + }, + { + "epoch": 0.2789312986046403, + "grad_norm": 0.5093575119972229, + "learning_rate": 0.0001640411416429669, + "loss": 0.1341, + "step": 11899 + }, + { + "epoch": 0.27895474017944527, + "grad_norm": 0.6629198789596558, + "learning_rate": 0.00016403548402398564, + "loss": 0.1722, + "step": 11900 + }, + { + "epoch": 0.27897818175425027, + "grad_norm": 0.5637006759643555, + "learning_rate": 0.00016402982605754599, + "loss": 0.1609, + "step": 11901 + }, + { + "epoch": 0.27900162332905526, + "grad_norm": 0.6494473814964294, + "learning_rate": 0.00016402416774367876, + "loss": 0.1206, + "step": 11902 + }, + { + "epoch": 0.27902506490386025, + "grad_norm": 0.32154932618141174, + "learning_rate": 0.00016401850908241453, + "loss": 0.3269, + "step": 11903 + }, + { + "epoch": 0.27904850647866525, + "grad_norm": 0.9597901701927185, + "learning_rate": 0.00016401285007378412, + "loss": 0.0714, + "step": 11904 + }, + { + "epoch": 0.27907194805347024, + "grad_norm": 0.2935883104801178, + "learning_rate": 0.00016400719071781817, + "loss": 0.0596, + "step": 11905 + }, + { + "epoch": 0.27909538962827524, + "grad_norm": 0.5101815462112427, + "learning_rate": 0.0001640015310145474, + "loss": 0.0634, + "step": 11906 + }, + { + "epoch": 0.27911883120308023, + "grad_norm": 0.5581859946250916, + "learning_rate": 0.00016399587096400256, + "loss": 0.1727, + "step": 11907 + }, + { + "epoch": 0.2791422727778852, + "grad_norm": 0.38730931282043457, + "learning_rate": 0.00016399021056621427, + "loss": 0.0658, + "step": 11908 + }, + { + "epoch": 0.2791657143526902, + "grad_norm": 0.31180161237716675, + "learning_rate": 0.00016398454982121332, + "loss": 0.0489, + "step": 11909 + }, + { + "epoch": 0.2791891559274952, + "grad_norm": 0.8050777912139893, + "learning_rate": 0.00016397888872903037, + "loss": 0.1754, + "step": 11910 + }, + { + "epoch": 0.2792125975023002, + "grad_norm": 0.19481819868087769, + "learning_rate": 0.00016397322728969624, + "loss": 0.0631, + "step": 11911 + }, + { + "epoch": 0.2792360390771052, + "grad_norm": 0.1419839859008789, + "learning_rate": 0.00016396756550324156, + "loss": 0.0201, + "step": 11912 + }, + { + "epoch": 0.2792594806519102, + "grad_norm": 0.7654271721839905, + "learning_rate": 0.00016396190336969704, + "loss": 0.0874, + "step": 11913 + }, + { + "epoch": 0.2792829222267152, + "grad_norm": 0.8640219569206238, + "learning_rate": 0.00016395624088909344, + "loss": 0.1946, + "step": 11914 + }, + { + "epoch": 0.2793063638015202, + "grad_norm": 0.5087112188339233, + "learning_rate": 0.0001639505780614615, + "loss": 0.1185, + "step": 11915 + }, + { + "epoch": 0.2793298053763252, + "grad_norm": 0.2901953160762787, + "learning_rate": 0.0001639449148868319, + "loss": 0.0505, + "step": 11916 + }, + { + "epoch": 0.2793532469511302, + "grad_norm": 0.8396941423416138, + "learning_rate": 0.0001639392513652354, + "loss": 0.1976, + "step": 11917 + }, + { + "epoch": 0.27937668852593517, + "grad_norm": 0.4903525412082672, + "learning_rate": 0.00016393358749670274, + "loss": 0.1455, + "step": 11918 + }, + { + "epoch": 0.27940013010074016, + "grad_norm": 0.14442048966884613, + "learning_rate": 0.00016392792328126464, + "loss": 0.0283, + "step": 11919 + }, + { + "epoch": 0.27942357167554516, + "grad_norm": 0.5456260442733765, + "learning_rate": 0.00016392225871895183, + "loss": 0.091, + "step": 11920 + }, + { + "epoch": 0.27944701325035015, + "grad_norm": 0.6180163025856018, + "learning_rate": 0.00016391659380979502, + "loss": 0.1846, + "step": 11921 + }, + { + "epoch": 0.27947045482515515, + "grad_norm": 0.427498996257782, + "learning_rate": 0.00016391092855382497, + "loss": 0.0788, + "step": 11922 + }, + { + "epoch": 0.27949389639996014, + "grad_norm": 0.7252967953681946, + "learning_rate": 0.00016390526295107245, + "loss": 0.1251, + "step": 11923 + }, + { + "epoch": 0.27951733797476513, + "grad_norm": 0.3045293092727661, + "learning_rate": 0.00016389959700156817, + "loss": 0.0686, + "step": 11924 + }, + { + "epoch": 0.27954077954957013, + "grad_norm": 0.46735483407974243, + "learning_rate": 0.0001638939307053429, + "loss": 0.4956, + "step": 11925 + }, + { + "epoch": 0.2795642211243751, + "grad_norm": 0.4451388716697693, + "learning_rate": 0.00016388826406242733, + "loss": 0.0829, + "step": 11926 + }, + { + "epoch": 0.2795876626991801, + "grad_norm": 0.6598297953605652, + "learning_rate": 0.00016388259707285227, + "loss": 0.488, + "step": 11927 + }, + { + "epoch": 0.2796111042739851, + "grad_norm": 0.5566425919532776, + "learning_rate": 0.0001638769297366484, + "loss": 0.1446, + "step": 11928 + }, + { + "epoch": 0.2796345458487901, + "grad_norm": 0.35733261704444885, + "learning_rate": 0.00016387126205384656, + "loss": 0.0719, + "step": 11929 + }, + { + "epoch": 0.2796579874235951, + "grad_norm": 0.41180145740509033, + "learning_rate": 0.00016386559402447747, + "loss": 0.144, + "step": 11930 + }, + { + "epoch": 0.2796814289984001, + "grad_norm": 0.17417238652706146, + "learning_rate": 0.00016385992564857182, + "loss": 0.0315, + "step": 11931 + }, + { + "epoch": 0.2797048705732051, + "grad_norm": 0.5887572765350342, + "learning_rate": 0.00016385425692616042, + "loss": 0.1417, + "step": 11932 + }, + { + "epoch": 0.2797283121480101, + "grad_norm": 0.5959784984588623, + "learning_rate": 0.00016384858785727406, + "loss": 0.1859, + "step": 11933 + }, + { + "epoch": 0.2797517537228151, + "grad_norm": 0.5595190525054932, + "learning_rate": 0.00016384291844194344, + "loss": 0.166, + "step": 11934 + }, + { + "epoch": 0.27977519529762007, + "grad_norm": 0.33191442489624023, + "learning_rate": 0.00016383724868019938, + "loss": 0.0789, + "step": 11935 + }, + { + "epoch": 0.27979863687242507, + "grad_norm": 0.4105977416038513, + "learning_rate": 0.00016383157857207257, + "loss": 0.107, + "step": 11936 + }, + { + "epoch": 0.27982207844723006, + "grad_norm": 0.6731806993484497, + "learning_rate": 0.00016382590811759386, + "loss": 0.1178, + "step": 11937 + }, + { + "epoch": 0.27984552002203505, + "grad_norm": 0.1573515683412552, + "learning_rate": 0.00016382023731679397, + "loss": 0.0348, + "step": 11938 + }, + { + "epoch": 0.27986896159684005, + "grad_norm": 0.42134127020835876, + "learning_rate": 0.0001638145661697037, + "loss": 0.0847, + "step": 11939 + }, + { + "epoch": 0.2798924031716451, + "grad_norm": 0.6772169470787048, + "learning_rate": 0.00016380889467635371, + "loss": 0.4624, + "step": 11940 + }, + { + "epoch": 0.2799158447464501, + "grad_norm": 1.087103247642517, + "learning_rate": 0.00016380322283677496, + "loss": 0.1475, + "step": 11941 + }, + { + "epoch": 0.2799392863212551, + "grad_norm": 0.9245229363441467, + "learning_rate": 0.00016379755065099808, + "loss": 0.191, + "step": 11942 + }, + { + "epoch": 0.2799627278960601, + "grad_norm": 0.3400633633136749, + "learning_rate": 0.00016379187811905392, + "loss": 0.0475, + "step": 11943 + }, + { + "epoch": 0.2799861694708651, + "grad_norm": 0.47959861159324646, + "learning_rate": 0.00016378620524097324, + "loss": 0.1216, + "step": 11944 + }, + { + "epoch": 0.28000961104567007, + "grad_norm": 0.2565518319606781, + "learning_rate": 0.0001637805320167868, + "loss": 0.0441, + "step": 11945 + }, + { + "epoch": 0.28003305262047506, + "grad_norm": 0.459965318441391, + "learning_rate": 0.0001637748584465254, + "loss": 0.0925, + "step": 11946 + }, + { + "epoch": 0.28005649419528006, + "grad_norm": 0.32654014229774475, + "learning_rate": 0.00016376918453021982, + "loss": 0.0433, + "step": 11947 + }, + { + "epoch": 0.28007993577008505, + "grad_norm": 0.1846461445093155, + "learning_rate": 0.00016376351026790083, + "loss": 0.0358, + "step": 11948 + }, + { + "epoch": 0.28010337734489005, + "grad_norm": 0.5760862231254578, + "learning_rate": 0.0001637578356595993, + "loss": 0.0855, + "step": 11949 + }, + { + "epoch": 0.28012681891969504, + "grad_norm": 0.627992570400238, + "learning_rate": 0.00016375216070534595, + "loss": 0.098, + "step": 11950 + }, + { + "epoch": 0.28015026049450004, + "grad_norm": 0.7291430830955505, + "learning_rate": 0.00016374648540517156, + "loss": 0.1812, + "step": 11951 + }, + { + "epoch": 0.28017370206930503, + "grad_norm": 0.45126107335090637, + "learning_rate": 0.00016374080975910694, + "loss": 0.0831, + "step": 11952 + }, + { + "epoch": 0.28019714364411, + "grad_norm": 0.23034527897834778, + "learning_rate": 0.0001637351337671829, + "loss": 0.0577, + "step": 11953 + }, + { + "epoch": 0.280220585218915, + "grad_norm": 0.13853509724140167, + "learning_rate": 0.00016372945742943023, + "loss": 0.0248, + "step": 11954 + }, + { + "epoch": 0.28024402679372, + "grad_norm": 0.41852906346321106, + "learning_rate": 0.00016372378074587976, + "loss": 0.084, + "step": 11955 + }, + { + "epoch": 0.280267468368525, + "grad_norm": 0.48947787284851074, + "learning_rate": 0.00016371810371656223, + "loss": 0.0947, + "step": 11956 + }, + { + "epoch": 0.28029090994333, + "grad_norm": 0.5482433438301086, + "learning_rate": 0.0001637124263415085, + "loss": 0.0885, + "step": 11957 + }, + { + "epoch": 0.280314351518135, + "grad_norm": 0.2880384624004364, + "learning_rate": 0.00016370674862074933, + "loss": 0.0448, + "step": 11958 + }, + { + "epoch": 0.28033779309294, + "grad_norm": 0.6213115453720093, + "learning_rate": 0.00016370107055431557, + "loss": 0.1312, + "step": 11959 + }, + { + "epoch": 0.280361234667745, + "grad_norm": 0.23900097608566284, + "learning_rate": 0.000163695392142238, + "loss": 0.0413, + "step": 11960 + }, + { + "epoch": 0.28038467624255, + "grad_norm": 0.7569602727890015, + "learning_rate": 0.00016368971338454746, + "loss": 0.6541, + "step": 11961 + }, + { + "epoch": 0.280408117817355, + "grad_norm": 0.7514211535453796, + "learning_rate": 0.00016368403428127472, + "loss": 0.1166, + "step": 11962 + }, + { + "epoch": 0.28043155939215997, + "grad_norm": 0.18580006062984467, + "learning_rate": 0.00016367835483245064, + "loss": 0.0271, + "step": 11963 + }, + { + "epoch": 0.28045500096696496, + "grad_norm": 0.6641671061515808, + "learning_rate": 0.000163672675038106, + "loss": 0.1044, + "step": 11964 + }, + { + "epoch": 0.28047844254176996, + "grad_norm": 0.21863549947738647, + "learning_rate": 0.00016366699489827165, + "loss": 0.0449, + "step": 11965 + }, + { + "epoch": 0.28050188411657495, + "grad_norm": 0.21364319324493408, + "learning_rate": 0.00016366131441297838, + "loss": 0.0596, + "step": 11966 + }, + { + "epoch": 0.28052532569137995, + "grad_norm": 0.5843938589096069, + "learning_rate": 0.00016365563358225704, + "loss": 0.1153, + "step": 11967 + }, + { + "epoch": 0.28054876726618494, + "grad_norm": 1.0453598499298096, + "learning_rate": 0.00016364995240613842, + "loss": 0.2116, + "step": 11968 + }, + { + "epoch": 0.28057220884098993, + "grad_norm": 0.6708529591560364, + "learning_rate": 0.00016364427088465342, + "loss": 0.1155, + "step": 11969 + }, + { + "epoch": 0.28059565041579493, + "grad_norm": 0.4742767810821533, + "learning_rate": 0.00016363858901783277, + "loss": 0.1382, + "step": 11970 + }, + { + "epoch": 0.2806190919905999, + "grad_norm": 1.0344620943069458, + "learning_rate": 0.00016363290680570735, + "loss": 0.0713, + "step": 11971 + }, + { + "epoch": 0.2806425335654049, + "grad_norm": 0.3546989858150482, + "learning_rate": 0.00016362722424830798, + "loss": 0.1169, + "step": 11972 + }, + { + "epoch": 0.2806659751402099, + "grad_norm": 0.6472330093383789, + "learning_rate": 0.00016362154134566553, + "loss": 0.793, + "step": 11973 + }, + { + "epoch": 0.2806894167150149, + "grad_norm": 0.6038250923156738, + "learning_rate": 0.0001636158580978108, + "loss": 0.1154, + "step": 11974 + }, + { + "epoch": 0.2807128582898199, + "grad_norm": 0.1691923588514328, + "learning_rate": 0.00016361017450477466, + "loss": 0.0416, + "step": 11975 + }, + { + "epoch": 0.2807362998646249, + "grad_norm": 0.2794003188610077, + "learning_rate": 0.0001636044905665879, + "loss": 0.0672, + "step": 11976 + }, + { + "epoch": 0.2807597414394299, + "grad_norm": 0.723242461681366, + "learning_rate": 0.00016359880628328134, + "loss": 0.1342, + "step": 11977 + }, + { + "epoch": 0.2807831830142349, + "grad_norm": 0.42661944031715393, + "learning_rate": 0.00016359312165488592, + "loss": 0.0897, + "step": 11978 + }, + { + "epoch": 0.2808066245890399, + "grad_norm": 0.7998427152633667, + "learning_rate": 0.00016358743668143245, + "loss": 0.3819, + "step": 11979 + }, + { + "epoch": 0.28083006616384487, + "grad_norm": 0.4883781373500824, + "learning_rate": 0.00016358175136295172, + "loss": 0.097, + "step": 11980 + }, + { + "epoch": 0.28085350773864987, + "grad_norm": 0.5550922751426697, + "learning_rate": 0.00016357606569947466, + "loss": 0.0752, + "step": 11981 + }, + { + "epoch": 0.28087694931345486, + "grad_norm": 0.5177724957466125, + "learning_rate": 0.00016357037969103204, + "loss": 0.1646, + "step": 11982 + }, + { + "epoch": 0.28090039088825985, + "grad_norm": 0.7174572944641113, + "learning_rate": 0.00016356469333765475, + "loss": 0.1253, + "step": 11983 + }, + { + "epoch": 0.28092383246306485, + "grad_norm": 0.2690947949886322, + "learning_rate": 0.0001635590066393737, + "loss": 0.0624, + "step": 11984 + }, + { + "epoch": 0.28094727403786984, + "grad_norm": 0.6594566702842712, + "learning_rate": 0.00016355331959621964, + "loss": 0.1752, + "step": 11985 + }, + { + "epoch": 0.28097071561267484, + "grad_norm": 0.5312314033508301, + "learning_rate": 0.0001635476322082235, + "loss": 0.073, + "step": 11986 + }, + { + "epoch": 0.28099415718747983, + "grad_norm": 0.0906430035829544, + "learning_rate": 0.00016354194447541613, + "loss": 0.0203, + "step": 11987 + }, + { + "epoch": 0.2810175987622848, + "grad_norm": 0.2964935004711151, + "learning_rate": 0.00016353625639782836, + "loss": 0.0314, + "step": 11988 + }, + { + "epoch": 0.2810410403370898, + "grad_norm": 0.7697287201881409, + "learning_rate": 0.00016353056797549108, + "loss": 0.0932, + "step": 11989 + }, + { + "epoch": 0.2810644819118948, + "grad_norm": 0.5427523255348206, + "learning_rate": 0.00016352487920843518, + "loss": 0.1337, + "step": 11990 + }, + { + "epoch": 0.2810879234866998, + "grad_norm": 0.515056848526001, + "learning_rate": 0.00016351919009669147, + "loss": 0.0875, + "step": 11991 + }, + { + "epoch": 0.28111136506150486, + "grad_norm": 0.2312411069869995, + "learning_rate": 0.00016351350064029084, + "loss": 0.0417, + "step": 11992 + }, + { + "epoch": 0.28113480663630985, + "grad_norm": 0.536949098110199, + "learning_rate": 0.0001635078108392642, + "loss": 0.1661, + "step": 11993 + }, + { + "epoch": 0.28115824821111485, + "grad_norm": 0.2811684310436249, + "learning_rate": 0.00016350212069364238, + "loss": 0.0561, + "step": 11994 + }, + { + "epoch": 0.28118168978591984, + "grad_norm": 0.3984359800815582, + "learning_rate": 0.00016349643020345628, + "loss": 0.0946, + "step": 11995 + }, + { + "epoch": 0.28120513136072484, + "grad_norm": 0.5473271608352661, + "learning_rate": 0.00016349073936873672, + "loss": 0.1325, + "step": 11996 + }, + { + "epoch": 0.28122857293552983, + "grad_norm": 0.4704720377922058, + "learning_rate": 0.00016348504818951466, + "loss": 0.1313, + "step": 11997 + }, + { + "epoch": 0.2812520145103348, + "grad_norm": 0.6452811360359192, + "learning_rate": 0.00016347935666582097, + "loss": 0.0982, + "step": 11998 + }, + { + "epoch": 0.2812754560851398, + "grad_norm": 0.9624441266059875, + "learning_rate": 0.00016347366479768646, + "loss": 0.1889, + "step": 11999 + }, + { + "epoch": 0.2812988976599448, + "grad_norm": 0.5074067115783691, + "learning_rate": 0.00016346797258514207, + "loss": 0.1862, + "step": 12000 + }, + { + "epoch": 0.2813223392347498, + "grad_norm": 0.3731517493724823, + "learning_rate": 0.00016346228002821867, + "loss": 0.0544, + "step": 12001 + }, + { + "epoch": 0.2813457808095548, + "grad_norm": 0.3433060050010681, + "learning_rate": 0.0001634565871269472, + "loss": 0.0527, + "step": 12002 + }, + { + "epoch": 0.2813692223843598, + "grad_norm": 0.5815003514289856, + "learning_rate": 0.00016345089388135846, + "loss": 0.8867, + "step": 12003 + }, + { + "epoch": 0.2813926639591648, + "grad_norm": 0.20958012342453003, + "learning_rate": 0.00016344520029148337, + "loss": 0.0371, + "step": 12004 + }, + { + "epoch": 0.2814161055339698, + "grad_norm": 0.47977215051651, + "learning_rate": 0.0001634395063573529, + "loss": 0.1042, + "step": 12005 + }, + { + "epoch": 0.2814395471087748, + "grad_norm": 0.5358937382698059, + "learning_rate": 0.00016343381207899784, + "loss": 0.0781, + "step": 12006 + }, + { + "epoch": 0.2814629886835798, + "grad_norm": 0.23820537328720093, + "learning_rate": 0.00016342811745644913, + "loss": 0.0544, + "step": 12007 + }, + { + "epoch": 0.28148643025838477, + "grad_norm": 0.3451705873012543, + "learning_rate": 0.0001634224224897377, + "loss": 0.0968, + "step": 12008 + }, + { + "epoch": 0.28150987183318976, + "grad_norm": 0.19230994582176208, + "learning_rate": 0.0001634167271788944, + "loss": 0.0447, + "step": 12009 + }, + { + "epoch": 0.28153331340799476, + "grad_norm": 0.6259166598320007, + "learning_rate": 0.00016341103152395014, + "loss": 0.0988, + "step": 12010 + }, + { + "epoch": 0.28155675498279975, + "grad_norm": 0.20955903828144073, + "learning_rate": 0.00016340533552493586, + "loss": 0.0481, + "step": 12011 + }, + { + "epoch": 0.28158019655760474, + "grad_norm": 0.5690682530403137, + "learning_rate": 0.00016339963918188247, + "loss": 0.2229, + "step": 12012 + }, + { + "epoch": 0.28160363813240974, + "grad_norm": 0.3799203038215637, + "learning_rate": 0.0001633939424948208, + "loss": 0.0497, + "step": 12013 + }, + { + "epoch": 0.28162707970721473, + "grad_norm": 0.6505701541900635, + "learning_rate": 0.00016338824546378186, + "loss": 0.2311, + "step": 12014 + }, + { + "epoch": 0.28165052128201973, + "grad_norm": 0.6688111424446106, + "learning_rate": 0.0001633825480887965, + "loss": 0.632, + "step": 12015 + }, + { + "epoch": 0.2816739628568247, + "grad_norm": 0.574636697769165, + "learning_rate": 0.00016337685036989566, + "loss": 0.1053, + "step": 12016 + }, + { + "epoch": 0.2816974044316297, + "grad_norm": 0.570975661277771, + "learning_rate": 0.00016337115230711027, + "loss": 0.1496, + "step": 12017 + }, + { + "epoch": 0.2817208460064347, + "grad_norm": 0.5150288939476013, + "learning_rate": 0.0001633654539004712, + "loss": 0.0839, + "step": 12018 + }, + { + "epoch": 0.2817442875812397, + "grad_norm": 0.3704228699207306, + "learning_rate": 0.00016335975515000936, + "loss": 0.0704, + "step": 12019 + }, + { + "epoch": 0.2817677291560447, + "grad_norm": 0.22372660040855408, + "learning_rate": 0.00016335405605575576, + "loss": 0.0467, + "step": 12020 + }, + { + "epoch": 0.2817911707308497, + "grad_norm": 0.6321881413459778, + "learning_rate": 0.00016334835661774126, + "loss": 0.1116, + "step": 12021 + }, + { + "epoch": 0.2818146123056547, + "grad_norm": 0.6834200024604797, + "learning_rate": 0.00016334265683599678, + "loss": 0.2109, + "step": 12022 + }, + { + "epoch": 0.2818380538804597, + "grad_norm": 0.29193148016929626, + "learning_rate": 0.0001633369567105533, + "loss": 0.0572, + "step": 12023 + }, + { + "epoch": 0.2818614954552647, + "grad_norm": 0.48701635003089905, + "learning_rate": 0.00016333125624144165, + "loss": 0.1266, + "step": 12024 + }, + { + "epoch": 0.28188493703006967, + "grad_norm": 0.5020018219947815, + "learning_rate": 0.00016332555542869287, + "loss": 0.1157, + "step": 12025 + }, + { + "epoch": 0.28190837860487467, + "grad_norm": 0.3083523213863373, + "learning_rate": 0.00016331985427233783, + "loss": 0.0694, + "step": 12026 + }, + { + "epoch": 0.28193182017967966, + "grad_norm": 0.9902756214141846, + "learning_rate": 0.00016331415277240748, + "loss": 0.2642, + "step": 12027 + }, + { + "epoch": 0.28195526175448465, + "grad_norm": 0.4343772828578949, + "learning_rate": 0.00016330845092893276, + "loss": 0.4557, + "step": 12028 + }, + { + "epoch": 0.28197870332928965, + "grad_norm": 0.25507134199142456, + "learning_rate": 0.0001633027487419446, + "loss": 0.0588, + "step": 12029 + }, + { + "epoch": 0.28200214490409464, + "grad_norm": 0.6008937954902649, + "learning_rate": 0.00016329704621147396, + "loss": 0.9238, + "step": 12030 + }, + { + "epoch": 0.28202558647889964, + "grad_norm": 0.3674595057964325, + "learning_rate": 0.00016329134333755175, + "loss": 0.0537, + "step": 12031 + }, + { + "epoch": 0.28204902805370463, + "grad_norm": 0.15947552025318146, + "learning_rate": 0.00016328564012020895, + "loss": 0.035, + "step": 12032 + }, + { + "epoch": 0.2820724696285096, + "grad_norm": 0.6303535103797913, + "learning_rate": 0.00016327993655947646, + "loss": 0.1334, + "step": 12033 + }, + { + "epoch": 0.2820959112033146, + "grad_norm": 0.45030948519706726, + "learning_rate": 0.00016327423265538528, + "loss": 0.0804, + "step": 12034 + }, + { + "epoch": 0.2821193527781196, + "grad_norm": 0.14487779140472412, + "learning_rate": 0.00016326852840796633, + "loss": 0.0446, + "step": 12035 + }, + { + "epoch": 0.2821427943529246, + "grad_norm": 0.6718847751617432, + "learning_rate": 0.00016326282381725057, + "loss": 0.0982, + "step": 12036 + }, + { + "epoch": 0.2821662359277296, + "grad_norm": 0.5170446038246155, + "learning_rate": 0.0001632571188832689, + "loss": 0.6348, + "step": 12037 + }, + { + "epoch": 0.2821896775025346, + "grad_norm": 0.396089106798172, + "learning_rate": 0.00016325141360605237, + "loss": 0.0344, + "step": 12038 + }, + { + "epoch": 0.2822131190773396, + "grad_norm": 0.5601168870925903, + "learning_rate": 0.0001632457079856319, + "loss": 0.1349, + "step": 12039 + }, + { + "epoch": 0.2822365606521446, + "grad_norm": 0.7356196641921997, + "learning_rate": 0.00016324000202203845, + "loss": 0.2116, + "step": 12040 + }, + { + "epoch": 0.2822600022269496, + "grad_norm": 0.515817403793335, + "learning_rate": 0.00016323429571530294, + "loss": 0.0946, + "step": 12041 + }, + { + "epoch": 0.2822834438017546, + "grad_norm": 0.6224789023399353, + "learning_rate": 0.0001632285890654564, + "loss": 0.1098, + "step": 12042 + }, + { + "epoch": 0.2823068853765596, + "grad_norm": 1.0102012157440186, + "learning_rate": 0.0001632228820725297, + "loss": 0.1598, + "step": 12043 + }, + { + "epoch": 0.2823303269513646, + "grad_norm": 0.43903085589408875, + "learning_rate": 0.0001632171747365539, + "loss": 0.083, + "step": 12044 + }, + { + "epoch": 0.2823537685261696, + "grad_norm": 0.6714254021644592, + "learning_rate": 0.0001632114670575599, + "loss": 0.1191, + "step": 12045 + }, + { + "epoch": 0.2823772101009746, + "grad_norm": 0.2416757345199585, + "learning_rate": 0.00016320575903557875, + "loss": 0.0681, + "step": 12046 + }, + { + "epoch": 0.2824006516757796, + "grad_norm": 0.787117063999176, + "learning_rate": 0.00016320005067064136, + "loss": 0.2399, + "step": 12047 + }, + { + "epoch": 0.2824240932505846, + "grad_norm": 0.5880342721939087, + "learning_rate": 0.0001631943419627787, + "loss": 0.1607, + "step": 12048 + }, + { + "epoch": 0.2824475348253896, + "grad_norm": 0.36961430311203003, + "learning_rate": 0.0001631886329120218, + "loss": 0.088, + "step": 12049 + }, + { + "epoch": 0.2824709764001946, + "grad_norm": 0.5239953994750977, + "learning_rate": 0.00016318292351840153, + "loss": 0.1042, + "step": 12050 + }, + { + "epoch": 0.2824944179749996, + "grad_norm": 0.38482779264450073, + "learning_rate": 0.000163177213781949, + "loss": 0.0842, + "step": 12051 + }, + { + "epoch": 0.2825178595498046, + "grad_norm": 0.6636601090431213, + "learning_rate": 0.00016317150370269513, + "loss": 0.86, + "step": 12052 + }, + { + "epoch": 0.28254130112460957, + "grad_norm": 0.2600485682487488, + "learning_rate": 0.0001631657932806709, + "loss": 0.0546, + "step": 12053 + }, + { + "epoch": 0.28256474269941456, + "grad_norm": 0.6543569564819336, + "learning_rate": 0.00016316008251590728, + "loss": 0.6418, + "step": 12054 + }, + { + "epoch": 0.28258818427421956, + "grad_norm": 0.49293985962867737, + "learning_rate": 0.0001631543714084353, + "loss": 0.0667, + "step": 12055 + }, + { + "epoch": 0.28261162584902455, + "grad_norm": 0.4405202567577362, + "learning_rate": 0.0001631486599582859, + "loss": 0.0836, + "step": 12056 + }, + { + "epoch": 0.28263506742382954, + "grad_norm": 0.3515811562538147, + "learning_rate": 0.00016314294816549012, + "loss": 0.087, + "step": 12057 + }, + { + "epoch": 0.28265850899863454, + "grad_norm": 1.1675227880477905, + "learning_rate": 0.0001631372360300789, + "loss": 0.1042, + "step": 12058 + }, + { + "epoch": 0.28268195057343953, + "grad_norm": 0.13491997122764587, + "learning_rate": 0.00016313152355208329, + "loss": 0.0332, + "step": 12059 + }, + { + "epoch": 0.2827053921482445, + "grad_norm": 0.44126880168914795, + "learning_rate": 0.00016312581073153423, + "loss": 0.083, + "step": 12060 + }, + { + "epoch": 0.2827288337230495, + "grad_norm": 0.19468112289905548, + "learning_rate": 0.00016312009756846277, + "loss": 0.033, + "step": 12061 + }, + { + "epoch": 0.2827522752978545, + "grad_norm": 0.9187101721763611, + "learning_rate": 0.00016311438406289987, + "loss": 0.1558, + "step": 12062 + }, + { + "epoch": 0.2827757168726595, + "grad_norm": 0.5068452954292297, + "learning_rate": 0.00016310867021487657, + "loss": 0.1066, + "step": 12063 + }, + { + "epoch": 0.2827991584474645, + "grad_norm": 0.2701241672039032, + "learning_rate": 0.00016310295602442384, + "loss": 0.0656, + "step": 12064 + }, + { + "epoch": 0.2828226000222695, + "grad_norm": 0.42001309990882874, + "learning_rate": 0.0001630972414915727, + "loss": 0.0787, + "step": 12065 + }, + { + "epoch": 0.2828460415970745, + "grad_norm": 0.7100896835327148, + "learning_rate": 0.00016309152661635415, + "loss": 0.6227, + "step": 12066 + }, + { + "epoch": 0.2828694831718795, + "grad_norm": 0.4692246615886688, + "learning_rate": 0.00016308581139879918, + "loss": 0.0817, + "step": 12067 + }, + { + "epoch": 0.2828929247466845, + "grad_norm": 0.6524267792701721, + "learning_rate": 0.00016308009583893885, + "loss": 0.1367, + "step": 12068 + }, + { + "epoch": 0.2829163663214895, + "grad_norm": 0.4060733914375305, + "learning_rate": 0.00016307437993680414, + "loss": 0.0556, + "step": 12069 + }, + { + "epoch": 0.28293980789629447, + "grad_norm": 0.09079059958457947, + "learning_rate": 0.0001630686636924261, + "loss": 0.0142, + "step": 12070 + }, + { + "epoch": 0.28296324947109947, + "grad_norm": 0.35822299122810364, + "learning_rate": 0.00016306294710583568, + "loss": 0.0946, + "step": 12071 + }, + { + "epoch": 0.28298669104590446, + "grad_norm": 0.5404653549194336, + "learning_rate": 0.0001630572301770639, + "loss": 0.0811, + "step": 12072 + }, + { + "epoch": 0.28301013262070945, + "grad_norm": 0.5846880078315735, + "learning_rate": 0.0001630515129061419, + "loss": 0.1848, + "step": 12073 + }, + { + "epoch": 0.28303357419551445, + "grad_norm": 0.24331021308898926, + "learning_rate": 0.00016304579529310056, + "loss": 0.0659, + "step": 12074 + }, + { + "epoch": 0.28305701577031944, + "grad_norm": 0.4383707344532013, + "learning_rate": 0.00016304007733797095, + "loss": 0.0961, + "step": 12075 + }, + { + "epoch": 0.28308045734512444, + "grad_norm": 0.09783904999494553, + "learning_rate": 0.00016303435904078416, + "loss": 0.0223, + "step": 12076 + }, + { + "epoch": 0.28310389891992943, + "grad_norm": 0.6741527318954468, + "learning_rate": 0.00016302864040157114, + "loss": 0.1421, + "step": 12077 + }, + { + "epoch": 0.2831273404947344, + "grad_norm": 0.4029207229614258, + "learning_rate": 0.00016302292142036294, + "loss": 0.0877, + "step": 12078 + }, + { + "epoch": 0.2831507820695394, + "grad_norm": 0.6255443692207336, + "learning_rate": 0.00016301720209719063, + "loss": 0.1692, + "step": 12079 + }, + { + "epoch": 0.2831742236443444, + "grad_norm": 0.23700793087482452, + "learning_rate": 0.00016301148243208515, + "loss": 0.0779, + "step": 12080 + }, + { + "epoch": 0.2831976652191494, + "grad_norm": 0.7630571126937866, + "learning_rate": 0.00016300576242507764, + "loss": 0.1565, + "step": 12081 + }, + { + "epoch": 0.2832211067939544, + "grad_norm": 0.5083204507827759, + "learning_rate": 0.0001630000420761991, + "loss": 0.108, + "step": 12082 + }, + { + "epoch": 0.2832445483687594, + "grad_norm": 0.5312826633453369, + "learning_rate": 0.00016299432138548053, + "loss": 0.1252, + "step": 12083 + }, + { + "epoch": 0.2832679899435644, + "grad_norm": 0.20838232338428497, + "learning_rate": 0.000162988600352953, + "loss": 0.0224, + "step": 12084 + }, + { + "epoch": 0.2832914315183694, + "grad_norm": 0.5635868310928345, + "learning_rate": 0.00016298287897864756, + "loss": 0.0991, + "step": 12085 + }, + { + "epoch": 0.2833148730931744, + "grad_norm": 0.36159032583236694, + "learning_rate": 0.00016297715726259525, + "loss": 0.0819, + "step": 12086 + }, + { + "epoch": 0.2833383146679794, + "grad_norm": 0.501473069190979, + "learning_rate": 0.00016297143520482713, + "loss": 0.0909, + "step": 12087 + }, + { + "epoch": 0.28336175624278437, + "grad_norm": 0.27112284302711487, + "learning_rate": 0.00016296571280537424, + "loss": 0.0393, + "step": 12088 + }, + { + "epoch": 0.28338519781758936, + "grad_norm": 0.40093377232551575, + "learning_rate": 0.00016295999006426759, + "loss": 0.0847, + "step": 12089 + }, + { + "epoch": 0.28340863939239436, + "grad_norm": 0.20526117086410522, + "learning_rate": 0.00016295426698153826, + "loss": 0.0346, + "step": 12090 + }, + { + "epoch": 0.28343208096719935, + "grad_norm": 0.10684511065483093, + "learning_rate": 0.00016294854355721733, + "loss": 0.024, + "step": 12091 + }, + { + "epoch": 0.28345552254200435, + "grad_norm": 0.35626664757728577, + "learning_rate": 0.00016294281979133582, + "loss": 0.0794, + "step": 12092 + }, + { + "epoch": 0.28347896411680934, + "grad_norm": 0.7080421447753906, + "learning_rate": 0.0001629370956839248, + "loss": 0.2057, + "step": 12093 + }, + { + "epoch": 0.2835024056916144, + "grad_norm": 0.7089306712150574, + "learning_rate": 0.00016293137123501533, + "loss": 0.7714, + "step": 12094 + }, + { + "epoch": 0.2835258472664194, + "grad_norm": 0.4562118947505951, + "learning_rate": 0.00016292564644463848, + "loss": 0.1416, + "step": 12095 + }, + { + "epoch": 0.2835492888412244, + "grad_norm": 0.3589307367801666, + "learning_rate": 0.0001629199213128253, + "loss": 0.0863, + "step": 12096 + }, + { + "epoch": 0.2835727304160294, + "grad_norm": 0.43533435463905334, + "learning_rate": 0.00016291419583960685, + "loss": 0.0791, + "step": 12097 + }, + { + "epoch": 0.28359617199083437, + "grad_norm": 0.5624638795852661, + "learning_rate": 0.00016290847002501418, + "loss": 0.0711, + "step": 12098 + }, + { + "epoch": 0.28361961356563936, + "grad_norm": 0.5714187026023865, + "learning_rate": 0.00016290274386907843, + "loss": 0.1276, + "step": 12099 + }, + { + "epoch": 0.28364305514044436, + "grad_norm": 0.6136371493339539, + "learning_rate": 0.0001628970173718306, + "loss": 0.6212, + "step": 12100 + }, + { + "epoch": 0.28366649671524935, + "grad_norm": 0.21055251359939575, + "learning_rate": 0.0001628912905333018, + "loss": 0.0469, + "step": 12101 + }, + { + "epoch": 0.28368993829005434, + "grad_norm": 0.3353757858276367, + "learning_rate": 0.00016288556335352307, + "loss": 0.0878, + "step": 12102 + }, + { + "epoch": 0.28371337986485934, + "grad_norm": 0.3767244517803192, + "learning_rate": 0.00016287983583252553, + "loss": 0.3345, + "step": 12103 + }, + { + "epoch": 0.28373682143966433, + "grad_norm": 0.7404871582984924, + "learning_rate": 0.0001628741079703402, + "loss": 0.1118, + "step": 12104 + }, + { + "epoch": 0.2837602630144693, + "grad_norm": 1.1322391033172607, + "learning_rate": 0.00016286837976699822, + "loss": 0.1749, + "step": 12105 + }, + { + "epoch": 0.2837837045892743, + "grad_norm": 0.17155881226062775, + "learning_rate": 0.00016286265122253063, + "loss": 0.0415, + "step": 12106 + }, + { + "epoch": 0.2838071461640793, + "grad_norm": 0.48084813356399536, + "learning_rate": 0.00016285692233696853, + "loss": 0.1418, + "step": 12107 + }, + { + "epoch": 0.2838305877388843, + "grad_norm": 0.12201489508152008, + "learning_rate": 0.000162851193110343, + "loss": 0.0244, + "step": 12108 + }, + { + "epoch": 0.2838540293136893, + "grad_norm": 0.629206120967865, + "learning_rate": 0.00016284546354268512, + "loss": 0.1452, + "step": 12109 + }, + { + "epoch": 0.2838774708884943, + "grad_norm": 0.5589656829833984, + "learning_rate": 0.000162839733634026, + "loss": 0.1, + "step": 12110 + }, + { + "epoch": 0.2839009124632993, + "grad_norm": 0.19953450560569763, + "learning_rate": 0.0001628340033843967, + "loss": 0.0258, + "step": 12111 + }, + { + "epoch": 0.2839243540381043, + "grad_norm": 0.3023931682109833, + "learning_rate": 0.00016282827279382836, + "loss": 0.0767, + "step": 12112 + }, + { + "epoch": 0.2839477956129093, + "grad_norm": 0.26231321692466736, + "learning_rate": 0.00016282254186235205, + "loss": 0.0722, + "step": 12113 + }, + { + "epoch": 0.2839712371877143, + "grad_norm": 0.45661914348602295, + "learning_rate": 0.00016281681058999886, + "loss": 0.1509, + "step": 12114 + }, + { + "epoch": 0.28399467876251927, + "grad_norm": 0.5509002208709717, + "learning_rate": 0.00016281107897679986, + "loss": 0.1886, + "step": 12115 + }, + { + "epoch": 0.28401812033732426, + "grad_norm": 0.5228161215782166, + "learning_rate": 0.00016280534702278618, + "loss": 0.201, + "step": 12116 + }, + { + "epoch": 0.28404156191212926, + "grad_norm": 0.36062049865722656, + "learning_rate": 0.00016279961472798893, + "loss": 0.0855, + "step": 12117 + }, + { + "epoch": 0.28406500348693425, + "grad_norm": 0.4260685443878174, + "learning_rate": 0.0001627938820924392, + "loss": 0.0772, + "step": 12118 + }, + { + "epoch": 0.28408844506173925, + "grad_norm": 0.5599150061607361, + "learning_rate": 0.00016278814911616814, + "loss": 0.0573, + "step": 12119 + }, + { + "epoch": 0.28411188663654424, + "grad_norm": 0.37051305174827576, + "learning_rate": 0.0001627824157992068, + "loss": 0.0856, + "step": 12120 + }, + { + "epoch": 0.28413532821134924, + "grad_norm": 0.35366085171699524, + "learning_rate": 0.00016277668214158627, + "loss": 0.0787, + "step": 12121 + }, + { + "epoch": 0.28415876978615423, + "grad_norm": 0.4819912314414978, + "learning_rate": 0.0001627709481433377, + "loss": 0.1751, + "step": 12122 + }, + { + "epoch": 0.2841822113609592, + "grad_norm": 0.3763159215450287, + "learning_rate": 0.00016276521380449222, + "loss": 0.049, + "step": 12123 + }, + { + "epoch": 0.2842056529357642, + "grad_norm": 0.22745539247989655, + "learning_rate": 0.00016275947912508095, + "loss": 0.044, + "step": 12124 + }, + { + "epoch": 0.2842290945105692, + "grad_norm": 0.4917092025279999, + "learning_rate": 0.00016275374410513495, + "loss": 0.1329, + "step": 12125 + }, + { + "epoch": 0.2842525360853742, + "grad_norm": 0.9647659659385681, + "learning_rate": 0.00016274800874468536, + "loss": 0.1367, + "step": 12126 + }, + { + "epoch": 0.2842759776601792, + "grad_norm": 0.14656774699687958, + "learning_rate": 0.00016274227304376333, + "loss": 0.0414, + "step": 12127 + }, + { + "epoch": 0.2842994192349842, + "grad_norm": 0.30851536989212036, + "learning_rate": 0.00016273653700239998, + "loss": 0.0569, + "step": 12128 + }, + { + "epoch": 0.2843228608097892, + "grad_norm": 0.06833087652921677, + "learning_rate": 0.0001627308006206264, + "loss": 0.0115, + "step": 12129 + }, + { + "epoch": 0.2843463023845942, + "grad_norm": 0.4246821403503418, + "learning_rate": 0.00016272506389847374, + "loss": 0.0833, + "step": 12130 + }, + { + "epoch": 0.2843697439593992, + "grad_norm": 0.9480371475219727, + "learning_rate": 0.00016271932683597308, + "loss": 0.1466, + "step": 12131 + }, + { + "epoch": 0.2843931855342042, + "grad_norm": 0.35484856367111206, + "learning_rate": 0.00016271358943315562, + "loss": 0.0837, + "step": 12132 + }, + { + "epoch": 0.28441662710900917, + "grad_norm": 0.6096807718276978, + "learning_rate": 0.00016270785169005247, + "loss": 0.1125, + "step": 12133 + }, + { + "epoch": 0.28444006868381416, + "grad_norm": 0.6313192248344421, + "learning_rate": 0.00016270211360669475, + "loss": 0.1929, + "step": 12134 + }, + { + "epoch": 0.28446351025861916, + "grad_norm": 0.6828420162200928, + "learning_rate": 0.00016269637518311357, + "loss": 0.08, + "step": 12135 + }, + { + "epoch": 0.28448695183342415, + "grad_norm": 0.7396661639213562, + "learning_rate": 0.00016269063641934014, + "loss": 0.255, + "step": 12136 + }, + { + "epoch": 0.28451039340822915, + "grad_norm": 0.3398934006690979, + "learning_rate": 0.0001626848973154055, + "loss": 0.1149, + "step": 12137 + }, + { + "epoch": 0.28453383498303414, + "grad_norm": 0.9112058281898499, + "learning_rate": 0.00016267915787134088, + "loss": 0.1601, + "step": 12138 + }, + { + "epoch": 0.28455727655783913, + "grad_norm": 0.6156110763549805, + "learning_rate": 0.0001626734180871774, + "loss": 0.1636, + "step": 12139 + }, + { + "epoch": 0.28458071813264413, + "grad_norm": 0.19116972386837006, + "learning_rate": 0.00016266767796294617, + "loss": 0.0609, + "step": 12140 + }, + { + "epoch": 0.2846041597074491, + "grad_norm": 0.42299243807792664, + "learning_rate": 0.00016266193749867833, + "loss": 0.0815, + "step": 12141 + }, + { + "epoch": 0.2846276012822541, + "grad_norm": 0.5914701223373413, + "learning_rate": 0.00016265619669440508, + "loss": 0.0743, + "step": 12142 + }, + { + "epoch": 0.2846510428570591, + "grad_norm": 0.9481020569801331, + "learning_rate": 0.00016265045555015757, + "loss": 0.2218, + "step": 12143 + }, + { + "epoch": 0.2846744844318641, + "grad_norm": 0.29291391372680664, + "learning_rate": 0.0001626447140659669, + "loss": 0.0806, + "step": 12144 + }, + { + "epoch": 0.28469792600666916, + "grad_norm": 0.5367618203163147, + "learning_rate": 0.00016263897224186427, + "loss": 0.1829, + "step": 12145 + }, + { + "epoch": 0.28472136758147415, + "grad_norm": 0.3413830101490021, + "learning_rate": 0.0001626332300778808, + "loss": 0.0934, + "step": 12146 + }, + { + "epoch": 0.28474480915627914, + "grad_norm": 0.5909628868103027, + "learning_rate": 0.00016262748757404764, + "loss": 0.6898, + "step": 12147 + }, + { + "epoch": 0.28476825073108414, + "grad_norm": 0.41545435786247253, + "learning_rate": 0.00016262174473039603, + "loss": 0.0764, + "step": 12148 + }, + { + "epoch": 0.28479169230588913, + "grad_norm": 0.5301070809364319, + "learning_rate": 0.000162616001546957, + "loss": 0.1164, + "step": 12149 + }, + { + "epoch": 0.2848151338806941, + "grad_norm": 0.13086821138858795, + "learning_rate": 0.00016261025802376185, + "loss": 0.0186, + "step": 12150 + }, + { + "epoch": 0.2848385754554991, + "grad_norm": 0.6231837868690491, + "learning_rate": 0.00016260451416084165, + "loss": 0.1109, + "step": 12151 + }, + { + "epoch": 0.2848620170303041, + "grad_norm": 0.293262243270874, + "learning_rate": 0.0001625987699582276, + "loss": 0.0835, + "step": 12152 + }, + { + "epoch": 0.2848854586051091, + "grad_norm": 0.27562764286994934, + "learning_rate": 0.00016259302541595086, + "loss": 0.0675, + "step": 12153 + }, + { + "epoch": 0.2849089001799141, + "grad_norm": 0.47595179080963135, + "learning_rate": 0.0001625872805340426, + "loss": 0.1202, + "step": 12154 + }, + { + "epoch": 0.2849323417547191, + "grad_norm": 0.20345495641231537, + "learning_rate": 0.00016258153531253399, + "loss": 0.0542, + "step": 12155 + }, + { + "epoch": 0.2849557833295241, + "grad_norm": 0.4235742390155792, + "learning_rate": 0.00016257578975145624, + "loss": 0.0821, + "step": 12156 + }, + { + "epoch": 0.2849792249043291, + "grad_norm": 0.5651290416717529, + "learning_rate": 0.00016257004385084048, + "loss": 0.1484, + "step": 12157 + }, + { + "epoch": 0.2850026664791341, + "grad_norm": 0.5630178451538086, + "learning_rate": 0.0001625642976107179, + "loss": 0.0764, + "step": 12158 + }, + { + "epoch": 0.2850261080539391, + "grad_norm": 0.7166889309883118, + "learning_rate": 0.00016255855103111964, + "loss": 0.1798, + "step": 12159 + }, + { + "epoch": 0.28504954962874407, + "grad_norm": 0.7869316339492798, + "learning_rate": 0.000162552804112077, + "loss": 0.1489, + "step": 12160 + }, + { + "epoch": 0.28507299120354906, + "grad_norm": 0.536260724067688, + "learning_rate": 0.00016254705685362104, + "loss": 0.1359, + "step": 12161 + }, + { + "epoch": 0.28509643277835406, + "grad_norm": 0.507412314414978, + "learning_rate": 0.00016254130925578302, + "loss": 0.1017, + "step": 12162 + }, + { + "epoch": 0.28511987435315905, + "grad_norm": 0.20714642107486725, + "learning_rate": 0.00016253556131859406, + "loss": 0.035, + "step": 12163 + }, + { + "epoch": 0.28514331592796405, + "grad_norm": 0.6062098741531372, + "learning_rate": 0.00016252981304208541, + "loss": 0.0954, + "step": 12164 + }, + { + "epoch": 0.28516675750276904, + "grad_norm": 0.20604762434959412, + "learning_rate": 0.00016252406442628824, + "loss": 0.0456, + "step": 12165 + }, + { + "epoch": 0.28519019907757404, + "grad_norm": 1.1134904623031616, + "learning_rate": 0.0001625183154712337, + "loss": 0.2542, + "step": 12166 + }, + { + "epoch": 0.28521364065237903, + "grad_norm": 0.6718322038650513, + "learning_rate": 0.0001625125661769531, + "loss": 0.1821, + "step": 12167 + }, + { + "epoch": 0.285237082227184, + "grad_norm": 0.1783459633588791, + "learning_rate": 0.0001625068165434775, + "loss": 0.0271, + "step": 12168 + }, + { + "epoch": 0.285260523801989, + "grad_norm": 0.4029799997806549, + "learning_rate": 0.00016250106657083815, + "loss": 0.1015, + "step": 12169 + }, + { + "epoch": 0.285283965376794, + "grad_norm": 0.5796549320220947, + "learning_rate": 0.0001624953162590663, + "loss": 0.0942, + "step": 12170 + }, + { + "epoch": 0.285307406951599, + "grad_norm": 0.6164954304695129, + "learning_rate": 0.00016248956560819305, + "loss": 0.0564, + "step": 12171 + }, + { + "epoch": 0.285330848526404, + "grad_norm": 0.5178478360176086, + "learning_rate": 0.0001624838146182497, + "loss": 0.1595, + "step": 12172 + }, + { + "epoch": 0.285354290101209, + "grad_norm": 0.6190565228462219, + "learning_rate": 0.0001624780632892674, + "loss": 0.1423, + "step": 12173 + }, + { + "epoch": 0.285377731676014, + "grad_norm": 0.12147079408168793, + "learning_rate": 0.00016247231162127735, + "loss": 0.0196, + "step": 12174 + }, + { + "epoch": 0.285401173250819, + "grad_norm": 0.42928433418273926, + "learning_rate": 0.00016246655961431084, + "loss": 0.0942, + "step": 12175 + }, + { + "epoch": 0.285424614825624, + "grad_norm": 0.5871038436889648, + "learning_rate": 0.00016246080726839898, + "loss": 0.7596, + "step": 12176 + }, + { + "epoch": 0.285448056400429, + "grad_norm": 0.41912949085235596, + "learning_rate": 0.00016245505458357303, + "loss": 0.067, + "step": 12177 + }, + { + "epoch": 0.28547149797523397, + "grad_norm": 0.7553650140762329, + "learning_rate": 0.00016244930155986419, + "loss": 0.1882, + "step": 12178 + }, + { + "epoch": 0.28549493955003896, + "grad_norm": 0.21892355382442474, + "learning_rate": 0.00016244354819730373, + "loss": 0.0543, + "step": 12179 + }, + { + "epoch": 0.28551838112484396, + "grad_norm": 0.6526315212249756, + "learning_rate": 0.00016243779449592275, + "loss": 0.1681, + "step": 12180 + }, + { + "epoch": 0.28554182269964895, + "grad_norm": 0.5223435759544373, + "learning_rate": 0.00016243204045575258, + "loss": 0.1025, + "step": 12181 + }, + { + "epoch": 0.28556526427445394, + "grad_norm": 0.5530139803886414, + "learning_rate": 0.0001624262860768244, + "loss": 0.1155, + "step": 12182 + }, + { + "epoch": 0.28558870584925894, + "grad_norm": 0.5354264974594116, + "learning_rate": 0.00016242053135916943, + "loss": 0.1288, + "step": 12183 + }, + { + "epoch": 0.28561214742406393, + "grad_norm": 0.1849498450756073, + "learning_rate": 0.00016241477630281892, + "loss": 0.0414, + "step": 12184 + }, + { + "epoch": 0.28563558899886893, + "grad_norm": 0.8771672248840332, + "learning_rate": 0.00016240902090780405, + "loss": 0.176, + "step": 12185 + }, + { + "epoch": 0.2856590305736739, + "grad_norm": 0.11800183355808258, + "learning_rate": 0.00016240326517415608, + "loss": 0.0187, + "step": 12186 + }, + { + "epoch": 0.2856824721484789, + "grad_norm": 0.2641812562942505, + "learning_rate": 0.00016239750910190627, + "loss": 0.0723, + "step": 12187 + }, + { + "epoch": 0.2857059137232839, + "grad_norm": 0.36431142687797546, + "learning_rate": 0.00016239175269108582, + "loss": 0.0674, + "step": 12188 + }, + { + "epoch": 0.2857293552980889, + "grad_norm": 0.4444428086280823, + "learning_rate": 0.00016238599594172595, + "loss": 0.4479, + "step": 12189 + }, + { + "epoch": 0.2857527968728939, + "grad_norm": 0.5727058053016663, + "learning_rate": 0.0001623802388538579, + "loss": 0.1439, + "step": 12190 + }, + { + "epoch": 0.2857762384476989, + "grad_norm": 0.5638153553009033, + "learning_rate": 0.00016237448142751291, + "loss": 0.6308, + "step": 12191 + }, + { + "epoch": 0.2857996800225039, + "grad_norm": 0.5253340601921082, + "learning_rate": 0.00016236872366272228, + "loss": 0.103, + "step": 12192 + }, + { + "epoch": 0.2858231215973089, + "grad_norm": 0.6200065016746521, + "learning_rate": 0.00016236296555951718, + "loss": 0.2172, + "step": 12193 + }, + { + "epoch": 0.2858465631721139, + "grad_norm": 0.4186246693134308, + "learning_rate": 0.00016235720711792886, + "loss": 0.0722, + "step": 12194 + }, + { + "epoch": 0.28587000474691887, + "grad_norm": 0.5894129872322083, + "learning_rate": 0.00016235144833798857, + "loss": 0.1698, + "step": 12195 + }, + { + "epoch": 0.2858934463217239, + "grad_norm": 0.6512038707733154, + "learning_rate": 0.0001623456892197276, + "loss": 0.5315, + "step": 12196 + }, + { + "epoch": 0.2859168878965289, + "grad_norm": 0.19113077223300934, + "learning_rate": 0.00016233992976317715, + "loss": 0.0388, + "step": 12197 + }, + { + "epoch": 0.2859403294713339, + "grad_norm": 0.6048861145973206, + "learning_rate": 0.0001623341699683685, + "loss": 0.1197, + "step": 12198 + }, + { + "epoch": 0.2859637710461389, + "grad_norm": 0.5845904350280762, + "learning_rate": 0.00016232840983533287, + "loss": 0.1027, + "step": 12199 + }, + { + "epoch": 0.2859872126209439, + "grad_norm": 0.5810973048210144, + "learning_rate": 0.00016232264936410156, + "loss": 0.2088, + "step": 12200 + }, + { + "epoch": 0.2860106541957489, + "grad_norm": 0.2110857516527176, + "learning_rate": 0.0001623168885547058, + "loss": 0.0366, + "step": 12201 + }, + { + "epoch": 0.2860340957705539, + "grad_norm": 0.5235446095466614, + "learning_rate": 0.00016231112740717682, + "loss": 0.0905, + "step": 12202 + }, + { + "epoch": 0.2860575373453589, + "grad_norm": 0.36626750230789185, + "learning_rate": 0.00016230536592154592, + "loss": 0.0704, + "step": 12203 + }, + { + "epoch": 0.2860809789201639, + "grad_norm": 0.5630609393119812, + "learning_rate": 0.0001622996040978444, + "loss": 0.0703, + "step": 12204 + }, + { + "epoch": 0.28610442049496887, + "grad_norm": 0.13477998971939087, + "learning_rate": 0.00016229384193610345, + "loss": 0.039, + "step": 12205 + }, + { + "epoch": 0.28612786206977386, + "grad_norm": 0.5545873045921326, + "learning_rate": 0.00016228807943635435, + "loss": 0.1687, + "step": 12206 + }, + { + "epoch": 0.28615130364457886, + "grad_norm": 0.7748442888259888, + "learning_rate": 0.0001622823165986284, + "loss": 0.169, + "step": 12207 + }, + { + "epoch": 0.28617474521938385, + "grad_norm": 1.0423568487167358, + "learning_rate": 0.0001622765534229568, + "loss": 0.0975, + "step": 12208 + }, + { + "epoch": 0.28619818679418885, + "grad_norm": 0.5720610022544861, + "learning_rate": 0.0001622707899093709, + "loss": 0.0697, + "step": 12209 + }, + { + "epoch": 0.28622162836899384, + "grad_norm": 0.1596565991640091, + "learning_rate": 0.00016226502605790194, + "loss": 0.0308, + "step": 12210 + }, + { + "epoch": 0.28624506994379884, + "grad_norm": 0.33259469270706177, + "learning_rate": 0.0001622592618685812, + "loss": 0.0815, + "step": 12211 + }, + { + "epoch": 0.28626851151860383, + "grad_norm": 0.2068132609128952, + "learning_rate": 0.00016225349734143998, + "loss": 0.0341, + "step": 12212 + }, + { + "epoch": 0.2862919530934088, + "grad_norm": 0.4155396819114685, + "learning_rate": 0.0001622477324765095, + "loss": 0.0848, + "step": 12213 + }, + { + "epoch": 0.2863153946682138, + "grad_norm": 0.5045011639595032, + "learning_rate": 0.00016224196727382108, + "loss": 0.4579, + "step": 12214 + }, + { + "epoch": 0.2863388362430188, + "grad_norm": 0.6031787991523743, + "learning_rate": 0.00016223620173340598, + "loss": 0.1275, + "step": 12215 + }, + { + "epoch": 0.2863622778178238, + "grad_norm": 0.8132888674736023, + "learning_rate": 0.00016223043585529556, + "loss": 0.0985, + "step": 12216 + }, + { + "epoch": 0.2863857193926288, + "grad_norm": 0.6800323128700256, + "learning_rate": 0.00016222466963952098, + "loss": 0.0762, + "step": 12217 + }, + { + "epoch": 0.2864091609674338, + "grad_norm": 0.5428168773651123, + "learning_rate": 0.00016221890308611362, + "loss": 0.1124, + "step": 12218 + }, + { + "epoch": 0.2864326025422388, + "grad_norm": 0.6867180466651917, + "learning_rate": 0.00016221313619510475, + "loss": 0.1874, + "step": 12219 + }, + { + "epoch": 0.2864560441170438, + "grad_norm": 0.5150183439254761, + "learning_rate": 0.00016220736896652564, + "loss": 0.0808, + "step": 12220 + }, + { + "epoch": 0.2864794856918488, + "grad_norm": 0.8570798635482788, + "learning_rate": 0.0001622016014004076, + "loss": 0.1418, + "step": 12221 + }, + { + "epoch": 0.2865029272666538, + "grad_norm": 0.42416059970855713, + "learning_rate": 0.00016219583349678192, + "loss": 0.1012, + "step": 12222 + }, + { + "epoch": 0.28652636884145877, + "grad_norm": 0.12636694312095642, + "learning_rate": 0.0001621900652556799, + "loss": 0.0215, + "step": 12223 + }, + { + "epoch": 0.28654981041626376, + "grad_norm": 0.516618013381958, + "learning_rate": 0.00016218429667713284, + "loss": 0.1612, + "step": 12224 + }, + { + "epoch": 0.28657325199106876, + "grad_norm": 0.7975360155105591, + "learning_rate": 0.00016217852776117205, + "loss": 0.9704, + "step": 12225 + }, + { + "epoch": 0.28659669356587375, + "grad_norm": 0.13976477086544037, + "learning_rate": 0.00016217275850782882, + "loss": 0.0264, + "step": 12226 + }, + { + "epoch": 0.28662013514067874, + "grad_norm": 0.5521924495697021, + "learning_rate": 0.0001621669889171344, + "loss": 0.1632, + "step": 12227 + }, + { + "epoch": 0.28664357671548374, + "grad_norm": 0.5774816870689392, + "learning_rate": 0.0001621612189891202, + "loss": 0.1416, + "step": 12228 + }, + { + "epoch": 0.28666701829028873, + "grad_norm": 0.5851475596427917, + "learning_rate": 0.00016215544872381747, + "loss": 0.1381, + "step": 12229 + }, + { + "epoch": 0.2866904598650937, + "grad_norm": 0.5189569592475891, + "learning_rate": 0.00016214967812125752, + "loss": 0.8935, + "step": 12230 + }, + { + "epoch": 0.2867139014398987, + "grad_norm": 0.6826216578483582, + "learning_rate": 0.0001621439071814717, + "loss": 0.1245, + "step": 12231 + }, + { + "epoch": 0.2867373430147037, + "grad_norm": 0.7929061055183411, + "learning_rate": 0.00016213813590449126, + "loss": 0.087, + "step": 12232 + }, + { + "epoch": 0.2867607845895087, + "grad_norm": 0.2859933376312256, + "learning_rate": 0.00016213236429034753, + "loss": 0.0671, + "step": 12233 + }, + { + "epoch": 0.2867842261643137, + "grad_norm": 0.48132461309432983, + "learning_rate": 0.00016212659233907187, + "loss": 0.1149, + "step": 12234 + }, + { + "epoch": 0.2868076677391187, + "grad_norm": 0.5992345213890076, + "learning_rate": 0.0001621208200506956, + "loss": 0.117, + "step": 12235 + }, + { + "epoch": 0.2868311093139237, + "grad_norm": 0.7539579272270203, + "learning_rate": 0.00016211504742524995, + "loss": 0.1787, + "step": 12236 + }, + { + "epoch": 0.2868545508887287, + "grad_norm": 0.4338873624801636, + "learning_rate": 0.00016210927446276636, + "loss": 0.1524, + "step": 12237 + }, + { + "epoch": 0.2868779924635337, + "grad_norm": 0.20520344376564026, + "learning_rate": 0.00016210350116327606, + "loss": 0.0556, + "step": 12238 + }, + { + "epoch": 0.2869014340383387, + "grad_norm": 0.6429004073143005, + "learning_rate": 0.00016209772752681046, + "loss": 0.1222, + "step": 12239 + }, + { + "epoch": 0.28692487561314367, + "grad_norm": 0.5691424012184143, + "learning_rate": 0.00016209195355340082, + "loss": 0.1227, + "step": 12240 + }, + { + "epoch": 0.28694831718794866, + "grad_norm": 0.2154150903224945, + "learning_rate": 0.00016208617924307853, + "loss": 0.0536, + "step": 12241 + }, + { + "epoch": 0.28697175876275366, + "grad_norm": 0.15040868520736694, + "learning_rate": 0.0001620804045958748, + "loss": 0.026, + "step": 12242 + }, + { + "epoch": 0.28699520033755865, + "grad_norm": 0.6003128886222839, + "learning_rate": 0.00016207462961182113, + "loss": 0.1059, + "step": 12243 + }, + { + "epoch": 0.28701864191236365, + "grad_norm": 0.580696702003479, + "learning_rate": 0.00016206885429094874, + "loss": 0.1166, + "step": 12244 + }, + { + "epoch": 0.28704208348716864, + "grad_norm": 0.36719000339508057, + "learning_rate": 0.000162063078633289, + "loss": 0.0697, + "step": 12245 + }, + { + "epoch": 0.28706552506197364, + "grad_norm": 0.4760476350784302, + "learning_rate": 0.00016205730263887326, + "loss": 0.1315, + "step": 12246 + }, + { + "epoch": 0.2870889666367787, + "grad_norm": 0.49059876799583435, + "learning_rate": 0.00016205152630773285, + "loss": 0.1131, + "step": 12247 + }, + { + "epoch": 0.2871124082115837, + "grad_norm": 0.5246067047119141, + "learning_rate": 0.00016204574963989913, + "loss": 0.0655, + "step": 12248 + }, + { + "epoch": 0.2871358497863887, + "grad_norm": 0.580705463886261, + "learning_rate": 0.00016203997263540342, + "loss": 0.0917, + "step": 12249 + }, + { + "epoch": 0.28715929136119367, + "grad_norm": 0.6358885765075684, + "learning_rate": 0.00016203419529427703, + "loss": 0.1526, + "step": 12250 + }, + { + "epoch": 0.28718273293599866, + "grad_norm": 0.5493634343147278, + "learning_rate": 0.0001620284176165514, + "loss": 0.141, + "step": 12251 + }, + { + "epoch": 0.28720617451080366, + "grad_norm": 0.48695552349090576, + "learning_rate": 0.0001620226396022578, + "loss": 0.1401, + "step": 12252 + }, + { + "epoch": 0.28722961608560865, + "grad_norm": 1.033470869064331, + "learning_rate": 0.00016201686125142767, + "loss": 0.1299, + "step": 12253 + }, + { + "epoch": 0.28725305766041365, + "grad_norm": 0.5252133011817932, + "learning_rate": 0.00016201108256409227, + "loss": 0.0913, + "step": 12254 + }, + { + "epoch": 0.28727649923521864, + "grad_norm": 0.5609156489372253, + "learning_rate": 0.00016200530354028297, + "loss": 0.1307, + "step": 12255 + }, + { + "epoch": 0.28729994081002364, + "grad_norm": 0.1808079183101654, + "learning_rate": 0.00016199952418003116, + "loss": 0.0208, + "step": 12256 + }, + { + "epoch": 0.28732338238482863, + "grad_norm": 0.4395678639411926, + "learning_rate": 0.00016199374448336822, + "loss": 0.4104, + "step": 12257 + }, + { + "epoch": 0.2873468239596336, + "grad_norm": 0.7974303960800171, + "learning_rate": 0.00016198796445032545, + "loss": 0.1424, + "step": 12258 + }, + { + "epoch": 0.2873702655344386, + "grad_norm": 0.8112886548042297, + "learning_rate": 0.00016198218408093427, + "loss": 0.5854, + "step": 12259 + }, + { + "epoch": 0.2873937071092436, + "grad_norm": 0.8814489841461182, + "learning_rate": 0.000161976403375226, + "loss": 0.1594, + "step": 12260 + }, + { + "epoch": 0.2874171486840486, + "grad_norm": 0.7466394305229187, + "learning_rate": 0.000161970622333232, + "loss": 0.2234, + "step": 12261 + }, + { + "epoch": 0.2874405902588536, + "grad_norm": 0.1684398055076599, + "learning_rate": 0.00016196484095498367, + "loss": 0.0345, + "step": 12262 + }, + { + "epoch": 0.2874640318336586, + "grad_norm": 0.5451651811599731, + "learning_rate": 0.00016195905924051238, + "loss": 0.1012, + "step": 12263 + }, + { + "epoch": 0.2874874734084636, + "grad_norm": 0.3444954752922058, + "learning_rate": 0.0001619532771898495, + "loss": 0.0481, + "step": 12264 + }, + { + "epoch": 0.2875109149832686, + "grad_norm": 0.48707738518714905, + "learning_rate": 0.0001619474948030264, + "loss": 0.0811, + "step": 12265 + }, + { + "epoch": 0.2875343565580736, + "grad_norm": 0.3391001224517822, + "learning_rate": 0.00016194171208007446, + "loss": 0.0335, + "step": 12266 + }, + { + "epoch": 0.2875577981328786, + "grad_norm": 0.9134679436683655, + "learning_rate": 0.000161935929021025, + "loss": 0.2675, + "step": 12267 + }, + { + "epoch": 0.28758123970768357, + "grad_norm": 0.5355250835418701, + "learning_rate": 0.00016193014562590946, + "loss": 0.1107, + "step": 12268 + }, + { + "epoch": 0.28760468128248856, + "grad_norm": 0.5687113404273987, + "learning_rate": 0.00016192436189475923, + "loss": 0.09, + "step": 12269 + }, + { + "epoch": 0.28762812285729356, + "grad_norm": 0.11176492273807526, + "learning_rate": 0.00016191857782760569, + "loss": 0.021, + "step": 12270 + }, + { + "epoch": 0.28765156443209855, + "grad_norm": 0.3965948820114136, + "learning_rate": 0.00016191279342448017, + "loss": 0.0908, + "step": 12271 + }, + { + "epoch": 0.28767500600690354, + "grad_norm": 0.29436618089675903, + "learning_rate": 0.0001619070086854141, + "loss": 0.0266, + "step": 12272 + }, + { + "epoch": 0.28769844758170854, + "grad_norm": 0.4702930450439453, + "learning_rate": 0.00016190122361043892, + "loss": 0.1043, + "step": 12273 + }, + { + "epoch": 0.28772188915651353, + "grad_norm": 0.9322939515113831, + "learning_rate": 0.0001618954381995859, + "loss": 0.1336, + "step": 12274 + }, + { + "epoch": 0.2877453307313185, + "grad_norm": 1.221298336982727, + "learning_rate": 0.00016188965245288647, + "loss": 0.1685, + "step": 12275 + }, + { + "epoch": 0.2877687723061235, + "grad_norm": 0.8025607466697693, + "learning_rate": 0.00016188386637037207, + "loss": 0.1192, + "step": 12276 + }, + { + "epoch": 0.2877922138809285, + "grad_norm": 0.5456843972206116, + "learning_rate": 0.00016187807995207408, + "loss": 0.0949, + "step": 12277 + }, + { + "epoch": 0.2878156554557335, + "grad_norm": 0.08600813150405884, + "learning_rate": 0.00016187229319802388, + "loss": 0.0137, + "step": 12278 + }, + { + "epoch": 0.2878390970305385, + "grad_norm": 0.6909818053245544, + "learning_rate": 0.0001618665061082529, + "loss": 0.1718, + "step": 12279 + }, + { + "epoch": 0.2878625386053435, + "grad_norm": 0.5116749405860901, + "learning_rate": 0.0001618607186827925, + "loss": 0.1506, + "step": 12280 + }, + { + "epoch": 0.2878859801801485, + "grad_norm": 0.2638907730579376, + "learning_rate": 0.00016185493092167407, + "loss": 0.0766, + "step": 12281 + }, + { + "epoch": 0.2879094217549535, + "grad_norm": 0.7177711725234985, + "learning_rate": 0.0001618491428249291, + "loss": 0.0935, + "step": 12282 + }, + { + "epoch": 0.2879328633297585, + "grad_norm": 0.5603464245796204, + "learning_rate": 0.0001618433543925889, + "loss": 0.1752, + "step": 12283 + }, + { + "epoch": 0.2879563049045635, + "grad_norm": 0.3333669602870941, + "learning_rate": 0.00016183756562468494, + "loss": 0.0465, + "step": 12284 + }, + { + "epoch": 0.28797974647936847, + "grad_norm": 0.41987717151641846, + "learning_rate": 0.00016183177652124862, + "loss": 0.1418, + "step": 12285 + }, + { + "epoch": 0.28800318805417346, + "grad_norm": 0.3484654724597931, + "learning_rate": 0.0001618259870823113, + "loss": 0.1033, + "step": 12286 + }, + { + "epoch": 0.28802662962897846, + "grad_norm": 0.4906713366508484, + "learning_rate": 0.00016182019730790446, + "loss": 0.1179, + "step": 12287 + }, + { + "epoch": 0.28805007120378345, + "grad_norm": 0.6677531003952026, + "learning_rate": 0.00016181440719805948, + "loss": 0.1472, + "step": 12288 + }, + { + "epoch": 0.28807351277858845, + "grad_norm": 0.3664349615573883, + "learning_rate": 0.00016180861675280786, + "loss": 0.0698, + "step": 12289 + }, + { + "epoch": 0.28809695435339344, + "grad_norm": 0.10518015176057816, + "learning_rate": 0.00016180282597218088, + "loss": 0.0184, + "step": 12290 + }, + { + "epoch": 0.28812039592819844, + "grad_norm": 0.6685925722122192, + "learning_rate": 0.00016179703485621002, + "loss": 0.2182, + "step": 12291 + }, + { + "epoch": 0.28814383750300343, + "grad_norm": 0.3160770833492279, + "learning_rate": 0.00016179124340492674, + "loss": 0.0436, + "step": 12292 + }, + { + "epoch": 0.2881672790778084, + "grad_norm": 0.4273220896720886, + "learning_rate": 0.00016178545161836243, + "loss": 0.1441, + "step": 12293 + }, + { + "epoch": 0.2881907206526134, + "grad_norm": 0.3049888610839844, + "learning_rate": 0.00016177965949654851, + "loss": 0.0667, + "step": 12294 + }, + { + "epoch": 0.2882141622274184, + "grad_norm": 0.6416013836860657, + "learning_rate": 0.00016177386703951645, + "loss": 1.0541, + "step": 12295 + }, + { + "epoch": 0.2882376038022234, + "grad_norm": 0.4820390045642853, + "learning_rate": 0.00016176807424729764, + "loss": 0.1603, + "step": 12296 + }, + { + "epoch": 0.2882610453770284, + "grad_norm": 0.22445066273212433, + "learning_rate": 0.0001617622811199235, + "loss": 0.0374, + "step": 12297 + }, + { + "epoch": 0.28828448695183345, + "grad_norm": 0.6129814386367798, + "learning_rate": 0.0001617564876574255, + "loss": 0.1035, + "step": 12298 + }, + { + "epoch": 0.28830792852663845, + "grad_norm": 0.5033628344535828, + "learning_rate": 0.00016175069385983507, + "loss": 0.1464, + "step": 12299 + }, + { + "epoch": 0.28833137010144344, + "grad_norm": 0.5038412809371948, + "learning_rate": 0.00016174489972718364, + "loss": 0.078, + "step": 12300 + }, + { + "epoch": 0.28835481167624843, + "grad_norm": 0.12883460521697998, + "learning_rate": 0.00016173910525950263, + "loss": 0.0285, + "step": 12301 + }, + { + "epoch": 0.28837825325105343, + "grad_norm": 0.5444929003715515, + "learning_rate": 0.0001617333104568235, + "loss": 0.0607, + "step": 12302 + }, + { + "epoch": 0.2884016948258584, + "grad_norm": 0.3977859616279602, + "learning_rate": 0.0001617275153191777, + "loss": 0.0788, + "step": 12303 + }, + { + "epoch": 0.2884251364006634, + "grad_norm": 0.6327039003372192, + "learning_rate": 0.00016172171984659664, + "loss": 0.1779, + "step": 12304 + }, + { + "epoch": 0.2884485779754684, + "grad_norm": 0.185080885887146, + "learning_rate": 0.00016171592403911182, + "loss": 0.0135, + "step": 12305 + }, + { + "epoch": 0.2884720195502734, + "grad_norm": 0.8935672044754028, + "learning_rate": 0.00016171012789675468, + "loss": 0.1351, + "step": 12306 + }, + { + "epoch": 0.2884954611250784, + "grad_norm": 0.45831671357154846, + "learning_rate": 0.00016170433141955663, + "loss": 0.1168, + "step": 12307 + }, + { + "epoch": 0.2885189026998834, + "grad_norm": 0.34188565611839294, + "learning_rate": 0.00016169853460754914, + "loss": 0.075, + "step": 12308 + }, + { + "epoch": 0.2885423442746884, + "grad_norm": 0.5400647521018982, + "learning_rate": 0.00016169273746076366, + "loss": 0.0625, + "step": 12309 + }, + { + "epoch": 0.2885657858494934, + "grad_norm": 0.3041689693927765, + "learning_rate": 0.00016168693997923165, + "loss": 0.0594, + "step": 12310 + }, + { + "epoch": 0.2885892274242984, + "grad_norm": 0.64150071144104, + "learning_rate": 0.0001616811421629846, + "loss": 0.591, + "step": 12311 + }, + { + "epoch": 0.2886126689991034, + "grad_norm": 0.3717145323753357, + "learning_rate": 0.0001616753440120539, + "loss": 0.1024, + "step": 12312 + }, + { + "epoch": 0.28863611057390837, + "grad_norm": 0.5146780610084534, + "learning_rate": 0.00016166954552647108, + "loss": 0.0969, + "step": 12313 + }, + { + "epoch": 0.28865955214871336, + "grad_norm": 0.09055506438016891, + "learning_rate": 0.00016166374670626755, + "loss": 0.0152, + "step": 12314 + }, + { + "epoch": 0.28868299372351836, + "grad_norm": 0.7906628251075745, + "learning_rate": 0.0001616579475514748, + "loss": 0.1258, + "step": 12315 + }, + { + "epoch": 0.28870643529832335, + "grad_norm": 0.39225634932518005, + "learning_rate": 0.00016165214806212433, + "loss": 0.0461, + "step": 12316 + }, + { + "epoch": 0.28872987687312834, + "grad_norm": 0.786801278591156, + "learning_rate": 0.0001616463482382475, + "loss": 0.1774, + "step": 12317 + }, + { + "epoch": 0.28875331844793334, + "grad_norm": 0.37341275811195374, + "learning_rate": 0.0001616405480798759, + "loss": 0.0709, + "step": 12318 + }, + { + "epoch": 0.28877676002273833, + "grad_norm": 0.18902067840099335, + "learning_rate": 0.00016163474758704096, + "loss": 0.033, + "step": 12319 + }, + { + "epoch": 0.2888002015975433, + "grad_norm": 0.6006562113761902, + "learning_rate": 0.00016162894675977413, + "loss": 0.0978, + "step": 12320 + }, + { + "epoch": 0.2888236431723483, + "grad_norm": 0.2732107937335968, + "learning_rate": 0.0001616231455981069, + "loss": 0.0837, + "step": 12321 + }, + { + "epoch": 0.2888470847471533, + "grad_norm": 0.6025024056434631, + "learning_rate": 0.00016161734410207075, + "loss": 0.1566, + "step": 12322 + }, + { + "epoch": 0.2888705263219583, + "grad_norm": 0.8710392713546753, + "learning_rate": 0.00016161154227169715, + "loss": 0.2585, + "step": 12323 + }, + { + "epoch": 0.2888939678967633, + "grad_norm": 0.6013713479042053, + "learning_rate": 0.00016160574010701758, + "loss": 0.1954, + "step": 12324 + }, + { + "epoch": 0.2889174094715683, + "grad_norm": 0.5945386290550232, + "learning_rate": 0.00016159993760806357, + "loss": 0.1945, + "step": 12325 + }, + { + "epoch": 0.2889408510463733, + "grad_norm": 0.07436303049325943, + "learning_rate": 0.00016159413477486654, + "loss": 0.0129, + "step": 12326 + }, + { + "epoch": 0.2889642926211783, + "grad_norm": 0.5683172941207886, + "learning_rate": 0.00016158833160745804, + "loss": 0.1473, + "step": 12327 + }, + { + "epoch": 0.2889877341959833, + "grad_norm": 0.4516993761062622, + "learning_rate": 0.00016158252810586946, + "loss": 0.0997, + "step": 12328 + }, + { + "epoch": 0.2890111757707883, + "grad_norm": 0.2863614857196808, + "learning_rate": 0.0001615767242701324, + "loss": 0.0794, + "step": 12329 + }, + { + "epoch": 0.28903461734559327, + "grad_norm": 0.31153154373168945, + "learning_rate": 0.0001615709201002783, + "loss": 0.0659, + "step": 12330 + }, + { + "epoch": 0.28905805892039826, + "grad_norm": 0.31304699182510376, + "learning_rate": 0.00016156511559633862, + "loss": 0.0366, + "step": 12331 + }, + { + "epoch": 0.28908150049520326, + "grad_norm": 0.5803194642066956, + "learning_rate": 0.00016155931075834492, + "loss": 0.2158, + "step": 12332 + }, + { + "epoch": 0.28910494207000825, + "grad_norm": 0.38457804918289185, + "learning_rate": 0.00016155350558632868, + "loss": 0.0888, + "step": 12333 + }, + { + "epoch": 0.28912838364481325, + "grad_norm": 0.17587214708328247, + "learning_rate": 0.00016154770008032133, + "loss": 0.0369, + "step": 12334 + }, + { + "epoch": 0.28915182521961824, + "grad_norm": 0.4932202994823456, + "learning_rate": 0.0001615418942403545, + "loss": 0.0762, + "step": 12335 + }, + { + "epoch": 0.28917526679442324, + "grad_norm": 0.5271374583244324, + "learning_rate": 0.00016153608806645958, + "loss": 0.1106, + "step": 12336 + }, + { + "epoch": 0.28919870836922823, + "grad_norm": 0.6610024571418762, + "learning_rate": 0.00016153028155866812, + "loss": 0.1499, + "step": 12337 + }, + { + "epoch": 0.2892221499440332, + "grad_norm": 0.5683757066726685, + "learning_rate": 0.00016152447471701163, + "loss": 0.1301, + "step": 12338 + }, + { + "epoch": 0.2892455915188382, + "grad_norm": 0.692779541015625, + "learning_rate": 0.00016151866754152163, + "loss": 0.1429, + "step": 12339 + }, + { + "epoch": 0.2892690330936432, + "grad_norm": 0.3063565790653229, + "learning_rate": 0.0001615128600322296, + "loss": 0.0616, + "step": 12340 + }, + { + "epoch": 0.2892924746684482, + "grad_norm": 0.6857776045799255, + "learning_rate": 0.00016150705218916702, + "loss": 0.1269, + "step": 12341 + }, + { + "epoch": 0.2893159162432532, + "grad_norm": 0.677049994468689, + "learning_rate": 0.00016150124401236547, + "loss": 0.0754, + "step": 12342 + }, + { + "epoch": 0.2893393578180582, + "grad_norm": 0.9065360426902771, + "learning_rate": 0.00016149543550185644, + "loss": 0.7856, + "step": 12343 + }, + { + "epoch": 0.2893627993928632, + "grad_norm": 1.2139978408813477, + "learning_rate": 0.00016148962665767146, + "loss": 0.1572, + "step": 12344 + }, + { + "epoch": 0.2893862409676682, + "grad_norm": 0.5549066066741943, + "learning_rate": 0.00016148381747984204, + "loss": 0.1109, + "step": 12345 + }, + { + "epoch": 0.2894096825424732, + "grad_norm": 0.8012135624885559, + "learning_rate": 0.0001614780079683997, + "loss": 0.1383, + "step": 12346 + }, + { + "epoch": 0.2894331241172782, + "grad_norm": 0.7506958842277527, + "learning_rate": 0.00016147219812337593, + "loss": 0.1057, + "step": 12347 + }, + { + "epoch": 0.28945656569208317, + "grad_norm": 0.5727964639663696, + "learning_rate": 0.00016146638794480232, + "loss": 0.1166, + "step": 12348 + }, + { + "epoch": 0.28948000726688816, + "grad_norm": 0.6767438650131226, + "learning_rate": 0.00016146057743271036, + "loss": 0.1207, + "step": 12349 + }, + { + "epoch": 0.2895034488416932, + "grad_norm": 0.18753531575202942, + "learning_rate": 0.00016145476658713156, + "loss": 0.0437, + "step": 12350 + }, + { + "epoch": 0.2895268904164982, + "grad_norm": 0.5329678058624268, + "learning_rate": 0.00016144895540809746, + "loss": 0.1333, + "step": 12351 + }, + { + "epoch": 0.2895503319913032, + "grad_norm": 0.39061278104782104, + "learning_rate": 0.00016144314389563962, + "loss": 0.0679, + "step": 12352 + }, + { + "epoch": 0.2895737735661082, + "grad_norm": 0.6342465877532959, + "learning_rate": 0.00016143733204978955, + "loss": 0.1788, + "step": 12353 + }, + { + "epoch": 0.2895972151409132, + "grad_norm": 0.18720178306102753, + "learning_rate": 0.00016143151987057876, + "loss": 0.0415, + "step": 12354 + }, + { + "epoch": 0.2896206567157182, + "grad_norm": 0.6415135264396667, + "learning_rate": 0.00016142570735803885, + "loss": 0.7965, + "step": 12355 + }, + { + "epoch": 0.2896440982905232, + "grad_norm": 0.6405467391014099, + "learning_rate": 0.0001614198945122013, + "loss": 0.6888, + "step": 12356 + }, + { + "epoch": 0.28966753986532817, + "grad_norm": 0.30606722831726074, + "learning_rate": 0.0001614140813330977, + "loss": 0.0353, + "step": 12357 + }, + { + "epoch": 0.28969098144013317, + "grad_norm": 1.0460630655288696, + "learning_rate": 0.00016140826782075959, + "loss": 0.3268, + "step": 12358 + }, + { + "epoch": 0.28971442301493816, + "grad_norm": 0.5489214658737183, + "learning_rate": 0.00016140245397521842, + "loss": 0.1346, + "step": 12359 + }, + { + "epoch": 0.28973786458974315, + "grad_norm": 0.3391777575016022, + "learning_rate": 0.00016139663979650585, + "loss": 0.0524, + "step": 12360 + }, + { + "epoch": 0.28976130616454815, + "grad_norm": 0.640933632850647, + "learning_rate": 0.00016139082528465339, + "loss": 0.1698, + "step": 12361 + }, + { + "epoch": 0.28978474773935314, + "grad_norm": 0.42846187949180603, + "learning_rate": 0.0001613850104396926, + "loss": 0.0627, + "step": 12362 + }, + { + "epoch": 0.28980818931415814, + "grad_norm": 0.43990257382392883, + "learning_rate": 0.00016137919526165498, + "loss": 0.1338, + "step": 12363 + }, + { + "epoch": 0.28983163088896313, + "grad_norm": 0.4046320915222168, + "learning_rate": 0.00016137337975057214, + "loss": 0.0413, + "step": 12364 + }, + { + "epoch": 0.2898550724637681, + "grad_norm": 0.7329965829849243, + "learning_rate": 0.00016136756390647562, + "loss": 0.1171, + "step": 12365 + }, + { + "epoch": 0.2898785140385731, + "grad_norm": 0.7305977940559387, + "learning_rate": 0.00016136174772939693, + "loss": 0.1831, + "step": 12366 + }, + { + "epoch": 0.2899019556133781, + "grad_norm": 0.19100159406661987, + "learning_rate": 0.00016135593121936775, + "loss": 0.0375, + "step": 12367 + }, + { + "epoch": 0.2899253971881831, + "grad_norm": 0.48577558994293213, + "learning_rate": 0.00016135011437641951, + "loss": 0.1042, + "step": 12368 + }, + { + "epoch": 0.2899488387629881, + "grad_norm": 0.5218377709388733, + "learning_rate": 0.00016134429720058384, + "loss": 0.1187, + "step": 12369 + }, + { + "epoch": 0.2899722803377931, + "grad_norm": 0.4189870059490204, + "learning_rate": 0.0001613384796918923, + "loss": 0.5759, + "step": 12370 + }, + { + "epoch": 0.2899957219125981, + "grad_norm": 0.6015213131904602, + "learning_rate": 0.0001613326618503764, + "loss": 0.1079, + "step": 12371 + }, + { + "epoch": 0.2900191634874031, + "grad_norm": 0.11203743517398834, + "learning_rate": 0.00016132684367606778, + "loss": 0.0193, + "step": 12372 + }, + { + "epoch": 0.2900426050622081, + "grad_norm": 0.7717220783233643, + "learning_rate": 0.00016132102516899799, + "loss": 0.1422, + "step": 12373 + }, + { + "epoch": 0.2900660466370131, + "grad_norm": 0.24026919901371002, + "learning_rate": 0.00016131520632919858, + "loss": 0.0456, + "step": 12374 + }, + { + "epoch": 0.29008948821181807, + "grad_norm": 0.4861062169075012, + "learning_rate": 0.00016130938715670117, + "loss": 0.1585, + "step": 12375 + }, + { + "epoch": 0.29011292978662306, + "grad_norm": 0.5011429190635681, + "learning_rate": 0.00016130356765153728, + "loss": 0.1045, + "step": 12376 + }, + { + "epoch": 0.29013637136142806, + "grad_norm": 0.5537569522857666, + "learning_rate": 0.00016129774781373848, + "loss": 0.174, + "step": 12377 + }, + { + "epoch": 0.29015981293623305, + "grad_norm": 0.4344937205314636, + "learning_rate": 0.00016129192764333637, + "loss": 0.1391, + "step": 12378 + }, + { + "epoch": 0.29018325451103805, + "grad_norm": 0.37736833095550537, + "learning_rate": 0.0001612861071403626, + "loss": 0.0958, + "step": 12379 + }, + { + "epoch": 0.29020669608584304, + "grad_norm": 0.20548875629901886, + "learning_rate": 0.00016128028630484864, + "loss": 0.033, + "step": 12380 + }, + { + "epoch": 0.29023013766064804, + "grad_norm": 0.7223261594772339, + "learning_rate": 0.00016127446513682612, + "loss": 0.1588, + "step": 12381 + }, + { + "epoch": 0.29025357923545303, + "grad_norm": 0.35747936367988586, + "learning_rate": 0.00016126864363632666, + "loss": 0.0679, + "step": 12382 + }, + { + "epoch": 0.290277020810258, + "grad_norm": 0.6424369215965271, + "learning_rate": 0.00016126282180338182, + "loss": 0.0937, + "step": 12383 + }, + { + "epoch": 0.290300462385063, + "grad_norm": 0.911343514919281, + "learning_rate": 0.00016125699963802313, + "loss": 0.2495, + "step": 12384 + }, + { + "epoch": 0.290323903959868, + "grad_norm": 0.4508140981197357, + "learning_rate": 0.0001612511771402823, + "loss": 0.1079, + "step": 12385 + }, + { + "epoch": 0.290347345534673, + "grad_norm": 0.5653804540634155, + "learning_rate": 0.00016124535431019085, + "loss": 0.0943, + "step": 12386 + }, + { + "epoch": 0.290370787109478, + "grad_norm": 0.462853342294693, + "learning_rate": 0.00016123953114778039, + "loss": 0.0859, + "step": 12387 + }, + { + "epoch": 0.290394228684283, + "grad_norm": 0.49192968010902405, + "learning_rate": 0.0001612337076530825, + "loss": 0.0969, + "step": 12388 + }, + { + "epoch": 0.290417670259088, + "grad_norm": 0.12806257605552673, + "learning_rate": 0.0001612278838261288, + "loss": 0.0198, + "step": 12389 + }, + { + "epoch": 0.290441111833893, + "grad_norm": 0.4911598563194275, + "learning_rate": 0.00016122205966695083, + "loss": 0.1258, + "step": 12390 + }, + { + "epoch": 0.290464553408698, + "grad_norm": 0.5385366082191467, + "learning_rate": 0.00016121623517558029, + "loss": 0.087, + "step": 12391 + }, + { + "epoch": 0.290487994983503, + "grad_norm": 0.6035380363464355, + "learning_rate": 0.00016121041035204874, + "loss": 0.0969, + "step": 12392 + }, + { + "epoch": 0.29051143655830797, + "grad_norm": 0.3177644610404968, + "learning_rate": 0.00016120458519638774, + "loss": 0.0602, + "step": 12393 + }, + { + "epoch": 0.29053487813311296, + "grad_norm": 0.6023731231689453, + "learning_rate": 0.000161198759708629, + "loss": 0.138, + "step": 12394 + }, + { + "epoch": 0.29055831970791796, + "grad_norm": 0.561822772026062, + "learning_rate": 0.00016119293388880404, + "loss": 0.1374, + "step": 12395 + }, + { + "epoch": 0.29058176128272295, + "grad_norm": 0.3199930191040039, + "learning_rate": 0.0001611871077369445, + "loss": 0.0539, + "step": 12396 + }, + { + "epoch": 0.29060520285752794, + "grad_norm": 0.4573383629322052, + "learning_rate": 0.00016118128125308195, + "loss": 0.1014, + "step": 12397 + }, + { + "epoch": 0.29062864443233294, + "grad_norm": 0.28502771258354187, + "learning_rate": 0.00016117545443724811, + "loss": 0.0624, + "step": 12398 + }, + { + "epoch": 0.29065208600713793, + "grad_norm": 0.5820116996765137, + "learning_rate": 0.0001611696272894745, + "loss": 0.4235, + "step": 12399 + }, + { + "epoch": 0.2906755275819429, + "grad_norm": 0.8747044205665588, + "learning_rate": 0.0001611637998097928, + "loss": 0.1817, + "step": 12400 + }, + { + "epoch": 0.290698969156748, + "grad_norm": 0.5163342952728271, + "learning_rate": 0.0001611579719982346, + "loss": 0.1099, + "step": 12401 + }, + { + "epoch": 0.29072241073155297, + "grad_norm": 0.19792024791240692, + "learning_rate": 0.0001611521438548315, + "loss": 0.0315, + "step": 12402 + }, + { + "epoch": 0.29074585230635797, + "grad_norm": 0.3515148162841797, + "learning_rate": 0.00016114631537961514, + "loss": 0.5035, + "step": 12403 + }, + { + "epoch": 0.29076929388116296, + "grad_norm": 0.5660316944122314, + "learning_rate": 0.0001611404865726172, + "loss": 0.126, + "step": 12404 + }, + { + "epoch": 0.29079273545596795, + "grad_norm": 0.48356378078460693, + "learning_rate": 0.00016113465743386923, + "loss": 0.1069, + "step": 12405 + }, + { + "epoch": 0.29081617703077295, + "grad_norm": 0.9672315120697021, + "learning_rate": 0.00016112882796340288, + "loss": 0.111, + "step": 12406 + }, + { + "epoch": 0.29083961860557794, + "grad_norm": 0.2700502574443817, + "learning_rate": 0.00016112299816124981, + "loss": 0.0863, + "step": 12407 + }, + { + "epoch": 0.29086306018038294, + "grad_norm": 0.5513252019882202, + "learning_rate": 0.00016111716802744166, + "loss": 0.0927, + "step": 12408 + }, + { + "epoch": 0.29088650175518793, + "grad_norm": 0.4707707166671753, + "learning_rate": 0.00016111133756201, + "loss": 0.0788, + "step": 12409 + }, + { + "epoch": 0.2909099433299929, + "grad_norm": 0.6255043148994446, + "learning_rate": 0.0001611055067649865, + "loss": 0.0864, + "step": 12410 + }, + { + "epoch": 0.2909333849047979, + "grad_norm": 0.5251614451408386, + "learning_rate": 0.0001610996756364028, + "loss": 0.1418, + "step": 12411 + }, + { + "epoch": 0.2909568264796029, + "grad_norm": 0.5146400928497314, + "learning_rate": 0.0001610938441762906, + "loss": 0.1852, + "step": 12412 + }, + { + "epoch": 0.2909802680544079, + "grad_norm": 0.5316494107246399, + "learning_rate": 0.0001610880123846814, + "loss": 0.1394, + "step": 12413 + }, + { + "epoch": 0.2910037096292129, + "grad_norm": 0.5174316763877869, + "learning_rate": 0.00016108218026160697, + "loss": 0.1362, + "step": 12414 + }, + { + "epoch": 0.2910271512040179, + "grad_norm": 0.48285990953445435, + "learning_rate": 0.0001610763478070989, + "loss": 0.1477, + "step": 12415 + }, + { + "epoch": 0.2910505927788229, + "grad_norm": 0.42658981680870056, + "learning_rate": 0.00016107051502118886, + "loss": 0.0805, + "step": 12416 + }, + { + "epoch": 0.2910740343536279, + "grad_norm": 0.8330734372138977, + "learning_rate": 0.00016106468190390852, + "loss": 0.094, + "step": 12417 + }, + { + "epoch": 0.2910974759284329, + "grad_norm": 0.9011593461036682, + "learning_rate": 0.00016105884845528945, + "loss": 0.104, + "step": 12418 + }, + { + "epoch": 0.2911209175032379, + "grad_norm": 0.5709006190299988, + "learning_rate": 0.00016105301467536336, + "loss": 0.0977, + "step": 12419 + }, + { + "epoch": 0.29114435907804287, + "grad_norm": 0.6294284462928772, + "learning_rate": 0.0001610471805641619, + "loss": 0.1645, + "step": 12420 + }, + { + "epoch": 0.29116780065284786, + "grad_norm": 0.48936375975608826, + "learning_rate": 0.00016104134612171674, + "loss": 0.1482, + "step": 12421 + }, + { + "epoch": 0.29119124222765286, + "grad_norm": 0.6839306354522705, + "learning_rate": 0.0001610355113480595, + "loss": 0.1481, + "step": 12422 + }, + { + "epoch": 0.29121468380245785, + "grad_norm": 0.9090079069137573, + "learning_rate": 0.00016102967624322185, + "loss": 0.127, + "step": 12423 + }, + { + "epoch": 0.29123812537726285, + "grad_norm": 0.4980638027191162, + "learning_rate": 0.0001610238408072355, + "loss": 0.1206, + "step": 12424 + }, + { + "epoch": 0.29126156695206784, + "grad_norm": 0.27438393235206604, + "learning_rate": 0.00016101800504013202, + "loss": 0.0365, + "step": 12425 + }, + { + "epoch": 0.29128500852687284, + "grad_norm": 0.6315383315086365, + "learning_rate": 0.00016101216894194315, + "loss": 0.1189, + "step": 12426 + }, + { + "epoch": 0.29130845010167783, + "grad_norm": 0.5214620232582092, + "learning_rate": 0.00016100633251270056, + "loss": 0.0943, + "step": 12427 + }, + { + "epoch": 0.2913318916764828, + "grad_norm": 0.546343207359314, + "learning_rate": 0.00016100049575243586, + "loss": 0.1242, + "step": 12428 + }, + { + "epoch": 0.2913553332512878, + "grad_norm": 0.633264422416687, + "learning_rate": 0.0001609946586611808, + "loss": 0.2194, + "step": 12429 + }, + { + "epoch": 0.2913787748260928, + "grad_norm": 0.4200579822063446, + "learning_rate": 0.00016098882123896698, + "loss": 0.1156, + "step": 12430 + }, + { + "epoch": 0.2914022164008978, + "grad_norm": 0.9216395616531372, + "learning_rate": 0.0001609829834858261, + "loss": 0.1217, + "step": 12431 + }, + { + "epoch": 0.2914256579757028, + "grad_norm": 0.4960973262786865, + "learning_rate": 0.00016097714540178984, + "loss": 0.0818, + "step": 12432 + }, + { + "epoch": 0.2914490995505078, + "grad_norm": 0.5658662915229797, + "learning_rate": 0.00016097130698688983, + "loss": 0.2022, + "step": 12433 + }, + { + "epoch": 0.2914725411253128, + "grad_norm": 0.1974749118089676, + "learning_rate": 0.00016096546824115784, + "loss": 0.0497, + "step": 12434 + }, + { + "epoch": 0.2914959827001178, + "grad_norm": 0.706924319267273, + "learning_rate": 0.0001609596291646255, + "loss": 0.132, + "step": 12435 + }, + { + "epoch": 0.2915194242749228, + "grad_norm": 0.4099717140197754, + "learning_rate": 0.0001609537897573245, + "loss": 0.5733, + "step": 12436 + }, + { + "epoch": 0.2915428658497278, + "grad_norm": 0.5492329001426697, + "learning_rate": 0.00016094795001928653, + "loss": 0.1091, + "step": 12437 + }, + { + "epoch": 0.29156630742453277, + "grad_norm": 0.35996013879776, + "learning_rate": 0.00016094210995054326, + "loss": 0.048, + "step": 12438 + }, + { + "epoch": 0.29158974899933776, + "grad_norm": 0.6825478076934814, + "learning_rate": 0.00016093626955112638, + "loss": 0.1273, + "step": 12439 + }, + { + "epoch": 0.29161319057414276, + "grad_norm": 0.626751184463501, + "learning_rate": 0.00016093042882106762, + "loss": 0.1588, + "step": 12440 + }, + { + "epoch": 0.29163663214894775, + "grad_norm": 0.5751283764839172, + "learning_rate": 0.0001609245877603986, + "loss": 0.2249, + "step": 12441 + }, + { + "epoch": 0.29166007372375274, + "grad_norm": 0.5664374232292175, + "learning_rate": 0.00016091874636915107, + "loss": 0.1806, + "step": 12442 + }, + { + "epoch": 0.29168351529855774, + "grad_norm": 0.4142511188983917, + "learning_rate": 0.0001609129046473567, + "loss": 0.1331, + "step": 12443 + }, + { + "epoch": 0.29170695687336273, + "grad_norm": 0.4607537090778351, + "learning_rate": 0.00016090706259504724, + "loss": 0.1317, + "step": 12444 + }, + { + "epoch": 0.2917303984481677, + "grad_norm": 0.6300249099731445, + "learning_rate": 0.0001609012202122543, + "loss": 0.1688, + "step": 12445 + }, + { + "epoch": 0.2917538400229727, + "grad_norm": 0.6109418869018555, + "learning_rate": 0.00016089537749900963, + "loss": 0.1584, + "step": 12446 + }, + { + "epoch": 0.2917772815977777, + "grad_norm": 0.5895828008651733, + "learning_rate": 0.00016088953445534494, + "loss": 0.1167, + "step": 12447 + }, + { + "epoch": 0.2918007231725827, + "grad_norm": 0.4136655926704407, + "learning_rate": 0.00016088369108129192, + "loss": 0.3111, + "step": 12448 + }, + { + "epoch": 0.2918241647473877, + "grad_norm": 0.45185911655426025, + "learning_rate": 0.00016087784737688228, + "loss": 0.1107, + "step": 12449 + }, + { + "epoch": 0.2918476063221927, + "grad_norm": 0.34425345063209534, + "learning_rate": 0.00016087200334214777, + "loss": 0.0832, + "step": 12450 + }, + { + "epoch": 0.2918710478969977, + "grad_norm": 0.10719354450702667, + "learning_rate": 0.00016086615897711998, + "loss": 0.0248, + "step": 12451 + }, + { + "epoch": 0.29189448947180274, + "grad_norm": 0.5468716025352478, + "learning_rate": 0.00016086031428183076, + "loss": 0.0909, + "step": 12452 + }, + { + "epoch": 0.29191793104660774, + "grad_norm": 0.18730059266090393, + "learning_rate": 0.00016085446925631175, + "loss": 0.0442, + "step": 12453 + }, + { + "epoch": 0.29194137262141273, + "grad_norm": 0.6164595484733582, + "learning_rate": 0.00016084862390059466, + "loss": 0.1092, + "step": 12454 + }, + { + "epoch": 0.2919648141962177, + "grad_norm": 0.45945650339126587, + "learning_rate": 0.00016084277821471122, + "loss": 0.0608, + "step": 12455 + }, + { + "epoch": 0.2919882557710227, + "grad_norm": 0.5058950185775757, + "learning_rate": 0.0001608369321986932, + "loss": 0.1266, + "step": 12456 + }, + { + "epoch": 0.2920116973458277, + "grad_norm": 0.40981554985046387, + "learning_rate": 0.00016083108585257224, + "loss": 0.0787, + "step": 12457 + }, + { + "epoch": 0.2920351389206327, + "grad_norm": 0.5045484304428101, + "learning_rate": 0.00016082523917638008, + "loss": 0.141, + "step": 12458 + }, + { + "epoch": 0.2920585804954377, + "grad_norm": 0.4133192300796509, + "learning_rate": 0.0001608193921701485, + "loss": 0.105, + "step": 12459 + }, + { + "epoch": 0.2920820220702427, + "grad_norm": 0.8638753890991211, + "learning_rate": 0.00016081354483390918, + "loss": 0.199, + "step": 12460 + }, + { + "epoch": 0.2921054636450477, + "grad_norm": 0.7417733669281006, + "learning_rate": 0.00016080769716769386, + "loss": 0.1905, + "step": 12461 + }, + { + "epoch": 0.2921289052198527, + "grad_norm": 0.537088930606842, + "learning_rate": 0.00016080184917153427, + "loss": 0.126, + "step": 12462 + }, + { + "epoch": 0.2921523467946577, + "grad_norm": 0.2809961140155792, + "learning_rate": 0.0001607960008454621, + "loss": 0.0707, + "step": 12463 + }, + { + "epoch": 0.2921757883694627, + "grad_norm": 0.6288930773735046, + "learning_rate": 0.00016079015218950913, + "loss": 0.1022, + "step": 12464 + }, + { + "epoch": 0.29219922994426767, + "grad_norm": 0.20020954310894012, + "learning_rate": 0.00016078430320370712, + "loss": 0.0372, + "step": 12465 + }, + { + "epoch": 0.29222267151907266, + "grad_norm": 0.39696168899536133, + "learning_rate": 0.00016077845388808774, + "loss": 0.0919, + "step": 12466 + }, + { + "epoch": 0.29224611309387766, + "grad_norm": 0.26581841707229614, + "learning_rate": 0.0001607726042426828, + "loss": 0.0626, + "step": 12467 + }, + { + "epoch": 0.29226955466868265, + "grad_norm": 0.3257412612438202, + "learning_rate": 0.00016076675426752394, + "loss": 0.0726, + "step": 12468 + }, + { + "epoch": 0.29229299624348765, + "grad_norm": 0.3852526545524597, + "learning_rate": 0.000160760903962643, + "loss": 0.0701, + "step": 12469 + }, + { + "epoch": 0.29231643781829264, + "grad_norm": 0.4033302962779999, + "learning_rate": 0.00016075505332807167, + "loss": 0.4277, + "step": 12470 + }, + { + "epoch": 0.29233987939309763, + "grad_norm": 0.385741651058197, + "learning_rate": 0.0001607492023638417, + "loss": 0.0763, + "step": 12471 + }, + { + "epoch": 0.29236332096790263, + "grad_norm": 0.7464970946311951, + "learning_rate": 0.0001607433510699849, + "loss": 0.1069, + "step": 12472 + }, + { + "epoch": 0.2923867625427076, + "grad_norm": 0.43344512581825256, + "learning_rate": 0.00016073749944653295, + "loss": 0.0841, + "step": 12473 + }, + { + "epoch": 0.2924102041175126, + "grad_norm": 0.5781944990158081, + "learning_rate": 0.0001607316474935176, + "loss": 0.0789, + "step": 12474 + }, + { + "epoch": 0.2924336456923176, + "grad_norm": 0.20249208807945251, + "learning_rate": 0.00016072579521097063, + "loss": 0.0217, + "step": 12475 + }, + { + "epoch": 0.2924570872671226, + "grad_norm": 0.3106077015399933, + "learning_rate": 0.0001607199425989238, + "loss": 0.0635, + "step": 12476 + }, + { + "epoch": 0.2924805288419276, + "grad_norm": 0.12565121054649353, + "learning_rate": 0.00016071408965740884, + "loss": 0.0328, + "step": 12477 + }, + { + "epoch": 0.2925039704167326, + "grad_norm": 0.5650808215141296, + "learning_rate": 0.00016070823638645755, + "loss": 0.0881, + "step": 12478 + }, + { + "epoch": 0.2925274119915376, + "grad_norm": 0.8227767944335938, + "learning_rate": 0.00016070238278610162, + "loss": 0.1329, + "step": 12479 + }, + { + "epoch": 0.2925508535663426, + "grad_norm": 0.22099274396896362, + "learning_rate": 0.0001606965288563729, + "loss": 0.0587, + "step": 12480 + }, + { + "epoch": 0.2925742951411476, + "grad_norm": 0.706093430519104, + "learning_rate": 0.00016069067459730305, + "loss": 0.1439, + "step": 12481 + }, + { + "epoch": 0.29259773671595257, + "grad_norm": 0.5442342162132263, + "learning_rate": 0.00016068482000892393, + "loss": 0.0816, + "step": 12482 + }, + { + "epoch": 0.29262117829075757, + "grad_norm": 0.39906734228134155, + "learning_rate": 0.00016067896509126727, + "loss": 0.0433, + "step": 12483 + }, + { + "epoch": 0.29264461986556256, + "grad_norm": 0.5284209251403809, + "learning_rate": 0.00016067310984436484, + "loss": 0.0625, + "step": 12484 + }, + { + "epoch": 0.29266806144036756, + "grad_norm": 0.4774622619152069, + "learning_rate": 0.0001606672542682484, + "loss": 0.1068, + "step": 12485 + }, + { + "epoch": 0.29269150301517255, + "grad_norm": 0.8401097655296326, + "learning_rate": 0.00016066139836294976, + "loss": 0.7079, + "step": 12486 + }, + { + "epoch": 0.29271494458997754, + "grad_norm": 0.4506610333919525, + "learning_rate": 0.00016065554212850062, + "loss": 0.0665, + "step": 12487 + }, + { + "epoch": 0.29273838616478254, + "grad_norm": 0.20442304015159607, + "learning_rate": 0.00016064968556493283, + "loss": 0.0181, + "step": 12488 + }, + { + "epoch": 0.29276182773958753, + "grad_norm": 0.5109789371490479, + "learning_rate": 0.00016064382867227815, + "loss": 0.1318, + "step": 12489 + }, + { + "epoch": 0.2927852693143925, + "grad_norm": 1.1936230659484863, + "learning_rate": 0.00016063797145056836, + "loss": 0.129, + "step": 12490 + }, + { + "epoch": 0.2928087108891975, + "grad_norm": 0.34831318259239197, + "learning_rate": 0.0001606321138998352, + "loss": 0.051, + "step": 12491 + }, + { + "epoch": 0.2928321524640025, + "grad_norm": 0.7703447341918945, + "learning_rate": 0.0001606262560201105, + "loss": 0.1524, + "step": 12492 + }, + { + "epoch": 0.2928555940388075, + "grad_norm": 0.44391539692878723, + "learning_rate": 0.00016062039781142602, + "loss": 0.0586, + "step": 12493 + }, + { + "epoch": 0.2928790356136125, + "grad_norm": 0.3139341175556183, + "learning_rate": 0.00016061453927381357, + "loss": 0.0656, + "step": 12494 + }, + { + "epoch": 0.2929024771884175, + "grad_norm": 0.6428908109664917, + "learning_rate": 0.0001606086804073049, + "loss": 0.142, + "step": 12495 + }, + { + "epoch": 0.2929259187632225, + "grad_norm": 0.535552442073822, + "learning_rate": 0.00016060282121193185, + "loss": 0.1185, + "step": 12496 + }, + { + "epoch": 0.2929493603380275, + "grad_norm": 0.13519099354743958, + "learning_rate": 0.0001605969616877262, + "loss": 0.0328, + "step": 12497 + }, + { + "epoch": 0.2929728019128325, + "grad_norm": 0.8370836973190308, + "learning_rate": 0.00016059110183471971, + "loss": 0.145, + "step": 12498 + }, + { + "epoch": 0.2929962434876375, + "grad_norm": 0.5775215029716492, + "learning_rate": 0.0001605852416529442, + "loss": 0.1363, + "step": 12499 + }, + { + "epoch": 0.29301968506244247, + "grad_norm": 0.4373376667499542, + "learning_rate": 0.00016057938114243147, + "loss": 0.1438, + "step": 12500 + }, + { + "epoch": 0.29304312663724746, + "grad_norm": 0.7440882921218872, + "learning_rate": 0.0001605735203032133, + "loss": 0.1164, + "step": 12501 + }, + { + "epoch": 0.29306656821205246, + "grad_norm": 0.13223880529403687, + "learning_rate": 0.00016056765913532151, + "loss": 0.031, + "step": 12502 + }, + { + "epoch": 0.2930900097868575, + "grad_norm": 0.3645975887775421, + "learning_rate": 0.00016056179763878793, + "loss": 0.064, + "step": 12503 + }, + { + "epoch": 0.2931134513616625, + "grad_norm": 0.9497394561767578, + "learning_rate": 0.00016055593581364432, + "loss": 0.1622, + "step": 12504 + }, + { + "epoch": 0.2931368929364675, + "grad_norm": 0.6696357727050781, + "learning_rate": 0.0001605500736599225, + "loss": 0.7985, + "step": 12505 + }, + { + "epoch": 0.2931603345112725, + "grad_norm": 0.5275449752807617, + "learning_rate": 0.00016054421117765423, + "loss": 0.1076, + "step": 12506 + }, + { + "epoch": 0.2931837760860775, + "grad_norm": 0.4973558187484741, + "learning_rate": 0.00016053834836687143, + "loss": 0.0984, + "step": 12507 + }, + { + "epoch": 0.2932072176608825, + "grad_norm": 0.5788658857345581, + "learning_rate": 0.00016053248522760583, + "loss": 0.0764, + "step": 12508 + }, + { + "epoch": 0.2932306592356875, + "grad_norm": 0.6098034381866455, + "learning_rate": 0.00016052662175988926, + "loss": 0.0926, + "step": 12509 + }, + { + "epoch": 0.29325410081049247, + "grad_norm": 0.9837958216667175, + "learning_rate": 0.00016052075796375352, + "loss": 0.2229, + "step": 12510 + }, + { + "epoch": 0.29327754238529746, + "grad_norm": 1.7368918657302856, + "learning_rate": 0.00016051489383923047, + "loss": 0.1035, + "step": 12511 + }, + { + "epoch": 0.29330098396010246, + "grad_norm": 0.3699105381965637, + "learning_rate": 0.00016050902938635192, + "loss": 0.0587, + "step": 12512 + }, + { + "epoch": 0.29332442553490745, + "grad_norm": 0.21459050476551056, + "learning_rate": 0.00016050316460514962, + "loss": 0.0247, + "step": 12513 + }, + { + "epoch": 0.29334786710971245, + "grad_norm": 0.1967732012271881, + "learning_rate": 0.0001604972994956555, + "loss": 0.0247, + "step": 12514 + }, + { + "epoch": 0.29337130868451744, + "grad_norm": 0.17767491936683655, + "learning_rate": 0.0001604914340579013, + "loss": 0.0309, + "step": 12515 + }, + { + "epoch": 0.29339475025932243, + "grad_norm": 0.5281283855438232, + "learning_rate": 0.0001604855682919189, + "loss": 0.0978, + "step": 12516 + }, + { + "epoch": 0.29341819183412743, + "grad_norm": 0.6460756063461304, + "learning_rate": 0.00016047970219774008, + "loss": 0.0686, + "step": 12517 + }, + { + "epoch": 0.2934416334089324, + "grad_norm": 0.6089462041854858, + "learning_rate": 0.00016047383577539673, + "loss": 0.6265, + "step": 12518 + }, + { + "epoch": 0.2934650749837374, + "grad_norm": 1.0996875762939453, + "learning_rate": 0.0001604679690249206, + "loss": 0.3657, + "step": 12519 + }, + { + "epoch": 0.2934885165585424, + "grad_norm": 0.6852155923843384, + "learning_rate": 0.00016046210194634362, + "loss": 0.1799, + "step": 12520 + }, + { + "epoch": 0.2935119581333474, + "grad_norm": 0.45474255084991455, + "learning_rate": 0.00016045623453969757, + "loss": 0.1269, + "step": 12521 + }, + { + "epoch": 0.2935353997081524, + "grad_norm": 0.5203748941421509, + "learning_rate": 0.00016045036680501426, + "loss": 0.0891, + "step": 12522 + }, + { + "epoch": 0.2935588412829574, + "grad_norm": 0.3957171142101288, + "learning_rate": 0.00016044449874232555, + "loss": 0.0427, + "step": 12523 + }, + { + "epoch": 0.2935822828577624, + "grad_norm": 1.1523250341415405, + "learning_rate": 0.0001604386303516633, + "loss": 0.2111, + "step": 12524 + }, + { + "epoch": 0.2936057244325674, + "grad_norm": 0.5512673854827881, + "learning_rate": 0.0001604327616330594, + "loss": 0.136, + "step": 12525 + }, + { + "epoch": 0.2936291660073724, + "grad_norm": 0.4629794955253601, + "learning_rate": 0.00016042689258654558, + "loss": 0.1174, + "step": 12526 + }, + { + "epoch": 0.29365260758217737, + "grad_norm": 0.23838108777999878, + "learning_rate": 0.00016042102321215374, + "loss": 0.0783, + "step": 12527 + }, + { + "epoch": 0.29367604915698237, + "grad_norm": 0.2685483396053314, + "learning_rate": 0.00016041515350991573, + "loss": 0.0502, + "step": 12528 + }, + { + "epoch": 0.29369949073178736, + "grad_norm": 0.4270746409893036, + "learning_rate": 0.00016040928347986345, + "loss": 0.5671, + "step": 12529 + }, + { + "epoch": 0.29372293230659235, + "grad_norm": 0.50142902135849, + "learning_rate": 0.00016040341312202863, + "loss": 0.1398, + "step": 12530 + }, + { + "epoch": 0.29374637388139735, + "grad_norm": 0.17122961580753326, + "learning_rate": 0.00016039754243644323, + "loss": 0.0557, + "step": 12531 + }, + { + "epoch": 0.29376981545620234, + "grad_norm": 1.3439007997512817, + "learning_rate": 0.00016039167142313905, + "loss": 0.1569, + "step": 12532 + }, + { + "epoch": 0.29379325703100734, + "grad_norm": 0.38588497042655945, + "learning_rate": 0.000160385800082148, + "loss": 0.0811, + "step": 12533 + }, + { + "epoch": 0.29381669860581233, + "grad_norm": 0.3732953369617462, + "learning_rate": 0.00016037992841350186, + "loss": 0.1015, + "step": 12534 + }, + { + "epoch": 0.2938401401806173, + "grad_norm": 0.3790472745895386, + "learning_rate": 0.00016037405641723256, + "loss": 0.0876, + "step": 12535 + }, + { + "epoch": 0.2938635817554223, + "grad_norm": 0.34946009516716003, + "learning_rate": 0.00016036818409337187, + "loss": 0.0737, + "step": 12536 + }, + { + "epoch": 0.2938870233302273, + "grad_norm": 0.47302594780921936, + "learning_rate": 0.00016036231144195178, + "loss": 0.1377, + "step": 12537 + }, + { + "epoch": 0.2939104649050323, + "grad_norm": 0.34501877427101135, + "learning_rate": 0.00016035643846300408, + "loss": 0.0726, + "step": 12538 + }, + { + "epoch": 0.2939339064798373, + "grad_norm": 0.11909965425729752, + "learning_rate": 0.00016035056515656065, + "loss": 0.0186, + "step": 12539 + }, + { + "epoch": 0.2939573480546423, + "grad_norm": 0.42779815196990967, + "learning_rate": 0.00016034469152265336, + "loss": 0.0934, + "step": 12540 + }, + { + "epoch": 0.2939807896294473, + "grad_norm": 0.42224636673927307, + "learning_rate": 0.00016033881756131405, + "loss": 0.0406, + "step": 12541 + }, + { + "epoch": 0.2940042312042523, + "grad_norm": 0.6880200505256653, + "learning_rate": 0.00016033294327257464, + "loss": 0.658, + "step": 12542 + }, + { + "epoch": 0.2940276727790573, + "grad_norm": 0.7132605314254761, + "learning_rate": 0.00016032706865646697, + "loss": 0.2321, + "step": 12543 + }, + { + "epoch": 0.2940511143538623, + "grad_norm": 0.7754214406013489, + "learning_rate": 0.00016032119371302294, + "loss": 0.1973, + "step": 12544 + }, + { + "epoch": 0.29407455592866727, + "grad_norm": 0.13345573842525482, + "learning_rate": 0.0001603153184422744, + "loss": 0.0226, + "step": 12545 + }, + { + "epoch": 0.29409799750347226, + "grad_norm": 0.8559541702270508, + "learning_rate": 0.00016030944284425326, + "loss": 0.1059, + "step": 12546 + }, + { + "epoch": 0.29412143907827726, + "grad_norm": 0.4952225387096405, + "learning_rate": 0.0001603035669189914, + "loss": 0.1529, + "step": 12547 + }, + { + "epoch": 0.29414488065308225, + "grad_norm": 0.25077858567237854, + "learning_rate": 0.00016029769066652064, + "loss": 0.0476, + "step": 12548 + }, + { + "epoch": 0.29416832222788725, + "grad_norm": 0.6478359699249268, + "learning_rate": 0.00016029181408687296, + "loss": 0.1079, + "step": 12549 + }, + { + "epoch": 0.29419176380269224, + "grad_norm": 0.6659058928489685, + "learning_rate": 0.00016028593718008023, + "loss": 0.115, + "step": 12550 + }, + { + "epoch": 0.29421520537749724, + "grad_norm": 0.5794030427932739, + "learning_rate": 0.00016028005994617428, + "loss": 0.8101, + "step": 12551 + }, + { + "epoch": 0.29423864695230223, + "grad_norm": 0.16273866593837738, + "learning_rate": 0.000160274182385187, + "loss": 0.0432, + "step": 12552 + }, + { + "epoch": 0.2942620885271072, + "grad_norm": 0.7373908162117004, + "learning_rate": 0.00016026830449715034, + "loss": 0.1126, + "step": 12553 + }, + { + "epoch": 0.2942855301019123, + "grad_norm": 0.19897788763046265, + "learning_rate": 0.00016026242628209615, + "loss": 0.0444, + "step": 12554 + }, + { + "epoch": 0.29430897167671727, + "grad_norm": 0.4699004590511322, + "learning_rate": 0.00016025654774005633, + "loss": 0.0621, + "step": 12555 + }, + { + "epoch": 0.29433241325152226, + "grad_norm": 0.8499754667282104, + "learning_rate": 0.00016025066887106283, + "loss": 0.171, + "step": 12556 + }, + { + "epoch": 0.29435585482632726, + "grad_norm": 0.5007830858230591, + "learning_rate": 0.0001602447896751475, + "loss": 0.1313, + "step": 12557 + }, + { + "epoch": 0.29437929640113225, + "grad_norm": 0.4875776767730713, + "learning_rate": 0.00016023891015234224, + "loss": 0.698, + "step": 12558 + }, + { + "epoch": 0.29440273797593725, + "grad_norm": 0.4127693772315979, + "learning_rate": 0.00016023303030267893, + "loss": 0.5806, + "step": 12559 + }, + { + "epoch": 0.29442617955074224, + "grad_norm": 0.4424225687980652, + "learning_rate": 0.0001602271501261895, + "loss": 0.0818, + "step": 12560 + }, + { + "epoch": 0.29444962112554723, + "grad_norm": 0.294494092464447, + "learning_rate": 0.00016022126962290592, + "loss": 0.0662, + "step": 12561 + }, + { + "epoch": 0.29447306270035223, + "grad_norm": 0.7820115685462952, + "learning_rate": 0.00016021538879286, + "loss": 0.1511, + "step": 12562 + }, + { + "epoch": 0.2944965042751572, + "grad_norm": 0.5030354857444763, + "learning_rate": 0.0001602095076360837, + "loss": 0.1603, + "step": 12563 + }, + { + "epoch": 0.2945199458499622, + "grad_norm": 0.3096480965614319, + "learning_rate": 0.00016020362615260889, + "loss": 0.056, + "step": 12564 + }, + { + "epoch": 0.2945433874247672, + "grad_norm": 0.6028105616569519, + "learning_rate": 0.00016019774434246754, + "loss": 0.8141, + "step": 12565 + }, + { + "epoch": 0.2945668289995722, + "grad_norm": 0.45057085156440735, + "learning_rate": 0.00016019186220569154, + "loss": 0.0986, + "step": 12566 + }, + { + "epoch": 0.2945902705743772, + "grad_norm": 0.44554799795150757, + "learning_rate": 0.00016018597974231278, + "loss": 0.7123, + "step": 12567 + }, + { + "epoch": 0.2946137121491822, + "grad_norm": 0.4949273467063904, + "learning_rate": 0.0001601800969523632, + "loss": 0.1049, + "step": 12568 + }, + { + "epoch": 0.2946371537239872, + "grad_norm": 0.6175662875175476, + "learning_rate": 0.00016017421383587475, + "loss": 0.1019, + "step": 12569 + }, + { + "epoch": 0.2946605952987922, + "grad_norm": 0.66237872838974, + "learning_rate": 0.00016016833039287932, + "loss": 0.1185, + "step": 12570 + }, + { + "epoch": 0.2946840368735972, + "grad_norm": 0.35300272703170776, + "learning_rate": 0.0001601624466234088, + "loss": 0.0716, + "step": 12571 + }, + { + "epoch": 0.29470747844840217, + "grad_norm": 0.598174512386322, + "learning_rate": 0.00016015656252749518, + "loss": 0.0889, + "step": 12572 + }, + { + "epoch": 0.29473092002320717, + "grad_norm": 0.387551486492157, + "learning_rate": 0.00016015067810517034, + "loss": 0.0648, + "step": 12573 + }, + { + "epoch": 0.29475436159801216, + "grad_norm": 0.6329023241996765, + "learning_rate": 0.00016014479335646627, + "loss": 0.1461, + "step": 12574 + }, + { + "epoch": 0.29477780317281715, + "grad_norm": 0.3878251314163208, + "learning_rate": 0.00016013890828141483, + "loss": 0.098, + "step": 12575 + }, + { + "epoch": 0.29480124474762215, + "grad_norm": 0.1479564607143402, + "learning_rate": 0.000160133022880048, + "loss": 0.0239, + "step": 12576 + }, + { + "epoch": 0.29482468632242714, + "grad_norm": 0.6024467945098877, + "learning_rate": 0.00016012713715239768, + "loss": 0.1516, + "step": 12577 + }, + { + "epoch": 0.29484812789723214, + "grad_norm": 0.6068441271781921, + "learning_rate": 0.00016012125109849583, + "loss": 0.1174, + "step": 12578 + }, + { + "epoch": 0.29487156947203713, + "grad_norm": 0.7089954018592834, + "learning_rate": 0.00016011536471837438, + "loss": 0.1951, + "step": 12579 + }, + { + "epoch": 0.2948950110468421, + "grad_norm": 0.8714286684989929, + "learning_rate": 0.00016010947801206528, + "loss": 0.0941, + "step": 12580 + }, + { + "epoch": 0.2949184526216471, + "grad_norm": 0.46483123302459717, + "learning_rate": 0.00016010359097960047, + "loss": 0.0681, + "step": 12581 + }, + { + "epoch": 0.2949418941964521, + "grad_norm": 0.6418969631195068, + "learning_rate": 0.00016009770362101187, + "loss": 0.1288, + "step": 12582 + }, + { + "epoch": 0.2949653357712571, + "grad_norm": 0.6845934987068176, + "learning_rate": 0.00016009181593633145, + "loss": 0.184, + "step": 12583 + }, + { + "epoch": 0.2949887773460621, + "grad_norm": 0.7532872557640076, + "learning_rate": 0.00016008592792559113, + "loss": 0.1495, + "step": 12584 + }, + { + "epoch": 0.2950122189208671, + "grad_norm": 0.43023672699928284, + "learning_rate": 0.00016008003958882286, + "loss": 0.1014, + "step": 12585 + }, + { + "epoch": 0.2950356604956721, + "grad_norm": 0.24596130847930908, + "learning_rate": 0.00016007415092605864, + "loss": 0.0397, + "step": 12586 + }, + { + "epoch": 0.2950591020704771, + "grad_norm": 0.6245896816253662, + "learning_rate": 0.0001600682619373304, + "loss": 0.1457, + "step": 12587 + }, + { + "epoch": 0.2950825436452821, + "grad_norm": 0.6497378945350647, + "learning_rate": 0.00016006237262267004, + "loss": 0.2034, + "step": 12588 + }, + { + "epoch": 0.2951059852200871, + "grad_norm": 0.3212549388408661, + "learning_rate": 0.00016005648298210957, + "loss": 0.0243, + "step": 12589 + }, + { + "epoch": 0.29512942679489207, + "grad_norm": 0.8965354561805725, + "learning_rate": 0.00016005059301568095, + "loss": 0.2646, + "step": 12590 + }, + { + "epoch": 0.29515286836969706, + "grad_norm": 0.281028151512146, + "learning_rate": 0.0001600447027234161, + "loss": 0.0371, + "step": 12591 + }, + { + "epoch": 0.29517630994450206, + "grad_norm": 0.20547284185886383, + "learning_rate": 0.00016003881210534704, + "loss": 0.0451, + "step": 12592 + }, + { + "epoch": 0.29519975151930705, + "grad_norm": 0.16344794631004333, + "learning_rate": 0.00016003292116150566, + "loss": 0.0415, + "step": 12593 + }, + { + "epoch": 0.29522319309411205, + "grad_norm": 0.4111344814300537, + "learning_rate": 0.00016002702989192398, + "loss": 0.0765, + "step": 12594 + }, + { + "epoch": 0.29524663466891704, + "grad_norm": 0.3104822337627411, + "learning_rate": 0.00016002113829663396, + "loss": 0.0952, + "step": 12595 + }, + { + "epoch": 0.29527007624372203, + "grad_norm": 0.3375358283519745, + "learning_rate": 0.00016001524637566754, + "loss": 0.0766, + "step": 12596 + }, + { + "epoch": 0.29529351781852703, + "grad_norm": 0.4134819805622101, + "learning_rate": 0.00016000935412905672, + "loss": 0.1066, + "step": 12597 + }, + { + "epoch": 0.295316959393332, + "grad_norm": 0.46933144330978394, + "learning_rate": 0.00016000346155683342, + "loss": 0.1001, + "step": 12598 + }, + { + "epoch": 0.295340400968137, + "grad_norm": 0.5574300289154053, + "learning_rate": 0.0001599975686590297, + "loss": 0.0944, + "step": 12599 + }, + { + "epoch": 0.295363842542942, + "grad_norm": 0.3876362442970276, + "learning_rate": 0.00015999167543567745, + "loss": 0.1196, + "step": 12600 + }, + { + "epoch": 0.295387284117747, + "grad_norm": 0.09142835438251495, + "learning_rate": 0.0001599857818868087, + "loss": 0.0244, + "step": 12601 + }, + { + "epoch": 0.295410725692552, + "grad_norm": 0.14391109347343445, + "learning_rate": 0.0001599798880124554, + "loss": 0.0459, + "step": 12602 + }, + { + "epoch": 0.295434167267357, + "grad_norm": 0.27799683809280396, + "learning_rate": 0.00015997399381264952, + "loss": 0.0466, + "step": 12603 + }, + { + "epoch": 0.295457608842162, + "grad_norm": 0.11969947814941406, + "learning_rate": 0.00015996809928742312, + "loss": 0.0195, + "step": 12604 + }, + { + "epoch": 0.29548105041696704, + "grad_norm": 0.4021434187889099, + "learning_rate": 0.0001599622044368081, + "loss": 0.0857, + "step": 12605 + }, + { + "epoch": 0.29550449199177203, + "grad_norm": 0.4717649817466736, + "learning_rate": 0.00015995630926083647, + "loss": 0.1021, + "step": 12606 + }, + { + "epoch": 0.29552793356657703, + "grad_norm": 0.4848802089691162, + "learning_rate": 0.0001599504137595402, + "loss": 0.1474, + "step": 12607 + }, + { + "epoch": 0.295551375141382, + "grad_norm": 0.44375336170196533, + "learning_rate": 0.00015994451793295133, + "loss": 0.0903, + "step": 12608 + }, + { + "epoch": 0.295574816716187, + "grad_norm": 0.5753815770149231, + "learning_rate": 0.00015993862178110182, + "loss": 0.1616, + "step": 12609 + }, + { + "epoch": 0.295598258290992, + "grad_norm": 0.2634436786174774, + "learning_rate": 0.00015993272530402365, + "loss": 0.0519, + "step": 12610 + }, + { + "epoch": 0.295621699865797, + "grad_norm": 0.7378050088882446, + "learning_rate": 0.00015992682850174883, + "loss": 0.117, + "step": 12611 + }, + { + "epoch": 0.295645141440602, + "grad_norm": 0.5783119797706604, + "learning_rate": 0.00015992093137430936, + "loss": 0.0934, + "step": 12612 + }, + { + "epoch": 0.295668583015407, + "grad_norm": 0.48559072613716125, + "learning_rate": 0.00015991503392173722, + "loss": 0.0636, + "step": 12613 + }, + { + "epoch": 0.295692024590212, + "grad_norm": 0.9213685989379883, + "learning_rate": 0.0001599091361440644, + "loss": 0.1663, + "step": 12614 + }, + { + "epoch": 0.295715466165017, + "grad_norm": 0.7450405359268188, + "learning_rate": 0.00015990323804132296, + "loss": 0.1748, + "step": 12615 + }, + { + "epoch": 0.295738907739822, + "grad_norm": 0.5079326033592224, + "learning_rate": 0.00015989733961354485, + "loss": 0.7083, + "step": 12616 + }, + { + "epoch": 0.29576234931462697, + "grad_norm": 0.2980318069458008, + "learning_rate": 0.0001598914408607621, + "loss": 0.0813, + "step": 12617 + }, + { + "epoch": 0.29578579088943197, + "grad_norm": 0.23281729221343994, + "learning_rate": 0.0001598855417830067, + "loss": 0.0523, + "step": 12618 + }, + { + "epoch": 0.29580923246423696, + "grad_norm": 0.4343240559101105, + "learning_rate": 0.00015987964238031068, + "loss": 0.1004, + "step": 12619 + }, + { + "epoch": 0.29583267403904195, + "grad_norm": 0.6315838694572449, + "learning_rate": 0.00015987374265270603, + "loss": 0.1057, + "step": 12620 + }, + { + "epoch": 0.29585611561384695, + "grad_norm": 0.2968744933605194, + "learning_rate": 0.00015986784260022473, + "loss": 0.0615, + "step": 12621 + }, + { + "epoch": 0.29587955718865194, + "grad_norm": 0.4790011942386627, + "learning_rate": 0.0001598619422228989, + "loss": 0.0638, + "step": 12622 + }, + { + "epoch": 0.29590299876345694, + "grad_norm": 0.14460371434688568, + "learning_rate": 0.00015985604152076046, + "loss": 0.0441, + "step": 12623 + }, + { + "epoch": 0.29592644033826193, + "grad_norm": 0.2986280024051666, + "learning_rate": 0.00015985014049384145, + "loss": 0.0521, + "step": 12624 + }, + { + "epoch": 0.2959498819130669, + "grad_norm": 0.6461062431335449, + "learning_rate": 0.00015984423914217391, + "loss": 0.1969, + "step": 12625 + }, + { + "epoch": 0.2959733234878719, + "grad_norm": 0.5366066694259644, + "learning_rate": 0.00015983833746578983, + "loss": 0.1732, + "step": 12626 + }, + { + "epoch": 0.2959967650626769, + "grad_norm": 0.7425017952919006, + "learning_rate": 0.00015983243546472122, + "loss": 0.1674, + "step": 12627 + }, + { + "epoch": 0.2960202066374819, + "grad_norm": 0.8603092432022095, + "learning_rate": 0.0001598265331390002, + "loss": 0.1751, + "step": 12628 + }, + { + "epoch": 0.2960436482122869, + "grad_norm": 0.49380335211753845, + "learning_rate": 0.00015982063048865868, + "loss": 0.1587, + "step": 12629 + }, + { + "epoch": 0.2960670897870919, + "grad_norm": 0.12221488356590271, + "learning_rate": 0.00015981472751372872, + "loss": 0.0271, + "step": 12630 + }, + { + "epoch": 0.2960905313618969, + "grad_norm": 0.4462064504623413, + "learning_rate": 0.0001598088242142424, + "loss": 0.119, + "step": 12631 + }, + { + "epoch": 0.2961139729367019, + "grad_norm": 0.40235188603401184, + "learning_rate": 0.00015980292059023174, + "loss": 0.0891, + "step": 12632 + }, + { + "epoch": 0.2961374145115069, + "grad_norm": 0.1217670664191246, + "learning_rate": 0.0001597970166417287, + "loss": 0.0308, + "step": 12633 + }, + { + "epoch": 0.2961608560863119, + "grad_norm": 0.5365267992019653, + "learning_rate": 0.00015979111236876542, + "loss": 0.4887, + "step": 12634 + }, + { + "epoch": 0.29618429766111687, + "grad_norm": 0.8301663994789124, + "learning_rate": 0.00015978520777137385, + "loss": 0.2024, + "step": 12635 + }, + { + "epoch": 0.29620773923592186, + "grad_norm": 0.24897873401641846, + "learning_rate": 0.00015977930284958607, + "loss": 0.1016, + "step": 12636 + }, + { + "epoch": 0.29623118081072686, + "grad_norm": 0.48585623502731323, + "learning_rate": 0.0001597733976034341, + "loss": 0.1086, + "step": 12637 + }, + { + "epoch": 0.29625462238553185, + "grad_norm": 0.4686746895313263, + "learning_rate": 0.00015976749203295, + "loss": 0.0846, + "step": 12638 + }, + { + "epoch": 0.29627806396033685, + "grad_norm": 0.48655498027801514, + "learning_rate": 0.0001597615861381658, + "loss": 0.1945, + "step": 12639 + }, + { + "epoch": 0.29630150553514184, + "grad_norm": 0.6267929077148438, + "learning_rate": 0.00015975567991911358, + "loss": 0.7758, + "step": 12640 + }, + { + "epoch": 0.29632494710994683, + "grad_norm": 0.31542500853538513, + "learning_rate": 0.00015974977337582534, + "loss": 0.0285, + "step": 12641 + }, + { + "epoch": 0.29634838868475183, + "grad_norm": 0.3413114845752716, + "learning_rate": 0.00015974386650833316, + "loss": 0.0518, + "step": 12642 + }, + { + "epoch": 0.2963718302595568, + "grad_norm": 0.718249499797821, + "learning_rate": 0.00015973795931666906, + "loss": 0.1578, + "step": 12643 + }, + { + "epoch": 0.2963952718343618, + "grad_norm": 0.7159321904182434, + "learning_rate": 0.00015973205180086513, + "loss": 0.2374, + "step": 12644 + }, + { + "epoch": 0.2964187134091668, + "grad_norm": 0.1371094435453415, + "learning_rate": 0.0001597261439609534, + "loss": 0.0297, + "step": 12645 + }, + { + "epoch": 0.2964421549839718, + "grad_norm": 0.4641745388507843, + "learning_rate": 0.00015972023579696596, + "loss": 0.1802, + "step": 12646 + }, + { + "epoch": 0.2964655965587768, + "grad_norm": 0.6289628744125366, + "learning_rate": 0.00015971432730893482, + "loss": 0.1772, + "step": 12647 + }, + { + "epoch": 0.2964890381335818, + "grad_norm": 0.4829002320766449, + "learning_rate": 0.00015970841849689203, + "loss": 0.0658, + "step": 12648 + }, + { + "epoch": 0.2965124797083868, + "grad_norm": 0.3983338475227356, + "learning_rate": 0.00015970250936086974, + "loss": 0.1265, + "step": 12649 + }, + { + "epoch": 0.2965359212831918, + "grad_norm": 0.1045491099357605, + "learning_rate": 0.00015969659990089992, + "loss": 0.0135, + "step": 12650 + }, + { + "epoch": 0.2965593628579968, + "grad_norm": 0.486054927110672, + "learning_rate": 0.00015969069011701468, + "loss": 0.1335, + "step": 12651 + }, + { + "epoch": 0.29658280443280177, + "grad_norm": 0.5235721468925476, + "learning_rate": 0.00015968478000924607, + "loss": 0.1176, + "step": 12652 + }, + { + "epoch": 0.29660624600760677, + "grad_norm": 0.11612416803836823, + "learning_rate": 0.0001596788695776262, + "loss": 0.0306, + "step": 12653 + }, + { + "epoch": 0.29662968758241176, + "grad_norm": 0.3819279074668884, + "learning_rate": 0.00015967295882218706, + "loss": 0.0742, + "step": 12654 + }, + { + "epoch": 0.29665312915721676, + "grad_norm": 0.5014808177947998, + "learning_rate": 0.0001596670477429608, + "loss": 0.1165, + "step": 12655 + }, + { + "epoch": 0.2966765707320218, + "grad_norm": 0.1135832667350769, + "learning_rate": 0.00015966113633997947, + "loss": 0.0195, + "step": 12656 + }, + { + "epoch": 0.2967000123068268, + "grad_norm": 0.5610535144805908, + "learning_rate": 0.00015965522461327506, + "loss": 0.2048, + "step": 12657 + }, + { + "epoch": 0.2967234538816318, + "grad_norm": 0.4207877218723297, + "learning_rate": 0.00015964931256287978, + "loss": 0.1015, + "step": 12658 + }, + { + "epoch": 0.2967468954564368, + "grad_norm": 0.26687031984329224, + "learning_rate": 0.00015964340018882566, + "loss": 0.0705, + "step": 12659 + }, + { + "epoch": 0.2967703370312418, + "grad_norm": 0.522560179233551, + "learning_rate": 0.00015963748749114477, + "loss": 0.1073, + "step": 12660 + }, + { + "epoch": 0.2967937786060468, + "grad_norm": 0.10656917095184326, + "learning_rate": 0.0001596315744698692, + "loss": 0.0185, + "step": 12661 + }, + { + "epoch": 0.29681722018085177, + "grad_norm": 0.1549176126718521, + "learning_rate": 0.000159625661125031, + "loss": 0.0277, + "step": 12662 + }, + { + "epoch": 0.29684066175565677, + "grad_norm": 0.6511155962944031, + "learning_rate": 0.0001596197474566623, + "loss": 0.776, + "step": 12663 + }, + { + "epoch": 0.29686410333046176, + "grad_norm": 0.10537836700677872, + "learning_rate": 0.0001596138334647952, + "loss": 0.0147, + "step": 12664 + }, + { + "epoch": 0.29688754490526675, + "grad_norm": 0.48966407775878906, + "learning_rate": 0.0001596079191494617, + "loss": 0.0906, + "step": 12665 + }, + { + "epoch": 0.29691098648007175, + "grad_norm": 0.506377637386322, + "learning_rate": 0.00015960200451069402, + "loss": 0.106, + "step": 12666 + }, + { + "epoch": 0.29693442805487674, + "grad_norm": 0.4072006046772003, + "learning_rate": 0.0001595960895485242, + "loss": 0.1133, + "step": 12667 + }, + { + "epoch": 0.29695786962968174, + "grad_norm": 0.23216985166072845, + "learning_rate": 0.00015959017426298426, + "loss": 0.0611, + "step": 12668 + }, + { + "epoch": 0.29698131120448673, + "grad_norm": 0.2040538489818573, + "learning_rate": 0.0001595842586541064, + "loss": 0.0303, + "step": 12669 + }, + { + "epoch": 0.2970047527792917, + "grad_norm": 0.4320656359195709, + "learning_rate": 0.00015957834272192266, + "loss": 0.1142, + "step": 12670 + }, + { + "epoch": 0.2970281943540967, + "grad_norm": 0.1992344856262207, + "learning_rate": 0.00015957242646646517, + "loss": 0.0442, + "step": 12671 + }, + { + "epoch": 0.2970516359289017, + "grad_norm": 0.41193562746047974, + "learning_rate": 0.000159566509887766, + "loss": 0.0863, + "step": 12672 + }, + { + "epoch": 0.2970750775037067, + "grad_norm": 0.5330110788345337, + "learning_rate": 0.00015956059298585727, + "loss": 0.4255, + "step": 12673 + }, + { + "epoch": 0.2970985190785117, + "grad_norm": 0.2915492057800293, + "learning_rate": 0.0001595546757607711, + "loss": 0.0466, + "step": 12674 + }, + { + "epoch": 0.2971219606533167, + "grad_norm": 0.22192376852035522, + "learning_rate": 0.0001595487582125396, + "loss": 0.0448, + "step": 12675 + }, + { + "epoch": 0.2971454022281217, + "grad_norm": 0.2118939459323883, + "learning_rate": 0.00015954284034119483, + "loss": 0.0449, + "step": 12676 + }, + { + "epoch": 0.2971688438029267, + "grad_norm": 0.5754262208938599, + "learning_rate": 0.00015953692214676896, + "loss": 0.1291, + "step": 12677 + }, + { + "epoch": 0.2971922853777317, + "grad_norm": 0.37506526708602905, + "learning_rate": 0.00015953100362929408, + "loss": 0.0674, + "step": 12678 + }, + { + "epoch": 0.2972157269525367, + "grad_norm": 0.45508190989494324, + "learning_rate": 0.00015952508478880227, + "loss": 0.0993, + "step": 12679 + }, + { + "epoch": 0.29723916852734167, + "grad_norm": 0.2667246460914612, + "learning_rate": 0.0001595191656253257, + "loss": 0.3171, + "step": 12680 + }, + { + "epoch": 0.29726261010214666, + "grad_norm": 0.3050786554813385, + "learning_rate": 0.0001595132461388965, + "loss": 0.0267, + "step": 12681 + }, + { + "epoch": 0.29728605167695166, + "grad_norm": 0.5019696354866028, + "learning_rate": 0.00015950732632954667, + "loss": 0.117, + "step": 12682 + }, + { + "epoch": 0.29730949325175665, + "grad_norm": 0.4881477355957031, + "learning_rate": 0.00015950140619730847, + "loss": 0.1037, + "step": 12683 + }, + { + "epoch": 0.29733293482656165, + "grad_norm": 0.5356399416923523, + "learning_rate": 0.00015949548574221395, + "loss": 0.1486, + "step": 12684 + }, + { + "epoch": 0.29735637640136664, + "grad_norm": 0.4098850190639496, + "learning_rate": 0.00015948956496429525, + "loss": 0.0783, + "step": 12685 + }, + { + "epoch": 0.29737981797617163, + "grad_norm": 0.7935230135917664, + "learning_rate": 0.00015948364386358452, + "loss": 0.1334, + "step": 12686 + }, + { + "epoch": 0.29740325955097663, + "grad_norm": 0.38513895869255066, + "learning_rate": 0.00015947772244011385, + "loss": 0.0923, + "step": 12687 + }, + { + "epoch": 0.2974267011257816, + "grad_norm": 0.4957554042339325, + "learning_rate": 0.00015947180069391536, + "loss": 0.1386, + "step": 12688 + }, + { + "epoch": 0.2974501427005866, + "grad_norm": 0.20853281021118164, + "learning_rate": 0.00015946587862502126, + "loss": 0.0494, + "step": 12689 + }, + { + "epoch": 0.2974735842753916, + "grad_norm": 0.4319738745689392, + "learning_rate": 0.0001594599562334636, + "loss": 0.056, + "step": 12690 + }, + { + "epoch": 0.2974970258501966, + "grad_norm": 0.22805297374725342, + "learning_rate": 0.00015945403351927455, + "loss": 0.0174, + "step": 12691 + }, + { + "epoch": 0.2975204674250016, + "grad_norm": 0.50872403383255, + "learning_rate": 0.00015944811048248622, + "loss": 0.1234, + "step": 12692 + }, + { + "epoch": 0.2975439089998066, + "grad_norm": 0.5062070488929749, + "learning_rate": 0.0001594421871231308, + "loss": 0.1206, + "step": 12693 + }, + { + "epoch": 0.2975673505746116, + "grad_norm": 0.1844753623008728, + "learning_rate": 0.0001594362634412404, + "loss": 0.0483, + "step": 12694 + }, + { + "epoch": 0.2975907921494166, + "grad_norm": 0.7295498847961426, + "learning_rate": 0.00015943033943684713, + "loss": 0.2023, + "step": 12695 + }, + { + "epoch": 0.2976142337242216, + "grad_norm": 0.7674392461776733, + "learning_rate": 0.0001594244151099832, + "loss": 0.129, + "step": 12696 + }, + { + "epoch": 0.29763767529902657, + "grad_norm": 0.4345487654209137, + "learning_rate": 0.00015941849046068073, + "loss": 0.0899, + "step": 12697 + }, + { + "epoch": 0.29766111687383157, + "grad_norm": 0.6501827836036682, + "learning_rate": 0.00015941256548897186, + "loss": 0.0904, + "step": 12698 + }, + { + "epoch": 0.29768455844863656, + "grad_norm": 0.37594038248062134, + "learning_rate": 0.00015940664019488873, + "loss": 0.0786, + "step": 12699 + }, + { + "epoch": 0.29770800002344155, + "grad_norm": 0.5104688405990601, + "learning_rate": 0.00015940071457846346, + "loss": 0.1582, + "step": 12700 + }, + { + "epoch": 0.29773144159824655, + "grad_norm": 0.4057215452194214, + "learning_rate": 0.0001593947886397283, + "loss": 0.0902, + "step": 12701 + }, + { + "epoch": 0.29775488317305154, + "grad_norm": 0.5018413662910461, + "learning_rate": 0.00015938886237871534, + "loss": 0.7187, + "step": 12702 + }, + { + "epoch": 0.29777832474785654, + "grad_norm": 0.32885751128196716, + "learning_rate": 0.0001593829357954567, + "loss": 0.0687, + "step": 12703 + }, + { + "epoch": 0.29780176632266153, + "grad_norm": 0.6374766230583191, + "learning_rate": 0.0001593770088899846, + "loss": 0.5078, + "step": 12704 + }, + { + "epoch": 0.2978252078974665, + "grad_norm": 0.22657962143421173, + "learning_rate": 0.00015937108166233122, + "loss": 0.043, + "step": 12705 + }, + { + "epoch": 0.2978486494722715, + "grad_norm": 0.19869813323020935, + "learning_rate": 0.00015936515411252864, + "loss": 0.032, + "step": 12706 + }, + { + "epoch": 0.2978720910470765, + "grad_norm": 0.42101815342903137, + "learning_rate": 0.0001593592262406091, + "loss": 0.057, + "step": 12707 + }, + { + "epoch": 0.29789553262188156, + "grad_norm": 0.9351232647895813, + "learning_rate": 0.0001593532980466047, + "loss": 0.1409, + "step": 12708 + }, + { + "epoch": 0.29791897419668656, + "grad_norm": 0.5528275370597839, + "learning_rate": 0.0001593473695305476, + "loss": 0.0918, + "step": 12709 + }, + { + "epoch": 0.29794241577149155, + "grad_norm": 0.08640924096107483, + "learning_rate": 0.00015934144069247006, + "loss": 0.0155, + "step": 12710 + }, + { + "epoch": 0.29796585734629655, + "grad_norm": 0.4882858693599701, + "learning_rate": 0.00015933551153240417, + "loss": 0.0574, + "step": 12711 + }, + { + "epoch": 0.29798929892110154, + "grad_norm": 1.0429540872573853, + "learning_rate": 0.00015932958205038214, + "loss": 0.2135, + "step": 12712 + }, + { + "epoch": 0.29801274049590654, + "grad_norm": 0.552433431148529, + "learning_rate": 0.00015932365224643612, + "loss": 0.1169, + "step": 12713 + }, + { + "epoch": 0.29803618207071153, + "grad_norm": 0.14552216231822968, + "learning_rate": 0.0001593177221205983, + "loss": 0.0296, + "step": 12714 + }, + { + "epoch": 0.2980596236455165, + "grad_norm": 0.2474149614572525, + "learning_rate": 0.00015931179167290088, + "loss": 0.0576, + "step": 12715 + }, + { + "epoch": 0.2980830652203215, + "grad_norm": 0.43205738067626953, + "learning_rate": 0.00015930586090337595, + "loss": 0.1014, + "step": 12716 + }, + { + "epoch": 0.2981065067951265, + "grad_norm": 0.4381226897239685, + "learning_rate": 0.00015929992981205578, + "loss": 0.0831, + "step": 12717 + }, + { + "epoch": 0.2981299483699315, + "grad_norm": 0.44936007261276245, + "learning_rate": 0.0001592939983989725, + "loss": 0.0657, + "step": 12718 + }, + { + "epoch": 0.2981533899447365, + "grad_norm": 0.9024969935417175, + "learning_rate": 0.00015928806666415838, + "loss": 0.7929, + "step": 12719 + }, + { + "epoch": 0.2981768315195415, + "grad_norm": 0.720978319644928, + "learning_rate": 0.0001592821346076455, + "loss": 0.2141, + "step": 12720 + }, + { + "epoch": 0.2982002730943465, + "grad_norm": 0.6903005242347717, + "learning_rate": 0.00015927620222946608, + "loss": 0.1917, + "step": 12721 + }, + { + "epoch": 0.2982237146691515, + "grad_norm": 0.22864410281181335, + "learning_rate": 0.00015927026952965235, + "loss": 0.0451, + "step": 12722 + }, + { + "epoch": 0.2982471562439565, + "grad_norm": 0.6329736709594727, + "learning_rate": 0.00015926433650823648, + "loss": 0.1212, + "step": 12723 + }, + { + "epoch": 0.2982705978187615, + "grad_norm": 0.3673239052295685, + "learning_rate": 0.0001592584031652506, + "loss": 0.074, + "step": 12724 + }, + { + "epoch": 0.29829403939356647, + "grad_norm": 0.6218797564506531, + "learning_rate": 0.000159252469500727, + "loss": 0.181, + "step": 12725 + }, + { + "epoch": 0.29831748096837146, + "grad_norm": 0.9974265098571777, + "learning_rate": 0.00015924653551469784, + "loss": 0.1235, + "step": 12726 + }, + { + "epoch": 0.29834092254317646, + "grad_norm": 0.34995511174201965, + "learning_rate": 0.00015924060120719527, + "loss": 0.0862, + "step": 12727 + }, + { + "epoch": 0.29836436411798145, + "grad_norm": 0.5453566908836365, + "learning_rate": 0.0001592346665782516, + "loss": 0.1487, + "step": 12728 + }, + { + "epoch": 0.29838780569278645, + "grad_norm": 0.2237498164176941, + "learning_rate": 0.00015922873162789892, + "loss": 0.0554, + "step": 12729 + }, + { + "epoch": 0.29841124726759144, + "grad_norm": 1.2989869117736816, + "learning_rate": 0.00015922279635616945, + "loss": 0.1642, + "step": 12730 + }, + { + "epoch": 0.29843468884239643, + "grad_norm": 0.4580506682395935, + "learning_rate": 0.00015921686076309546, + "loss": 0.1201, + "step": 12731 + }, + { + "epoch": 0.29845813041720143, + "grad_norm": 0.3321481943130493, + "learning_rate": 0.0001592109248487091, + "loss": 0.1104, + "step": 12732 + }, + { + "epoch": 0.2984815719920064, + "grad_norm": 0.22388137876987457, + "learning_rate": 0.00015920498861304261, + "loss": 0.0528, + "step": 12733 + }, + { + "epoch": 0.2985050135668114, + "grad_norm": 0.6395736932754517, + "learning_rate": 0.0001591990520561282, + "loss": 0.1097, + "step": 12734 + }, + { + "epoch": 0.2985284551416164, + "grad_norm": 0.5122646689414978, + "learning_rate": 0.00015919311517799804, + "loss": 0.1046, + "step": 12735 + }, + { + "epoch": 0.2985518967164214, + "grad_norm": 0.38962364196777344, + "learning_rate": 0.0001591871779786844, + "loss": 0.1092, + "step": 12736 + }, + { + "epoch": 0.2985753382912264, + "grad_norm": 0.6380918025970459, + "learning_rate": 0.00015918124045821944, + "loss": 0.1777, + "step": 12737 + }, + { + "epoch": 0.2985987798660314, + "grad_norm": 0.5569185614585876, + "learning_rate": 0.00015917530261663542, + "loss": 0.097, + "step": 12738 + }, + { + "epoch": 0.2986222214408364, + "grad_norm": 0.4341198801994324, + "learning_rate": 0.00015916936445396456, + "loss": 0.0921, + "step": 12739 + }, + { + "epoch": 0.2986456630156414, + "grad_norm": 0.5526326298713684, + "learning_rate": 0.00015916342597023904, + "loss": 0.1147, + "step": 12740 + }, + { + "epoch": 0.2986691045904464, + "grad_norm": 0.08772391825914383, + "learning_rate": 0.00015915748716549113, + "loss": 0.0211, + "step": 12741 + }, + { + "epoch": 0.29869254616525137, + "grad_norm": 0.6619718074798584, + "learning_rate": 0.00015915154803975302, + "loss": 0.1397, + "step": 12742 + }, + { + "epoch": 0.29871598774005637, + "grad_norm": 0.2787722945213318, + "learning_rate": 0.0001591456085930569, + "loss": 0.07, + "step": 12743 + }, + { + "epoch": 0.29873942931486136, + "grad_norm": 0.21312177181243896, + "learning_rate": 0.0001591396688254351, + "loss": 0.0374, + "step": 12744 + }, + { + "epoch": 0.29876287088966635, + "grad_norm": 0.18713201582431793, + "learning_rate": 0.00015913372873691982, + "loss": 0.0293, + "step": 12745 + }, + { + "epoch": 0.29878631246447135, + "grad_norm": 0.5385304689407349, + "learning_rate": 0.0001591277883275432, + "loss": 0.7719, + "step": 12746 + }, + { + "epoch": 0.29880975403927634, + "grad_norm": 0.8172031044960022, + "learning_rate": 0.00015912184759733758, + "loss": 0.1611, + "step": 12747 + }, + { + "epoch": 0.29883319561408134, + "grad_norm": 0.14419732987880707, + "learning_rate": 0.00015911590654633513, + "loss": 0.0287, + "step": 12748 + }, + { + "epoch": 0.29885663718888633, + "grad_norm": 0.4849438965320587, + "learning_rate": 0.00015910996517456811, + "loss": 0.1677, + "step": 12749 + }, + { + "epoch": 0.2988800787636913, + "grad_norm": 0.44850555062294006, + "learning_rate": 0.00015910402348206877, + "loss": 0.0722, + "step": 12750 + }, + { + "epoch": 0.2989035203384963, + "grad_norm": 0.22389690577983856, + "learning_rate": 0.00015909808146886934, + "loss": 0.045, + "step": 12751 + }, + { + "epoch": 0.2989269619133013, + "grad_norm": 0.20034050941467285, + "learning_rate": 0.00015909213913500202, + "loss": 0.0286, + "step": 12752 + }, + { + "epoch": 0.2989504034881063, + "grad_norm": 0.4821363091468811, + "learning_rate": 0.00015908619648049915, + "loss": 0.0689, + "step": 12753 + }, + { + "epoch": 0.2989738450629113, + "grad_norm": 1.098353385925293, + "learning_rate": 0.0001590802535053929, + "loss": 0.2194, + "step": 12754 + }, + { + "epoch": 0.2989972866377163, + "grad_norm": 0.10954206436872482, + "learning_rate": 0.00015907431020971552, + "loss": 0.029, + "step": 12755 + }, + { + "epoch": 0.2990207282125213, + "grad_norm": 0.15905150771141052, + "learning_rate": 0.00015906836659349924, + "loss": 0.0397, + "step": 12756 + }, + { + "epoch": 0.2990441697873263, + "grad_norm": 0.7659737467765808, + "learning_rate": 0.00015906242265677637, + "loss": 0.0856, + "step": 12757 + }, + { + "epoch": 0.2990676113621313, + "grad_norm": 0.26199832558631897, + "learning_rate": 0.00015905647839957915, + "loss": 0.0604, + "step": 12758 + }, + { + "epoch": 0.29909105293693633, + "grad_norm": 0.15994253754615784, + "learning_rate": 0.00015905053382193982, + "loss": 0.0287, + "step": 12759 + }, + { + "epoch": 0.2991144945117413, + "grad_norm": 0.8442479372024536, + "learning_rate": 0.00015904458892389062, + "loss": 0.1923, + "step": 12760 + }, + { + "epoch": 0.2991379360865463, + "grad_norm": 0.28909987211227417, + "learning_rate": 0.0001590386437054638, + "loss": 0.3351, + "step": 12761 + }, + { + "epoch": 0.2991613776613513, + "grad_norm": 0.7711607217788696, + "learning_rate": 0.00015903269816669167, + "loss": 0.0924, + "step": 12762 + }, + { + "epoch": 0.2991848192361563, + "grad_norm": 0.7785350680351257, + "learning_rate": 0.00015902675230760645, + "loss": 0.0998, + "step": 12763 + }, + { + "epoch": 0.2992082608109613, + "grad_norm": 0.7204457521438599, + "learning_rate": 0.00015902080612824042, + "loss": 0.1376, + "step": 12764 + }, + { + "epoch": 0.2992317023857663, + "grad_norm": 0.6098506450653076, + "learning_rate": 0.00015901485962862584, + "loss": 0.1957, + "step": 12765 + }, + { + "epoch": 0.2992551439605713, + "grad_norm": 0.4770206809043884, + "learning_rate": 0.00015900891280879495, + "loss": 0.1077, + "step": 12766 + }, + { + "epoch": 0.2992785855353763, + "grad_norm": 1.0204719305038452, + "learning_rate": 0.00015900296566878, + "loss": 0.1701, + "step": 12767 + }, + { + "epoch": 0.2993020271101813, + "grad_norm": 0.22730514407157898, + "learning_rate": 0.00015899701820861338, + "loss": 0.0472, + "step": 12768 + }, + { + "epoch": 0.2993254686849863, + "grad_norm": 0.7658883333206177, + "learning_rate": 0.00015899107042832724, + "loss": 0.1474, + "step": 12769 + }, + { + "epoch": 0.29934891025979127, + "grad_norm": 0.2564321756362915, + "learning_rate": 0.00015898512232795392, + "loss": 0.0463, + "step": 12770 + }, + { + "epoch": 0.29937235183459626, + "grad_norm": 0.34035056829452515, + "learning_rate": 0.00015897917390752565, + "loss": 0.0796, + "step": 12771 + }, + { + "epoch": 0.29939579340940126, + "grad_norm": 0.47875383496284485, + "learning_rate": 0.0001589732251670747, + "loss": 0.5353, + "step": 12772 + }, + { + "epoch": 0.29941923498420625, + "grad_norm": 0.7775318026542664, + "learning_rate": 0.00015896727610663342, + "loss": 0.1678, + "step": 12773 + }, + { + "epoch": 0.29944267655901124, + "grad_norm": 0.6356266140937805, + "learning_rate": 0.00015896132672623404, + "loss": 0.1156, + "step": 12774 + }, + { + "epoch": 0.29946611813381624, + "grad_norm": 0.5083867311477661, + "learning_rate": 0.0001589553770259088, + "loss": 0.6738, + "step": 12775 + }, + { + "epoch": 0.29948955970862123, + "grad_norm": 0.23826977610588074, + "learning_rate": 0.00015894942700569006, + "loss": 0.0581, + "step": 12776 + }, + { + "epoch": 0.29951300128342623, + "grad_norm": 0.13413959741592407, + "learning_rate": 0.00015894347666561004, + "loss": 0.0281, + "step": 12777 + }, + { + "epoch": 0.2995364428582312, + "grad_norm": 0.6547167301177979, + "learning_rate": 0.0001589375260057011, + "loss": 0.1261, + "step": 12778 + }, + { + "epoch": 0.2995598844330362, + "grad_norm": 0.2380829006433487, + "learning_rate": 0.00015893157502599542, + "loss": 0.0187, + "step": 12779 + }, + { + "epoch": 0.2995833260078412, + "grad_norm": 0.3649323284626007, + "learning_rate": 0.00015892562372652543, + "loss": 0.1001, + "step": 12780 + }, + { + "epoch": 0.2996067675826462, + "grad_norm": 0.6621320843696594, + "learning_rate": 0.00015891967210732332, + "loss": 0.4852, + "step": 12781 + }, + { + "epoch": 0.2996302091574512, + "grad_norm": 0.6483266949653625, + "learning_rate": 0.0001589137201684214, + "loss": 0.7136, + "step": 12782 + }, + { + "epoch": 0.2996536507322562, + "grad_norm": 0.39267697930336, + "learning_rate": 0.00015890776790985201, + "loss": 0.1065, + "step": 12783 + }, + { + "epoch": 0.2996770923070612, + "grad_norm": 0.10654965043067932, + "learning_rate": 0.0001589018153316474, + "loss": 0.0165, + "step": 12784 + }, + { + "epoch": 0.2997005338818662, + "grad_norm": 0.5090387463569641, + "learning_rate": 0.00015889586243383988, + "loss": 0.1229, + "step": 12785 + }, + { + "epoch": 0.2997239754566712, + "grad_norm": 0.5655834078788757, + "learning_rate": 0.00015888990921646175, + "loss": 0.1146, + "step": 12786 + }, + { + "epoch": 0.29974741703147617, + "grad_norm": 0.4360744059085846, + "learning_rate": 0.0001588839556795453, + "loss": 0.0941, + "step": 12787 + }, + { + "epoch": 0.29977085860628117, + "grad_norm": 0.4781060814857483, + "learning_rate": 0.0001588780018231229, + "loss": 0.1178, + "step": 12788 + }, + { + "epoch": 0.29979430018108616, + "grad_norm": 0.5374126434326172, + "learning_rate": 0.00015887204764722677, + "loss": 0.1015, + "step": 12789 + }, + { + "epoch": 0.29981774175589115, + "grad_norm": 0.3985086977481842, + "learning_rate": 0.00015886609315188924, + "loss": 0.1334, + "step": 12790 + }, + { + "epoch": 0.29984118333069615, + "grad_norm": 0.6046473383903503, + "learning_rate": 0.00015886013833714267, + "loss": 0.0465, + "step": 12791 + }, + { + "epoch": 0.29986462490550114, + "grad_norm": 0.12377425283193588, + "learning_rate": 0.0001588541832030193, + "loss": 0.0147, + "step": 12792 + }, + { + "epoch": 0.29988806648030614, + "grad_norm": 0.43439048528671265, + "learning_rate": 0.0001588482277495515, + "loss": 0.0847, + "step": 12793 + }, + { + "epoch": 0.29991150805511113, + "grad_norm": 0.90751051902771, + "learning_rate": 0.00015884227197677158, + "loss": 0.2235, + "step": 12794 + }, + { + "epoch": 0.2999349496299161, + "grad_norm": 0.4843693971633911, + "learning_rate": 0.0001588363158847118, + "loss": 0.0488, + "step": 12795 + }, + { + "epoch": 0.2999583912047211, + "grad_norm": 0.1849023997783661, + "learning_rate": 0.00015883035947340454, + "loss": 0.0316, + "step": 12796 + }, + { + "epoch": 0.2999818327795261, + "grad_norm": 0.23939470946788788, + "learning_rate": 0.00015882440274288205, + "loss": 0.0596, + "step": 12797 + }, + { + "epoch": 0.3000052743543311, + "grad_norm": 0.37508857250213623, + "learning_rate": 0.00015881844569317678, + "loss": 0.0484, + "step": 12798 + }, + { + "epoch": 0.3000287159291361, + "grad_norm": 0.4775731563568115, + "learning_rate": 0.0001588124883243209, + "loss": 0.0536, + "step": 12799 + }, + { + "epoch": 0.3000521575039411, + "grad_norm": 0.4977845251560211, + "learning_rate": 0.00015880653063634682, + "loss": 0.1684, + "step": 12800 + }, + { + "epoch": 0.3000755990787461, + "grad_norm": 0.669281542301178, + "learning_rate": 0.00015880057262928688, + "loss": 0.1898, + "step": 12801 + }, + { + "epoch": 0.3000990406535511, + "grad_norm": 0.34395894408226013, + "learning_rate": 0.00015879461430317338, + "loss": 0.1047, + "step": 12802 + }, + { + "epoch": 0.3001224822283561, + "grad_norm": 0.4607917368412018, + "learning_rate": 0.0001587886556580386, + "loss": 0.6703, + "step": 12803 + }, + { + "epoch": 0.3001459238031611, + "grad_norm": 0.5773283243179321, + "learning_rate": 0.00015878269669391493, + "loss": 0.1313, + "step": 12804 + }, + { + "epoch": 0.30016936537796607, + "grad_norm": 0.8519477844238281, + "learning_rate": 0.00015877673741083475, + "loss": 0.2187, + "step": 12805 + }, + { + "epoch": 0.30019280695277106, + "grad_norm": 0.22015611827373505, + "learning_rate": 0.00015877077780883028, + "loss": 0.0653, + "step": 12806 + }, + { + "epoch": 0.30021624852757606, + "grad_norm": 0.7480697631835938, + "learning_rate": 0.00015876481788793396, + "loss": 0.1395, + "step": 12807 + }, + { + "epoch": 0.30023969010238105, + "grad_norm": 0.3203582167625427, + "learning_rate": 0.0001587588576481781, + "loss": 0.0915, + "step": 12808 + }, + { + "epoch": 0.30026313167718605, + "grad_norm": 0.6284352540969849, + "learning_rate": 0.00015875289708959498, + "loss": 0.6606, + "step": 12809 + }, + { + "epoch": 0.3002865732519911, + "grad_norm": 0.29750531911849976, + "learning_rate": 0.00015874693621221696, + "loss": 0.067, + "step": 12810 + }, + { + "epoch": 0.3003100148267961, + "grad_norm": 0.5213223099708557, + "learning_rate": 0.00015874097501607646, + "loss": 0.099, + "step": 12811 + }, + { + "epoch": 0.3003334564016011, + "grad_norm": 0.9848500490188599, + "learning_rate": 0.00015873501350120579, + "loss": 0.1305, + "step": 12812 + }, + { + "epoch": 0.3003568979764061, + "grad_norm": 0.46449121832847595, + "learning_rate": 0.00015872905166763727, + "loss": 0.1817, + "step": 12813 + }, + { + "epoch": 0.3003803395512111, + "grad_norm": 0.5618071556091309, + "learning_rate": 0.00015872308951540327, + "loss": 0.1786, + "step": 12814 + }, + { + "epoch": 0.30040378112601607, + "grad_norm": 0.21046461164951324, + "learning_rate": 0.00015871712704453614, + "loss": 0.0377, + "step": 12815 + }, + { + "epoch": 0.30042722270082106, + "grad_norm": 0.741597056388855, + "learning_rate": 0.00015871116425506817, + "loss": 0.2389, + "step": 12816 + }, + { + "epoch": 0.30045066427562606, + "grad_norm": 0.529133141040802, + "learning_rate": 0.00015870520114703184, + "loss": 0.0682, + "step": 12817 + }, + { + "epoch": 0.30047410585043105, + "grad_norm": 0.365027517080307, + "learning_rate": 0.0001586992377204594, + "loss": 0.1072, + "step": 12818 + }, + { + "epoch": 0.30049754742523604, + "grad_norm": 0.24577802419662476, + "learning_rate": 0.0001586932739753833, + "loss": 0.0357, + "step": 12819 + }, + { + "epoch": 0.30052098900004104, + "grad_norm": 0.3239877223968506, + "learning_rate": 0.00015868730991183577, + "loss": 0.081, + "step": 12820 + }, + { + "epoch": 0.30054443057484603, + "grad_norm": 0.6454901099205017, + "learning_rate": 0.0001586813455298493, + "loss": 0.1137, + "step": 12821 + }, + { + "epoch": 0.300567872149651, + "grad_norm": 0.5647878050804138, + "learning_rate": 0.00015867538082945618, + "loss": 0.1063, + "step": 12822 + }, + { + "epoch": 0.300591313724456, + "grad_norm": 0.16484063863754272, + "learning_rate": 0.0001586694158106888, + "loss": 0.0372, + "step": 12823 + }, + { + "epoch": 0.300614755299261, + "grad_norm": 0.13163769245147705, + "learning_rate": 0.00015866345047357952, + "loss": 0.0323, + "step": 12824 + }, + { + "epoch": 0.300638196874066, + "grad_norm": 0.48961177468299866, + "learning_rate": 0.0001586574848181607, + "loss": 0.1302, + "step": 12825 + }, + { + "epoch": 0.300661638448871, + "grad_norm": 0.9726395606994629, + "learning_rate": 0.00015865151884446473, + "loss": 0.1812, + "step": 12826 + }, + { + "epoch": 0.300685080023676, + "grad_norm": 0.46793535351753235, + "learning_rate": 0.00015864555255252396, + "loss": 0.8819, + "step": 12827 + }, + { + "epoch": 0.300708521598481, + "grad_norm": 0.11788898706436157, + "learning_rate": 0.0001586395859423708, + "loss": 0.0363, + "step": 12828 + }, + { + "epoch": 0.300731963173286, + "grad_norm": 0.5930625796318054, + "learning_rate": 0.00015863361901403752, + "loss": 0.1797, + "step": 12829 + }, + { + "epoch": 0.300755404748091, + "grad_norm": 0.5138011574745178, + "learning_rate": 0.00015862765176755664, + "loss": 0.1118, + "step": 12830 + }, + { + "epoch": 0.300778846322896, + "grad_norm": 0.24447618424892426, + "learning_rate": 0.00015862168420296047, + "loss": 0.0427, + "step": 12831 + }, + { + "epoch": 0.30080228789770097, + "grad_norm": 0.6459386944770813, + "learning_rate": 0.00015861571632028138, + "loss": 0.1362, + "step": 12832 + }, + { + "epoch": 0.30082572947250596, + "grad_norm": 0.48154133558273315, + "learning_rate": 0.00015860974811955178, + "loss": 0.1112, + "step": 12833 + }, + { + "epoch": 0.30084917104731096, + "grad_norm": 0.40304240584373474, + "learning_rate": 0.00015860377960080402, + "loss": 0.1079, + "step": 12834 + }, + { + "epoch": 0.30087261262211595, + "grad_norm": 0.3674716651439667, + "learning_rate": 0.00015859781076407053, + "loss": 0.5934, + "step": 12835 + }, + { + "epoch": 0.30089605419692095, + "grad_norm": 0.4628383219242096, + "learning_rate": 0.00015859184160938364, + "loss": 0.1247, + "step": 12836 + }, + { + "epoch": 0.30091949577172594, + "grad_norm": 0.1432303935289383, + "learning_rate": 0.0001585858721367758, + "loss": 0.0219, + "step": 12837 + }, + { + "epoch": 0.30094293734653094, + "grad_norm": 0.1293913722038269, + "learning_rate": 0.00015857990234627932, + "loss": 0.0328, + "step": 12838 + }, + { + "epoch": 0.30096637892133593, + "grad_norm": 0.531825065612793, + "learning_rate": 0.00015857393223792668, + "loss": 0.094, + "step": 12839 + }, + { + "epoch": 0.3009898204961409, + "grad_norm": 0.8584846258163452, + "learning_rate": 0.00015856796181175025, + "loss": 0.1031, + "step": 12840 + }, + { + "epoch": 0.3010132620709459, + "grad_norm": 0.49288398027420044, + "learning_rate": 0.00015856199106778238, + "loss": 0.1179, + "step": 12841 + }, + { + "epoch": 0.3010367036457509, + "grad_norm": 0.7169408798217773, + "learning_rate": 0.0001585560200060555, + "loss": 0.1051, + "step": 12842 + }, + { + "epoch": 0.3010601452205559, + "grad_norm": 0.5031992793083191, + "learning_rate": 0.00015855004862660204, + "loss": 0.5306, + "step": 12843 + }, + { + "epoch": 0.3010835867953609, + "grad_norm": 0.1440614014863968, + "learning_rate": 0.00015854407692945434, + "loss": 0.0235, + "step": 12844 + }, + { + "epoch": 0.3011070283701659, + "grad_norm": 0.27675995230674744, + "learning_rate": 0.00015853810491464488, + "loss": 0.0544, + "step": 12845 + }, + { + "epoch": 0.3011304699449709, + "grad_norm": 0.39565184712409973, + "learning_rate": 0.00015853213258220595, + "loss": 0.159, + "step": 12846 + }, + { + "epoch": 0.3011539115197759, + "grad_norm": 0.8677770495414734, + "learning_rate": 0.00015852615993217005, + "loss": 0.1235, + "step": 12847 + }, + { + "epoch": 0.3011773530945809, + "grad_norm": 0.5900951027870178, + "learning_rate": 0.00015852018696456958, + "loss": 0.609, + "step": 12848 + }, + { + "epoch": 0.3012007946693859, + "grad_norm": 0.5720300674438477, + "learning_rate": 0.00015851421367943692, + "loss": 0.1245, + "step": 12849 + }, + { + "epoch": 0.30122423624419087, + "grad_norm": 0.6106486320495605, + "learning_rate": 0.00015850824007680449, + "loss": 0.1181, + "step": 12850 + }, + { + "epoch": 0.30124767781899586, + "grad_norm": 0.4762175977230072, + "learning_rate": 0.0001585022661567047, + "loss": 0.135, + "step": 12851 + }, + { + "epoch": 0.30127111939380086, + "grad_norm": 0.3630441427230835, + "learning_rate": 0.00015849629191916996, + "loss": 0.0588, + "step": 12852 + }, + { + "epoch": 0.30129456096860585, + "grad_norm": 0.4838625192642212, + "learning_rate": 0.0001584903173642327, + "loss": 0.1042, + "step": 12853 + }, + { + "epoch": 0.30131800254341085, + "grad_norm": 0.15697582066059113, + "learning_rate": 0.00015848434249192533, + "loss": 0.0314, + "step": 12854 + }, + { + "epoch": 0.30134144411821584, + "grad_norm": 0.22619779407978058, + "learning_rate": 0.00015847836730228032, + "loss": 0.0651, + "step": 12855 + }, + { + "epoch": 0.30136488569302083, + "grad_norm": 0.5965614318847656, + "learning_rate": 0.00015847239179533, + "loss": 0.0694, + "step": 12856 + }, + { + "epoch": 0.30138832726782583, + "grad_norm": 0.4838550388813019, + "learning_rate": 0.00015846641597110685, + "loss": 0.0975, + "step": 12857 + }, + { + "epoch": 0.3014117688426308, + "grad_norm": 0.19887152314186096, + "learning_rate": 0.00015846043982964324, + "loss": 0.0541, + "step": 12858 + }, + { + "epoch": 0.3014352104174358, + "grad_norm": 0.4191058576107025, + "learning_rate": 0.0001584544633709717, + "loss": 0.0628, + "step": 12859 + }, + { + "epoch": 0.3014586519922408, + "grad_norm": 0.5692657232284546, + "learning_rate": 0.00015844848659512457, + "loss": 0.0821, + "step": 12860 + }, + { + "epoch": 0.30148209356704586, + "grad_norm": 0.7061968445777893, + "learning_rate": 0.00015844250950213432, + "loss": 0.1748, + "step": 12861 + }, + { + "epoch": 0.30150553514185086, + "grad_norm": 0.12848149240016937, + "learning_rate": 0.00015843653209203338, + "loss": 0.0157, + "step": 12862 + }, + { + "epoch": 0.30152897671665585, + "grad_norm": 0.4911331534385681, + "learning_rate": 0.00015843055436485414, + "loss": 0.1084, + "step": 12863 + }, + { + "epoch": 0.30155241829146084, + "grad_norm": 0.4573552906513214, + "learning_rate": 0.00015842457632062908, + "loss": 0.1502, + "step": 12864 + }, + { + "epoch": 0.30157585986626584, + "grad_norm": 0.6586965918540955, + "learning_rate": 0.00015841859795939065, + "loss": 0.1371, + "step": 12865 + }, + { + "epoch": 0.30159930144107083, + "grad_norm": 0.6591430902481079, + "learning_rate": 0.00015841261928117124, + "loss": 0.1389, + "step": 12866 + }, + { + "epoch": 0.3016227430158758, + "grad_norm": 0.42237141728401184, + "learning_rate": 0.00015840664028600335, + "loss": 0.3323, + "step": 12867 + }, + { + "epoch": 0.3016461845906808, + "grad_norm": 0.40482938289642334, + "learning_rate": 0.00015840066097391934, + "loss": 0.0707, + "step": 12868 + }, + { + "epoch": 0.3016696261654858, + "grad_norm": 0.21978460252285004, + "learning_rate": 0.00015839468134495172, + "loss": 0.0313, + "step": 12869 + }, + { + "epoch": 0.3016930677402908, + "grad_norm": 0.4580993056297302, + "learning_rate": 0.00015838870139913292, + "loss": 0.0667, + "step": 12870 + }, + { + "epoch": 0.3017165093150958, + "grad_norm": 0.4466887414455414, + "learning_rate": 0.0001583827211364954, + "loss": 0.0796, + "step": 12871 + }, + { + "epoch": 0.3017399508899008, + "grad_norm": 0.5050243139266968, + "learning_rate": 0.00015837674055707156, + "loss": 0.1013, + "step": 12872 + }, + { + "epoch": 0.3017633924647058, + "grad_norm": 0.2111518383026123, + "learning_rate": 0.0001583707596608939, + "loss": 0.0505, + "step": 12873 + }, + { + "epoch": 0.3017868340395108, + "grad_norm": 0.6651328802108765, + "learning_rate": 0.00015836477844799487, + "loss": 0.1113, + "step": 12874 + }, + { + "epoch": 0.3018102756143158, + "grad_norm": 0.16768716275691986, + "learning_rate": 0.0001583587969184069, + "loss": 0.0165, + "step": 12875 + }, + { + "epoch": 0.3018337171891208, + "grad_norm": 0.15469539165496826, + "learning_rate": 0.00015835281507216244, + "loss": 0.0221, + "step": 12876 + }, + { + "epoch": 0.30185715876392577, + "grad_norm": 0.42456135153770447, + "learning_rate": 0.000158346832909294, + "loss": 0.0649, + "step": 12877 + }, + { + "epoch": 0.30188060033873076, + "grad_norm": 0.23329272866249084, + "learning_rate": 0.000158340850429834, + "loss": 0.0233, + "step": 12878 + }, + { + "epoch": 0.30190404191353576, + "grad_norm": 0.26721158623695374, + "learning_rate": 0.00015833486763381486, + "loss": 0.0521, + "step": 12879 + }, + { + "epoch": 0.30192748348834075, + "grad_norm": 0.1660016030073166, + "learning_rate": 0.00015832888452126914, + "loss": 0.0255, + "step": 12880 + }, + { + "epoch": 0.30195092506314575, + "grad_norm": 0.6511425971984863, + "learning_rate": 0.00015832290109222923, + "loss": 0.1043, + "step": 12881 + }, + { + "epoch": 0.30197436663795074, + "grad_norm": 0.5589607954025269, + "learning_rate": 0.0001583169173467276, + "loss": 0.0569, + "step": 12882 + }, + { + "epoch": 0.30199780821275574, + "grad_norm": 0.17418549954891205, + "learning_rate": 0.00015831093328479677, + "loss": 0.032, + "step": 12883 + }, + { + "epoch": 0.30202124978756073, + "grad_norm": 0.6183736324310303, + "learning_rate": 0.00015830494890646917, + "loss": 0.0698, + "step": 12884 + }, + { + "epoch": 0.3020446913623657, + "grad_norm": 0.5348994135856628, + "learning_rate": 0.00015829896421177728, + "loss": 0.1181, + "step": 12885 + }, + { + "epoch": 0.3020681329371707, + "grad_norm": 0.6170745491981506, + "learning_rate": 0.00015829297920075356, + "loss": 0.0743, + "step": 12886 + }, + { + "epoch": 0.3020915745119757, + "grad_norm": 0.3776453137397766, + "learning_rate": 0.00015828699387343052, + "loss": 0.0302, + "step": 12887 + }, + { + "epoch": 0.3021150160867807, + "grad_norm": 0.5191347599029541, + "learning_rate": 0.00015828100822984055, + "loss": 0.0926, + "step": 12888 + }, + { + "epoch": 0.3021384576615857, + "grad_norm": 0.7517353296279907, + "learning_rate": 0.00015827502227001622, + "loss": 0.2203, + "step": 12889 + }, + { + "epoch": 0.3021618992363907, + "grad_norm": 0.5277005434036255, + "learning_rate": 0.00015826903599399001, + "loss": 0.1548, + "step": 12890 + }, + { + "epoch": 0.3021853408111957, + "grad_norm": 0.4983639121055603, + "learning_rate": 0.00015826304940179435, + "loss": 0.1002, + "step": 12891 + }, + { + "epoch": 0.3022087823860007, + "grad_norm": 0.16695000231266022, + "learning_rate": 0.00015825706249346174, + "loss": 0.0326, + "step": 12892 + }, + { + "epoch": 0.3022322239608057, + "grad_norm": 0.4456450939178467, + "learning_rate": 0.00015825107526902472, + "loss": 0.1173, + "step": 12893 + }, + { + "epoch": 0.3022556655356107, + "grad_norm": 0.33013495802879333, + "learning_rate": 0.00015824508772851567, + "loss": 0.0768, + "step": 12894 + }, + { + "epoch": 0.30227910711041567, + "grad_norm": 0.5084787607192993, + "learning_rate": 0.00015823909987196716, + "loss": 0.1685, + "step": 12895 + }, + { + "epoch": 0.30230254868522066, + "grad_norm": 0.23439054191112518, + "learning_rate": 0.00015823311169941163, + "loss": 0.0487, + "step": 12896 + }, + { + "epoch": 0.30232599026002566, + "grad_norm": 0.6281304359436035, + "learning_rate": 0.00015822712321088162, + "loss": 0.1505, + "step": 12897 + }, + { + "epoch": 0.30234943183483065, + "grad_norm": 0.31229910254478455, + "learning_rate": 0.00015822113440640961, + "loss": 0.1144, + "step": 12898 + }, + { + "epoch": 0.30237287340963565, + "grad_norm": 0.30305254459381104, + "learning_rate": 0.0001582151452860281, + "loss": 0.0494, + "step": 12899 + }, + { + "epoch": 0.30239631498444064, + "grad_norm": 0.5427597165107727, + "learning_rate": 0.00015820915584976953, + "loss": 0.0611, + "step": 12900 + }, + { + "epoch": 0.30241975655924563, + "grad_norm": 0.48605167865753174, + "learning_rate": 0.00015820316609766648, + "loss": 0.088, + "step": 12901 + }, + { + "epoch": 0.30244319813405063, + "grad_norm": 0.3741571307182312, + "learning_rate": 0.0001581971760297514, + "loss": 0.0685, + "step": 12902 + }, + { + "epoch": 0.3024666397088556, + "grad_norm": 0.33359792828559875, + "learning_rate": 0.0001581911856460568, + "loss": 0.0527, + "step": 12903 + }, + { + "epoch": 0.3024900812836606, + "grad_norm": 0.34567829966545105, + "learning_rate": 0.00015818519494661522, + "loss": 0.0499, + "step": 12904 + }, + { + "epoch": 0.3025135228584656, + "grad_norm": 0.7053372263908386, + "learning_rate": 0.0001581792039314591, + "loss": 0.1157, + "step": 12905 + }, + { + "epoch": 0.3025369644332706, + "grad_norm": 0.5654634833335876, + "learning_rate": 0.000158173212600621, + "loss": 0.1158, + "step": 12906 + }, + { + "epoch": 0.3025604060080756, + "grad_norm": 0.3816078007221222, + "learning_rate": 0.0001581672209541334, + "loss": 0.0875, + "step": 12907 + }, + { + "epoch": 0.3025838475828806, + "grad_norm": 0.7664957642555237, + "learning_rate": 0.00015816122899202885, + "loss": 0.1683, + "step": 12908 + }, + { + "epoch": 0.3026072891576856, + "grad_norm": 0.6909460425376892, + "learning_rate": 0.00015815523671433983, + "loss": 0.2208, + "step": 12909 + }, + { + "epoch": 0.3026307307324906, + "grad_norm": 0.10011948645114899, + "learning_rate": 0.00015814924412109886, + "loss": 0.017, + "step": 12910 + }, + { + "epoch": 0.3026541723072956, + "grad_norm": 0.3071857988834381, + "learning_rate": 0.00015814325121233844, + "loss": 0.0293, + "step": 12911 + }, + { + "epoch": 0.3026776138821006, + "grad_norm": 0.720521092414856, + "learning_rate": 0.0001581372579880911, + "loss": 0.8184, + "step": 12912 + }, + { + "epoch": 0.3027010554569056, + "grad_norm": 0.1236601322889328, + "learning_rate": 0.0001581312644483894, + "loss": 0.0179, + "step": 12913 + }, + { + "epoch": 0.3027244970317106, + "grad_norm": 0.7987655401229858, + "learning_rate": 0.0001581252705932658, + "loss": 0.1584, + "step": 12914 + }, + { + "epoch": 0.3027479386065156, + "grad_norm": 0.3213689625263214, + "learning_rate": 0.00015811927642275285, + "loss": 0.0969, + "step": 12915 + }, + { + "epoch": 0.3027713801813206, + "grad_norm": 0.6004984378814697, + "learning_rate": 0.00015811328193688306, + "loss": 0.6284, + "step": 12916 + }, + { + "epoch": 0.3027948217561256, + "grad_norm": 0.5041795372962952, + "learning_rate": 0.00015810728713568897, + "loss": 0.0823, + "step": 12917 + }, + { + "epoch": 0.3028182633309306, + "grad_norm": 0.7809000611305237, + "learning_rate": 0.00015810129201920315, + "loss": 0.1288, + "step": 12918 + }, + { + "epoch": 0.3028417049057356, + "grad_norm": 0.37874487042427063, + "learning_rate": 0.000158095296587458, + "loss": 0.0954, + "step": 12919 + }, + { + "epoch": 0.3028651464805406, + "grad_norm": 0.47603240609169006, + "learning_rate": 0.00015808930084048621, + "loss": 0.1101, + "step": 12920 + }, + { + "epoch": 0.3028885880553456, + "grad_norm": 0.42069539427757263, + "learning_rate": 0.0001580833047783202, + "loss": 0.0572, + "step": 12921 + }, + { + "epoch": 0.30291202963015057, + "grad_norm": 0.3300012946128845, + "learning_rate": 0.00015807730840099258, + "loss": 0.0631, + "step": 12922 + }, + { + "epoch": 0.30293547120495556, + "grad_norm": 0.6641491651535034, + "learning_rate": 0.00015807131170853585, + "loss": 0.1469, + "step": 12923 + }, + { + "epoch": 0.30295891277976056, + "grad_norm": 0.8680782318115234, + "learning_rate": 0.00015806531470098252, + "loss": 0.2175, + "step": 12924 + }, + { + "epoch": 0.30298235435456555, + "grad_norm": 0.682295560836792, + "learning_rate": 0.00015805931737836518, + "loss": 0.1781, + "step": 12925 + }, + { + "epoch": 0.30300579592937055, + "grad_norm": 0.559569239616394, + "learning_rate": 0.00015805331974071634, + "loss": 0.2129, + "step": 12926 + }, + { + "epoch": 0.30302923750417554, + "grad_norm": 0.4870198965072632, + "learning_rate": 0.00015804732178806855, + "loss": 0.1234, + "step": 12927 + }, + { + "epoch": 0.30305267907898054, + "grad_norm": 0.24414201080799103, + "learning_rate": 0.0001580413235204544, + "loss": 0.0516, + "step": 12928 + }, + { + "epoch": 0.30307612065378553, + "grad_norm": 0.5425345301628113, + "learning_rate": 0.00015803532493790634, + "loss": 0.0649, + "step": 12929 + }, + { + "epoch": 0.3030995622285905, + "grad_norm": 0.2634201645851135, + "learning_rate": 0.00015802932604045702, + "loss": 0.0704, + "step": 12930 + }, + { + "epoch": 0.3031230038033955, + "grad_norm": 0.23662176728248596, + "learning_rate": 0.0001580233268281389, + "loss": 0.0509, + "step": 12931 + }, + { + "epoch": 0.3031464453782005, + "grad_norm": 0.37433475255966187, + "learning_rate": 0.00015801732730098462, + "loss": 0.0858, + "step": 12932 + }, + { + "epoch": 0.3031698869530055, + "grad_norm": 0.5551442503929138, + "learning_rate": 0.0001580113274590267, + "loss": 0.1574, + "step": 12933 + }, + { + "epoch": 0.3031933285278105, + "grad_norm": 0.6667510271072388, + "learning_rate": 0.0001580053273022976, + "loss": 0.122, + "step": 12934 + }, + { + "epoch": 0.3032167701026155, + "grad_norm": 0.42351940274238586, + "learning_rate": 0.00015799932683083006, + "loss": 0.1497, + "step": 12935 + }, + { + "epoch": 0.3032402116774205, + "grad_norm": 0.21217790246009827, + "learning_rate": 0.00015799332604465651, + "loss": 0.0579, + "step": 12936 + }, + { + "epoch": 0.3032636532522255, + "grad_norm": 0.6308515071868896, + "learning_rate": 0.00015798732494380952, + "loss": 0.1006, + "step": 12937 + }, + { + "epoch": 0.3032870948270305, + "grad_norm": 0.4300665557384491, + "learning_rate": 0.00015798132352832172, + "loss": 0.0955, + "step": 12938 + }, + { + "epoch": 0.3033105364018355, + "grad_norm": 0.507297933101654, + "learning_rate": 0.00015797532179822558, + "loss": 0.1378, + "step": 12939 + }, + { + "epoch": 0.30333397797664047, + "grad_norm": 0.7432830929756165, + "learning_rate": 0.00015796931975355376, + "loss": 0.6603, + "step": 12940 + }, + { + "epoch": 0.30335741955144546, + "grad_norm": 0.2975146174430847, + "learning_rate": 0.00015796331739433874, + "loss": 0.0618, + "step": 12941 + }, + { + "epoch": 0.30338086112625046, + "grad_norm": 0.3799106478691101, + "learning_rate": 0.00015795731472061317, + "loss": 0.0782, + "step": 12942 + }, + { + "epoch": 0.30340430270105545, + "grad_norm": 0.48156315088272095, + "learning_rate": 0.00015795131173240953, + "loss": 0.1048, + "step": 12943 + }, + { + "epoch": 0.30342774427586044, + "grad_norm": 0.18033188581466675, + "learning_rate": 0.00015794530842976048, + "loss": 0.0301, + "step": 12944 + }, + { + "epoch": 0.30345118585066544, + "grad_norm": 0.41799694299697876, + "learning_rate": 0.00015793930481269855, + "loss": 0.0898, + "step": 12945 + }, + { + "epoch": 0.30347462742547043, + "grad_norm": 0.12916308641433716, + "learning_rate": 0.00015793330088125632, + "loss": 0.0416, + "step": 12946 + }, + { + "epoch": 0.30349806900027543, + "grad_norm": 0.12249912321567535, + "learning_rate": 0.00015792729663546636, + "loss": 0.0262, + "step": 12947 + }, + { + "epoch": 0.3035215105750804, + "grad_norm": 0.38328519463539124, + "learning_rate": 0.00015792129207536126, + "loss": 0.1249, + "step": 12948 + }, + { + "epoch": 0.3035449521498854, + "grad_norm": 0.5000482201576233, + "learning_rate": 0.00015791528720097362, + "loss": 0.1448, + "step": 12949 + }, + { + "epoch": 0.3035683937246904, + "grad_norm": 0.3854873478412628, + "learning_rate": 0.00015790928201233598, + "loss": 0.0637, + "step": 12950 + }, + { + "epoch": 0.3035918352994954, + "grad_norm": 0.3523722290992737, + "learning_rate": 0.00015790327650948096, + "loss": 0.0975, + "step": 12951 + }, + { + "epoch": 0.3036152768743004, + "grad_norm": 0.49409133195877075, + "learning_rate": 0.00015789727069244117, + "loss": 0.1333, + "step": 12952 + }, + { + "epoch": 0.3036387184491054, + "grad_norm": 0.6239104866981506, + "learning_rate": 0.00015789126456124908, + "loss": 0.6574, + "step": 12953 + }, + { + "epoch": 0.3036621600239104, + "grad_norm": 0.1331377476453781, + "learning_rate": 0.00015788525811593745, + "loss": 0.0271, + "step": 12954 + }, + { + "epoch": 0.3036856015987154, + "grad_norm": 0.5306746959686279, + "learning_rate": 0.0001578792513565387, + "loss": 0.1519, + "step": 12955 + }, + { + "epoch": 0.3037090431735204, + "grad_norm": 0.5888978838920593, + "learning_rate": 0.00015787324428308557, + "loss": 0.1492, + "step": 12956 + }, + { + "epoch": 0.30373248474832537, + "grad_norm": 0.1929481327533722, + "learning_rate": 0.00015786723689561053, + "loss": 0.0534, + "step": 12957 + }, + { + "epoch": 0.30375592632313037, + "grad_norm": 0.4273107945919037, + "learning_rate": 0.0001578612291941463, + "loss": 0.0692, + "step": 12958 + }, + { + "epoch": 0.30377936789793536, + "grad_norm": 0.29543548822402954, + "learning_rate": 0.00015785522117872539, + "loss": 0.0634, + "step": 12959 + }, + { + "epoch": 0.30380280947274035, + "grad_norm": 0.47272294759750366, + "learning_rate": 0.0001578492128493804, + "loss": 0.0832, + "step": 12960 + }, + { + "epoch": 0.30382625104754535, + "grad_norm": 0.5298578143119812, + "learning_rate": 0.00015784320420614395, + "loss": 0.0383, + "step": 12961 + }, + { + "epoch": 0.30384969262235034, + "grad_norm": 0.47266653180122375, + "learning_rate": 0.00015783719524904867, + "loss": 0.0407, + "step": 12962 + }, + { + "epoch": 0.3038731341971554, + "grad_norm": 0.2993675172328949, + "learning_rate": 0.00015783118597812717, + "loss": 0.0791, + "step": 12963 + }, + { + "epoch": 0.3038965757719604, + "grad_norm": 0.6541202664375305, + "learning_rate": 0.00015782517639341198, + "loss": 0.1413, + "step": 12964 + }, + { + "epoch": 0.3039200173467654, + "grad_norm": 0.49751001596450806, + "learning_rate": 0.00015781916649493577, + "loss": 0.1301, + "step": 12965 + }, + { + "epoch": 0.3039434589215704, + "grad_norm": 0.36241763830184937, + "learning_rate": 0.00015781315628273117, + "loss": 0.0796, + "step": 12966 + }, + { + "epoch": 0.30396690049637537, + "grad_norm": 0.18431955575942993, + "learning_rate": 0.00015780714575683072, + "loss": 0.0321, + "step": 12967 + }, + { + "epoch": 0.30399034207118036, + "grad_norm": 0.5047779083251953, + "learning_rate": 0.00015780113491726709, + "loss": 0.1219, + "step": 12968 + }, + { + "epoch": 0.30401378364598536, + "grad_norm": 0.573615550994873, + "learning_rate": 0.00015779512376407287, + "loss": 0.1028, + "step": 12969 + }, + { + "epoch": 0.30403722522079035, + "grad_norm": 0.21569646894931793, + "learning_rate": 0.0001577891122972807, + "loss": 0.0344, + "step": 12970 + }, + { + "epoch": 0.30406066679559535, + "grad_norm": 0.4519837200641632, + "learning_rate": 0.00015778310051692314, + "loss": 0.0777, + "step": 12971 + }, + { + "epoch": 0.30408410837040034, + "grad_norm": 0.23390205204486847, + "learning_rate": 0.00015777708842303287, + "loss": 0.0554, + "step": 12972 + }, + { + "epoch": 0.30410754994520534, + "grad_norm": 0.8762259483337402, + "learning_rate": 0.00015777107601564253, + "loss": 0.1728, + "step": 12973 + }, + { + "epoch": 0.30413099152001033, + "grad_norm": 0.24449597299098969, + "learning_rate": 0.00015776506329478465, + "loss": 0.0329, + "step": 12974 + }, + { + "epoch": 0.3041544330948153, + "grad_norm": 0.452431321144104, + "learning_rate": 0.00015775905026049194, + "loss": 0.1788, + "step": 12975 + }, + { + "epoch": 0.3041778746696203, + "grad_norm": 0.5646793246269226, + "learning_rate": 0.000157753036912797, + "loss": 0.0876, + "step": 12976 + }, + { + "epoch": 0.3042013162444253, + "grad_norm": 0.18886886537075043, + "learning_rate": 0.00015774702325173245, + "loss": 0.036, + "step": 12977 + }, + { + "epoch": 0.3042247578192303, + "grad_norm": 0.45663508772850037, + "learning_rate": 0.00015774100927733094, + "loss": 0.0774, + "step": 12978 + }, + { + "epoch": 0.3042481993940353, + "grad_norm": 0.20672208070755005, + "learning_rate": 0.0001577349949896251, + "loss": 0.0347, + "step": 12979 + }, + { + "epoch": 0.3042716409688403, + "grad_norm": 0.48657047748565674, + "learning_rate": 0.0001577289803886475, + "loss": 0.1036, + "step": 12980 + }, + { + "epoch": 0.3042950825436453, + "grad_norm": 1.06768000125885, + "learning_rate": 0.00015772296547443088, + "loss": 0.1433, + "step": 12981 + }, + { + "epoch": 0.3043185241184503, + "grad_norm": 0.8413928151130676, + "learning_rate": 0.0001577169502470078, + "loss": 0.2083, + "step": 12982 + }, + { + "epoch": 0.3043419656932553, + "grad_norm": 0.1619560271501541, + "learning_rate": 0.00015771093470641096, + "loss": 0.0333, + "step": 12983 + }, + { + "epoch": 0.3043654072680603, + "grad_norm": 0.19868232309818268, + "learning_rate": 0.00015770491885267295, + "loss": 0.0408, + "step": 12984 + }, + { + "epoch": 0.30438884884286527, + "grad_norm": 0.4848967492580414, + "learning_rate": 0.00015769890268582643, + "loss": 0.3829, + "step": 12985 + }, + { + "epoch": 0.30441229041767026, + "grad_norm": 0.5533047318458557, + "learning_rate": 0.000157692886205904, + "loss": 0.1416, + "step": 12986 + }, + { + "epoch": 0.30443573199247526, + "grad_norm": 0.21333208680152893, + "learning_rate": 0.0001576868694129384, + "loss": 0.048, + "step": 12987 + }, + { + "epoch": 0.30445917356728025, + "grad_norm": 0.5587214827537537, + "learning_rate": 0.00015768085230696222, + "loss": 0.0955, + "step": 12988 + }, + { + "epoch": 0.30448261514208524, + "grad_norm": 0.3829978406429291, + "learning_rate": 0.00015767483488800807, + "loss": 0.0553, + "step": 12989 + }, + { + "epoch": 0.30450605671689024, + "grad_norm": 0.311143159866333, + "learning_rate": 0.0001576688171561087, + "loss": 0.062, + "step": 12990 + }, + { + "epoch": 0.30452949829169523, + "grad_norm": 1.7573922872543335, + "learning_rate": 0.00015766279911129667, + "loss": 0.1769, + "step": 12991 + }, + { + "epoch": 0.3045529398665002, + "grad_norm": 0.4717737138271332, + "learning_rate": 0.0001576567807536047, + "loss": 0.1262, + "step": 12992 + }, + { + "epoch": 0.3045763814413052, + "grad_norm": 0.6286865472793579, + "learning_rate": 0.00015765076208306536, + "loss": 0.1088, + "step": 12993 + }, + { + "epoch": 0.3045998230161102, + "grad_norm": 0.5822294354438782, + "learning_rate": 0.00015764474309971143, + "loss": 0.1125, + "step": 12994 + }, + { + "epoch": 0.3046232645909152, + "grad_norm": 0.5483118295669556, + "learning_rate": 0.00015763872380357546, + "loss": 0.1096, + "step": 12995 + }, + { + "epoch": 0.3046467061657202, + "grad_norm": 0.6471817493438721, + "learning_rate": 0.0001576327041946902, + "loss": 0.1013, + "step": 12996 + }, + { + "epoch": 0.3046701477405252, + "grad_norm": 0.34957119822502136, + "learning_rate": 0.0001576266842730882, + "loss": 0.0451, + "step": 12997 + }, + { + "epoch": 0.3046935893153302, + "grad_norm": 0.4468453824520111, + "learning_rate": 0.00015762066403880222, + "loss": 0.0772, + "step": 12998 + }, + { + "epoch": 0.3047170308901352, + "grad_norm": 0.4554196298122406, + "learning_rate": 0.00015761464349186492, + "loss": 0.0824, + "step": 12999 + }, + { + "epoch": 0.3047404724649402, + "grad_norm": 0.3825385570526123, + "learning_rate": 0.00015760862263230894, + "loss": 0.0464, + "step": 13000 + }, + { + "epoch": 0.3047639140397452, + "grad_norm": 0.22735275328159332, + "learning_rate": 0.00015760260146016695, + "loss": 0.0445, + "step": 13001 + }, + { + "epoch": 0.30478735561455017, + "grad_norm": 0.6400476098060608, + "learning_rate": 0.0001575965799754716, + "loss": 0.1538, + "step": 13002 + }, + { + "epoch": 0.30481079718935516, + "grad_norm": 0.7041304707527161, + "learning_rate": 0.0001575905581782556, + "loss": 0.1441, + "step": 13003 + }, + { + "epoch": 0.30483423876416016, + "grad_norm": 0.3091954290866852, + "learning_rate": 0.00015758453606855158, + "loss": 0.0795, + "step": 13004 + }, + { + "epoch": 0.30485768033896515, + "grad_norm": 0.8779503703117371, + "learning_rate": 0.0001575785136463923, + "loss": 0.1462, + "step": 13005 + }, + { + "epoch": 0.30488112191377015, + "grad_norm": 0.16682122647762299, + "learning_rate": 0.0001575724909118104, + "loss": 0.0269, + "step": 13006 + }, + { + "epoch": 0.30490456348857514, + "grad_norm": 1.2440831661224365, + "learning_rate": 0.0001575664678648385, + "loss": 0.1652, + "step": 13007 + }, + { + "epoch": 0.30492800506338014, + "grad_norm": 0.6385425329208374, + "learning_rate": 0.00015756044450550936, + "loss": 0.8564, + "step": 13008 + }, + { + "epoch": 0.30495144663818513, + "grad_norm": 0.6573658585548401, + "learning_rate": 0.0001575544208338556, + "loss": 0.0811, + "step": 13009 + }, + { + "epoch": 0.3049748882129901, + "grad_norm": 0.201863095164299, + "learning_rate": 0.00015754839684990995, + "loss": 0.0609, + "step": 13010 + }, + { + "epoch": 0.3049983297877951, + "grad_norm": 0.19078369438648224, + "learning_rate": 0.00015754237255370508, + "loss": 0.0283, + "step": 13011 + }, + { + "epoch": 0.3050217713626001, + "grad_norm": 0.8533262014389038, + "learning_rate": 0.00015753634794527364, + "loss": 0.1475, + "step": 13012 + }, + { + "epoch": 0.3050452129374051, + "grad_norm": 0.46068570017814636, + "learning_rate": 0.00015753032302464842, + "loss": 0.1169, + "step": 13013 + }, + { + "epoch": 0.30506865451221016, + "grad_norm": 0.2797105312347412, + "learning_rate": 0.00015752429779186203, + "loss": 0.0493, + "step": 13014 + }, + { + "epoch": 0.30509209608701515, + "grad_norm": 0.5879504680633545, + "learning_rate": 0.00015751827224694717, + "loss": 0.1152, + "step": 13015 + }, + { + "epoch": 0.30511553766182015, + "grad_norm": 0.5717185735702515, + "learning_rate": 0.00015751224638993654, + "loss": 0.127, + "step": 13016 + }, + { + "epoch": 0.30513897923662514, + "grad_norm": 0.32755404710769653, + "learning_rate": 0.00015750622022086286, + "loss": 0.0937, + "step": 13017 + }, + { + "epoch": 0.30516242081143014, + "grad_norm": 0.6159470081329346, + "learning_rate": 0.0001575001937397588, + "loss": 0.1423, + "step": 13018 + }, + { + "epoch": 0.30518586238623513, + "grad_norm": 0.5789000988006592, + "learning_rate": 0.0001574941669466571, + "loss": 0.0977, + "step": 13019 + }, + { + "epoch": 0.3052093039610401, + "grad_norm": 0.6776860356330872, + "learning_rate": 0.0001574881398415904, + "loss": 0.0799, + "step": 13020 + }, + { + "epoch": 0.3052327455358451, + "grad_norm": 0.463583767414093, + "learning_rate": 0.00015748211242459147, + "loss": 0.0893, + "step": 13021 + }, + { + "epoch": 0.3052561871106501, + "grad_norm": 0.48348331451416016, + "learning_rate": 0.00015747608469569296, + "loss": 0.1432, + "step": 13022 + }, + { + "epoch": 0.3052796286854551, + "grad_norm": 0.39341971278190613, + "learning_rate": 0.0001574700566549276, + "loss": 0.108, + "step": 13023 + }, + { + "epoch": 0.3053030702602601, + "grad_norm": 0.38774025440216064, + "learning_rate": 0.0001574640283023281, + "loss": 0.1, + "step": 13024 + }, + { + "epoch": 0.3053265118350651, + "grad_norm": 0.5105014443397522, + "learning_rate": 0.00015745799963792716, + "loss": 0.1505, + "step": 13025 + }, + { + "epoch": 0.3053499534098701, + "grad_norm": 0.27238139510154724, + "learning_rate": 0.0001574519706617575, + "loss": 0.0351, + "step": 13026 + }, + { + "epoch": 0.3053733949846751, + "grad_norm": 0.5434048175811768, + "learning_rate": 0.00015744594137385182, + "loss": 0.1553, + "step": 13027 + }, + { + "epoch": 0.3053968365594801, + "grad_norm": 0.6422064304351807, + "learning_rate": 0.00015743991177424285, + "loss": 0.087, + "step": 13028 + }, + { + "epoch": 0.3054202781342851, + "grad_norm": 0.44393390417099, + "learning_rate": 0.00015743388186296333, + "loss": 0.0722, + "step": 13029 + }, + { + "epoch": 0.30544371970909007, + "grad_norm": 0.4608132839202881, + "learning_rate": 0.00015742785164004594, + "loss": 0.0853, + "step": 13030 + }, + { + "epoch": 0.30546716128389506, + "grad_norm": 0.7756025791168213, + "learning_rate": 0.00015742182110552343, + "loss": 0.0985, + "step": 13031 + }, + { + "epoch": 0.30549060285870006, + "grad_norm": 0.40931418538093567, + "learning_rate": 0.00015741579025942848, + "loss": 0.0805, + "step": 13032 + }, + { + "epoch": 0.30551404443350505, + "grad_norm": 0.3521917462348938, + "learning_rate": 0.00015740975910179383, + "loss": 0.0925, + "step": 13033 + }, + { + "epoch": 0.30553748600831004, + "grad_norm": 0.5210956335067749, + "learning_rate": 0.00015740372763265223, + "loss": 0.442, + "step": 13034 + }, + { + "epoch": 0.30556092758311504, + "grad_norm": 0.5624454617500305, + "learning_rate": 0.00015739769585203638, + "loss": 0.5746, + "step": 13035 + }, + { + "epoch": 0.30558436915792003, + "grad_norm": 0.8152930736541748, + "learning_rate": 0.00015739166375997903, + "loss": 0.0878, + "step": 13036 + }, + { + "epoch": 0.305607810732725, + "grad_norm": 0.7475395202636719, + "learning_rate": 0.00015738563135651287, + "loss": 0.1179, + "step": 13037 + }, + { + "epoch": 0.30563125230753, + "grad_norm": 0.240336611866951, + "learning_rate": 0.0001573795986416707, + "loss": 0.0437, + "step": 13038 + }, + { + "epoch": 0.305654693882335, + "grad_norm": 1.2033889293670654, + "learning_rate": 0.0001573735656154852, + "loss": 0.1568, + "step": 13039 + }, + { + "epoch": 0.30567813545714, + "grad_norm": 0.17552436888217926, + "learning_rate": 0.0001573675322779891, + "loss": 0.0352, + "step": 13040 + }, + { + "epoch": 0.305701577031945, + "grad_norm": 0.21257849037647247, + "learning_rate": 0.00015736149862921516, + "loss": 0.0604, + "step": 13041 + }, + { + "epoch": 0.30572501860675, + "grad_norm": 0.5702955722808838, + "learning_rate": 0.00015735546466919616, + "loss": 0.5896, + "step": 13042 + }, + { + "epoch": 0.305748460181555, + "grad_norm": 0.16111978888511658, + "learning_rate": 0.00015734943039796476, + "loss": 0.0322, + "step": 13043 + }, + { + "epoch": 0.30577190175636, + "grad_norm": 0.8930695652961731, + "learning_rate": 0.00015734339581555373, + "loss": 0.1395, + "step": 13044 + }, + { + "epoch": 0.305795343331165, + "grad_norm": 0.6200088262557983, + "learning_rate": 0.00015733736092199584, + "loss": 0.2473, + "step": 13045 + }, + { + "epoch": 0.30581878490597, + "grad_norm": 0.1345527321100235, + "learning_rate": 0.0001573313257173238, + "loss": 0.029, + "step": 13046 + }, + { + "epoch": 0.30584222648077497, + "grad_norm": 0.5297858715057373, + "learning_rate": 0.00015732529020157042, + "loss": 0.1484, + "step": 13047 + }, + { + "epoch": 0.30586566805557996, + "grad_norm": 0.37099069356918335, + "learning_rate": 0.00015731925437476835, + "loss": 0.06, + "step": 13048 + }, + { + "epoch": 0.30588910963038496, + "grad_norm": 0.4602002203464508, + "learning_rate": 0.00015731321823695042, + "loss": 0.0618, + "step": 13049 + }, + { + "epoch": 0.30591255120518995, + "grad_norm": 0.7969696521759033, + "learning_rate": 0.00015730718178814934, + "loss": 0.1085, + "step": 13050 + }, + { + "epoch": 0.30593599277999495, + "grad_norm": 0.27700144052505493, + "learning_rate": 0.0001573011450283979, + "loss": 0.0311, + "step": 13051 + }, + { + "epoch": 0.30595943435479994, + "grad_norm": 0.7346398830413818, + "learning_rate": 0.0001572951079577288, + "loss": 0.5643, + "step": 13052 + }, + { + "epoch": 0.30598287592960494, + "grad_norm": 0.34738001227378845, + "learning_rate": 0.0001572890705761749, + "loss": 0.095, + "step": 13053 + }, + { + "epoch": 0.30600631750440993, + "grad_norm": 0.9838518500328064, + "learning_rate": 0.00015728303288376885, + "loss": 0.1837, + "step": 13054 + }, + { + "epoch": 0.3060297590792149, + "grad_norm": 0.6069453358650208, + "learning_rate": 0.00015727699488054343, + "loss": 0.1106, + "step": 13055 + }, + { + "epoch": 0.3060532006540199, + "grad_norm": 0.6064300537109375, + "learning_rate": 0.0001572709565665315, + "loss": 0.14, + "step": 13056 + }, + { + "epoch": 0.3060766422288249, + "grad_norm": 0.49738427996635437, + "learning_rate": 0.00015726491794176565, + "loss": 0.0767, + "step": 13057 + }, + { + "epoch": 0.3061000838036299, + "grad_norm": 0.4452340006828308, + "learning_rate": 0.00015725887900627883, + "loss": 0.1019, + "step": 13058 + }, + { + "epoch": 0.3061235253784349, + "grad_norm": 0.6797120571136475, + "learning_rate": 0.00015725283976010366, + "loss": 0.1223, + "step": 13059 + }, + { + "epoch": 0.3061469669532399, + "grad_norm": 0.7516437768936157, + "learning_rate": 0.00015724680020327302, + "loss": 0.0966, + "step": 13060 + }, + { + "epoch": 0.3061704085280449, + "grad_norm": 0.7760113477706909, + "learning_rate": 0.0001572407603358196, + "loss": 0.1486, + "step": 13061 + }, + { + "epoch": 0.3061938501028499, + "grad_norm": 0.42888256907463074, + "learning_rate": 0.00015723472015777623, + "loss": 0.0815, + "step": 13062 + }, + { + "epoch": 0.3062172916776549, + "grad_norm": 0.5716139674186707, + "learning_rate": 0.00015722867966917566, + "loss": 0.0855, + "step": 13063 + }, + { + "epoch": 0.3062407332524599, + "grad_norm": 0.5973283648490906, + "learning_rate": 0.00015722263887005063, + "loss": 0.6695, + "step": 13064 + }, + { + "epoch": 0.3062641748272649, + "grad_norm": 0.7930651903152466, + "learning_rate": 0.00015721659776043399, + "loss": 0.1222, + "step": 13065 + }, + { + "epoch": 0.3062876164020699, + "grad_norm": 0.42995932698249817, + "learning_rate": 0.00015721055634035843, + "loss": 0.0766, + "step": 13066 + }, + { + "epoch": 0.3063110579768749, + "grad_norm": 0.8686266541481018, + "learning_rate": 0.00015720451460985682, + "loss": 0.1751, + "step": 13067 + }, + { + "epoch": 0.3063344995516799, + "grad_norm": 0.41935858130455017, + "learning_rate": 0.00015719847256896193, + "loss": 0.0709, + "step": 13068 + }, + { + "epoch": 0.3063579411264849, + "grad_norm": 0.22206401824951172, + "learning_rate": 0.0001571924302177065, + "loss": 0.0354, + "step": 13069 + }, + { + "epoch": 0.3063813827012899, + "grad_norm": 0.6161923408508301, + "learning_rate": 0.0001571863875561233, + "loss": 0.16, + "step": 13070 + }, + { + "epoch": 0.3064048242760949, + "grad_norm": 0.6406531929969788, + "learning_rate": 0.0001571803445842452, + "loss": 0.2017, + "step": 13071 + }, + { + "epoch": 0.3064282658508999, + "grad_norm": 0.6651013493537903, + "learning_rate": 0.0001571743013021049, + "loss": 0.1365, + "step": 13072 + }, + { + "epoch": 0.3064517074257049, + "grad_norm": 0.9793004393577576, + "learning_rate": 0.00015716825770973526, + "loss": 0.1234, + "step": 13073 + }, + { + "epoch": 0.30647514900050987, + "grad_norm": 0.6069460511207581, + "learning_rate": 0.00015716221380716907, + "loss": 0.1287, + "step": 13074 + }, + { + "epoch": 0.30649859057531487, + "grad_norm": 1.2707692384719849, + "learning_rate": 0.00015715616959443906, + "loss": 0.1501, + "step": 13075 + }, + { + "epoch": 0.30652203215011986, + "grad_norm": 0.5492902398109436, + "learning_rate": 0.00015715012507157807, + "loss": 0.1454, + "step": 13076 + }, + { + "epoch": 0.30654547372492486, + "grad_norm": 0.3560415208339691, + "learning_rate": 0.0001571440802386189, + "loss": 0.0591, + "step": 13077 + }, + { + "epoch": 0.30656891529972985, + "grad_norm": 0.419950932264328, + "learning_rate": 0.00015713803509559434, + "loss": 0.0812, + "step": 13078 + }, + { + "epoch": 0.30659235687453484, + "grad_norm": 0.6390207409858704, + "learning_rate": 0.00015713198964253724, + "loss": 0.1316, + "step": 13079 + }, + { + "epoch": 0.30661579844933984, + "grad_norm": 0.16244551539421082, + "learning_rate": 0.0001571259438794803, + "loss": 0.0338, + "step": 13080 + }, + { + "epoch": 0.30663924002414483, + "grad_norm": 0.4127865433692932, + "learning_rate": 0.00015711989780645643, + "loss": 0.0915, + "step": 13081 + }, + { + "epoch": 0.3066626815989498, + "grad_norm": 0.6942306160926819, + "learning_rate": 0.00015711385142349834, + "loss": 0.7703, + "step": 13082 + }, + { + "epoch": 0.3066861231737548, + "grad_norm": 0.7731118202209473, + "learning_rate": 0.0001571078047306389, + "loss": 0.1277, + "step": 13083 + }, + { + "epoch": 0.3067095647485598, + "grad_norm": 0.30366089940071106, + "learning_rate": 0.00015710175772791095, + "loss": 0.0447, + "step": 13084 + }, + { + "epoch": 0.3067330063233648, + "grad_norm": 0.1538270115852356, + "learning_rate": 0.0001570957104153472, + "loss": 0.0248, + "step": 13085 + }, + { + "epoch": 0.3067564478981698, + "grad_norm": 0.5579497814178467, + "learning_rate": 0.00015708966279298053, + "loss": 0.0592, + "step": 13086 + }, + { + "epoch": 0.3067798894729748, + "grad_norm": 0.29123207926750183, + "learning_rate": 0.00015708361486084377, + "loss": 0.0615, + "step": 13087 + }, + { + "epoch": 0.3068033310477798, + "grad_norm": 0.950859546661377, + "learning_rate": 0.0001570775666189697, + "loss": 0.1579, + "step": 13088 + }, + { + "epoch": 0.3068267726225848, + "grad_norm": 0.39863064885139465, + "learning_rate": 0.00015707151806739113, + "loss": 0.0629, + "step": 13089 + }, + { + "epoch": 0.3068502141973898, + "grad_norm": 0.4157956540584564, + "learning_rate": 0.00015706546920614094, + "loss": 0.125, + "step": 13090 + }, + { + "epoch": 0.3068736557721948, + "grad_norm": 0.5243772864341736, + "learning_rate": 0.00015705942003525186, + "loss": 0.1347, + "step": 13091 + }, + { + "epoch": 0.30689709734699977, + "grad_norm": 0.6210592985153198, + "learning_rate": 0.0001570533705547568, + "loss": 0.1403, + "step": 13092 + }, + { + "epoch": 0.30692053892180476, + "grad_norm": 0.3944600820541382, + "learning_rate": 0.00015704732076468852, + "loss": 0.1063, + "step": 13093 + }, + { + "epoch": 0.30694398049660976, + "grad_norm": 0.503948450088501, + "learning_rate": 0.0001570412706650799, + "loss": 0.1409, + "step": 13094 + }, + { + "epoch": 0.30696742207141475, + "grad_norm": 0.579464316368103, + "learning_rate": 0.00015703522025596373, + "loss": 0.6555, + "step": 13095 + }, + { + "epoch": 0.30699086364621975, + "grad_norm": 0.37498027086257935, + "learning_rate": 0.00015702916953737285, + "loss": 0.1122, + "step": 13096 + }, + { + "epoch": 0.30701430522102474, + "grad_norm": 0.22923853993415833, + "learning_rate": 0.0001570231185093401, + "loss": 0.0482, + "step": 13097 + }, + { + "epoch": 0.30703774679582974, + "grad_norm": 0.3588472306728363, + "learning_rate": 0.00015701706717189827, + "loss": 0.053, + "step": 13098 + }, + { + "epoch": 0.30706118837063473, + "grad_norm": 0.4133758246898651, + "learning_rate": 0.00015701101552508028, + "loss": 0.0235, + "step": 13099 + }, + { + "epoch": 0.3070846299454397, + "grad_norm": 0.5925237536430359, + "learning_rate": 0.00015700496356891889, + "loss": 0.0795, + "step": 13100 + }, + { + "epoch": 0.3071080715202447, + "grad_norm": 0.2530617117881775, + "learning_rate": 0.00015699891130344695, + "loss": 0.0617, + "step": 13101 + }, + { + "epoch": 0.3071315130950497, + "grad_norm": 0.20441249012947083, + "learning_rate": 0.00015699285872869734, + "loss": 0.0299, + "step": 13102 + }, + { + "epoch": 0.3071549546698547, + "grad_norm": 0.3018794059753418, + "learning_rate": 0.0001569868058447029, + "loss": 0.074, + "step": 13103 + }, + { + "epoch": 0.3071783962446597, + "grad_norm": 0.7458487153053284, + "learning_rate": 0.00015698075265149642, + "loss": 0.2481, + "step": 13104 + }, + { + "epoch": 0.3072018378194647, + "grad_norm": 0.48465144634246826, + "learning_rate": 0.00015697469914911075, + "loss": 0.112, + "step": 13105 + }, + { + "epoch": 0.3072252793942697, + "grad_norm": 0.1466599553823471, + "learning_rate": 0.0001569686453375788, + "loss": 0.0447, + "step": 13106 + }, + { + "epoch": 0.3072487209690747, + "grad_norm": 0.7191744446754456, + "learning_rate": 0.00015696259121693334, + "loss": 0.1879, + "step": 13107 + }, + { + "epoch": 0.3072721625438797, + "grad_norm": 0.29912829399108887, + "learning_rate": 0.00015695653678720727, + "loss": 0.0719, + "step": 13108 + }, + { + "epoch": 0.3072956041186847, + "grad_norm": 0.6467163562774658, + "learning_rate": 0.00015695048204843344, + "loss": 0.1419, + "step": 13109 + }, + { + "epoch": 0.30731904569348967, + "grad_norm": 0.7802058458328247, + "learning_rate": 0.00015694442700064473, + "loss": 0.1031, + "step": 13110 + }, + { + "epoch": 0.30734248726829466, + "grad_norm": 0.5652786493301392, + "learning_rate": 0.00015693837164387393, + "loss": 0.1743, + "step": 13111 + }, + { + "epoch": 0.30736592884309966, + "grad_norm": 0.3624497950077057, + "learning_rate": 0.0001569323159781539, + "loss": 0.0906, + "step": 13112 + }, + { + "epoch": 0.30738937041790465, + "grad_norm": 0.7957281470298767, + "learning_rate": 0.00015692626000351752, + "loss": 0.1546, + "step": 13113 + }, + { + "epoch": 0.30741281199270964, + "grad_norm": 0.3245597183704376, + "learning_rate": 0.00015692020371999767, + "loss": 0.3929, + "step": 13114 + }, + { + "epoch": 0.30743625356751464, + "grad_norm": 0.6789974570274353, + "learning_rate": 0.0001569141471276272, + "loss": 0.8406, + "step": 13115 + }, + { + "epoch": 0.30745969514231963, + "grad_norm": 0.2593456208705902, + "learning_rate": 0.000156908090226439, + "loss": 0.0282, + "step": 13116 + }, + { + "epoch": 0.3074831367171247, + "grad_norm": 0.27285200357437134, + "learning_rate": 0.00015690203301646583, + "loss": 0.0561, + "step": 13117 + }, + { + "epoch": 0.3075065782919297, + "grad_norm": 0.25619709491729736, + "learning_rate": 0.00015689597549774069, + "loss": 0.0626, + "step": 13118 + }, + { + "epoch": 0.30753001986673467, + "grad_norm": 0.45829492807388306, + "learning_rate": 0.00015688991767029634, + "loss": 0.1296, + "step": 13119 + }, + { + "epoch": 0.30755346144153967, + "grad_norm": 0.4376920461654663, + "learning_rate": 0.00015688385953416573, + "loss": 0.4624, + "step": 13120 + }, + { + "epoch": 0.30757690301634466, + "grad_norm": 0.5078352093696594, + "learning_rate": 0.0001568778010893817, + "loss": 0.1217, + "step": 13121 + }, + { + "epoch": 0.30760034459114965, + "grad_norm": 0.7809985876083374, + "learning_rate": 0.0001568717423359771, + "loss": 0.2103, + "step": 13122 + }, + { + "epoch": 0.30762378616595465, + "grad_norm": 0.5271496176719666, + "learning_rate": 0.00015686568327398482, + "loss": 0.5876, + "step": 13123 + }, + { + "epoch": 0.30764722774075964, + "grad_norm": 0.7063848376274109, + "learning_rate": 0.00015685962390343775, + "loss": 0.1433, + "step": 13124 + }, + { + "epoch": 0.30767066931556464, + "grad_norm": 0.33775201439857483, + "learning_rate": 0.0001568535642243688, + "loss": 0.0676, + "step": 13125 + }, + { + "epoch": 0.30769411089036963, + "grad_norm": 0.34402501583099365, + "learning_rate": 0.00015684750423681075, + "loss": 0.0485, + "step": 13126 + }, + { + "epoch": 0.3077175524651746, + "grad_norm": 0.4594564437866211, + "learning_rate": 0.0001568414439407966, + "loss": 0.0601, + "step": 13127 + }, + { + "epoch": 0.3077409940399796, + "grad_norm": 0.11875545233488083, + "learning_rate": 0.00015683538333635915, + "loss": 0.0195, + "step": 13128 + }, + { + "epoch": 0.3077644356147846, + "grad_norm": 0.5942839980125427, + "learning_rate": 0.00015682932242353133, + "loss": 0.1339, + "step": 13129 + }, + { + "epoch": 0.3077878771895896, + "grad_norm": 0.46675601601600647, + "learning_rate": 0.000156823261202346, + "loss": 0.0637, + "step": 13130 + }, + { + "epoch": 0.3078113187643946, + "grad_norm": 0.6995079517364502, + "learning_rate": 0.00015681719967283606, + "loss": 0.1962, + "step": 13131 + }, + { + "epoch": 0.3078347603391996, + "grad_norm": 0.9107058048248291, + "learning_rate": 0.00015681113783503435, + "loss": 0.2223, + "step": 13132 + }, + { + "epoch": 0.3078582019140046, + "grad_norm": 0.5930392742156982, + "learning_rate": 0.00015680507568897387, + "loss": 0.0686, + "step": 13133 + }, + { + "epoch": 0.3078816434888096, + "grad_norm": 1.1442204713821411, + "learning_rate": 0.00015679901323468744, + "loss": 0.1643, + "step": 13134 + }, + { + "epoch": 0.3079050850636146, + "grad_norm": 0.88096684217453, + "learning_rate": 0.00015679295047220794, + "loss": 0.2487, + "step": 13135 + }, + { + "epoch": 0.3079285266384196, + "grad_norm": 0.4116250276565552, + "learning_rate": 0.00015678688740156834, + "loss": 0.0633, + "step": 13136 + }, + { + "epoch": 0.30795196821322457, + "grad_norm": 0.4369061291217804, + "learning_rate": 0.00015678082402280144, + "loss": 0.1171, + "step": 13137 + }, + { + "epoch": 0.30797540978802956, + "grad_norm": 0.3877302408218384, + "learning_rate": 0.00015677476033594022, + "loss": 0.0391, + "step": 13138 + }, + { + "epoch": 0.30799885136283456, + "grad_norm": 0.4627762734889984, + "learning_rate": 0.00015676869634101754, + "loss": 0.1673, + "step": 13139 + }, + { + "epoch": 0.30802229293763955, + "grad_norm": 0.6210778951644897, + "learning_rate": 0.00015676263203806632, + "loss": 0.1547, + "step": 13140 + }, + { + "epoch": 0.30804573451244455, + "grad_norm": 0.29222774505615234, + "learning_rate": 0.00015675656742711948, + "loss": 0.3547, + "step": 13141 + }, + { + "epoch": 0.30806917608724954, + "grad_norm": 0.579045295715332, + "learning_rate": 0.00015675050250820992, + "loss": 0.1036, + "step": 13142 + }, + { + "epoch": 0.30809261766205454, + "grad_norm": 0.6284055709838867, + "learning_rate": 0.00015674443728137049, + "loss": 0.1993, + "step": 13143 + }, + { + "epoch": 0.30811605923685953, + "grad_norm": 0.5500726699829102, + "learning_rate": 0.00015673837174663417, + "loss": 0.1572, + "step": 13144 + }, + { + "epoch": 0.3081395008116645, + "grad_norm": 0.5830731987953186, + "learning_rate": 0.00015673230590403385, + "loss": 0.1822, + "step": 13145 + }, + { + "epoch": 0.3081629423864695, + "grad_norm": 0.6940301656723022, + "learning_rate": 0.00015672623975360246, + "loss": 0.2454, + "step": 13146 + }, + { + "epoch": 0.3081863839612745, + "grad_norm": 0.42496222257614136, + "learning_rate": 0.00015672017329537287, + "loss": 0.069, + "step": 13147 + }, + { + "epoch": 0.3082098255360795, + "grad_norm": 0.6357430219650269, + "learning_rate": 0.00015671410652937806, + "loss": 0.1295, + "step": 13148 + }, + { + "epoch": 0.3082332671108845, + "grad_norm": 1.0647187232971191, + "learning_rate": 0.00015670803945565088, + "loss": 0.1942, + "step": 13149 + }, + { + "epoch": 0.3082567086856895, + "grad_norm": 0.373109370470047, + "learning_rate": 0.0001567019720742243, + "loss": 0.092, + "step": 13150 + }, + { + "epoch": 0.3082801502604945, + "grad_norm": 0.6087711453437805, + "learning_rate": 0.00015669590438513123, + "loss": 0.095, + "step": 13151 + }, + { + "epoch": 0.3083035918352995, + "grad_norm": 0.3697814345359802, + "learning_rate": 0.00015668983638840458, + "loss": 0.0806, + "step": 13152 + }, + { + "epoch": 0.3083270334101045, + "grad_norm": 1.3860574960708618, + "learning_rate": 0.0001566837680840773, + "loss": 0.1812, + "step": 13153 + }, + { + "epoch": 0.3083504749849095, + "grad_norm": 0.5398190021514893, + "learning_rate": 0.00015667769947218227, + "loss": 0.6566, + "step": 13154 + }, + { + "epoch": 0.30837391655971447, + "grad_norm": 0.5228239893913269, + "learning_rate": 0.00015667163055275247, + "loss": 0.1285, + "step": 13155 + }, + { + "epoch": 0.30839735813451946, + "grad_norm": 0.545597493648529, + "learning_rate": 0.00015666556132582076, + "loss": 0.1224, + "step": 13156 + }, + { + "epoch": 0.30842079970932446, + "grad_norm": 0.5721386671066284, + "learning_rate": 0.0001566594917914202, + "loss": 0.1166, + "step": 13157 + }, + { + "epoch": 0.30844424128412945, + "grad_norm": 0.1386866420507431, + "learning_rate": 0.00015665342194958362, + "loss": 0.032, + "step": 13158 + }, + { + "epoch": 0.30846768285893444, + "grad_norm": 0.4572538137435913, + "learning_rate": 0.00015664735180034395, + "loss": 0.1046, + "step": 13159 + }, + { + "epoch": 0.30849112443373944, + "grad_norm": 0.5417520999908447, + "learning_rate": 0.00015664128134373418, + "loss": 0.077, + "step": 13160 + }, + { + "epoch": 0.30851456600854443, + "grad_norm": 0.4588961601257324, + "learning_rate": 0.00015663521057978722, + "loss": 0.141, + "step": 13161 + }, + { + "epoch": 0.3085380075833494, + "grad_norm": 0.2678750157356262, + "learning_rate": 0.00015662913950853602, + "loss": 0.0375, + "step": 13162 + }, + { + "epoch": 0.3085614491581544, + "grad_norm": 0.1871780902147293, + "learning_rate": 0.0001566230681300135, + "loss": 0.0495, + "step": 13163 + }, + { + "epoch": 0.3085848907329594, + "grad_norm": 0.3687104284763336, + "learning_rate": 0.00015661699644425262, + "loss": 0.1213, + "step": 13164 + }, + { + "epoch": 0.3086083323077644, + "grad_norm": 0.2708037793636322, + "learning_rate": 0.00015661092445128633, + "loss": 0.0426, + "step": 13165 + }, + { + "epoch": 0.3086317738825694, + "grad_norm": 0.21390770375728607, + "learning_rate": 0.0001566048521511476, + "loss": 0.0303, + "step": 13166 + }, + { + "epoch": 0.3086552154573744, + "grad_norm": 0.6466007232666016, + "learning_rate": 0.00015659877954386935, + "loss": 0.1192, + "step": 13167 + }, + { + "epoch": 0.30867865703217945, + "grad_norm": 0.4316050708293915, + "learning_rate": 0.00015659270662948447, + "loss": 0.1026, + "step": 13168 + }, + { + "epoch": 0.30870209860698444, + "grad_norm": 0.6102323532104492, + "learning_rate": 0.00015658663340802602, + "loss": 0.6227, + "step": 13169 + }, + { + "epoch": 0.30872554018178944, + "grad_norm": 0.5114749670028687, + "learning_rate": 0.0001565805598795269, + "loss": 0.637, + "step": 13170 + }, + { + "epoch": 0.30874898175659443, + "grad_norm": 0.2775486707687378, + "learning_rate": 0.00015657448604402006, + "loss": 0.0548, + "step": 13171 + }, + { + "epoch": 0.3087724233313994, + "grad_norm": 0.15360093116760254, + "learning_rate": 0.00015656841190153848, + "loss": 0.0285, + "step": 13172 + }, + { + "epoch": 0.3087958649062044, + "grad_norm": 0.16999895870685577, + "learning_rate": 0.00015656233745211512, + "loss": 0.0196, + "step": 13173 + }, + { + "epoch": 0.3088193064810094, + "grad_norm": 0.6702501773834229, + "learning_rate": 0.0001565562626957829, + "loss": 0.148, + "step": 13174 + }, + { + "epoch": 0.3088427480558144, + "grad_norm": 0.3963305652141571, + "learning_rate": 0.00015655018763257483, + "loss": 0.0817, + "step": 13175 + }, + { + "epoch": 0.3088661896306194, + "grad_norm": 0.37839826941490173, + "learning_rate": 0.00015654411226252385, + "loss": 0.0548, + "step": 13176 + }, + { + "epoch": 0.3088896312054244, + "grad_norm": 0.4673936367034912, + "learning_rate": 0.0001565380365856629, + "loss": 0.1265, + "step": 13177 + }, + { + "epoch": 0.3089130727802294, + "grad_norm": 0.36956164240837097, + "learning_rate": 0.000156531960602025, + "loss": 0.0624, + "step": 13178 + }, + { + "epoch": 0.3089365143550344, + "grad_norm": 0.3254948854446411, + "learning_rate": 0.0001565258843116431, + "loss": 0.0629, + "step": 13179 + }, + { + "epoch": 0.3089599559298394, + "grad_norm": 0.43286389112472534, + "learning_rate": 0.00015651980771455012, + "loss": 0.0996, + "step": 13180 + }, + { + "epoch": 0.3089833975046444, + "grad_norm": 0.7116671800613403, + "learning_rate": 0.0001565137308107791, + "loss": 0.4914, + "step": 13181 + }, + { + "epoch": 0.30900683907944937, + "grad_norm": 0.2280140072107315, + "learning_rate": 0.000156507653600363, + "loss": 0.0847, + "step": 13182 + }, + { + "epoch": 0.30903028065425436, + "grad_norm": 0.5623407959938049, + "learning_rate": 0.0001565015760833348, + "loss": 0.1352, + "step": 13183 + }, + { + "epoch": 0.30905372222905936, + "grad_norm": 0.4580025374889374, + "learning_rate": 0.0001564954982597274, + "loss": 0.1343, + "step": 13184 + }, + { + "epoch": 0.30907716380386435, + "grad_norm": 0.596321165561676, + "learning_rate": 0.00015648942012957387, + "loss": 0.1591, + "step": 13185 + }, + { + "epoch": 0.30910060537866935, + "grad_norm": 1.0979413986206055, + "learning_rate": 0.00015648334169290714, + "loss": 0.1376, + "step": 13186 + }, + { + "epoch": 0.30912404695347434, + "grad_norm": 0.07391642779111862, + "learning_rate": 0.00015647726294976025, + "loss": 0.0156, + "step": 13187 + }, + { + "epoch": 0.30914748852827934, + "grad_norm": 0.5201169848442078, + "learning_rate": 0.0001564711839001661, + "loss": 0.0998, + "step": 13188 + }, + { + "epoch": 0.30917093010308433, + "grad_norm": 0.4918419122695923, + "learning_rate": 0.00015646510454415773, + "loss": 0.1162, + "step": 13189 + }, + { + "epoch": 0.3091943716778893, + "grad_norm": 0.3781612813472748, + "learning_rate": 0.00015645902488176816, + "loss": 0.0901, + "step": 13190 + }, + { + "epoch": 0.3092178132526943, + "grad_norm": 0.8545082807540894, + "learning_rate": 0.00015645294491303027, + "loss": 0.1731, + "step": 13191 + }, + { + "epoch": 0.3092412548274993, + "grad_norm": 0.15670554339885712, + "learning_rate": 0.00015644686463797713, + "loss": 0.0384, + "step": 13192 + }, + { + "epoch": 0.3092646964023043, + "grad_norm": 0.6743385195732117, + "learning_rate": 0.00015644078405664173, + "loss": 0.503, + "step": 13193 + }, + { + "epoch": 0.3092881379771093, + "grad_norm": 0.12424062937498093, + "learning_rate": 0.00015643470316905705, + "loss": 0.0193, + "step": 13194 + }, + { + "epoch": 0.3093115795519143, + "grad_norm": 0.4828340709209442, + "learning_rate": 0.00015642862197525606, + "loss": 0.0967, + "step": 13195 + }, + { + "epoch": 0.3093350211267193, + "grad_norm": 0.19067923724651337, + "learning_rate": 0.0001564225404752718, + "loss": 0.0287, + "step": 13196 + }, + { + "epoch": 0.3093584627015243, + "grad_norm": 0.39986753463745117, + "learning_rate": 0.00015641645866913727, + "loss": 0.1154, + "step": 13197 + }, + { + "epoch": 0.3093819042763293, + "grad_norm": 0.5715112090110779, + "learning_rate": 0.00015641037655688542, + "loss": 0.1121, + "step": 13198 + }, + { + "epoch": 0.3094053458511343, + "grad_norm": 0.4902917742729187, + "learning_rate": 0.00015640429413854928, + "loss": 0.1382, + "step": 13199 + }, + { + "epoch": 0.30942878742593927, + "grad_norm": 0.6884393095970154, + "learning_rate": 0.00015639821141416187, + "loss": 0.0929, + "step": 13200 + }, + { + "epoch": 0.30945222900074426, + "grad_norm": 0.5707409381866455, + "learning_rate": 0.00015639212838375617, + "loss": 0.666, + "step": 13201 + }, + { + "epoch": 0.30947567057554926, + "grad_norm": 0.6191786527633667, + "learning_rate": 0.00015638604504736517, + "loss": 0.1322, + "step": 13202 + }, + { + "epoch": 0.30949911215035425, + "grad_norm": 0.8290018439292908, + "learning_rate": 0.00015637996140502195, + "loss": 0.1382, + "step": 13203 + }, + { + "epoch": 0.30952255372515924, + "grad_norm": 0.21831218898296356, + "learning_rate": 0.00015637387745675948, + "loss": 0.0352, + "step": 13204 + }, + { + "epoch": 0.30954599529996424, + "grad_norm": 0.1651589721441269, + "learning_rate": 0.00015636779320261072, + "loss": 0.0363, + "step": 13205 + }, + { + "epoch": 0.30956943687476923, + "grad_norm": 0.43105602264404297, + "learning_rate": 0.00015636170864260875, + "loss": 0.0819, + "step": 13206 + }, + { + "epoch": 0.3095928784495742, + "grad_norm": 0.4505831301212311, + "learning_rate": 0.00015635562377678657, + "loss": 0.1677, + "step": 13207 + }, + { + "epoch": 0.3096163200243792, + "grad_norm": 0.5992639660835266, + "learning_rate": 0.00015634953860517716, + "loss": 0.117, + "step": 13208 + }, + { + "epoch": 0.3096397615991842, + "grad_norm": 0.5621866583824158, + "learning_rate": 0.00015634345312781362, + "loss": 0.1543, + "step": 13209 + }, + { + "epoch": 0.3096632031739892, + "grad_norm": 0.759595513343811, + "learning_rate": 0.0001563373673447289, + "loss": 0.1489, + "step": 13210 + }, + { + "epoch": 0.3096866447487942, + "grad_norm": 0.3694675862789154, + "learning_rate": 0.000156331281255956, + "loss": 0.0445, + "step": 13211 + }, + { + "epoch": 0.3097100863235992, + "grad_norm": 0.3144325911998749, + "learning_rate": 0.00015632519486152803, + "loss": 0.0773, + "step": 13212 + }, + { + "epoch": 0.3097335278984042, + "grad_norm": 0.26961368322372437, + "learning_rate": 0.00015631910816147796, + "loss": 0.0591, + "step": 13213 + }, + { + "epoch": 0.3097569694732092, + "grad_norm": 0.24930764734745026, + "learning_rate": 0.00015631302115583885, + "loss": 0.0374, + "step": 13214 + }, + { + "epoch": 0.3097804110480142, + "grad_norm": 0.23977436125278473, + "learning_rate": 0.00015630693384464364, + "loss": 0.0621, + "step": 13215 + }, + { + "epoch": 0.3098038526228192, + "grad_norm": 0.8225262761116028, + "learning_rate": 0.00015630084622792547, + "loss": 0.2326, + "step": 13216 + }, + { + "epoch": 0.30982729419762417, + "grad_norm": 0.44961801171302795, + "learning_rate": 0.0001562947583057173, + "loss": 0.0451, + "step": 13217 + }, + { + "epoch": 0.30985073577242916, + "grad_norm": 0.515877366065979, + "learning_rate": 0.0001562886700780522, + "loss": 0.126, + "step": 13218 + }, + { + "epoch": 0.3098741773472342, + "grad_norm": 0.7579277753829956, + "learning_rate": 0.0001562825815449632, + "loss": 0.0736, + "step": 13219 + }, + { + "epoch": 0.3098976189220392, + "grad_norm": 0.27705705165863037, + "learning_rate": 0.00015627649270648334, + "loss": 0.0757, + "step": 13220 + }, + { + "epoch": 0.3099210604968442, + "grad_norm": 0.4700813889503479, + "learning_rate": 0.00015627040356264561, + "loss": 0.094, + "step": 13221 + }, + { + "epoch": 0.3099445020716492, + "grad_norm": 0.656024158000946, + "learning_rate": 0.00015626431411348314, + "loss": 0.0975, + "step": 13222 + }, + { + "epoch": 0.3099679436464542, + "grad_norm": 0.18973934650421143, + "learning_rate": 0.00015625822435902888, + "loss": 0.0312, + "step": 13223 + }, + { + "epoch": 0.3099913852212592, + "grad_norm": 0.17256736755371094, + "learning_rate": 0.0001562521342993159, + "loss": 0.0409, + "step": 13224 + }, + { + "epoch": 0.3100148267960642, + "grad_norm": 0.5346209406852722, + "learning_rate": 0.00015624604393437732, + "loss": 0.0865, + "step": 13225 + }, + { + "epoch": 0.3100382683708692, + "grad_norm": 0.6169013977050781, + "learning_rate": 0.00015623995326424607, + "loss": 0.0884, + "step": 13226 + }, + { + "epoch": 0.31006170994567417, + "grad_norm": 0.6207048296928406, + "learning_rate": 0.00015623386228895528, + "loss": 0.7023, + "step": 13227 + }, + { + "epoch": 0.31008515152047916, + "grad_norm": 0.3612131178379059, + "learning_rate": 0.00015622777100853796, + "loss": 0.1, + "step": 13228 + }, + { + "epoch": 0.31010859309528416, + "grad_norm": 0.4187875986099243, + "learning_rate": 0.00015622167942302718, + "loss": 0.1073, + "step": 13229 + }, + { + "epoch": 0.31013203467008915, + "grad_norm": 0.2892153561115265, + "learning_rate": 0.00015621558753245598, + "loss": 0.1123, + "step": 13230 + }, + { + "epoch": 0.31015547624489415, + "grad_norm": 0.88747239112854, + "learning_rate": 0.0001562094953368574, + "loss": 0.1861, + "step": 13231 + }, + { + "epoch": 0.31017891781969914, + "grad_norm": 0.47248512506484985, + "learning_rate": 0.00015620340283626453, + "loss": 0.1395, + "step": 13232 + }, + { + "epoch": 0.31020235939450413, + "grad_norm": 0.5152300596237183, + "learning_rate": 0.00015619731003071045, + "loss": 0.0957, + "step": 13233 + }, + { + "epoch": 0.31022580096930913, + "grad_norm": 0.252215176820755, + "learning_rate": 0.00015619121692022817, + "loss": 0.0578, + "step": 13234 + }, + { + "epoch": 0.3102492425441141, + "grad_norm": 0.5252723693847656, + "learning_rate": 0.00015618512350485075, + "loss": 0.1092, + "step": 13235 + }, + { + "epoch": 0.3102726841189191, + "grad_norm": 0.4371017813682556, + "learning_rate": 0.00015617902978461127, + "loss": 0.0582, + "step": 13236 + }, + { + "epoch": 0.3102961256937241, + "grad_norm": 0.7217769026756287, + "learning_rate": 0.00015617293575954282, + "loss": 0.1179, + "step": 13237 + }, + { + "epoch": 0.3103195672685291, + "grad_norm": 0.5319877862930298, + "learning_rate": 0.0001561668414296784, + "loss": 0.1299, + "step": 13238 + }, + { + "epoch": 0.3103430088433341, + "grad_norm": 0.5125715136528015, + "learning_rate": 0.00015616074679505117, + "loss": 0.1186, + "step": 13239 + }, + { + "epoch": 0.3103664504181391, + "grad_norm": 0.15742425620555878, + "learning_rate": 0.00015615465185569413, + "loss": 0.0359, + "step": 13240 + }, + { + "epoch": 0.3103898919929441, + "grad_norm": 0.37714430689811707, + "learning_rate": 0.00015614855661164035, + "loss": 0.0853, + "step": 13241 + }, + { + "epoch": 0.3104133335677491, + "grad_norm": 0.10728469491004944, + "learning_rate": 0.00015614246106292293, + "loss": 0.0276, + "step": 13242 + }, + { + "epoch": 0.3104367751425541, + "grad_norm": 0.373904824256897, + "learning_rate": 0.00015613636520957495, + "loss": 0.1202, + "step": 13243 + }, + { + "epoch": 0.31046021671735907, + "grad_norm": 0.4595508873462677, + "learning_rate": 0.00015613026905162946, + "loss": 0.0572, + "step": 13244 + }, + { + "epoch": 0.31048365829216407, + "grad_norm": 0.6621421575546265, + "learning_rate": 0.00015612417258911956, + "loss": 0.0726, + "step": 13245 + }, + { + "epoch": 0.31050709986696906, + "grad_norm": 0.6046280264854431, + "learning_rate": 0.00015611807582207834, + "loss": 0.1265, + "step": 13246 + }, + { + "epoch": 0.31053054144177406, + "grad_norm": 0.5085354447364807, + "learning_rate": 0.00015611197875053882, + "loss": 0.1469, + "step": 13247 + }, + { + "epoch": 0.31055398301657905, + "grad_norm": 0.831255316734314, + "learning_rate": 0.00015610588137453414, + "loss": 0.2166, + "step": 13248 + }, + { + "epoch": 0.31057742459138404, + "grad_norm": 0.2599966526031494, + "learning_rate": 0.0001560997836940974, + "loss": 0.0581, + "step": 13249 + }, + { + "epoch": 0.31060086616618904, + "grad_norm": 0.19946283102035522, + "learning_rate": 0.00015609368570926164, + "loss": 0.0273, + "step": 13250 + }, + { + "epoch": 0.31062430774099403, + "grad_norm": 0.4662315547466278, + "learning_rate": 0.00015608758742005995, + "loss": 0.093, + "step": 13251 + }, + { + "epoch": 0.310647749315799, + "grad_norm": 0.5443528294563293, + "learning_rate": 0.00015608148882652545, + "loss": 0.0864, + "step": 13252 + }, + { + "epoch": 0.310671190890604, + "grad_norm": 0.22686296701431274, + "learning_rate": 0.0001560753899286912, + "loss": 0.0414, + "step": 13253 + }, + { + "epoch": 0.310694632465409, + "grad_norm": 0.7070357203483582, + "learning_rate": 0.00015606929072659034, + "loss": 0.2054, + "step": 13254 + }, + { + "epoch": 0.310718074040214, + "grad_norm": 0.5301008224487305, + "learning_rate": 0.00015606319122025587, + "loss": 0.144, + "step": 13255 + }, + { + "epoch": 0.310741515615019, + "grad_norm": 0.5012030601501465, + "learning_rate": 0.000156057091409721, + "loss": 0.1056, + "step": 13256 + }, + { + "epoch": 0.310764957189824, + "grad_norm": 0.28243038058280945, + "learning_rate": 0.00015605099129501878, + "loss": 0.0802, + "step": 13257 + }, + { + "epoch": 0.310788398764629, + "grad_norm": 0.5350571870803833, + "learning_rate": 0.00015604489087618227, + "loss": 0.1199, + "step": 13258 + }, + { + "epoch": 0.310811840339434, + "grad_norm": 0.5633931756019592, + "learning_rate": 0.00015603879015324462, + "loss": 0.0616, + "step": 13259 + }, + { + "epoch": 0.310835281914239, + "grad_norm": 0.43501022458076477, + "learning_rate": 0.00015603268912623891, + "loss": 0.1343, + "step": 13260 + }, + { + "epoch": 0.310858723489044, + "grad_norm": 0.38579902052879333, + "learning_rate": 0.00015602658779519828, + "loss": 0.3856, + "step": 13261 + }, + { + "epoch": 0.31088216506384897, + "grad_norm": 0.39182907342910767, + "learning_rate": 0.0001560204861601558, + "loss": 0.0758, + "step": 13262 + }, + { + "epoch": 0.31090560663865396, + "grad_norm": 0.6125289797782898, + "learning_rate": 0.0001560143842211446, + "loss": 0.1286, + "step": 13263 + }, + { + "epoch": 0.31092904821345896, + "grad_norm": 0.45756179094314575, + "learning_rate": 0.00015600828197819775, + "loss": 0.112, + "step": 13264 + }, + { + "epoch": 0.31095248978826395, + "grad_norm": 0.5877964496612549, + "learning_rate": 0.0001560021794313484, + "loss": 0.6426, + "step": 13265 + }, + { + "epoch": 0.31097593136306895, + "grad_norm": 0.3411884009838104, + "learning_rate": 0.00015599607658062963, + "loss": 0.111, + "step": 13266 + }, + { + "epoch": 0.31099937293787394, + "grad_norm": 0.5980194807052612, + "learning_rate": 0.00015598997342607457, + "loss": 0.1694, + "step": 13267 + }, + { + "epoch": 0.31102281451267894, + "grad_norm": 0.1711965650320053, + "learning_rate": 0.00015598386996771638, + "loss": 0.0326, + "step": 13268 + }, + { + "epoch": 0.31104625608748393, + "grad_norm": 1.0316849946975708, + "learning_rate": 0.00015597776620558814, + "loss": 0.2008, + "step": 13269 + }, + { + "epoch": 0.311069697662289, + "grad_norm": 0.435947448015213, + "learning_rate": 0.0001559716621397229, + "loss": 0.0527, + "step": 13270 + }, + { + "epoch": 0.311093139237094, + "grad_norm": 0.6590081453323364, + "learning_rate": 0.0001559655577701539, + "loss": 0.1681, + "step": 13271 + }, + { + "epoch": 0.31111658081189897, + "grad_norm": 0.5315685272216797, + "learning_rate": 0.0001559594530969142, + "loss": 0.1671, + "step": 13272 + }, + { + "epoch": 0.31114002238670396, + "grad_norm": 0.265797883272171, + "learning_rate": 0.00015595334812003694, + "loss": 0.0526, + "step": 13273 + }, + { + "epoch": 0.31116346396150896, + "grad_norm": 0.4086514711380005, + "learning_rate": 0.00015594724283955524, + "loss": 0.0828, + "step": 13274 + }, + { + "epoch": 0.31118690553631395, + "grad_norm": 0.2389136701822281, + "learning_rate": 0.0001559411372555022, + "loss": 0.0682, + "step": 13275 + }, + { + "epoch": 0.31121034711111895, + "grad_norm": 0.478805273771286, + "learning_rate": 0.000155935031367911, + "loss": 0.6814, + "step": 13276 + }, + { + "epoch": 0.31123378868592394, + "grad_norm": 0.6641390323638916, + "learning_rate": 0.00015592892517681477, + "loss": 0.1857, + "step": 13277 + }, + { + "epoch": 0.31125723026072893, + "grad_norm": 0.6963496208190918, + "learning_rate": 0.00015592281868224655, + "loss": 0.1284, + "step": 13278 + }, + { + "epoch": 0.31128067183553393, + "grad_norm": 0.3480536937713623, + "learning_rate": 0.00015591671188423963, + "loss": 0.0424, + "step": 13279 + }, + { + "epoch": 0.3113041134103389, + "grad_norm": 0.267228901386261, + "learning_rate": 0.00015591060478282703, + "loss": 0.0767, + "step": 13280 + }, + { + "epoch": 0.3113275549851439, + "grad_norm": 0.6091201901435852, + "learning_rate": 0.00015590449737804186, + "loss": 0.6672, + "step": 13281 + }, + { + "epoch": 0.3113509965599489, + "grad_norm": 0.6032604575157166, + "learning_rate": 0.00015589838966991736, + "loss": 0.1762, + "step": 13282 + }, + { + "epoch": 0.3113744381347539, + "grad_norm": 0.5299364924430847, + "learning_rate": 0.00015589228165848663, + "loss": 0.6742, + "step": 13283 + }, + { + "epoch": 0.3113978797095589, + "grad_norm": 0.4160088896751404, + "learning_rate": 0.00015588617334378281, + "loss": 0.1252, + "step": 13284 + }, + { + "epoch": 0.3114213212843639, + "grad_norm": 0.6271082758903503, + "learning_rate": 0.00015588006472583903, + "loss": 0.0657, + "step": 13285 + }, + { + "epoch": 0.3114447628591689, + "grad_norm": 0.13641886413097382, + "learning_rate": 0.00015587395580468844, + "loss": 0.0269, + "step": 13286 + }, + { + "epoch": 0.3114682044339739, + "grad_norm": 0.5857502222061157, + "learning_rate": 0.00015586784658036423, + "loss": 0.1428, + "step": 13287 + }, + { + "epoch": 0.3114916460087789, + "grad_norm": 0.6753922700881958, + "learning_rate": 0.0001558617370528995, + "loss": 0.1871, + "step": 13288 + }, + { + "epoch": 0.31151508758358387, + "grad_norm": 0.131351500749588, + "learning_rate": 0.0001558556272223274, + "loss": 0.0428, + "step": 13289 + }, + { + "epoch": 0.31153852915838887, + "grad_norm": 0.6505394577980042, + "learning_rate": 0.00015584951708868107, + "loss": 0.1061, + "step": 13290 + }, + { + "epoch": 0.31156197073319386, + "grad_norm": 0.277459979057312, + "learning_rate": 0.00015584340665199373, + "loss": 0.0325, + "step": 13291 + }, + { + "epoch": 0.31158541230799885, + "grad_norm": 0.5989922881126404, + "learning_rate": 0.0001558372959122985, + "loss": 0.1224, + "step": 13292 + }, + { + "epoch": 0.31160885388280385, + "grad_norm": 0.35881146788597107, + "learning_rate": 0.00015583118486962852, + "loss": 0.1073, + "step": 13293 + }, + { + "epoch": 0.31163229545760884, + "grad_norm": 0.24278275668621063, + "learning_rate": 0.00015582507352401693, + "loss": 0.0664, + "step": 13294 + }, + { + "epoch": 0.31165573703241384, + "grad_norm": 0.41464853286743164, + "learning_rate": 0.00015581896187549695, + "loss": 0.0946, + "step": 13295 + }, + { + "epoch": 0.31167917860721883, + "grad_norm": 0.3701625168323517, + "learning_rate": 0.00015581284992410174, + "loss": 0.095, + "step": 13296 + }, + { + "epoch": 0.3117026201820238, + "grad_norm": 0.7179785966873169, + "learning_rate": 0.00015580673766986443, + "loss": 0.1595, + "step": 13297 + }, + { + "epoch": 0.3117260617568288, + "grad_norm": 0.4706164300441742, + "learning_rate": 0.00015580062511281816, + "loss": 0.0758, + "step": 13298 + }, + { + "epoch": 0.3117495033316338, + "grad_norm": 0.3826671242713928, + "learning_rate": 0.00015579451225299616, + "loss": 0.0621, + "step": 13299 + }, + { + "epoch": 0.3117729449064388, + "grad_norm": 0.3890572190284729, + "learning_rate": 0.0001557883990904315, + "loss": 0.0807, + "step": 13300 + }, + { + "epoch": 0.3117963864812438, + "grad_norm": 0.3442225158214569, + "learning_rate": 0.00015578228562515748, + "loss": 0.0877, + "step": 13301 + }, + { + "epoch": 0.3118198280560488, + "grad_norm": 0.3345697522163391, + "learning_rate": 0.0001557761718572072, + "loss": 0.0964, + "step": 13302 + }, + { + "epoch": 0.3118432696308538, + "grad_norm": 0.8022957444190979, + "learning_rate": 0.00015577005778661385, + "loss": 0.1294, + "step": 13303 + }, + { + "epoch": 0.3118667112056588, + "grad_norm": 0.48116379976272583, + "learning_rate": 0.00015576394341341058, + "loss": 0.1124, + "step": 13304 + }, + { + "epoch": 0.3118901527804638, + "grad_norm": 0.20251132547855377, + "learning_rate": 0.0001557578287376306, + "loss": 0.0519, + "step": 13305 + }, + { + "epoch": 0.3119135943552688, + "grad_norm": 0.32664361596107483, + "learning_rate": 0.0001557517137593071, + "loss": 0.0455, + "step": 13306 + }, + { + "epoch": 0.31193703593007377, + "grad_norm": 0.39705178141593933, + "learning_rate": 0.00015574559847847318, + "loss": 0.0908, + "step": 13307 + }, + { + "epoch": 0.31196047750487876, + "grad_norm": 0.41951704025268555, + "learning_rate": 0.0001557394828951621, + "loss": 0.0712, + "step": 13308 + }, + { + "epoch": 0.31198391907968376, + "grad_norm": 0.8533743619918823, + "learning_rate": 0.000155733367009407, + "loss": 0.0632, + "step": 13309 + }, + { + "epoch": 0.31200736065448875, + "grad_norm": 0.6361277103424072, + "learning_rate": 0.00015572725082124112, + "loss": 0.535, + "step": 13310 + }, + { + "epoch": 0.31203080222929375, + "grad_norm": 1.0339274406433105, + "learning_rate": 0.0001557211343306976, + "loss": 0.1435, + "step": 13311 + }, + { + "epoch": 0.31205424380409874, + "grad_norm": 0.3146798312664032, + "learning_rate": 0.00015571501753780962, + "loss": 0.0351, + "step": 13312 + }, + { + "epoch": 0.31207768537890374, + "grad_norm": 0.5493374466896057, + "learning_rate": 0.0001557089004426104, + "loss": 0.1503, + "step": 13313 + }, + { + "epoch": 0.31210112695370873, + "grad_norm": 0.4281613826751709, + "learning_rate": 0.00015570278304513307, + "loss": 0.1161, + "step": 13314 + }, + { + "epoch": 0.3121245685285137, + "grad_norm": 0.3343510329723358, + "learning_rate": 0.00015569666534541094, + "loss": 0.064, + "step": 13315 + }, + { + "epoch": 0.3121480101033187, + "grad_norm": 0.511125922203064, + "learning_rate": 0.00015569054734347712, + "loss": 0.0951, + "step": 13316 + }, + { + "epoch": 0.3121714516781237, + "grad_norm": 0.4475346505641937, + "learning_rate": 0.0001556844290393648, + "loss": 0.0761, + "step": 13317 + }, + { + "epoch": 0.3121948932529287, + "grad_norm": 0.736570417881012, + "learning_rate": 0.00015567831043310725, + "loss": 0.1703, + "step": 13318 + }, + { + "epoch": 0.3122183348277337, + "grad_norm": 0.5251924395561218, + "learning_rate": 0.00015567219152473757, + "loss": 0.088, + "step": 13319 + }, + { + "epoch": 0.3122417764025387, + "grad_norm": 0.9765495657920837, + "learning_rate": 0.00015566607231428905, + "loss": 0.0915, + "step": 13320 + }, + { + "epoch": 0.31226521797734375, + "grad_norm": 0.2908739447593689, + "learning_rate": 0.00015565995280179484, + "loss": 0.0934, + "step": 13321 + }, + { + "epoch": 0.31228865955214874, + "grad_norm": 0.35285574197769165, + "learning_rate": 0.00015565383298728818, + "loss": 0.0555, + "step": 13322 + }, + { + "epoch": 0.31231210112695373, + "grad_norm": 0.5937958359718323, + "learning_rate": 0.00015564771287080223, + "loss": 0.0716, + "step": 13323 + }, + { + "epoch": 0.31233554270175873, + "grad_norm": 0.4912911057472229, + "learning_rate": 0.00015564159245237024, + "loss": 0.1074, + "step": 13324 + }, + { + "epoch": 0.3123589842765637, + "grad_norm": 0.5289623141288757, + "learning_rate": 0.00015563547173202543, + "loss": 0.1289, + "step": 13325 + }, + { + "epoch": 0.3123824258513687, + "grad_norm": 0.6270827054977417, + "learning_rate": 0.00015562935070980092, + "loss": 0.1526, + "step": 13326 + }, + { + "epoch": 0.3124058674261737, + "grad_norm": 0.4659956693649292, + "learning_rate": 0.00015562322938573005, + "loss": 0.0634, + "step": 13327 + }, + { + "epoch": 0.3124293090009787, + "grad_norm": 0.9413291215896606, + "learning_rate": 0.00015561710775984595, + "loss": 0.1787, + "step": 13328 + }, + { + "epoch": 0.3124527505757837, + "grad_norm": 0.559782862663269, + "learning_rate": 0.00015561098583218185, + "loss": 0.1454, + "step": 13329 + }, + { + "epoch": 0.3124761921505887, + "grad_norm": 0.77032071352005, + "learning_rate": 0.000155604863602771, + "loss": 0.5984, + "step": 13330 + }, + { + "epoch": 0.3124996337253937, + "grad_norm": 0.7633260488510132, + "learning_rate": 0.00015559874107164658, + "loss": 0.1031, + "step": 13331 + }, + { + "epoch": 0.3125230753001987, + "grad_norm": 0.5057719349861145, + "learning_rate": 0.00015559261823884182, + "loss": 0.0909, + "step": 13332 + }, + { + "epoch": 0.3125465168750037, + "grad_norm": 0.12037025392055511, + "learning_rate": 0.00015558649510438998, + "loss": 0.0146, + "step": 13333 + }, + { + "epoch": 0.31256995844980867, + "grad_norm": 0.5522967576980591, + "learning_rate": 0.0001555803716683242, + "loss": 0.1031, + "step": 13334 + }, + { + "epoch": 0.31259340002461367, + "grad_norm": 0.17185916006565094, + "learning_rate": 0.00015557424793067782, + "loss": 0.0306, + "step": 13335 + }, + { + "epoch": 0.31261684159941866, + "grad_norm": 0.3096795678138733, + "learning_rate": 0.000155568123891484, + "loss": 0.0386, + "step": 13336 + }, + { + "epoch": 0.31264028317422365, + "grad_norm": 0.1507689207792282, + "learning_rate": 0.00015556199955077594, + "loss": 0.0366, + "step": 13337 + }, + { + "epoch": 0.31266372474902865, + "grad_norm": 0.4929085671901703, + "learning_rate": 0.00015555587490858694, + "loss": 0.1107, + "step": 13338 + }, + { + "epoch": 0.31268716632383364, + "grad_norm": 0.8124352097511292, + "learning_rate": 0.0001555497499649502, + "loss": 0.1476, + "step": 13339 + }, + { + "epoch": 0.31271060789863864, + "grad_norm": 0.28700992465019226, + "learning_rate": 0.00015554362471989895, + "loss": 0.0474, + "step": 13340 + }, + { + "epoch": 0.31273404947344363, + "grad_norm": 0.1671523153781891, + "learning_rate": 0.00015553749917346642, + "loss": 0.0351, + "step": 13341 + }, + { + "epoch": 0.3127574910482486, + "grad_norm": 0.4978989362716675, + "learning_rate": 0.0001555313733256859, + "loss": 0.0693, + "step": 13342 + }, + { + "epoch": 0.3127809326230536, + "grad_norm": 0.32578617334365845, + "learning_rate": 0.00015552524717659053, + "loss": 0.0634, + "step": 13343 + }, + { + "epoch": 0.3128043741978586, + "grad_norm": 0.11109765619039536, + "learning_rate": 0.00015551912072621363, + "loss": 0.0086, + "step": 13344 + }, + { + "epoch": 0.3128278157726636, + "grad_norm": 0.29625871777534485, + "learning_rate": 0.0001555129939745884, + "loss": 0.0796, + "step": 13345 + }, + { + "epoch": 0.3128512573474686, + "grad_norm": 0.19824473559856415, + "learning_rate": 0.00015550686692174815, + "loss": 0.0171, + "step": 13346 + }, + { + "epoch": 0.3128746989222736, + "grad_norm": 0.5042000412940979, + "learning_rate": 0.00015550073956772603, + "loss": 0.0716, + "step": 13347 + }, + { + "epoch": 0.3128981404970786, + "grad_norm": 0.5346840023994446, + "learning_rate": 0.00015549461191255537, + "loss": 0.0712, + "step": 13348 + }, + { + "epoch": 0.3129215820718836, + "grad_norm": 0.3837422728538513, + "learning_rate": 0.00015548848395626939, + "loss": 0.0495, + "step": 13349 + }, + { + "epoch": 0.3129450236466886, + "grad_norm": 0.3715527653694153, + "learning_rate": 0.0001554823556989013, + "loss": 0.1027, + "step": 13350 + }, + { + "epoch": 0.3129684652214936, + "grad_norm": 0.6681236028671265, + "learning_rate": 0.0001554762271404844, + "loss": 0.6055, + "step": 13351 + }, + { + "epoch": 0.31299190679629857, + "grad_norm": 0.15202227234840393, + "learning_rate": 0.00015547009828105197, + "loss": 0.0291, + "step": 13352 + }, + { + "epoch": 0.31301534837110356, + "grad_norm": 0.398305207490921, + "learning_rate": 0.0001554639691206372, + "loss": 0.0585, + "step": 13353 + }, + { + "epoch": 0.31303878994590856, + "grad_norm": 0.44292399287223816, + "learning_rate": 0.00015545783965927336, + "loss": 0.1107, + "step": 13354 + }, + { + "epoch": 0.31306223152071355, + "grad_norm": 0.7732597589492798, + "learning_rate": 0.00015545170989699373, + "loss": 0.1176, + "step": 13355 + }, + { + "epoch": 0.31308567309551855, + "grad_norm": 0.5292845964431763, + "learning_rate": 0.0001554455798338316, + "loss": 0.1571, + "step": 13356 + }, + { + "epoch": 0.31310911467032354, + "grad_norm": 0.5842336416244507, + "learning_rate": 0.00015543944946982014, + "loss": 0.1636, + "step": 13357 + }, + { + "epoch": 0.31313255624512853, + "grad_norm": 0.40438708662986755, + "learning_rate": 0.00015543331880499267, + "loss": 0.0529, + "step": 13358 + }, + { + "epoch": 0.31315599781993353, + "grad_norm": 0.20741751790046692, + "learning_rate": 0.00015542718783938248, + "loss": 0.0423, + "step": 13359 + }, + { + "epoch": 0.3131794393947385, + "grad_norm": 0.5672323107719421, + "learning_rate": 0.0001554210565730228, + "loss": 0.1241, + "step": 13360 + }, + { + "epoch": 0.3132028809695435, + "grad_norm": 0.525500476360321, + "learning_rate": 0.00015541492500594693, + "loss": 0.1347, + "step": 13361 + }, + { + "epoch": 0.3132263225443485, + "grad_norm": 0.8236292600631714, + "learning_rate": 0.0001554087931381881, + "loss": 0.1634, + "step": 13362 + }, + { + "epoch": 0.3132497641191535, + "grad_norm": 0.43015947937965393, + "learning_rate": 0.00015540266096977957, + "loss": 0.1127, + "step": 13363 + }, + { + "epoch": 0.3132732056939585, + "grad_norm": 0.454070508480072, + "learning_rate": 0.0001553965285007547, + "loss": 0.0808, + "step": 13364 + }, + { + "epoch": 0.3132966472687635, + "grad_norm": 0.39311015605926514, + "learning_rate": 0.00015539039573114668, + "loss": 0.1138, + "step": 13365 + }, + { + "epoch": 0.3133200888435685, + "grad_norm": 0.6133456826210022, + "learning_rate": 0.00015538426266098883, + "loss": 0.1201, + "step": 13366 + }, + { + "epoch": 0.3133435304183735, + "grad_norm": 0.12822064757347107, + "learning_rate": 0.0001553781292903144, + "loss": 0.0184, + "step": 13367 + }, + { + "epoch": 0.3133669719931785, + "grad_norm": 0.649433434009552, + "learning_rate": 0.00015537199561915669, + "loss": 0.1244, + "step": 13368 + }, + { + "epoch": 0.3133904135679835, + "grad_norm": 0.8049819469451904, + "learning_rate": 0.00015536586164754897, + "loss": 0.139, + "step": 13369 + }, + { + "epoch": 0.31341385514278847, + "grad_norm": 0.7416811585426331, + "learning_rate": 0.00015535972737552452, + "loss": 0.2538, + "step": 13370 + }, + { + "epoch": 0.31343729671759346, + "grad_norm": 0.3337767720222473, + "learning_rate": 0.00015535359280311665, + "loss": 0.0666, + "step": 13371 + }, + { + "epoch": 0.3134607382923985, + "grad_norm": 0.4560546875, + "learning_rate": 0.0001553474579303586, + "loss": 0.1036, + "step": 13372 + }, + { + "epoch": 0.3134841798672035, + "grad_norm": 0.3489971458911896, + "learning_rate": 0.00015534132275728374, + "loss": 0.0584, + "step": 13373 + }, + { + "epoch": 0.3135076214420085, + "grad_norm": 0.5396766066551208, + "learning_rate": 0.00015533518728392526, + "loss": 0.1182, + "step": 13374 + }, + { + "epoch": 0.3135310630168135, + "grad_norm": 0.5821441411972046, + "learning_rate": 0.00015532905151031652, + "loss": 0.0964, + "step": 13375 + }, + { + "epoch": 0.3135545045916185, + "grad_norm": 0.3754849433898926, + "learning_rate": 0.00015532291543649077, + "loss": 0.0496, + "step": 13376 + }, + { + "epoch": 0.3135779461664235, + "grad_norm": 0.37354421615600586, + "learning_rate": 0.00015531677906248134, + "loss": 0.068, + "step": 13377 + }, + { + "epoch": 0.3136013877412285, + "grad_norm": 0.14432942867279053, + "learning_rate": 0.00015531064238832152, + "loss": 0.0164, + "step": 13378 + }, + { + "epoch": 0.31362482931603347, + "grad_norm": 0.3322993516921997, + "learning_rate": 0.0001553045054140446, + "loss": 0.0956, + "step": 13379 + }, + { + "epoch": 0.31364827089083847, + "grad_norm": 0.2219143807888031, + "learning_rate": 0.00015529836813968386, + "loss": 0.0307, + "step": 13380 + }, + { + "epoch": 0.31367171246564346, + "grad_norm": 0.4942905604839325, + "learning_rate": 0.00015529223056527262, + "loss": 0.0485, + "step": 13381 + }, + { + "epoch": 0.31369515404044845, + "grad_norm": 0.339200884103775, + "learning_rate": 0.0001552860926908442, + "loss": 0.0622, + "step": 13382 + }, + { + "epoch": 0.31371859561525345, + "grad_norm": 0.5302256941795349, + "learning_rate": 0.00015527995451643188, + "loss": 0.2227, + "step": 13383 + }, + { + "epoch": 0.31374203719005844, + "grad_norm": 0.7855716347694397, + "learning_rate": 0.00015527381604206898, + "loss": 0.0618, + "step": 13384 + }, + { + "epoch": 0.31376547876486344, + "grad_norm": 0.3469829261302948, + "learning_rate": 0.00015526767726778878, + "loss": 0.0532, + "step": 13385 + }, + { + "epoch": 0.31378892033966843, + "grad_norm": 0.43196961283683777, + "learning_rate": 0.0001552615381936246, + "loss": 0.0737, + "step": 13386 + }, + { + "epoch": 0.3138123619144734, + "grad_norm": 0.44748926162719727, + "learning_rate": 0.0001552553988196098, + "loss": 0.4768, + "step": 13387 + }, + { + "epoch": 0.3138358034892784, + "grad_norm": 0.6978621482849121, + "learning_rate": 0.0001552492591457776, + "loss": 0.0651, + "step": 13388 + }, + { + "epoch": 0.3138592450640834, + "grad_norm": 0.10974444448947906, + "learning_rate": 0.0001552431191721614, + "loss": 0.02, + "step": 13389 + }, + { + "epoch": 0.3138826866388884, + "grad_norm": 0.5597538948059082, + "learning_rate": 0.0001552369788987945, + "loss": 0.1127, + "step": 13390 + }, + { + "epoch": 0.3139061282136934, + "grad_norm": 0.3391117751598358, + "learning_rate": 0.00015523083832571016, + "loss": 0.0518, + "step": 13391 + }, + { + "epoch": 0.3139295697884984, + "grad_norm": 0.1668066829442978, + "learning_rate": 0.00015522469745294176, + "loss": 0.0345, + "step": 13392 + }, + { + "epoch": 0.3139530113633034, + "grad_norm": 0.5525522232055664, + "learning_rate": 0.00015521855628052259, + "loss": 0.1096, + "step": 13393 + }, + { + "epoch": 0.3139764529381084, + "grad_norm": 0.7719630599021912, + "learning_rate": 0.000155212414808486, + "loss": 0.2579, + "step": 13394 + }, + { + "epoch": 0.3139998945129134, + "grad_norm": 0.41554656624794006, + "learning_rate": 0.00015520627303686527, + "loss": 0.1244, + "step": 13395 + }, + { + "epoch": 0.3140233360877184, + "grad_norm": 0.6629656553268433, + "learning_rate": 0.00015520013096569378, + "loss": 0.1074, + "step": 13396 + }, + { + "epoch": 0.31404677766252337, + "grad_norm": 0.6109104156494141, + "learning_rate": 0.0001551939885950048, + "loss": 0.1345, + "step": 13397 + }, + { + "epoch": 0.31407021923732836, + "grad_norm": 0.5313289761543274, + "learning_rate": 0.00015518784592483172, + "loss": 0.1192, + "step": 13398 + }, + { + "epoch": 0.31409366081213336, + "grad_norm": 0.31233853101730347, + "learning_rate": 0.00015518170295520777, + "loss": 0.079, + "step": 13399 + }, + { + "epoch": 0.31411710238693835, + "grad_norm": 0.26653873920440674, + "learning_rate": 0.0001551755596861664, + "loss": 0.0401, + "step": 13400 + }, + { + "epoch": 0.31414054396174335, + "grad_norm": 0.3624722361564636, + "learning_rate": 0.00015516941611774088, + "loss": 0.0512, + "step": 13401 + }, + { + "epoch": 0.31416398553654834, + "grad_norm": 0.22676461935043335, + "learning_rate": 0.00015516327224996456, + "loss": 0.02, + "step": 13402 + }, + { + "epoch": 0.31418742711135333, + "grad_norm": 0.8027969598770142, + "learning_rate": 0.00015515712808287079, + "loss": 0.0934, + "step": 13403 + }, + { + "epoch": 0.31421086868615833, + "grad_norm": 0.40706297755241394, + "learning_rate": 0.00015515098361649287, + "loss": 0.0651, + "step": 13404 + }, + { + "epoch": 0.3142343102609633, + "grad_norm": 0.6522338390350342, + "learning_rate": 0.00015514483885086414, + "loss": 0.1151, + "step": 13405 + }, + { + "epoch": 0.3142577518357683, + "grad_norm": 0.38975849747657776, + "learning_rate": 0.000155138693786018, + "loss": 0.0559, + "step": 13406 + }, + { + "epoch": 0.3142811934105733, + "grad_norm": 0.5844976902008057, + "learning_rate": 0.00015513254842198775, + "loss": 0.1117, + "step": 13407 + }, + { + "epoch": 0.3143046349853783, + "grad_norm": 0.49992048740386963, + "learning_rate": 0.00015512640275880675, + "loss": 0.1275, + "step": 13408 + }, + { + "epoch": 0.3143280765601833, + "grad_norm": 0.3177293539047241, + "learning_rate": 0.00015512025679650834, + "loss": 0.0729, + "step": 13409 + }, + { + "epoch": 0.3143515181349883, + "grad_norm": 0.34968823194503784, + "learning_rate": 0.00015511411053512584, + "loss": 0.049, + "step": 13410 + }, + { + "epoch": 0.3143749597097933, + "grad_norm": 0.2863573133945465, + "learning_rate": 0.00015510796397469264, + "loss": 0.0488, + "step": 13411 + }, + { + "epoch": 0.3143984012845983, + "grad_norm": 0.24241362512111664, + "learning_rate": 0.00015510181711524208, + "loss": 0.0602, + "step": 13412 + }, + { + "epoch": 0.3144218428594033, + "grad_norm": 0.5164082050323486, + "learning_rate": 0.00015509566995680748, + "loss": 0.1004, + "step": 13413 + }, + { + "epoch": 0.31444528443420827, + "grad_norm": 0.48118534684181213, + "learning_rate": 0.0001550895224994223, + "loss": 0.4043, + "step": 13414 + }, + { + "epoch": 0.31446872600901327, + "grad_norm": 0.22200073301792145, + "learning_rate": 0.0001550833747431198, + "loss": 0.039, + "step": 13415 + }, + { + "epoch": 0.31449216758381826, + "grad_norm": 0.5413135290145874, + "learning_rate": 0.0001550772266879333, + "loss": 0.1419, + "step": 13416 + }, + { + "epoch": 0.31451560915862325, + "grad_norm": 0.3249415159225464, + "learning_rate": 0.00015507107833389623, + "loss": 0.0595, + "step": 13417 + }, + { + "epoch": 0.31453905073342825, + "grad_norm": 0.6285447478294373, + "learning_rate": 0.00015506492968104198, + "loss": 0.1153, + "step": 13418 + }, + { + "epoch": 0.31456249230823324, + "grad_norm": 0.22240114212036133, + "learning_rate": 0.00015505878072940386, + "loss": 0.0576, + "step": 13419 + }, + { + "epoch": 0.31458593388303824, + "grad_norm": 0.5953702330589294, + "learning_rate": 0.00015505263147901527, + "loss": 0.4788, + "step": 13420 + }, + { + "epoch": 0.31460937545784323, + "grad_norm": 0.14414046704769135, + "learning_rate": 0.00015504648192990952, + "loss": 0.0266, + "step": 13421 + }, + { + "epoch": 0.3146328170326482, + "grad_norm": 0.9223267436027527, + "learning_rate": 0.00015504033208212006, + "loss": 0.1539, + "step": 13422 + }, + { + "epoch": 0.3146562586074533, + "grad_norm": 0.7046468257904053, + "learning_rate": 0.00015503418193568014, + "loss": 0.0822, + "step": 13423 + }, + { + "epoch": 0.31467970018225827, + "grad_norm": 0.5693347454071045, + "learning_rate": 0.00015502803149062324, + "loss": 0.1192, + "step": 13424 + }, + { + "epoch": 0.31470314175706326, + "grad_norm": 0.26394277811050415, + "learning_rate": 0.00015502188074698268, + "loss": 0.0308, + "step": 13425 + }, + { + "epoch": 0.31472658333186826, + "grad_norm": 0.12739042937755585, + "learning_rate": 0.0001550157297047919, + "loss": 0.0327, + "step": 13426 + }, + { + "epoch": 0.31475002490667325, + "grad_norm": 0.6487534642219543, + "learning_rate": 0.00015500957836408414, + "loss": 0.1418, + "step": 13427 + }, + { + "epoch": 0.31477346648147825, + "grad_norm": 0.6626314520835876, + "learning_rate": 0.0001550034267248929, + "loss": 0.1404, + "step": 13428 + }, + { + "epoch": 0.31479690805628324, + "grad_norm": 0.37463510036468506, + "learning_rate": 0.00015499727478725153, + "loss": 0.0792, + "step": 13429 + }, + { + "epoch": 0.31482034963108824, + "grad_norm": 0.9990425109863281, + "learning_rate": 0.0001549911225511934, + "loss": 0.2015, + "step": 13430 + }, + { + "epoch": 0.31484379120589323, + "grad_norm": 0.24942168593406677, + "learning_rate": 0.00015498497001675186, + "loss": 0.0552, + "step": 13431 + }, + { + "epoch": 0.3148672327806982, + "grad_norm": 0.33140867948532104, + "learning_rate": 0.00015497881718396037, + "loss": 0.0744, + "step": 13432 + }, + { + "epoch": 0.3148906743555032, + "grad_norm": 0.5747031569480896, + "learning_rate": 0.00015497266405285225, + "loss": 0.0908, + "step": 13433 + }, + { + "epoch": 0.3149141159303082, + "grad_norm": 0.46533796191215515, + "learning_rate": 0.00015496651062346092, + "loss": 0.1087, + "step": 13434 + }, + { + "epoch": 0.3149375575051132, + "grad_norm": 0.674754798412323, + "learning_rate": 0.00015496035689581975, + "loss": 0.1314, + "step": 13435 + }, + { + "epoch": 0.3149609990799182, + "grad_norm": 0.5078964829444885, + "learning_rate": 0.00015495420286996213, + "loss": 0.1269, + "step": 13436 + }, + { + "epoch": 0.3149844406547232, + "grad_norm": 0.38522017002105713, + "learning_rate": 0.00015494804854592145, + "loss": 0.1136, + "step": 13437 + }, + { + "epoch": 0.3150078822295282, + "grad_norm": 0.36127549409866333, + "learning_rate": 0.00015494189392373117, + "loss": 0.0954, + "step": 13438 + }, + { + "epoch": 0.3150313238043332, + "grad_norm": 0.12806937098503113, + "learning_rate": 0.0001549357390034246, + "loss": 0.0107, + "step": 13439 + }, + { + "epoch": 0.3150547653791382, + "grad_norm": 0.3007606267929077, + "learning_rate": 0.00015492958378503512, + "loss": 0.0569, + "step": 13440 + }, + { + "epoch": 0.3150782069539432, + "grad_norm": 0.24566856026649475, + "learning_rate": 0.00015492342826859624, + "loss": 0.0477, + "step": 13441 + }, + { + "epoch": 0.31510164852874817, + "grad_norm": 0.49849268794059753, + "learning_rate": 0.00015491727245414125, + "loss": 0.14, + "step": 13442 + }, + { + "epoch": 0.31512509010355316, + "grad_norm": 0.1647481620311737, + "learning_rate": 0.0001549111163417036, + "loss": 0.0396, + "step": 13443 + }, + { + "epoch": 0.31514853167835816, + "grad_norm": 0.643129825592041, + "learning_rate": 0.00015490495993131673, + "loss": 0.1439, + "step": 13444 + }, + { + "epoch": 0.31517197325316315, + "grad_norm": 0.27974024415016174, + "learning_rate": 0.00015489880322301395, + "loss": 0.0565, + "step": 13445 + }, + { + "epoch": 0.31519541482796815, + "grad_norm": 0.4152795672416687, + "learning_rate": 0.00015489264621682875, + "loss": 0.1102, + "step": 13446 + }, + { + "epoch": 0.31521885640277314, + "grad_norm": 0.1418505162000656, + "learning_rate": 0.00015488648891279452, + "loss": 0.0379, + "step": 13447 + }, + { + "epoch": 0.31524229797757813, + "grad_norm": 0.25020870566368103, + "learning_rate": 0.00015488033131094462, + "loss": 0.0347, + "step": 13448 + }, + { + "epoch": 0.31526573955238313, + "grad_norm": 0.6465421319007874, + "learning_rate": 0.0001548741734113125, + "loss": 0.0577, + "step": 13449 + }, + { + "epoch": 0.3152891811271881, + "grad_norm": 0.4840943515300751, + "learning_rate": 0.0001548680152139316, + "loss": 0.1097, + "step": 13450 + }, + { + "epoch": 0.3153126227019931, + "grad_norm": 0.45933762192726135, + "learning_rate": 0.0001548618567188353, + "loss": 0.1151, + "step": 13451 + }, + { + "epoch": 0.3153360642767981, + "grad_norm": 0.16798032820224762, + "learning_rate": 0.00015485569792605703, + "loss": 0.0476, + "step": 13452 + }, + { + "epoch": 0.3153595058516031, + "grad_norm": 0.4251447916030884, + "learning_rate": 0.00015484953883563018, + "loss": 0.0991, + "step": 13453 + }, + { + "epoch": 0.3153829474264081, + "grad_norm": 0.6336937546730042, + "learning_rate": 0.0001548433794475882, + "loss": 0.2192, + "step": 13454 + }, + { + "epoch": 0.3154063890012131, + "grad_norm": 0.5281418561935425, + "learning_rate": 0.0001548372197619645, + "loss": 0.1396, + "step": 13455 + }, + { + "epoch": 0.3154298305760181, + "grad_norm": 0.5489970445632935, + "learning_rate": 0.0001548310597787925, + "loss": 0.142, + "step": 13456 + }, + { + "epoch": 0.3154532721508231, + "grad_norm": 0.33522072434425354, + "learning_rate": 0.00015482489949810563, + "loss": 0.0691, + "step": 13457 + }, + { + "epoch": 0.3154767137256281, + "grad_norm": 0.08412305265665054, + "learning_rate": 0.00015481873891993732, + "loss": 0.0178, + "step": 13458 + }, + { + "epoch": 0.31550015530043307, + "grad_norm": 0.6399938464164734, + "learning_rate": 0.00015481257804432098, + "loss": 0.1172, + "step": 13459 + }, + { + "epoch": 0.31552359687523807, + "grad_norm": 0.5365886688232422, + "learning_rate": 0.00015480641687129007, + "loss": 0.1433, + "step": 13460 + }, + { + "epoch": 0.31554703845004306, + "grad_norm": 0.3992297649383545, + "learning_rate": 0.00015480025540087799, + "loss": 0.1068, + "step": 13461 + }, + { + "epoch": 0.31557048002484805, + "grad_norm": 0.48322513699531555, + "learning_rate": 0.00015479409363311822, + "loss": 0.1912, + "step": 13462 + }, + { + "epoch": 0.31559392159965305, + "grad_norm": 0.16190484166145325, + "learning_rate": 0.0001547879315680441, + "loss": 0.0393, + "step": 13463 + }, + { + "epoch": 0.31561736317445804, + "grad_norm": 0.5049543976783752, + "learning_rate": 0.00015478176920568916, + "loss": 0.1124, + "step": 13464 + }, + { + "epoch": 0.31564080474926304, + "grad_norm": 0.22951722145080566, + "learning_rate": 0.00015477560654608675, + "loss": 0.0422, + "step": 13465 + }, + { + "epoch": 0.31566424632406803, + "grad_norm": 0.10595162957906723, + "learning_rate": 0.0001547694435892704, + "loss": 0.016, + "step": 13466 + }, + { + "epoch": 0.315687687898873, + "grad_norm": 0.6825752258300781, + "learning_rate": 0.0001547632803352735, + "loss": 0.1753, + "step": 13467 + }, + { + "epoch": 0.315711129473678, + "grad_norm": 0.47798582911491394, + "learning_rate": 0.0001547571167841295, + "loss": 0.7297, + "step": 13468 + }, + { + "epoch": 0.315734571048483, + "grad_norm": 0.40673255920410156, + "learning_rate": 0.00015475095293587186, + "loss": 0.0594, + "step": 13469 + }, + { + "epoch": 0.315758012623288, + "grad_norm": 0.4206765294075012, + "learning_rate": 0.000154744788790534, + "loss": 0.1413, + "step": 13470 + }, + { + "epoch": 0.315781454198093, + "grad_norm": 0.24760794639587402, + "learning_rate": 0.00015473862434814937, + "loss": 0.0347, + "step": 13471 + }, + { + "epoch": 0.315804895772898, + "grad_norm": 0.4364030659198761, + "learning_rate": 0.00015473245960875143, + "loss": 0.0753, + "step": 13472 + }, + { + "epoch": 0.315828337347703, + "grad_norm": 0.5757149457931519, + "learning_rate": 0.00015472629457237362, + "loss": 0.1093, + "step": 13473 + }, + { + "epoch": 0.315851778922508, + "grad_norm": 0.34201619029045105, + "learning_rate": 0.0001547201292390494, + "loss": 0.1004, + "step": 13474 + }, + { + "epoch": 0.31587522049731304, + "grad_norm": 0.5158131718635559, + "learning_rate": 0.00015471396360881222, + "loss": 0.1281, + "step": 13475 + }, + { + "epoch": 0.31589866207211803, + "grad_norm": 0.6910024285316467, + "learning_rate": 0.0001547077976816955, + "loss": 0.0787, + "step": 13476 + }, + { + "epoch": 0.315922103646923, + "grad_norm": 0.5200402736663818, + "learning_rate": 0.0001547016314577328, + "loss": 0.1353, + "step": 13477 + }, + { + "epoch": 0.315945545221728, + "grad_norm": 0.4286903142929077, + "learning_rate": 0.00015469546493695745, + "loss": 0.0497, + "step": 13478 + }, + { + "epoch": 0.315968986796533, + "grad_norm": 0.4646199941635132, + "learning_rate": 0.00015468929811940303, + "loss": 0.0929, + "step": 13479 + }, + { + "epoch": 0.315992428371338, + "grad_norm": 0.19133129715919495, + "learning_rate": 0.0001546831310051029, + "loss": 0.0483, + "step": 13480 + }, + { + "epoch": 0.316015869946143, + "grad_norm": 0.41061198711395264, + "learning_rate": 0.00015467696359409055, + "loss": 0.1302, + "step": 13481 + }, + { + "epoch": 0.316039311520948, + "grad_norm": 0.2216186225414276, + "learning_rate": 0.0001546707958863995, + "loss": 0.0433, + "step": 13482 + }, + { + "epoch": 0.316062753095753, + "grad_norm": 0.9013839364051819, + "learning_rate": 0.00015466462788206315, + "loss": 0.1473, + "step": 13483 + }, + { + "epoch": 0.316086194670558, + "grad_norm": 0.45770135521888733, + "learning_rate": 0.000154658459581115, + "loss": 0.0971, + "step": 13484 + }, + { + "epoch": 0.316109636245363, + "grad_norm": 0.6846947073936462, + "learning_rate": 0.00015465229098358848, + "loss": 0.5846, + "step": 13485 + }, + { + "epoch": 0.316133077820168, + "grad_norm": 0.555837869644165, + "learning_rate": 0.00015464612208951714, + "loss": 0.1043, + "step": 13486 + }, + { + "epoch": 0.31615651939497297, + "grad_norm": 0.5782040357589722, + "learning_rate": 0.00015463995289893438, + "loss": 0.1574, + "step": 13487 + }, + { + "epoch": 0.31617996096977796, + "grad_norm": 0.8870754837989807, + "learning_rate": 0.00015463378341187372, + "loss": 0.1957, + "step": 13488 + }, + { + "epoch": 0.31620340254458296, + "grad_norm": 0.4091506898403168, + "learning_rate": 0.00015462761362836859, + "loss": 0.0978, + "step": 13489 + }, + { + "epoch": 0.31622684411938795, + "grad_norm": 1.0146561861038208, + "learning_rate": 0.0001546214435484525, + "loss": 0.1076, + "step": 13490 + }, + { + "epoch": 0.31625028569419295, + "grad_norm": 0.5447062849998474, + "learning_rate": 0.0001546152731721589, + "loss": 0.1312, + "step": 13491 + }, + { + "epoch": 0.31627372726899794, + "grad_norm": 0.7762374877929688, + "learning_rate": 0.00015460910249952132, + "loss": 0.146, + "step": 13492 + }, + { + "epoch": 0.31629716884380293, + "grad_norm": 0.8044276237487793, + "learning_rate": 0.0001546029315305732, + "loss": 0.1657, + "step": 13493 + }, + { + "epoch": 0.31632061041860793, + "grad_norm": 0.10982617735862732, + "learning_rate": 0.00015459676026534805, + "loss": 0.0297, + "step": 13494 + }, + { + "epoch": 0.3163440519934129, + "grad_norm": 0.22944481670856476, + "learning_rate": 0.00015459058870387937, + "loss": 0.0495, + "step": 13495 + }, + { + "epoch": 0.3163674935682179, + "grad_norm": 0.43258824944496155, + "learning_rate": 0.00015458441684620055, + "loss": 0.0714, + "step": 13496 + }, + { + "epoch": 0.3163909351430229, + "grad_norm": 0.3174508810043335, + "learning_rate": 0.0001545782446923452, + "loss": 0.0267, + "step": 13497 + }, + { + "epoch": 0.3164143767178279, + "grad_norm": 0.2539365887641907, + "learning_rate": 0.00015457207224234674, + "loss": 0.0677, + "step": 13498 + }, + { + "epoch": 0.3164378182926329, + "grad_norm": 0.3829181492328644, + "learning_rate": 0.00015456589949623873, + "loss": 0.0722, + "step": 13499 + }, + { + "epoch": 0.3164612598674379, + "grad_norm": 0.2725624740123749, + "learning_rate": 0.0001545597264540546, + "loss": 0.0634, + "step": 13500 + }, + { + "epoch": 0.3164847014422429, + "grad_norm": 0.5421427488327026, + "learning_rate": 0.0001545535531158278, + "loss": 0.0996, + "step": 13501 + }, + { + "epoch": 0.3165081430170479, + "grad_norm": 0.6806311011314392, + "learning_rate": 0.00015454737948159193, + "loss": 0.0971, + "step": 13502 + }, + { + "epoch": 0.3165315845918529, + "grad_norm": 0.2512824237346649, + "learning_rate": 0.00015454120555138045, + "loss": 0.0481, + "step": 13503 + }, + { + "epoch": 0.31655502616665787, + "grad_norm": 1.0405066013336182, + "learning_rate": 0.00015453503132522683, + "loss": 0.2168, + "step": 13504 + }, + { + "epoch": 0.31657846774146287, + "grad_norm": 0.4005858302116394, + "learning_rate": 0.00015452885680316463, + "loss": 0.038, + "step": 13505 + }, + { + "epoch": 0.31660190931626786, + "grad_norm": 0.396589994430542, + "learning_rate": 0.00015452268198522728, + "loss": 0.0836, + "step": 13506 + }, + { + "epoch": 0.31662535089107285, + "grad_norm": 0.22885358333587646, + "learning_rate": 0.00015451650687144835, + "loss": 0.0301, + "step": 13507 + }, + { + "epoch": 0.31664879246587785, + "grad_norm": 0.5344467759132385, + "learning_rate": 0.0001545103314618613, + "loss": 0.1232, + "step": 13508 + }, + { + "epoch": 0.31667223404068284, + "grad_norm": 0.6650007963180542, + "learning_rate": 0.00015450415575649968, + "loss": 0.1062, + "step": 13509 + }, + { + "epoch": 0.31669567561548784, + "grad_norm": 0.6798468232154846, + "learning_rate": 0.00015449797975539698, + "loss": 0.1554, + "step": 13510 + }, + { + "epoch": 0.31671911719029283, + "grad_norm": 1.16547691822052, + "learning_rate": 0.00015449180345858673, + "loss": 0.1543, + "step": 13511 + }, + { + "epoch": 0.3167425587650978, + "grad_norm": 0.2231171876192093, + "learning_rate": 0.00015448562686610237, + "loss": 0.0432, + "step": 13512 + }, + { + "epoch": 0.3167660003399028, + "grad_norm": 0.6346610188484192, + "learning_rate": 0.0001544794499779775, + "loss": 0.7486, + "step": 13513 + }, + { + "epoch": 0.3167894419147078, + "grad_norm": 0.4179147779941559, + "learning_rate": 0.0001544732727942456, + "loss": 0.0849, + "step": 13514 + }, + { + "epoch": 0.3168128834895128, + "grad_norm": 0.3569653332233429, + "learning_rate": 0.00015446709531494017, + "loss": 0.0943, + "step": 13515 + }, + { + "epoch": 0.3168363250643178, + "grad_norm": 0.4943106770515442, + "learning_rate": 0.00015446091754009477, + "loss": 0.0885, + "step": 13516 + }, + { + "epoch": 0.3168597666391228, + "grad_norm": 0.5490493178367615, + "learning_rate": 0.0001544547394697429, + "loss": 0.1915, + "step": 13517 + }, + { + "epoch": 0.3168832082139278, + "grad_norm": 0.23346032202243805, + "learning_rate": 0.00015444856110391806, + "loss": 0.0536, + "step": 13518 + }, + { + "epoch": 0.3169066497887328, + "grad_norm": 0.19498178362846375, + "learning_rate": 0.00015444238244265382, + "loss": 0.0347, + "step": 13519 + }, + { + "epoch": 0.3169300913635378, + "grad_norm": 0.673994243144989, + "learning_rate": 0.0001544362034859837, + "loss": 0.5396, + "step": 13520 + }, + { + "epoch": 0.3169535329383428, + "grad_norm": 0.338032066822052, + "learning_rate": 0.00015443002423394116, + "loss": 0.0567, + "step": 13521 + }, + { + "epoch": 0.31697697451314777, + "grad_norm": 0.48718854784965515, + "learning_rate": 0.0001544238446865598, + "loss": 0.6261, + "step": 13522 + }, + { + "epoch": 0.31700041608795276, + "grad_norm": 0.7086297273635864, + "learning_rate": 0.00015441766484387313, + "loss": 0.1342, + "step": 13523 + }, + { + "epoch": 0.31702385766275776, + "grad_norm": 0.12649953365325928, + "learning_rate": 0.0001544114847059147, + "loss": 0.0282, + "step": 13524 + }, + { + "epoch": 0.31704729923756275, + "grad_norm": 0.39768165349960327, + "learning_rate": 0.000154405304272718, + "loss": 0.584, + "step": 13525 + }, + { + "epoch": 0.3170707408123678, + "grad_norm": 0.15918874740600586, + "learning_rate": 0.0001543991235443166, + "loss": 0.0445, + "step": 13526 + }, + { + "epoch": 0.3170941823871728, + "grad_norm": 0.25171202421188354, + "learning_rate": 0.00015439294252074398, + "loss": 0.2168, + "step": 13527 + }, + { + "epoch": 0.3171176239619778, + "grad_norm": 0.4042486548423767, + "learning_rate": 0.0001543867612020338, + "loss": 0.1113, + "step": 13528 + }, + { + "epoch": 0.3171410655367828, + "grad_norm": 0.3953385353088379, + "learning_rate": 0.00015438057958821948, + "loss": 0.0646, + "step": 13529 + }, + { + "epoch": 0.3171645071115878, + "grad_norm": 0.36037343740463257, + "learning_rate": 0.00015437439767933465, + "loss": 0.1029, + "step": 13530 + }, + { + "epoch": 0.3171879486863928, + "grad_norm": 0.11296293139457703, + "learning_rate": 0.00015436821547541274, + "loss": 0.0208, + "step": 13531 + }, + { + "epoch": 0.31721139026119777, + "grad_norm": 0.12314922362565994, + "learning_rate": 0.00015436203297648738, + "loss": 0.0267, + "step": 13532 + }, + { + "epoch": 0.31723483183600276, + "grad_norm": 0.816228985786438, + "learning_rate": 0.00015435585018259214, + "loss": 0.0767, + "step": 13533 + }, + { + "epoch": 0.31725827341080776, + "grad_norm": 0.30735647678375244, + "learning_rate": 0.00015434966709376052, + "loss": 0.0766, + "step": 13534 + }, + { + "epoch": 0.31728171498561275, + "grad_norm": 0.38878241181373596, + "learning_rate": 0.00015434348371002606, + "loss": 0.0576, + "step": 13535 + }, + { + "epoch": 0.31730515656041774, + "grad_norm": 0.3440639078617096, + "learning_rate": 0.00015433730003142234, + "loss": 0.0643, + "step": 13536 + }, + { + "epoch": 0.31732859813522274, + "grad_norm": 0.5807008147239685, + "learning_rate": 0.0001543311160579829, + "loss": 0.156, + "step": 13537 + }, + { + "epoch": 0.31735203971002773, + "grad_norm": 1.5451292991638184, + "learning_rate": 0.0001543249317897413, + "loss": 0.1534, + "step": 13538 + }, + { + "epoch": 0.31737548128483273, + "grad_norm": 0.4192039668560028, + "learning_rate": 0.0001543187472267311, + "loss": 0.0583, + "step": 13539 + }, + { + "epoch": 0.3173989228596377, + "grad_norm": 0.361210435628891, + "learning_rate": 0.0001543125623689858, + "loss": 0.0841, + "step": 13540 + }, + { + "epoch": 0.3174223644344427, + "grad_norm": 0.33309927582740784, + "learning_rate": 0.00015430637721653903, + "loss": 0.0541, + "step": 13541 + }, + { + "epoch": 0.3174458060092477, + "grad_norm": 0.28789839148521423, + "learning_rate": 0.00015430019176942435, + "loss": 0.0498, + "step": 13542 + }, + { + "epoch": 0.3174692475840527, + "grad_norm": 0.2149842530488968, + "learning_rate": 0.00015429400602767532, + "loss": 0.0346, + "step": 13543 + }, + { + "epoch": 0.3174926891588577, + "grad_norm": 0.6071495413780212, + "learning_rate": 0.00015428781999132549, + "loss": 0.1149, + "step": 13544 + }, + { + "epoch": 0.3175161307336627, + "grad_norm": 0.2579508125782013, + "learning_rate": 0.0001542816336604084, + "loss": 0.0581, + "step": 13545 + }, + { + "epoch": 0.3175395723084677, + "grad_norm": 0.3179555833339691, + "learning_rate": 0.00015427544703495763, + "loss": 0.0643, + "step": 13546 + }, + { + "epoch": 0.3175630138832727, + "grad_norm": 0.1845601201057434, + "learning_rate": 0.00015426926011500675, + "loss": 0.018, + "step": 13547 + }, + { + "epoch": 0.3175864554580777, + "grad_norm": 0.4992149770259857, + "learning_rate": 0.00015426307290058934, + "loss": 0.5782, + "step": 13548 + }, + { + "epoch": 0.31760989703288267, + "grad_norm": 0.53528892993927, + "learning_rate": 0.000154256885391739, + "loss": 0.1169, + "step": 13549 + }, + { + "epoch": 0.31763333860768767, + "grad_norm": 0.4554833471775055, + "learning_rate": 0.0001542506975884893, + "loss": 0.1177, + "step": 13550 + }, + { + "epoch": 0.31765678018249266, + "grad_norm": 0.46483880281448364, + "learning_rate": 0.00015424450949087372, + "loss": 0.4893, + "step": 13551 + }, + { + "epoch": 0.31768022175729765, + "grad_norm": 0.1444980949163437, + "learning_rate": 0.00015423832109892594, + "loss": 0.0484, + "step": 13552 + }, + { + "epoch": 0.31770366333210265, + "grad_norm": 0.7000336647033691, + "learning_rate": 0.0001542321324126795, + "loss": 0.7327, + "step": 13553 + }, + { + "epoch": 0.31772710490690764, + "grad_norm": 0.3168487250804901, + "learning_rate": 0.000154225943432168, + "loss": 0.057, + "step": 13554 + }, + { + "epoch": 0.31775054648171264, + "grad_norm": 0.9296610355377197, + "learning_rate": 0.00015421975415742497, + "loss": 0.1414, + "step": 13555 + }, + { + "epoch": 0.31777398805651763, + "grad_norm": 0.6547183990478516, + "learning_rate": 0.00015421356458848406, + "loss": 0.1383, + "step": 13556 + }, + { + "epoch": 0.3177974296313226, + "grad_norm": 0.4168558716773987, + "learning_rate": 0.0001542073747253788, + "loss": 0.0716, + "step": 13557 + }, + { + "epoch": 0.3178208712061276, + "grad_norm": 0.4591078758239746, + "learning_rate": 0.0001542011845681428, + "loss": 0.101, + "step": 13558 + }, + { + "epoch": 0.3178443127809326, + "grad_norm": 0.3932187557220459, + "learning_rate": 0.0001541949941168097, + "loss": 0.0876, + "step": 13559 + }, + { + "epoch": 0.3178677543557376, + "grad_norm": 0.24962715804576874, + "learning_rate": 0.000154188803371413, + "loss": 0.0614, + "step": 13560 + }, + { + "epoch": 0.3178911959305426, + "grad_norm": 0.20776323974132538, + "learning_rate": 0.00015418261233198633, + "loss": 0.0283, + "step": 13561 + }, + { + "epoch": 0.3179146375053476, + "grad_norm": 0.3587252199649811, + "learning_rate": 0.00015417642099856328, + "loss": 0.0658, + "step": 13562 + }, + { + "epoch": 0.3179380790801526, + "grad_norm": 0.6687178015708923, + "learning_rate": 0.00015417022937117746, + "loss": 0.9611, + "step": 13563 + }, + { + "epoch": 0.3179615206549576, + "grad_norm": 1.1555025577545166, + "learning_rate": 0.00015416403744986242, + "loss": 0.1357, + "step": 13564 + }, + { + "epoch": 0.3179849622297626, + "grad_norm": 0.7482980489730835, + "learning_rate": 0.0001541578452346518, + "loss": 0.1497, + "step": 13565 + }, + { + "epoch": 0.3180084038045676, + "grad_norm": 0.4323042035102844, + "learning_rate": 0.00015415165272557923, + "loss": 0.1095, + "step": 13566 + }, + { + "epoch": 0.31803184537937257, + "grad_norm": 0.11011112481355667, + "learning_rate": 0.00015414545992267824, + "loss": 0.0324, + "step": 13567 + }, + { + "epoch": 0.31805528695417756, + "grad_norm": 0.8342926502227783, + "learning_rate": 0.00015413926682598245, + "loss": 0.0986, + "step": 13568 + }, + { + "epoch": 0.31807872852898256, + "grad_norm": 0.12074059993028641, + "learning_rate": 0.00015413307343552547, + "loss": 0.0145, + "step": 13569 + }, + { + "epoch": 0.31810217010378755, + "grad_norm": 0.6420939564704895, + "learning_rate": 0.00015412687975134096, + "loss": 0.1405, + "step": 13570 + }, + { + "epoch": 0.31812561167859255, + "grad_norm": 0.060864776372909546, + "learning_rate": 0.00015412068577346245, + "loss": 0.0187, + "step": 13571 + }, + { + "epoch": 0.31814905325339754, + "grad_norm": 0.569367527961731, + "learning_rate": 0.00015411449150192356, + "loss": 0.8632, + "step": 13572 + }, + { + "epoch": 0.31817249482820253, + "grad_norm": 0.340533584356308, + "learning_rate": 0.00015410829693675795, + "loss": 0.0619, + "step": 13573 + }, + { + "epoch": 0.31819593640300753, + "grad_norm": 0.2071538269519806, + "learning_rate": 0.00015410210207799917, + "loss": 0.0344, + "step": 13574 + }, + { + "epoch": 0.3182193779778125, + "grad_norm": 0.449432909488678, + "learning_rate": 0.00015409590692568088, + "loss": 0.0592, + "step": 13575 + }, + { + "epoch": 0.3182428195526175, + "grad_norm": 0.5664825439453125, + "learning_rate": 0.00015408971147983668, + "loss": 0.0843, + "step": 13576 + }, + { + "epoch": 0.31826626112742257, + "grad_norm": 0.2360914945602417, + "learning_rate": 0.00015408351574050017, + "loss": 0.0521, + "step": 13577 + }, + { + "epoch": 0.31828970270222756, + "grad_norm": 0.15496563911437988, + "learning_rate": 0.000154077319707705, + "loss": 0.0314, + "step": 13578 + }, + { + "epoch": 0.31831314427703256, + "grad_norm": 0.7976778745651245, + "learning_rate": 0.00015407112338148478, + "loss": 0.1501, + "step": 13579 + }, + { + "epoch": 0.31833658585183755, + "grad_norm": 0.48246026039123535, + "learning_rate": 0.0001540649267618731, + "loss": 0.0863, + "step": 13580 + }, + { + "epoch": 0.31836002742664254, + "grad_norm": 0.6502967476844788, + "learning_rate": 0.0001540587298489036, + "loss": 0.1185, + "step": 13581 + }, + { + "epoch": 0.31838346900144754, + "grad_norm": 0.7147530317306519, + "learning_rate": 0.00015405253264260993, + "loss": 0.1611, + "step": 13582 + }, + { + "epoch": 0.31840691057625253, + "grad_norm": 0.4216141700744629, + "learning_rate": 0.0001540463351430257, + "loss": 0.0803, + "step": 13583 + }, + { + "epoch": 0.3184303521510575, + "grad_norm": 0.7063252925872803, + "learning_rate": 0.00015404013735018454, + "loss": 0.7804, + "step": 13584 + }, + { + "epoch": 0.3184537937258625, + "grad_norm": 0.7387251257896423, + "learning_rate": 0.00015403393926412007, + "loss": 0.17, + "step": 13585 + }, + { + "epoch": 0.3184772353006675, + "grad_norm": 0.6286988258361816, + "learning_rate": 0.00015402774088486593, + "loss": 0.1238, + "step": 13586 + }, + { + "epoch": 0.3185006768754725, + "grad_norm": 0.39882880449295044, + "learning_rate": 0.00015402154221245571, + "loss": 0.0526, + "step": 13587 + }, + { + "epoch": 0.3185241184502775, + "grad_norm": 0.3033360540866852, + "learning_rate": 0.00015401534324692312, + "loss": 0.0631, + "step": 13588 + }, + { + "epoch": 0.3185475600250825, + "grad_norm": 0.5979171395301819, + "learning_rate": 0.0001540091439883018, + "loss": 0.0913, + "step": 13589 + }, + { + "epoch": 0.3185710015998875, + "grad_norm": 0.6779133081436157, + "learning_rate": 0.0001540029444366253, + "loss": 0.1976, + "step": 13590 + }, + { + "epoch": 0.3185944431746925, + "grad_norm": 0.5342113375663757, + "learning_rate": 0.00015399674459192728, + "loss": 0.5139, + "step": 13591 + }, + { + "epoch": 0.3186178847494975, + "grad_norm": 0.5256993174552917, + "learning_rate": 0.00015399054445424144, + "loss": 0.0866, + "step": 13592 + }, + { + "epoch": 0.3186413263243025, + "grad_norm": 0.45626839995384216, + "learning_rate": 0.00015398434402360135, + "loss": 0.0537, + "step": 13593 + }, + { + "epoch": 0.31866476789910747, + "grad_norm": 0.6828694343566895, + "learning_rate": 0.00015397814330004073, + "loss": 0.1054, + "step": 13594 + }, + { + "epoch": 0.31868820947391246, + "grad_norm": 0.34495243430137634, + "learning_rate": 0.00015397194228359317, + "loss": 0.0604, + "step": 13595 + }, + { + "epoch": 0.31871165104871746, + "grad_norm": 0.29929524660110474, + "learning_rate": 0.00015396574097429236, + "loss": 0.0404, + "step": 13596 + }, + { + "epoch": 0.31873509262352245, + "grad_norm": 0.5554851293563843, + "learning_rate": 0.00015395953937217188, + "loss": 0.1133, + "step": 13597 + }, + { + "epoch": 0.31875853419832745, + "grad_norm": 0.7900151610374451, + "learning_rate": 0.00015395333747726545, + "loss": 0.1739, + "step": 13598 + }, + { + "epoch": 0.31878197577313244, + "grad_norm": 0.3895375430583954, + "learning_rate": 0.0001539471352896067, + "loss": 0.078, + "step": 13599 + }, + { + "epoch": 0.31880541734793744, + "grad_norm": 0.6424814462661743, + "learning_rate": 0.00015394093280922925, + "loss": 0.188, + "step": 13600 + }, + { + "epoch": 0.31882885892274243, + "grad_norm": 0.5953755974769592, + "learning_rate": 0.0001539347300361668, + "loss": 0.1021, + "step": 13601 + }, + { + "epoch": 0.3188523004975474, + "grad_norm": 0.3058798015117645, + "learning_rate": 0.00015392852697045299, + "loss": 0.0707, + "step": 13602 + }, + { + "epoch": 0.3188757420723524, + "grad_norm": 1.7177919149398804, + "learning_rate": 0.00015392232361212147, + "loss": 0.4308, + "step": 13603 + }, + { + "epoch": 0.3188991836471574, + "grad_norm": 0.6945607662200928, + "learning_rate": 0.0001539161199612059, + "loss": 0.0885, + "step": 13604 + }, + { + "epoch": 0.3189226252219624, + "grad_norm": 0.510008692741394, + "learning_rate": 0.00015390991601773997, + "loss": 0.0621, + "step": 13605 + }, + { + "epoch": 0.3189460667967674, + "grad_norm": 0.33553674817085266, + "learning_rate": 0.0001539037117817573, + "loss": 0.0806, + "step": 13606 + }, + { + "epoch": 0.3189695083715724, + "grad_norm": 0.5444328784942627, + "learning_rate": 0.00015389750725329157, + "loss": 0.0913, + "step": 13607 + }, + { + "epoch": 0.3189929499463774, + "grad_norm": 0.3221682906150818, + "learning_rate": 0.00015389130243237647, + "loss": 0.0646, + "step": 13608 + }, + { + "epoch": 0.3190163915211824, + "grad_norm": 0.8434786200523376, + "learning_rate": 0.00015388509731904563, + "loss": 0.1146, + "step": 13609 + }, + { + "epoch": 0.3190398330959874, + "grad_norm": 0.5479951500892639, + "learning_rate": 0.00015387889191333275, + "loss": 0.1009, + "step": 13610 + }, + { + "epoch": 0.3190632746707924, + "grad_norm": 0.8242790102958679, + "learning_rate": 0.0001538726862152715, + "loss": 0.1616, + "step": 13611 + }, + { + "epoch": 0.31908671624559737, + "grad_norm": 0.37306100130081177, + "learning_rate": 0.0001538664802248955, + "loss": 0.0924, + "step": 13612 + }, + { + "epoch": 0.31911015782040236, + "grad_norm": 0.7727534174919128, + "learning_rate": 0.00015386027394223846, + "loss": 0.2608, + "step": 13613 + }, + { + "epoch": 0.31913359939520736, + "grad_norm": 0.6085309982299805, + "learning_rate": 0.0001538540673673341, + "loss": 0.1023, + "step": 13614 + }, + { + "epoch": 0.31915704097001235, + "grad_norm": 0.38524940609931946, + "learning_rate": 0.00015384786050021604, + "loss": 0.0345, + "step": 13615 + }, + { + "epoch": 0.31918048254481735, + "grad_norm": 0.8578136563301086, + "learning_rate": 0.00015384165334091798, + "loss": 0.153, + "step": 13616 + }, + { + "epoch": 0.31920392411962234, + "grad_norm": 0.2952553331851959, + "learning_rate": 0.0001538354458894736, + "loss": 0.0677, + "step": 13617 + }, + { + "epoch": 0.31922736569442733, + "grad_norm": 0.6879053711891174, + "learning_rate": 0.00015382923814591655, + "loss": 0.1188, + "step": 13618 + }, + { + "epoch": 0.31925080726923233, + "grad_norm": 0.3478337228298187, + "learning_rate": 0.00015382303011028054, + "loss": 0.0802, + "step": 13619 + }, + { + "epoch": 0.3192742488440373, + "grad_norm": 0.7247452139854431, + "learning_rate": 0.0001538168217825993, + "loss": 0.0621, + "step": 13620 + }, + { + "epoch": 0.3192976904188423, + "grad_norm": 0.3898468613624573, + "learning_rate": 0.00015381061316290645, + "loss": 0.0876, + "step": 13621 + }, + { + "epoch": 0.3193211319936473, + "grad_norm": 0.3209215998649597, + "learning_rate": 0.00015380440425123568, + "loss": 0.059, + "step": 13622 + }, + { + "epoch": 0.3193445735684523, + "grad_norm": 0.3148517906665802, + "learning_rate": 0.0001537981950476207, + "loss": 0.3325, + "step": 13623 + }, + { + "epoch": 0.3193680151432573, + "grad_norm": 0.6952115297317505, + "learning_rate": 0.00015379198555209525, + "loss": 0.1545, + "step": 13624 + }, + { + "epoch": 0.3193914567180623, + "grad_norm": 0.4203796684741974, + "learning_rate": 0.00015378577576469294, + "loss": 0.1123, + "step": 13625 + }, + { + "epoch": 0.3194148982928673, + "grad_norm": 0.5755981802940369, + "learning_rate": 0.00015377956568544748, + "loss": 0.1169, + "step": 13626 + }, + { + "epoch": 0.3194383398676723, + "grad_norm": 0.3087066411972046, + "learning_rate": 0.0001537733553143926, + "loss": 0.0755, + "step": 13627 + }, + { + "epoch": 0.31946178144247733, + "grad_norm": 0.6728754639625549, + "learning_rate": 0.00015376714465156202, + "loss": 0.1164, + "step": 13628 + }, + { + "epoch": 0.3194852230172823, + "grad_norm": 0.4042263627052307, + "learning_rate": 0.00015376093369698932, + "loss": 0.0674, + "step": 13629 + }, + { + "epoch": 0.3195086645920873, + "grad_norm": 0.13739940524101257, + "learning_rate": 0.0001537547224507083, + "loss": 0.0437, + "step": 13630 + }, + { + "epoch": 0.3195321061668923, + "grad_norm": 0.6107300519943237, + "learning_rate": 0.00015374851091275272, + "loss": 0.0948, + "step": 13631 + }, + { + "epoch": 0.3195555477416973, + "grad_norm": 0.3522806167602539, + "learning_rate": 0.00015374229908315614, + "loss": 0.0488, + "step": 13632 + }, + { + "epoch": 0.3195789893165023, + "grad_norm": 0.267511248588562, + "learning_rate": 0.00015373608696195236, + "loss": 0.0478, + "step": 13633 + }, + { + "epoch": 0.3196024308913073, + "grad_norm": 0.541746199131012, + "learning_rate": 0.00015372987454917507, + "loss": 0.0883, + "step": 13634 + }, + { + "epoch": 0.3196258724661123, + "grad_norm": 0.8366674780845642, + "learning_rate": 0.00015372366184485795, + "loss": 0.1504, + "step": 13635 + }, + { + "epoch": 0.3196493140409173, + "grad_norm": 0.3367016017436981, + "learning_rate": 0.00015371744884903473, + "loss": 0.088, + "step": 13636 + }, + { + "epoch": 0.3196727556157223, + "grad_norm": 0.3114806115627289, + "learning_rate": 0.00015371123556173912, + "loss": 0.0775, + "step": 13637 + }, + { + "epoch": 0.3196961971905273, + "grad_norm": 0.6130706071853638, + "learning_rate": 0.00015370502198300482, + "loss": 0.0855, + "step": 13638 + }, + { + "epoch": 0.31971963876533227, + "grad_norm": 0.4853875935077667, + "learning_rate": 0.00015369880811286562, + "loss": 0.1186, + "step": 13639 + }, + { + "epoch": 0.31974308034013726, + "grad_norm": 0.26463785767555237, + "learning_rate": 0.00015369259395135515, + "loss": 0.0369, + "step": 13640 + }, + { + "epoch": 0.31976652191494226, + "grad_norm": 0.5745686888694763, + "learning_rate": 0.00015368637949850713, + "loss": 0.1762, + "step": 13641 + }, + { + "epoch": 0.31978996348974725, + "grad_norm": 0.562163770198822, + "learning_rate": 0.00015368016475435532, + "loss": 0.1434, + "step": 13642 + }, + { + "epoch": 0.31981340506455225, + "grad_norm": 0.4627060294151306, + "learning_rate": 0.00015367394971893344, + "loss": 0.0843, + "step": 13643 + }, + { + "epoch": 0.31983684663935724, + "grad_norm": 0.5183565616607666, + "learning_rate": 0.00015366773439227517, + "loss": 0.0979, + "step": 13644 + }, + { + "epoch": 0.31986028821416224, + "grad_norm": 0.5137754082679749, + "learning_rate": 0.00015366151877441427, + "loss": 0.1236, + "step": 13645 + }, + { + "epoch": 0.31988372978896723, + "grad_norm": 0.5552082061767578, + "learning_rate": 0.0001536553028653845, + "loss": 0.0974, + "step": 13646 + }, + { + "epoch": 0.3199071713637722, + "grad_norm": 0.38750872015953064, + "learning_rate": 0.00015364908666521952, + "loss": 0.0643, + "step": 13647 + }, + { + "epoch": 0.3199306129385772, + "grad_norm": 0.6554803848266602, + "learning_rate": 0.00015364287017395308, + "loss": 0.1128, + "step": 13648 + }, + { + "epoch": 0.3199540545133822, + "grad_norm": 0.10396208614110947, + "learning_rate": 0.00015363665339161893, + "loss": 0.0242, + "step": 13649 + }, + { + "epoch": 0.3199774960881872, + "grad_norm": 0.1396925300359726, + "learning_rate": 0.0001536304363182508, + "loss": 0.0222, + "step": 13650 + }, + { + "epoch": 0.3200009376629922, + "grad_norm": 0.5405993461608887, + "learning_rate": 0.0001536242189538824, + "loss": 0.0775, + "step": 13651 + }, + { + "epoch": 0.3200243792377972, + "grad_norm": 0.5906836986541748, + "learning_rate": 0.00015361800129854748, + "loss": 0.8929, + "step": 13652 + }, + { + "epoch": 0.3200478208126022, + "grad_norm": 0.18031272292137146, + "learning_rate": 0.0001536117833522798, + "loss": 0.0324, + "step": 13653 + }, + { + "epoch": 0.3200712623874072, + "grad_norm": 0.5175034999847412, + "learning_rate": 0.000153605565115113, + "loss": 0.088, + "step": 13654 + }, + { + "epoch": 0.3200947039622122, + "grad_norm": 0.4357362985610962, + "learning_rate": 0.00015359934658708098, + "loss": 0.1201, + "step": 13655 + }, + { + "epoch": 0.3201181455370172, + "grad_norm": 0.6070231795310974, + "learning_rate": 0.00015359312776821738, + "loss": 0.0816, + "step": 13656 + }, + { + "epoch": 0.32014158711182217, + "grad_norm": 1.2421046495437622, + "learning_rate": 0.0001535869086585559, + "loss": 0.2048, + "step": 13657 + }, + { + "epoch": 0.32016502868662716, + "grad_norm": 0.8637301325798035, + "learning_rate": 0.00015358068925813045, + "loss": 0.1356, + "step": 13658 + }, + { + "epoch": 0.32018847026143216, + "grad_norm": 0.319308876991272, + "learning_rate": 0.00015357446956697462, + "loss": 0.0567, + "step": 13659 + }, + { + "epoch": 0.32021191183623715, + "grad_norm": 0.1443524807691574, + "learning_rate": 0.00015356824958512219, + "loss": 0.0359, + "step": 13660 + }, + { + "epoch": 0.32023535341104215, + "grad_norm": 0.6777157187461853, + "learning_rate": 0.00015356202931260694, + "loss": 0.5279, + "step": 13661 + }, + { + "epoch": 0.32025879498584714, + "grad_norm": 0.5914331674575806, + "learning_rate": 0.00015355580874946268, + "loss": 0.1194, + "step": 13662 + }, + { + "epoch": 0.32028223656065213, + "grad_norm": 0.6424688696861267, + "learning_rate": 0.00015354958789572302, + "loss": 0.5437, + "step": 13663 + }, + { + "epoch": 0.32030567813545713, + "grad_norm": 0.4046505391597748, + "learning_rate": 0.0001535433667514218, + "loss": 0.0513, + "step": 13664 + }, + { + "epoch": 0.3203291197102621, + "grad_norm": 0.25686946511268616, + "learning_rate": 0.0001535371453165928, + "loss": 0.0693, + "step": 13665 + }, + { + "epoch": 0.3203525612850671, + "grad_norm": 0.4290587306022644, + "learning_rate": 0.00015353092359126974, + "loss": 0.0979, + "step": 13666 + }, + { + "epoch": 0.3203760028598721, + "grad_norm": 0.4876640737056732, + "learning_rate": 0.00015352470157548633, + "loss": 0.1272, + "step": 13667 + }, + { + "epoch": 0.3203994444346771, + "grad_norm": 0.807052493095398, + "learning_rate": 0.00015351847926927645, + "loss": 0.129, + "step": 13668 + }, + { + "epoch": 0.3204228860094821, + "grad_norm": 0.33147770166397095, + "learning_rate": 0.00015351225667267377, + "loss": 0.0797, + "step": 13669 + }, + { + "epoch": 0.3204463275842871, + "grad_norm": 0.5410645604133606, + "learning_rate": 0.00015350603378571208, + "loss": 0.0553, + "step": 13670 + }, + { + "epoch": 0.3204697691590921, + "grad_norm": 0.28140878677368164, + "learning_rate": 0.00015349981060842517, + "loss": 0.0492, + "step": 13671 + }, + { + "epoch": 0.3204932107338971, + "grad_norm": 0.07381321489810944, + "learning_rate": 0.00015349358714084673, + "loss": 0.0081, + "step": 13672 + }, + { + "epoch": 0.3205166523087021, + "grad_norm": 0.5750092267990112, + "learning_rate": 0.0001534873633830106, + "loss": 0.0991, + "step": 13673 + }, + { + "epoch": 0.32054009388350707, + "grad_norm": 0.6024866700172424, + "learning_rate": 0.00015348113933495055, + "loss": 0.1222, + "step": 13674 + }, + { + "epoch": 0.32056353545831207, + "grad_norm": 0.6842116713523865, + "learning_rate": 0.00015347491499670036, + "loss": 0.1655, + "step": 13675 + }, + { + "epoch": 0.32058697703311706, + "grad_norm": 0.8677282929420471, + "learning_rate": 0.00015346869036829373, + "loss": 0.1833, + "step": 13676 + }, + { + "epoch": 0.32061041860792205, + "grad_norm": 0.5869843363761902, + "learning_rate": 0.0001534624654497645, + "loss": 0.1785, + "step": 13677 + }, + { + "epoch": 0.32063386018272705, + "grad_norm": 0.1272657811641693, + "learning_rate": 0.00015345624024114645, + "loss": 0.022, + "step": 13678 + }, + { + "epoch": 0.3206573017575321, + "grad_norm": 0.12081409245729446, + "learning_rate": 0.0001534500147424733, + "loss": 0.0195, + "step": 13679 + }, + { + "epoch": 0.3206807433323371, + "grad_norm": 0.8912476897239685, + "learning_rate": 0.00015344378895377884, + "loss": 0.1835, + "step": 13680 + }, + { + "epoch": 0.3207041849071421, + "grad_norm": 0.673549234867096, + "learning_rate": 0.00015343756287509693, + "loss": 0.2373, + "step": 13681 + }, + { + "epoch": 0.3207276264819471, + "grad_norm": 0.5358778238296509, + "learning_rate": 0.00015343133650646133, + "loss": 0.1091, + "step": 13682 + }, + { + "epoch": 0.3207510680567521, + "grad_norm": 0.7756301760673523, + "learning_rate": 0.00015342510984790573, + "loss": 0.1415, + "step": 13683 + }, + { + "epoch": 0.32077450963155707, + "grad_norm": 0.39149728417396545, + "learning_rate": 0.00015341888289946402, + "loss": 0.0894, + "step": 13684 + }, + { + "epoch": 0.32079795120636206, + "grad_norm": 1.1174488067626953, + "learning_rate": 0.00015341265566116993, + "loss": 0.0213, + "step": 13685 + }, + { + "epoch": 0.32082139278116706, + "grad_norm": 0.7465423941612244, + "learning_rate": 0.00015340642813305727, + "loss": 0.7285, + "step": 13686 + }, + { + "epoch": 0.32084483435597205, + "grad_norm": 0.12682008743286133, + "learning_rate": 0.00015340020031515985, + "loss": 0.0186, + "step": 13687 + }, + { + "epoch": 0.32086827593077705, + "grad_norm": 0.29132404923439026, + "learning_rate": 0.00015339397220751145, + "loss": 0.0428, + "step": 13688 + }, + { + "epoch": 0.32089171750558204, + "grad_norm": 0.5699378848075867, + "learning_rate": 0.00015338774381014583, + "loss": 0.0901, + "step": 13689 + }, + { + "epoch": 0.32091515908038704, + "grad_norm": 0.6305656433105469, + "learning_rate": 0.0001533815151230968, + "loss": 0.1272, + "step": 13690 + }, + { + "epoch": 0.32093860065519203, + "grad_norm": 0.694115161895752, + "learning_rate": 0.0001533752861463982, + "loss": 0.8147, + "step": 13691 + }, + { + "epoch": 0.320962042229997, + "grad_norm": 0.3943326771259308, + "learning_rate": 0.00015336905688008377, + "loss": 0.0527, + "step": 13692 + }, + { + "epoch": 0.320985483804802, + "grad_norm": 0.5336909890174866, + "learning_rate": 0.00015336282732418733, + "loss": 0.114, + "step": 13693 + }, + { + "epoch": 0.321008925379607, + "grad_norm": 0.42815035581588745, + "learning_rate": 0.00015335659747874272, + "loss": 0.1095, + "step": 13694 + }, + { + "epoch": 0.321032366954412, + "grad_norm": 0.12626220285892487, + "learning_rate": 0.00015335036734378366, + "loss": 0.0291, + "step": 13695 + }, + { + "epoch": 0.321055808529217, + "grad_norm": 0.32080885767936707, + "learning_rate": 0.00015334413691934408, + "loss": 0.0886, + "step": 13696 + }, + { + "epoch": 0.321079250104022, + "grad_norm": 0.18715092539787292, + "learning_rate": 0.00015333790620545768, + "loss": 0.0527, + "step": 13697 + }, + { + "epoch": 0.321102691678827, + "grad_norm": 0.5506422519683838, + "learning_rate": 0.0001533316752021583, + "loss": 0.0797, + "step": 13698 + }, + { + "epoch": 0.321126133253632, + "grad_norm": 0.5408821702003479, + "learning_rate": 0.00015332544390947973, + "loss": 0.1738, + "step": 13699 + }, + { + "epoch": 0.321149574828437, + "grad_norm": 0.6586996912956238, + "learning_rate": 0.00015331921232745584, + "loss": 0.1883, + "step": 13700 + }, + { + "epoch": 0.321173016403242, + "grad_norm": 0.5965974926948547, + "learning_rate": 0.00015331298045612036, + "loss": 0.1618, + "step": 13701 + }, + { + "epoch": 0.32119645797804697, + "grad_norm": 0.1361141800880432, + "learning_rate": 0.0001533067482955072, + "loss": 0.0168, + "step": 13702 + }, + { + "epoch": 0.32121989955285196, + "grad_norm": 0.13203582167625427, + "learning_rate": 0.0001533005158456501, + "loss": 0.025, + "step": 13703 + }, + { + "epoch": 0.32124334112765696, + "grad_norm": 0.4700043797492981, + "learning_rate": 0.00015329428310658288, + "loss": 0.0869, + "step": 13704 + }, + { + "epoch": 0.32126678270246195, + "grad_norm": 0.3924386501312256, + "learning_rate": 0.00015328805007833938, + "loss": 0.0798, + "step": 13705 + }, + { + "epoch": 0.32129022427726694, + "grad_norm": 0.6144112348556519, + "learning_rate": 0.00015328181676095348, + "loss": 0.0973, + "step": 13706 + }, + { + "epoch": 0.32131366585207194, + "grad_norm": 0.3378109335899353, + "learning_rate": 0.0001532755831544589, + "loss": 0.0738, + "step": 13707 + }, + { + "epoch": 0.32133710742687693, + "grad_norm": 0.5880793333053589, + "learning_rate": 0.0001532693492588895, + "loss": 0.6172, + "step": 13708 + }, + { + "epoch": 0.3213605490016819, + "grad_norm": 0.47455817461013794, + "learning_rate": 0.00015326311507427915, + "loss": 0.0524, + "step": 13709 + }, + { + "epoch": 0.3213839905764869, + "grad_norm": 0.16246217489242554, + "learning_rate": 0.00015325688060066162, + "loss": 0.0457, + "step": 13710 + }, + { + "epoch": 0.3214074321512919, + "grad_norm": 1.0476391315460205, + "learning_rate": 0.00015325064583807075, + "loss": 0.1382, + "step": 13711 + }, + { + "epoch": 0.3214308737260969, + "grad_norm": 0.27143850922584534, + "learning_rate": 0.0001532444107865404, + "loss": 0.0907, + "step": 13712 + }, + { + "epoch": 0.3214543153009019, + "grad_norm": 0.43458202481269836, + "learning_rate": 0.00015323817544610438, + "loss": 0.078, + "step": 13713 + }, + { + "epoch": 0.3214777568757069, + "grad_norm": 0.40163955092430115, + "learning_rate": 0.0001532319398167965, + "loss": 0.0765, + "step": 13714 + }, + { + "epoch": 0.3215011984505119, + "grad_norm": 0.37633031606674194, + "learning_rate": 0.0001532257038986506, + "loss": 0.0483, + "step": 13715 + }, + { + "epoch": 0.3215246400253169, + "grad_norm": 0.3226276934146881, + "learning_rate": 0.00015321946769170056, + "loss": 0.0575, + "step": 13716 + }, + { + "epoch": 0.3215480816001219, + "grad_norm": 0.5881257057189941, + "learning_rate": 0.00015321323119598025, + "loss": 0.1333, + "step": 13717 + }, + { + "epoch": 0.3215715231749269, + "grad_norm": 0.7818975448608398, + "learning_rate": 0.00015320699441152336, + "loss": 0.1369, + "step": 13718 + }, + { + "epoch": 0.32159496474973187, + "grad_norm": 0.4576185643672943, + "learning_rate": 0.00015320075733836385, + "loss": 0.0768, + "step": 13719 + }, + { + "epoch": 0.32161840632453687, + "grad_norm": 0.605598509311676, + "learning_rate": 0.00015319451997653557, + "loss": 0.11, + "step": 13720 + }, + { + "epoch": 0.32164184789934186, + "grad_norm": 0.4425242841243744, + "learning_rate": 0.00015318828232607228, + "loss": 0.0635, + "step": 13721 + }, + { + "epoch": 0.32166528947414685, + "grad_norm": 0.6478126049041748, + "learning_rate": 0.0001531820443870079, + "loss": 0.1354, + "step": 13722 + }, + { + "epoch": 0.32168873104895185, + "grad_norm": 0.3267136216163635, + "learning_rate": 0.00015317580615937626, + "loss": 0.0556, + "step": 13723 + }, + { + "epoch": 0.32171217262375684, + "grad_norm": 0.86734539270401, + "learning_rate": 0.00015316956764321117, + "loss": 0.7414, + "step": 13724 + }, + { + "epoch": 0.32173561419856184, + "grad_norm": 0.6841928958892822, + "learning_rate": 0.00015316332883854653, + "loss": 0.2019, + "step": 13725 + }, + { + "epoch": 0.32175905577336683, + "grad_norm": 0.2626076340675354, + "learning_rate": 0.0001531570897454162, + "loss": 0.0426, + "step": 13726 + }, + { + "epoch": 0.3217824973481718, + "grad_norm": 0.8637951612472534, + "learning_rate": 0.00015315085036385394, + "loss": 0.0935, + "step": 13727 + }, + { + "epoch": 0.3218059389229768, + "grad_norm": 0.4578690528869629, + "learning_rate": 0.00015314461069389373, + "loss": 0.0676, + "step": 13728 + }, + { + "epoch": 0.3218293804977818, + "grad_norm": 0.2204992175102234, + "learning_rate": 0.00015313837073556934, + "loss": 0.0326, + "step": 13729 + }, + { + "epoch": 0.32185282207258686, + "grad_norm": 0.553305983543396, + "learning_rate": 0.00015313213048891467, + "loss": 0.1283, + "step": 13730 + }, + { + "epoch": 0.32187626364739186, + "grad_norm": 0.559839129447937, + "learning_rate": 0.00015312588995396355, + "loss": 0.2169, + "step": 13731 + }, + { + "epoch": 0.32189970522219685, + "grad_norm": 0.5994709134101868, + "learning_rate": 0.00015311964913074987, + "loss": 0.2192, + "step": 13732 + }, + { + "epoch": 0.32192314679700185, + "grad_norm": 0.21534807980060577, + "learning_rate": 0.00015311340801930746, + "loss": 0.0313, + "step": 13733 + }, + { + "epoch": 0.32194658837180684, + "grad_norm": 1.5735175609588623, + "learning_rate": 0.00015310716661967022, + "loss": 0.0838, + "step": 13734 + }, + { + "epoch": 0.32197002994661184, + "grad_norm": 0.7320497035980225, + "learning_rate": 0.000153100924931872, + "loss": 0.0986, + "step": 13735 + }, + { + "epoch": 0.32199347152141683, + "grad_norm": 0.4101369082927704, + "learning_rate": 0.0001530946829559467, + "loss": 0.0599, + "step": 13736 + }, + { + "epoch": 0.3220169130962218, + "grad_norm": 0.208892822265625, + "learning_rate": 0.00015308844069192814, + "loss": 0.0381, + "step": 13737 + }, + { + "epoch": 0.3220403546710268, + "grad_norm": 0.2830946743488312, + "learning_rate": 0.0001530821981398502, + "loss": 0.0637, + "step": 13738 + }, + { + "epoch": 0.3220637962458318, + "grad_norm": 0.5880311727523804, + "learning_rate": 0.00015307595529974677, + "loss": 0.1362, + "step": 13739 + }, + { + "epoch": 0.3220872378206368, + "grad_norm": 0.49357035756111145, + "learning_rate": 0.00015306971217165166, + "loss": 0.0885, + "step": 13740 + }, + { + "epoch": 0.3221106793954418, + "grad_norm": 0.3806760013103485, + "learning_rate": 0.00015306346875559884, + "loss": 0.0536, + "step": 13741 + }, + { + "epoch": 0.3221341209702468, + "grad_norm": 0.4574997127056122, + "learning_rate": 0.00015305722505162218, + "loss": 0.101, + "step": 13742 + }, + { + "epoch": 0.3221575625450518, + "grad_norm": 0.6108567118644714, + "learning_rate": 0.00015305098105975547, + "loss": 0.1053, + "step": 13743 + }, + { + "epoch": 0.3221810041198568, + "grad_norm": 0.4028269946575165, + "learning_rate": 0.00015304473678003268, + "loss": 0.0888, + "step": 13744 + }, + { + "epoch": 0.3222044456946618, + "grad_norm": 0.25779595971107483, + "learning_rate": 0.00015303849221248763, + "loss": 0.0556, + "step": 13745 + }, + { + "epoch": 0.3222278872694668, + "grad_norm": 0.1971444934606552, + "learning_rate": 0.00015303224735715426, + "loss": 0.0446, + "step": 13746 + }, + { + "epoch": 0.32225132884427177, + "grad_norm": 0.18177306652069092, + "learning_rate": 0.00015302600221406637, + "loss": 0.0548, + "step": 13747 + }, + { + "epoch": 0.32227477041907676, + "grad_norm": 0.19158458709716797, + "learning_rate": 0.00015301975678325795, + "loss": 0.0422, + "step": 13748 + }, + { + "epoch": 0.32229821199388176, + "grad_norm": 0.505990743637085, + "learning_rate": 0.00015301351106476285, + "loss": 0.1445, + "step": 13749 + }, + { + "epoch": 0.32232165356868675, + "grad_norm": 0.1627722829580307, + "learning_rate": 0.0001530072650586149, + "loss": 0.0308, + "step": 13750 + }, + { + "epoch": 0.32234509514349174, + "grad_norm": 0.23602977395057678, + "learning_rate": 0.0001530010187648481, + "loss": 0.0602, + "step": 13751 + }, + { + "epoch": 0.32236853671829674, + "grad_norm": 0.27186119556427, + "learning_rate": 0.00015299477218349626, + "loss": 0.0772, + "step": 13752 + }, + { + "epoch": 0.32239197829310173, + "grad_norm": 0.5115479826927185, + "learning_rate": 0.00015298852531459324, + "loss": 0.0805, + "step": 13753 + }, + { + "epoch": 0.3224154198679067, + "grad_norm": 0.4917759597301483, + "learning_rate": 0.00015298227815817306, + "loss": 0.1212, + "step": 13754 + }, + { + "epoch": 0.3224388614427117, + "grad_norm": 0.47815218567848206, + "learning_rate": 0.00015297603071426954, + "loss": 0.122, + "step": 13755 + }, + { + "epoch": 0.3224623030175167, + "grad_norm": 0.81573486328125, + "learning_rate": 0.00015296978298291655, + "loss": 0.1965, + "step": 13756 + }, + { + "epoch": 0.3224857445923217, + "grad_norm": 0.5290364027023315, + "learning_rate": 0.00015296353496414806, + "loss": 0.1138, + "step": 13757 + }, + { + "epoch": 0.3225091861671267, + "grad_norm": 0.37467342615127563, + "learning_rate": 0.00015295728665799794, + "loss": 0.0711, + "step": 13758 + }, + { + "epoch": 0.3225326277419317, + "grad_norm": 0.36220696568489075, + "learning_rate": 0.00015295103806450006, + "loss": 0.0412, + "step": 13759 + }, + { + "epoch": 0.3225560693167367, + "grad_norm": 0.22214293479919434, + "learning_rate": 0.0001529447891836884, + "loss": 0.0187, + "step": 13760 + }, + { + "epoch": 0.3225795108915417, + "grad_norm": 0.2148287147283554, + "learning_rate": 0.0001529385400155968, + "loss": 0.0318, + "step": 13761 + }, + { + "epoch": 0.3226029524663467, + "grad_norm": 0.1879064440727234, + "learning_rate": 0.00015293229056025922, + "loss": 0.0449, + "step": 13762 + }, + { + "epoch": 0.3226263940411517, + "grad_norm": 0.4193907380104065, + "learning_rate": 0.00015292604081770952, + "loss": 0.1033, + "step": 13763 + }, + { + "epoch": 0.32264983561595667, + "grad_norm": 0.5478887557983398, + "learning_rate": 0.00015291979078798162, + "loss": 0.0781, + "step": 13764 + }, + { + "epoch": 0.32267327719076166, + "grad_norm": 0.6860342025756836, + "learning_rate": 0.00015291354047110948, + "loss": 0.0734, + "step": 13765 + }, + { + "epoch": 0.32269671876556666, + "grad_norm": 0.3641621470451355, + "learning_rate": 0.00015290728986712697, + "loss": 0.085, + "step": 13766 + }, + { + "epoch": 0.32272016034037165, + "grad_norm": 0.6038793325424194, + "learning_rate": 0.00015290103897606802, + "loss": 0.1482, + "step": 13767 + }, + { + "epoch": 0.32274360191517665, + "grad_norm": 0.4920172095298767, + "learning_rate": 0.0001528947877979665, + "loss": 0.0902, + "step": 13768 + }, + { + "epoch": 0.32276704348998164, + "grad_norm": 0.6967554092407227, + "learning_rate": 0.00015288853633285644, + "loss": 0.1352, + "step": 13769 + }, + { + "epoch": 0.32279048506478664, + "grad_norm": 0.7989881038665771, + "learning_rate": 0.00015288228458077167, + "loss": 0.1919, + "step": 13770 + }, + { + "epoch": 0.32281392663959163, + "grad_norm": 0.19973331689834595, + "learning_rate": 0.00015287603254174613, + "loss": 0.0492, + "step": 13771 + }, + { + "epoch": 0.3228373682143966, + "grad_norm": 0.5939799547195435, + "learning_rate": 0.00015286978021581375, + "loss": 0.1421, + "step": 13772 + }, + { + "epoch": 0.3228608097892016, + "grad_norm": 0.24939416348934174, + "learning_rate": 0.00015286352760300847, + "loss": 0.031, + "step": 13773 + }, + { + "epoch": 0.3228842513640066, + "grad_norm": 0.41972556710243225, + "learning_rate": 0.00015285727470336423, + "loss": 0.06, + "step": 13774 + }, + { + "epoch": 0.3229076929388116, + "grad_norm": 0.22783739864826202, + "learning_rate": 0.0001528510215169149, + "loss": 0.0276, + "step": 13775 + }, + { + "epoch": 0.3229311345136166, + "grad_norm": 0.20486560463905334, + "learning_rate": 0.00015284476804369441, + "loss": 0.0411, + "step": 13776 + }, + { + "epoch": 0.3229545760884216, + "grad_norm": 0.8163740038871765, + "learning_rate": 0.00015283851428373675, + "loss": 0.2274, + "step": 13777 + }, + { + "epoch": 0.3229780176632266, + "grad_norm": 0.6088862419128418, + "learning_rate": 0.00015283226023707583, + "loss": 0.6877, + "step": 13778 + }, + { + "epoch": 0.3230014592380316, + "grad_norm": 0.22283953428268433, + "learning_rate": 0.0001528260059037456, + "loss": 0.0706, + "step": 13779 + }, + { + "epoch": 0.3230249008128366, + "grad_norm": 0.457570344209671, + "learning_rate": 0.00015281975128377995, + "loss": 0.134, + "step": 13780 + }, + { + "epoch": 0.32304834238764163, + "grad_norm": 0.4598759412765503, + "learning_rate": 0.00015281349637721284, + "loss": 0.0872, + "step": 13781 + }, + { + "epoch": 0.3230717839624466, + "grad_norm": 0.3966560959815979, + "learning_rate": 0.00015280724118407822, + "loss": 0.1098, + "step": 13782 + }, + { + "epoch": 0.3230952255372516, + "grad_norm": 0.3260590434074402, + "learning_rate": 0.00015280098570441002, + "loss": 0.0364, + "step": 13783 + }, + { + "epoch": 0.3231186671120566, + "grad_norm": 0.7543467283248901, + "learning_rate": 0.00015279472993824222, + "loss": 0.1725, + "step": 13784 + }, + { + "epoch": 0.3231421086868616, + "grad_norm": 0.5587853193283081, + "learning_rate": 0.00015278847388560868, + "loss": 0.1619, + "step": 13785 + }, + { + "epoch": 0.3231655502616666, + "grad_norm": 0.23068781197071075, + "learning_rate": 0.00015278221754654343, + "loss": 0.0743, + "step": 13786 + }, + { + "epoch": 0.3231889918364716, + "grad_norm": 0.16718412935733795, + "learning_rate": 0.00015277596092108038, + "loss": 0.0247, + "step": 13787 + }, + { + "epoch": 0.3232124334112766, + "grad_norm": 0.7454635500907898, + "learning_rate": 0.00015276970400925346, + "loss": 0.725, + "step": 13788 + }, + { + "epoch": 0.3232358749860816, + "grad_norm": 0.584363579750061, + "learning_rate": 0.00015276344681109665, + "loss": 0.5774, + "step": 13789 + }, + { + "epoch": 0.3232593165608866, + "grad_norm": 0.3830904960632324, + "learning_rate": 0.0001527571893266439, + "loss": 0.0705, + "step": 13790 + }, + { + "epoch": 0.3232827581356916, + "grad_norm": 0.5584062933921814, + "learning_rate": 0.00015275093155592916, + "loss": 0.0852, + "step": 13791 + }, + { + "epoch": 0.32330619971049657, + "grad_norm": 0.5044223666191101, + "learning_rate": 0.0001527446734989864, + "loss": 0.0828, + "step": 13792 + }, + { + "epoch": 0.32332964128530156, + "grad_norm": 0.5950095057487488, + "learning_rate": 0.00015273841515584956, + "loss": 0.205, + "step": 13793 + }, + { + "epoch": 0.32335308286010656, + "grad_norm": 0.27348360419273376, + "learning_rate": 0.00015273215652655253, + "loss": 0.0552, + "step": 13794 + }, + { + "epoch": 0.32337652443491155, + "grad_norm": 0.09702886641025543, + "learning_rate": 0.0001527258976111294, + "loss": 0.0158, + "step": 13795 + }, + { + "epoch": 0.32339996600971654, + "grad_norm": 0.31004253029823303, + "learning_rate": 0.00015271963840961407, + "loss": 0.0373, + "step": 13796 + }, + { + "epoch": 0.32342340758452154, + "grad_norm": 0.7844216823577881, + "learning_rate": 0.0001527133789220405, + "loss": 0.1929, + "step": 13797 + }, + { + "epoch": 0.32344684915932653, + "grad_norm": 0.5753737092018127, + "learning_rate": 0.0001527071191484426, + "loss": 0.6054, + "step": 13798 + }, + { + "epoch": 0.3234702907341315, + "grad_norm": 0.32779932022094727, + "learning_rate": 0.00015270085908885444, + "loss": 0.0485, + "step": 13799 + }, + { + "epoch": 0.3234937323089365, + "grad_norm": 0.8955798745155334, + "learning_rate": 0.0001526945987433099, + "loss": 0.1429, + "step": 13800 + }, + { + "epoch": 0.3235171738837415, + "grad_norm": 0.4795452058315277, + "learning_rate": 0.000152688338111843, + "loss": 0.0849, + "step": 13801 + }, + { + "epoch": 0.3235406154585465, + "grad_norm": 0.24454964697360992, + "learning_rate": 0.00015268207719448774, + "loss": 0.0482, + "step": 13802 + }, + { + "epoch": 0.3235640570333515, + "grad_norm": 0.26198625564575195, + "learning_rate": 0.000152675815991278, + "loss": 0.0467, + "step": 13803 + }, + { + "epoch": 0.3235874986081565, + "grad_norm": 1.0700784921646118, + "learning_rate": 0.0001526695545022478, + "loss": 0.2677, + "step": 13804 + }, + { + "epoch": 0.3236109401829615, + "grad_norm": 0.7732657194137573, + "learning_rate": 0.00015266329272743115, + "loss": 0.1462, + "step": 13805 + }, + { + "epoch": 0.3236343817577665, + "grad_norm": 0.6403403282165527, + "learning_rate": 0.00015265703066686195, + "loss": 0.1137, + "step": 13806 + }, + { + "epoch": 0.3236578233325715, + "grad_norm": 0.34942924976348877, + "learning_rate": 0.00015265076832057423, + "loss": 0.0668, + "step": 13807 + }, + { + "epoch": 0.3236812649073765, + "grad_norm": 0.11277329921722412, + "learning_rate": 0.00015264450568860198, + "loss": 0.0278, + "step": 13808 + }, + { + "epoch": 0.32370470648218147, + "grad_norm": 0.5009021759033203, + "learning_rate": 0.00015263824277097915, + "loss": 0.1057, + "step": 13809 + }, + { + "epoch": 0.32372814805698646, + "grad_norm": 0.46761274337768555, + "learning_rate": 0.00015263197956773974, + "loss": 0.116, + "step": 13810 + }, + { + "epoch": 0.32375158963179146, + "grad_norm": 0.5717591643333435, + "learning_rate": 0.00015262571607891774, + "loss": 0.1456, + "step": 13811 + }, + { + "epoch": 0.32377503120659645, + "grad_norm": 0.4446314871311188, + "learning_rate": 0.00015261945230454708, + "loss": 0.1194, + "step": 13812 + }, + { + "epoch": 0.32379847278140145, + "grad_norm": 0.5800995230674744, + "learning_rate": 0.00015261318824466184, + "loss": 0.1099, + "step": 13813 + }, + { + "epoch": 0.32382191435620644, + "grad_norm": 0.45918264985084534, + "learning_rate": 0.00015260692389929594, + "loss": 0.0798, + "step": 13814 + }, + { + "epoch": 0.32384535593101144, + "grad_norm": 0.6750203371047974, + "learning_rate": 0.00015260065926848342, + "loss": 0.092, + "step": 13815 + }, + { + "epoch": 0.32386879750581643, + "grad_norm": 0.4889063835144043, + "learning_rate": 0.00015259439435225821, + "loss": 0.1098, + "step": 13816 + }, + { + "epoch": 0.3238922390806214, + "grad_norm": 0.9743533134460449, + "learning_rate": 0.00015258812915065433, + "loss": 0.1818, + "step": 13817 + }, + { + "epoch": 0.3239156806554264, + "grad_norm": 0.38455551862716675, + "learning_rate": 0.00015258186366370584, + "loss": 0.1445, + "step": 13818 + }, + { + "epoch": 0.3239391222302314, + "grad_norm": 0.15449480712413788, + "learning_rate": 0.00015257559789144662, + "loss": 0.0189, + "step": 13819 + }, + { + "epoch": 0.3239625638050364, + "grad_norm": 0.5923640727996826, + "learning_rate": 0.00015256933183391077, + "loss": 0.1258, + "step": 13820 + }, + { + "epoch": 0.3239860053798414, + "grad_norm": 0.5155543684959412, + "learning_rate": 0.00015256306549113222, + "loss": 0.089, + "step": 13821 + }, + { + "epoch": 0.3240094469546464, + "grad_norm": 0.274826318025589, + "learning_rate": 0.000152556798863145, + "loss": 0.0692, + "step": 13822 + }, + { + "epoch": 0.3240328885294514, + "grad_norm": 0.6284314393997192, + "learning_rate": 0.0001525505319499831, + "loss": 0.1316, + "step": 13823 + }, + { + "epoch": 0.3240563301042564, + "grad_norm": 0.5988786220550537, + "learning_rate": 0.00015254426475168055, + "loss": 0.1313, + "step": 13824 + }, + { + "epoch": 0.3240797716790614, + "grad_norm": 0.17121383547782898, + "learning_rate": 0.00015253799726827136, + "loss": 0.0398, + "step": 13825 + }, + { + "epoch": 0.3241032132538664, + "grad_norm": 0.3249247372150421, + "learning_rate": 0.00015253172949978948, + "loss": 0.0645, + "step": 13826 + }, + { + "epoch": 0.32412665482867137, + "grad_norm": 0.5711403489112854, + "learning_rate": 0.000152525461446269, + "loss": 0.0818, + "step": 13827 + }, + { + "epoch": 0.32415009640347636, + "grad_norm": 0.2153405398130417, + "learning_rate": 0.00015251919310774387, + "loss": 0.0489, + "step": 13828 + }, + { + "epoch": 0.32417353797828136, + "grad_norm": 0.4379107654094696, + "learning_rate": 0.00015251292448424813, + "loss": 0.112, + "step": 13829 + }, + { + "epoch": 0.32419697955308635, + "grad_norm": 0.38265088200569153, + "learning_rate": 0.00015250665557581578, + "loss": 0.0474, + "step": 13830 + }, + { + "epoch": 0.32422042112789135, + "grad_norm": 0.1715056449174881, + "learning_rate": 0.00015250038638248083, + "loss": 0.0361, + "step": 13831 + }, + { + "epoch": 0.3242438627026964, + "grad_norm": 0.17996329069137573, + "learning_rate": 0.0001524941169042773, + "loss": 0.0346, + "step": 13832 + }, + { + "epoch": 0.3242673042775014, + "grad_norm": 0.48317664861679077, + "learning_rate": 0.00015248784714123924, + "loss": 0.119, + "step": 13833 + }, + { + "epoch": 0.3242907458523064, + "grad_norm": 0.1581982672214508, + "learning_rate": 0.00015248157709340065, + "loss": 0.0346, + "step": 13834 + }, + { + "epoch": 0.3243141874271114, + "grad_norm": 0.5898411870002747, + "learning_rate": 0.00015247530676079554, + "loss": 0.1422, + "step": 13835 + }, + { + "epoch": 0.32433762900191637, + "grad_norm": 0.30456069111824036, + "learning_rate": 0.00015246903614345793, + "loss": 0.0828, + "step": 13836 + }, + { + "epoch": 0.32436107057672137, + "grad_norm": 0.793882429599762, + "learning_rate": 0.00015246276524142184, + "loss": 0.1951, + "step": 13837 + }, + { + "epoch": 0.32438451215152636, + "grad_norm": 0.36811816692352295, + "learning_rate": 0.00015245649405472132, + "loss": 0.0708, + "step": 13838 + }, + { + "epoch": 0.32440795372633136, + "grad_norm": 0.2841481864452362, + "learning_rate": 0.0001524502225833904, + "loss": 0.0658, + "step": 13839 + }, + { + "epoch": 0.32443139530113635, + "grad_norm": 0.14184290170669556, + "learning_rate": 0.00015244395082746311, + "loss": 0.0307, + "step": 13840 + }, + { + "epoch": 0.32445483687594134, + "grad_norm": 0.5820318460464478, + "learning_rate": 0.00015243767878697347, + "loss": 0.0556, + "step": 13841 + }, + { + "epoch": 0.32447827845074634, + "grad_norm": 0.7004052996635437, + "learning_rate": 0.00015243140646195547, + "loss": 0.1022, + "step": 13842 + }, + { + "epoch": 0.32450172002555133, + "grad_norm": 0.6105347275733948, + "learning_rate": 0.0001524251338524432, + "loss": 0.1123, + "step": 13843 + }, + { + "epoch": 0.3245251616003563, + "grad_norm": 0.4075625538825989, + "learning_rate": 0.0001524188609584707, + "loss": 0.1084, + "step": 13844 + }, + { + "epoch": 0.3245486031751613, + "grad_norm": 0.5692236423492432, + "learning_rate": 0.000152412587780072, + "loss": 0.1406, + "step": 13845 + }, + { + "epoch": 0.3245720447499663, + "grad_norm": 0.36832761764526367, + "learning_rate": 0.00015240631431728108, + "loss": 0.0775, + "step": 13846 + }, + { + "epoch": 0.3245954863247713, + "grad_norm": 0.47477802634239197, + "learning_rate": 0.00015240004057013207, + "loss": 0.1381, + "step": 13847 + }, + { + "epoch": 0.3246189278995763, + "grad_norm": 0.1682032197713852, + "learning_rate": 0.0001523937665386589, + "loss": 0.0263, + "step": 13848 + }, + { + "epoch": 0.3246423694743813, + "grad_norm": 0.5702695846557617, + "learning_rate": 0.00015238749222289575, + "loss": 0.094, + "step": 13849 + }, + { + "epoch": 0.3246658110491863, + "grad_norm": 0.36575230956077576, + "learning_rate": 0.00015238121762287655, + "loss": 0.1188, + "step": 13850 + }, + { + "epoch": 0.3246892526239913, + "grad_norm": 0.2912333011627197, + "learning_rate": 0.0001523749427386354, + "loss": 0.0542, + "step": 13851 + }, + { + "epoch": 0.3247126941987963, + "grad_norm": 0.648343563079834, + "learning_rate": 0.00015236866757020634, + "loss": 0.0771, + "step": 13852 + }, + { + "epoch": 0.3247361357736013, + "grad_norm": 0.4049101769924164, + "learning_rate": 0.00015236239211762344, + "loss": 0.0998, + "step": 13853 + }, + { + "epoch": 0.32475957734840627, + "grad_norm": 0.45967739820480347, + "learning_rate": 0.00015235611638092072, + "loss": 0.0672, + "step": 13854 + }, + { + "epoch": 0.32478301892321126, + "grad_norm": 0.2876662611961365, + "learning_rate": 0.0001523498403601322, + "loss": 0.0318, + "step": 13855 + }, + { + "epoch": 0.32480646049801626, + "grad_norm": 0.4712375998497009, + "learning_rate": 0.00015234356405529202, + "loss": 0.0976, + "step": 13856 + }, + { + "epoch": 0.32482990207282125, + "grad_norm": 0.33649906516075134, + "learning_rate": 0.00015233728746643418, + "loss": 0.094, + "step": 13857 + }, + { + "epoch": 0.32485334364762625, + "grad_norm": 0.5236013531684875, + "learning_rate": 0.00015233101059359272, + "loss": 0.1163, + "step": 13858 + }, + { + "epoch": 0.32487678522243124, + "grad_norm": 0.4921607971191406, + "learning_rate": 0.00015232473343680176, + "loss": 0.1269, + "step": 13859 + }, + { + "epoch": 0.32490022679723624, + "grad_norm": 0.29269465804100037, + "learning_rate": 0.00015231845599609532, + "loss": 0.064, + "step": 13860 + }, + { + "epoch": 0.32492366837204123, + "grad_norm": 0.2993770241737366, + "learning_rate": 0.00015231217827150743, + "loss": 0.0864, + "step": 13861 + }, + { + "epoch": 0.3249471099468462, + "grad_norm": 0.4233585298061371, + "learning_rate": 0.00015230590026307222, + "loss": 0.0568, + "step": 13862 + }, + { + "epoch": 0.3249705515216512, + "grad_norm": 0.7293097972869873, + "learning_rate": 0.00015229962197082372, + "loss": 0.1717, + "step": 13863 + }, + { + "epoch": 0.3249939930964562, + "grad_norm": 0.2850983440876007, + "learning_rate": 0.00015229334339479598, + "loss": 0.0516, + "step": 13864 + }, + { + "epoch": 0.3250174346712612, + "grad_norm": 0.5235785245895386, + "learning_rate": 0.00015228706453502313, + "loss": 0.109, + "step": 13865 + }, + { + "epoch": 0.3250408762460662, + "grad_norm": 0.6199187636375427, + "learning_rate": 0.0001522807853915392, + "loss": 0.1286, + "step": 13866 + }, + { + "epoch": 0.3250643178208712, + "grad_norm": 1.1237616539001465, + "learning_rate": 0.0001522745059643782, + "loss": 0.3094, + "step": 13867 + }, + { + "epoch": 0.3250877593956762, + "grad_norm": 1.1319236755371094, + "learning_rate": 0.0001522682262535743, + "loss": 0.1609, + "step": 13868 + }, + { + "epoch": 0.3251112009704812, + "grad_norm": 0.2941162884235382, + "learning_rate": 0.00015226194625916155, + "loss": 0.0782, + "step": 13869 + }, + { + "epoch": 0.3251346425452862, + "grad_norm": 0.10230308771133423, + "learning_rate": 0.00015225566598117397, + "loss": 0.0261, + "step": 13870 + }, + { + "epoch": 0.3251580841200912, + "grad_norm": 0.5566446185112, + "learning_rate": 0.00015224938541964571, + "loss": 0.0789, + "step": 13871 + }, + { + "epoch": 0.32518152569489617, + "grad_norm": 0.3358168303966522, + "learning_rate": 0.0001522431045746108, + "loss": 0.0508, + "step": 13872 + }, + { + "epoch": 0.32520496726970116, + "grad_norm": 0.3159034848213196, + "learning_rate": 0.00015223682344610336, + "loss": 0.1096, + "step": 13873 + }, + { + "epoch": 0.32522840884450616, + "grad_norm": 0.47943904995918274, + "learning_rate": 0.0001522305420341574, + "loss": 0.0997, + "step": 13874 + }, + { + "epoch": 0.32525185041931115, + "grad_norm": 0.8513184189796448, + "learning_rate": 0.0001522242603388071, + "loss": 0.1249, + "step": 13875 + }, + { + "epoch": 0.32527529199411614, + "grad_norm": 0.46543392539024353, + "learning_rate": 0.00015221797836008648, + "loss": 0.0565, + "step": 13876 + }, + { + "epoch": 0.32529873356892114, + "grad_norm": 0.8383069634437561, + "learning_rate": 0.00015221169609802964, + "loss": 0.1313, + "step": 13877 + }, + { + "epoch": 0.32532217514372613, + "grad_norm": 0.7235292792320251, + "learning_rate": 0.00015220541355267066, + "loss": 0.1434, + "step": 13878 + }, + { + "epoch": 0.3253456167185311, + "grad_norm": 0.440456360578537, + "learning_rate": 0.00015219913072404362, + "loss": 0.1246, + "step": 13879 + }, + { + "epoch": 0.3253690582933361, + "grad_norm": 0.20264755189418793, + "learning_rate": 0.00015219284761218264, + "loss": 0.0438, + "step": 13880 + }, + { + "epoch": 0.3253924998681411, + "grad_norm": 0.07426350563764572, + "learning_rate": 0.00015218656421712182, + "loss": 0.0162, + "step": 13881 + }, + { + "epoch": 0.3254159414429461, + "grad_norm": 0.6388759613037109, + "learning_rate": 0.00015218028053889522, + "loss": 0.2131, + "step": 13882 + }, + { + "epoch": 0.3254393830177511, + "grad_norm": 0.517620861530304, + "learning_rate": 0.00015217399657753696, + "loss": 0.0923, + "step": 13883 + }, + { + "epoch": 0.32546282459255615, + "grad_norm": 0.568168580532074, + "learning_rate": 0.00015216771233308115, + "loss": 0.1224, + "step": 13884 + }, + { + "epoch": 0.32548626616736115, + "grad_norm": 0.6455985307693481, + "learning_rate": 0.00015216142780556183, + "loss": 0.1676, + "step": 13885 + }, + { + "epoch": 0.32550970774216614, + "grad_norm": 0.4727780222892761, + "learning_rate": 0.0001521551429950131, + "loss": 0.0966, + "step": 13886 + }, + { + "epoch": 0.32553314931697114, + "grad_norm": 0.5481367707252502, + "learning_rate": 0.00015214885790146916, + "loss": 0.2421, + "step": 13887 + }, + { + "epoch": 0.32555659089177613, + "grad_norm": 0.1651555895805359, + "learning_rate": 0.00015214257252496406, + "loss": 0.0285, + "step": 13888 + }, + { + "epoch": 0.3255800324665811, + "grad_norm": 0.17914098501205444, + "learning_rate": 0.00015213628686553188, + "loss": 0.0578, + "step": 13889 + }, + { + "epoch": 0.3256034740413861, + "grad_norm": 0.2737352252006531, + "learning_rate": 0.00015213000092320674, + "loss": 0.0402, + "step": 13890 + }, + { + "epoch": 0.3256269156161911, + "grad_norm": 0.3496340215206146, + "learning_rate": 0.0001521237146980227, + "loss": 0.0615, + "step": 13891 + }, + { + "epoch": 0.3256503571909961, + "grad_norm": 0.43224960565567017, + "learning_rate": 0.00015211742819001397, + "loss": 0.0995, + "step": 13892 + }, + { + "epoch": 0.3256737987658011, + "grad_norm": 0.26451176404953003, + "learning_rate": 0.0001521111413992146, + "loss": 0.0657, + "step": 13893 + }, + { + "epoch": 0.3256972403406061, + "grad_norm": 0.5201084613800049, + "learning_rate": 0.00015210485432565874, + "loss": 0.1296, + "step": 13894 + }, + { + "epoch": 0.3257206819154111, + "grad_norm": 0.1740075796842575, + "learning_rate": 0.00015209856696938044, + "loss": 0.0418, + "step": 13895 + }, + { + "epoch": 0.3257441234902161, + "grad_norm": 0.24449366331100464, + "learning_rate": 0.00015209227933041385, + "loss": 0.0615, + "step": 13896 + }, + { + "epoch": 0.3257675650650211, + "grad_norm": 0.41854751110076904, + "learning_rate": 0.0001520859914087931, + "loss": 0.1367, + "step": 13897 + }, + { + "epoch": 0.3257910066398261, + "grad_norm": 0.5065606832504272, + "learning_rate": 0.00015207970320455226, + "loss": 0.1022, + "step": 13898 + }, + { + "epoch": 0.32581444821463107, + "grad_norm": 0.5704189538955688, + "learning_rate": 0.0001520734147177255, + "loss": 0.0963, + "step": 13899 + }, + { + "epoch": 0.32583788978943606, + "grad_norm": 0.4750918745994568, + "learning_rate": 0.00015206712594834697, + "loss": 0.0697, + "step": 13900 + }, + { + "epoch": 0.32586133136424106, + "grad_norm": 0.4871262311935425, + "learning_rate": 0.00015206083689645072, + "loss": 0.1164, + "step": 13901 + }, + { + "epoch": 0.32588477293904605, + "grad_norm": 0.24083475768566132, + "learning_rate": 0.00015205454756207093, + "loss": 0.0345, + "step": 13902 + }, + { + "epoch": 0.32590821451385105, + "grad_norm": 0.4586699306964874, + "learning_rate": 0.00015204825794524167, + "loss": 0.1034, + "step": 13903 + }, + { + "epoch": 0.32593165608865604, + "grad_norm": 0.9767122864723206, + "learning_rate": 0.0001520419680459971, + "loss": 0.1342, + "step": 13904 + }, + { + "epoch": 0.32595509766346104, + "grad_norm": 0.6997212171554565, + "learning_rate": 0.00015203567786437137, + "loss": 0.7156, + "step": 13905 + }, + { + "epoch": 0.32597853923826603, + "grad_norm": 0.48251602053642273, + "learning_rate": 0.00015202938740039858, + "loss": 0.0759, + "step": 13906 + }, + { + "epoch": 0.326001980813071, + "grad_norm": 0.5313964486122131, + "learning_rate": 0.00015202309665411287, + "loss": 0.5969, + "step": 13907 + }, + { + "epoch": 0.326025422387876, + "grad_norm": 0.34946420788764954, + "learning_rate": 0.00015201680562554833, + "loss": 0.0382, + "step": 13908 + }, + { + "epoch": 0.326048863962681, + "grad_norm": 0.25283512473106384, + "learning_rate": 0.00015201051431473924, + "loss": 0.0522, + "step": 13909 + }, + { + "epoch": 0.326072305537486, + "grad_norm": 0.5904913544654846, + "learning_rate": 0.00015200422272171956, + "loss": 0.0932, + "step": 13910 + }, + { + "epoch": 0.326095747112291, + "grad_norm": 0.5150827169418335, + "learning_rate": 0.00015199793084652352, + "loss": 0.1346, + "step": 13911 + }, + { + "epoch": 0.326119188687096, + "grad_norm": 0.18793600797653198, + "learning_rate": 0.00015199163868918525, + "loss": 0.0322, + "step": 13912 + }, + { + "epoch": 0.326142630261901, + "grad_norm": 0.6348475217819214, + "learning_rate": 0.00015198534624973888, + "loss": 0.0985, + "step": 13913 + }, + { + "epoch": 0.326166071836706, + "grad_norm": 0.42641761898994446, + "learning_rate": 0.00015197905352821856, + "loss": 0.0889, + "step": 13914 + }, + { + "epoch": 0.326189513411511, + "grad_norm": 0.33164554834365845, + "learning_rate": 0.00015197276052465844, + "loss": 0.0511, + "step": 13915 + }, + { + "epoch": 0.326212954986316, + "grad_norm": 0.7012106776237488, + "learning_rate": 0.00015196646723909267, + "loss": 0.9954, + "step": 13916 + }, + { + "epoch": 0.32623639656112097, + "grad_norm": 0.38898080587387085, + "learning_rate": 0.00015196017367155538, + "loss": 0.0893, + "step": 13917 + }, + { + "epoch": 0.32625983813592596, + "grad_norm": 0.1906972974538803, + "learning_rate": 0.00015195387982208072, + "loss": 0.0386, + "step": 13918 + }, + { + "epoch": 0.32628327971073096, + "grad_norm": 0.15756674110889435, + "learning_rate": 0.00015194758569070285, + "loss": 0.028, + "step": 13919 + }, + { + "epoch": 0.32630672128553595, + "grad_norm": 0.5970812439918518, + "learning_rate": 0.00015194129127745593, + "loss": 0.1123, + "step": 13920 + }, + { + "epoch": 0.32633016286034094, + "grad_norm": 0.4769560396671295, + "learning_rate": 0.00015193499658237408, + "loss": 0.0652, + "step": 13921 + }, + { + "epoch": 0.32635360443514594, + "grad_norm": 0.6183115839958191, + "learning_rate": 0.00015192870160549152, + "loss": 0.1527, + "step": 13922 + }, + { + "epoch": 0.32637704600995093, + "grad_norm": 0.897916316986084, + "learning_rate": 0.00015192240634684232, + "loss": 0.1355, + "step": 13923 + }, + { + "epoch": 0.3264004875847559, + "grad_norm": 0.1707799881696701, + "learning_rate": 0.0001519161108064607, + "loss": 0.0337, + "step": 13924 + }, + { + "epoch": 0.3264239291595609, + "grad_norm": 0.17538626492023468, + "learning_rate": 0.00015190981498438083, + "loss": 0.0285, + "step": 13925 + }, + { + "epoch": 0.3264473707343659, + "grad_norm": 0.6376181244850159, + "learning_rate": 0.0001519035188806368, + "loss": 0.6016, + "step": 13926 + }, + { + "epoch": 0.3264708123091709, + "grad_norm": 0.7751206755638123, + "learning_rate": 0.00015189722249526284, + "loss": 0.2141, + "step": 13927 + }, + { + "epoch": 0.3264942538839759, + "grad_norm": 0.2177172154188156, + "learning_rate": 0.00015189092582829306, + "loss": 0.0405, + "step": 13928 + }, + { + "epoch": 0.3265176954587809, + "grad_norm": 0.6725398898124695, + "learning_rate": 0.00015188462887976169, + "loss": 0.1497, + "step": 13929 + }, + { + "epoch": 0.3265411370335859, + "grad_norm": 0.4778507649898529, + "learning_rate": 0.00015187833164970286, + "loss": 0.0869, + "step": 13930 + }, + { + "epoch": 0.3265645786083909, + "grad_norm": 0.4727945029735565, + "learning_rate": 0.00015187203413815075, + "loss": 0.1632, + "step": 13931 + }, + { + "epoch": 0.3265880201831959, + "grad_norm": 0.17374923825263977, + "learning_rate": 0.0001518657363451395, + "loss": 0.0159, + "step": 13932 + }, + { + "epoch": 0.3266114617580009, + "grad_norm": 0.7262431979179382, + "learning_rate": 0.00015185943827070335, + "loss": 0.102, + "step": 13933 + }, + { + "epoch": 0.32663490333280587, + "grad_norm": 0.42591574788093567, + "learning_rate": 0.00015185313991487637, + "loss": 0.0982, + "step": 13934 + }, + { + "epoch": 0.3266583449076109, + "grad_norm": 0.47767239809036255, + "learning_rate": 0.00015184684127769285, + "loss": 0.1007, + "step": 13935 + }, + { + "epoch": 0.3266817864824159, + "grad_norm": 0.4899515211582184, + "learning_rate": 0.00015184054235918687, + "loss": 0.0811, + "step": 13936 + }, + { + "epoch": 0.3267052280572209, + "grad_norm": 0.6247353553771973, + "learning_rate": 0.00015183424315939268, + "loss": 0.7058, + "step": 13937 + }, + { + "epoch": 0.3267286696320259, + "grad_norm": 0.22755078971385956, + "learning_rate": 0.00015182794367834442, + "loss": 0.023, + "step": 13938 + }, + { + "epoch": 0.3267521112068309, + "grad_norm": 0.1750766783952713, + "learning_rate": 0.00015182164391607628, + "loss": 0.0361, + "step": 13939 + }, + { + "epoch": 0.3267755527816359, + "grad_norm": 0.43087634444236755, + "learning_rate": 0.00015181534387262243, + "loss": 0.1102, + "step": 13940 + }, + { + "epoch": 0.3267989943564409, + "grad_norm": 0.26229360699653625, + "learning_rate": 0.0001518090435480171, + "loss": 0.0588, + "step": 13941 + }, + { + "epoch": 0.3268224359312459, + "grad_norm": 0.47346198558807373, + "learning_rate": 0.00015180274294229442, + "loss": 0.0885, + "step": 13942 + }, + { + "epoch": 0.3268458775060509, + "grad_norm": 1.0949211120605469, + "learning_rate": 0.00015179644205548856, + "loss": 0.1248, + "step": 13943 + }, + { + "epoch": 0.32686931908085587, + "grad_norm": 0.8069592118263245, + "learning_rate": 0.00015179014088763384, + "loss": 0.1512, + "step": 13944 + }, + { + "epoch": 0.32689276065566086, + "grad_norm": 0.7317711710929871, + "learning_rate": 0.0001517838394387643, + "loss": 0.1456, + "step": 13945 + }, + { + "epoch": 0.32691620223046586, + "grad_norm": 0.5193055272102356, + "learning_rate": 0.0001517775377089142, + "loss": 0.1159, + "step": 13946 + }, + { + "epoch": 0.32693964380527085, + "grad_norm": 0.44428080320358276, + "learning_rate": 0.00015177123569811771, + "loss": 0.0675, + "step": 13947 + }, + { + "epoch": 0.32696308538007585, + "grad_norm": 0.5790316462516785, + "learning_rate": 0.00015176493340640908, + "loss": 0.1022, + "step": 13948 + }, + { + "epoch": 0.32698652695488084, + "grad_norm": 0.7844952344894409, + "learning_rate": 0.00015175863083382245, + "loss": 0.1376, + "step": 13949 + }, + { + "epoch": 0.32700996852968583, + "grad_norm": 1.014565348625183, + "learning_rate": 0.00015175232798039198, + "loss": 0.1697, + "step": 13950 + }, + { + "epoch": 0.32703341010449083, + "grad_norm": 0.440457284450531, + "learning_rate": 0.00015174602484615198, + "loss": 0.531, + "step": 13951 + }, + { + "epoch": 0.3270568516792958, + "grad_norm": 0.5941112637519836, + "learning_rate": 0.00015173972143113658, + "loss": 0.1613, + "step": 13952 + }, + { + "epoch": 0.3270802932541008, + "grad_norm": 1.5212476253509521, + "learning_rate": 0.00015173341773538, + "loss": 0.2499, + "step": 13953 + }, + { + "epoch": 0.3271037348289058, + "grad_norm": 0.5948570966720581, + "learning_rate": 0.00015172711375891645, + "loss": 0.1489, + "step": 13954 + }, + { + "epoch": 0.3271271764037108, + "grad_norm": 0.4957319498062134, + "learning_rate": 0.00015172080950178013, + "loss": 0.1148, + "step": 13955 + }, + { + "epoch": 0.3271506179785158, + "grad_norm": 0.6809244155883789, + "learning_rate": 0.0001517145049640052, + "loss": 0.1212, + "step": 13956 + }, + { + "epoch": 0.3271740595533208, + "grad_norm": 0.3972876965999603, + "learning_rate": 0.00015170820014562595, + "loss": 0.051, + "step": 13957 + }, + { + "epoch": 0.3271975011281258, + "grad_norm": 0.4934951961040497, + "learning_rate": 0.00015170189504667655, + "loss": 0.0622, + "step": 13958 + }, + { + "epoch": 0.3272209427029308, + "grad_norm": 0.6543289422988892, + "learning_rate": 0.00015169558966719122, + "loss": 0.1944, + "step": 13959 + }, + { + "epoch": 0.3272443842777358, + "grad_norm": 0.4649565517902374, + "learning_rate": 0.0001516892840072041, + "loss": 0.0867, + "step": 13960 + }, + { + "epoch": 0.3272678258525408, + "grad_norm": 0.20536139607429504, + "learning_rate": 0.00015168297806674957, + "loss": 0.0404, + "step": 13961 + }, + { + "epoch": 0.32729126742734577, + "grad_norm": 0.550249457359314, + "learning_rate": 0.0001516766718458617, + "loss": 0.1472, + "step": 13962 + }, + { + "epoch": 0.32731470900215076, + "grad_norm": 0.4014947712421417, + "learning_rate": 0.00015167036534457476, + "loss": 0.1078, + "step": 13963 + }, + { + "epoch": 0.32733815057695576, + "grad_norm": 0.5974758863449097, + "learning_rate": 0.00015166405856292293, + "loss": 0.0904, + "step": 13964 + }, + { + "epoch": 0.32736159215176075, + "grad_norm": 0.13160322606563568, + "learning_rate": 0.00015165775150094053, + "loss": 0.022, + "step": 13965 + }, + { + "epoch": 0.32738503372656574, + "grad_norm": 0.49878621101379395, + "learning_rate": 0.0001516514441586617, + "loss": 0.1282, + "step": 13966 + }, + { + "epoch": 0.32740847530137074, + "grad_norm": 0.412924587726593, + "learning_rate": 0.00015164513653612068, + "loss": 0.066, + "step": 13967 + }, + { + "epoch": 0.32743191687617573, + "grad_norm": 0.6282274127006531, + "learning_rate": 0.00015163882863335166, + "loss": 0.1498, + "step": 13968 + }, + { + "epoch": 0.3274553584509807, + "grad_norm": 0.42523109912872314, + "learning_rate": 0.00015163252045038894, + "loss": 0.124, + "step": 13969 + }, + { + "epoch": 0.3274788000257857, + "grad_norm": 0.577539324760437, + "learning_rate": 0.00015162621198726672, + "loss": 0.1709, + "step": 13970 + }, + { + "epoch": 0.3275022416005907, + "grad_norm": 0.5597982406616211, + "learning_rate": 0.00015161990324401918, + "loss": 0.1593, + "step": 13971 + }, + { + "epoch": 0.3275256831753957, + "grad_norm": 0.15403051674365997, + "learning_rate": 0.00015161359422068062, + "loss": 0.0281, + "step": 13972 + }, + { + "epoch": 0.3275491247502007, + "grad_norm": 0.5362935066223145, + "learning_rate": 0.00015160728491728524, + "loss": 0.0803, + "step": 13973 + }, + { + "epoch": 0.3275725663250057, + "grad_norm": 0.6491706967353821, + "learning_rate": 0.00015160097533386727, + "loss": 0.0727, + "step": 13974 + }, + { + "epoch": 0.3275960078998107, + "grad_norm": 0.07791583985090256, + "learning_rate": 0.00015159466547046098, + "loss": 0.0248, + "step": 13975 + }, + { + "epoch": 0.3276194494746157, + "grad_norm": 0.5226130485534668, + "learning_rate": 0.00015158835532710061, + "loss": 0.0877, + "step": 13976 + }, + { + "epoch": 0.3276428910494207, + "grad_norm": 0.2339596301317215, + "learning_rate": 0.0001515820449038203, + "loss": 0.0295, + "step": 13977 + }, + { + "epoch": 0.3276663326242257, + "grad_norm": 0.1278640776872635, + "learning_rate": 0.00015157573420065444, + "loss": 0.0203, + "step": 13978 + }, + { + "epoch": 0.32768977419903067, + "grad_norm": 0.8412493467330933, + "learning_rate": 0.00015156942321763715, + "loss": 0.1118, + "step": 13979 + }, + { + "epoch": 0.32771321577383566, + "grad_norm": 0.141577810049057, + "learning_rate": 0.00015156311195480273, + "loss": 0.0242, + "step": 13980 + }, + { + "epoch": 0.32773665734864066, + "grad_norm": 0.6456174254417419, + "learning_rate": 0.0001515568004121854, + "loss": 0.6202, + "step": 13981 + }, + { + "epoch": 0.32776009892344565, + "grad_norm": 1.4733999967575073, + "learning_rate": 0.00015155048858981945, + "loss": 0.1078, + "step": 13982 + }, + { + "epoch": 0.32778354049825065, + "grad_norm": 0.646073579788208, + "learning_rate": 0.00015154417648773907, + "loss": 0.1224, + "step": 13983 + }, + { + "epoch": 0.32780698207305564, + "grad_norm": 0.3692658245563507, + "learning_rate": 0.00015153786410597857, + "loss": 0.0644, + "step": 13984 + }, + { + "epoch": 0.32783042364786064, + "grad_norm": 0.7248607277870178, + "learning_rate": 0.00015153155144457213, + "loss": 0.1679, + "step": 13985 + }, + { + "epoch": 0.3278538652226657, + "grad_norm": 0.1377331018447876, + "learning_rate": 0.00015152523850355407, + "loss": 0.021, + "step": 13986 + }, + { + "epoch": 0.3278773067974707, + "grad_norm": 0.19562210142612457, + "learning_rate": 0.00015151892528295863, + "loss": 0.0333, + "step": 13987 + }, + { + "epoch": 0.3279007483722757, + "grad_norm": 0.3954848349094391, + "learning_rate": 0.00015151261178282004, + "loss": 0.1, + "step": 13988 + }, + { + "epoch": 0.32792418994708067, + "grad_norm": 0.2460498809814453, + "learning_rate": 0.00015150629800317255, + "loss": 0.0407, + "step": 13989 + }, + { + "epoch": 0.32794763152188566, + "grad_norm": 0.1150926798582077, + "learning_rate": 0.00015149998394405048, + "loss": 0.0291, + "step": 13990 + }, + { + "epoch": 0.32797107309669066, + "grad_norm": 0.5202431082725525, + "learning_rate": 0.00015149366960548802, + "loss": 0.1107, + "step": 13991 + }, + { + "epoch": 0.32799451467149565, + "grad_norm": 1.089245319366455, + "learning_rate": 0.00015148735498751945, + "loss": 0.1379, + "step": 13992 + }, + { + "epoch": 0.32801795624630065, + "grad_norm": 0.6598648428916931, + "learning_rate": 0.00015148104009017908, + "loss": 0.1878, + "step": 13993 + }, + { + "epoch": 0.32804139782110564, + "grad_norm": 0.21554839611053467, + "learning_rate": 0.00015147472491350113, + "loss": 0.0555, + "step": 13994 + }, + { + "epoch": 0.32806483939591063, + "grad_norm": 0.8562152981758118, + "learning_rate": 0.00015146840945751984, + "loss": 0.2152, + "step": 13995 + }, + { + "epoch": 0.32808828097071563, + "grad_norm": 0.36874622106552124, + "learning_rate": 0.00015146209372226955, + "loss": 0.043, + "step": 13996 + }, + { + "epoch": 0.3281117225455206, + "grad_norm": 0.38003087043762207, + "learning_rate": 0.00015145577770778448, + "loss": 0.0857, + "step": 13997 + }, + { + "epoch": 0.3281351641203256, + "grad_norm": 0.53378826379776, + "learning_rate": 0.0001514494614140989, + "loss": 0.7788, + "step": 13998 + }, + { + "epoch": 0.3281586056951306, + "grad_norm": 0.9769792556762695, + "learning_rate": 0.00015144314484124714, + "loss": 0.1858, + "step": 13999 + }, + { + "epoch": 0.3281820472699356, + "grad_norm": 0.4610154330730438, + "learning_rate": 0.00015143682798926338, + "loss": 0.0888, + "step": 14000 + }, + { + "epoch": 0.3282054888447406, + "grad_norm": 0.29465505480766296, + "learning_rate": 0.00015143051085818193, + "loss": 0.4155, + "step": 14001 + }, + { + "epoch": 0.3282289304195456, + "grad_norm": 0.5175170302391052, + "learning_rate": 0.00015142419344803712, + "loss": 0.0947, + "step": 14002 + }, + { + "epoch": 0.3282523719943506, + "grad_norm": 0.9412294030189514, + "learning_rate": 0.0001514178757588632, + "loss": 0.2971, + "step": 14003 + }, + { + "epoch": 0.3282758135691556, + "grad_norm": 0.4616345763206482, + "learning_rate": 0.00015141155779069442, + "loss": 0.0952, + "step": 14004 + }, + { + "epoch": 0.3282992551439606, + "grad_norm": 0.510924220085144, + "learning_rate": 0.0001514052395435651, + "loss": 0.1172, + "step": 14005 + }, + { + "epoch": 0.32832269671876557, + "grad_norm": 0.5209413170814514, + "learning_rate": 0.0001513989210175095, + "loss": 0.4777, + "step": 14006 + }, + { + "epoch": 0.32834613829357057, + "grad_norm": 0.4882621765136719, + "learning_rate": 0.00015139260221256188, + "loss": 0.1129, + "step": 14007 + }, + { + "epoch": 0.32836957986837556, + "grad_norm": 0.4479592740535736, + "learning_rate": 0.0001513862831287566, + "loss": 0.1228, + "step": 14008 + }, + { + "epoch": 0.32839302144318055, + "grad_norm": 0.30609527230262756, + "learning_rate": 0.00015137996376612787, + "loss": 0.0524, + "step": 14009 + }, + { + "epoch": 0.32841646301798555, + "grad_norm": 0.6215572953224182, + "learning_rate": 0.00015137364412471003, + "loss": 0.04, + "step": 14010 + }, + { + "epoch": 0.32843990459279054, + "grad_norm": 0.4992368221282959, + "learning_rate": 0.00015136732420453734, + "loss": 0.1105, + "step": 14011 + }, + { + "epoch": 0.32846334616759554, + "grad_norm": 0.5185089707374573, + "learning_rate": 0.00015136100400564413, + "loss": 0.7276, + "step": 14012 + }, + { + "epoch": 0.32848678774240053, + "grad_norm": 0.6223841309547424, + "learning_rate": 0.00015135468352806464, + "loss": 0.1066, + "step": 14013 + }, + { + "epoch": 0.3285102293172055, + "grad_norm": 0.2875397205352783, + "learning_rate": 0.0001513483627718332, + "loss": 0.0632, + "step": 14014 + }, + { + "epoch": 0.3285336708920105, + "grad_norm": 0.5205550789833069, + "learning_rate": 0.0001513420417369841, + "loss": 0.1262, + "step": 14015 + }, + { + "epoch": 0.3285571124668155, + "grad_norm": 0.5365009307861328, + "learning_rate": 0.00015133572042355163, + "loss": 0.1028, + "step": 14016 + }, + { + "epoch": 0.3285805540416205, + "grad_norm": 0.4516582190990448, + "learning_rate": 0.0001513293988315701, + "loss": 0.6147, + "step": 14017 + }, + { + "epoch": 0.3286039956164255, + "grad_norm": 0.4874345362186432, + "learning_rate": 0.00015132307696107383, + "loss": 0.0598, + "step": 14018 + }, + { + "epoch": 0.3286274371912305, + "grad_norm": 0.7510636448860168, + "learning_rate": 0.00015131675481209708, + "loss": 0.137, + "step": 14019 + }, + { + "epoch": 0.3286508787660355, + "grad_norm": 0.40198367834091187, + "learning_rate": 0.00015131043238467417, + "loss": 0.1006, + "step": 14020 + }, + { + "epoch": 0.3286743203408405, + "grad_norm": 0.6083430051803589, + "learning_rate": 0.00015130410967883944, + "loss": 0.141, + "step": 14021 + }, + { + "epoch": 0.3286977619156455, + "grad_norm": 0.13584807515144348, + "learning_rate": 0.00015129778669462716, + "loss": 0.0219, + "step": 14022 + }, + { + "epoch": 0.3287212034904505, + "grad_norm": 0.6064188480377197, + "learning_rate": 0.00015129146343207165, + "loss": 0.1718, + "step": 14023 + }, + { + "epoch": 0.32874464506525547, + "grad_norm": 0.43189629912376404, + "learning_rate": 0.00015128513989120718, + "loss": 0.0416, + "step": 14024 + }, + { + "epoch": 0.32876808664006046, + "grad_norm": 0.6185154914855957, + "learning_rate": 0.00015127881607206813, + "loss": 0.2053, + "step": 14025 + }, + { + "epoch": 0.32879152821486546, + "grad_norm": 0.666584312915802, + "learning_rate": 0.00015127249197468876, + "loss": 0.1779, + "step": 14026 + }, + { + "epoch": 0.32881496978967045, + "grad_norm": 0.08827901631593704, + "learning_rate": 0.00015126616759910343, + "loss": 0.0241, + "step": 14027 + }, + { + "epoch": 0.32883841136447545, + "grad_norm": 0.3567838668823242, + "learning_rate": 0.00015125984294534644, + "loss": 0.0982, + "step": 14028 + }, + { + "epoch": 0.32886185293928044, + "grad_norm": 0.7630572319030762, + "learning_rate": 0.00015125351801345208, + "loss": 0.2258, + "step": 14029 + }, + { + "epoch": 0.32888529451408544, + "grad_norm": 0.8593761324882507, + "learning_rate": 0.00015124719280345473, + "loss": 0.1388, + "step": 14030 + }, + { + "epoch": 0.32890873608889043, + "grad_norm": 0.4562183618545532, + "learning_rate": 0.0001512408673153886, + "loss": 0.0964, + "step": 14031 + }, + { + "epoch": 0.3289321776636954, + "grad_norm": 0.20314672589302063, + "learning_rate": 0.00015123454154928814, + "loss": 0.0437, + "step": 14032 + }, + { + "epoch": 0.3289556192385004, + "grad_norm": 0.6330533027648926, + "learning_rate": 0.00015122821550518758, + "loss": 0.1227, + "step": 14033 + }, + { + "epoch": 0.3289790608133054, + "grad_norm": 0.5138949751853943, + "learning_rate": 0.00015122188918312133, + "loss": 0.1294, + "step": 14034 + }, + { + "epoch": 0.3290025023881104, + "grad_norm": 0.3266935646533966, + "learning_rate": 0.00015121556258312362, + "loss": 0.0628, + "step": 14035 + }, + { + "epoch": 0.3290259439629154, + "grad_norm": 0.4032166004180908, + "learning_rate": 0.00015120923570522888, + "loss": 0.0958, + "step": 14036 + }, + { + "epoch": 0.32904938553772045, + "grad_norm": 0.5241751670837402, + "learning_rate": 0.00015120290854947133, + "loss": 0.1195, + "step": 14037 + }, + { + "epoch": 0.32907282711252545, + "grad_norm": 0.20672906935214996, + "learning_rate": 0.0001511965811158854, + "loss": 0.0461, + "step": 14038 + }, + { + "epoch": 0.32909626868733044, + "grad_norm": 0.954149603843689, + "learning_rate": 0.00015119025340450533, + "loss": 0.1877, + "step": 14039 + }, + { + "epoch": 0.32911971026213543, + "grad_norm": 0.865323007106781, + "learning_rate": 0.00015118392541536556, + "loss": 0.2453, + "step": 14040 + }, + { + "epoch": 0.32914315183694043, + "grad_norm": 0.3874357044696808, + "learning_rate": 0.00015117759714850036, + "loss": 0.0663, + "step": 14041 + }, + { + "epoch": 0.3291665934117454, + "grad_norm": 0.664949357509613, + "learning_rate": 0.00015117126860394407, + "loss": 0.1473, + "step": 14042 + }, + { + "epoch": 0.3291900349865504, + "grad_norm": 0.18836882710456848, + "learning_rate": 0.000151164939781731, + "loss": 0.0238, + "step": 14043 + }, + { + "epoch": 0.3292134765613554, + "grad_norm": 0.4486582577228546, + "learning_rate": 0.00015115861068189557, + "loss": 0.0719, + "step": 14044 + }, + { + "epoch": 0.3292369181361604, + "grad_norm": 0.9646596908569336, + "learning_rate": 0.00015115228130447209, + "loss": 0.1553, + "step": 14045 + }, + { + "epoch": 0.3292603597109654, + "grad_norm": 0.6203328371047974, + "learning_rate": 0.0001511459516494949, + "loss": 0.0679, + "step": 14046 + }, + { + "epoch": 0.3292838012857704, + "grad_norm": 0.14279258251190186, + "learning_rate": 0.0001511396217169983, + "loss": 0.0271, + "step": 14047 + }, + { + "epoch": 0.3293072428605754, + "grad_norm": 0.25989389419555664, + "learning_rate": 0.0001511332915070167, + "loss": 0.0529, + "step": 14048 + }, + { + "epoch": 0.3293306844353804, + "grad_norm": 0.6119366884231567, + "learning_rate": 0.0001511269610195844, + "loss": 0.1699, + "step": 14049 + }, + { + "epoch": 0.3293541260101854, + "grad_norm": 0.6115597486495972, + "learning_rate": 0.0001511206302547358, + "loss": 0.1133, + "step": 14050 + }, + { + "epoch": 0.32937756758499037, + "grad_norm": 0.1398988962173462, + "learning_rate": 0.00015111429921250524, + "loss": 0.0217, + "step": 14051 + }, + { + "epoch": 0.32940100915979537, + "grad_norm": 0.15551115572452545, + "learning_rate": 0.00015110796789292698, + "loss": 0.0282, + "step": 14052 + }, + { + "epoch": 0.32942445073460036, + "grad_norm": 0.5958268642425537, + "learning_rate": 0.00015110163629603553, + "loss": 0.1431, + "step": 14053 + }, + { + "epoch": 0.32944789230940535, + "grad_norm": 0.5557399988174438, + "learning_rate": 0.00015109530442186512, + "loss": 0.1556, + "step": 14054 + }, + { + "epoch": 0.32947133388421035, + "grad_norm": 0.38703835010528564, + "learning_rate": 0.00015108897227045015, + "loss": 0.0744, + "step": 14055 + }, + { + "epoch": 0.32949477545901534, + "grad_norm": 0.6122789978981018, + "learning_rate": 0.00015108263984182498, + "loss": 0.1025, + "step": 14056 + }, + { + "epoch": 0.32951821703382034, + "grad_norm": 0.44474226236343384, + "learning_rate": 0.00015107630713602399, + "loss": 0.0665, + "step": 14057 + }, + { + "epoch": 0.32954165860862533, + "grad_norm": 0.3791883885860443, + "learning_rate": 0.00015106997415308154, + "loss": 0.0793, + "step": 14058 + }, + { + "epoch": 0.3295651001834303, + "grad_norm": 0.31301024556159973, + "learning_rate": 0.00015106364089303192, + "loss": 0.0679, + "step": 14059 + }, + { + "epoch": 0.3295885417582353, + "grad_norm": 0.6446639895439148, + "learning_rate": 0.0001510573073559096, + "loss": 0.1265, + "step": 14060 + }, + { + "epoch": 0.3296119833330403, + "grad_norm": 0.22063088417053223, + "learning_rate": 0.00015105097354174886, + "loss": 0.0342, + "step": 14061 + }, + { + "epoch": 0.3296354249078453, + "grad_norm": 0.38454726338386536, + "learning_rate": 0.0001510446394505841, + "loss": 0.0828, + "step": 14062 + }, + { + "epoch": 0.3296588664826503, + "grad_norm": 0.50630122423172, + "learning_rate": 0.0001510383050824497, + "loss": 0.1415, + "step": 14063 + }, + { + "epoch": 0.3296823080574553, + "grad_norm": 0.5458316802978516, + "learning_rate": 0.00015103197043738002, + "loss": 0.0951, + "step": 14064 + }, + { + "epoch": 0.3297057496322603, + "grad_norm": 0.7377055287361145, + "learning_rate": 0.00015102563551540945, + "loss": 0.1257, + "step": 14065 + }, + { + "epoch": 0.3297291912070653, + "grad_norm": 0.5359458327293396, + "learning_rate": 0.00015101930031657235, + "loss": 0.1093, + "step": 14066 + }, + { + "epoch": 0.3297526327818703, + "grad_norm": 0.6233419179916382, + "learning_rate": 0.00015101296484090307, + "loss": 0.1106, + "step": 14067 + }, + { + "epoch": 0.3297760743566753, + "grad_norm": 0.5416567325592041, + "learning_rate": 0.000151006629088436, + "loss": 0.1315, + "step": 14068 + }, + { + "epoch": 0.32979951593148027, + "grad_norm": 0.801149308681488, + "learning_rate": 0.00015100029305920555, + "loss": 0.1418, + "step": 14069 + }, + { + "epoch": 0.32982295750628526, + "grad_norm": 0.5923015475273132, + "learning_rate": 0.00015099395675324608, + "loss": 0.1307, + "step": 14070 + }, + { + "epoch": 0.32984639908109026, + "grad_norm": 0.5268746018409729, + "learning_rate": 0.00015098762017059193, + "loss": 0.0854, + "step": 14071 + }, + { + "epoch": 0.32986984065589525, + "grad_norm": 0.48076653480529785, + "learning_rate": 0.00015098128331127758, + "loss": 0.1014, + "step": 14072 + }, + { + "epoch": 0.32989328223070025, + "grad_norm": 0.1781310737133026, + "learning_rate": 0.0001509749461753373, + "loss": 0.023, + "step": 14073 + }, + { + "epoch": 0.32991672380550524, + "grad_norm": 0.46674084663391113, + "learning_rate": 0.00015096860876280554, + "loss": 0.0974, + "step": 14074 + }, + { + "epoch": 0.32994016538031024, + "grad_norm": 0.17792800068855286, + "learning_rate": 0.00015096227107371668, + "loss": 0.0163, + "step": 14075 + }, + { + "epoch": 0.32996360695511523, + "grad_norm": 0.2094222903251648, + "learning_rate": 0.0001509559331081051, + "loss": 0.0444, + "step": 14076 + }, + { + "epoch": 0.3299870485299202, + "grad_norm": 0.41940411925315857, + "learning_rate": 0.00015094959486600522, + "loss": 0.053, + "step": 14077 + }, + { + "epoch": 0.3300104901047252, + "grad_norm": 0.179127499461174, + "learning_rate": 0.00015094325634745136, + "loss": 0.0532, + "step": 14078 + }, + { + "epoch": 0.3300339316795302, + "grad_norm": 0.8107694983482361, + "learning_rate": 0.00015093691755247801, + "loss": 0.1417, + "step": 14079 + }, + { + "epoch": 0.3300573732543352, + "grad_norm": 0.78984534740448, + "learning_rate": 0.00015093057848111947, + "loss": 0.1194, + "step": 14080 + }, + { + "epoch": 0.3300808148291402, + "grad_norm": 0.4720235764980316, + "learning_rate": 0.0001509242391334102, + "loss": 0.1187, + "step": 14081 + }, + { + "epoch": 0.3301042564039452, + "grad_norm": 0.19140873849391937, + "learning_rate": 0.00015091789950938457, + "loss": 0.0507, + "step": 14082 + }, + { + "epoch": 0.3301276979787502, + "grad_norm": 0.8961655497550964, + "learning_rate": 0.00015091155960907698, + "loss": 0.1661, + "step": 14083 + }, + { + "epoch": 0.3301511395535552, + "grad_norm": 0.42292243242263794, + "learning_rate": 0.00015090521943252188, + "loss": 0.0769, + "step": 14084 + }, + { + "epoch": 0.3301745811283602, + "grad_norm": 0.16501279175281525, + "learning_rate": 0.0001508988789797536, + "loss": 0.0417, + "step": 14085 + }, + { + "epoch": 0.3301980227031652, + "grad_norm": 0.40883681178092957, + "learning_rate": 0.00015089253825080655, + "loss": 0.0798, + "step": 14086 + }, + { + "epoch": 0.33022146427797017, + "grad_norm": 0.4812222123146057, + "learning_rate": 0.00015088619724571513, + "loss": 0.0929, + "step": 14087 + }, + { + "epoch": 0.3302449058527752, + "grad_norm": 0.5724904537200928, + "learning_rate": 0.00015087985596451384, + "loss": 0.1892, + "step": 14088 + }, + { + "epoch": 0.3302683474275802, + "grad_norm": 0.497938334941864, + "learning_rate": 0.000150873514407237, + "loss": 0.1388, + "step": 14089 + }, + { + "epoch": 0.3302917890023852, + "grad_norm": 0.3581421971321106, + "learning_rate": 0.00015086717257391903, + "loss": 0.067, + "step": 14090 + }, + { + "epoch": 0.3303152305771902, + "grad_norm": 0.338069885969162, + "learning_rate": 0.0001508608304645944, + "loss": 0.067, + "step": 14091 + }, + { + "epoch": 0.3303386721519952, + "grad_norm": 0.38893479108810425, + "learning_rate": 0.00015085448807929737, + "loss": 0.0899, + "step": 14092 + }, + { + "epoch": 0.3303621137268002, + "grad_norm": 0.6625792384147644, + "learning_rate": 0.00015084814541806252, + "loss": 0.1443, + "step": 14093 + }, + { + "epoch": 0.3303855553016052, + "grad_norm": 0.15761129558086395, + "learning_rate": 0.00015084180248092421, + "loss": 0.0248, + "step": 14094 + }, + { + "epoch": 0.3304089968764102, + "grad_norm": 0.3691204786300659, + "learning_rate": 0.00015083545926791687, + "loss": 0.0665, + "step": 14095 + }, + { + "epoch": 0.33043243845121517, + "grad_norm": 0.27553778886795044, + "learning_rate": 0.00015082911577907483, + "loss": 0.0768, + "step": 14096 + }, + { + "epoch": 0.33045588002602017, + "grad_norm": 0.4397900700569153, + "learning_rate": 0.00015082277201443263, + "loss": 0.127, + "step": 14097 + }, + { + "epoch": 0.33047932160082516, + "grad_norm": 0.45108675956726074, + "learning_rate": 0.00015081642797402464, + "loss": 0.4195, + "step": 14098 + }, + { + "epoch": 0.33050276317563015, + "grad_norm": 0.6114065051078796, + "learning_rate": 0.00015081008365788528, + "loss": 0.1068, + "step": 14099 + }, + { + "epoch": 0.33052620475043515, + "grad_norm": 0.6927388906478882, + "learning_rate": 0.00015080373906604897, + "loss": 0.0721, + "step": 14100 + }, + { + "epoch": 0.33054964632524014, + "grad_norm": 0.5508027076721191, + "learning_rate": 0.00015079739419855014, + "loss": 0.1904, + "step": 14101 + }, + { + "epoch": 0.33057308790004514, + "grad_norm": 0.5100421905517578, + "learning_rate": 0.00015079104905542322, + "loss": 0.0956, + "step": 14102 + }, + { + "epoch": 0.33059652947485013, + "grad_norm": 0.346426784992218, + "learning_rate": 0.00015078470363670264, + "loss": 0.0858, + "step": 14103 + }, + { + "epoch": 0.3306199710496551, + "grad_norm": 0.5040380358695984, + "learning_rate": 0.0001507783579424228, + "loss": 0.0839, + "step": 14104 + }, + { + "epoch": 0.3306434126244601, + "grad_norm": 0.09119637310504913, + "learning_rate": 0.00015077201197261824, + "loss": 0.0122, + "step": 14105 + }, + { + "epoch": 0.3306668541992651, + "grad_norm": 0.5156355500221252, + "learning_rate": 0.00015076566572732325, + "loss": 0.1002, + "step": 14106 + }, + { + "epoch": 0.3306902957740701, + "grad_norm": 0.46669405698776245, + "learning_rate": 0.00015075931920657235, + "loss": 0.1025, + "step": 14107 + }, + { + "epoch": 0.3307137373488751, + "grad_norm": 0.32597047090530396, + "learning_rate": 0.0001507529724104, + "loss": 0.059, + "step": 14108 + }, + { + "epoch": 0.3307371789236801, + "grad_norm": 0.14206308126449585, + "learning_rate": 0.00015074662533884052, + "loss": 0.0153, + "step": 14109 + }, + { + "epoch": 0.3307606204984851, + "grad_norm": 0.7857551574707031, + "learning_rate": 0.00015074027799192845, + "loss": 0.1957, + "step": 14110 + }, + { + "epoch": 0.3307840620732901, + "grad_norm": 0.4823411703109741, + "learning_rate": 0.00015073393036969823, + "loss": 0.1387, + "step": 14111 + }, + { + "epoch": 0.3308075036480951, + "grad_norm": 0.5709383487701416, + "learning_rate": 0.00015072758247218425, + "loss": 0.1348, + "step": 14112 + }, + { + "epoch": 0.3308309452229001, + "grad_norm": 0.48312845826148987, + "learning_rate": 0.000150721234299421, + "loss": 0.0857, + "step": 14113 + }, + { + "epoch": 0.33085438679770507, + "grad_norm": 0.3870198726654053, + "learning_rate": 0.00015071488585144288, + "loss": 0.0787, + "step": 14114 + }, + { + "epoch": 0.33087782837251006, + "grad_norm": 0.38778606057167053, + "learning_rate": 0.00015070853712828436, + "loss": 0.0853, + "step": 14115 + }, + { + "epoch": 0.33090126994731506, + "grad_norm": 0.7589985132217407, + "learning_rate": 0.00015070218812997995, + "loss": 0.851, + "step": 14116 + }, + { + "epoch": 0.33092471152212005, + "grad_norm": 0.310213565826416, + "learning_rate": 0.00015069583885656397, + "loss": 0.117, + "step": 14117 + }, + { + "epoch": 0.33094815309692505, + "grad_norm": 0.45902761816978455, + "learning_rate": 0.000150689489308071, + "loss": 0.4175, + "step": 14118 + }, + { + "epoch": 0.33097159467173004, + "grad_norm": 0.5516729354858398, + "learning_rate": 0.0001506831394845354, + "loss": 0.1229, + "step": 14119 + }, + { + "epoch": 0.33099503624653503, + "grad_norm": 0.33722564578056335, + "learning_rate": 0.0001506767893859917, + "loss": 0.0338, + "step": 14120 + }, + { + "epoch": 0.33101847782134003, + "grad_norm": 0.28791603446006775, + "learning_rate": 0.00015067043901247426, + "loss": 0.0595, + "step": 14121 + }, + { + "epoch": 0.331041919396145, + "grad_norm": 1.1965643167495728, + "learning_rate": 0.0001506640883640176, + "loss": 0.1758, + "step": 14122 + }, + { + "epoch": 0.33106536097095, + "grad_norm": 0.40267202258110046, + "learning_rate": 0.0001506577374406562, + "loss": 0.0753, + "step": 14123 + }, + { + "epoch": 0.331088802545755, + "grad_norm": 0.47547900676727295, + "learning_rate": 0.0001506513862424245, + "loss": 0.0515, + "step": 14124 + }, + { + "epoch": 0.33111224412056, + "grad_norm": 0.2262568175792694, + "learning_rate": 0.00015064503476935694, + "loss": 0.0383, + "step": 14125 + }, + { + "epoch": 0.331135685695365, + "grad_norm": 0.20974518358707428, + "learning_rate": 0.000150638683021488, + "loss": 0.0397, + "step": 14126 + }, + { + "epoch": 0.33115912727017, + "grad_norm": 0.2702401578426361, + "learning_rate": 0.00015063233099885218, + "loss": 0.0503, + "step": 14127 + }, + { + "epoch": 0.331182568844975, + "grad_norm": 0.7438949346542358, + "learning_rate": 0.00015062597870148383, + "loss": 0.1386, + "step": 14128 + }, + { + "epoch": 0.33120601041978, + "grad_norm": 0.4631155729293823, + "learning_rate": 0.00015061962612941752, + "loss": 0.1335, + "step": 14129 + }, + { + "epoch": 0.331229451994585, + "grad_norm": 0.33302661776542664, + "learning_rate": 0.0001506132732826877, + "loss": 0.0776, + "step": 14130 + }, + { + "epoch": 0.33125289356939, + "grad_norm": 0.1805279552936554, + "learning_rate": 0.00015060692016132885, + "loss": 0.0498, + "step": 14131 + }, + { + "epoch": 0.33127633514419497, + "grad_norm": 0.3272063732147217, + "learning_rate": 0.00015060056676537544, + "loss": 0.0712, + "step": 14132 + }, + { + "epoch": 0.33129977671899996, + "grad_norm": 0.465030699968338, + "learning_rate": 0.00015059421309486192, + "loss": 0.1223, + "step": 14133 + }, + { + "epoch": 0.33132321829380496, + "grad_norm": 0.994282066822052, + "learning_rate": 0.00015058785914982275, + "loss": 0.0761, + "step": 14134 + }, + { + "epoch": 0.33134665986860995, + "grad_norm": 0.3506230115890503, + "learning_rate": 0.00015058150493029246, + "loss": 0.0547, + "step": 14135 + }, + { + "epoch": 0.33137010144341494, + "grad_norm": 0.6341761946678162, + "learning_rate": 0.0001505751504363055, + "loss": 0.1762, + "step": 14136 + }, + { + "epoch": 0.33139354301821994, + "grad_norm": 0.6956007480621338, + "learning_rate": 0.00015056879566789636, + "loss": 0.1544, + "step": 14137 + }, + { + "epoch": 0.33141698459302493, + "grad_norm": 0.9155228734016418, + "learning_rate": 0.0001505624406250995, + "loss": 0.2265, + "step": 14138 + }, + { + "epoch": 0.33144042616783, + "grad_norm": 0.4846557676792145, + "learning_rate": 0.0001505560853079494, + "loss": 0.0706, + "step": 14139 + }, + { + "epoch": 0.331463867742635, + "grad_norm": 0.36081311106681824, + "learning_rate": 0.00015054972971648058, + "loss": 0.0726, + "step": 14140 + }, + { + "epoch": 0.33148730931743997, + "grad_norm": 0.37041589617729187, + "learning_rate": 0.0001505433738507275, + "loss": 0.062, + "step": 14141 + }, + { + "epoch": 0.33151075089224497, + "grad_norm": 0.6325158476829529, + "learning_rate": 0.0001505370177107247, + "loss": 0.1572, + "step": 14142 + }, + { + "epoch": 0.33153419246704996, + "grad_norm": 0.489793062210083, + "learning_rate": 0.00015053066129650655, + "loss": 0.1439, + "step": 14143 + }, + { + "epoch": 0.33155763404185495, + "grad_norm": 0.42402899265289307, + "learning_rate": 0.00015052430460810762, + "loss": 0.0507, + "step": 14144 + }, + { + "epoch": 0.33158107561665995, + "grad_norm": 0.11598457396030426, + "learning_rate": 0.00015051794764556244, + "loss": 0.0334, + "step": 14145 + }, + { + "epoch": 0.33160451719146494, + "grad_norm": 0.17293256521224976, + "learning_rate": 0.00015051159040890542, + "loss": 0.0461, + "step": 14146 + }, + { + "epoch": 0.33162795876626994, + "grad_norm": 0.624396026134491, + "learning_rate": 0.00015050523289817107, + "loss": 0.682, + "step": 14147 + }, + { + "epoch": 0.33165140034107493, + "grad_norm": 0.35569119453430176, + "learning_rate": 0.00015049887511339394, + "loss": 0.0462, + "step": 14148 + }, + { + "epoch": 0.3316748419158799, + "grad_norm": 0.35994115471839905, + "learning_rate": 0.0001504925170546085, + "loss": 0.0724, + "step": 14149 + }, + { + "epoch": 0.3316982834906849, + "grad_norm": 0.594034731388092, + "learning_rate": 0.00015048615872184922, + "loss": 0.1784, + "step": 14150 + }, + { + "epoch": 0.3317217250654899, + "grad_norm": 0.48663944005966187, + "learning_rate": 0.00015047980011515061, + "loss": 0.0782, + "step": 14151 + }, + { + "epoch": 0.3317451666402949, + "grad_norm": 0.3150002956390381, + "learning_rate": 0.00015047344123454725, + "loss": 0.0419, + "step": 14152 + }, + { + "epoch": 0.3317686082150999, + "grad_norm": 0.9013993144035339, + "learning_rate": 0.0001504670820800735, + "loss": 0.1841, + "step": 14153 + }, + { + "epoch": 0.3317920497899049, + "grad_norm": 0.6261090040206909, + "learning_rate": 0.000150460722651764, + "loss": 0.1341, + "step": 14154 + }, + { + "epoch": 0.3318154913647099, + "grad_norm": 0.2504160702228546, + "learning_rate": 0.0001504543629496532, + "loss": 0.0764, + "step": 14155 + }, + { + "epoch": 0.3318389329395149, + "grad_norm": 0.5449961423873901, + "learning_rate": 0.00015044800297377554, + "loss": 0.0912, + "step": 14156 + }, + { + "epoch": 0.3318623745143199, + "grad_norm": 0.4327603578567505, + "learning_rate": 0.00015044164272416566, + "loss": 0.088, + "step": 14157 + }, + { + "epoch": 0.3318858160891249, + "grad_norm": 0.33151715993881226, + "learning_rate": 0.00015043528220085798, + "loss": 0.061, + "step": 14158 + }, + { + "epoch": 0.33190925766392987, + "grad_norm": 0.6058264970779419, + "learning_rate": 0.00015042892140388706, + "loss": 0.1307, + "step": 14159 + }, + { + "epoch": 0.33193269923873486, + "grad_norm": 0.3957872986793518, + "learning_rate": 0.00015042256033328738, + "loss": 0.0769, + "step": 14160 + }, + { + "epoch": 0.33195614081353986, + "grad_norm": 0.699434757232666, + "learning_rate": 0.0001504161989890935, + "loss": 0.1717, + "step": 14161 + }, + { + "epoch": 0.33197958238834485, + "grad_norm": 0.16648215055465698, + "learning_rate": 0.0001504098373713399, + "loss": 0.0438, + "step": 14162 + }, + { + "epoch": 0.33200302396314985, + "grad_norm": 0.5822352766990662, + "learning_rate": 0.00015040347548006108, + "loss": 0.0963, + "step": 14163 + }, + { + "epoch": 0.33202646553795484, + "grad_norm": 0.7626376152038574, + "learning_rate": 0.0001503971133152916, + "loss": 0.0784, + "step": 14164 + }, + { + "epoch": 0.33204990711275983, + "grad_norm": 0.29590365290641785, + "learning_rate": 0.00015039075087706593, + "loss": 0.077, + "step": 14165 + }, + { + "epoch": 0.33207334868756483, + "grad_norm": 0.1112067848443985, + "learning_rate": 0.00015038438816541863, + "loss": 0.0263, + "step": 14166 + }, + { + "epoch": 0.3320967902623698, + "grad_norm": 0.47136515378952026, + "learning_rate": 0.00015037802518038428, + "loss": 0.0644, + "step": 14167 + }, + { + "epoch": 0.3321202318371748, + "grad_norm": 0.1832766979932785, + "learning_rate": 0.00015037166192199733, + "loss": 0.0386, + "step": 14168 + }, + { + "epoch": 0.3321436734119798, + "grad_norm": 0.4197937250137329, + "learning_rate": 0.0001503652983902923, + "loss": 0.0644, + "step": 14169 + }, + { + "epoch": 0.3321671149867848, + "grad_norm": 0.2557789981365204, + "learning_rate": 0.00015035893458530377, + "loss": 0.0512, + "step": 14170 + }, + { + "epoch": 0.3321905565615898, + "grad_norm": 0.15787285566329956, + "learning_rate": 0.00015035257050706623, + "loss": 0.0482, + "step": 14171 + }, + { + "epoch": 0.3322139981363948, + "grad_norm": 0.5429120659828186, + "learning_rate": 0.00015034620615561424, + "loss": 0.1066, + "step": 14172 + }, + { + "epoch": 0.3322374397111998, + "grad_norm": 0.11089646816253662, + "learning_rate": 0.00015033984153098233, + "loss": 0.0199, + "step": 14173 + }, + { + "epoch": 0.3322608812860048, + "grad_norm": 0.46529316902160645, + "learning_rate": 0.00015033347663320502, + "loss": 0.0797, + "step": 14174 + }, + { + "epoch": 0.3322843228608098, + "grad_norm": 0.7086760401725769, + "learning_rate": 0.00015032711146231682, + "loss": 0.6724, + "step": 14175 + }, + { + "epoch": 0.33230776443561477, + "grad_norm": 0.1513485312461853, + "learning_rate": 0.0001503207460183523, + "loss": 0.0496, + "step": 14176 + }, + { + "epoch": 0.33233120601041977, + "grad_norm": 0.5448741316795349, + "learning_rate": 0.00015031438030134605, + "loss": 0.1004, + "step": 14177 + }, + { + "epoch": 0.33235464758522476, + "grad_norm": 0.48312556743621826, + "learning_rate": 0.0001503080143113325, + "loss": 0.0665, + "step": 14178 + }, + { + "epoch": 0.33237808916002975, + "grad_norm": 0.42114129662513733, + "learning_rate": 0.00015030164804834625, + "loss": 0.0851, + "step": 14179 + }, + { + "epoch": 0.33240153073483475, + "grad_norm": 0.5763155221939087, + "learning_rate": 0.0001502952815124219, + "loss": 0.0888, + "step": 14180 + }, + { + "epoch": 0.33242497230963974, + "grad_norm": 0.691920280456543, + "learning_rate": 0.0001502889147035939, + "loss": 0.1156, + "step": 14181 + }, + { + "epoch": 0.33244841388444474, + "grad_norm": 0.49919670820236206, + "learning_rate": 0.00015028254762189684, + "loss": 0.0892, + "step": 14182 + }, + { + "epoch": 0.33247185545924973, + "grad_norm": 0.1387040764093399, + "learning_rate": 0.00015027618026736524, + "loss": 0.04, + "step": 14183 + }, + { + "epoch": 0.3324952970340547, + "grad_norm": 0.08353344351053238, + "learning_rate": 0.0001502698126400337, + "loss": 0.0274, + "step": 14184 + }, + { + "epoch": 0.3325187386088597, + "grad_norm": 0.3808317184448242, + "learning_rate": 0.00015026344473993673, + "loss": 0.1119, + "step": 14185 + }, + { + "epoch": 0.3325421801836647, + "grad_norm": 0.9292437434196472, + "learning_rate": 0.0001502570765671089, + "loss": 0.1933, + "step": 14186 + }, + { + "epoch": 0.3325656217584697, + "grad_norm": 0.38883107900619507, + "learning_rate": 0.00015025070812158477, + "loss": 0.0764, + "step": 14187 + }, + { + "epoch": 0.3325890633332747, + "grad_norm": 0.6328309178352356, + "learning_rate": 0.00015024433940339887, + "loss": 0.1255, + "step": 14188 + }, + { + "epoch": 0.3326125049080797, + "grad_norm": 0.2076379805803299, + "learning_rate": 0.00015023797041258577, + "loss": 0.0951, + "step": 14189 + }, + { + "epoch": 0.33263594648288475, + "grad_norm": 0.18411359190940857, + "learning_rate": 0.00015023160114918004, + "loss": 0.0564, + "step": 14190 + }, + { + "epoch": 0.33265938805768974, + "grad_norm": 0.4847903251647949, + "learning_rate": 0.00015022523161321622, + "loss": 0.7061, + "step": 14191 + }, + { + "epoch": 0.33268282963249474, + "grad_norm": 0.5153317451477051, + "learning_rate": 0.00015021886180472888, + "loss": 0.0795, + "step": 14192 + }, + { + "epoch": 0.33270627120729973, + "grad_norm": 0.37176334857940674, + "learning_rate": 0.00015021249172375264, + "loss": 0.427, + "step": 14193 + }, + { + "epoch": 0.3327297127821047, + "grad_norm": 0.4742012023925781, + "learning_rate": 0.00015020612137032192, + "loss": 0.0729, + "step": 14194 + }, + { + "epoch": 0.3327531543569097, + "grad_norm": 0.41411104798316956, + "learning_rate": 0.00015019975074447142, + "loss": 0.0987, + "step": 14195 + }, + { + "epoch": 0.3327765959317147, + "grad_norm": 0.47717759013175964, + "learning_rate": 0.00015019337984623565, + "loss": 0.0747, + "step": 14196 + }, + { + "epoch": 0.3328000375065197, + "grad_norm": 0.4547243118286133, + "learning_rate": 0.0001501870086756492, + "loss": 0.1031, + "step": 14197 + }, + { + "epoch": 0.3328234790813247, + "grad_norm": 0.5557578206062317, + "learning_rate": 0.00015018063723274663, + "loss": 0.0609, + "step": 14198 + }, + { + "epoch": 0.3328469206561297, + "grad_norm": 0.5487815737724304, + "learning_rate": 0.00015017426551756252, + "loss": 0.1635, + "step": 14199 + }, + { + "epoch": 0.3328703622309347, + "grad_norm": 0.5871159434318542, + "learning_rate": 0.00015016789353013144, + "loss": 0.1491, + "step": 14200 + }, + { + "epoch": 0.3328938038057397, + "grad_norm": 0.6131447553634644, + "learning_rate": 0.0001501615212704879, + "loss": 0.1241, + "step": 14201 + }, + { + "epoch": 0.3329172453805447, + "grad_norm": 0.4465358257293701, + "learning_rate": 0.00015015514873866658, + "loss": 0.0609, + "step": 14202 + }, + { + "epoch": 0.3329406869553497, + "grad_norm": 1.5762355327606201, + "learning_rate": 0.000150148775934702, + "loss": 0.1719, + "step": 14203 + }, + { + "epoch": 0.33296412853015467, + "grad_norm": 0.14883500337600708, + "learning_rate": 0.00015014240285862878, + "loss": 0.0216, + "step": 14204 + }, + { + "epoch": 0.33298757010495966, + "grad_norm": 0.27474403381347656, + "learning_rate": 0.00015013602951048145, + "loss": 0.0458, + "step": 14205 + }, + { + "epoch": 0.33301101167976466, + "grad_norm": 0.2289172261953354, + "learning_rate": 0.00015012965589029462, + "loss": 0.0581, + "step": 14206 + }, + { + "epoch": 0.33303445325456965, + "grad_norm": 0.3516371548175812, + "learning_rate": 0.00015012328199810285, + "loss": 0.0719, + "step": 14207 + }, + { + "epoch": 0.33305789482937465, + "grad_norm": 0.4024115800857544, + "learning_rate": 0.00015011690783394075, + "loss": 0.1122, + "step": 14208 + }, + { + "epoch": 0.33308133640417964, + "grad_norm": 0.40379059314727783, + "learning_rate": 0.0001501105333978429, + "loss": 0.0826, + "step": 14209 + }, + { + "epoch": 0.33310477797898463, + "grad_norm": 0.4093494415283203, + "learning_rate": 0.0001501041586898439, + "loss": 0.0993, + "step": 14210 + }, + { + "epoch": 0.33312821955378963, + "grad_norm": 0.3651464581489563, + "learning_rate": 0.0001500977837099783, + "loss": 0.043, + "step": 14211 + }, + { + "epoch": 0.3331516611285946, + "grad_norm": 0.40132516622543335, + "learning_rate": 0.00015009140845828074, + "loss": 0.0865, + "step": 14212 + }, + { + "epoch": 0.3331751027033996, + "grad_norm": 0.3705034852027893, + "learning_rate": 0.00015008503293478578, + "loss": 0.0585, + "step": 14213 + }, + { + "epoch": 0.3331985442782046, + "grad_norm": 0.4845488965511322, + "learning_rate": 0.000150078657139528, + "loss": 0.1158, + "step": 14214 + }, + { + "epoch": 0.3332219858530096, + "grad_norm": 0.3424425423145294, + "learning_rate": 0.00015007228107254207, + "loss": 0.0654, + "step": 14215 + }, + { + "epoch": 0.3332454274278146, + "grad_norm": 0.6879425644874573, + "learning_rate": 0.00015006590473386252, + "loss": 0.148, + "step": 14216 + }, + { + "epoch": 0.3332688690026196, + "grad_norm": 0.295646995306015, + "learning_rate": 0.00015005952812352394, + "loss": 0.0764, + "step": 14217 + }, + { + "epoch": 0.3332923105774246, + "grad_norm": 1.1543853282928467, + "learning_rate": 0.00015005315124156098, + "loss": 0.1308, + "step": 14218 + }, + { + "epoch": 0.3333157521522296, + "grad_norm": 0.5934743285179138, + "learning_rate": 0.00015004677408800822, + "loss": 0.1575, + "step": 14219 + }, + { + "epoch": 0.3333391937270346, + "grad_norm": 0.6272156238555908, + "learning_rate": 0.00015004039666290022, + "loss": 0.1416, + "step": 14220 + }, + { + "epoch": 0.33336263530183957, + "grad_norm": 0.3915150761604309, + "learning_rate": 0.00015003401896627162, + "loss": 0.0963, + "step": 14221 + }, + { + "epoch": 0.33338607687664457, + "grad_norm": 0.429423063993454, + "learning_rate": 0.00015002764099815704, + "loss": 0.0855, + "step": 14222 + }, + { + "epoch": 0.33340951845144956, + "grad_norm": 0.139067143201828, + "learning_rate": 0.00015002126275859108, + "loss": 0.029, + "step": 14223 + }, + { + "epoch": 0.33343296002625455, + "grad_norm": 0.42807793617248535, + "learning_rate": 0.00015001488424760834, + "loss": 0.0977, + "step": 14224 + }, + { + "epoch": 0.33345640160105955, + "grad_norm": 0.868190348148346, + "learning_rate": 0.00015000850546524345, + "loss": 0.1334, + "step": 14225 + }, + { + "epoch": 0.33347984317586454, + "grad_norm": 0.5348873734474182, + "learning_rate": 0.00015000212641153097, + "loss": 0.1278, + "step": 14226 + }, + { + "epoch": 0.33350328475066954, + "grad_norm": 0.8118534684181213, + "learning_rate": 0.00014999574708650558, + "loss": 0.134, + "step": 14227 + }, + { + "epoch": 0.33352672632547453, + "grad_norm": 0.5095423460006714, + "learning_rate": 0.00014998936749020182, + "loss": 0.1635, + "step": 14228 + }, + { + "epoch": 0.3335501679002795, + "grad_norm": 0.5206950306892395, + "learning_rate": 0.00014998298762265437, + "loss": 0.0676, + "step": 14229 + }, + { + "epoch": 0.3335736094750845, + "grad_norm": 0.3534298837184906, + "learning_rate": 0.00014997660748389784, + "loss": 0.0397, + "step": 14230 + }, + { + "epoch": 0.3335970510498895, + "grad_norm": 0.5179874897003174, + "learning_rate": 0.0001499702270739668, + "loss": 0.1934, + "step": 14231 + }, + { + "epoch": 0.3336204926246945, + "grad_norm": 0.708382248878479, + "learning_rate": 0.0001499638463928959, + "loss": 0.1048, + "step": 14232 + }, + { + "epoch": 0.3336439341994995, + "grad_norm": 0.7864917516708374, + "learning_rate": 0.00014995746544071978, + "loss": 0.2118, + "step": 14233 + }, + { + "epoch": 0.3336673757743045, + "grad_norm": 0.7559961080551147, + "learning_rate": 0.00014995108421747306, + "loss": 0.176, + "step": 14234 + }, + { + "epoch": 0.3336908173491095, + "grad_norm": 0.4278261959552765, + "learning_rate": 0.00014994470272319035, + "loss": 0.1465, + "step": 14235 + }, + { + "epoch": 0.3337142589239145, + "grad_norm": 0.27070048451423645, + "learning_rate": 0.00014993832095790626, + "loss": 0.0566, + "step": 14236 + }, + { + "epoch": 0.3337377004987195, + "grad_norm": 0.09528899192810059, + "learning_rate": 0.00014993193892165544, + "loss": 0.0176, + "step": 14237 + }, + { + "epoch": 0.3337611420735245, + "grad_norm": 0.904759407043457, + "learning_rate": 0.00014992555661447253, + "loss": 0.2749, + "step": 14238 + }, + { + "epoch": 0.33378458364832947, + "grad_norm": 0.5005319714546204, + "learning_rate": 0.00014991917403639214, + "loss": 0.1334, + "step": 14239 + }, + { + "epoch": 0.33380802522313446, + "grad_norm": 0.10078892856836319, + "learning_rate": 0.0001499127911874489, + "loss": 0.0235, + "step": 14240 + }, + { + "epoch": 0.33383146679793946, + "grad_norm": 0.39033347368240356, + "learning_rate": 0.00014990640806767747, + "loss": 0.1016, + "step": 14241 + }, + { + "epoch": 0.3338549083727445, + "grad_norm": 0.619540274143219, + "learning_rate": 0.00014990002467711244, + "loss": 0.1229, + "step": 14242 + }, + { + "epoch": 0.3338783499475495, + "grad_norm": 0.2856772541999817, + "learning_rate": 0.0001498936410157885, + "loss": 0.0641, + "step": 14243 + }, + { + "epoch": 0.3339017915223545, + "grad_norm": 0.5252335667610168, + "learning_rate": 0.00014988725708374022, + "loss": 0.1451, + "step": 14244 + }, + { + "epoch": 0.3339252330971595, + "grad_norm": 0.31087520718574524, + "learning_rate": 0.0001498808728810023, + "loss": 0.0991, + "step": 14245 + }, + { + "epoch": 0.3339486746719645, + "grad_norm": 0.8435632586479187, + "learning_rate": 0.00014987448840760937, + "loss": 0.098, + "step": 14246 + }, + { + "epoch": 0.3339721162467695, + "grad_norm": 0.3912420868873596, + "learning_rate": 0.0001498681036635961, + "loss": 0.0771, + "step": 14247 + }, + { + "epoch": 0.3339955578215745, + "grad_norm": 0.21027742326259613, + "learning_rate": 0.00014986171864899702, + "loss": 0.0534, + "step": 14248 + }, + { + "epoch": 0.33401899939637947, + "grad_norm": 0.6631720662117004, + "learning_rate": 0.00014985533336384692, + "loss": 0.1929, + "step": 14249 + }, + { + "epoch": 0.33404244097118446, + "grad_norm": 0.4333011209964752, + "learning_rate": 0.00014984894780818034, + "loss": 0.0625, + "step": 14250 + }, + { + "epoch": 0.33406588254598946, + "grad_norm": 0.18164575099945068, + "learning_rate": 0.000149842561982032, + "loss": 0.0354, + "step": 14251 + }, + { + "epoch": 0.33408932412079445, + "grad_norm": 0.33536696434020996, + "learning_rate": 0.00014983617588543648, + "loss": 0.074, + "step": 14252 + }, + { + "epoch": 0.33411276569559945, + "grad_norm": 0.7725064754486084, + "learning_rate": 0.0001498297895184285, + "loss": 0.1438, + "step": 14253 + }, + { + "epoch": 0.33413620727040444, + "grad_norm": 0.4366337060928345, + "learning_rate": 0.00014982340288104267, + "loss": 0.1405, + "step": 14254 + }, + { + "epoch": 0.33415964884520943, + "grad_norm": 0.4805119037628174, + "learning_rate": 0.00014981701597331363, + "loss": 0.0777, + "step": 14255 + }, + { + "epoch": 0.33418309042001443, + "grad_norm": 0.5824275612831116, + "learning_rate": 0.00014981062879527608, + "loss": 0.0907, + "step": 14256 + }, + { + "epoch": 0.3342065319948194, + "grad_norm": 0.12195051461458206, + "learning_rate": 0.00014980424134696468, + "loss": 0.0152, + "step": 14257 + }, + { + "epoch": 0.3342299735696244, + "grad_norm": 0.17439478635787964, + "learning_rate": 0.000149797853628414, + "loss": 0.0415, + "step": 14258 + }, + { + "epoch": 0.3342534151444294, + "grad_norm": 0.5716749429702759, + "learning_rate": 0.00014979146563965883, + "loss": 0.1806, + "step": 14259 + }, + { + "epoch": 0.3342768567192344, + "grad_norm": 0.12305747717618942, + "learning_rate": 0.00014978507738073378, + "loss": 0.0214, + "step": 14260 + }, + { + "epoch": 0.3343002982940394, + "grad_norm": 0.5335317254066467, + "learning_rate": 0.00014977868885167344, + "loss": 0.1038, + "step": 14261 + }, + { + "epoch": 0.3343237398688444, + "grad_norm": 0.482656866312027, + "learning_rate": 0.00014977230005251257, + "loss": 0.106, + "step": 14262 + }, + { + "epoch": 0.3343471814436494, + "grad_norm": 0.8953620195388794, + "learning_rate": 0.0001497659109832858, + "loss": 0.1146, + "step": 14263 + }, + { + "epoch": 0.3343706230184544, + "grad_norm": 0.3170217275619507, + "learning_rate": 0.0001497595216440278, + "loss": 0.0937, + "step": 14264 + }, + { + "epoch": 0.3343940645932594, + "grad_norm": 0.25274011492729187, + "learning_rate": 0.00014975313203477323, + "loss": 0.0422, + "step": 14265 + }, + { + "epoch": 0.33441750616806437, + "grad_norm": 0.3163764476776123, + "learning_rate": 0.00014974674215555676, + "loss": 0.4703, + "step": 14266 + }, + { + "epoch": 0.33444094774286937, + "grad_norm": 0.5637715458869934, + "learning_rate": 0.00014974035200641307, + "loss": 0.1478, + "step": 14267 + }, + { + "epoch": 0.33446438931767436, + "grad_norm": 0.566281795501709, + "learning_rate": 0.00014973396158737682, + "loss": 0.6197, + "step": 14268 + }, + { + "epoch": 0.33448783089247935, + "grad_norm": 0.5565474033355713, + "learning_rate": 0.00014972757089848273, + "loss": 0.1461, + "step": 14269 + }, + { + "epoch": 0.33451127246728435, + "grad_norm": 0.2181837111711502, + "learning_rate": 0.0001497211799397654, + "loss": 0.0571, + "step": 14270 + }, + { + "epoch": 0.33453471404208934, + "grad_norm": 0.5907274484634399, + "learning_rate": 0.0001497147887112596, + "loss": 0.1161, + "step": 14271 + }, + { + "epoch": 0.33455815561689434, + "grad_norm": 0.24586911499500275, + "learning_rate": 0.00014970839721299993, + "loss": 0.0408, + "step": 14272 + }, + { + "epoch": 0.33458159719169933, + "grad_norm": 0.7044366002082825, + "learning_rate": 0.0001497020054450211, + "loss": 0.1347, + "step": 14273 + }, + { + "epoch": 0.3346050387665043, + "grad_norm": 0.5915036201477051, + "learning_rate": 0.00014969561340735777, + "loss": 0.1508, + "step": 14274 + }, + { + "epoch": 0.3346284803413093, + "grad_norm": 0.10729502141475677, + "learning_rate": 0.00014968922110004468, + "loss": 0.0129, + "step": 14275 + }, + { + "epoch": 0.3346519219161143, + "grad_norm": 0.6422460079193115, + "learning_rate": 0.00014968282852311645, + "loss": 0.1466, + "step": 14276 + }, + { + "epoch": 0.3346753634909193, + "grad_norm": 0.6110458374023438, + "learning_rate": 0.00014967643567660784, + "loss": 0.1136, + "step": 14277 + }, + { + "epoch": 0.3346988050657243, + "grad_norm": 0.4868372678756714, + "learning_rate": 0.00014967004256055347, + "loss": 0.0827, + "step": 14278 + }, + { + "epoch": 0.3347222466405293, + "grad_norm": 0.35101696848869324, + "learning_rate": 0.00014966364917498804, + "loss": 0.0444, + "step": 14279 + }, + { + "epoch": 0.3347456882153343, + "grad_norm": 0.4084185063838959, + "learning_rate": 0.00014965725551994625, + "loss": 0.0552, + "step": 14280 + }, + { + "epoch": 0.3347691297901393, + "grad_norm": 0.23082371056079865, + "learning_rate": 0.0001496508615954628, + "loss": 0.0463, + "step": 14281 + }, + { + "epoch": 0.3347925713649443, + "grad_norm": 0.5910502076148987, + "learning_rate": 0.00014964446740157238, + "loss": 0.1643, + "step": 14282 + }, + { + "epoch": 0.3348160129397493, + "grad_norm": 0.3378433883190155, + "learning_rate": 0.0001496380729383097, + "loss": 0.0436, + "step": 14283 + }, + { + "epoch": 0.33483945451455427, + "grad_norm": 0.1664891242980957, + "learning_rate": 0.0001496316782057094, + "loss": 0.016, + "step": 14284 + }, + { + "epoch": 0.33486289608935926, + "grad_norm": 0.2526436150074005, + "learning_rate": 0.00014962528320380626, + "loss": 0.0625, + "step": 14285 + }, + { + "epoch": 0.33488633766416426, + "grad_norm": 0.9622742533683777, + "learning_rate": 0.00014961888793263493, + "loss": 0.1407, + "step": 14286 + }, + { + "epoch": 0.33490977923896925, + "grad_norm": 0.7818254828453064, + "learning_rate": 0.0001496124923922301, + "loss": 0.2295, + "step": 14287 + }, + { + "epoch": 0.33493322081377425, + "grad_norm": 0.4023752808570862, + "learning_rate": 0.00014960609658262652, + "loss": 0.0606, + "step": 14288 + }, + { + "epoch": 0.33495666238857924, + "grad_norm": 0.5054852962493896, + "learning_rate": 0.00014959970050385885, + "loss": 0.0741, + "step": 14289 + }, + { + "epoch": 0.33498010396338423, + "grad_norm": 0.16521716117858887, + "learning_rate": 0.0001495933041559618, + "loss": 0.0256, + "step": 14290 + }, + { + "epoch": 0.33500354553818923, + "grad_norm": 0.31245705485343933, + "learning_rate": 0.0001495869075389701, + "loss": 0.0896, + "step": 14291 + }, + { + "epoch": 0.3350269871129942, + "grad_norm": 0.42522376775741577, + "learning_rate": 0.00014958051065291846, + "loss": 0.1505, + "step": 14292 + }, + { + "epoch": 0.3350504286877993, + "grad_norm": 0.5816555619239807, + "learning_rate": 0.00014957411349784155, + "loss": 0.1132, + "step": 14293 + }, + { + "epoch": 0.33507387026260427, + "grad_norm": 0.6736772060394287, + "learning_rate": 0.00014956771607377411, + "loss": 0.1439, + "step": 14294 + }, + { + "epoch": 0.33509731183740926, + "grad_norm": 0.4071950614452362, + "learning_rate": 0.00014956131838075084, + "loss": 0.0933, + "step": 14295 + }, + { + "epoch": 0.33512075341221426, + "grad_norm": 0.6739944815635681, + "learning_rate": 0.00014955492041880648, + "loss": 0.1181, + "step": 14296 + }, + { + "epoch": 0.33514419498701925, + "grad_norm": 0.5112942457199097, + "learning_rate": 0.00014954852218797572, + "loss": 0.1141, + "step": 14297 + }, + { + "epoch": 0.33516763656182424, + "grad_norm": 0.33201560378074646, + "learning_rate": 0.00014954212368829332, + "loss": 0.0185, + "step": 14298 + }, + { + "epoch": 0.33519107813662924, + "grad_norm": 0.5795794129371643, + "learning_rate": 0.0001495357249197939, + "loss": 0.1193, + "step": 14299 + }, + { + "epoch": 0.33521451971143423, + "grad_norm": 0.7152203917503357, + "learning_rate": 0.0001495293258825123, + "loss": 0.1492, + "step": 14300 + }, + { + "epoch": 0.3352379612862392, + "grad_norm": 0.13200290501117706, + "learning_rate": 0.00014952292657648315, + "loss": 0.0321, + "step": 14301 + }, + { + "epoch": 0.3352614028610442, + "grad_norm": 0.24647338688373566, + "learning_rate": 0.00014951652700174125, + "loss": 0.0524, + "step": 14302 + }, + { + "epoch": 0.3352848444358492, + "grad_norm": 0.23573452234268188, + "learning_rate": 0.00014951012715832122, + "loss": 0.0621, + "step": 14303 + }, + { + "epoch": 0.3353082860106542, + "grad_norm": 0.17521105706691742, + "learning_rate": 0.0001495037270462579, + "loss": 0.0385, + "step": 14304 + }, + { + "epoch": 0.3353317275854592, + "grad_norm": 0.3659318685531616, + "learning_rate": 0.00014949732666558595, + "loss": 0.1151, + "step": 14305 + }, + { + "epoch": 0.3353551691602642, + "grad_norm": 0.8185169696807861, + "learning_rate": 0.00014949092601634009, + "loss": 0.3134, + "step": 14306 + }, + { + "epoch": 0.3353786107350692, + "grad_norm": 0.5369742512702942, + "learning_rate": 0.00014948452509855507, + "loss": 0.1594, + "step": 14307 + }, + { + "epoch": 0.3354020523098742, + "grad_norm": 0.27376553416252136, + "learning_rate": 0.00014947812391226568, + "loss": 0.0931, + "step": 14308 + }, + { + "epoch": 0.3354254938846792, + "grad_norm": 0.5435239672660828, + "learning_rate": 0.00014947172245750656, + "loss": 0.107, + "step": 14309 + }, + { + "epoch": 0.3354489354594842, + "grad_norm": 0.14728844165802002, + "learning_rate": 0.0001494653207343125, + "loss": 0.0348, + "step": 14310 + }, + { + "epoch": 0.33547237703428917, + "grad_norm": 0.36926084756851196, + "learning_rate": 0.0001494589187427182, + "loss": 0.327, + "step": 14311 + }, + { + "epoch": 0.33549581860909417, + "grad_norm": 0.26402226090431213, + "learning_rate": 0.00014945251648275842, + "loss": 0.051, + "step": 14312 + }, + { + "epoch": 0.33551926018389916, + "grad_norm": 0.5360609889030457, + "learning_rate": 0.0001494461139544679, + "loss": 0.117, + "step": 14313 + }, + { + "epoch": 0.33554270175870415, + "grad_norm": 0.4670916795730591, + "learning_rate": 0.0001494397111578814, + "loss": 0.1288, + "step": 14314 + }, + { + "epoch": 0.33556614333350915, + "grad_norm": 0.45395150780677795, + "learning_rate": 0.0001494333080930336, + "loss": 0.1593, + "step": 14315 + }, + { + "epoch": 0.33558958490831414, + "grad_norm": 0.488434761762619, + "learning_rate": 0.00014942690475995928, + "loss": 0.1262, + "step": 14316 + }, + { + "epoch": 0.33561302648311914, + "grad_norm": 0.2624407410621643, + "learning_rate": 0.0001494205011586932, + "loss": 0.071, + "step": 14317 + }, + { + "epoch": 0.33563646805792413, + "grad_norm": 0.13492386043071747, + "learning_rate": 0.0001494140972892701, + "loss": 0.0297, + "step": 14318 + }, + { + "epoch": 0.3356599096327291, + "grad_norm": 0.575899064540863, + "learning_rate": 0.00014940769315172469, + "loss": 0.1315, + "step": 14319 + }, + { + "epoch": 0.3356833512075341, + "grad_norm": 0.49321848154067993, + "learning_rate": 0.00014940128874609178, + "loss": 0.1488, + "step": 14320 + }, + { + "epoch": 0.3357067927823391, + "grad_norm": 0.6092843413352966, + "learning_rate": 0.00014939488407240606, + "loss": 0.1851, + "step": 14321 + }, + { + "epoch": 0.3357302343571441, + "grad_norm": 0.3352751135826111, + "learning_rate": 0.00014938847913070234, + "loss": 0.0863, + "step": 14322 + }, + { + "epoch": 0.3357536759319491, + "grad_norm": 0.4107615351676941, + "learning_rate": 0.00014938207392101532, + "loss": 0.0824, + "step": 14323 + }, + { + "epoch": 0.3357771175067541, + "grad_norm": 0.18946604430675507, + "learning_rate": 0.0001493756684433798, + "loss": 0.0452, + "step": 14324 + }, + { + "epoch": 0.3358005590815591, + "grad_norm": 0.4873697757720947, + "learning_rate": 0.0001493692626978305, + "loss": 0.103, + "step": 14325 + }, + { + "epoch": 0.3358240006563641, + "grad_norm": 0.3799408972263336, + "learning_rate": 0.00014936285668440218, + "loss": 0.07, + "step": 14326 + }, + { + "epoch": 0.3358474422311691, + "grad_norm": 0.3624351918697357, + "learning_rate": 0.00014935645040312965, + "loss": 0.136, + "step": 14327 + }, + { + "epoch": 0.3358708838059741, + "grad_norm": 0.7537160515785217, + "learning_rate": 0.0001493500438540476, + "loss": 0.6314, + "step": 14328 + }, + { + "epoch": 0.33589432538077907, + "grad_norm": 0.7250707149505615, + "learning_rate": 0.00014934363703719082, + "loss": 0.123, + "step": 14329 + }, + { + "epoch": 0.33591776695558406, + "grad_norm": 0.49683907628059387, + "learning_rate": 0.00014933722995259409, + "loss": 0.0946, + "step": 14330 + }, + { + "epoch": 0.33594120853038906, + "grad_norm": 0.6227467656135559, + "learning_rate": 0.00014933082260029217, + "loss": 0.8247, + "step": 14331 + }, + { + "epoch": 0.33596465010519405, + "grad_norm": 0.648182213306427, + "learning_rate": 0.00014932441498031982, + "loss": 0.1349, + "step": 14332 + }, + { + "epoch": 0.33598809167999905, + "grad_norm": 0.12627087533473969, + "learning_rate": 0.0001493180070927118, + "loss": 0.0225, + "step": 14333 + }, + { + "epoch": 0.33601153325480404, + "grad_norm": 0.3600304424762726, + "learning_rate": 0.00014931159893750287, + "loss": 0.0606, + "step": 14334 + }, + { + "epoch": 0.33603497482960903, + "grad_norm": 0.7362254858016968, + "learning_rate": 0.00014930519051472782, + "loss": 0.0684, + "step": 14335 + }, + { + "epoch": 0.33605841640441403, + "grad_norm": 0.40960708260536194, + "learning_rate": 0.00014929878182442144, + "loss": 0.1126, + "step": 14336 + }, + { + "epoch": 0.336081857979219, + "grad_norm": 0.33630451560020447, + "learning_rate": 0.0001492923728666185, + "loss": 0.0803, + "step": 14337 + }, + { + "epoch": 0.336105299554024, + "grad_norm": 0.5159075856208801, + "learning_rate": 0.00014928596364135373, + "loss": 0.1464, + "step": 14338 + }, + { + "epoch": 0.336128741128829, + "grad_norm": 0.35051363706588745, + "learning_rate": 0.00014927955414866195, + "loss": 0.0894, + "step": 14339 + }, + { + "epoch": 0.336152182703634, + "grad_norm": 0.44663017988204956, + "learning_rate": 0.0001492731443885779, + "loss": 0.0854, + "step": 14340 + }, + { + "epoch": 0.336175624278439, + "grad_norm": 0.595766007900238, + "learning_rate": 0.00014926673436113638, + "loss": 0.1777, + "step": 14341 + }, + { + "epoch": 0.336199065853244, + "grad_norm": 0.6702379584312439, + "learning_rate": 0.0001492603240663722, + "loss": 0.607, + "step": 14342 + }, + { + "epoch": 0.336222507428049, + "grad_norm": 0.18881237506866455, + "learning_rate": 0.0001492539135043201, + "loss": 0.0423, + "step": 14343 + }, + { + "epoch": 0.33624594900285404, + "grad_norm": 0.4953882098197937, + "learning_rate": 0.0001492475026750149, + "loss": 0.161, + "step": 14344 + }, + { + "epoch": 0.33626939057765903, + "grad_norm": 0.5719220042228699, + "learning_rate": 0.0001492410915784914, + "loss": 0.7203, + "step": 14345 + }, + { + "epoch": 0.336292832152464, + "grad_norm": 0.7694931626319885, + "learning_rate": 0.00014923468021478428, + "loss": 0.101, + "step": 14346 + }, + { + "epoch": 0.336316273727269, + "grad_norm": 0.23967139422893524, + "learning_rate": 0.0001492282685839284, + "loss": 0.0458, + "step": 14347 + }, + { + "epoch": 0.336339715302074, + "grad_norm": 0.3254272937774658, + "learning_rate": 0.0001492218566859586, + "loss": 0.0771, + "step": 14348 + }, + { + "epoch": 0.336363156876879, + "grad_norm": 0.2599112093448639, + "learning_rate": 0.0001492154445209096, + "loss": 0.0378, + "step": 14349 + }, + { + "epoch": 0.336386598451684, + "grad_norm": 0.5368724465370178, + "learning_rate": 0.0001492090320888162, + "loss": 0.0931, + "step": 14350 + }, + { + "epoch": 0.336410040026489, + "grad_norm": 0.8815504908561707, + "learning_rate": 0.00014920261938971324, + "loss": 0.1894, + "step": 14351 + }, + { + "epoch": 0.336433481601294, + "grad_norm": 0.5723749995231628, + "learning_rate": 0.00014919620642363547, + "loss": 0.8131, + "step": 14352 + }, + { + "epoch": 0.336456923176099, + "grad_norm": 0.6462680697441101, + "learning_rate": 0.0001491897931906177, + "loss": 0.0567, + "step": 14353 + }, + { + "epoch": 0.336480364750904, + "grad_norm": 0.16904376447200775, + "learning_rate": 0.0001491833796906947, + "loss": 0.0449, + "step": 14354 + }, + { + "epoch": 0.336503806325709, + "grad_norm": 0.2918047308921814, + "learning_rate": 0.00014917696592390135, + "loss": 0.0659, + "step": 14355 + }, + { + "epoch": 0.33652724790051397, + "grad_norm": 0.5766916871070862, + "learning_rate": 0.0001491705518902724, + "loss": 0.094, + "step": 14356 + }, + { + "epoch": 0.33655068947531896, + "grad_norm": 0.29869508743286133, + "learning_rate": 0.0001491641375898426, + "loss": 0.057, + "step": 14357 + }, + { + "epoch": 0.33657413105012396, + "grad_norm": 0.4714539647102356, + "learning_rate": 0.00014915772302264685, + "loss": 0.0859, + "step": 14358 + }, + { + "epoch": 0.33659757262492895, + "grad_norm": 0.41270625591278076, + "learning_rate": 0.0001491513081887199, + "loss": 0.0599, + "step": 14359 + }, + { + "epoch": 0.33662101419973395, + "grad_norm": 0.37881341576576233, + "learning_rate": 0.00014914489308809654, + "loss": 0.0626, + "step": 14360 + }, + { + "epoch": 0.33664445577453894, + "grad_norm": 0.836367130279541, + "learning_rate": 0.00014913847772081166, + "loss": 0.0949, + "step": 14361 + }, + { + "epoch": 0.33666789734934394, + "grad_norm": 0.12952779233455658, + "learning_rate": 0.0001491320620869, + "loss": 0.0223, + "step": 14362 + }, + { + "epoch": 0.33669133892414893, + "grad_norm": 0.7695896625518799, + "learning_rate": 0.00014912564618639636, + "loss": 0.1713, + "step": 14363 + }, + { + "epoch": 0.3367147804989539, + "grad_norm": 0.29964107275009155, + "learning_rate": 0.00014911923001933563, + "loss": 0.0329, + "step": 14364 + }, + { + "epoch": 0.3367382220737589, + "grad_norm": 1.0424180030822754, + "learning_rate": 0.00014911281358575255, + "loss": 0.2952, + "step": 14365 + }, + { + "epoch": 0.3367616636485639, + "grad_norm": 0.10809537023305893, + "learning_rate": 0.00014910639688568196, + "loss": 0.0137, + "step": 14366 + }, + { + "epoch": 0.3367851052233689, + "grad_norm": 1.5020099878311157, + "learning_rate": 0.0001490999799191587, + "loss": 0.149, + "step": 14367 + }, + { + "epoch": 0.3368085467981739, + "grad_norm": 0.3253007233142853, + "learning_rate": 0.00014909356268621754, + "loss": 0.0549, + "step": 14368 + }, + { + "epoch": 0.3368319883729789, + "grad_norm": 0.10720523446798325, + "learning_rate": 0.00014908714518689336, + "loss": 0.0075, + "step": 14369 + }, + { + "epoch": 0.3368554299477839, + "grad_norm": 0.43159428238868713, + "learning_rate": 0.00014908072742122092, + "loss": 0.0714, + "step": 14370 + }, + { + "epoch": 0.3368788715225889, + "grad_norm": 0.15344412624835968, + "learning_rate": 0.00014907430938923511, + "loss": 0.0301, + "step": 14371 + }, + { + "epoch": 0.3369023130973939, + "grad_norm": 0.19046248495578766, + "learning_rate": 0.00014906789109097067, + "loss": 0.0351, + "step": 14372 + }, + { + "epoch": 0.3369257546721989, + "grad_norm": 0.4491516947746277, + "learning_rate": 0.00014906147252646249, + "loss": 0.1013, + "step": 14373 + }, + { + "epoch": 0.33694919624700387, + "grad_norm": 0.06017376109957695, + "learning_rate": 0.0001490550536957454, + "loss": 0.008, + "step": 14374 + }, + { + "epoch": 0.33697263782180886, + "grad_norm": 0.6657983064651489, + "learning_rate": 0.00014904863459885422, + "loss": 0.5584, + "step": 14375 + }, + { + "epoch": 0.33699607939661386, + "grad_norm": 0.15221144258975983, + "learning_rate": 0.00014904221523582374, + "loss": 0.0324, + "step": 14376 + }, + { + "epoch": 0.33701952097141885, + "grad_norm": 0.4025869071483612, + "learning_rate": 0.00014903579560668885, + "loss": 0.0842, + "step": 14377 + }, + { + "epoch": 0.33704296254622385, + "grad_norm": 0.39444878697395325, + "learning_rate": 0.0001490293757114843, + "loss": 0.0826, + "step": 14378 + }, + { + "epoch": 0.33706640412102884, + "grad_norm": 0.4728447496891022, + "learning_rate": 0.00014902295555024498, + "loss": 0.1071, + "step": 14379 + }, + { + "epoch": 0.33708984569583383, + "grad_norm": 0.2960957884788513, + "learning_rate": 0.00014901653512300577, + "loss": 0.0458, + "step": 14380 + }, + { + "epoch": 0.33711328727063883, + "grad_norm": 0.5591578483581543, + "learning_rate": 0.00014901011442980147, + "loss": 0.079, + "step": 14381 + }, + { + "epoch": 0.3371367288454438, + "grad_norm": 0.4833153784275055, + "learning_rate": 0.00014900369347066686, + "loss": 0.1243, + "step": 14382 + }, + { + "epoch": 0.3371601704202488, + "grad_norm": 0.42259442806243896, + "learning_rate": 0.0001489972722456369, + "loss": 0.0812, + "step": 14383 + }, + { + "epoch": 0.3371836119950538, + "grad_norm": 0.2242971807718277, + "learning_rate": 0.0001489908507547463, + "loss": 0.0245, + "step": 14384 + }, + { + "epoch": 0.3372070535698588, + "grad_norm": 0.4366845190525055, + "learning_rate": 0.00014898442899802998, + "loss": 0.0793, + "step": 14385 + }, + { + "epoch": 0.3372304951446638, + "grad_norm": 0.7608450055122375, + "learning_rate": 0.00014897800697552278, + "loss": 0.1115, + "step": 14386 + }, + { + "epoch": 0.3372539367194688, + "grad_norm": 0.4275430142879486, + "learning_rate": 0.00014897158468725957, + "loss": 0.1222, + "step": 14387 + }, + { + "epoch": 0.3372773782942738, + "grad_norm": 0.21297118067741394, + "learning_rate": 0.0001489651621332751, + "loss": 0.044, + "step": 14388 + }, + { + "epoch": 0.3373008198690788, + "grad_norm": 0.6209856867790222, + "learning_rate": 0.00014895873931360432, + "loss": 0.0762, + "step": 14389 + }, + { + "epoch": 0.3373242614438838, + "grad_norm": 0.20524804294109344, + "learning_rate": 0.00014895231622828202, + "loss": 0.0508, + "step": 14390 + }, + { + "epoch": 0.33734770301868877, + "grad_norm": 0.2692919671535492, + "learning_rate": 0.0001489458928773431, + "loss": 0.0376, + "step": 14391 + }, + { + "epoch": 0.33737114459349377, + "grad_norm": 0.19566699862480164, + "learning_rate": 0.00014893946926082237, + "loss": 0.0551, + "step": 14392 + }, + { + "epoch": 0.33739458616829876, + "grad_norm": 0.39787325263023376, + "learning_rate": 0.00014893304537875476, + "loss": 0.0728, + "step": 14393 + }, + { + "epoch": 0.33741802774310375, + "grad_norm": 0.2110229730606079, + "learning_rate": 0.00014892662123117503, + "loss": 0.0442, + "step": 14394 + }, + { + "epoch": 0.3374414693179088, + "grad_norm": 0.1843656599521637, + "learning_rate": 0.00014892019681811807, + "loss": 0.037, + "step": 14395 + }, + { + "epoch": 0.3374649108927138, + "grad_norm": 0.49671849608421326, + "learning_rate": 0.0001489137721396187, + "loss": 0.113, + "step": 14396 + }, + { + "epoch": 0.3374883524675188, + "grad_norm": 0.48967599868774414, + "learning_rate": 0.00014890734719571192, + "loss": 0.1153, + "step": 14397 + }, + { + "epoch": 0.3375117940423238, + "grad_norm": 0.5258532166481018, + "learning_rate": 0.00014890092198643245, + "loss": 0.0825, + "step": 14398 + }, + { + "epoch": 0.3375352356171288, + "grad_norm": 0.4013495445251465, + "learning_rate": 0.00014889449651181522, + "loss": 0.0782, + "step": 14399 + }, + { + "epoch": 0.3375586771919338, + "grad_norm": 0.22995273768901825, + "learning_rate": 0.00014888807077189506, + "loss": 0.0342, + "step": 14400 + }, + { + "epoch": 0.33758211876673877, + "grad_norm": 0.4061196744441986, + "learning_rate": 0.00014888164476670686, + "loss": 0.0814, + "step": 14401 + }, + { + "epoch": 0.33760556034154376, + "grad_norm": 0.46758607029914856, + "learning_rate": 0.00014887521849628546, + "loss": 0.1195, + "step": 14402 + }, + { + "epoch": 0.33762900191634876, + "grad_norm": 0.8430513143539429, + "learning_rate": 0.0001488687919606658, + "loss": 0.0609, + "step": 14403 + }, + { + "epoch": 0.33765244349115375, + "grad_norm": 0.20716401934623718, + "learning_rate": 0.00014886236515988265, + "loss": 0.0311, + "step": 14404 + }, + { + "epoch": 0.33767588506595875, + "grad_norm": 0.47790369391441345, + "learning_rate": 0.00014885593809397098, + "loss": 0.0727, + "step": 14405 + }, + { + "epoch": 0.33769932664076374, + "grad_norm": 0.3434901237487793, + "learning_rate": 0.00014884951076296558, + "loss": 0.0804, + "step": 14406 + }, + { + "epoch": 0.33772276821556874, + "grad_norm": 0.4851107597351074, + "learning_rate": 0.00014884308316690138, + "loss": 0.0961, + "step": 14407 + }, + { + "epoch": 0.33774620979037373, + "grad_norm": 0.38474366068840027, + "learning_rate": 0.0001488366553058132, + "loss": 0.0696, + "step": 14408 + }, + { + "epoch": 0.3377696513651787, + "grad_norm": 1.0100898742675781, + "learning_rate": 0.000148830227179736, + "loss": 0.0505, + "step": 14409 + }, + { + "epoch": 0.3377930929399837, + "grad_norm": 0.4414658546447754, + "learning_rate": 0.00014882379878870464, + "loss": 0.5715, + "step": 14410 + }, + { + "epoch": 0.3378165345147887, + "grad_norm": 0.4918535053730011, + "learning_rate": 0.0001488173701327539, + "loss": 0.0802, + "step": 14411 + }, + { + "epoch": 0.3378399760895937, + "grad_norm": 0.373759388923645, + "learning_rate": 0.0001488109412119188, + "loss": 0.1517, + "step": 14412 + }, + { + "epoch": 0.3378634176643987, + "grad_norm": 0.42484724521636963, + "learning_rate": 0.00014880451202623412, + "loss": 0.1647, + "step": 14413 + }, + { + "epoch": 0.3378868592392037, + "grad_norm": 0.4691443145275116, + "learning_rate": 0.00014879808257573478, + "loss": 0.157, + "step": 14414 + }, + { + "epoch": 0.3379103008140087, + "grad_norm": 0.2201077789068222, + "learning_rate": 0.0001487916528604557, + "loss": 0.0611, + "step": 14415 + }, + { + "epoch": 0.3379337423888137, + "grad_norm": 0.5813938975334167, + "learning_rate": 0.00014878522288043174, + "loss": 0.1514, + "step": 14416 + }, + { + "epoch": 0.3379571839636187, + "grad_norm": 0.20095938444137573, + "learning_rate": 0.00014877879263569776, + "loss": 0.0484, + "step": 14417 + }, + { + "epoch": 0.3379806255384237, + "grad_norm": 0.20856846868991852, + "learning_rate": 0.0001487723621262887, + "loss": 0.0256, + "step": 14418 + }, + { + "epoch": 0.33800406711322867, + "grad_norm": 0.25849613547325134, + "learning_rate": 0.00014876593135223945, + "loss": 0.0631, + "step": 14419 + }, + { + "epoch": 0.33802750868803366, + "grad_norm": 0.6298059821128845, + "learning_rate": 0.00014875950031358485, + "loss": 0.1186, + "step": 14420 + }, + { + "epoch": 0.33805095026283866, + "grad_norm": 0.16218659281730652, + "learning_rate": 0.00014875306901035985, + "loss": 0.0454, + "step": 14421 + }, + { + "epoch": 0.33807439183764365, + "grad_norm": 0.14937467873096466, + "learning_rate": 0.0001487466374425993, + "loss": 0.0278, + "step": 14422 + }, + { + "epoch": 0.33809783341244865, + "grad_norm": 0.6607574820518494, + "learning_rate": 0.00014874020561033815, + "loss": 0.0834, + "step": 14423 + }, + { + "epoch": 0.33812127498725364, + "grad_norm": 0.4814370572566986, + "learning_rate": 0.00014873377351361128, + "loss": 0.0787, + "step": 14424 + }, + { + "epoch": 0.33814471656205863, + "grad_norm": 0.7015681266784668, + "learning_rate": 0.0001487273411524536, + "loss": 0.1825, + "step": 14425 + }, + { + "epoch": 0.33816815813686363, + "grad_norm": 0.6849742531776428, + "learning_rate": 0.00014872090852689995, + "loss": 0.1718, + "step": 14426 + }, + { + "epoch": 0.3381915997116686, + "grad_norm": 0.509787380695343, + "learning_rate": 0.0001487144756369853, + "loss": 0.1204, + "step": 14427 + }, + { + "epoch": 0.3382150412864736, + "grad_norm": 0.465219646692276, + "learning_rate": 0.00014870804248274454, + "loss": 0.1099, + "step": 14428 + }, + { + "epoch": 0.3382384828612786, + "grad_norm": 0.4332326352596283, + "learning_rate": 0.00014870160906421258, + "loss": 0.0636, + "step": 14429 + }, + { + "epoch": 0.3382619244360836, + "grad_norm": 0.4208941161632538, + "learning_rate": 0.0001486951753814243, + "loss": 0.0933, + "step": 14430 + }, + { + "epoch": 0.3382853660108886, + "grad_norm": 0.48447003960609436, + "learning_rate": 0.00014868874143441463, + "loss": 0.0686, + "step": 14431 + }, + { + "epoch": 0.3383088075856936, + "grad_norm": 0.3971906900405884, + "learning_rate": 0.0001486823072232185, + "loss": 0.0726, + "step": 14432 + }, + { + "epoch": 0.3383322491604986, + "grad_norm": 0.7349472045898438, + "learning_rate": 0.00014867587274787077, + "loss": 0.0698, + "step": 14433 + }, + { + "epoch": 0.3383556907353036, + "grad_norm": 0.4231961667537689, + "learning_rate": 0.00014866943800840644, + "loss": 0.0757, + "step": 14434 + }, + { + "epoch": 0.3383791323101086, + "grad_norm": 0.6702290177345276, + "learning_rate": 0.0001486630030048603, + "loss": 0.118, + "step": 14435 + }, + { + "epoch": 0.33840257388491357, + "grad_norm": 0.5575558543205261, + "learning_rate": 0.00014865656773726738, + "loss": 0.6842, + "step": 14436 + }, + { + "epoch": 0.33842601545971857, + "grad_norm": 0.38484805822372437, + "learning_rate": 0.00014865013220566255, + "loss": 0.0758, + "step": 14437 + }, + { + "epoch": 0.33844945703452356, + "grad_norm": 0.14820916950702667, + "learning_rate": 0.00014864369641008076, + "loss": 0.0317, + "step": 14438 + }, + { + "epoch": 0.33847289860932855, + "grad_norm": 0.37372252345085144, + "learning_rate": 0.00014863726035055686, + "loss": 0.1194, + "step": 14439 + }, + { + "epoch": 0.33849634018413355, + "grad_norm": 0.8519591093063354, + "learning_rate": 0.00014863082402712582, + "loss": 0.5689, + "step": 14440 + }, + { + "epoch": 0.33851978175893854, + "grad_norm": 0.5112820863723755, + "learning_rate": 0.0001486243874398226, + "loss": 0.1211, + "step": 14441 + }, + { + "epoch": 0.33854322333374354, + "grad_norm": 0.19006697833538055, + "learning_rate": 0.00014861795058868207, + "loss": 0.0528, + "step": 14442 + }, + { + "epoch": 0.33856666490854853, + "grad_norm": 0.36114397644996643, + "learning_rate": 0.00014861151347373919, + "loss": 0.0606, + "step": 14443 + }, + { + "epoch": 0.3385901064833535, + "grad_norm": 0.8766967058181763, + "learning_rate": 0.00014860507609502883, + "loss": 0.1641, + "step": 14444 + }, + { + "epoch": 0.3386135480581585, + "grad_norm": 0.579889178276062, + "learning_rate": 0.00014859863845258597, + "loss": 0.0935, + "step": 14445 + }, + { + "epoch": 0.33863698963296357, + "grad_norm": 0.4216156005859375, + "learning_rate": 0.00014859220054644554, + "loss": 0.1369, + "step": 14446 + }, + { + "epoch": 0.33866043120776856, + "grad_norm": 0.28833338618278503, + "learning_rate": 0.0001485857623766425, + "loss": 0.0693, + "step": 14447 + }, + { + "epoch": 0.33868387278257356, + "grad_norm": 0.8139455914497375, + "learning_rate": 0.0001485793239432117, + "loss": 0.1672, + "step": 14448 + }, + { + "epoch": 0.33870731435737855, + "grad_norm": 0.15462274849414825, + "learning_rate": 0.00014857288524618812, + "loss": 0.0116, + "step": 14449 + }, + { + "epoch": 0.33873075593218355, + "grad_norm": 0.7648093700408936, + "learning_rate": 0.0001485664462856067, + "loss": 0.1064, + "step": 14450 + }, + { + "epoch": 0.33875419750698854, + "grad_norm": 0.41319435834884644, + "learning_rate": 0.00014856000706150243, + "loss": 0.0469, + "step": 14451 + }, + { + "epoch": 0.33877763908179354, + "grad_norm": 0.9632378220558167, + "learning_rate": 0.00014855356757391012, + "loss": 0.2409, + "step": 14452 + }, + { + "epoch": 0.33880108065659853, + "grad_norm": 0.6441839933395386, + "learning_rate": 0.00014854712782286483, + "loss": 0.1054, + "step": 14453 + }, + { + "epoch": 0.3388245222314035, + "grad_norm": 0.3749234974384308, + "learning_rate": 0.00014854068780840147, + "loss": 0.0789, + "step": 14454 + }, + { + "epoch": 0.3388479638062085, + "grad_norm": 0.5011959671974182, + "learning_rate": 0.00014853424753055494, + "loss": 0.1164, + "step": 14455 + }, + { + "epoch": 0.3388714053810135, + "grad_norm": 0.20332121849060059, + "learning_rate": 0.00014852780698936026, + "loss": 0.0432, + "step": 14456 + }, + { + "epoch": 0.3388948469558185, + "grad_norm": 0.6088206768035889, + "learning_rate": 0.0001485213661848523, + "loss": 0.0908, + "step": 14457 + }, + { + "epoch": 0.3389182885306235, + "grad_norm": 0.7884685397148132, + "learning_rate": 0.00014851492511706605, + "loss": 0.1118, + "step": 14458 + }, + { + "epoch": 0.3389417301054285, + "grad_norm": 0.7094026803970337, + "learning_rate": 0.00014850848378603645, + "loss": 0.1211, + "step": 14459 + }, + { + "epoch": 0.3389651716802335, + "grad_norm": 0.2094200849533081, + "learning_rate": 0.00014850204219179846, + "loss": 0.0359, + "step": 14460 + }, + { + "epoch": 0.3389886132550385, + "grad_norm": 0.6198296546936035, + "learning_rate": 0.000148495600334387, + "loss": 0.137, + "step": 14461 + }, + { + "epoch": 0.3390120548298435, + "grad_norm": 0.23559610545635223, + "learning_rate": 0.0001484891582138371, + "loss": 0.0409, + "step": 14462 + }, + { + "epoch": 0.3390354964046485, + "grad_norm": 0.4499877393245697, + "learning_rate": 0.00014848271583018364, + "loss": 0.0927, + "step": 14463 + }, + { + "epoch": 0.33905893797945347, + "grad_norm": 0.5300232172012329, + "learning_rate": 0.0001484762731834616, + "loss": 0.1642, + "step": 14464 + }, + { + "epoch": 0.33908237955425846, + "grad_norm": 0.44631338119506836, + "learning_rate": 0.0001484698302737059, + "loss": 0.1152, + "step": 14465 + }, + { + "epoch": 0.33910582112906346, + "grad_norm": 0.40180736780166626, + "learning_rate": 0.0001484633871009516, + "loss": 0.0421, + "step": 14466 + }, + { + "epoch": 0.33912926270386845, + "grad_norm": 0.47243377566337585, + "learning_rate": 0.00014845694366523356, + "loss": 0.083, + "step": 14467 + }, + { + "epoch": 0.33915270427867344, + "grad_norm": 0.6343084573745728, + "learning_rate": 0.00014845049996658682, + "loss": 0.0797, + "step": 14468 + }, + { + "epoch": 0.33917614585347844, + "grad_norm": 0.7293429374694824, + "learning_rate": 0.00014844405600504624, + "loss": 0.1034, + "step": 14469 + }, + { + "epoch": 0.33919958742828343, + "grad_norm": 0.5642808079719543, + "learning_rate": 0.00014843761178064694, + "loss": 0.0834, + "step": 14470 + }, + { + "epoch": 0.3392230290030884, + "grad_norm": 0.7525211572647095, + "learning_rate": 0.00014843116729342372, + "loss": 0.1098, + "step": 14471 + }, + { + "epoch": 0.3392464705778934, + "grad_norm": 0.1448264867067337, + "learning_rate": 0.00014842472254341163, + "loss": 0.0144, + "step": 14472 + }, + { + "epoch": 0.3392699121526984, + "grad_norm": 0.45970186591148376, + "learning_rate": 0.00014841827753064567, + "loss": 0.1098, + "step": 14473 + }, + { + "epoch": 0.3392933537275034, + "grad_norm": 0.6090550422668457, + "learning_rate": 0.00014841183225516074, + "loss": 0.2639, + "step": 14474 + }, + { + "epoch": 0.3393167953023084, + "grad_norm": 0.1673402041196823, + "learning_rate": 0.0001484053867169919, + "loss": 0.0325, + "step": 14475 + }, + { + "epoch": 0.3393402368771134, + "grad_norm": 0.35233181715011597, + "learning_rate": 0.00014839894091617405, + "loss": 0.0616, + "step": 14476 + }, + { + "epoch": 0.3393636784519184, + "grad_norm": 0.20060382783412933, + "learning_rate": 0.00014839249485274213, + "loss": 0.026, + "step": 14477 + }, + { + "epoch": 0.3393871200267234, + "grad_norm": 0.20032480359077454, + "learning_rate": 0.00014838604852673122, + "loss": 0.0465, + "step": 14478 + }, + { + "epoch": 0.3394105616015284, + "grad_norm": 0.455503910779953, + "learning_rate": 0.00014837960193817625, + "loss": 0.0489, + "step": 14479 + }, + { + "epoch": 0.3394340031763334, + "grad_norm": 0.5221225619316101, + "learning_rate": 0.0001483731550871122, + "loss": 0.0758, + "step": 14480 + }, + { + "epoch": 0.33945744475113837, + "grad_norm": 0.16432812809944153, + "learning_rate": 0.00014836670797357403, + "loss": 0.0455, + "step": 14481 + }, + { + "epoch": 0.33948088632594337, + "grad_norm": 0.6791388392448425, + "learning_rate": 0.00014836026059759675, + "loss": 0.0852, + "step": 14482 + }, + { + "epoch": 0.33950432790074836, + "grad_norm": 0.18520264327526093, + "learning_rate": 0.00014835381295921538, + "loss": 0.037, + "step": 14483 + }, + { + "epoch": 0.33952776947555335, + "grad_norm": 0.2756800055503845, + "learning_rate": 0.0001483473650584648, + "loss": 0.0252, + "step": 14484 + }, + { + "epoch": 0.33955121105035835, + "grad_norm": 0.6006042957305908, + "learning_rate": 0.0001483409168953801, + "loss": 0.1327, + "step": 14485 + }, + { + "epoch": 0.33957465262516334, + "grad_norm": 0.5670390725135803, + "learning_rate": 0.00014833446846999621, + "loss": 0.0876, + "step": 14486 + }, + { + "epoch": 0.33959809419996834, + "grad_norm": 0.9108706116676331, + "learning_rate": 0.0001483280197823481, + "loss": 0.1768, + "step": 14487 + }, + { + "epoch": 0.33962153577477333, + "grad_norm": 0.957915723323822, + "learning_rate": 0.00014832157083247084, + "loss": 0.2662, + "step": 14488 + }, + { + "epoch": 0.3396449773495783, + "grad_norm": 0.3178956210613251, + "learning_rate": 0.00014831512162039937, + "loss": 0.0658, + "step": 14489 + }, + { + "epoch": 0.3396684189243833, + "grad_norm": 0.5371880531311035, + "learning_rate": 0.00014830867214616868, + "loss": 0.1009, + "step": 14490 + }, + { + "epoch": 0.3396918604991883, + "grad_norm": 0.16398201882839203, + "learning_rate": 0.0001483022224098138, + "loss": 0.0252, + "step": 14491 + }, + { + "epoch": 0.3397153020739933, + "grad_norm": 0.1357940137386322, + "learning_rate": 0.0001482957724113697, + "loss": 0.0373, + "step": 14492 + }, + { + "epoch": 0.3397387436487983, + "grad_norm": 0.3151426613330841, + "learning_rate": 0.00014828932215087134, + "loss": 0.0398, + "step": 14493 + }, + { + "epoch": 0.3397621852236033, + "grad_norm": 0.1073824092745781, + "learning_rate": 0.0001482828716283538, + "loss": 0.0216, + "step": 14494 + }, + { + "epoch": 0.3397856267984083, + "grad_norm": 0.6479809284210205, + "learning_rate": 0.00014827642084385204, + "loss": 0.2186, + "step": 14495 + }, + { + "epoch": 0.3398090683732133, + "grad_norm": 0.6423466801643372, + "learning_rate": 0.00014826996979740105, + "loss": 0.1461, + "step": 14496 + }, + { + "epoch": 0.33983250994801834, + "grad_norm": 0.3332687020301819, + "learning_rate": 0.00014826351848903584, + "loss": 0.1791, + "step": 14497 + }, + { + "epoch": 0.33985595152282333, + "grad_norm": 0.38143882155418396, + "learning_rate": 0.00014825706691879144, + "loss": 0.0827, + "step": 14498 + }, + { + "epoch": 0.3398793930976283, + "grad_norm": 0.7906175851821899, + "learning_rate": 0.00014825061508670282, + "loss": 0.2592, + "step": 14499 + }, + { + "epoch": 0.3399028346724333, + "grad_norm": 0.25235700607299805, + "learning_rate": 0.000148244162992805, + "loss": 0.0913, + "step": 14500 + }, + { + "epoch": 0.3399262762472383, + "grad_norm": 0.4223902225494385, + "learning_rate": 0.00014823771063713303, + "loss": 0.062, + "step": 14501 + }, + { + "epoch": 0.3399497178220433, + "grad_norm": 0.6623263359069824, + "learning_rate": 0.00014823125801972185, + "loss": 0.1671, + "step": 14502 + }, + { + "epoch": 0.3399731593968483, + "grad_norm": 0.5593002438545227, + "learning_rate": 0.00014822480514060655, + "loss": 0.0777, + "step": 14503 + }, + { + "epoch": 0.3399966009716533, + "grad_norm": 0.3138365149497986, + "learning_rate": 0.00014821835199982207, + "loss": 0.0626, + "step": 14504 + }, + { + "epoch": 0.3400200425464583, + "grad_norm": 0.6381233930587769, + "learning_rate": 0.00014821189859740344, + "loss": 0.1731, + "step": 14505 + }, + { + "epoch": 0.3400434841212633, + "grad_norm": 0.31866297125816345, + "learning_rate": 0.0001482054449333857, + "loss": 0.0322, + "step": 14506 + }, + { + "epoch": 0.3400669256960683, + "grad_norm": 0.3907194435596466, + "learning_rate": 0.0001481989910078039, + "loss": 0.1283, + "step": 14507 + }, + { + "epoch": 0.3400903672708733, + "grad_norm": 0.5035475492477417, + "learning_rate": 0.000148192536820693, + "loss": 0.5828, + "step": 14508 + }, + { + "epoch": 0.34011380884567827, + "grad_norm": 0.516897439956665, + "learning_rate": 0.000148186082372088, + "loss": 0.1021, + "step": 14509 + }, + { + "epoch": 0.34013725042048326, + "grad_norm": 0.39694637060165405, + "learning_rate": 0.00014817962766202402, + "loss": 0.1314, + "step": 14510 + }, + { + "epoch": 0.34016069199528826, + "grad_norm": 0.620603621006012, + "learning_rate": 0.00014817317269053603, + "loss": 0.1089, + "step": 14511 + }, + { + "epoch": 0.34018413357009325, + "grad_norm": 0.1728098839521408, + "learning_rate": 0.000148166717457659, + "loss": 0.034, + "step": 14512 + }, + { + "epoch": 0.34020757514489824, + "grad_norm": 0.46972963213920593, + "learning_rate": 0.00014816026196342802, + "loss": 0.1052, + "step": 14513 + }, + { + "epoch": 0.34023101671970324, + "grad_norm": 0.5710077881813049, + "learning_rate": 0.00014815380620787814, + "loss": 0.1606, + "step": 14514 + }, + { + "epoch": 0.34025445829450823, + "grad_norm": 0.4868013858795166, + "learning_rate": 0.0001481473501910443, + "loss": 0.1314, + "step": 14515 + }, + { + "epoch": 0.3402778998693132, + "grad_norm": 0.6396695971488953, + "learning_rate": 0.0001481408939129616, + "loss": 0.6541, + "step": 14516 + }, + { + "epoch": 0.3403013414441182, + "grad_norm": 0.7767796516418457, + "learning_rate": 0.0001481344373736651, + "loss": 0.1685, + "step": 14517 + }, + { + "epoch": 0.3403247830189232, + "grad_norm": 0.4540571868419647, + "learning_rate": 0.00014812798057318974, + "loss": 0.104, + "step": 14518 + }, + { + "epoch": 0.3403482245937282, + "grad_norm": 0.35780057311058044, + "learning_rate": 0.00014812152351157063, + "loss": 0.113, + "step": 14519 + }, + { + "epoch": 0.3403716661685332, + "grad_norm": 0.49940457940101624, + "learning_rate": 0.00014811506618884278, + "loss": 0.0939, + "step": 14520 + }, + { + "epoch": 0.3403951077433382, + "grad_norm": 0.10503692924976349, + "learning_rate": 0.0001481086086050412, + "loss": 0.0241, + "step": 14521 + }, + { + "epoch": 0.3404185493181432, + "grad_norm": 0.2606598734855652, + "learning_rate": 0.000148102150760201, + "loss": 0.0678, + "step": 14522 + }, + { + "epoch": 0.3404419908929482, + "grad_norm": 0.8609730005264282, + "learning_rate": 0.00014809569265435715, + "loss": 0.1762, + "step": 14523 + }, + { + "epoch": 0.3404654324677532, + "grad_norm": 0.24389150738716125, + "learning_rate": 0.0001480892342875447, + "loss": 0.0318, + "step": 14524 + }, + { + "epoch": 0.3404888740425582, + "grad_norm": 0.3534943461418152, + "learning_rate": 0.00014808277565979874, + "loss": 0.1281, + "step": 14525 + }, + { + "epoch": 0.34051231561736317, + "grad_norm": 0.33229392766952515, + "learning_rate": 0.0001480763167711543, + "loss": 0.1259, + "step": 14526 + }, + { + "epoch": 0.34053575719216816, + "grad_norm": 0.3423255980014801, + "learning_rate": 0.0001480698576216464, + "loss": 0.0654, + "step": 14527 + }, + { + "epoch": 0.34055919876697316, + "grad_norm": 0.0625494047999382, + "learning_rate": 0.00014806339821131007, + "loss": 0.0142, + "step": 14528 + }, + { + "epoch": 0.34058264034177815, + "grad_norm": 0.13465316593647003, + "learning_rate": 0.00014805693854018042, + "loss": 0.0275, + "step": 14529 + }, + { + "epoch": 0.34060608191658315, + "grad_norm": 0.5750711560249329, + "learning_rate": 0.00014805047860829248, + "loss": 0.1422, + "step": 14530 + }, + { + "epoch": 0.34062952349138814, + "grad_norm": 0.39508748054504395, + "learning_rate": 0.0001480440184156813, + "loss": 0.1127, + "step": 14531 + }, + { + "epoch": 0.34065296506619314, + "grad_norm": 0.20489685237407684, + "learning_rate": 0.00014803755796238187, + "loss": 0.0607, + "step": 14532 + }, + { + "epoch": 0.34067640664099813, + "grad_norm": 0.37507086992263794, + "learning_rate": 0.00014803109724842935, + "loss": 0.0572, + "step": 14533 + }, + { + "epoch": 0.3406998482158031, + "grad_norm": 0.5214030742645264, + "learning_rate": 0.0001480246362738587, + "loss": 0.0883, + "step": 14534 + }, + { + "epoch": 0.3407232897906081, + "grad_norm": 0.12696139514446259, + "learning_rate": 0.00014801817503870508, + "loss": 0.0156, + "step": 14535 + }, + { + "epoch": 0.3407467313654131, + "grad_norm": 0.985528290271759, + "learning_rate": 0.00014801171354300347, + "loss": 0.131, + "step": 14536 + }, + { + "epoch": 0.3407701729402181, + "grad_norm": 0.1219584196805954, + "learning_rate": 0.00014800525178678895, + "loss": 0.0199, + "step": 14537 + }, + { + "epoch": 0.3407936145150231, + "grad_norm": 0.240436851978302, + "learning_rate": 0.00014799878977009659, + "loss": 0.0605, + "step": 14538 + }, + { + "epoch": 0.3408170560898281, + "grad_norm": 0.6461689472198486, + "learning_rate": 0.00014799232749296144, + "loss": 0.1788, + "step": 14539 + }, + { + "epoch": 0.3408404976646331, + "grad_norm": 0.6625802516937256, + "learning_rate": 0.00014798586495541857, + "loss": 0.1833, + "step": 14540 + }, + { + "epoch": 0.3408639392394381, + "grad_norm": 0.6759414076805115, + "learning_rate": 0.00014797940215750303, + "loss": 0.1179, + "step": 14541 + }, + { + "epoch": 0.3408873808142431, + "grad_norm": 0.8923348188400269, + "learning_rate": 0.00014797293909924993, + "loss": 0.2329, + "step": 14542 + }, + { + "epoch": 0.3409108223890481, + "grad_norm": 0.4755546450614929, + "learning_rate": 0.0001479664757806943, + "loss": 0.1183, + "step": 14543 + }, + { + "epoch": 0.34093426396385307, + "grad_norm": 0.6981374025344849, + "learning_rate": 0.00014796001220187123, + "loss": 0.1213, + "step": 14544 + }, + { + "epoch": 0.34095770553865806, + "grad_norm": 0.4035606384277344, + "learning_rate": 0.00014795354836281578, + "loss": 0.0762, + "step": 14545 + }, + { + "epoch": 0.34098114711346306, + "grad_norm": 0.7750647068023682, + "learning_rate": 0.00014794708426356306, + "loss": 0.2021, + "step": 14546 + }, + { + "epoch": 0.34100458868826805, + "grad_norm": 0.3379955589771271, + "learning_rate": 0.00014794061990414806, + "loss": 0.0659, + "step": 14547 + }, + { + "epoch": 0.3410280302630731, + "grad_norm": 0.3012698292732239, + "learning_rate": 0.00014793415528460596, + "loss": 0.0718, + "step": 14548 + }, + { + "epoch": 0.3410514718378781, + "grad_norm": 0.17124052345752716, + "learning_rate": 0.00014792769040497175, + "loss": 0.0484, + "step": 14549 + }, + { + "epoch": 0.3410749134126831, + "grad_norm": 0.5218636989593506, + "learning_rate": 0.00014792122526528057, + "loss": 0.2334, + "step": 14550 + }, + { + "epoch": 0.3410983549874881, + "grad_norm": 0.3770613968372345, + "learning_rate": 0.00014791475986556743, + "loss": 0.1278, + "step": 14551 + }, + { + "epoch": 0.3411217965622931, + "grad_norm": 0.2068384438753128, + "learning_rate": 0.00014790829420586752, + "loss": 0.0342, + "step": 14552 + }, + { + "epoch": 0.3411452381370981, + "grad_norm": 0.4161568284034729, + "learning_rate": 0.0001479018282862158, + "loss": 0.0782, + "step": 14553 + }, + { + "epoch": 0.34116867971190307, + "grad_norm": 0.3609873354434967, + "learning_rate": 0.00014789536210664742, + "loss": 0.1097, + "step": 14554 + }, + { + "epoch": 0.34119212128670806, + "grad_norm": 0.5691697001457214, + "learning_rate": 0.0001478888956671975, + "loss": 0.6094, + "step": 14555 + }, + { + "epoch": 0.34121556286151306, + "grad_norm": 0.5758916139602661, + "learning_rate": 0.00014788242896790105, + "loss": 0.1138, + "step": 14556 + }, + { + "epoch": 0.34123900443631805, + "grad_norm": 0.2930554449558258, + "learning_rate": 0.00014787596200879319, + "loss": 0.0601, + "step": 14557 + }, + { + "epoch": 0.34126244601112304, + "grad_norm": 0.4320099651813507, + "learning_rate": 0.00014786949478990904, + "loss": 0.1562, + "step": 14558 + }, + { + "epoch": 0.34128588758592804, + "grad_norm": 0.13540393114089966, + "learning_rate": 0.00014786302731128366, + "loss": 0.0335, + "step": 14559 + }, + { + "epoch": 0.34130932916073303, + "grad_norm": 0.19687193632125854, + "learning_rate": 0.00014785655957295213, + "loss": 0.0351, + "step": 14560 + }, + { + "epoch": 0.341332770735538, + "grad_norm": 0.10588680952787399, + "learning_rate": 0.00014785009157494957, + "loss": 0.0157, + "step": 14561 + }, + { + "epoch": 0.341356212310343, + "grad_norm": 0.3123134672641754, + "learning_rate": 0.0001478436233173111, + "loss": 0.0842, + "step": 14562 + }, + { + "epoch": 0.341379653885148, + "grad_norm": 0.5624749064445496, + "learning_rate": 0.00014783715480007174, + "loss": 0.1067, + "step": 14563 + }, + { + "epoch": 0.341403095459953, + "grad_norm": 0.15437962114810944, + "learning_rate": 0.00014783068602326665, + "loss": 0.0416, + "step": 14564 + }, + { + "epoch": 0.341426537034758, + "grad_norm": 0.4183327555656433, + "learning_rate": 0.00014782421698693093, + "loss": 0.115, + "step": 14565 + }, + { + "epoch": 0.341449978609563, + "grad_norm": 0.44121572375297546, + "learning_rate": 0.0001478177476910996, + "loss": 0.1318, + "step": 14566 + }, + { + "epoch": 0.341473420184368, + "grad_norm": 0.4261903762817383, + "learning_rate": 0.0001478112781358079, + "loss": 0.0583, + "step": 14567 + }, + { + "epoch": 0.341496861759173, + "grad_norm": 0.41163182258605957, + "learning_rate": 0.00014780480832109083, + "loss": 0.0652, + "step": 14568 + }, + { + "epoch": 0.341520303333978, + "grad_norm": 0.5407329201698303, + "learning_rate": 0.00014779833824698355, + "loss": 0.0987, + "step": 14569 + }, + { + "epoch": 0.341543744908783, + "grad_norm": 0.8517464399337769, + "learning_rate": 0.00014779186791352112, + "loss": 0.1436, + "step": 14570 + }, + { + "epoch": 0.34156718648358797, + "grad_norm": 0.569412350654602, + "learning_rate": 0.00014778539732073867, + "loss": 0.1238, + "step": 14571 + }, + { + "epoch": 0.34159062805839296, + "grad_norm": 0.656011700630188, + "learning_rate": 0.0001477789264686713, + "loss": 0.6103, + "step": 14572 + }, + { + "epoch": 0.34161406963319796, + "grad_norm": 0.43119385838508606, + "learning_rate": 0.00014777245535735413, + "loss": 0.1356, + "step": 14573 + }, + { + "epoch": 0.34163751120800295, + "grad_norm": 0.625858724117279, + "learning_rate": 0.0001477659839868223, + "loss": 0.1135, + "step": 14574 + }, + { + "epoch": 0.34166095278280795, + "grad_norm": 0.4281964898109436, + "learning_rate": 0.00014775951235711089, + "loss": 0.1006, + "step": 14575 + }, + { + "epoch": 0.34168439435761294, + "grad_norm": 0.49138447642326355, + "learning_rate": 0.000147753040468255, + "loss": 0.6728, + "step": 14576 + }, + { + "epoch": 0.34170783593241794, + "grad_norm": 0.24828322231769562, + "learning_rate": 0.0001477465683202898, + "loss": 0.0569, + "step": 14577 + }, + { + "epoch": 0.34173127750722293, + "grad_norm": 0.5436041951179504, + "learning_rate": 0.00014774009591325034, + "loss": 0.1097, + "step": 14578 + }, + { + "epoch": 0.3417547190820279, + "grad_norm": 0.3370283842086792, + "learning_rate": 0.00014773362324717182, + "loss": 0.0708, + "step": 14579 + }, + { + "epoch": 0.3417781606568329, + "grad_norm": 0.3852444589138031, + "learning_rate": 0.0001477271503220893, + "loss": 0.1112, + "step": 14580 + }, + { + "epoch": 0.3418016022316379, + "grad_norm": 0.4543953835964203, + "learning_rate": 0.00014772067713803794, + "loss": 0.0924, + "step": 14581 + }, + { + "epoch": 0.3418250438064429, + "grad_norm": 0.35828903317451477, + "learning_rate": 0.0001477142036950528, + "loss": 0.0686, + "step": 14582 + }, + { + "epoch": 0.3418484853812479, + "grad_norm": 0.7877360582351685, + "learning_rate": 0.00014770772999316912, + "loss": 0.1663, + "step": 14583 + }, + { + "epoch": 0.3418719269560529, + "grad_norm": 0.7977624535560608, + "learning_rate": 0.00014770125603242188, + "loss": 0.217, + "step": 14584 + }, + { + "epoch": 0.3418953685308579, + "grad_norm": 0.546941876411438, + "learning_rate": 0.00014769478181284634, + "loss": 0.1556, + "step": 14585 + }, + { + "epoch": 0.3419188101056629, + "grad_norm": 0.14126278460025787, + "learning_rate": 0.00014768830733447751, + "loss": 0.0231, + "step": 14586 + }, + { + "epoch": 0.3419422516804679, + "grad_norm": 0.27489280700683594, + "learning_rate": 0.00014768183259735064, + "loss": 0.0494, + "step": 14587 + }, + { + "epoch": 0.3419656932552729, + "grad_norm": 0.3812173008918762, + "learning_rate": 0.00014767535760150082, + "loss": 0.0941, + "step": 14588 + }, + { + "epoch": 0.34198913483007787, + "grad_norm": 0.5805615186691284, + "learning_rate": 0.0001476688823469631, + "loss": 0.14, + "step": 14589 + }, + { + "epoch": 0.34201257640488286, + "grad_norm": 0.15087737143039703, + "learning_rate": 0.00014766240683377274, + "loss": 0.0333, + "step": 14590 + }, + { + "epoch": 0.34203601797968786, + "grad_norm": 0.5232334733009338, + "learning_rate": 0.00014765593106196479, + "loss": 0.1214, + "step": 14591 + }, + { + "epoch": 0.34205945955449285, + "grad_norm": 0.3788069486618042, + "learning_rate": 0.0001476494550315744, + "loss": 0.1179, + "step": 14592 + }, + { + "epoch": 0.34208290112929784, + "grad_norm": 0.7444151639938354, + "learning_rate": 0.00014764297874263676, + "loss": 0.2266, + "step": 14593 + }, + { + "epoch": 0.34210634270410284, + "grad_norm": 0.7100487351417542, + "learning_rate": 0.000147636502195187, + "loss": 0.0663, + "step": 14594 + }, + { + "epoch": 0.34212978427890783, + "grad_norm": 0.35700932145118713, + "learning_rate": 0.00014763002538926021, + "loss": 0.1288, + "step": 14595 + }, + { + "epoch": 0.34215322585371283, + "grad_norm": 1.5075539350509644, + "learning_rate": 0.0001476235483248916, + "loss": 0.2712, + "step": 14596 + }, + { + "epoch": 0.3421766674285178, + "grad_norm": 0.42776796221733093, + "learning_rate": 0.00014761707100211622, + "loss": 0.0833, + "step": 14597 + }, + { + "epoch": 0.3422001090033228, + "grad_norm": 0.22452501952648163, + "learning_rate": 0.00014761059342096927, + "loss": 0.0315, + "step": 14598 + }, + { + "epoch": 0.3422235505781278, + "grad_norm": 0.3427448868751526, + "learning_rate": 0.00014760411558148594, + "loss": 0.0298, + "step": 14599 + }, + { + "epoch": 0.34224699215293286, + "grad_norm": 0.5811211466789246, + "learning_rate": 0.00014759763748370134, + "loss": 0.1572, + "step": 14600 + }, + { + "epoch": 0.34227043372773786, + "grad_norm": 0.36553001403808594, + "learning_rate": 0.0001475911591276506, + "loss": 0.0765, + "step": 14601 + }, + { + "epoch": 0.34229387530254285, + "grad_norm": 0.4126490652561188, + "learning_rate": 0.0001475846805133689, + "loss": 0.077, + "step": 14602 + }, + { + "epoch": 0.34231731687734784, + "grad_norm": 0.3292762339115143, + "learning_rate": 0.0001475782016408914, + "loss": 0.0271, + "step": 14603 + }, + { + "epoch": 0.34234075845215284, + "grad_norm": 0.44342634081840515, + "learning_rate": 0.00014757172251025325, + "loss": 0.1066, + "step": 14604 + }, + { + "epoch": 0.34236420002695783, + "grad_norm": 0.15043258666992188, + "learning_rate": 0.00014756524312148955, + "loss": 0.0247, + "step": 14605 + }, + { + "epoch": 0.3423876416017628, + "grad_norm": 0.49074694514274597, + "learning_rate": 0.00014755876347463553, + "loss": 0.1631, + "step": 14606 + }, + { + "epoch": 0.3424110831765678, + "grad_norm": 0.09426666051149368, + "learning_rate": 0.00014755228356972633, + "loss": 0.025, + "step": 14607 + }, + { + "epoch": 0.3424345247513728, + "grad_norm": 0.4371212124824524, + "learning_rate": 0.0001475458034067971, + "loss": 0.1231, + "step": 14608 + }, + { + "epoch": 0.3424579663261778, + "grad_norm": 0.5836547017097473, + "learning_rate": 0.00014753932298588296, + "loss": 0.1495, + "step": 14609 + }, + { + "epoch": 0.3424814079009828, + "grad_norm": 0.4498034417629242, + "learning_rate": 0.0001475328423070192, + "loss": 0.1355, + "step": 14610 + }, + { + "epoch": 0.3425048494757878, + "grad_norm": 0.2992309629917145, + "learning_rate": 0.00014752636137024082, + "loss": 0.0644, + "step": 14611 + }, + { + "epoch": 0.3425282910505928, + "grad_norm": 0.6564181447029114, + "learning_rate": 0.00014751988017558312, + "loss": 0.7522, + "step": 14612 + }, + { + "epoch": 0.3425517326253978, + "grad_norm": 0.8838860988616943, + "learning_rate": 0.00014751339872308119, + "loss": 0.1862, + "step": 14613 + }, + { + "epoch": 0.3425751742002028, + "grad_norm": 0.4420505166053772, + "learning_rate": 0.0001475069170127702, + "loss": 0.1235, + "step": 14614 + }, + { + "epoch": 0.3425986157750078, + "grad_norm": 0.6746765375137329, + "learning_rate": 0.00014750043504468537, + "loss": 0.172, + "step": 14615 + }, + { + "epoch": 0.34262205734981277, + "grad_norm": 0.4049280881881714, + "learning_rate": 0.00014749395281886187, + "loss": 0.1002, + "step": 14616 + }, + { + "epoch": 0.34264549892461776, + "grad_norm": 0.37861013412475586, + "learning_rate": 0.0001474874703353348, + "loss": 0.0785, + "step": 14617 + }, + { + "epoch": 0.34266894049942276, + "grad_norm": 0.2018602192401886, + "learning_rate": 0.0001474809875941394, + "loss": 0.0593, + "step": 14618 + }, + { + "epoch": 0.34269238207422775, + "grad_norm": 0.8609459400177002, + "learning_rate": 0.00014747450459531085, + "loss": 0.2474, + "step": 14619 + }, + { + "epoch": 0.34271582364903275, + "grad_norm": 0.5429098010063171, + "learning_rate": 0.00014746802133888426, + "loss": 0.1677, + "step": 14620 + }, + { + "epoch": 0.34273926522383774, + "grad_norm": 0.4367018938064575, + "learning_rate": 0.00014746153782489488, + "loss": 0.1118, + "step": 14621 + }, + { + "epoch": 0.34276270679864274, + "grad_norm": 0.46612608432769775, + "learning_rate": 0.00014745505405337785, + "loss": 0.0702, + "step": 14622 + }, + { + "epoch": 0.34278614837344773, + "grad_norm": 0.5327589511871338, + "learning_rate": 0.0001474485700243684, + "loss": 0.1174, + "step": 14623 + }, + { + "epoch": 0.3428095899482527, + "grad_norm": 0.8003571033477783, + "learning_rate": 0.0001474420857379016, + "loss": 0.1435, + "step": 14624 + }, + { + "epoch": 0.3428330315230577, + "grad_norm": 0.4363585412502289, + "learning_rate": 0.00014743560119401277, + "loss": 0.077, + "step": 14625 + }, + { + "epoch": 0.3428564730978627, + "grad_norm": 0.38427427411079407, + "learning_rate": 0.00014742911639273703, + "loss": 0.0657, + "step": 14626 + }, + { + "epoch": 0.3428799146726677, + "grad_norm": 0.1222960352897644, + "learning_rate": 0.00014742263133410953, + "loss": 0.0289, + "step": 14627 + }, + { + "epoch": 0.3429033562474727, + "grad_norm": 0.33397507667541504, + "learning_rate": 0.0001474161460181655, + "loss": 0.0851, + "step": 14628 + }, + { + "epoch": 0.3429267978222777, + "grad_norm": 0.8071443438529968, + "learning_rate": 0.00014740966044494018, + "loss": 0.2452, + "step": 14629 + }, + { + "epoch": 0.3429502393970827, + "grad_norm": 0.23296944797039032, + "learning_rate": 0.00014740317461446866, + "loss": 0.0685, + "step": 14630 + }, + { + "epoch": 0.3429736809718877, + "grad_norm": 0.3512900173664093, + "learning_rate": 0.00014739668852678623, + "loss": 0.3045, + "step": 14631 + }, + { + "epoch": 0.3429971225466927, + "grad_norm": 0.7419899702072144, + "learning_rate": 0.000147390202181928, + "loss": 0.2038, + "step": 14632 + }, + { + "epoch": 0.3430205641214977, + "grad_norm": 1.5805563926696777, + "learning_rate": 0.0001473837155799292, + "loss": 0.2697, + "step": 14633 + }, + { + "epoch": 0.34304400569630267, + "grad_norm": 0.2574872374534607, + "learning_rate": 0.00014737722872082504, + "loss": 0.0906, + "step": 14634 + }, + { + "epoch": 0.34306744727110766, + "grad_norm": 0.7720604538917542, + "learning_rate": 0.00014737074160465073, + "loss": 0.1858, + "step": 14635 + }, + { + "epoch": 0.34309088884591266, + "grad_norm": 0.6311544179916382, + "learning_rate": 0.00014736425423144138, + "loss": 0.6661, + "step": 14636 + }, + { + "epoch": 0.34311433042071765, + "grad_norm": 0.6459153890609741, + "learning_rate": 0.00014735776660123231, + "loss": 0.1593, + "step": 14637 + }, + { + "epoch": 0.34313777199552264, + "grad_norm": 0.0828884169459343, + "learning_rate": 0.00014735127871405866, + "loss": 0.0131, + "step": 14638 + }, + { + "epoch": 0.34316121357032764, + "grad_norm": 0.2552131414413452, + "learning_rate": 0.0001473447905699556, + "loss": 0.0692, + "step": 14639 + }, + { + "epoch": 0.34318465514513263, + "grad_norm": 0.5074358582496643, + "learning_rate": 0.0001473383021689584, + "loss": 0.1193, + "step": 14640 + }, + { + "epoch": 0.3432080967199376, + "grad_norm": 0.21459649503231049, + "learning_rate": 0.00014733181351110224, + "loss": 0.0435, + "step": 14641 + }, + { + "epoch": 0.3432315382947426, + "grad_norm": 0.1396803855895996, + "learning_rate": 0.00014732532459642235, + "loss": 0.0265, + "step": 14642 + }, + { + "epoch": 0.3432549798695476, + "grad_norm": 0.6913257837295532, + "learning_rate": 0.0001473188354249539, + "loss": 0.2277, + "step": 14643 + }, + { + "epoch": 0.3432784214443526, + "grad_norm": 0.12846031785011292, + "learning_rate": 0.0001473123459967321, + "loss": 0.0252, + "step": 14644 + }, + { + "epoch": 0.3433018630191576, + "grad_norm": 0.685760498046875, + "learning_rate": 0.0001473058563117922, + "loss": 0.1229, + "step": 14645 + }, + { + "epoch": 0.3433253045939626, + "grad_norm": 0.36478346586227417, + "learning_rate": 0.00014729936637016938, + "loss": 0.0793, + "step": 14646 + }, + { + "epoch": 0.3433487461687676, + "grad_norm": 0.5712181925773621, + "learning_rate": 0.0001472928761718989, + "loss": 0.5407, + "step": 14647 + }, + { + "epoch": 0.3433721877435726, + "grad_norm": 0.4362117648124695, + "learning_rate": 0.00014728638571701594, + "loss": 0.0905, + "step": 14648 + }, + { + "epoch": 0.3433956293183776, + "grad_norm": 0.5594156384468079, + "learning_rate": 0.0001472798950055557, + "loss": 0.0932, + "step": 14649 + }, + { + "epoch": 0.3434190708931826, + "grad_norm": 0.3330077826976776, + "learning_rate": 0.00014727340403755343, + "loss": 0.0565, + "step": 14650 + }, + { + "epoch": 0.3434425124679876, + "grad_norm": 0.3999558389186859, + "learning_rate": 0.00014726691281304432, + "loss": 0.0666, + "step": 14651 + }, + { + "epoch": 0.3434659540427926, + "grad_norm": 0.8367995619773865, + "learning_rate": 0.00014726042133206364, + "loss": 0.253, + "step": 14652 + }, + { + "epoch": 0.3434893956175976, + "grad_norm": 0.23587650060653687, + "learning_rate": 0.00014725392959464656, + "loss": 0.0539, + "step": 14653 + }, + { + "epoch": 0.3435128371924026, + "grad_norm": 0.37072300910949707, + "learning_rate": 0.00014724743760082836, + "loss": 0.0955, + "step": 14654 + }, + { + "epoch": 0.3435362787672076, + "grad_norm": 0.45232996344566345, + "learning_rate": 0.0001472409453506442, + "loss": 0.1244, + "step": 14655 + }, + { + "epoch": 0.3435597203420126, + "grad_norm": 0.3649516999721527, + "learning_rate": 0.00014723445284412938, + "loss": 0.0351, + "step": 14656 + }, + { + "epoch": 0.3435831619168176, + "grad_norm": 0.22817742824554443, + "learning_rate": 0.00014722796008131907, + "loss": 0.0378, + "step": 14657 + }, + { + "epoch": 0.3436066034916226, + "grad_norm": 0.40676918625831604, + "learning_rate": 0.00014722146706224853, + "loss": 0.0716, + "step": 14658 + }, + { + "epoch": 0.3436300450664276, + "grad_norm": 0.657470703125, + "learning_rate": 0.00014721497378695296, + "loss": 0.0894, + "step": 14659 + }, + { + "epoch": 0.3436534866412326, + "grad_norm": 0.11996231228113174, + "learning_rate": 0.00014720848025546763, + "loss": 0.0245, + "step": 14660 + }, + { + "epoch": 0.34367692821603757, + "grad_norm": 0.15127748250961304, + "learning_rate": 0.00014720198646782777, + "loss": 0.0376, + "step": 14661 + }, + { + "epoch": 0.34370036979084256, + "grad_norm": 0.46159347891807556, + "learning_rate": 0.0001471954924240686, + "loss": 0.0968, + "step": 14662 + }, + { + "epoch": 0.34372381136564756, + "grad_norm": 0.9527338743209839, + "learning_rate": 0.00014718899812422533, + "loss": 0.5597, + "step": 14663 + }, + { + "epoch": 0.34374725294045255, + "grad_norm": 0.9468867778778076, + "learning_rate": 0.00014718250356833326, + "loss": 0.6208, + "step": 14664 + }, + { + "epoch": 0.34377069451525755, + "grad_norm": 0.37335488200187683, + "learning_rate": 0.00014717600875642758, + "loss": 0.1122, + "step": 14665 + }, + { + "epoch": 0.34379413609006254, + "grad_norm": 0.42321115732192993, + "learning_rate": 0.00014716951368854358, + "loss": 0.0752, + "step": 14666 + }, + { + "epoch": 0.34381757766486754, + "grad_norm": 0.443668395280838, + "learning_rate": 0.00014716301836471645, + "loss": 0.1044, + "step": 14667 + }, + { + "epoch": 0.34384101923967253, + "grad_norm": 0.6056934595108032, + "learning_rate": 0.00014715652278498147, + "loss": 0.1325, + "step": 14668 + }, + { + "epoch": 0.3438644608144775, + "grad_norm": 0.6208871603012085, + "learning_rate": 0.00014715002694937386, + "loss": 0.116, + "step": 14669 + }, + { + "epoch": 0.3438879023892825, + "grad_norm": 0.34472906589508057, + "learning_rate": 0.00014714353085792887, + "loss": 0.3179, + "step": 14670 + }, + { + "epoch": 0.3439113439640875, + "grad_norm": 0.1547999382019043, + "learning_rate": 0.00014713703451068176, + "loss": 0.0393, + "step": 14671 + }, + { + "epoch": 0.3439347855388925, + "grad_norm": 0.34467723965644836, + "learning_rate": 0.00014713053790766777, + "loss": 0.0464, + "step": 14672 + }, + { + "epoch": 0.3439582271136975, + "grad_norm": 0.5655249357223511, + "learning_rate": 0.00014712404104892218, + "loss": 0.0963, + "step": 14673 + }, + { + "epoch": 0.3439816686885025, + "grad_norm": 0.1252972036600113, + "learning_rate": 0.00014711754393448022, + "loss": 0.037, + "step": 14674 + }, + { + "epoch": 0.3440051102633075, + "grad_norm": 0.5305308103561401, + "learning_rate": 0.00014711104656437712, + "loss": 0.1432, + "step": 14675 + }, + { + "epoch": 0.3440285518381125, + "grad_norm": 0.28904008865356445, + "learning_rate": 0.00014710454893864815, + "loss": 0.0298, + "step": 14676 + }, + { + "epoch": 0.3440519934129175, + "grad_norm": 0.7828628420829773, + "learning_rate": 0.00014709805105732858, + "loss": 0.1015, + "step": 14677 + }, + { + "epoch": 0.3440754349877225, + "grad_norm": 0.34799087047576904, + "learning_rate": 0.0001470915529204537, + "loss": 0.0978, + "step": 14678 + }, + { + "epoch": 0.34409887656252747, + "grad_norm": 0.2325655072927475, + "learning_rate": 0.0001470850545280587, + "loss": 0.0368, + "step": 14679 + }, + { + "epoch": 0.34412231813733246, + "grad_norm": 0.549537718296051, + "learning_rate": 0.00014707855588017886, + "loss": 0.6733, + "step": 14680 + }, + { + "epoch": 0.34414575971213746, + "grad_norm": 0.5748422741889954, + "learning_rate": 0.00014707205697684944, + "loss": 0.109, + "step": 14681 + }, + { + "epoch": 0.34416920128694245, + "grad_norm": 0.5450202822685242, + "learning_rate": 0.00014706555781810574, + "loss": 0.8019, + "step": 14682 + }, + { + "epoch": 0.34419264286174744, + "grad_norm": 0.267489492893219, + "learning_rate": 0.000147059058403983, + "loss": 0.0705, + "step": 14683 + }, + { + "epoch": 0.34421608443655244, + "grad_norm": 0.3671042323112488, + "learning_rate": 0.00014705255873451647, + "loss": 0.0553, + "step": 14684 + }, + { + "epoch": 0.34423952601135743, + "grad_norm": 0.6077080965042114, + "learning_rate": 0.00014704605880974146, + "loss": 0.1401, + "step": 14685 + }, + { + "epoch": 0.3442629675861624, + "grad_norm": 0.18345075845718384, + "learning_rate": 0.0001470395586296932, + "loss": 0.0483, + "step": 14686 + }, + { + "epoch": 0.3442864091609674, + "grad_norm": 0.5073060989379883, + "learning_rate": 0.00014703305819440695, + "loss": 0.0922, + "step": 14687 + }, + { + "epoch": 0.3443098507357724, + "grad_norm": 0.11280051618814468, + "learning_rate": 0.00014702655750391803, + "loss": 0.0279, + "step": 14688 + }, + { + "epoch": 0.3443332923105774, + "grad_norm": 0.08932406455278397, + "learning_rate": 0.00014702005655826167, + "loss": 0.0159, + "step": 14689 + }, + { + "epoch": 0.3443567338853824, + "grad_norm": 0.16460709273815155, + "learning_rate": 0.00014701355535747315, + "loss": 0.0392, + "step": 14690 + }, + { + "epoch": 0.3443801754601874, + "grad_norm": 0.364240437746048, + "learning_rate": 0.0001470070539015878, + "loss": 0.1003, + "step": 14691 + }, + { + "epoch": 0.3444036170349924, + "grad_norm": 0.7314023375511169, + "learning_rate": 0.00014700055219064084, + "loss": 0.14, + "step": 14692 + }, + { + "epoch": 0.3444270586097974, + "grad_norm": 0.5790005922317505, + "learning_rate": 0.00014699405022466753, + "loss": 0.1571, + "step": 14693 + }, + { + "epoch": 0.3444505001846024, + "grad_norm": 0.37831467390060425, + "learning_rate": 0.00014698754800370321, + "loss": 0.0441, + "step": 14694 + }, + { + "epoch": 0.3444739417594074, + "grad_norm": 0.44590625166893005, + "learning_rate": 0.00014698104552778314, + "loss": 0.2973, + "step": 14695 + }, + { + "epoch": 0.34449738333421237, + "grad_norm": 0.23485594987869263, + "learning_rate": 0.00014697454279694259, + "loss": 0.0415, + "step": 14696 + }, + { + "epoch": 0.34452082490901736, + "grad_norm": 0.34199926257133484, + "learning_rate": 0.00014696803981121683, + "loss": 0.0827, + "step": 14697 + }, + { + "epoch": 0.34454426648382236, + "grad_norm": 0.43409618735313416, + "learning_rate": 0.0001469615365706412, + "loss": 0.3726, + "step": 14698 + }, + { + "epoch": 0.34456770805862735, + "grad_norm": 0.5366619825363159, + "learning_rate": 0.00014695503307525094, + "loss": 0.0882, + "step": 14699 + }, + { + "epoch": 0.34459114963343235, + "grad_norm": 0.1427312195301056, + "learning_rate": 0.00014694852932508131, + "loss": 0.0181, + "step": 14700 + }, + { + "epoch": 0.34461459120823734, + "grad_norm": 0.36506158113479614, + "learning_rate": 0.00014694202532016768, + "loss": 0.0526, + "step": 14701 + }, + { + "epoch": 0.3446380327830424, + "grad_norm": 0.2131536304950714, + "learning_rate": 0.0001469355210605453, + "loss": 0.0271, + "step": 14702 + }, + { + "epoch": 0.3446614743578474, + "grad_norm": 0.46555137634277344, + "learning_rate": 0.00014692901654624944, + "loss": 0.0833, + "step": 14703 + }, + { + "epoch": 0.3446849159326524, + "grad_norm": 0.2792479395866394, + "learning_rate": 0.00014692251177731545, + "loss": 0.0572, + "step": 14704 + }, + { + "epoch": 0.3447083575074574, + "grad_norm": 0.5071874856948853, + "learning_rate": 0.00014691600675377858, + "loss": 0.1439, + "step": 14705 + }, + { + "epoch": 0.34473179908226237, + "grad_norm": 0.40035730600357056, + "learning_rate": 0.00014690950147567414, + "loss": 0.0565, + "step": 14706 + }, + { + "epoch": 0.34475524065706736, + "grad_norm": 0.6810459494590759, + "learning_rate": 0.00014690299594303737, + "loss": 0.098, + "step": 14707 + }, + { + "epoch": 0.34477868223187236, + "grad_norm": 0.966132402420044, + "learning_rate": 0.0001468964901559037, + "loss": 0.2133, + "step": 14708 + }, + { + "epoch": 0.34480212380667735, + "grad_norm": 0.6163320541381836, + "learning_rate": 0.0001468899841143083, + "loss": 0.1225, + "step": 14709 + }, + { + "epoch": 0.34482556538148235, + "grad_norm": 0.6079674959182739, + "learning_rate": 0.0001468834778182866, + "loss": 0.1047, + "step": 14710 + }, + { + "epoch": 0.34484900695628734, + "grad_norm": 0.45404520630836487, + "learning_rate": 0.0001468769712678738, + "loss": 0.0793, + "step": 14711 + }, + { + "epoch": 0.34487244853109233, + "grad_norm": 0.4504353106021881, + "learning_rate": 0.0001468704644631052, + "loss": 0.0936, + "step": 14712 + }, + { + "epoch": 0.34489589010589733, + "grad_norm": 0.5725010633468628, + "learning_rate": 0.00014686395740401615, + "loss": 0.1375, + "step": 14713 + }, + { + "epoch": 0.3449193316807023, + "grad_norm": 0.5616909265518188, + "learning_rate": 0.00014685745009064197, + "loss": 0.1153, + "step": 14714 + }, + { + "epoch": 0.3449427732555073, + "grad_norm": 0.3079863488674164, + "learning_rate": 0.00014685094252301796, + "loss": 0.0547, + "step": 14715 + }, + { + "epoch": 0.3449662148303123, + "grad_norm": 1.3560594320297241, + "learning_rate": 0.00014684443470117938, + "loss": 0.0764, + "step": 14716 + }, + { + "epoch": 0.3449896564051173, + "grad_norm": 0.19358424842357635, + "learning_rate": 0.00014683792662516161, + "loss": 0.025, + "step": 14717 + }, + { + "epoch": 0.3450130979799223, + "grad_norm": 0.4201611876487732, + "learning_rate": 0.00014683141829499995, + "loss": 0.0615, + "step": 14718 + }, + { + "epoch": 0.3450365395547273, + "grad_norm": 0.5083842277526855, + "learning_rate": 0.00014682490971072964, + "loss": 0.0776, + "step": 14719 + }, + { + "epoch": 0.3450599811295323, + "grad_norm": 0.4781157076358795, + "learning_rate": 0.0001468184008723861, + "loss": 0.1033, + "step": 14720 + }, + { + "epoch": 0.3450834227043373, + "grad_norm": 0.543398380279541, + "learning_rate": 0.00014681189178000458, + "loss": 0.0746, + "step": 14721 + }, + { + "epoch": 0.3451068642791423, + "grad_norm": 0.6340222358703613, + "learning_rate": 0.00014680538243362043, + "loss": 0.1349, + "step": 14722 + }, + { + "epoch": 0.3451303058539473, + "grad_norm": 0.6395598649978638, + "learning_rate": 0.00014679887283326897, + "loss": 0.1706, + "step": 14723 + }, + { + "epoch": 0.34515374742875227, + "grad_norm": 0.3945360481739044, + "learning_rate": 0.0001467923629789855, + "loss": 0.0604, + "step": 14724 + }, + { + "epoch": 0.34517718900355726, + "grad_norm": 0.5667845010757446, + "learning_rate": 0.00014678585287080534, + "loss": 0.1369, + "step": 14725 + }, + { + "epoch": 0.34520063057836226, + "grad_norm": 0.4703423082828522, + "learning_rate": 0.00014677934250876382, + "loss": 0.4779, + "step": 14726 + }, + { + "epoch": 0.34522407215316725, + "grad_norm": 1.0109822750091553, + "learning_rate": 0.0001467728318928963, + "loss": 0.165, + "step": 14727 + }, + { + "epoch": 0.34524751372797224, + "grad_norm": 0.7416578531265259, + "learning_rate": 0.00014676632102323805, + "loss": 0.2164, + "step": 14728 + }, + { + "epoch": 0.34527095530277724, + "grad_norm": 0.2605361342430115, + "learning_rate": 0.00014675980989982445, + "loss": 0.0613, + "step": 14729 + }, + { + "epoch": 0.34529439687758223, + "grad_norm": 0.27275070548057556, + "learning_rate": 0.0001467532985226908, + "loss": 0.0377, + "step": 14730 + }, + { + "epoch": 0.3453178384523872, + "grad_norm": 0.1544209122657776, + "learning_rate": 0.00014674678689187242, + "loss": 0.029, + "step": 14731 + }, + { + "epoch": 0.3453412800271922, + "grad_norm": 0.09759155660867691, + "learning_rate": 0.00014674027500740467, + "loss": 0.0128, + "step": 14732 + }, + { + "epoch": 0.3453647216019972, + "grad_norm": 0.46689313650131226, + "learning_rate": 0.0001467337628693229, + "loss": 0.0932, + "step": 14733 + }, + { + "epoch": 0.3453881631768022, + "grad_norm": 0.5153895616531372, + "learning_rate": 0.0001467272504776624, + "loss": 0.1067, + "step": 14734 + }, + { + "epoch": 0.3454116047516072, + "grad_norm": 0.4404832124710083, + "learning_rate": 0.00014672073783245852, + "loss": 0.0526, + "step": 14735 + }, + { + "epoch": 0.3454350463264122, + "grad_norm": 0.46897608041763306, + "learning_rate": 0.00014671422493374661, + "loss": 0.1306, + "step": 14736 + }, + { + "epoch": 0.3454584879012172, + "grad_norm": 0.7499039173126221, + "learning_rate": 0.00014670771178156198, + "loss": 0.198, + "step": 14737 + }, + { + "epoch": 0.3454819294760222, + "grad_norm": 0.5874882340431213, + "learning_rate": 0.00014670119837594, + "loss": 0.0501, + "step": 14738 + }, + { + "epoch": 0.3455053710508272, + "grad_norm": 0.4440721273422241, + "learning_rate": 0.00014669468471691606, + "loss": 0.1119, + "step": 14739 + }, + { + "epoch": 0.3455288126256322, + "grad_norm": 0.625505268573761, + "learning_rate": 0.0001466881708045254, + "loss": 0.7452, + "step": 14740 + }, + { + "epoch": 0.34555225420043717, + "grad_norm": 0.6758635640144348, + "learning_rate": 0.0001466816566388034, + "loss": 0.1284, + "step": 14741 + }, + { + "epoch": 0.34557569577524216, + "grad_norm": 0.4248811602592468, + "learning_rate": 0.00014667514221978546, + "loss": 0.1109, + "step": 14742 + }, + { + "epoch": 0.34559913735004716, + "grad_norm": 0.23147501051425934, + "learning_rate": 0.00014666862754750682, + "loss": 0.0525, + "step": 14743 + }, + { + "epoch": 0.34562257892485215, + "grad_norm": 0.18166625499725342, + "learning_rate": 0.00014666211262200296, + "loss": 0.0407, + "step": 14744 + }, + { + "epoch": 0.34564602049965715, + "grad_norm": 0.690414309501648, + "learning_rate": 0.0001466555974433091, + "loss": 0.1616, + "step": 14745 + }, + { + "epoch": 0.34566946207446214, + "grad_norm": 0.4848153293132782, + "learning_rate": 0.00014664908201146074, + "loss": 0.1441, + "step": 14746 + }, + { + "epoch": 0.34569290364926714, + "grad_norm": 1.0040289163589478, + "learning_rate": 0.0001466425663264931, + "loss": 0.1158, + "step": 14747 + }, + { + "epoch": 0.34571634522407213, + "grad_norm": 0.12778641283512115, + "learning_rate": 0.00014663605038844158, + "loss": 0.0203, + "step": 14748 + }, + { + "epoch": 0.3457397867988771, + "grad_norm": 0.1433655470609665, + "learning_rate": 0.00014662953419734156, + "loss": 0.0428, + "step": 14749 + }, + { + "epoch": 0.3457632283736821, + "grad_norm": 0.5940723419189453, + "learning_rate": 0.00014662301775322838, + "loss": 0.0848, + "step": 14750 + }, + { + "epoch": 0.3457866699484871, + "grad_norm": 0.5009192228317261, + "learning_rate": 0.00014661650105613736, + "loss": 0.1005, + "step": 14751 + }, + { + "epoch": 0.3458101115232921, + "grad_norm": 0.7212299704551697, + "learning_rate": 0.00014660998410610394, + "loss": 0.1258, + "step": 14752 + }, + { + "epoch": 0.34583355309809716, + "grad_norm": 0.4568786919116974, + "learning_rate": 0.0001466034669031634, + "loss": 0.1213, + "step": 14753 + }, + { + "epoch": 0.34585699467290215, + "grad_norm": 0.6767177581787109, + "learning_rate": 0.00014659694944735115, + "loss": 0.4893, + "step": 14754 + }, + { + "epoch": 0.34588043624770715, + "grad_norm": 0.5810200572013855, + "learning_rate": 0.00014659043173870253, + "loss": 0.078, + "step": 14755 + }, + { + "epoch": 0.34590387782251214, + "grad_norm": 0.2626597285270691, + "learning_rate": 0.00014658391377725294, + "loss": 0.0561, + "step": 14756 + }, + { + "epoch": 0.34592731939731713, + "grad_norm": 0.21177583932876587, + "learning_rate": 0.00014657739556303768, + "loss": 0.0327, + "step": 14757 + }, + { + "epoch": 0.34595076097212213, + "grad_norm": 0.6707342267036438, + "learning_rate": 0.0001465708770960922, + "loss": 0.0999, + "step": 14758 + }, + { + "epoch": 0.3459742025469271, + "grad_norm": 0.6258993148803711, + "learning_rate": 0.00014656435837645185, + "loss": 0.2413, + "step": 14759 + }, + { + "epoch": 0.3459976441217321, + "grad_norm": 0.5313205718994141, + "learning_rate": 0.00014655783940415194, + "loss": 0.1157, + "step": 14760 + }, + { + "epoch": 0.3460210856965371, + "grad_norm": 0.5160676836967468, + "learning_rate": 0.0001465513201792279, + "loss": 0.0596, + "step": 14761 + }, + { + "epoch": 0.3460445272713421, + "grad_norm": 0.25781503319740295, + "learning_rate": 0.0001465448007017151, + "loss": 0.0581, + "step": 14762 + }, + { + "epoch": 0.3460679688461471, + "grad_norm": 0.37821459770202637, + "learning_rate": 0.0001465382809716489, + "loss": 0.4322, + "step": 14763 + }, + { + "epoch": 0.3460914104209521, + "grad_norm": 0.10765458643436432, + "learning_rate": 0.00014653176098906466, + "loss": 0.03, + "step": 14764 + }, + { + "epoch": 0.3461148519957571, + "grad_norm": 0.09915535897016525, + "learning_rate": 0.00014652524075399782, + "loss": 0.0247, + "step": 14765 + }, + { + "epoch": 0.3461382935705621, + "grad_norm": 1.0404589176177979, + "learning_rate": 0.00014651872026648366, + "loss": 0.1227, + "step": 14766 + }, + { + "epoch": 0.3461617351453671, + "grad_norm": 0.41505759954452515, + "learning_rate": 0.0001465121995265576, + "loss": 0.1016, + "step": 14767 + }, + { + "epoch": 0.34618517672017207, + "grad_norm": 0.19937999546527863, + "learning_rate": 0.00014650567853425513, + "loss": 0.0184, + "step": 14768 + }, + { + "epoch": 0.34620861829497707, + "grad_norm": 0.4238491356372833, + "learning_rate": 0.00014649915728961148, + "loss": 0.0543, + "step": 14769 + }, + { + "epoch": 0.34623205986978206, + "grad_norm": 1.0200430154800415, + "learning_rate": 0.00014649263579266208, + "loss": 0.2661, + "step": 14770 + }, + { + "epoch": 0.34625550144458705, + "grad_norm": 0.581354558467865, + "learning_rate": 0.00014648611404344237, + "loss": 0.0443, + "step": 14771 + }, + { + "epoch": 0.34627894301939205, + "grad_norm": 0.33383530378341675, + "learning_rate": 0.00014647959204198768, + "loss": 0.0672, + "step": 14772 + }, + { + "epoch": 0.34630238459419704, + "grad_norm": 0.7000291347503662, + "learning_rate": 0.0001464730697883334, + "loss": 0.6768, + "step": 14773 + }, + { + "epoch": 0.34632582616900204, + "grad_norm": 0.47690749168395996, + "learning_rate": 0.00014646654728251498, + "loss": 0.0642, + "step": 14774 + }, + { + "epoch": 0.34634926774380703, + "grad_norm": 0.39546138048171997, + "learning_rate": 0.00014646002452456776, + "loss": 0.095, + "step": 14775 + }, + { + "epoch": 0.346372709318612, + "grad_norm": 0.185487300157547, + "learning_rate": 0.0001464535015145271, + "loss": 0.0369, + "step": 14776 + }, + { + "epoch": 0.346396150893417, + "grad_norm": 0.6115723848342896, + "learning_rate": 0.00014644697825242844, + "loss": 0.5305, + "step": 14777 + }, + { + "epoch": 0.346419592468222, + "grad_norm": 0.3064315617084503, + "learning_rate": 0.00014644045473830722, + "loss": 0.0419, + "step": 14778 + }, + { + "epoch": 0.346443034043027, + "grad_norm": 0.43628814816474915, + "learning_rate": 0.00014643393097219876, + "loss": 0.1145, + "step": 14779 + }, + { + "epoch": 0.346466475617832, + "grad_norm": 0.49428799748420715, + "learning_rate": 0.00014642740695413846, + "loss": 0.144, + "step": 14780 + }, + { + "epoch": 0.346489917192637, + "grad_norm": 0.48461097478866577, + "learning_rate": 0.00014642088268416176, + "loss": 0.1125, + "step": 14781 + }, + { + "epoch": 0.346513358767442, + "grad_norm": 0.8882256150245667, + "learning_rate": 0.00014641435816230405, + "loss": 0.16, + "step": 14782 + }, + { + "epoch": 0.346536800342247, + "grad_norm": 0.923366367816925, + "learning_rate": 0.00014640783338860073, + "loss": 0.1471, + "step": 14783 + }, + { + "epoch": 0.346560241917052, + "grad_norm": 0.4061291515827179, + "learning_rate": 0.0001464013083630872, + "loss": 0.0973, + "step": 14784 + }, + { + "epoch": 0.346583683491857, + "grad_norm": 0.14388523995876312, + "learning_rate": 0.00014639478308579886, + "loss": 0.0251, + "step": 14785 + }, + { + "epoch": 0.34660712506666197, + "grad_norm": 0.6107262372970581, + "learning_rate": 0.0001463882575567711, + "loss": 0.225, + "step": 14786 + }, + { + "epoch": 0.34663056664146696, + "grad_norm": 0.2246498167514801, + "learning_rate": 0.0001463817317760394, + "loss": 0.036, + "step": 14787 + }, + { + "epoch": 0.34665400821627196, + "grad_norm": 0.3520158529281616, + "learning_rate": 0.00014637520574363913, + "loss": 0.0586, + "step": 14788 + }, + { + "epoch": 0.34667744979107695, + "grad_norm": 0.5759493112564087, + "learning_rate": 0.00014636867945960562, + "loss": 0.5532, + "step": 14789 + }, + { + "epoch": 0.34670089136588195, + "grad_norm": 0.3824179470539093, + "learning_rate": 0.0001463621529239744, + "loss": 0.1408, + "step": 14790 + }, + { + "epoch": 0.34672433294068694, + "grad_norm": 0.38148701190948486, + "learning_rate": 0.00014635562613678083, + "loss": 0.0994, + "step": 14791 + }, + { + "epoch": 0.34674777451549194, + "grad_norm": 0.8063398599624634, + "learning_rate": 0.0001463490990980603, + "loss": 0.1625, + "step": 14792 + }, + { + "epoch": 0.34677121609029693, + "grad_norm": 0.9186820387840271, + "learning_rate": 0.0001463425718078483, + "loss": 0.1813, + "step": 14793 + }, + { + "epoch": 0.3467946576651019, + "grad_norm": 0.4592430591583252, + "learning_rate": 0.00014633604426618016, + "loss": 0.0955, + "step": 14794 + }, + { + "epoch": 0.3468180992399069, + "grad_norm": 0.24782665073871613, + "learning_rate": 0.00014632951647309137, + "loss": 0.08, + "step": 14795 + }, + { + "epoch": 0.3468415408147119, + "grad_norm": 0.44271254539489746, + "learning_rate": 0.0001463229884286173, + "loss": 0.118, + "step": 14796 + }, + { + "epoch": 0.3468649823895169, + "grad_norm": 0.13066944479942322, + "learning_rate": 0.0001463164601327934, + "loss": 0.0206, + "step": 14797 + }, + { + "epoch": 0.3468884239643219, + "grad_norm": 0.3751559853553772, + "learning_rate": 0.0001463099315856551, + "loss": 0.0887, + "step": 14798 + }, + { + "epoch": 0.3469118655391269, + "grad_norm": 0.5798384547233582, + "learning_rate": 0.00014630340278723778, + "loss": 0.5408, + "step": 14799 + }, + { + "epoch": 0.3469353071139319, + "grad_norm": 1.1839230060577393, + "learning_rate": 0.00014629687373757692, + "loss": 0.2185, + "step": 14800 + }, + { + "epoch": 0.3469587486887369, + "grad_norm": 0.2842055559158325, + "learning_rate": 0.0001462903444367079, + "loss": 0.0648, + "step": 14801 + }, + { + "epoch": 0.3469821902635419, + "grad_norm": 0.47225379943847656, + "learning_rate": 0.0001462838148846662, + "loss": 0.1434, + "step": 14802 + }, + { + "epoch": 0.3470056318383469, + "grad_norm": 0.4255422055721283, + "learning_rate": 0.0001462772850814872, + "loss": 0.1427, + "step": 14803 + }, + { + "epoch": 0.3470290734131519, + "grad_norm": 0.37630516290664673, + "learning_rate": 0.00014627075502720636, + "loss": 0.0902, + "step": 14804 + }, + { + "epoch": 0.3470525149879569, + "grad_norm": 0.4361509382724762, + "learning_rate": 0.00014626422472185908, + "loss": 0.0488, + "step": 14805 + }, + { + "epoch": 0.3470759565627619, + "grad_norm": 0.5063683986663818, + "learning_rate": 0.00014625769416548083, + "loss": 0.1046, + "step": 14806 + }, + { + "epoch": 0.3470993981375669, + "grad_norm": 0.24163082242012024, + "learning_rate": 0.00014625116335810704, + "loss": 0.0712, + "step": 14807 + }, + { + "epoch": 0.3471228397123719, + "grad_norm": 0.5401737689971924, + "learning_rate": 0.00014624463229977313, + "loss": 0.7148, + "step": 14808 + }, + { + "epoch": 0.3471462812871769, + "grad_norm": 0.5672754049301147, + "learning_rate": 0.00014623810099051454, + "loss": 0.6484, + "step": 14809 + }, + { + "epoch": 0.3471697228619819, + "grad_norm": 0.6716971397399902, + "learning_rate": 0.0001462315694303667, + "loss": 0.0714, + "step": 14810 + }, + { + "epoch": 0.3471931644367869, + "grad_norm": 0.31649041175842285, + "learning_rate": 0.00014622503761936508, + "loss": 0.0492, + "step": 14811 + }, + { + "epoch": 0.3472166060115919, + "grad_norm": 0.18609528243541718, + "learning_rate": 0.00014621850555754512, + "loss": 0.048, + "step": 14812 + }, + { + "epoch": 0.34724004758639687, + "grad_norm": 0.22557757794857025, + "learning_rate": 0.00014621197324494225, + "loss": 0.0653, + "step": 14813 + }, + { + "epoch": 0.34726348916120187, + "grad_norm": 0.12530362606048584, + "learning_rate": 0.00014620544068159189, + "loss": 0.0291, + "step": 14814 + }, + { + "epoch": 0.34728693073600686, + "grad_norm": 0.454828679561615, + "learning_rate": 0.0001461989078675295, + "loss": 0.1008, + "step": 14815 + }, + { + "epoch": 0.34731037231081185, + "grad_norm": 0.6274641156196594, + "learning_rate": 0.00014619237480279057, + "loss": 0.149, + "step": 14816 + }, + { + "epoch": 0.34733381388561685, + "grad_norm": 0.3220435678958893, + "learning_rate": 0.00014618584148741052, + "loss": 0.0489, + "step": 14817 + }, + { + "epoch": 0.34735725546042184, + "grad_norm": 0.6446823477745056, + "learning_rate": 0.00014617930792142477, + "loss": 0.0964, + "step": 14818 + }, + { + "epoch": 0.34738069703522684, + "grad_norm": 0.09135536849498749, + "learning_rate": 0.0001461727741048688, + "loss": 0.0237, + "step": 14819 + }, + { + "epoch": 0.34740413861003183, + "grad_norm": 0.38385242223739624, + "learning_rate": 0.00014616624003777805, + "loss": 0.0983, + "step": 14820 + }, + { + "epoch": 0.3474275801848368, + "grad_norm": 0.40865010023117065, + "learning_rate": 0.000146159705720188, + "loss": 0.084, + "step": 14821 + }, + { + "epoch": 0.3474510217596418, + "grad_norm": 0.4291025698184967, + "learning_rate": 0.0001461531711521341, + "loss": 0.5, + "step": 14822 + }, + { + "epoch": 0.3474744633344468, + "grad_norm": 0.47661587595939636, + "learning_rate": 0.00014614663633365178, + "loss": 0.0858, + "step": 14823 + }, + { + "epoch": 0.3474979049092518, + "grad_norm": 0.534820556640625, + "learning_rate": 0.00014614010126477652, + "loss": 0.0989, + "step": 14824 + }, + { + "epoch": 0.3475213464840568, + "grad_norm": 0.4743317663669586, + "learning_rate": 0.00014613356594554374, + "loss": 0.1104, + "step": 14825 + }, + { + "epoch": 0.3475447880588618, + "grad_norm": 0.2551210820674896, + "learning_rate": 0.00014612703037598902, + "loss": 0.0481, + "step": 14826 + }, + { + "epoch": 0.3475682296336668, + "grad_norm": 0.4444148540496826, + "learning_rate": 0.00014612049455614765, + "loss": 0.0758, + "step": 14827 + }, + { + "epoch": 0.3475916712084718, + "grad_norm": 0.3214248716831207, + "learning_rate": 0.0001461139584860552, + "loss": 0.1126, + "step": 14828 + }, + { + "epoch": 0.3476151127832768, + "grad_norm": 0.3942975401878357, + "learning_rate": 0.0001461074221657471, + "loss": 0.116, + "step": 14829 + }, + { + "epoch": 0.3476385543580818, + "grad_norm": 0.740032970905304, + "learning_rate": 0.00014610088559525886, + "loss": 0.114, + "step": 14830 + }, + { + "epoch": 0.34766199593288677, + "grad_norm": 0.47563475370407104, + "learning_rate": 0.00014609434877462592, + "loss": 0.1007, + "step": 14831 + }, + { + "epoch": 0.34768543750769176, + "grad_norm": 0.6015611886978149, + "learning_rate": 0.00014608781170388375, + "loss": 0.1067, + "step": 14832 + }, + { + "epoch": 0.34770887908249676, + "grad_norm": 0.12942737340927124, + "learning_rate": 0.0001460812743830678, + "loss": 0.0273, + "step": 14833 + }, + { + "epoch": 0.34773232065730175, + "grad_norm": 0.4564424753189087, + "learning_rate": 0.00014607473681221356, + "loss": 0.0825, + "step": 14834 + }, + { + "epoch": 0.34775576223210675, + "grad_norm": 0.28557297587394714, + "learning_rate": 0.0001460681989913565, + "loss": 0.0621, + "step": 14835 + }, + { + "epoch": 0.34777920380691174, + "grad_norm": 0.560547947883606, + "learning_rate": 0.00014606166092053213, + "loss": 0.1055, + "step": 14836 + }, + { + "epoch": 0.34780264538171674, + "grad_norm": 0.6840406656265259, + "learning_rate": 0.00014605512259977587, + "loss": 0.7002, + "step": 14837 + }, + { + "epoch": 0.34782608695652173, + "grad_norm": 0.43568721413612366, + "learning_rate": 0.00014604858402912323, + "loss": 0.09, + "step": 14838 + }, + { + "epoch": 0.3478495285313267, + "grad_norm": 0.41321104764938354, + "learning_rate": 0.00014604204520860965, + "loss": 0.0635, + "step": 14839 + }, + { + "epoch": 0.3478729701061317, + "grad_norm": 0.31243783235549927, + "learning_rate": 0.00014603550613827065, + "loss": 0.0505, + "step": 14840 + }, + { + "epoch": 0.3478964116809367, + "grad_norm": 0.536637544631958, + "learning_rate": 0.00014602896681814172, + "loss": 0.1745, + "step": 14841 + }, + { + "epoch": 0.3479198532557417, + "grad_norm": 0.419322669506073, + "learning_rate": 0.0001460224272482583, + "loss": 0.1136, + "step": 14842 + }, + { + "epoch": 0.3479432948305467, + "grad_norm": 0.35221394896507263, + "learning_rate": 0.00014601588742865591, + "loss": 0.0774, + "step": 14843 + }, + { + "epoch": 0.3479667364053517, + "grad_norm": 0.44680991768836975, + "learning_rate": 0.00014600934735937003, + "loss": 0.0717, + "step": 14844 + }, + { + "epoch": 0.3479901779801567, + "grad_norm": 0.5418670773506165, + "learning_rate": 0.0001460028070404361, + "loss": 0.1267, + "step": 14845 + }, + { + "epoch": 0.3480136195549617, + "grad_norm": 0.44308462738990784, + "learning_rate": 0.00014599626647188965, + "loss": 0.0384, + "step": 14846 + }, + { + "epoch": 0.3480370611297667, + "grad_norm": 0.6119072437286377, + "learning_rate": 0.00014598972565376618, + "loss": 0.8149, + "step": 14847 + }, + { + "epoch": 0.3480605027045717, + "grad_norm": 0.5539621114730835, + "learning_rate": 0.00014598318458610118, + "loss": 0.0832, + "step": 14848 + }, + { + "epoch": 0.34808394427937667, + "grad_norm": 0.6015462875366211, + "learning_rate": 0.0001459766432689301, + "loss": 0.1565, + "step": 14849 + }, + { + "epoch": 0.34810738585418166, + "grad_norm": 0.49168843030929565, + "learning_rate": 0.00014597010170228848, + "loss": 0.0831, + "step": 14850 + }, + { + "epoch": 0.34813082742898666, + "grad_norm": 0.7138320207595825, + "learning_rate": 0.0001459635598862118, + "loss": 0.1807, + "step": 14851 + }, + { + "epoch": 0.34815426900379165, + "grad_norm": 0.17069761455059052, + "learning_rate": 0.00014595701782073555, + "loss": 0.0261, + "step": 14852 + }, + { + "epoch": 0.34817771057859664, + "grad_norm": 0.2963777780532837, + "learning_rate": 0.00014595047550589524, + "loss": 0.0849, + "step": 14853 + }, + { + "epoch": 0.34820115215340164, + "grad_norm": 0.5627407431602478, + "learning_rate": 0.00014594393294172634, + "loss": 0.1094, + "step": 14854 + }, + { + "epoch": 0.3482245937282067, + "grad_norm": 0.5761268734931946, + "learning_rate": 0.00014593739012826434, + "loss": 0.8838, + "step": 14855 + }, + { + "epoch": 0.3482480353030117, + "grad_norm": 0.2729710042476654, + "learning_rate": 0.0001459308470655448, + "loss": 0.0492, + "step": 14856 + }, + { + "epoch": 0.3482714768778167, + "grad_norm": 0.5345137715339661, + "learning_rate": 0.0001459243037536032, + "loss": 0.1806, + "step": 14857 + }, + { + "epoch": 0.34829491845262167, + "grad_norm": 0.13288253545761108, + "learning_rate": 0.00014591776019247501, + "loss": 0.0191, + "step": 14858 + }, + { + "epoch": 0.34831836002742667, + "grad_norm": 0.29390841722488403, + "learning_rate": 0.00014591121638219578, + "loss": 0.0748, + "step": 14859 + }, + { + "epoch": 0.34834180160223166, + "grad_norm": 0.4729017913341522, + "learning_rate": 0.000145904672322801, + "loss": 0.1086, + "step": 14860 + }, + { + "epoch": 0.34836524317703665, + "grad_norm": 0.37274104356765747, + "learning_rate": 0.00014589812801432618, + "loss": 0.0794, + "step": 14861 + }, + { + "epoch": 0.34838868475184165, + "grad_norm": 0.6242573857307434, + "learning_rate": 0.0001458915834568068, + "loss": 0.1714, + "step": 14862 + }, + { + "epoch": 0.34841212632664664, + "grad_norm": 0.2040790319442749, + "learning_rate": 0.00014588503865027846, + "loss": 0.0849, + "step": 14863 + }, + { + "epoch": 0.34843556790145164, + "grad_norm": 0.5473454594612122, + "learning_rate": 0.00014587849359477654, + "loss": 0.816, + "step": 14864 + }, + { + "epoch": 0.34845900947625663, + "grad_norm": 0.42211154103279114, + "learning_rate": 0.00014587194829033666, + "loss": 0.0861, + "step": 14865 + }, + { + "epoch": 0.3484824510510616, + "grad_norm": 0.520911455154419, + "learning_rate": 0.0001458654027369943, + "loss": 0.1041, + "step": 14866 + }, + { + "epoch": 0.3485058926258666, + "grad_norm": 0.2501066327095032, + "learning_rate": 0.00014585885693478497, + "loss": 0.0438, + "step": 14867 + }, + { + "epoch": 0.3485293342006716, + "grad_norm": 0.1924593448638916, + "learning_rate": 0.00014585231088374418, + "loss": 0.0315, + "step": 14868 + }, + { + "epoch": 0.3485527757754766, + "grad_norm": 0.0688888281583786, + "learning_rate": 0.0001458457645839075, + "loss": 0.0083, + "step": 14869 + }, + { + "epoch": 0.3485762173502816, + "grad_norm": 0.6925193667411804, + "learning_rate": 0.00014583921803531036, + "loss": 0.5653, + "step": 14870 + }, + { + "epoch": 0.3485996589250866, + "grad_norm": 0.5050672888755798, + "learning_rate": 0.00014583267123798835, + "loss": 0.1368, + "step": 14871 + }, + { + "epoch": 0.3486231004998916, + "grad_norm": 1.0788770914077759, + "learning_rate": 0.00014582612419197698, + "loss": 0.1945, + "step": 14872 + }, + { + "epoch": 0.3486465420746966, + "grad_norm": 0.7473042607307434, + "learning_rate": 0.0001458195768973118, + "loss": 0.1498, + "step": 14873 + }, + { + "epoch": 0.3486699836495016, + "grad_norm": 0.5875660181045532, + "learning_rate": 0.00014581302935402828, + "loss": 0.1494, + "step": 14874 + }, + { + "epoch": 0.3486934252243066, + "grad_norm": 0.6435931324958801, + "learning_rate": 0.00014580648156216198, + "loss": 0.0607, + "step": 14875 + }, + { + "epoch": 0.34871686679911157, + "grad_norm": 0.643822431564331, + "learning_rate": 0.0001457999335217484, + "loss": 0.1966, + "step": 14876 + }, + { + "epoch": 0.34874030837391656, + "grad_norm": 0.6516477465629578, + "learning_rate": 0.00014579338523282314, + "loss": 0.0646, + "step": 14877 + }, + { + "epoch": 0.34876374994872156, + "grad_norm": 0.5935094952583313, + "learning_rate": 0.00014578683669542165, + "loss": 0.1048, + "step": 14878 + }, + { + "epoch": 0.34878719152352655, + "grad_norm": 0.11037276685237885, + "learning_rate": 0.00014578028790957948, + "loss": 0.0251, + "step": 14879 + }, + { + "epoch": 0.34881063309833155, + "grad_norm": 0.9192205667495728, + "learning_rate": 0.00014577373887533223, + "loss": 0.1082, + "step": 14880 + }, + { + "epoch": 0.34883407467313654, + "grad_norm": 0.621097207069397, + "learning_rate": 0.00014576718959271535, + "loss": 0.1596, + "step": 14881 + }, + { + "epoch": 0.34885751624794153, + "grad_norm": 0.6369591951370239, + "learning_rate": 0.00014576064006176443, + "loss": 0.1777, + "step": 14882 + }, + { + "epoch": 0.34888095782274653, + "grad_norm": 1.0098998546600342, + "learning_rate": 0.000145754090282515, + "loss": 0.159, + "step": 14883 + }, + { + "epoch": 0.3489043993975515, + "grad_norm": 0.3810213506221771, + "learning_rate": 0.00014574754025500254, + "loss": 0.1051, + "step": 14884 + }, + { + "epoch": 0.3489278409723565, + "grad_norm": 0.5646651387214661, + "learning_rate": 0.00014574098997926268, + "loss": 0.1576, + "step": 14885 + }, + { + "epoch": 0.3489512825471615, + "grad_norm": 0.2033083587884903, + "learning_rate": 0.00014573443945533094, + "loss": 0.0408, + "step": 14886 + }, + { + "epoch": 0.3489747241219665, + "grad_norm": 0.5571937561035156, + "learning_rate": 0.0001457278886832428, + "loss": 0.1263, + "step": 14887 + }, + { + "epoch": 0.3489981656967715, + "grad_norm": 0.5842241048812866, + "learning_rate": 0.0001457213376630339, + "loss": 0.0661, + "step": 14888 + }, + { + "epoch": 0.3490216072715765, + "grad_norm": 0.14338193833827972, + "learning_rate": 0.00014571478639473969, + "loss": 0.0324, + "step": 14889 + }, + { + "epoch": 0.3490450488463815, + "grad_norm": 0.44909918308258057, + "learning_rate": 0.0001457082348783958, + "loss": 0.4785, + "step": 14890 + }, + { + "epoch": 0.3490684904211865, + "grad_norm": 0.4651401937007904, + "learning_rate": 0.0001457016831140377, + "loss": 0.0858, + "step": 14891 + }, + { + "epoch": 0.3490919319959915, + "grad_norm": 0.5687074661254883, + "learning_rate": 0.00014569513110170102, + "loss": 0.1321, + "step": 14892 + }, + { + "epoch": 0.3491153735707965, + "grad_norm": 0.44393691420555115, + "learning_rate": 0.0001456885788414213, + "loss": 0.085, + "step": 14893 + }, + { + "epoch": 0.34913881514560147, + "grad_norm": 0.27865204215049744, + "learning_rate": 0.00014568202633323397, + "loss": 0.0316, + "step": 14894 + }, + { + "epoch": 0.34916225672040646, + "grad_norm": 0.2587791085243225, + "learning_rate": 0.00014567547357717474, + "loss": 0.0853, + "step": 14895 + }, + { + "epoch": 0.34918569829521146, + "grad_norm": 0.5067504048347473, + "learning_rate": 0.00014566892057327912, + "loss": 0.0912, + "step": 14896 + }, + { + "epoch": 0.34920913987001645, + "grad_norm": 0.44160494208335876, + "learning_rate": 0.00014566236732158262, + "loss": 0.0438, + "step": 14897 + }, + { + "epoch": 0.34923258144482144, + "grad_norm": 0.4899040460586548, + "learning_rate": 0.00014565581382212085, + "loss": 0.1357, + "step": 14898 + }, + { + "epoch": 0.34925602301962644, + "grad_norm": 0.45115962624549866, + "learning_rate": 0.0001456492600749294, + "loss": 0.0929, + "step": 14899 + }, + { + "epoch": 0.34927946459443143, + "grad_norm": 0.4258163869380951, + "learning_rate": 0.0001456427060800437, + "loss": 0.1307, + "step": 14900 + }, + { + "epoch": 0.3493029061692364, + "grad_norm": 0.2688123285770416, + "learning_rate": 0.0001456361518374994, + "loss": 0.0461, + "step": 14901 + }, + { + "epoch": 0.3493263477440414, + "grad_norm": 0.3487761914730072, + "learning_rate": 0.0001456295973473321, + "loss": 0.0678, + "step": 14902 + }, + { + "epoch": 0.3493497893188464, + "grad_norm": 0.9039412140846252, + "learning_rate": 0.0001456230426095773, + "loss": 0.1998, + "step": 14903 + }, + { + "epoch": 0.3493732308936514, + "grad_norm": 0.5256901383399963, + "learning_rate": 0.00014561648762427058, + "loss": 0.0647, + "step": 14904 + }, + { + "epoch": 0.3493966724684564, + "grad_norm": 0.2279919981956482, + "learning_rate": 0.00014560993239144755, + "loss": 0.0626, + "step": 14905 + }, + { + "epoch": 0.34942011404326145, + "grad_norm": 1.0236186981201172, + "learning_rate": 0.0001456033769111437, + "loss": 0.0947, + "step": 14906 + }, + { + "epoch": 0.34944355561806645, + "grad_norm": 0.4677368998527527, + "learning_rate": 0.00014559682118339467, + "loss": 0.0937, + "step": 14907 + }, + { + "epoch": 0.34946699719287144, + "grad_norm": 0.5734500288963318, + "learning_rate": 0.000145590265208236, + "loss": 0.7066, + "step": 14908 + }, + { + "epoch": 0.34949043876767644, + "grad_norm": 0.42770490050315857, + "learning_rate": 0.00014558370898570328, + "loss": 0.0908, + "step": 14909 + }, + { + "epoch": 0.34951388034248143, + "grad_norm": 0.4901486337184906, + "learning_rate": 0.00014557715251583205, + "loss": 0.1363, + "step": 14910 + }, + { + "epoch": 0.3495373219172864, + "grad_norm": 0.5408167243003845, + "learning_rate": 0.00014557059579865795, + "loss": 0.1263, + "step": 14911 + }, + { + "epoch": 0.3495607634920914, + "grad_norm": 0.28684353828430176, + "learning_rate": 0.0001455640388342165, + "loss": 0.0464, + "step": 14912 + }, + { + "epoch": 0.3495842050668964, + "grad_norm": 0.5032060742378235, + "learning_rate": 0.00014555748162254324, + "loss": 0.0945, + "step": 14913 + }, + { + "epoch": 0.3496076466417014, + "grad_norm": 0.44964703917503357, + "learning_rate": 0.00014555092416367385, + "loss": 0.1084, + "step": 14914 + }, + { + "epoch": 0.3496310882165064, + "grad_norm": 0.4959893524646759, + "learning_rate": 0.00014554436645764385, + "loss": 0.0791, + "step": 14915 + }, + { + "epoch": 0.3496545297913114, + "grad_norm": 0.23456989228725433, + "learning_rate": 0.00014553780850448886, + "loss": 0.0425, + "step": 14916 + }, + { + "epoch": 0.3496779713661164, + "grad_norm": 0.38572752475738525, + "learning_rate": 0.00014553125030424444, + "loss": 0.1468, + "step": 14917 + }, + { + "epoch": 0.3497014129409214, + "grad_norm": 0.23837655782699585, + "learning_rate": 0.00014552469185694618, + "loss": 0.0442, + "step": 14918 + }, + { + "epoch": 0.3497248545157264, + "grad_norm": 0.3716752827167511, + "learning_rate": 0.0001455181331626296, + "loss": 0.087, + "step": 14919 + }, + { + "epoch": 0.3497482960905314, + "grad_norm": 0.2925557494163513, + "learning_rate": 0.00014551157422133042, + "loss": 0.0248, + "step": 14920 + }, + { + "epoch": 0.34977173766533637, + "grad_norm": 0.8304538130760193, + "learning_rate": 0.00014550501503308414, + "loss": 0.1759, + "step": 14921 + }, + { + "epoch": 0.34979517924014136, + "grad_norm": 0.6160115003585815, + "learning_rate": 0.00014549845559792634, + "loss": 0.0926, + "step": 14922 + }, + { + "epoch": 0.34981862081494636, + "grad_norm": 0.6146506071090698, + "learning_rate": 0.00014549189591589267, + "loss": 0.1196, + "step": 14923 + }, + { + "epoch": 0.34984206238975135, + "grad_norm": 0.40688905119895935, + "learning_rate": 0.0001454853359870187, + "loss": 0.1063, + "step": 14924 + }, + { + "epoch": 0.34986550396455635, + "grad_norm": 0.6181090474128723, + "learning_rate": 0.00014547877581134, + "loss": 0.2256, + "step": 14925 + }, + { + "epoch": 0.34988894553936134, + "grad_norm": 0.6654247045516968, + "learning_rate": 0.00014547221538889217, + "loss": 0.1351, + "step": 14926 + }, + { + "epoch": 0.34991238711416633, + "grad_norm": 0.2442571371793747, + "learning_rate": 0.00014546565471971085, + "loss": 0.0323, + "step": 14927 + }, + { + "epoch": 0.34993582868897133, + "grad_norm": 0.6345497965812683, + "learning_rate": 0.00014545909380383164, + "loss": 0.1906, + "step": 14928 + }, + { + "epoch": 0.3499592702637763, + "grad_norm": 0.610644519329071, + "learning_rate": 0.00014545253264129003, + "loss": 0.6501, + "step": 14929 + }, + { + "epoch": 0.3499827118385813, + "grad_norm": 0.6068000197410583, + "learning_rate": 0.00014544597123212176, + "loss": 0.1015, + "step": 14930 + }, + { + "epoch": 0.3500061534133863, + "grad_norm": 0.572171151638031, + "learning_rate": 0.00014543940957636233, + "loss": 0.1624, + "step": 14931 + }, + { + "epoch": 0.3500295949881913, + "grad_norm": 0.5639334917068481, + "learning_rate": 0.0001454328476740474, + "loss": 0.1082, + "step": 14932 + }, + { + "epoch": 0.3500530365629963, + "grad_norm": 0.21385438740253448, + "learning_rate": 0.0001454262855252126, + "loss": 0.0439, + "step": 14933 + }, + { + "epoch": 0.3500764781378013, + "grad_norm": 0.8569440841674805, + "learning_rate": 0.00014541972312989346, + "loss": 0.1016, + "step": 14934 + }, + { + "epoch": 0.3500999197126063, + "grad_norm": 0.5610707402229309, + "learning_rate": 0.0001454131604881256, + "loss": 0.1227, + "step": 14935 + }, + { + "epoch": 0.3501233612874113, + "grad_norm": 0.38308990001678467, + "learning_rate": 0.0001454065975999447, + "loss": 0.1043, + "step": 14936 + }, + { + "epoch": 0.3501468028622163, + "grad_norm": 0.45667195320129395, + "learning_rate": 0.0001454000344653863, + "loss": 0.4175, + "step": 14937 + }, + { + "epoch": 0.35017024443702127, + "grad_norm": 0.3854065537452698, + "learning_rate": 0.00014539347108448604, + "loss": 0.0584, + "step": 14938 + }, + { + "epoch": 0.35019368601182627, + "grad_norm": 0.6741488575935364, + "learning_rate": 0.00014538690745727955, + "loss": 0.0942, + "step": 14939 + }, + { + "epoch": 0.35021712758663126, + "grad_norm": 0.3334841728210449, + "learning_rate": 0.0001453803435838024, + "loss": 0.2469, + "step": 14940 + }, + { + "epoch": 0.35024056916143625, + "grad_norm": 0.4865803122520447, + "learning_rate": 0.00014537377946409024, + "loss": 0.0767, + "step": 14941 + }, + { + "epoch": 0.35026401073624125, + "grad_norm": 0.1660880446434021, + "learning_rate": 0.00014536721509817867, + "loss": 0.0318, + "step": 14942 + }, + { + "epoch": 0.35028745231104624, + "grad_norm": 0.19445353746414185, + "learning_rate": 0.00014536065048610333, + "loss": 0.0293, + "step": 14943 + }, + { + "epoch": 0.35031089388585124, + "grad_norm": 0.4525719881057739, + "learning_rate": 0.00014535408562789984, + "loss": 0.0835, + "step": 14944 + }, + { + "epoch": 0.35033433546065623, + "grad_norm": 0.3527919352054596, + "learning_rate": 0.00014534752052360373, + "loss": 0.0525, + "step": 14945 + }, + { + "epoch": 0.3503577770354612, + "grad_norm": 0.35686758160591125, + "learning_rate": 0.00014534095517325076, + "loss": 0.0633, + "step": 14946 + }, + { + "epoch": 0.3503812186102662, + "grad_norm": 0.6165280342102051, + "learning_rate": 0.0001453343895768765, + "loss": 0.8517, + "step": 14947 + }, + { + "epoch": 0.3504046601850712, + "grad_norm": 0.7579771280288696, + "learning_rate": 0.00014532782373451655, + "loss": 0.1209, + "step": 14948 + }, + { + "epoch": 0.3504281017598762, + "grad_norm": 0.20382286608219147, + "learning_rate": 0.00014532125764620655, + "loss": 0.0426, + "step": 14949 + }, + { + "epoch": 0.3504515433346812, + "grad_norm": 0.43105262517929077, + "learning_rate": 0.00014531469131198214, + "loss": 0.0816, + "step": 14950 + }, + { + "epoch": 0.3504749849094862, + "grad_norm": 0.46958082914352417, + "learning_rate": 0.0001453081247318789, + "loss": 0.0962, + "step": 14951 + }, + { + "epoch": 0.3504984264842912, + "grad_norm": 0.11233362555503845, + "learning_rate": 0.00014530155790593255, + "loss": 0.0303, + "step": 14952 + }, + { + "epoch": 0.3505218680590962, + "grad_norm": 0.4129209518432617, + "learning_rate": 0.0001452949908341787, + "loss": 0.1137, + "step": 14953 + }, + { + "epoch": 0.3505453096339012, + "grad_norm": 0.5102824568748474, + "learning_rate": 0.0001452884235166529, + "loss": 0.1225, + "step": 14954 + }, + { + "epoch": 0.3505687512087062, + "grad_norm": 0.8736830949783325, + "learning_rate": 0.00014528185595339083, + "loss": 0.2069, + "step": 14955 + }, + { + "epoch": 0.35059219278351117, + "grad_norm": 0.22834132611751556, + "learning_rate": 0.00014527528814442818, + "loss": 0.0529, + "step": 14956 + }, + { + "epoch": 0.3506156343583162, + "grad_norm": 0.8172510862350464, + "learning_rate": 0.00014526872008980052, + "loss": 0.1692, + "step": 14957 + }, + { + "epoch": 0.3506390759331212, + "grad_norm": 0.40477892756462097, + "learning_rate": 0.00014526215178954353, + "loss": 0.1114, + "step": 14958 + }, + { + "epoch": 0.3506625175079262, + "grad_norm": 0.2562006711959839, + "learning_rate": 0.00014525558324369285, + "loss": 0.0397, + "step": 14959 + }, + { + "epoch": 0.3506859590827312, + "grad_norm": 0.6897245645523071, + "learning_rate": 0.00014524901445228406, + "loss": 0.1442, + "step": 14960 + }, + { + "epoch": 0.3507094006575362, + "grad_norm": 0.4898776113986969, + "learning_rate": 0.00014524244541535284, + "loss": 0.1143, + "step": 14961 + }, + { + "epoch": 0.3507328422323412, + "grad_norm": 0.39861050248146057, + "learning_rate": 0.00014523587613293488, + "loss": 0.0702, + "step": 14962 + }, + { + "epoch": 0.3507562838071462, + "grad_norm": 0.687713086605072, + "learning_rate": 0.00014522930660506575, + "loss": 0.1488, + "step": 14963 + }, + { + "epoch": 0.3507797253819512, + "grad_norm": 1.1389527320861816, + "learning_rate": 0.00014522273683178114, + "loss": 0.1543, + "step": 14964 + }, + { + "epoch": 0.3508031669567562, + "grad_norm": 0.2202020138502121, + "learning_rate": 0.0001452161668131167, + "loss": 0.0577, + "step": 14965 + }, + { + "epoch": 0.35082660853156117, + "grad_norm": 0.39092159271240234, + "learning_rate": 0.00014520959654910806, + "loss": 0.0837, + "step": 14966 + }, + { + "epoch": 0.35085005010636616, + "grad_norm": 0.6378992795944214, + "learning_rate": 0.00014520302603979085, + "loss": 0.1126, + "step": 14967 + }, + { + "epoch": 0.35087349168117116, + "grad_norm": 0.373457670211792, + "learning_rate": 0.00014519645528520079, + "loss": 0.1037, + "step": 14968 + }, + { + "epoch": 0.35089693325597615, + "grad_norm": 0.3851023018360138, + "learning_rate": 0.0001451898842853735, + "loss": 0.1298, + "step": 14969 + }, + { + "epoch": 0.35092037483078115, + "grad_norm": 0.658237636089325, + "learning_rate": 0.00014518331304034458, + "loss": 0.069, + "step": 14970 + }, + { + "epoch": 0.35094381640558614, + "grad_norm": 0.24463576078414917, + "learning_rate": 0.00014517674155014977, + "loss": 0.035, + "step": 14971 + }, + { + "epoch": 0.35096725798039113, + "grad_norm": 0.13629500567913055, + "learning_rate": 0.00014517016981482469, + "loss": 0.0393, + "step": 14972 + }, + { + "epoch": 0.35099069955519613, + "grad_norm": 0.3637765049934387, + "learning_rate": 0.000145163597834405, + "loss": 0.0512, + "step": 14973 + }, + { + "epoch": 0.3510141411300011, + "grad_norm": 0.6630548238754272, + "learning_rate": 0.00014515702560892634, + "loss": 0.7769, + "step": 14974 + }, + { + "epoch": 0.3510375827048061, + "grad_norm": 0.7984098196029663, + "learning_rate": 0.00014515045313842438, + "loss": 0.1584, + "step": 14975 + }, + { + "epoch": 0.3510610242796111, + "grad_norm": 0.4255906939506531, + "learning_rate": 0.0001451438804229348, + "loss": 0.1113, + "step": 14976 + }, + { + "epoch": 0.3510844658544161, + "grad_norm": 0.4607278108596802, + "learning_rate": 0.00014513730746249328, + "loss": 0.0878, + "step": 14977 + }, + { + "epoch": 0.3511079074292211, + "grad_norm": 0.4606483280658722, + "learning_rate": 0.00014513073425713544, + "loss": 0.1095, + "step": 14978 + }, + { + "epoch": 0.3511313490040261, + "grad_norm": 0.17898257076740265, + "learning_rate": 0.00014512416080689696, + "loss": 0.0501, + "step": 14979 + }, + { + "epoch": 0.3511547905788311, + "grad_norm": 0.47848275303840637, + "learning_rate": 0.0001451175871118135, + "loss": 0.0916, + "step": 14980 + }, + { + "epoch": 0.3511782321536361, + "grad_norm": 0.683998167514801, + "learning_rate": 0.0001451110131719208, + "loss": 0.6308, + "step": 14981 + }, + { + "epoch": 0.3512016737284411, + "grad_norm": 0.3625466823577881, + "learning_rate": 0.00014510443898725446, + "loss": 0.0685, + "step": 14982 + }, + { + "epoch": 0.35122511530324607, + "grad_norm": 0.5219454169273376, + "learning_rate": 0.00014509786455785013, + "loss": 0.1179, + "step": 14983 + }, + { + "epoch": 0.35124855687805107, + "grad_norm": 0.3258061110973358, + "learning_rate": 0.00014509128988374355, + "loss": 0.0622, + "step": 14984 + }, + { + "epoch": 0.35127199845285606, + "grad_norm": 0.4949590861797333, + "learning_rate": 0.00014508471496497036, + "loss": 0.1567, + "step": 14985 + }, + { + "epoch": 0.35129544002766105, + "grad_norm": 0.1975620836019516, + "learning_rate": 0.0001450781398015662, + "loss": 0.0465, + "step": 14986 + }, + { + "epoch": 0.35131888160246605, + "grad_norm": 0.49355682730674744, + "learning_rate": 0.00014507156439356683, + "loss": 0.1159, + "step": 14987 + }, + { + "epoch": 0.35134232317727104, + "grad_norm": 0.4227064549922943, + "learning_rate": 0.00014506498874100786, + "loss": 0.1148, + "step": 14988 + }, + { + "epoch": 0.35136576475207604, + "grad_norm": 0.7655077576637268, + "learning_rate": 0.000145058412843925, + "loss": 0.0935, + "step": 14989 + }, + { + "epoch": 0.35138920632688103, + "grad_norm": 0.6353089213371277, + "learning_rate": 0.00014505183670235393, + "loss": 0.137, + "step": 14990 + }, + { + "epoch": 0.351412647901686, + "grad_norm": 0.21057382225990295, + "learning_rate": 0.00014504526031633037, + "loss": 0.0173, + "step": 14991 + }, + { + "epoch": 0.351436089476491, + "grad_norm": 0.36299774050712585, + "learning_rate": 0.0001450386836858899, + "loss": 0.0754, + "step": 14992 + }, + { + "epoch": 0.351459531051296, + "grad_norm": 0.5503971576690674, + "learning_rate": 0.00014503210681106828, + "loss": 0.0859, + "step": 14993 + }, + { + "epoch": 0.351482972626101, + "grad_norm": 0.5711585879325867, + "learning_rate": 0.00014502552969190117, + "loss": 0.1004, + "step": 14994 + }, + { + "epoch": 0.351506414200906, + "grad_norm": 0.5020803809165955, + "learning_rate": 0.00014501895232842428, + "loss": 0.1736, + "step": 14995 + }, + { + "epoch": 0.351529855775711, + "grad_norm": 0.5942709445953369, + "learning_rate": 0.0001450123747206733, + "loss": 0.1443, + "step": 14996 + }, + { + "epoch": 0.351553297350516, + "grad_norm": 0.14312687516212463, + "learning_rate": 0.0001450057968686839, + "loss": 0.0407, + "step": 14997 + }, + { + "epoch": 0.351576738925321, + "grad_norm": 0.8024514317512512, + "learning_rate": 0.00014499921877249178, + "loss": 0.1996, + "step": 14998 + }, + { + "epoch": 0.351600180500126, + "grad_norm": 0.36313119530677795, + "learning_rate": 0.00014499264043213263, + "loss": 0.0783, + "step": 14999 + }, + { + "epoch": 0.351623622074931, + "grad_norm": 0.37743738293647766, + "learning_rate": 0.00014498606184764219, + "loss": 0.0899, + "step": 15000 + }, + { + "epoch": 0.35164706364973597, + "grad_norm": 0.4134349822998047, + "learning_rate": 0.0001449794830190561, + "loss": 0.1351, + "step": 15001 + }, + { + "epoch": 0.35167050522454096, + "grad_norm": 0.49805742502212524, + "learning_rate": 0.00014497290394641001, + "loss": 0.0634, + "step": 15002 + }, + { + "epoch": 0.35169394679934596, + "grad_norm": 1.915297269821167, + "learning_rate": 0.00014496632462973974, + "loss": 0.0889, + "step": 15003 + }, + { + "epoch": 0.35171738837415095, + "grad_norm": 0.8589234352111816, + "learning_rate": 0.0001449597450690809, + "loss": 0.7042, + "step": 15004 + }, + { + "epoch": 0.35174082994895595, + "grad_norm": 0.09460244327783585, + "learning_rate": 0.0001449531652644692, + "loss": 0.0278, + "step": 15005 + }, + { + "epoch": 0.35176427152376094, + "grad_norm": 0.6466801762580872, + "learning_rate": 0.0001449465852159404, + "loss": 0.1703, + "step": 15006 + }, + { + "epoch": 0.35178771309856594, + "grad_norm": 0.5613182783126831, + "learning_rate": 0.00014494000492353018, + "loss": 0.1103, + "step": 15007 + }, + { + "epoch": 0.35181115467337093, + "grad_norm": 0.3918820917606354, + "learning_rate": 0.00014493342438727419, + "loss": 0.1043, + "step": 15008 + }, + { + "epoch": 0.351834596248176, + "grad_norm": 0.3726392090320587, + "learning_rate": 0.00014492684360720818, + "loss": 0.0895, + "step": 15009 + }, + { + "epoch": 0.351858037822981, + "grad_norm": 0.7329981327056885, + "learning_rate": 0.00014492026258336785, + "loss": 0.1061, + "step": 15010 + }, + { + "epoch": 0.35188147939778597, + "grad_norm": 0.3305954039096832, + "learning_rate": 0.00014491368131578893, + "loss": 0.0434, + "step": 15011 + }, + { + "epoch": 0.35190492097259096, + "grad_norm": 0.36958053708076477, + "learning_rate": 0.0001449070998045071, + "loss": 0.0922, + "step": 15012 + }, + { + "epoch": 0.35192836254739596, + "grad_norm": 0.6081244349479675, + "learning_rate": 0.00014490051804955808, + "loss": 0.0531, + "step": 15013 + }, + { + "epoch": 0.35195180412220095, + "grad_norm": 1.0978232622146606, + "learning_rate": 0.0001448939360509776, + "loss": 0.1571, + "step": 15014 + }, + { + "epoch": 0.35197524569700595, + "grad_norm": 0.22884586453437805, + "learning_rate": 0.00014488735380880136, + "loss": 0.0358, + "step": 15015 + }, + { + "epoch": 0.35199868727181094, + "grad_norm": 0.5532789826393127, + "learning_rate": 0.00014488077132306508, + "loss": 0.151, + "step": 15016 + }, + { + "epoch": 0.35202212884661593, + "grad_norm": 0.23709040880203247, + "learning_rate": 0.00014487418859380447, + "loss": 0.0488, + "step": 15017 + }, + { + "epoch": 0.35204557042142093, + "grad_norm": 0.1980396956205368, + "learning_rate": 0.00014486760562105523, + "loss": 0.0668, + "step": 15018 + }, + { + "epoch": 0.3520690119962259, + "grad_norm": 0.22290416061878204, + "learning_rate": 0.00014486102240485313, + "loss": 0.0245, + "step": 15019 + }, + { + "epoch": 0.3520924535710309, + "grad_norm": 0.24652379751205444, + "learning_rate": 0.00014485443894523384, + "loss": 0.0449, + "step": 15020 + }, + { + "epoch": 0.3521158951458359, + "grad_norm": 0.5312551856040955, + "learning_rate": 0.00014484785524223313, + "loss": 0.1073, + "step": 15021 + }, + { + "epoch": 0.3521393367206409, + "grad_norm": 1.405056357383728, + "learning_rate": 0.00014484127129588665, + "loss": 0.1574, + "step": 15022 + }, + { + "epoch": 0.3521627782954459, + "grad_norm": 0.462545245885849, + "learning_rate": 0.00014483468710623022, + "loss": 0.0496, + "step": 15023 + }, + { + "epoch": 0.3521862198702509, + "grad_norm": 0.1942295879125595, + "learning_rate": 0.00014482810267329947, + "loss": 0.0354, + "step": 15024 + }, + { + "epoch": 0.3522096614450559, + "grad_norm": 0.39287158846855164, + "learning_rate": 0.00014482151799713022, + "loss": 0.1205, + "step": 15025 + }, + { + "epoch": 0.3522331030198609, + "grad_norm": 0.4900822639465332, + "learning_rate": 0.00014481493307775812, + "loss": 0.1271, + "step": 15026 + }, + { + "epoch": 0.3522565445946659, + "grad_norm": 0.2629556357860565, + "learning_rate": 0.00014480834791521895, + "loss": 0.0661, + "step": 15027 + }, + { + "epoch": 0.35227998616947087, + "grad_norm": 0.3167760968208313, + "learning_rate": 0.00014480176250954843, + "loss": 0.0485, + "step": 15028 + }, + { + "epoch": 0.35230342774427587, + "grad_norm": 0.7902346849441528, + "learning_rate": 0.00014479517686078228, + "loss": 0.1998, + "step": 15029 + }, + { + "epoch": 0.35232686931908086, + "grad_norm": 0.13850203156471252, + "learning_rate": 0.00014478859096895622, + "loss": 0.0188, + "step": 15030 + }, + { + "epoch": 0.35235031089388585, + "grad_norm": 0.3849312365055084, + "learning_rate": 0.00014478200483410603, + "loss": 0.1205, + "step": 15031 + }, + { + "epoch": 0.35237375246869085, + "grad_norm": 0.5117472410202026, + "learning_rate": 0.00014477541845626742, + "loss": 0.0672, + "step": 15032 + }, + { + "epoch": 0.35239719404349584, + "grad_norm": 0.16560831665992737, + "learning_rate": 0.0001447688318354761, + "loss": 0.0156, + "step": 15033 + }, + { + "epoch": 0.35242063561830084, + "grad_norm": 0.5052921175956726, + "learning_rate": 0.00014476224497176788, + "loss": 0.0672, + "step": 15034 + }, + { + "epoch": 0.35244407719310583, + "grad_norm": 0.444282591342926, + "learning_rate": 0.00014475565786517845, + "loss": 0.0906, + "step": 15035 + }, + { + "epoch": 0.3524675187679108, + "grad_norm": 1.126820683479309, + "learning_rate": 0.0001447490705157436, + "loss": 0.2096, + "step": 15036 + }, + { + "epoch": 0.3524909603427158, + "grad_norm": 0.496479332447052, + "learning_rate": 0.00014474248292349896, + "loss": 0.1547, + "step": 15037 + }, + { + "epoch": 0.3525144019175208, + "grad_norm": 0.7787871956825256, + "learning_rate": 0.0001447358950884804, + "loss": 0.7415, + "step": 15038 + }, + { + "epoch": 0.3525378434923258, + "grad_norm": 0.5764645338058472, + "learning_rate": 0.0001447293070107236, + "loss": 0.7656, + "step": 15039 + }, + { + "epoch": 0.3525612850671308, + "grad_norm": 0.538112223148346, + "learning_rate": 0.0001447227186902643, + "loss": 0.111, + "step": 15040 + }, + { + "epoch": 0.3525847266419358, + "grad_norm": 0.48761889338493347, + "learning_rate": 0.0001447161301271383, + "loss": 0.1435, + "step": 15041 + }, + { + "epoch": 0.3526081682167408, + "grad_norm": 0.5019400715827942, + "learning_rate": 0.00014470954132138133, + "loss": 0.1386, + "step": 15042 + }, + { + "epoch": 0.3526316097915458, + "grad_norm": 0.568202793598175, + "learning_rate": 0.0001447029522730291, + "loss": 0.5422, + "step": 15043 + }, + { + "epoch": 0.3526550513663508, + "grad_norm": 0.45571279525756836, + "learning_rate": 0.0001446963629821174, + "loss": 0.0617, + "step": 15044 + }, + { + "epoch": 0.3526784929411558, + "grad_norm": 0.20191581547260284, + "learning_rate": 0.000144689773448682, + "loss": 0.0505, + "step": 15045 + }, + { + "epoch": 0.35270193451596077, + "grad_norm": 0.2602747678756714, + "learning_rate": 0.0001446831836727586, + "loss": 0.0522, + "step": 15046 + }, + { + "epoch": 0.35272537609076576, + "grad_norm": 0.25920939445495605, + "learning_rate": 0.00014467659365438303, + "loss": 0.0335, + "step": 15047 + }, + { + "epoch": 0.35274881766557076, + "grad_norm": 0.7053400278091431, + "learning_rate": 0.00014467000339359097, + "loss": 0.1657, + "step": 15048 + }, + { + "epoch": 0.35277225924037575, + "grad_norm": 0.48826172947883606, + "learning_rate": 0.00014466341289041826, + "loss": 0.0626, + "step": 15049 + }, + { + "epoch": 0.35279570081518075, + "grad_norm": 0.837035596370697, + "learning_rate": 0.00014465682214490059, + "loss": 0.2786, + "step": 15050 + }, + { + "epoch": 0.35281914238998574, + "grad_norm": 0.6012053489685059, + "learning_rate": 0.00014465023115707373, + "loss": 0.2174, + "step": 15051 + }, + { + "epoch": 0.35284258396479073, + "grad_norm": 0.21462984383106232, + "learning_rate": 0.00014464363992697346, + "loss": 0.0553, + "step": 15052 + }, + { + "epoch": 0.35286602553959573, + "grad_norm": 0.2123178392648697, + "learning_rate": 0.00014463704845463554, + "loss": 0.0332, + "step": 15053 + }, + { + "epoch": 0.3528894671144007, + "grad_norm": 0.25788310170173645, + "learning_rate": 0.00014463045674009578, + "loss": 0.0515, + "step": 15054 + }, + { + "epoch": 0.3529129086892057, + "grad_norm": 0.40029576420783997, + "learning_rate": 0.0001446238647833899, + "loss": 0.0859, + "step": 15055 + }, + { + "epoch": 0.3529363502640107, + "grad_norm": 0.21088507771492004, + "learning_rate": 0.00014461727258455363, + "loss": 0.0449, + "step": 15056 + }, + { + "epoch": 0.3529597918388157, + "grad_norm": 0.16843441128730774, + "learning_rate": 0.00014461068014362286, + "loss": 0.0475, + "step": 15057 + }, + { + "epoch": 0.3529832334136207, + "grad_norm": 0.5140364766120911, + "learning_rate": 0.00014460408746063322, + "loss": 0.0462, + "step": 15058 + }, + { + "epoch": 0.3530066749884257, + "grad_norm": 0.6061533093452454, + "learning_rate": 0.00014459749453562054, + "loss": 0.1151, + "step": 15059 + }, + { + "epoch": 0.35303011656323074, + "grad_norm": 0.31631308794021606, + "learning_rate": 0.00014459090136862064, + "loss": 0.0918, + "step": 15060 + }, + { + "epoch": 0.35305355813803574, + "grad_norm": 0.7017638683319092, + "learning_rate": 0.00014458430795966925, + "loss": 0.1525, + "step": 15061 + }, + { + "epoch": 0.35307699971284073, + "grad_norm": 0.3284108638763428, + "learning_rate": 0.00014457771430880214, + "loss": 0.0876, + "step": 15062 + }, + { + "epoch": 0.3531004412876457, + "grad_norm": 0.2371806800365448, + "learning_rate": 0.0001445711204160551, + "loss": 0.0506, + "step": 15063 + }, + { + "epoch": 0.3531238828624507, + "grad_norm": 0.38791581988334656, + "learning_rate": 0.00014456452628146393, + "loss": 0.0874, + "step": 15064 + }, + { + "epoch": 0.3531473244372557, + "grad_norm": 0.3697226643562317, + "learning_rate": 0.00014455793190506437, + "loss": 0.0897, + "step": 15065 + }, + { + "epoch": 0.3531707660120607, + "grad_norm": 0.3020305633544922, + "learning_rate": 0.00014455133728689221, + "loss": 0.3383, + "step": 15066 + }, + { + "epoch": 0.3531942075868657, + "grad_norm": 0.5164265036582947, + "learning_rate": 0.00014454474242698323, + "loss": 0.0783, + "step": 15067 + }, + { + "epoch": 0.3532176491616707, + "grad_norm": 0.6772429347038269, + "learning_rate": 0.00014453814732537327, + "loss": 0.7959, + "step": 15068 + }, + { + "epoch": 0.3532410907364757, + "grad_norm": 0.462376207113266, + "learning_rate": 0.00014453155198209804, + "loss": 0.0648, + "step": 15069 + }, + { + "epoch": 0.3532645323112807, + "grad_norm": 0.19377577304840088, + "learning_rate": 0.00014452495639719336, + "loss": 0.0445, + "step": 15070 + }, + { + "epoch": 0.3532879738860857, + "grad_norm": 0.306883841753006, + "learning_rate": 0.00014451836057069502, + "loss": 0.0776, + "step": 15071 + }, + { + "epoch": 0.3533114154608907, + "grad_norm": 0.3280450999736786, + "learning_rate": 0.00014451176450263877, + "loss": 0.0719, + "step": 15072 + }, + { + "epoch": 0.35333485703569567, + "grad_norm": 0.9637665748596191, + "learning_rate": 0.00014450516819306049, + "loss": 0.158, + "step": 15073 + }, + { + "epoch": 0.35335829861050067, + "grad_norm": 0.5720147490501404, + "learning_rate": 0.0001444985716419959, + "loss": 0.1028, + "step": 15074 + }, + { + "epoch": 0.35338174018530566, + "grad_norm": 0.17266836762428284, + "learning_rate": 0.00014449197484948077, + "loss": 0.042, + "step": 15075 + }, + { + "epoch": 0.35340518176011065, + "grad_norm": 0.6636739373207092, + "learning_rate": 0.00014448537781555098, + "loss": 0.8143, + "step": 15076 + }, + { + "epoch": 0.35342862333491565, + "grad_norm": 0.21905548870563507, + "learning_rate": 0.00014447878054024224, + "loss": 0.0573, + "step": 15077 + }, + { + "epoch": 0.35345206490972064, + "grad_norm": 0.4776175916194916, + "learning_rate": 0.0001444721830235904, + "loss": 0.0803, + "step": 15078 + }, + { + "epoch": 0.35347550648452564, + "grad_norm": 0.6038962602615356, + "learning_rate": 0.00014446558526563127, + "loss": 0.0945, + "step": 15079 + }, + { + "epoch": 0.35349894805933063, + "grad_norm": 0.6438377499580383, + "learning_rate": 0.0001444589872664006, + "loss": 0.8557, + "step": 15080 + }, + { + "epoch": 0.3535223896341356, + "grad_norm": 0.5257653594017029, + "learning_rate": 0.0001444523890259342, + "loss": 0.0777, + "step": 15081 + }, + { + "epoch": 0.3535458312089406, + "grad_norm": 0.3911442160606384, + "learning_rate": 0.00014444579054426795, + "loss": 0.5123, + "step": 15082 + }, + { + "epoch": 0.3535692727837456, + "grad_norm": 0.8575892448425293, + "learning_rate": 0.00014443919182143754, + "loss": 0.1896, + "step": 15083 + }, + { + "epoch": 0.3535927143585506, + "grad_norm": 0.5837143659591675, + "learning_rate": 0.00014443259285747882, + "loss": 0.1105, + "step": 15084 + }, + { + "epoch": 0.3536161559333556, + "grad_norm": 0.7363370656967163, + "learning_rate": 0.00014442599365242762, + "loss": 0.1003, + "step": 15085 + }, + { + "epoch": 0.3536395975081606, + "grad_norm": 0.5548307299613953, + "learning_rate": 0.00014441939420631973, + "loss": 0.0671, + "step": 15086 + }, + { + "epoch": 0.3536630390829656, + "grad_norm": 0.30014532804489136, + "learning_rate": 0.00014441279451919096, + "loss": 0.1091, + "step": 15087 + }, + { + "epoch": 0.3536864806577706, + "grad_norm": 0.4080555737018585, + "learning_rate": 0.0001444061945910771, + "loss": 0.1058, + "step": 15088 + }, + { + "epoch": 0.3537099222325756, + "grad_norm": 0.3373178541660309, + "learning_rate": 0.000144399594422014, + "loss": 0.072, + "step": 15089 + }, + { + "epoch": 0.3537333638073806, + "grad_norm": 0.9845951795578003, + "learning_rate": 0.00014439299401203742, + "loss": 0.2259, + "step": 15090 + }, + { + "epoch": 0.35375680538218557, + "grad_norm": 0.08338478207588196, + "learning_rate": 0.00014438639336118324, + "loss": 0.0107, + "step": 15091 + }, + { + "epoch": 0.35378024695699056, + "grad_norm": 0.43270963430404663, + "learning_rate": 0.00014437979246948723, + "loss": 0.1022, + "step": 15092 + }, + { + "epoch": 0.35380368853179556, + "grad_norm": 1.0165185928344727, + "learning_rate": 0.00014437319133698523, + "loss": 0.1101, + "step": 15093 + }, + { + "epoch": 0.35382713010660055, + "grad_norm": 0.43936195969581604, + "learning_rate": 0.00014436658996371305, + "loss": 0.0856, + "step": 15094 + }, + { + "epoch": 0.35385057168140555, + "grad_norm": 0.28952452540397644, + "learning_rate": 0.00014435998834970646, + "loss": 0.0575, + "step": 15095 + }, + { + "epoch": 0.35387401325621054, + "grad_norm": 1.4952658414840698, + "learning_rate": 0.00014435338649500137, + "loss": 0.2002, + "step": 15096 + }, + { + "epoch": 0.35389745483101553, + "grad_norm": 0.4972577393054962, + "learning_rate": 0.00014434678439963355, + "loss": 0.1017, + "step": 15097 + }, + { + "epoch": 0.35392089640582053, + "grad_norm": 0.401324987411499, + "learning_rate": 0.0001443401820636388, + "loss": 0.121, + "step": 15098 + }, + { + "epoch": 0.3539443379806255, + "grad_norm": 0.12440859526395798, + "learning_rate": 0.000144333579487053, + "loss": 0.0182, + "step": 15099 + }, + { + "epoch": 0.3539677795554305, + "grad_norm": 0.21510227024555206, + "learning_rate": 0.00014432697666991198, + "loss": 0.0389, + "step": 15100 + }, + { + "epoch": 0.3539912211302355, + "grad_norm": 0.6273371577262878, + "learning_rate": 0.0001443203736122515, + "loss": 0.0646, + "step": 15101 + }, + { + "epoch": 0.3540146627050405, + "grad_norm": 0.11809872835874557, + "learning_rate": 0.00014431377031410745, + "loss": 0.0162, + "step": 15102 + }, + { + "epoch": 0.3540381042798455, + "grad_norm": 0.5180671215057373, + "learning_rate": 0.0001443071667755156, + "loss": 0.1539, + "step": 15103 + }, + { + "epoch": 0.3540615458546505, + "grad_norm": 0.7176443934440613, + "learning_rate": 0.00014430056299651183, + "loss": 0.6413, + "step": 15104 + }, + { + "epoch": 0.3540849874294555, + "grad_norm": 0.637573778629303, + "learning_rate": 0.000144293958977132, + "loss": 0.1307, + "step": 15105 + }, + { + "epoch": 0.3541084290042605, + "grad_norm": 0.3324727416038513, + "learning_rate": 0.00014428735471741185, + "loss": 0.067, + "step": 15106 + }, + { + "epoch": 0.3541318705790655, + "grad_norm": 0.28656530380249023, + "learning_rate": 0.00014428075021738728, + "loss": 0.0516, + "step": 15107 + }, + { + "epoch": 0.35415531215387047, + "grad_norm": 0.27497589588165283, + "learning_rate": 0.00014427414547709414, + "loss": 0.0731, + "step": 15108 + }, + { + "epoch": 0.35417875372867547, + "grad_norm": 0.4901493787765503, + "learning_rate": 0.00014426754049656824, + "loss": 0.1388, + "step": 15109 + }, + { + "epoch": 0.35420219530348046, + "grad_norm": 0.6494714021682739, + "learning_rate": 0.0001442609352758454, + "loss": 0.1201, + "step": 15110 + }, + { + "epoch": 0.3542256368782855, + "grad_norm": 0.5663119554519653, + "learning_rate": 0.0001442543298149615, + "loss": 0.1701, + "step": 15111 + }, + { + "epoch": 0.3542490784530905, + "grad_norm": 0.6380974650382996, + "learning_rate": 0.00014424772411395236, + "loss": 0.6965, + "step": 15112 + }, + { + "epoch": 0.3542725200278955, + "grad_norm": 0.3207346796989441, + "learning_rate": 0.0001442411181728538, + "loss": 0.0635, + "step": 15113 + }, + { + "epoch": 0.3542959616027005, + "grad_norm": 0.5910550355911255, + "learning_rate": 0.0001442345119917017, + "loss": 0.0634, + "step": 15114 + }, + { + "epoch": 0.3543194031775055, + "grad_norm": 0.19789020717144012, + "learning_rate": 0.00014422790557053192, + "loss": 0.0469, + "step": 15115 + }, + { + "epoch": 0.3543428447523105, + "grad_norm": 0.38305193185806274, + "learning_rate": 0.00014422129890938023, + "loss": 0.1208, + "step": 15116 + }, + { + "epoch": 0.3543662863271155, + "grad_norm": 0.40153953433036804, + "learning_rate": 0.00014421469200828256, + "loss": 0.0482, + "step": 15117 + }, + { + "epoch": 0.35438972790192047, + "grad_norm": 0.44134101271629333, + "learning_rate": 0.00014420808486727475, + "loss": 0.1005, + "step": 15118 + }, + { + "epoch": 0.35441316947672546, + "grad_norm": 0.5363313555717468, + "learning_rate": 0.0001442014774863926, + "loss": 0.1524, + "step": 15119 + }, + { + "epoch": 0.35443661105153046, + "grad_norm": 0.1467013657093048, + "learning_rate": 0.00014419486986567197, + "loss": 0.0229, + "step": 15120 + }, + { + "epoch": 0.35446005262633545, + "grad_norm": 0.18747682869434357, + "learning_rate": 0.00014418826200514876, + "loss": 0.0427, + "step": 15121 + }, + { + "epoch": 0.35448349420114045, + "grad_norm": 0.5515857934951782, + "learning_rate": 0.00014418165390485882, + "loss": 0.1052, + "step": 15122 + }, + { + "epoch": 0.35450693577594544, + "grad_norm": 0.412339985370636, + "learning_rate": 0.00014417504556483795, + "loss": 0.1151, + "step": 15123 + }, + { + "epoch": 0.35453037735075044, + "grad_norm": 0.3309311866760254, + "learning_rate": 0.00014416843698512204, + "loss": 0.0443, + "step": 15124 + }, + { + "epoch": 0.35455381892555543, + "grad_norm": 0.5518073439598083, + "learning_rate": 0.00014416182816574696, + "loss": 0.1441, + "step": 15125 + }, + { + "epoch": 0.3545772605003604, + "grad_norm": 0.1439562439918518, + "learning_rate": 0.00014415521910674856, + "loss": 0.0368, + "step": 15126 + }, + { + "epoch": 0.3546007020751654, + "grad_norm": 0.7292405366897583, + "learning_rate": 0.0001441486098081627, + "loss": 0.1247, + "step": 15127 + }, + { + "epoch": 0.3546241436499704, + "grad_norm": 0.21444235742092133, + "learning_rate": 0.00014414200027002526, + "loss": 0.0314, + "step": 15128 + }, + { + "epoch": 0.3546475852247754, + "grad_norm": 0.33850032091140747, + "learning_rate": 0.00014413539049237204, + "loss": 0.087, + "step": 15129 + }, + { + "epoch": 0.3546710267995804, + "grad_norm": 0.5582752227783203, + "learning_rate": 0.000144128780475239, + "loss": 0.0945, + "step": 15130 + }, + { + "epoch": 0.3546944683743854, + "grad_norm": 1.1324026584625244, + "learning_rate": 0.00014412217021866192, + "loss": 0.1132, + "step": 15131 + }, + { + "epoch": 0.3547179099491904, + "grad_norm": 0.46830660104751587, + "learning_rate": 0.0001441155597226767, + "loss": 0.098, + "step": 15132 + }, + { + "epoch": 0.3547413515239954, + "grad_norm": 0.23089861869812012, + "learning_rate": 0.00014410894898731922, + "loss": 0.0452, + "step": 15133 + }, + { + "epoch": 0.3547647930988004, + "grad_norm": 0.4310033321380615, + "learning_rate": 0.00014410233801262536, + "loss": 0.0769, + "step": 15134 + }, + { + "epoch": 0.3547882346736054, + "grad_norm": 0.16181951761245728, + "learning_rate": 0.00014409572679863097, + "loss": 0.0306, + "step": 15135 + }, + { + "epoch": 0.35481167624841037, + "grad_norm": 0.513137698173523, + "learning_rate": 0.00014408911534537192, + "loss": 0.0643, + "step": 15136 + }, + { + "epoch": 0.35483511782321536, + "grad_norm": 0.22616297006607056, + "learning_rate": 0.00014408250365288412, + "loss": 0.0269, + "step": 15137 + }, + { + "epoch": 0.35485855939802036, + "grad_norm": 0.4556826651096344, + "learning_rate": 0.00014407589172120338, + "loss": 0.0732, + "step": 15138 + }, + { + "epoch": 0.35488200097282535, + "grad_norm": 0.43222254514694214, + "learning_rate": 0.0001440692795503656, + "loss": 0.0631, + "step": 15139 + }, + { + "epoch": 0.35490544254763035, + "grad_norm": 0.4588768184185028, + "learning_rate": 0.0001440626671404067, + "loss": 0.0955, + "step": 15140 + }, + { + "epoch": 0.35492888412243534, + "grad_norm": 0.5594522356987, + "learning_rate": 0.00014405605449136253, + "loss": 0.1426, + "step": 15141 + }, + { + "epoch": 0.35495232569724033, + "grad_norm": 0.4619494080543518, + "learning_rate": 0.00014404944160326894, + "loss": 0.0923, + "step": 15142 + }, + { + "epoch": 0.35497576727204533, + "grad_norm": 0.6746577620506287, + "learning_rate": 0.00014404282847616187, + "loss": 0.1057, + "step": 15143 + }, + { + "epoch": 0.3549992088468503, + "grad_norm": 0.7930972576141357, + "learning_rate": 0.00014403621511007713, + "loss": 0.1862, + "step": 15144 + }, + { + "epoch": 0.3550226504216553, + "grad_norm": 0.9337666034698486, + "learning_rate": 0.00014402960150505068, + "loss": 0.1754, + "step": 15145 + }, + { + "epoch": 0.3550460919964603, + "grad_norm": 0.5665156841278076, + "learning_rate": 0.0001440229876611184, + "loss": 0.0873, + "step": 15146 + }, + { + "epoch": 0.3550695335712653, + "grad_norm": 0.42232322692871094, + "learning_rate": 0.00014401637357831613, + "loss": 0.0964, + "step": 15147 + }, + { + "epoch": 0.3550929751460703, + "grad_norm": 0.11331086605787277, + "learning_rate": 0.00014400975925667975, + "loss": 0.0298, + "step": 15148 + }, + { + "epoch": 0.3551164167208753, + "grad_norm": 0.7162601351737976, + "learning_rate": 0.00014400314469624522, + "loss": 0.6966, + "step": 15149 + }, + { + "epoch": 0.3551398582956803, + "grad_norm": 1.0760544538497925, + "learning_rate": 0.00014399652989704838, + "loss": 0.2022, + "step": 15150 + }, + { + "epoch": 0.3551632998704853, + "grad_norm": 0.40971285104751587, + "learning_rate": 0.0001439899148591251, + "loss": 0.0692, + "step": 15151 + }, + { + "epoch": 0.3551867414452903, + "grad_norm": 0.6567769646644592, + "learning_rate": 0.00014398329958251137, + "loss": 0.1927, + "step": 15152 + }, + { + "epoch": 0.35521018302009527, + "grad_norm": 0.49205124378204346, + "learning_rate": 0.000143976684067243, + "loss": 0.1139, + "step": 15153 + }, + { + "epoch": 0.35523362459490027, + "grad_norm": 0.6298203468322754, + "learning_rate": 0.00014397006831335583, + "loss": 0.0967, + "step": 15154 + }, + { + "epoch": 0.35525706616970526, + "grad_norm": 0.5301697254180908, + "learning_rate": 0.00014396345232088592, + "loss": 0.1079, + "step": 15155 + }, + { + "epoch": 0.35528050774451025, + "grad_norm": 0.4715832471847534, + "learning_rate": 0.00014395683608986904, + "loss": 0.1283, + "step": 15156 + }, + { + "epoch": 0.35530394931931525, + "grad_norm": 0.37162670493125916, + "learning_rate": 0.00014395021962034118, + "loss": 0.0816, + "step": 15157 + }, + { + "epoch": 0.35532739089412024, + "grad_norm": 0.2681678235530853, + "learning_rate": 0.00014394360291233814, + "loss": 0.0381, + "step": 15158 + }, + { + "epoch": 0.35535083246892524, + "grad_norm": 0.1845708191394806, + "learning_rate": 0.0001439369859658959, + "loss": 0.0277, + "step": 15159 + }, + { + "epoch": 0.35537427404373023, + "grad_norm": 0.3440912067890167, + "learning_rate": 0.00014393036878105033, + "loss": 0.0941, + "step": 15160 + }, + { + "epoch": 0.3553977156185352, + "grad_norm": 0.5356665849685669, + "learning_rate": 0.00014392375135783734, + "loss": 0.1335, + "step": 15161 + }, + { + "epoch": 0.3554211571933403, + "grad_norm": 0.6923215389251709, + "learning_rate": 0.00014391713369629283, + "loss": 0.0845, + "step": 15162 + }, + { + "epoch": 0.35544459876814527, + "grad_norm": 0.1173660010099411, + "learning_rate": 0.00014391051579645275, + "loss": 0.0274, + "step": 15163 + }, + { + "epoch": 0.35546804034295026, + "grad_norm": 0.2800653278827667, + "learning_rate": 0.00014390389765835292, + "loss": 0.3411, + "step": 15164 + }, + { + "epoch": 0.35549148191775526, + "grad_norm": 0.4597272276878357, + "learning_rate": 0.00014389727928202937, + "loss": 0.1643, + "step": 15165 + }, + { + "epoch": 0.35551492349256025, + "grad_norm": 0.22344234585762024, + "learning_rate": 0.00014389066066751793, + "loss": 0.0535, + "step": 15166 + }, + { + "epoch": 0.35553836506736525, + "grad_norm": 0.4153842329978943, + "learning_rate": 0.0001438840418148545, + "loss": 0.066, + "step": 15167 + }, + { + "epoch": 0.35556180664217024, + "grad_norm": 0.11016908288002014, + "learning_rate": 0.00014387742272407506, + "loss": 0.0135, + "step": 15168 + }, + { + "epoch": 0.35558524821697524, + "grad_norm": 0.7008939385414124, + "learning_rate": 0.00014387080339521547, + "loss": 0.8011, + "step": 15169 + }, + { + "epoch": 0.35560868979178023, + "grad_norm": 0.24297381937503815, + "learning_rate": 0.00014386418382831168, + "loss": 0.0315, + "step": 15170 + }, + { + "epoch": 0.3556321313665852, + "grad_norm": 0.5456092953681946, + "learning_rate": 0.0001438575640233996, + "loss": 0.1019, + "step": 15171 + }, + { + "epoch": 0.3556555729413902, + "grad_norm": 0.5308582782745361, + "learning_rate": 0.00014385094398051515, + "loss": 0.0701, + "step": 15172 + }, + { + "epoch": 0.3556790145161952, + "grad_norm": 0.6652831435203552, + "learning_rate": 0.00014384432369969423, + "loss": 0.1729, + "step": 15173 + }, + { + "epoch": 0.3557024560910002, + "grad_norm": 0.5655859112739563, + "learning_rate": 0.00014383770318097277, + "loss": 0.1331, + "step": 15174 + }, + { + "epoch": 0.3557258976658052, + "grad_norm": 0.5956331491470337, + "learning_rate": 0.0001438310824243867, + "loss": 0.1581, + "step": 15175 + }, + { + "epoch": 0.3557493392406102, + "grad_norm": 0.46585899591445923, + "learning_rate": 0.00014382446142997197, + "loss": 0.1012, + "step": 15176 + }, + { + "epoch": 0.3557727808154152, + "grad_norm": 0.4846222698688507, + "learning_rate": 0.00014381784019776447, + "loss": 0.0881, + "step": 15177 + }, + { + "epoch": 0.3557962223902202, + "grad_norm": 0.4599037170410156, + "learning_rate": 0.00014381121872780012, + "loss": 0.1203, + "step": 15178 + }, + { + "epoch": 0.3558196639650252, + "grad_norm": 0.5452141165733337, + "learning_rate": 0.00014380459702011488, + "loss": 0.3651, + "step": 15179 + }, + { + "epoch": 0.3558431055398302, + "grad_norm": 0.32908326387405396, + "learning_rate": 0.00014379797507474464, + "loss": 0.0703, + "step": 15180 + }, + { + "epoch": 0.35586654711463517, + "grad_norm": 0.509022057056427, + "learning_rate": 0.00014379135289172538, + "loss": 0.0919, + "step": 15181 + }, + { + "epoch": 0.35588998868944016, + "grad_norm": 0.5197207927703857, + "learning_rate": 0.00014378473047109303, + "loss": 0.0782, + "step": 15182 + }, + { + "epoch": 0.35591343026424516, + "grad_norm": 0.4196588099002838, + "learning_rate": 0.00014377810781288346, + "loss": 0.0952, + "step": 15183 + }, + { + "epoch": 0.35593687183905015, + "grad_norm": 0.6528346538543701, + "learning_rate": 0.00014377148491713266, + "loss": 0.1609, + "step": 15184 + }, + { + "epoch": 0.35596031341385515, + "grad_norm": 0.19673925638198853, + "learning_rate": 0.00014376486178387655, + "loss": 0.0438, + "step": 15185 + }, + { + "epoch": 0.35598375498866014, + "grad_norm": 0.20056107640266418, + "learning_rate": 0.00014375823841315108, + "loss": 0.0341, + "step": 15186 + }, + { + "epoch": 0.35600719656346513, + "grad_norm": 0.5191015601158142, + "learning_rate": 0.00014375161480499214, + "loss": 0.096, + "step": 15187 + }, + { + "epoch": 0.35603063813827013, + "grad_norm": 0.6214020252227783, + "learning_rate": 0.00014374499095943578, + "loss": 0.066, + "step": 15188 + }, + { + "epoch": 0.3560540797130751, + "grad_norm": 0.5305129289627075, + "learning_rate": 0.0001437383668765178, + "loss": 0.0949, + "step": 15189 + }, + { + "epoch": 0.3560775212878801, + "grad_norm": 0.15492591261863708, + "learning_rate": 0.00014373174255627425, + "loss": 0.0372, + "step": 15190 + }, + { + "epoch": 0.3561009628626851, + "grad_norm": 0.5356788635253906, + "learning_rate": 0.00014372511799874105, + "loss": 0.0996, + "step": 15191 + }, + { + "epoch": 0.3561244044374901, + "grad_norm": 0.7288193106651306, + "learning_rate": 0.00014371849320395409, + "loss": 0.1973, + "step": 15192 + }, + { + "epoch": 0.3561478460122951, + "grad_norm": 0.8446459174156189, + "learning_rate": 0.00014371186817194935, + "loss": 0.1349, + "step": 15193 + }, + { + "epoch": 0.3561712875871001, + "grad_norm": 0.2445070445537567, + "learning_rate": 0.0001437052429027628, + "loss": 0.0497, + "step": 15194 + }, + { + "epoch": 0.3561947291619051, + "grad_norm": 0.17670336365699768, + "learning_rate": 0.0001436986173964304, + "loss": 0.0187, + "step": 15195 + }, + { + "epoch": 0.3562181707367101, + "grad_norm": 0.19270002841949463, + "learning_rate": 0.00014369199165298807, + "loss": 0.039, + "step": 15196 + }, + { + "epoch": 0.3562416123115151, + "grad_norm": 0.6576757431030273, + "learning_rate": 0.00014368536567247174, + "loss": 0.1091, + "step": 15197 + }, + { + "epoch": 0.35626505388632007, + "grad_norm": 0.9919494390487671, + "learning_rate": 0.00014367873945491738, + "loss": 0.1958, + "step": 15198 + }, + { + "epoch": 0.35628849546112507, + "grad_norm": 0.5033446550369263, + "learning_rate": 0.00014367211300036096, + "loss": 0.0899, + "step": 15199 + }, + { + "epoch": 0.35631193703593006, + "grad_norm": 0.7648343443870544, + "learning_rate": 0.00014366548630883843, + "loss": 0.1902, + "step": 15200 + }, + { + "epoch": 0.35633537861073505, + "grad_norm": 0.5147693157196045, + "learning_rate": 0.00014365885938038577, + "loss": 0.1011, + "step": 15201 + }, + { + "epoch": 0.35635882018554005, + "grad_norm": 0.2276890128850937, + "learning_rate": 0.00014365223221503886, + "loss": 0.0562, + "step": 15202 + }, + { + "epoch": 0.35638226176034504, + "grad_norm": 0.6075115203857422, + "learning_rate": 0.00014364560481283377, + "loss": 0.7122, + "step": 15203 + }, + { + "epoch": 0.35640570333515004, + "grad_norm": 0.39033129811286926, + "learning_rate": 0.00014363897717380636, + "loss": 0.1273, + "step": 15204 + }, + { + "epoch": 0.35642914490995503, + "grad_norm": 0.36645281314849854, + "learning_rate": 0.00014363234929799264, + "loss": 0.0657, + "step": 15205 + }, + { + "epoch": 0.35645258648476, + "grad_norm": 0.2671982944011688, + "learning_rate": 0.00014362572118542855, + "loss": 0.0656, + "step": 15206 + }, + { + "epoch": 0.356476028059565, + "grad_norm": 0.53596031665802, + "learning_rate": 0.0001436190928361501, + "loss": 0.5876, + "step": 15207 + }, + { + "epoch": 0.35649946963437, + "grad_norm": 0.5672889351844788, + "learning_rate": 0.0001436124642501932, + "loss": 0.097, + "step": 15208 + }, + { + "epoch": 0.356522911209175, + "grad_norm": 0.6004342436790466, + "learning_rate": 0.00014360583542759388, + "loss": 0.1512, + "step": 15209 + }, + { + "epoch": 0.35654635278398, + "grad_norm": 0.5090895295143127, + "learning_rate": 0.00014359920636838807, + "loss": 0.1059, + "step": 15210 + }, + { + "epoch": 0.356569794358785, + "grad_norm": 0.3716277778148651, + "learning_rate": 0.00014359257707261168, + "loss": 0.0614, + "step": 15211 + }, + { + "epoch": 0.35659323593359, + "grad_norm": 0.2210218459367752, + "learning_rate": 0.00014358594754030078, + "loss": 0.0408, + "step": 15212 + }, + { + "epoch": 0.35661667750839504, + "grad_norm": 0.20070846378803253, + "learning_rate": 0.00014357931777149132, + "loss": 0.0434, + "step": 15213 + }, + { + "epoch": 0.35664011908320004, + "grad_norm": 0.8141675591468811, + "learning_rate": 0.00014357268776621923, + "loss": 0.2418, + "step": 15214 + }, + { + "epoch": 0.35666356065800503, + "grad_norm": 0.45190852880477905, + "learning_rate": 0.00014356605752452052, + "loss": 0.1067, + "step": 15215 + }, + { + "epoch": 0.35668700223281, + "grad_norm": 0.5143461227416992, + "learning_rate": 0.00014355942704643118, + "loss": 0.1153, + "step": 15216 + }, + { + "epoch": 0.356710443807615, + "grad_norm": 0.18856261670589447, + "learning_rate": 0.00014355279633198714, + "loss": 0.0487, + "step": 15217 + }, + { + "epoch": 0.35673388538242, + "grad_norm": 0.3239424228668213, + "learning_rate": 0.00014354616538122435, + "loss": 0.084, + "step": 15218 + }, + { + "epoch": 0.356757326957225, + "grad_norm": 0.35858818888664246, + "learning_rate": 0.00014353953419417895, + "loss": 0.0723, + "step": 15219 + }, + { + "epoch": 0.35678076853203, + "grad_norm": 0.4605001211166382, + "learning_rate": 0.00014353290277088675, + "loss": 0.5996, + "step": 15220 + }, + { + "epoch": 0.356804210106835, + "grad_norm": 0.22118288278579712, + "learning_rate": 0.0001435262711113838, + "loss": 0.0414, + "step": 15221 + }, + { + "epoch": 0.35682765168164, + "grad_norm": 0.17671610414981842, + "learning_rate": 0.00014351963921570608, + "loss": 0.0251, + "step": 15222 + }, + { + "epoch": 0.356851093256445, + "grad_norm": 0.46974098682403564, + "learning_rate": 0.00014351300708388958, + "loss": 0.1072, + "step": 15223 + }, + { + "epoch": 0.35687453483125, + "grad_norm": 0.1483990103006363, + "learning_rate": 0.00014350637471597025, + "loss": 0.0279, + "step": 15224 + }, + { + "epoch": 0.356897976406055, + "grad_norm": 0.793464720249176, + "learning_rate": 0.00014349974211198417, + "loss": 0.1362, + "step": 15225 + }, + { + "epoch": 0.35692141798085997, + "grad_norm": 0.11782175302505493, + "learning_rate": 0.00014349310927196722, + "loss": 0.0174, + "step": 15226 + }, + { + "epoch": 0.35694485955566496, + "grad_norm": 0.16794787347316742, + "learning_rate": 0.00014348647619595543, + "loss": 0.0255, + "step": 15227 + }, + { + "epoch": 0.35696830113046996, + "grad_norm": 0.550499677658081, + "learning_rate": 0.00014347984288398484, + "loss": 0.0763, + "step": 15228 + }, + { + "epoch": 0.35699174270527495, + "grad_norm": 0.33286479115486145, + "learning_rate": 0.00014347320933609138, + "loss": 0.0553, + "step": 15229 + }, + { + "epoch": 0.35701518428007994, + "grad_norm": 0.6034372448921204, + "learning_rate": 0.00014346657555231105, + "loss": 0.131, + "step": 15230 + }, + { + "epoch": 0.35703862585488494, + "grad_norm": 0.3512215316295624, + "learning_rate": 0.00014345994153267986, + "loss": 0.0621, + "step": 15231 + }, + { + "epoch": 0.35706206742968993, + "grad_norm": 0.3588886260986328, + "learning_rate": 0.0001434533072772338, + "loss": 0.0616, + "step": 15232 + }, + { + "epoch": 0.3570855090044949, + "grad_norm": 0.3451363146305084, + "learning_rate": 0.0001434466727860089, + "loss": 0.3435, + "step": 15233 + }, + { + "epoch": 0.3571089505792999, + "grad_norm": 0.3546528220176697, + "learning_rate": 0.0001434400380590411, + "loss": 0.1085, + "step": 15234 + }, + { + "epoch": 0.3571323921541049, + "grad_norm": 0.6773254871368408, + "learning_rate": 0.00014343340309636645, + "loss": 0.1647, + "step": 15235 + }, + { + "epoch": 0.3571558337289099, + "grad_norm": 0.10920088738203049, + "learning_rate": 0.00014342676789802096, + "loss": 0.0118, + "step": 15236 + }, + { + "epoch": 0.3571792753037149, + "grad_norm": 0.45223379135131836, + "learning_rate": 0.00014342013246404056, + "loss": 0.0572, + "step": 15237 + }, + { + "epoch": 0.3572027168785199, + "grad_norm": 0.45413222908973694, + "learning_rate": 0.0001434134967944613, + "loss": 0.0561, + "step": 15238 + }, + { + "epoch": 0.3572261584533249, + "grad_norm": 0.1587178111076355, + "learning_rate": 0.00014340686088931925, + "loss": 0.0425, + "step": 15239 + }, + { + "epoch": 0.3572496000281299, + "grad_norm": 0.1561165750026703, + "learning_rate": 0.0001434002247486503, + "loss": 0.036, + "step": 15240 + }, + { + "epoch": 0.3572730416029349, + "grad_norm": 0.49838489294052124, + "learning_rate": 0.0001433935883724905, + "loss": 0.1141, + "step": 15241 + }, + { + "epoch": 0.3572964831777399, + "grad_norm": 1.550628423690796, + "learning_rate": 0.0001433869517608759, + "loss": 0.1668, + "step": 15242 + }, + { + "epoch": 0.35731992475254487, + "grad_norm": 0.21121181547641754, + "learning_rate": 0.00014338031491384245, + "loss": 0.0336, + "step": 15243 + }, + { + "epoch": 0.35734336632734987, + "grad_norm": 0.5421642065048218, + "learning_rate": 0.0001433736778314262, + "loss": 0.1027, + "step": 15244 + }, + { + "epoch": 0.35736680790215486, + "grad_norm": 0.09898748248815536, + "learning_rate": 0.0001433670405136632, + "loss": 0.0157, + "step": 15245 + }, + { + "epoch": 0.35739024947695985, + "grad_norm": 0.6433912515640259, + "learning_rate": 0.00014336040296058936, + "loss": 0.199, + "step": 15246 + }, + { + "epoch": 0.35741369105176485, + "grad_norm": 0.421297550201416, + "learning_rate": 0.00014335376517224076, + "loss": 0.0453, + "step": 15247 + }, + { + "epoch": 0.35743713262656984, + "grad_norm": 0.46486955881118774, + "learning_rate": 0.00014334712714865347, + "loss": 0.1759, + "step": 15248 + }, + { + "epoch": 0.35746057420137484, + "grad_norm": 0.41373538970947266, + "learning_rate": 0.0001433404888898634, + "loss": 0.1408, + "step": 15249 + }, + { + "epoch": 0.35748401577617983, + "grad_norm": 0.9117533564567566, + "learning_rate": 0.0001433338503959066, + "loss": 0.2198, + "step": 15250 + }, + { + "epoch": 0.3575074573509848, + "grad_norm": 0.10621833056211472, + "learning_rate": 0.00014332721166681917, + "loss": 0.0156, + "step": 15251 + }, + { + "epoch": 0.3575308989257898, + "grad_norm": 0.11236737668514252, + "learning_rate": 0.00014332057270263703, + "loss": 0.0204, + "step": 15252 + }, + { + "epoch": 0.3575543405005948, + "grad_norm": 0.2488178014755249, + "learning_rate": 0.00014331393350339622, + "loss": 0.0698, + "step": 15253 + }, + { + "epoch": 0.3575777820753998, + "grad_norm": 0.6624580025672913, + "learning_rate": 0.00014330729406913283, + "loss": 0.1178, + "step": 15254 + }, + { + "epoch": 0.3576012236502048, + "grad_norm": 0.4273901879787445, + "learning_rate": 0.00014330065439988285, + "loss": 0.0583, + "step": 15255 + }, + { + "epoch": 0.3576246652250098, + "grad_norm": 0.31488898396492004, + "learning_rate": 0.00014329401449568227, + "loss": 0.0549, + "step": 15256 + }, + { + "epoch": 0.3576481067998148, + "grad_norm": 0.1835309863090515, + "learning_rate": 0.00014328737435656717, + "loss": 0.0292, + "step": 15257 + }, + { + "epoch": 0.3576715483746198, + "grad_norm": 0.16840051114559174, + "learning_rate": 0.00014328073398257358, + "loss": 0.0164, + "step": 15258 + }, + { + "epoch": 0.3576949899494248, + "grad_norm": 0.6392549872398376, + "learning_rate": 0.00014327409337373746, + "loss": 0.2192, + "step": 15259 + }, + { + "epoch": 0.3577184315242298, + "grad_norm": 0.40889209508895874, + "learning_rate": 0.00014326745253009492, + "loss": 0.076, + "step": 15260 + }, + { + "epoch": 0.35774187309903477, + "grad_norm": 0.6789216995239258, + "learning_rate": 0.00014326081145168198, + "loss": 0.1655, + "step": 15261 + }, + { + "epoch": 0.35776531467383976, + "grad_norm": 0.4584147334098816, + "learning_rate": 0.00014325417013853463, + "loss": 0.1014, + "step": 15262 + }, + { + "epoch": 0.35778875624864476, + "grad_norm": 0.40571653842926025, + "learning_rate": 0.00014324752859068895, + "loss": 0.0658, + "step": 15263 + }, + { + "epoch": 0.3578121978234498, + "grad_norm": 0.44210124015808105, + "learning_rate": 0.00014324088680818097, + "loss": 0.0921, + "step": 15264 + }, + { + "epoch": 0.3578356393982548, + "grad_norm": 0.3554103374481201, + "learning_rate": 0.00014323424479104673, + "loss": 0.102, + "step": 15265 + }, + { + "epoch": 0.3578590809730598, + "grad_norm": 0.6097589135169983, + "learning_rate": 0.00014322760253932223, + "loss": 0.1412, + "step": 15266 + }, + { + "epoch": 0.3578825225478648, + "grad_norm": 0.39525744318962097, + "learning_rate": 0.00014322096005304358, + "loss": 0.1053, + "step": 15267 + }, + { + "epoch": 0.3579059641226698, + "grad_norm": 0.31594017148017883, + "learning_rate": 0.00014321431733224678, + "loss": 0.0651, + "step": 15268 + }, + { + "epoch": 0.3579294056974748, + "grad_norm": 0.6587033271789551, + "learning_rate": 0.00014320767437696786, + "loss": 0.7401, + "step": 15269 + }, + { + "epoch": 0.3579528472722798, + "grad_norm": 0.25848832726478577, + "learning_rate": 0.0001432010311872429, + "loss": 0.035, + "step": 15270 + }, + { + "epoch": 0.35797628884708477, + "grad_norm": 0.09802556782960892, + "learning_rate": 0.0001431943877631079, + "loss": 0.019, + "step": 15271 + }, + { + "epoch": 0.35799973042188976, + "grad_norm": 0.5452908277511597, + "learning_rate": 0.00014318774410459895, + "loss": 0.1093, + "step": 15272 + }, + { + "epoch": 0.35802317199669476, + "grad_norm": 0.6469303965568542, + "learning_rate": 0.0001431811002117521, + "loss": 0.1575, + "step": 15273 + }, + { + "epoch": 0.35804661357149975, + "grad_norm": 0.47661092877388, + "learning_rate": 0.00014317445608460342, + "loss": 0.1304, + "step": 15274 + }, + { + "epoch": 0.35807005514630474, + "grad_norm": 0.9082770943641663, + "learning_rate": 0.00014316781172318886, + "loss": 0.1326, + "step": 15275 + }, + { + "epoch": 0.35809349672110974, + "grad_norm": 0.27309486269950867, + "learning_rate": 0.00014316116712754457, + "loss": 0.0497, + "step": 15276 + }, + { + "epoch": 0.35811693829591473, + "grad_norm": 0.2425425499677658, + "learning_rate": 0.00014315452229770657, + "loss": 0.0476, + "step": 15277 + }, + { + "epoch": 0.3581403798707197, + "grad_norm": 0.5106315016746521, + "learning_rate": 0.00014314787723371093, + "loss": 0.0722, + "step": 15278 + }, + { + "epoch": 0.3581638214455247, + "grad_norm": 0.19255001842975616, + "learning_rate": 0.00014314123193559367, + "loss": 0.0455, + "step": 15279 + }, + { + "epoch": 0.3581872630203297, + "grad_norm": 0.44503453373908997, + "learning_rate": 0.0001431345864033909, + "loss": 0.0587, + "step": 15280 + }, + { + "epoch": 0.3582107045951347, + "grad_norm": 0.40987589955329895, + "learning_rate": 0.00014312794063713863, + "loss": 0.1071, + "step": 15281 + }, + { + "epoch": 0.3582341461699397, + "grad_norm": 0.6188866496086121, + "learning_rate": 0.00014312129463687296, + "loss": 0.0907, + "step": 15282 + }, + { + "epoch": 0.3582575877447447, + "grad_norm": 0.537573516368866, + "learning_rate": 0.00014311464840262994, + "loss": 0.0795, + "step": 15283 + }, + { + "epoch": 0.3582810293195497, + "grad_norm": 0.49550142884254456, + "learning_rate": 0.0001431080019344456, + "loss": 0.1378, + "step": 15284 + }, + { + "epoch": 0.3583044708943547, + "grad_norm": 0.5168811678886414, + "learning_rate": 0.00014310135523235603, + "loss": 0.0792, + "step": 15285 + }, + { + "epoch": 0.3583279124691597, + "grad_norm": 0.6790927052497864, + "learning_rate": 0.0001430947082963973, + "loss": 0.6148, + "step": 15286 + }, + { + "epoch": 0.3583513540439647, + "grad_norm": 0.2500784993171692, + "learning_rate": 0.0001430880611266055, + "loss": 0.0776, + "step": 15287 + }, + { + "epoch": 0.35837479561876967, + "grad_norm": 0.6594416499137878, + "learning_rate": 0.00014308141372301663, + "loss": 0.0603, + "step": 15288 + }, + { + "epoch": 0.35839823719357466, + "grad_norm": 0.38919076323509216, + "learning_rate": 0.0001430747660856668, + "loss": 0.0888, + "step": 15289 + }, + { + "epoch": 0.35842167876837966, + "grad_norm": 0.3943278193473816, + "learning_rate": 0.0001430681182145921, + "loss": 0.0708, + "step": 15290 + }, + { + "epoch": 0.35844512034318465, + "grad_norm": 0.5721248984336853, + "learning_rate": 0.00014306147010982855, + "loss": 0.1732, + "step": 15291 + }, + { + "epoch": 0.35846856191798965, + "grad_norm": 0.5662621855735779, + "learning_rate": 0.0001430548217714123, + "loss": 0.0964, + "step": 15292 + }, + { + "epoch": 0.35849200349279464, + "grad_norm": 0.36976152658462524, + "learning_rate": 0.00014304817319937935, + "loss": 0.0627, + "step": 15293 + }, + { + "epoch": 0.35851544506759964, + "grad_norm": 0.11097455024719238, + "learning_rate": 0.0001430415243937658, + "loss": 0.0259, + "step": 15294 + }, + { + "epoch": 0.35853888664240463, + "grad_norm": 0.36751189827919006, + "learning_rate": 0.0001430348753546077, + "loss": 0.0959, + "step": 15295 + }, + { + "epoch": 0.3585623282172096, + "grad_norm": 0.30158403515815735, + "learning_rate": 0.00014302822608194119, + "loss": 0.0621, + "step": 15296 + }, + { + "epoch": 0.3585857697920146, + "grad_norm": 0.5528475642204285, + "learning_rate": 0.00014302157657580233, + "loss": 0.1645, + "step": 15297 + }, + { + "epoch": 0.3586092113668196, + "grad_norm": 0.38427621126174927, + "learning_rate": 0.00014301492683622716, + "loss": 0.0392, + "step": 15298 + }, + { + "epoch": 0.3586326529416246, + "grad_norm": 0.5603936314582825, + "learning_rate": 0.0001430082768632518, + "loss": 0.6416, + "step": 15299 + }, + { + "epoch": 0.3586560945164296, + "grad_norm": 1.2139415740966797, + "learning_rate": 0.0001430016266569123, + "loss": 0.218, + "step": 15300 + }, + { + "epoch": 0.3586795360912346, + "grad_norm": 0.46276989579200745, + "learning_rate": 0.00014299497621724478, + "loss": 0.0536, + "step": 15301 + }, + { + "epoch": 0.3587029776660396, + "grad_norm": 1.0506082773208618, + "learning_rate": 0.0001429883255442853, + "loss": 0.2906, + "step": 15302 + }, + { + "epoch": 0.3587264192408446, + "grad_norm": 0.23107579350471497, + "learning_rate": 0.00014298167463806997, + "loss": 0.0448, + "step": 15303 + }, + { + "epoch": 0.3587498608156496, + "grad_norm": 0.7187661528587341, + "learning_rate": 0.00014297502349863486, + "loss": 0.1594, + "step": 15304 + }, + { + "epoch": 0.3587733023904546, + "grad_norm": 0.5334590673446655, + "learning_rate": 0.00014296837212601608, + "loss": 0.0826, + "step": 15305 + }, + { + "epoch": 0.35879674396525957, + "grad_norm": 0.5873621106147766, + "learning_rate": 0.0001429617205202497, + "loss": 0.1391, + "step": 15306 + }, + { + "epoch": 0.35882018554006456, + "grad_norm": 0.5735127925872803, + "learning_rate": 0.0001429550686813718, + "loss": 0.1302, + "step": 15307 + }, + { + "epoch": 0.35884362711486956, + "grad_norm": 0.5172568559646606, + "learning_rate": 0.00014294841660941846, + "loss": 0.1128, + "step": 15308 + }, + { + "epoch": 0.35886706868967455, + "grad_norm": 0.42384567856788635, + "learning_rate": 0.00014294176430442585, + "loss": 0.0871, + "step": 15309 + }, + { + "epoch": 0.35889051026447955, + "grad_norm": 0.2565728425979614, + "learning_rate": 0.00014293511176643003, + "loss": 0.0571, + "step": 15310 + }, + { + "epoch": 0.35891395183928454, + "grad_norm": 0.4718148112297058, + "learning_rate": 0.00014292845899546705, + "loss": 0.0835, + "step": 15311 + }, + { + "epoch": 0.35893739341408953, + "grad_norm": 0.5453210473060608, + "learning_rate": 0.00014292180599157308, + "loss": 0.175, + "step": 15312 + }, + { + "epoch": 0.35896083498889453, + "grad_norm": 0.5794983506202698, + "learning_rate": 0.00014291515275478416, + "loss": 0.69, + "step": 15313 + }, + { + "epoch": 0.3589842765636995, + "grad_norm": 0.5295314192771912, + "learning_rate": 0.0001429084992851364, + "loss": 0.3944, + "step": 15314 + }, + { + "epoch": 0.3590077181385046, + "grad_norm": 0.19800806045532227, + "learning_rate": 0.00014290184558266592, + "loss": 0.0467, + "step": 15315 + }, + { + "epoch": 0.35903115971330957, + "grad_norm": 0.6884468197822571, + "learning_rate": 0.00014289519164740886, + "loss": 0.6193, + "step": 15316 + }, + { + "epoch": 0.35905460128811456, + "grad_norm": 0.4323298931121826, + "learning_rate": 0.00014288853747940124, + "loss": 0.0882, + "step": 15317 + }, + { + "epoch": 0.35907804286291956, + "grad_norm": 0.10683068633079529, + "learning_rate": 0.00014288188307867925, + "loss": 0.025, + "step": 15318 + }, + { + "epoch": 0.35910148443772455, + "grad_norm": 0.5182574987411499, + "learning_rate": 0.00014287522844527893, + "loss": 0.7418, + "step": 15319 + }, + { + "epoch": 0.35912492601252954, + "grad_norm": 0.15186654031276703, + "learning_rate": 0.0001428685735792364, + "loss": 0.028, + "step": 15320 + }, + { + "epoch": 0.35914836758733454, + "grad_norm": 0.09992299228906631, + "learning_rate": 0.0001428619184805878, + "loss": 0.0235, + "step": 15321 + }, + { + "epoch": 0.35917180916213953, + "grad_norm": 0.40164923667907715, + "learning_rate": 0.00014285526314936924, + "loss": 0.1034, + "step": 15322 + }, + { + "epoch": 0.3591952507369445, + "grad_norm": 0.3598696291446686, + "learning_rate": 0.0001428486075856168, + "loss": 0.0522, + "step": 15323 + }, + { + "epoch": 0.3592186923117495, + "grad_norm": 0.09116311371326447, + "learning_rate": 0.0001428419517893666, + "loss": 0.0108, + "step": 15324 + }, + { + "epoch": 0.3592421338865545, + "grad_norm": 0.17264948785305023, + "learning_rate": 0.00014283529576065483, + "loss": 0.0438, + "step": 15325 + }, + { + "epoch": 0.3592655754613595, + "grad_norm": 0.6690117716789246, + "learning_rate": 0.00014282863949951742, + "loss": 0.6352, + "step": 15326 + }, + { + "epoch": 0.3592890170361645, + "grad_norm": 0.5591263175010681, + "learning_rate": 0.0001428219830059907, + "loss": 0.5171, + "step": 15327 + }, + { + "epoch": 0.3593124586109695, + "grad_norm": 0.2566153109073639, + "learning_rate": 0.0001428153262801107, + "loss": 0.0415, + "step": 15328 + }, + { + "epoch": 0.3593359001857745, + "grad_norm": 0.5326455235481262, + "learning_rate": 0.0001428086693219135, + "loss": 0.1248, + "step": 15329 + }, + { + "epoch": 0.3593593417605795, + "grad_norm": 0.5499171614646912, + "learning_rate": 0.00014280201213143527, + "loss": 0.14, + "step": 15330 + }, + { + "epoch": 0.3593827833353845, + "grad_norm": 0.451225221157074, + "learning_rate": 0.00014279535470871211, + "loss": 0.0453, + "step": 15331 + }, + { + "epoch": 0.3594062249101895, + "grad_norm": 0.7115644812583923, + "learning_rate": 0.00014278869705378018, + "loss": 0.1033, + "step": 15332 + }, + { + "epoch": 0.35942966648499447, + "grad_norm": 0.8526344895362854, + "learning_rate": 0.00014278203916667553, + "loss": 0.1721, + "step": 15333 + }, + { + "epoch": 0.35945310805979946, + "grad_norm": 0.09148505330085754, + "learning_rate": 0.00014277538104743437, + "loss": 0.0078, + "step": 15334 + }, + { + "epoch": 0.35947654963460446, + "grad_norm": 0.9384901523590088, + "learning_rate": 0.00014276872269609282, + "loss": 0.0975, + "step": 15335 + }, + { + "epoch": 0.35949999120940945, + "grad_norm": 0.6269135475158691, + "learning_rate": 0.00014276206411268692, + "loss": 0.5808, + "step": 15336 + }, + { + "epoch": 0.35952343278421445, + "grad_norm": 0.5489304065704346, + "learning_rate": 0.00014275540529725287, + "loss": 0.0778, + "step": 15337 + }, + { + "epoch": 0.35954687435901944, + "grad_norm": 3.1785573959350586, + "learning_rate": 0.00014274874624982677, + "loss": 0.1081, + "step": 15338 + }, + { + "epoch": 0.35957031593382444, + "grad_norm": 0.35968807339668274, + "learning_rate": 0.00014274208697044477, + "loss": 0.0585, + "step": 15339 + }, + { + "epoch": 0.35959375750862943, + "grad_norm": 0.4387403428554535, + "learning_rate": 0.00014273542745914307, + "loss": 0.0795, + "step": 15340 + }, + { + "epoch": 0.3596171990834344, + "grad_norm": 0.4100063145160675, + "learning_rate": 0.00014272876771595768, + "loss": 0.0591, + "step": 15341 + }, + { + "epoch": 0.3596406406582394, + "grad_norm": 0.1595812737941742, + "learning_rate": 0.00014272210774092482, + "loss": 0.041, + "step": 15342 + }, + { + "epoch": 0.3596640822330444, + "grad_norm": 0.6753712892532349, + "learning_rate": 0.0001427154475340806, + "loss": 0.0894, + "step": 15343 + }, + { + "epoch": 0.3596875238078494, + "grad_norm": 0.47392988204956055, + "learning_rate": 0.00014270878709546114, + "loss": 0.0787, + "step": 15344 + }, + { + "epoch": 0.3597109653826544, + "grad_norm": 0.22838374972343445, + "learning_rate": 0.0001427021264251026, + "loss": 0.0753, + "step": 15345 + }, + { + "epoch": 0.3597344069574594, + "grad_norm": 0.16867755353450775, + "learning_rate": 0.00014269546552304115, + "loss": 0.0246, + "step": 15346 + }, + { + "epoch": 0.3597578485322644, + "grad_norm": 0.18387871980667114, + "learning_rate": 0.0001426888043893129, + "loss": 0.0367, + "step": 15347 + }, + { + "epoch": 0.3597812901070694, + "grad_norm": 0.41114601492881775, + "learning_rate": 0.00014268214302395397, + "loss": 0.0989, + "step": 15348 + }, + { + "epoch": 0.3598047316818744, + "grad_norm": 0.4660171568393707, + "learning_rate": 0.00014267548142700054, + "loss": 0.1041, + "step": 15349 + }, + { + "epoch": 0.3598281732566794, + "grad_norm": 0.6649132370948792, + "learning_rate": 0.00014266881959848878, + "loss": 0.1138, + "step": 15350 + }, + { + "epoch": 0.35985161483148437, + "grad_norm": 0.5181355476379395, + "learning_rate": 0.00014266215753845477, + "loss": 0.1052, + "step": 15351 + }, + { + "epoch": 0.35987505640628936, + "grad_norm": 0.6274743676185608, + "learning_rate": 0.00014265549524693468, + "loss": 0.0801, + "step": 15352 + }, + { + "epoch": 0.35989849798109436, + "grad_norm": 0.5499054193496704, + "learning_rate": 0.00014264883272396473, + "loss": 0.0802, + "step": 15353 + }, + { + "epoch": 0.35992193955589935, + "grad_norm": 0.49564287066459656, + "learning_rate": 0.000142642169969581, + "loss": 0.4685, + "step": 15354 + }, + { + "epoch": 0.35994538113070434, + "grad_norm": 0.6722387671470642, + "learning_rate": 0.00014263550698381962, + "loss": 0.1094, + "step": 15355 + }, + { + "epoch": 0.35996882270550934, + "grad_norm": 0.6925005316734314, + "learning_rate": 0.0001426288437667168, + "loss": 0.0366, + "step": 15356 + }, + { + "epoch": 0.35999226428031433, + "grad_norm": 0.6131452918052673, + "learning_rate": 0.00014262218031830867, + "loss": 0.0868, + "step": 15357 + }, + { + "epoch": 0.36001570585511933, + "grad_norm": 0.46522650122642517, + "learning_rate": 0.0001426155166386314, + "loss": 0.0813, + "step": 15358 + }, + { + "epoch": 0.3600391474299243, + "grad_norm": 0.22044090926647186, + "learning_rate": 0.00014260885272772113, + "loss": 0.064, + "step": 15359 + }, + { + "epoch": 0.3600625890047293, + "grad_norm": 0.49458345770835876, + "learning_rate": 0.00014260218858561406, + "loss": 0.1416, + "step": 15360 + }, + { + "epoch": 0.3600860305795343, + "grad_norm": 0.6275799870491028, + "learning_rate": 0.00014259552421234625, + "loss": 0.1394, + "step": 15361 + }, + { + "epoch": 0.3601094721543393, + "grad_norm": 0.4284748435020447, + "learning_rate": 0.000142588859607954, + "loss": 0.0666, + "step": 15362 + }, + { + "epoch": 0.3601329137291443, + "grad_norm": 0.379987895488739, + "learning_rate": 0.00014258219477247333, + "loss": 0.1174, + "step": 15363 + }, + { + "epoch": 0.3601563553039493, + "grad_norm": 0.12703640758991241, + "learning_rate": 0.00014257552970594052, + "loss": 0.0157, + "step": 15364 + }, + { + "epoch": 0.3601797968787543, + "grad_norm": 0.5403708815574646, + "learning_rate": 0.0001425688644083917, + "loss": 0.1568, + "step": 15365 + }, + { + "epoch": 0.3602032384535593, + "grad_norm": 0.6093202829360962, + "learning_rate": 0.00014256219887986299, + "loss": 0.141, + "step": 15366 + }, + { + "epoch": 0.36022668002836433, + "grad_norm": 0.2759878933429718, + "learning_rate": 0.0001425555331203906, + "loss": 0.069, + "step": 15367 + }, + { + "epoch": 0.3602501216031693, + "grad_norm": 0.43945541977882385, + "learning_rate": 0.00014254886713001072, + "loss": 0.0853, + "step": 15368 + }, + { + "epoch": 0.3602735631779743, + "grad_norm": 0.40602439641952515, + "learning_rate": 0.00014254220090875944, + "loss": 0.0799, + "step": 15369 + }, + { + "epoch": 0.3602970047527793, + "grad_norm": 0.48789480328559875, + "learning_rate": 0.000142535534456673, + "loss": 0.091, + "step": 15370 + }, + { + "epoch": 0.3603204463275843, + "grad_norm": 0.6774864196777344, + "learning_rate": 0.0001425288677737876, + "loss": 0.8151, + "step": 15371 + }, + { + "epoch": 0.3603438879023893, + "grad_norm": 0.5559812784194946, + "learning_rate": 0.00014252220086013933, + "loss": 0.1906, + "step": 15372 + }, + { + "epoch": 0.3603673294771943, + "grad_norm": 0.2272338718175888, + "learning_rate": 0.00014251553371576441, + "loss": 0.0219, + "step": 15373 + }, + { + "epoch": 0.3603907710519993, + "grad_norm": 0.10960028320550919, + "learning_rate": 0.00014250886634069898, + "loss": 0.0089, + "step": 15374 + }, + { + "epoch": 0.3604142126268043, + "grad_norm": 0.3851814568042755, + "learning_rate": 0.00014250219873497926, + "loss": 0.1046, + "step": 15375 + }, + { + "epoch": 0.3604376542016093, + "grad_norm": 0.5804701447486877, + "learning_rate": 0.00014249553089864142, + "loss": 0.1207, + "step": 15376 + }, + { + "epoch": 0.3604610957764143, + "grad_norm": 0.6870311498641968, + "learning_rate": 0.00014248886283172165, + "loss": 0.1303, + "step": 15377 + }, + { + "epoch": 0.36048453735121927, + "grad_norm": 0.16749493777751923, + "learning_rate": 0.00014248219453425612, + "loss": 0.0207, + "step": 15378 + }, + { + "epoch": 0.36050797892602426, + "grad_norm": 0.3882644772529602, + "learning_rate": 0.000142475526006281, + "loss": 0.0667, + "step": 15379 + }, + { + "epoch": 0.36053142050082926, + "grad_norm": 0.2890281081199646, + "learning_rate": 0.00014246885724783246, + "loss": 0.0618, + "step": 15380 + }, + { + "epoch": 0.36055486207563425, + "grad_norm": 0.6012424230575562, + "learning_rate": 0.0001424621882589467, + "loss": 0.139, + "step": 15381 + }, + { + "epoch": 0.36057830365043925, + "grad_norm": 0.5110245943069458, + "learning_rate": 0.00014245551903965993, + "loss": 0.1394, + "step": 15382 + }, + { + "epoch": 0.36060174522524424, + "grad_norm": 0.47190776467323303, + "learning_rate": 0.0001424488495900083, + "loss": 0.0896, + "step": 15383 + }, + { + "epoch": 0.36062518680004924, + "grad_norm": 0.46198371052742004, + "learning_rate": 0.00014244217991002807, + "loss": 0.1249, + "step": 15384 + }, + { + "epoch": 0.36064862837485423, + "grad_norm": 0.4575066566467285, + "learning_rate": 0.00014243550999975535, + "loss": 0.0795, + "step": 15385 + }, + { + "epoch": 0.3606720699496592, + "grad_norm": 0.4812944829463959, + "learning_rate": 0.00014242883985922635, + "loss": 0.1225, + "step": 15386 + }, + { + "epoch": 0.3606955115244642, + "grad_norm": 0.5427215099334717, + "learning_rate": 0.00014242216948847728, + "loss": 0.0899, + "step": 15387 + }, + { + "epoch": 0.3607189530992692, + "grad_norm": 0.2719856798648834, + "learning_rate": 0.00014241549888754432, + "loss": 0.0651, + "step": 15388 + }, + { + "epoch": 0.3607423946740742, + "grad_norm": 0.6005908250808716, + "learning_rate": 0.00014240882805646368, + "loss": 0.6743, + "step": 15389 + }, + { + "epoch": 0.3607658362488792, + "grad_norm": 0.4858058989048004, + "learning_rate": 0.00014240215699527154, + "loss": 0.0487, + "step": 15390 + }, + { + "epoch": 0.3607892778236842, + "grad_norm": 0.39194256067276, + "learning_rate": 0.0001423954857040041, + "loss": 0.0767, + "step": 15391 + }, + { + "epoch": 0.3608127193984892, + "grad_norm": 0.5989351272583008, + "learning_rate": 0.00014238881418269757, + "loss": 0.4598, + "step": 15392 + }, + { + "epoch": 0.3608361609732942, + "grad_norm": 0.5252747535705566, + "learning_rate": 0.00014238214243138814, + "loss": 0.6096, + "step": 15393 + }, + { + "epoch": 0.3608596025480992, + "grad_norm": 0.6567450165748596, + "learning_rate": 0.000142375470450112, + "loss": 0.1235, + "step": 15394 + }, + { + "epoch": 0.3608830441229042, + "grad_norm": 0.32863178849220276, + "learning_rate": 0.0001423687982389054, + "loss": 0.0807, + "step": 15395 + }, + { + "epoch": 0.36090648569770917, + "grad_norm": 0.4764567017555237, + "learning_rate": 0.0001423621257978045, + "loss": 0.0728, + "step": 15396 + }, + { + "epoch": 0.36092992727251416, + "grad_norm": 0.1588173508644104, + "learning_rate": 0.0001423554531268455, + "loss": 0.0351, + "step": 15397 + }, + { + "epoch": 0.36095336884731916, + "grad_norm": 0.4871889352798462, + "learning_rate": 0.00014234878022606465, + "loss": 0.0408, + "step": 15398 + }, + { + "epoch": 0.36097681042212415, + "grad_norm": 0.4059670567512512, + "learning_rate": 0.00014234210709549805, + "loss": 0.0942, + "step": 15399 + }, + { + "epoch": 0.36100025199692914, + "grad_norm": 0.4839153289794922, + "learning_rate": 0.00014233543373518205, + "loss": 0.5231, + "step": 15400 + }, + { + "epoch": 0.36102369357173414, + "grad_norm": 0.16790691018104553, + "learning_rate": 0.0001423287601451528, + "loss": 0.0326, + "step": 15401 + }, + { + "epoch": 0.36104713514653913, + "grad_norm": 0.4869921803474426, + "learning_rate": 0.00014232208632544648, + "loss": 0.1138, + "step": 15402 + }, + { + "epoch": 0.3610705767213441, + "grad_norm": 0.46274372935295105, + "learning_rate": 0.00014231541227609934, + "loss": 0.0832, + "step": 15403 + }, + { + "epoch": 0.3610940182961491, + "grad_norm": 0.680107057094574, + "learning_rate": 0.0001423087379971476, + "loss": 0.2055, + "step": 15404 + }, + { + "epoch": 0.3611174598709541, + "grad_norm": 0.3234509229660034, + "learning_rate": 0.00014230206348862744, + "loss": 0.0352, + "step": 15405 + }, + { + "epoch": 0.3611409014457591, + "grad_norm": 0.27925801277160645, + "learning_rate": 0.0001422953887505751, + "loss": 0.0553, + "step": 15406 + }, + { + "epoch": 0.3611643430205641, + "grad_norm": 0.17809361219406128, + "learning_rate": 0.0001422887137830268, + "loss": 0.0288, + "step": 15407 + }, + { + "epoch": 0.3611877845953691, + "grad_norm": 0.32103845477104187, + "learning_rate": 0.00014228203858601874, + "loss": 0.0645, + "step": 15408 + }, + { + "epoch": 0.3612112261701741, + "grad_norm": 0.7683951258659363, + "learning_rate": 0.00014227536315958715, + "loss": 0.0875, + "step": 15409 + }, + { + "epoch": 0.3612346677449791, + "grad_norm": 0.36740732192993164, + "learning_rate": 0.00014226868750376824, + "loss": 0.0941, + "step": 15410 + }, + { + "epoch": 0.3612581093197841, + "grad_norm": 0.28306806087493896, + "learning_rate": 0.00014226201161859824, + "loss": 0.0342, + "step": 15411 + }, + { + "epoch": 0.3612815508945891, + "grad_norm": 0.5644633173942566, + "learning_rate": 0.00014225533550411338, + "loss": 0.0729, + "step": 15412 + }, + { + "epoch": 0.36130499246939407, + "grad_norm": 0.6787816286087036, + "learning_rate": 0.0001422486591603499, + "loss": 0.1071, + "step": 15413 + }, + { + "epoch": 0.36132843404419906, + "grad_norm": 0.18079590797424316, + "learning_rate": 0.000142241982587344, + "loss": 0.035, + "step": 15414 + }, + { + "epoch": 0.36135187561900406, + "grad_norm": 0.5102014541625977, + "learning_rate": 0.0001422353057851319, + "loss": 0.0906, + "step": 15415 + }, + { + "epoch": 0.36137531719380905, + "grad_norm": 0.35396048426628113, + "learning_rate": 0.00014222862875374988, + "loss": 0.0658, + "step": 15416 + }, + { + "epoch": 0.36139875876861405, + "grad_norm": 0.35683003067970276, + "learning_rate": 0.0001422219514932341, + "loss": 0.0352, + "step": 15417 + }, + { + "epoch": 0.3614222003434191, + "grad_norm": 0.4346896708011627, + "learning_rate": 0.0001422152740036208, + "loss": 0.1322, + "step": 15418 + }, + { + "epoch": 0.3614456419182241, + "grad_norm": 0.5763952732086182, + "learning_rate": 0.00014220859628494628, + "loss": 0.1248, + "step": 15419 + }, + { + "epoch": 0.3614690834930291, + "grad_norm": 0.25745880603790283, + "learning_rate": 0.00014220191833724674, + "loss": 0.0368, + "step": 15420 + }, + { + "epoch": 0.3614925250678341, + "grad_norm": 0.6148568391799927, + "learning_rate": 0.00014219524016055837, + "loss": 0.7009, + "step": 15421 + }, + { + "epoch": 0.3615159666426391, + "grad_norm": 0.31205040216445923, + "learning_rate": 0.00014218856175491747, + "loss": 0.0327, + "step": 15422 + }, + { + "epoch": 0.36153940821744407, + "grad_norm": 0.31133726239204407, + "learning_rate": 0.0001421818831203602, + "loss": 0.2731, + "step": 15423 + }, + { + "epoch": 0.36156284979224906, + "grad_norm": 0.13235197961330414, + "learning_rate": 0.0001421752042569229, + "loss": 0.0177, + "step": 15424 + }, + { + "epoch": 0.36158629136705406, + "grad_norm": 0.3039718568325043, + "learning_rate": 0.00014216852516464173, + "loss": 0.0462, + "step": 15425 + }, + { + "epoch": 0.36160973294185905, + "grad_norm": 0.9410830140113831, + "learning_rate": 0.00014216184584355296, + "loss": 0.1623, + "step": 15426 + }, + { + "epoch": 0.36163317451666405, + "grad_norm": 0.4579418897628784, + "learning_rate": 0.00014215516629369286, + "loss": 0.1076, + "step": 15427 + }, + { + "epoch": 0.36165661609146904, + "grad_norm": 0.19180527329444885, + "learning_rate": 0.00014214848651509759, + "loss": 0.027, + "step": 15428 + }, + { + "epoch": 0.36168005766627404, + "grad_norm": 0.17806024849414825, + "learning_rate": 0.00014214180650780347, + "loss": 0.0466, + "step": 15429 + }, + { + "epoch": 0.36170349924107903, + "grad_norm": 0.9922959208488464, + "learning_rate": 0.0001421351262718467, + "loss": 0.21, + "step": 15430 + }, + { + "epoch": 0.361726940815884, + "grad_norm": 0.4311802387237549, + "learning_rate": 0.0001421284458072636, + "loss": 0.0659, + "step": 15431 + }, + { + "epoch": 0.361750382390689, + "grad_norm": 0.6630748510360718, + "learning_rate": 0.00014212176511409032, + "loss": 0.8527, + "step": 15432 + }, + { + "epoch": 0.361773823965494, + "grad_norm": 0.20913435518741608, + "learning_rate": 0.00014211508419236318, + "loss": 0.0368, + "step": 15433 + }, + { + "epoch": 0.361797265540299, + "grad_norm": 0.599877119064331, + "learning_rate": 0.0001421084030421184, + "loss": 0.086, + "step": 15434 + }, + { + "epoch": 0.361820707115104, + "grad_norm": 0.46963608264923096, + "learning_rate": 0.00014210172166339227, + "loss": 0.1287, + "step": 15435 + }, + { + "epoch": 0.361844148689909, + "grad_norm": 0.5754266381263733, + "learning_rate": 0.000142095040056221, + "loss": 0.1466, + "step": 15436 + }, + { + "epoch": 0.361867590264714, + "grad_norm": 0.5173337459564209, + "learning_rate": 0.00014208835822064083, + "loss": 0.671, + "step": 15437 + }, + { + "epoch": 0.361891031839519, + "grad_norm": 0.5156481862068176, + "learning_rate": 0.00014208167615668805, + "loss": 0.0878, + "step": 15438 + }, + { + "epoch": 0.361914473414324, + "grad_norm": 0.5691326856613159, + "learning_rate": 0.00014207499386439896, + "loss": 0.8158, + "step": 15439 + }, + { + "epoch": 0.361937914989129, + "grad_norm": 0.6252326369285583, + "learning_rate": 0.00014206831134380972, + "loss": 0.0956, + "step": 15440 + }, + { + "epoch": 0.36196135656393397, + "grad_norm": 0.6128419637680054, + "learning_rate": 0.00014206162859495665, + "loss": 0.0765, + "step": 15441 + }, + { + "epoch": 0.36198479813873896, + "grad_norm": 0.5915762782096863, + "learning_rate": 0.00014205494561787601, + "loss": 0.1684, + "step": 15442 + }, + { + "epoch": 0.36200823971354396, + "grad_norm": 0.6020991206169128, + "learning_rate": 0.00014204826241260408, + "loss": 0.1074, + "step": 15443 + }, + { + "epoch": 0.36203168128834895, + "grad_norm": 0.28269994258880615, + "learning_rate": 0.00014204157897917705, + "loss": 0.0573, + "step": 15444 + }, + { + "epoch": 0.36205512286315394, + "grad_norm": 0.37044936418533325, + "learning_rate": 0.00014203489531763126, + "loss": 0.0787, + "step": 15445 + }, + { + "epoch": 0.36207856443795894, + "grad_norm": 0.16076311469078064, + "learning_rate": 0.00014202821142800294, + "loss": 0.0369, + "step": 15446 + }, + { + "epoch": 0.36210200601276393, + "grad_norm": 0.900645911693573, + "learning_rate": 0.00014202152731032834, + "loss": 0.0839, + "step": 15447 + }, + { + "epoch": 0.3621254475875689, + "grad_norm": 0.42584794759750366, + "learning_rate": 0.00014201484296464377, + "loss": 0.4873, + "step": 15448 + }, + { + "epoch": 0.3621488891623739, + "grad_norm": 0.49261239171028137, + "learning_rate": 0.0001420081583909855, + "loss": 0.0935, + "step": 15449 + }, + { + "epoch": 0.3621723307371789, + "grad_norm": 0.2802157700061798, + "learning_rate": 0.00014200147358938973, + "loss": 0.0605, + "step": 15450 + }, + { + "epoch": 0.3621957723119839, + "grad_norm": 0.5676197409629822, + "learning_rate": 0.00014199478855989284, + "loss": 0.1306, + "step": 15451 + }, + { + "epoch": 0.3622192138867889, + "grad_norm": 0.4899095296859741, + "learning_rate": 0.00014198810330253104, + "loss": 0.0642, + "step": 15452 + }, + { + "epoch": 0.3622426554615939, + "grad_norm": 0.16815681755542755, + "learning_rate": 0.00014198141781734055, + "loss": 0.0174, + "step": 15453 + }, + { + "epoch": 0.3622660970363989, + "grad_norm": 0.38084885478019714, + "learning_rate": 0.00014197473210435775, + "loss": 0.3133, + "step": 15454 + }, + { + "epoch": 0.3622895386112039, + "grad_norm": 0.754723310470581, + "learning_rate": 0.00014196804616361887, + "loss": 0.1433, + "step": 15455 + }, + { + "epoch": 0.3623129801860089, + "grad_norm": 0.46513840556144714, + "learning_rate": 0.0001419613599951602, + "loss": 0.5836, + "step": 15456 + }, + { + "epoch": 0.3623364217608139, + "grad_norm": 0.41133084893226624, + "learning_rate": 0.000141954673599018, + "loss": 0.0699, + "step": 15457 + }, + { + "epoch": 0.36235986333561887, + "grad_norm": 0.5161324143409729, + "learning_rate": 0.00014194798697522858, + "loss": 0.0992, + "step": 15458 + }, + { + "epoch": 0.36238330491042386, + "grad_norm": 0.2719495892524719, + "learning_rate": 0.0001419413001238282, + "loss": 0.0778, + "step": 15459 + }, + { + "epoch": 0.36240674648522886, + "grad_norm": 0.383050799369812, + "learning_rate": 0.00014193461304485308, + "loss": 0.0795, + "step": 15460 + }, + { + "epoch": 0.36243018806003385, + "grad_norm": 0.5854926705360413, + "learning_rate": 0.00014192792573833966, + "loss": 0.659, + "step": 15461 + }, + { + "epoch": 0.36245362963483885, + "grad_norm": 0.46871864795684814, + "learning_rate": 0.00014192123820432412, + "loss": 0.087, + "step": 15462 + }, + { + "epoch": 0.36247707120964384, + "grad_norm": 0.877346396446228, + "learning_rate": 0.0001419145504428427, + "loss": 0.5713, + "step": 15463 + }, + { + "epoch": 0.36250051278444884, + "grad_norm": 0.3975719213485718, + "learning_rate": 0.00014190786245393182, + "loss": 0.0471, + "step": 15464 + }, + { + "epoch": 0.36252395435925383, + "grad_norm": 0.17519856989383698, + "learning_rate": 0.00014190117423762769, + "loss": 0.0309, + "step": 15465 + }, + { + "epoch": 0.3625473959340588, + "grad_norm": 0.17185275256633759, + "learning_rate": 0.00014189448579396657, + "loss": 0.0142, + "step": 15466 + }, + { + "epoch": 0.3625708375088638, + "grad_norm": 0.31481415033340454, + "learning_rate": 0.0001418877971229848, + "loss": 0.0397, + "step": 15467 + }, + { + "epoch": 0.3625942790836688, + "grad_norm": 0.47840416431427, + "learning_rate": 0.00014188110822471872, + "loss": 0.1169, + "step": 15468 + }, + { + "epoch": 0.36261772065847386, + "grad_norm": 0.4243775010108948, + "learning_rate": 0.0001418744190992045, + "loss": 0.0879, + "step": 15469 + }, + { + "epoch": 0.36264116223327886, + "grad_norm": 0.6687907576560974, + "learning_rate": 0.00014186772974647855, + "loss": 0.6746, + "step": 15470 + }, + { + "epoch": 0.36266460380808385, + "grad_norm": 0.5714360475540161, + "learning_rate": 0.0001418610401665771, + "loss": 0.1214, + "step": 15471 + }, + { + "epoch": 0.36268804538288885, + "grad_norm": 0.5768030881881714, + "learning_rate": 0.00014185435035953648, + "loss": 0.1512, + "step": 15472 + }, + { + "epoch": 0.36271148695769384, + "grad_norm": 0.5132684111595154, + "learning_rate": 0.00014184766032539296, + "loss": 0.1164, + "step": 15473 + }, + { + "epoch": 0.36273492853249883, + "grad_norm": 0.10796112567186356, + "learning_rate": 0.00014184097006418288, + "loss": 0.0263, + "step": 15474 + }, + { + "epoch": 0.36275837010730383, + "grad_norm": 0.5885524153709412, + "learning_rate": 0.0001418342795759425, + "loss": 0.0582, + "step": 15475 + }, + { + "epoch": 0.3627818116821088, + "grad_norm": 0.5009608268737793, + "learning_rate": 0.00014182758886070818, + "loss": 0.0734, + "step": 15476 + }, + { + "epoch": 0.3628052532569138, + "grad_norm": 0.5063156485557556, + "learning_rate": 0.00014182089791851617, + "loss": 0.0862, + "step": 15477 + }, + { + "epoch": 0.3628286948317188, + "grad_norm": 0.6871361136436462, + "learning_rate": 0.0001418142067494028, + "loss": 0.1249, + "step": 15478 + }, + { + "epoch": 0.3628521364065238, + "grad_norm": 0.4815708100795746, + "learning_rate": 0.0001418075153534043, + "loss": 0.1166, + "step": 15479 + }, + { + "epoch": 0.3628755779813288, + "grad_norm": 0.6352282762527466, + "learning_rate": 0.00014180082373055713, + "loss": 0.1853, + "step": 15480 + }, + { + "epoch": 0.3628990195561338, + "grad_norm": 0.5028215646743774, + "learning_rate": 0.00014179413188089747, + "loss": 0.0877, + "step": 15481 + }, + { + "epoch": 0.3629224611309388, + "grad_norm": 0.595508337020874, + "learning_rate": 0.0001417874398044617, + "loss": 0.0784, + "step": 15482 + }, + { + "epoch": 0.3629459027057438, + "grad_norm": 0.3693205714225769, + "learning_rate": 0.0001417807475012861, + "loss": 0.0914, + "step": 15483 + }, + { + "epoch": 0.3629693442805488, + "grad_norm": 0.3818419277667999, + "learning_rate": 0.00014177405497140696, + "loss": 0.0921, + "step": 15484 + }, + { + "epoch": 0.3629927858553538, + "grad_norm": 0.10460471361875534, + "learning_rate": 0.00014176736221486067, + "loss": 0.0168, + "step": 15485 + }, + { + "epoch": 0.36301622743015877, + "grad_norm": 0.48093530535697937, + "learning_rate": 0.00014176066923168347, + "loss": 0.1093, + "step": 15486 + }, + { + "epoch": 0.36303966900496376, + "grad_norm": 0.6060812473297119, + "learning_rate": 0.00014175397602191173, + "loss": 0.7612, + "step": 15487 + }, + { + "epoch": 0.36306311057976876, + "grad_norm": 0.2066987305879593, + "learning_rate": 0.0001417472825855817, + "loss": 0.0285, + "step": 15488 + }, + { + "epoch": 0.36308655215457375, + "grad_norm": 0.21271029114723206, + "learning_rate": 0.00014174058892272976, + "loss": 0.038, + "step": 15489 + }, + { + "epoch": 0.36310999372937874, + "grad_norm": 0.7001416683197021, + "learning_rate": 0.00014173389503339222, + "loss": 0.1142, + "step": 15490 + }, + { + "epoch": 0.36313343530418374, + "grad_norm": 0.6319087147712708, + "learning_rate": 0.00014172720091760538, + "loss": 0.1826, + "step": 15491 + }, + { + "epoch": 0.36315687687898873, + "grad_norm": 0.45454370975494385, + "learning_rate": 0.00014172050657540558, + "loss": 0.0631, + "step": 15492 + }, + { + "epoch": 0.3631803184537937, + "grad_norm": 0.6966940760612488, + "learning_rate": 0.00014171381200682917, + "loss": 0.1627, + "step": 15493 + }, + { + "epoch": 0.3632037600285987, + "grad_norm": 0.1196429654955864, + "learning_rate": 0.00014170711721191238, + "loss": 0.026, + "step": 15494 + }, + { + "epoch": 0.3632272016034037, + "grad_norm": 0.6124982237815857, + "learning_rate": 0.00014170042219069164, + "loss": 0.1107, + "step": 15495 + }, + { + "epoch": 0.3632506431782087, + "grad_norm": 0.39186903834342957, + "learning_rate": 0.00014169372694320322, + "loss": 0.4877, + "step": 15496 + }, + { + "epoch": 0.3632740847530137, + "grad_norm": 0.8362382054328918, + "learning_rate": 0.0001416870314694835, + "loss": 0.1884, + "step": 15497 + }, + { + "epoch": 0.3632975263278187, + "grad_norm": 0.5507273077964783, + "learning_rate": 0.0001416803357695687, + "loss": 0.1554, + "step": 15498 + }, + { + "epoch": 0.3633209679026237, + "grad_norm": 0.5155247449874878, + "learning_rate": 0.0001416736398434953, + "loss": 0.0977, + "step": 15499 + }, + { + "epoch": 0.3633444094774287, + "grad_norm": 0.20177076756954193, + "learning_rate": 0.00014166694369129956, + "loss": 0.0328, + "step": 15500 + }, + { + "epoch": 0.3633678510522337, + "grad_norm": 0.5185129642486572, + "learning_rate": 0.00014166024731301777, + "loss": 0.075, + "step": 15501 + }, + { + "epoch": 0.3633912926270387, + "grad_norm": 0.4916629493236542, + "learning_rate": 0.0001416535507086863, + "loss": 0.0945, + "step": 15502 + }, + { + "epoch": 0.36341473420184367, + "grad_norm": 0.6445127725601196, + "learning_rate": 0.00014164685387834153, + "loss": 0.6851, + "step": 15503 + }, + { + "epoch": 0.36343817577664866, + "grad_norm": 0.42061522603034973, + "learning_rate": 0.00014164015682201973, + "loss": 0.0568, + "step": 15504 + }, + { + "epoch": 0.36346161735145366, + "grad_norm": 0.46476954221725464, + "learning_rate": 0.0001416334595397573, + "loss": 0.086, + "step": 15505 + }, + { + "epoch": 0.36348505892625865, + "grad_norm": 0.5573275089263916, + "learning_rate": 0.00014162676203159053, + "loss": 0.1381, + "step": 15506 + }, + { + "epoch": 0.36350850050106365, + "grad_norm": 0.5159194469451904, + "learning_rate": 0.0001416200642975558, + "loss": 0.133, + "step": 15507 + }, + { + "epoch": 0.36353194207586864, + "grad_norm": 0.18718531727790833, + "learning_rate": 0.0001416133663376894, + "loss": 0.0423, + "step": 15508 + }, + { + "epoch": 0.36355538365067364, + "grad_norm": 0.22966425120830536, + "learning_rate": 0.0001416066681520277, + "loss": 0.0446, + "step": 15509 + }, + { + "epoch": 0.36357882522547863, + "grad_norm": 0.5579330921173096, + "learning_rate": 0.00014159996974060708, + "loss": 0.0686, + "step": 15510 + }, + { + "epoch": 0.3636022668002836, + "grad_norm": 0.4915512502193451, + "learning_rate": 0.00014159327110346384, + "loss": 0.5406, + "step": 15511 + }, + { + "epoch": 0.3636257083750886, + "grad_norm": 0.4210815131664276, + "learning_rate": 0.00014158657224063435, + "loss": 0.0877, + "step": 15512 + }, + { + "epoch": 0.3636491499498936, + "grad_norm": 0.1165648102760315, + "learning_rate": 0.00014157987315215493, + "loss": 0.0172, + "step": 15513 + }, + { + "epoch": 0.3636725915246986, + "grad_norm": 0.6098243594169617, + "learning_rate": 0.00014157317383806195, + "loss": 0.0949, + "step": 15514 + }, + { + "epoch": 0.3636960330995036, + "grad_norm": 0.43626412749290466, + "learning_rate": 0.00014156647429839177, + "loss": 0.0943, + "step": 15515 + }, + { + "epoch": 0.3637194746743086, + "grad_norm": 0.5694379210472107, + "learning_rate": 0.00014155977453318073, + "loss": 0.1397, + "step": 15516 + }, + { + "epoch": 0.3637429162491136, + "grad_norm": 0.5609940886497498, + "learning_rate": 0.00014155307454246518, + "loss": 0.1412, + "step": 15517 + }, + { + "epoch": 0.3637663578239186, + "grad_norm": 0.6588655710220337, + "learning_rate": 0.0001415463743262815, + "loss": 0.0278, + "step": 15518 + }, + { + "epoch": 0.3637897993987236, + "grad_norm": 0.3755647838115692, + "learning_rate": 0.000141539673884666, + "loss": 0.0835, + "step": 15519 + }, + { + "epoch": 0.36381324097352863, + "grad_norm": 0.44741469621658325, + "learning_rate": 0.00014153297321765505, + "loss": 0.1131, + "step": 15520 + }, + { + "epoch": 0.3638366825483336, + "grad_norm": 0.31486237049102783, + "learning_rate": 0.00014152627232528504, + "loss": 0.0985, + "step": 15521 + }, + { + "epoch": 0.3638601241231386, + "grad_norm": 0.6240379214286804, + "learning_rate": 0.00014151957120759235, + "loss": 0.1184, + "step": 15522 + }, + { + "epoch": 0.3638835656979436, + "grad_norm": 0.05595661699771881, + "learning_rate": 0.00014151286986461322, + "loss": 0.0188, + "step": 15523 + }, + { + "epoch": 0.3639070072727486, + "grad_norm": 0.22889845073223114, + "learning_rate": 0.00014150616829638414, + "loss": 0.0463, + "step": 15524 + }, + { + "epoch": 0.3639304488475536, + "grad_norm": 0.3050382137298584, + "learning_rate": 0.00014149946650294143, + "loss": 0.054, + "step": 15525 + }, + { + "epoch": 0.3639538904223586, + "grad_norm": 0.6004399061203003, + "learning_rate": 0.0001414927644843214, + "loss": 0.0597, + "step": 15526 + }, + { + "epoch": 0.3639773319971636, + "grad_norm": 0.4075731337070465, + "learning_rate": 0.0001414860622405605, + "loss": 0.0539, + "step": 15527 + }, + { + "epoch": 0.3640007735719686, + "grad_norm": 0.43292534351348877, + "learning_rate": 0.00014147935977169507, + "loss": 0.1308, + "step": 15528 + }, + { + "epoch": 0.3640242151467736, + "grad_norm": 0.870259702205658, + "learning_rate": 0.00014147265707776144, + "loss": 0.1711, + "step": 15529 + }, + { + "epoch": 0.36404765672157857, + "grad_norm": 0.41374942660331726, + "learning_rate": 0.00014146595415879602, + "loss": 0.5278, + "step": 15530 + }, + { + "epoch": 0.36407109829638357, + "grad_norm": 0.36656567454338074, + "learning_rate": 0.00014145925101483517, + "loss": 0.4885, + "step": 15531 + }, + { + "epoch": 0.36409453987118856, + "grad_norm": 0.338455468416214, + "learning_rate": 0.00014145254764591524, + "loss": 0.0516, + "step": 15532 + }, + { + "epoch": 0.36411798144599355, + "grad_norm": 0.06661961227655411, + "learning_rate": 0.00014144584405207262, + "loss": 0.0159, + "step": 15533 + }, + { + "epoch": 0.36414142302079855, + "grad_norm": 0.31963661313056946, + "learning_rate": 0.0001414391402333437, + "loss": 0.0358, + "step": 15534 + }, + { + "epoch": 0.36416486459560354, + "grad_norm": 0.17980465292930603, + "learning_rate": 0.00014143243618976485, + "loss": 0.0286, + "step": 15535 + }, + { + "epoch": 0.36418830617040854, + "grad_norm": 0.6548458933830261, + "learning_rate": 0.00014142573192137242, + "loss": 0.466, + "step": 15536 + }, + { + "epoch": 0.36421174774521353, + "grad_norm": 0.5241584181785583, + "learning_rate": 0.00014141902742820283, + "loss": 0.0896, + "step": 15537 + }, + { + "epoch": 0.3642351893200185, + "grad_norm": 0.5094272494316101, + "learning_rate": 0.0001414123227102924, + "loss": 0.1229, + "step": 15538 + }, + { + "epoch": 0.3642586308948235, + "grad_norm": 0.6009736657142639, + "learning_rate": 0.00014140561776767753, + "loss": 0.1095, + "step": 15539 + }, + { + "epoch": 0.3642820724696285, + "grad_norm": 0.4593852758407593, + "learning_rate": 0.00014139891260039463, + "loss": 0.1295, + "step": 15540 + }, + { + "epoch": 0.3643055140444335, + "grad_norm": 0.14119097590446472, + "learning_rate": 0.0001413922072084801, + "loss": 0.0319, + "step": 15541 + }, + { + "epoch": 0.3643289556192385, + "grad_norm": 0.2960364520549774, + "learning_rate": 0.00014138550159197027, + "loss": 0.0863, + "step": 15542 + }, + { + "epoch": 0.3643523971940435, + "grad_norm": 0.9205546975135803, + "learning_rate": 0.00014137879575090153, + "loss": 0.1134, + "step": 15543 + }, + { + "epoch": 0.3643758387688485, + "grad_norm": 0.27685192227363586, + "learning_rate": 0.0001413720896853103, + "loss": 0.0519, + "step": 15544 + }, + { + "epoch": 0.3643992803436535, + "grad_norm": 0.3800085484981537, + "learning_rate": 0.00014136538339523292, + "loss": 0.0378, + "step": 15545 + }, + { + "epoch": 0.3644227219184585, + "grad_norm": 0.5069848299026489, + "learning_rate": 0.00014135867688070582, + "loss": 0.1679, + "step": 15546 + }, + { + "epoch": 0.3644461634932635, + "grad_norm": 0.17480358481407166, + "learning_rate": 0.00014135197014176538, + "loss": 0.0167, + "step": 15547 + }, + { + "epoch": 0.36446960506806847, + "grad_norm": 0.429055780172348, + "learning_rate": 0.000141345263178448, + "loss": 0.0519, + "step": 15548 + }, + { + "epoch": 0.36449304664287346, + "grad_norm": 0.6494378447532654, + "learning_rate": 0.00014133855599079005, + "loss": 0.1485, + "step": 15549 + }, + { + "epoch": 0.36451648821767846, + "grad_norm": 0.859595537185669, + "learning_rate": 0.00014133184857882794, + "loss": 0.9873, + "step": 15550 + }, + { + "epoch": 0.36453992979248345, + "grad_norm": 0.3261584937572479, + "learning_rate": 0.00014132514094259804, + "loss": 0.0824, + "step": 15551 + }, + { + "epoch": 0.36456337136728845, + "grad_norm": 0.7191302180290222, + "learning_rate": 0.00014131843308213675, + "loss": 0.1252, + "step": 15552 + }, + { + "epoch": 0.36458681294209344, + "grad_norm": 0.6073728799819946, + "learning_rate": 0.0001413117249974805, + "loss": 0.061, + "step": 15553 + }, + { + "epoch": 0.36461025451689844, + "grad_norm": 0.11187606304883957, + "learning_rate": 0.00014130501668866568, + "loss": 0.0274, + "step": 15554 + }, + { + "epoch": 0.36463369609170343, + "grad_norm": 0.545403778553009, + "learning_rate": 0.00014129830815572868, + "loss": 0.203, + "step": 15555 + }, + { + "epoch": 0.3646571376665084, + "grad_norm": 0.6345133781433105, + "learning_rate": 0.0001412915993987059, + "loss": 0.1939, + "step": 15556 + }, + { + "epoch": 0.3646805792413134, + "grad_norm": 0.676239013671875, + "learning_rate": 0.00014128489041763368, + "loss": 0.1965, + "step": 15557 + }, + { + "epoch": 0.3647040208161184, + "grad_norm": 0.5502156019210815, + "learning_rate": 0.00014127818121254855, + "loss": 0.1447, + "step": 15558 + }, + { + "epoch": 0.3647274623909234, + "grad_norm": 0.5754110217094421, + "learning_rate": 0.00014127147178348684, + "loss": 0.0795, + "step": 15559 + }, + { + "epoch": 0.3647509039657284, + "grad_norm": 0.19957618415355682, + "learning_rate": 0.00014126476213048494, + "loss": 0.0176, + "step": 15560 + }, + { + "epoch": 0.3647743455405334, + "grad_norm": 0.3950125575065613, + "learning_rate": 0.0001412580522535793, + "loss": 0.1333, + "step": 15561 + }, + { + "epoch": 0.3647977871153384, + "grad_norm": 0.5374683141708374, + "learning_rate": 0.00014125134215280628, + "loss": 0.0854, + "step": 15562 + }, + { + "epoch": 0.3648212286901434, + "grad_norm": 0.15224280953407288, + "learning_rate": 0.00014124463182820237, + "loss": 0.04, + "step": 15563 + }, + { + "epoch": 0.3648446702649484, + "grad_norm": 0.25748854875564575, + "learning_rate": 0.00014123792127980389, + "loss": 0.0509, + "step": 15564 + }, + { + "epoch": 0.3648681118397534, + "grad_norm": 0.3485444188117981, + "learning_rate": 0.00014123121050764727, + "loss": 0.0707, + "step": 15565 + }, + { + "epoch": 0.36489155341455837, + "grad_norm": 0.3507135212421417, + "learning_rate": 0.00014122449951176899, + "loss": 0.0883, + "step": 15566 + }, + { + "epoch": 0.36491499498936336, + "grad_norm": 0.24336092174053192, + "learning_rate": 0.00014121778829220537, + "loss": 0.0437, + "step": 15567 + }, + { + "epoch": 0.36493843656416836, + "grad_norm": 0.3824610710144043, + "learning_rate": 0.0001412110768489929, + "loss": 0.1213, + "step": 15568 + }, + { + "epoch": 0.36496187813897335, + "grad_norm": 0.1869295835494995, + "learning_rate": 0.00014120436518216798, + "loss": 0.0254, + "step": 15569 + }, + { + "epoch": 0.36498531971377834, + "grad_norm": 0.42795369029045105, + "learning_rate": 0.000141197653291767, + "loss": 0.0702, + "step": 15570 + }, + { + "epoch": 0.3650087612885834, + "grad_norm": 0.16904529929161072, + "learning_rate": 0.0001411909411778264, + "loss": 0.0299, + "step": 15571 + }, + { + "epoch": 0.3650322028633884, + "grad_norm": 0.08587750792503357, + "learning_rate": 0.00014118422884038258, + "loss": 0.0119, + "step": 15572 + }, + { + "epoch": 0.3650556444381934, + "grad_norm": 0.2696940302848816, + "learning_rate": 0.000141177516279472, + "loss": 0.0509, + "step": 15573 + }, + { + "epoch": 0.3650790860129984, + "grad_norm": 0.3238760828971863, + "learning_rate": 0.00014117080349513103, + "loss": 0.0455, + "step": 15574 + }, + { + "epoch": 0.36510252758780337, + "grad_norm": 0.4656485915184021, + "learning_rate": 0.00014116409048739613, + "loss": 0.0976, + "step": 15575 + }, + { + "epoch": 0.36512596916260837, + "grad_norm": 0.22092017531394958, + "learning_rate": 0.00014115737725630374, + "loss": 0.0369, + "step": 15576 + }, + { + "epoch": 0.36514941073741336, + "grad_norm": 0.23279747366905212, + "learning_rate": 0.00014115066380189023, + "loss": 0.0297, + "step": 15577 + }, + { + "epoch": 0.36517285231221835, + "grad_norm": 0.43240103125572205, + "learning_rate": 0.0001411439501241921, + "loss": 0.1261, + "step": 15578 + }, + { + "epoch": 0.36519629388702335, + "grad_norm": 0.37797242403030396, + "learning_rate": 0.00014113723622324572, + "loss": 0.0586, + "step": 15579 + }, + { + "epoch": 0.36521973546182834, + "grad_norm": 0.5008968114852905, + "learning_rate": 0.00014113052209908754, + "loss": 0.5904, + "step": 15580 + }, + { + "epoch": 0.36524317703663334, + "grad_norm": 0.5816229581832886, + "learning_rate": 0.000141123807751754, + "loss": 0.1093, + "step": 15581 + }, + { + "epoch": 0.36526661861143833, + "grad_norm": 0.2913031578063965, + "learning_rate": 0.0001411170931812815, + "loss": 0.039, + "step": 15582 + }, + { + "epoch": 0.3652900601862433, + "grad_norm": 0.7362316846847534, + "learning_rate": 0.00014111037838770654, + "loss": 0.1581, + "step": 15583 + }, + { + "epoch": 0.3653135017610483, + "grad_norm": 0.25019556283950806, + "learning_rate": 0.00014110366337106544, + "loss": 0.0382, + "step": 15584 + }, + { + "epoch": 0.3653369433358533, + "grad_norm": 0.1989336609840393, + "learning_rate": 0.00014109694813139477, + "loss": 0.0337, + "step": 15585 + }, + { + "epoch": 0.3653603849106583, + "grad_norm": 0.4083382785320282, + "learning_rate": 0.00014109023266873087, + "loss": 0.0566, + "step": 15586 + }, + { + "epoch": 0.3653838264854633, + "grad_norm": 0.501508891582489, + "learning_rate": 0.00014108351698311023, + "loss": 0.1235, + "step": 15587 + }, + { + "epoch": 0.3654072680602683, + "grad_norm": 1.0835509300231934, + "learning_rate": 0.00014107680107456927, + "loss": 0.142, + "step": 15588 + }, + { + "epoch": 0.3654307096350733, + "grad_norm": 0.4671345353126526, + "learning_rate": 0.00014107008494314443, + "loss": 0.1031, + "step": 15589 + }, + { + "epoch": 0.3654541512098783, + "grad_norm": 0.7984727025032043, + "learning_rate": 0.00014106336858887215, + "loss": 0.0812, + "step": 15590 + }, + { + "epoch": 0.3654775927846833, + "grad_norm": 0.4791213572025299, + "learning_rate": 0.00014105665201178888, + "loss": 0.4564, + "step": 15591 + }, + { + "epoch": 0.3655010343594883, + "grad_norm": 0.3646423816680908, + "learning_rate": 0.0001410499352119311, + "loss": 0.0937, + "step": 15592 + }, + { + "epoch": 0.36552447593429327, + "grad_norm": 0.494535356760025, + "learning_rate": 0.00014104321818933515, + "loss": 0.1208, + "step": 15593 + }, + { + "epoch": 0.36554791750909826, + "grad_norm": 0.3979901671409607, + "learning_rate": 0.00014103650094403756, + "loss": 0.0412, + "step": 15594 + }, + { + "epoch": 0.36557135908390326, + "grad_norm": 0.6108511090278625, + "learning_rate": 0.0001410297834760748, + "loss": 0.1463, + "step": 15595 + }, + { + "epoch": 0.36559480065870825, + "grad_norm": 0.272246778011322, + "learning_rate": 0.00014102306578548323, + "loss": 0.055, + "step": 15596 + }, + { + "epoch": 0.36561824223351325, + "grad_norm": 0.5621969103813171, + "learning_rate": 0.00014101634787229937, + "loss": 0.081, + "step": 15597 + }, + { + "epoch": 0.36564168380831824, + "grad_norm": 0.322858989238739, + "learning_rate": 0.00014100962973655968, + "loss": 0.0519, + "step": 15598 + }, + { + "epoch": 0.36566512538312324, + "grad_norm": 0.21088740229606628, + "learning_rate": 0.00014100291137830056, + "loss": 0.0424, + "step": 15599 + }, + { + "epoch": 0.36568856695792823, + "grad_norm": 0.7277148365974426, + "learning_rate": 0.0001409961927975585, + "loss": 0.1507, + "step": 15600 + }, + { + "epoch": 0.3657120085327332, + "grad_norm": 1.1843903064727783, + "learning_rate": 0.00014098947399436993, + "loss": 0.1823, + "step": 15601 + }, + { + "epoch": 0.3657354501075382, + "grad_norm": 0.3574422001838684, + "learning_rate": 0.00014098275496877136, + "loss": 0.0538, + "step": 15602 + }, + { + "epoch": 0.3657588916823432, + "grad_norm": 0.5907973647117615, + "learning_rate": 0.00014097603572079916, + "loss": 0.1531, + "step": 15603 + }, + { + "epoch": 0.3657823332571482, + "grad_norm": 0.38119399547576904, + "learning_rate": 0.0001409693162504899, + "loss": 0.1092, + "step": 15604 + }, + { + "epoch": 0.3658057748319532, + "grad_norm": 0.8487752079963684, + "learning_rate": 0.0001409625965578799, + "loss": 0.1226, + "step": 15605 + }, + { + "epoch": 0.3658292164067582, + "grad_norm": 0.1618645191192627, + "learning_rate": 0.00014095587664300573, + "loss": 0.0332, + "step": 15606 + }, + { + "epoch": 0.3658526579815632, + "grad_norm": 0.13462670147418976, + "learning_rate": 0.00014094915650590383, + "loss": 0.0306, + "step": 15607 + }, + { + "epoch": 0.3658760995563682, + "grad_norm": 0.32581254839897156, + "learning_rate": 0.00014094243614661066, + "loss": 0.0426, + "step": 15608 + }, + { + "epoch": 0.3658995411311732, + "grad_norm": 0.7577112913131714, + "learning_rate": 0.00014093571556516267, + "loss": 0.1512, + "step": 15609 + }, + { + "epoch": 0.3659229827059782, + "grad_norm": 0.45786193013191223, + "learning_rate": 0.00014092899476159635, + "loss": 0.1343, + "step": 15610 + }, + { + "epoch": 0.36594642428078317, + "grad_norm": 0.8131635785102844, + "learning_rate": 0.00014092227373594816, + "loss": 0.0369, + "step": 15611 + }, + { + "epoch": 0.36596986585558816, + "grad_norm": 0.39076101779937744, + "learning_rate": 0.00014091555248825453, + "loss": 0.0647, + "step": 15612 + }, + { + "epoch": 0.36599330743039316, + "grad_norm": 0.22979308664798737, + "learning_rate": 0.00014090883101855196, + "loss": 0.039, + "step": 15613 + }, + { + "epoch": 0.36601674900519815, + "grad_norm": 0.35675856471061707, + "learning_rate": 0.00014090210932687697, + "loss": 0.0605, + "step": 15614 + }, + { + "epoch": 0.36604019058000314, + "grad_norm": 0.14706005156040192, + "learning_rate": 0.00014089538741326595, + "loss": 0.0307, + "step": 15615 + }, + { + "epoch": 0.36606363215480814, + "grad_norm": 0.6547741889953613, + "learning_rate": 0.00014088866527775542, + "loss": 0.0842, + "step": 15616 + }, + { + "epoch": 0.36608707372961313, + "grad_norm": 0.5225555300712585, + "learning_rate": 0.00014088194292038187, + "loss": 0.1228, + "step": 15617 + }, + { + "epoch": 0.3661105153044181, + "grad_norm": 0.5966149568557739, + "learning_rate": 0.00014087522034118168, + "loss": 0.1849, + "step": 15618 + }, + { + "epoch": 0.3661339568792231, + "grad_norm": 0.48523178696632385, + "learning_rate": 0.0001408684975401914, + "loss": 0.1102, + "step": 15619 + }, + { + "epoch": 0.3661573984540281, + "grad_norm": 0.28900137543678284, + "learning_rate": 0.00014086177451744756, + "loss": 0.0551, + "step": 15620 + }, + { + "epoch": 0.3661808400288331, + "grad_norm": 0.177583247423172, + "learning_rate": 0.00014085505127298657, + "loss": 0.0486, + "step": 15621 + }, + { + "epoch": 0.36620428160363816, + "grad_norm": 0.3253939151763916, + "learning_rate": 0.0001408483278068449, + "loss": 0.0575, + "step": 15622 + }, + { + "epoch": 0.36622772317844315, + "grad_norm": 0.5517032742500305, + "learning_rate": 0.0001408416041190591, + "loss": 0.1679, + "step": 15623 + }, + { + "epoch": 0.36625116475324815, + "grad_norm": 0.3296574354171753, + "learning_rate": 0.00014083488020966554, + "loss": 0.0489, + "step": 15624 + }, + { + "epoch": 0.36627460632805314, + "grad_norm": 1.1884357929229736, + "learning_rate": 0.0001408281560787008, + "loss": 0.0945, + "step": 15625 + }, + { + "epoch": 0.36629804790285814, + "grad_norm": 0.5592599511146545, + "learning_rate": 0.00014082143172620135, + "loss": 0.1136, + "step": 15626 + }, + { + "epoch": 0.36632148947766313, + "grad_norm": 0.4707602560520172, + "learning_rate": 0.00014081470715220367, + "loss": 0.0815, + "step": 15627 + }, + { + "epoch": 0.3663449310524681, + "grad_norm": 0.46029922366142273, + "learning_rate": 0.0001408079823567442, + "loss": 0.5277, + "step": 15628 + }, + { + "epoch": 0.3663683726272731, + "grad_norm": 0.3080255687236786, + "learning_rate": 0.00014080125733985951, + "loss": 0.0525, + "step": 15629 + }, + { + "epoch": 0.3663918142020781, + "grad_norm": 0.5816525220870972, + "learning_rate": 0.000140794532101586, + "loss": 0.0896, + "step": 15630 + }, + { + "epoch": 0.3664152557768831, + "grad_norm": 0.08673421293497086, + "learning_rate": 0.00014078780664196025, + "loss": 0.0128, + "step": 15631 + }, + { + "epoch": 0.3664386973516881, + "grad_norm": 0.4386349022388458, + "learning_rate": 0.00014078108096101874, + "loss": 0.065, + "step": 15632 + }, + { + "epoch": 0.3664621389264931, + "grad_norm": 0.6446323394775391, + "learning_rate": 0.00014077435505879793, + "loss": 0.1606, + "step": 15633 + }, + { + "epoch": 0.3664855805012981, + "grad_norm": 0.4838044345378876, + "learning_rate": 0.0001407676289353343, + "loss": 0.0849, + "step": 15634 + }, + { + "epoch": 0.3665090220761031, + "grad_norm": 0.2495291382074356, + "learning_rate": 0.00014076090259066437, + "loss": 0.0534, + "step": 15635 + }, + { + "epoch": 0.3665324636509081, + "grad_norm": 0.3168972134590149, + "learning_rate": 0.00014075417602482461, + "loss": 0.0606, + "step": 15636 + }, + { + "epoch": 0.3665559052257131, + "grad_norm": 0.1348678320646286, + "learning_rate": 0.0001407474492378516, + "loss": 0.0144, + "step": 15637 + }, + { + "epoch": 0.36657934680051807, + "grad_norm": 0.7818272709846497, + "learning_rate": 0.00014074072222978174, + "loss": 0.1073, + "step": 15638 + }, + { + "epoch": 0.36660278837532306, + "grad_norm": 0.4256865680217743, + "learning_rate": 0.0001407339950006516, + "loss": 0.0918, + "step": 15639 + }, + { + "epoch": 0.36662622995012806, + "grad_norm": 0.5562074780464172, + "learning_rate": 0.00014072726755049768, + "loss": 0.0907, + "step": 15640 + }, + { + "epoch": 0.36664967152493305, + "grad_norm": 0.5605216026306152, + "learning_rate": 0.0001407205398793564, + "loss": 0.0912, + "step": 15641 + }, + { + "epoch": 0.36667311309973805, + "grad_norm": 0.5251980423927307, + "learning_rate": 0.00014071381198726437, + "loss": 0.1264, + "step": 15642 + }, + { + "epoch": 0.36669655467454304, + "grad_norm": 0.50408536195755, + "learning_rate": 0.00014070708387425805, + "loss": 0.0942, + "step": 15643 + }, + { + "epoch": 0.36671999624934803, + "grad_norm": 0.7406957745552063, + "learning_rate": 0.0001407003555403739, + "loss": 0.1555, + "step": 15644 + }, + { + "epoch": 0.36674343782415303, + "grad_norm": 0.540463387966156, + "learning_rate": 0.00014069362698564854, + "loss": 0.0974, + "step": 15645 + }, + { + "epoch": 0.366766879398958, + "grad_norm": 0.5115364193916321, + "learning_rate": 0.0001406868982101184, + "loss": 0.0957, + "step": 15646 + }, + { + "epoch": 0.366790320973763, + "grad_norm": 0.4497169852256775, + "learning_rate": 0.00014068016921382, + "loss": 0.1083, + "step": 15647 + }, + { + "epoch": 0.366813762548568, + "grad_norm": 0.11463009566068649, + "learning_rate": 0.00014067343999678983, + "loss": 0.0135, + "step": 15648 + }, + { + "epoch": 0.366837204123373, + "grad_norm": 0.4508204162120819, + "learning_rate": 0.00014066671055906445, + "loss": 0.1027, + "step": 15649 + }, + { + "epoch": 0.366860645698178, + "grad_norm": 0.808026909828186, + "learning_rate": 0.00014065998090068038, + "loss": 0.1163, + "step": 15650 + }, + { + "epoch": 0.366884087272983, + "grad_norm": 0.47886940836906433, + "learning_rate": 0.0001406532510216741, + "loss": 0.0963, + "step": 15651 + }, + { + "epoch": 0.366907528847788, + "grad_norm": 0.5793461799621582, + "learning_rate": 0.00014064652092208213, + "loss": 0.0833, + "step": 15652 + }, + { + "epoch": 0.366930970422593, + "grad_norm": 0.6637849807739258, + "learning_rate": 0.000140639790601941, + "loss": 0.1397, + "step": 15653 + }, + { + "epoch": 0.366954411997398, + "grad_norm": 0.5858306288719177, + "learning_rate": 0.0001406330600612872, + "loss": 0.1571, + "step": 15654 + }, + { + "epoch": 0.36697785357220297, + "grad_norm": 0.865097165107727, + "learning_rate": 0.0001406263293001573, + "loss": 0.1308, + "step": 15655 + }, + { + "epoch": 0.36700129514700797, + "grad_norm": 0.33646589517593384, + "learning_rate": 0.0001406195983185878, + "loss": 0.0771, + "step": 15656 + }, + { + "epoch": 0.36702473672181296, + "grad_norm": 0.5230391621589661, + "learning_rate": 0.0001406128671166152, + "loss": 0.1705, + "step": 15657 + }, + { + "epoch": 0.36704817829661796, + "grad_norm": 0.4527624547481537, + "learning_rate": 0.00014060613569427604, + "loss": 0.0821, + "step": 15658 + }, + { + "epoch": 0.36707161987142295, + "grad_norm": 0.8595010638237, + "learning_rate": 0.00014059940405160686, + "loss": 0.635, + "step": 15659 + }, + { + "epoch": 0.36709506144622794, + "grad_norm": 0.6062330007553101, + "learning_rate": 0.00014059267218864411, + "loss": 0.6451, + "step": 15660 + }, + { + "epoch": 0.36711850302103294, + "grad_norm": 0.2700575292110443, + "learning_rate": 0.00014058594010542445, + "loss": 0.0703, + "step": 15661 + }, + { + "epoch": 0.36714194459583793, + "grad_norm": 0.6856800317764282, + "learning_rate": 0.00014057920780198433, + "loss": 0.1477, + "step": 15662 + }, + { + "epoch": 0.3671653861706429, + "grad_norm": 0.988795816898346, + "learning_rate": 0.00014057247527836026, + "loss": 0.4235, + "step": 15663 + }, + { + "epoch": 0.3671888277454479, + "grad_norm": 0.4080718457698822, + "learning_rate": 0.00014056574253458882, + "loss": 0.0688, + "step": 15664 + }, + { + "epoch": 0.3672122693202529, + "grad_norm": 0.7141121029853821, + "learning_rate": 0.00014055900957070654, + "loss": 0.1315, + "step": 15665 + }, + { + "epoch": 0.3672357108950579, + "grad_norm": 0.46973201632499695, + "learning_rate": 0.0001405522763867499, + "loss": 0.744, + "step": 15666 + }, + { + "epoch": 0.3672591524698629, + "grad_norm": 0.562660813331604, + "learning_rate": 0.00014054554298275547, + "loss": 0.173, + "step": 15667 + }, + { + "epoch": 0.3672825940446679, + "grad_norm": 0.46259400248527527, + "learning_rate": 0.0001405388093587598, + "loss": 0.1483, + "step": 15668 + }, + { + "epoch": 0.3673060356194729, + "grad_norm": 0.3516181409358978, + "learning_rate": 0.0001405320755147994, + "loss": 0.0582, + "step": 15669 + }, + { + "epoch": 0.3673294771942779, + "grad_norm": 0.1406058371067047, + "learning_rate": 0.0001405253414509108, + "loss": 0.03, + "step": 15670 + }, + { + "epoch": 0.3673529187690829, + "grad_norm": 0.2960655987262726, + "learning_rate": 0.00014051860716713058, + "loss": 0.0312, + "step": 15671 + }, + { + "epoch": 0.3673763603438879, + "grad_norm": 0.4982856214046478, + "learning_rate": 0.00014051187266349527, + "loss": 0.0743, + "step": 15672 + }, + { + "epoch": 0.3673998019186929, + "grad_norm": 0.37555837631225586, + "learning_rate": 0.00014050513794004136, + "loss": 0.3791, + "step": 15673 + }, + { + "epoch": 0.3674232434934979, + "grad_norm": 0.3977750241756439, + "learning_rate": 0.0001404984029968055, + "loss": 0.0816, + "step": 15674 + }, + { + "epoch": 0.3674466850683029, + "grad_norm": 0.2400004118680954, + "learning_rate": 0.00014049166783382412, + "loss": 0.0633, + "step": 15675 + }, + { + "epoch": 0.3674701266431079, + "grad_norm": 0.33935806155204773, + "learning_rate": 0.0001404849324511338, + "loss": 0.0948, + "step": 15676 + }, + { + "epoch": 0.3674935682179129, + "grad_norm": 0.7917968034744263, + "learning_rate": 0.00014047819684877112, + "loss": 0.4347, + "step": 15677 + }, + { + "epoch": 0.3675170097927179, + "grad_norm": 0.4559464156627655, + "learning_rate": 0.00014047146102677262, + "loss": 0.0797, + "step": 15678 + }, + { + "epoch": 0.3675404513675229, + "grad_norm": 0.11926800012588501, + "learning_rate": 0.0001404647249851748, + "loss": 0.0164, + "step": 15679 + }, + { + "epoch": 0.3675638929423279, + "grad_norm": 0.7176985740661621, + "learning_rate": 0.00014045798872401426, + "loss": 0.1587, + "step": 15680 + }, + { + "epoch": 0.3675873345171329, + "grad_norm": 0.19977350533008575, + "learning_rate": 0.00014045125224332753, + "loss": 0.0302, + "step": 15681 + }, + { + "epoch": 0.3676107760919379, + "grad_norm": 0.1531054526567459, + "learning_rate": 0.00014044451554315118, + "loss": 0.0394, + "step": 15682 + }, + { + "epoch": 0.36763421766674287, + "grad_norm": 0.7286527752876282, + "learning_rate": 0.00014043777862352173, + "loss": 0.0837, + "step": 15683 + }, + { + "epoch": 0.36765765924154786, + "grad_norm": 0.23311936855316162, + "learning_rate": 0.0001404310414844758, + "loss": 0.0432, + "step": 15684 + }, + { + "epoch": 0.36768110081635286, + "grad_norm": 0.15097454190254211, + "learning_rate": 0.00014042430412604983, + "loss": 0.0349, + "step": 15685 + }, + { + "epoch": 0.36770454239115785, + "grad_norm": 0.44809386134147644, + "learning_rate": 0.0001404175665482805, + "loss": 0.1281, + "step": 15686 + }, + { + "epoch": 0.36772798396596285, + "grad_norm": 1.4547187089920044, + "learning_rate": 0.00014041082875120432, + "loss": 0.2244, + "step": 15687 + }, + { + "epoch": 0.36775142554076784, + "grad_norm": 0.5763091444969177, + "learning_rate": 0.00014040409073485784, + "loss": 0.1741, + "step": 15688 + }, + { + "epoch": 0.36777486711557283, + "grad_norm": 0.6363410353660583, + "learning_rate": 0.0001403973524992776, + "loss": 0.0946, + "step": 15689 + }, + { + "epoch": 0.36779830869037783, + "grad_norm": 0.5935019850730896, + "learning_rate": 0.00014039061404450025, + "loss": 0.0926, + "step": 15690 + }, + { + "epoch": 0.3678217502651828, + "grad_norm": 0.5978676676750183, + "learning_rate": 0.00014038387537056222, + "loss": 0.1306, + "step": 15691 + }, + { + "epoch": 0.3678451918399878, + "grad_norm": 0.287048876285553, + "learning_rate": 0.0001403771364775002, + "loss": 0.0626, + "step": 15692 + }, + { + "epoch": 0.3678686334147928, + "grad_norm": 0.4811996817588806, + "learning_rate": 0.00014037039736535068, + "loss": 0.0931, + "step": 15693 + }, + { + "epoch": 0.3678920749895978, + "grad_norm": 0.24521414935588837, + "learning_rate": 0.0001403636580341503, + "loss": 0.0282, + "step": 15694 + }, + { + "epoch": 0.3679155165644028, + "grad_norm": 0.8201712965965271, + "learning_rate": 0.0001403569184839355, + "loss": 0.1301, + "step": 15695 + }, + { + "epoch": 0.3679389581392078, + "grad_norm": 0.5978713631629944, + "learning_rate": 0.00014035017871474295, + "loss": 0.1236, + "step": 15696 + }, + { + "epoch": 0.3679623997140128, + "grad_norm": 0.5310399532318115, + "learning_rate": 0.0001403434387266092, + "loss": 0.0579, + "step": 15697 + }, + { + "epoch": 0.3679858412888178, + "grad_norm": 0.4245557188987732, + "learning_rate": 0.00014033669851957082, + "loss": 0.0665, + "step": 15698 + }, + { + "epoch": 0.3680092828636228, + "grad_norm": 1.0509849786758423, + "learning_rate": 0.00014032995809366437, + "loss": 0.1993, + "step": 15699 + }, + { + "epoch": 0.36803272443842777, + "grad_norm": 0.3764999806880951, + "learning_rate": 0.00014032321744892644, + "loss": 0.0803, + "step": 15700 + }, + { + "epoch": 0.36805616601323277, + "grad_norm": 0.5524464845657349, + "learning_rate": 0.00014031647658539363, + "loss": 0.667, + "step": 15701 + }, + { + "epoch": 0.36807960758803776, + "grad_norm": 0.5427095293998718, + "learning_rate": 0.00014030973550310243, + "loss": 0.1742, + "step": 15702 + }, + { + "epoch": 0.36810304916284275, + "grad_norm": 0.22043101489543915, + "learning_rate": 0.0001403029942020895, + "loss": 0.0341, + "step": 15703 + }, + { + "epoch": 0.36812649073764775, + "grad_norm": 0.3937276601791382, + "learning_rate": 0.00014029625268239138, + "loss": 0.0862, + "step": 15704 + }, + { + "epoch": 0.36814993231245274, + "grad_norm": 0.3440457880496979, + "learning_rate": 0.00014028951094404468, + "loss": 0.0792, + "step": 15705 + }, + { + "epoch": 0.36817337388725774, + "grad_norm": 0.4949117600917816, + "learning_rate": 0.00014028276898708593, + "loss": 0.0955, + "step": 15706 + }, + { + "epoch": 0.36819681546206273, + "grad_norm": 0.48723405599594116, + "learning_rate": 0.00014027602681155175, + "loss": 0.1224, + "step": 15707 + }, + { + "epoch": 0.3682202570368677, + "grad_norm": 0.5810900926589966, + "learning_rate": 0.00014026928441747874, + "loss": 0.143, + "step": 15708 + }, + { + "epoch": 0.3682436986116727, + "grad_norm": 0.30988582968711853, + "learning_rate": 0.00014026254180490345, + "loss": 0.0503, + "step": 15709 + }, + { + "epoch": 0.3682671401864777, + "grad_norm": 0.26761624217033386, + "learning_rate": 0.00014025579897386248, + "loss": 0.0542, + "step": 15710 + }, + { + "epoch": 0.3682905817612827, + "grad_norm": 0.545335590839386, + "learning_rate": 0.0001402490559243924, + "loss": 0.102, + "step": 15711 + }, + { + "epoch": 0.3683140233360877, + "grad_norm": 0.1420464664697647, + "learning_rate": 0.00014024231265652982, + "loss": 0.0242, + "step": 15712 + }, + { + "epoch": 0.3683374649108927, + "grad_norm": 0.7197564840316772, + "learning_rate": 0.00014023556917031133, + "loss": 0.1304, + "step": 15713 + }, + { + "epoch": 0.3683609064856977, + "grad_norm": 0.19550766050815582, + "learning_rate": 0.0001402288254657735, + "loss": 0.0298, + "step": 15714 + }, + { + "epoch": 0.3683843480605027, + "grad_norm": 0.6190772652626038, + "learning_rate": 0.0001402220815429529, + "loss": 0.1279, + "step": 15715 + }, + { + "epoch": 0.3684077896353077, + "grad_norm": 0.5742803812026978, + "learning_rate": 0.0001402153374018862, + "loss": 0.6642, + "step": 15716 + }, + { + "epoch": 0.3684312312101127, + "grad_norm": 0.3439716398715973, + "learning_rate": 0.0001402085930426099, + "loss": 0.044, + "step": 15717 + }, + { + "epoch": 0.36845467278491767, + "grad_norm": 0.3218945562839508, + "learning_rate": 0.00014020184846516071, + "loss": 0.0799, + "step": 15718 + }, + { + "epoch": 0.36847811435972266, + "grad_norm": 0.16024072468280792, + "learning_rate": 0.00014019510366957514, + "loss": 0.0231, + "step": 15719 + }, + { + "epoch": 0.36850155593452766, + "grad_norm": 0.6780598163604736, + "learning_rate": 0.00014018835865588978, + "loss": 0.0387, + "step": 15720 + }, + { + "epoch": 0.36852499750933265, + "grad_norm": 0.6467726826667786, + "learning_rate": 0.00014018161342414125, + "loss": 0.1407, + "step": 15721 + }, + { + "epoch": 0.36854843908413765, + "grad_norm": 0.5249501466751099, + "learning_rate": 0.0001401748679743662, + "loss": 0.1439, + "step": 15722 + }, + { + "epoch": 0.36857188065894264, + "grad_norm": 0.5149779915809631, + "learning_rate": 0.00014016812230660116, + "loss": 0.654, + "step": 15723 + }, + { + "epoch": 0.3685953222337477, + "grad_norm": 0.3883340656757355, + "learning_rate": 0.00014016137642088275, + "loss": 0.0801, + "step": 15724 + }, + { + "epoch": 0.3686187638085527, + "grad_norm": 0.4186076819896698, + "learning_rate": 0.00014015463031724762, + "loss": 0.0934, + "step": 15725 + }, + { + "epoch": 0.3686422053833577, + "grad_norm": 0.9316620826721191, + "learning_rate": 0.0001401478839957323, + "loss": 0.158, + "step": 15726 + }, + { + "epoch": 0.3686656469581627, + "grad_norm": 0.3597972095012665, + "learning_rate": 0.00014014113745637342, + "loss": 0.0675, + "step": 15727 + }, + { + "epoch": 0.36868908853296767, + "grad_norm": 0.6425904631614685, + "learning_rate": 0.00014013439069920762, + "loss": 0.0596, + "step": 15728 + }, + { + "epoch": 0.36871253010777266, + "grad_norm": 0.11600496619939804, + "learning_rate": 0.0001401276437242715, + "loss": 0.0151, + "step": 15729 + }, + { + "epoch": 0.36873597168257766, + "grad_norm": 0.17142656445503235, + "learning_rate": 0.00014012089653160162, + "loss": 0.0348, + "step": 15730 + }, + { + "epoch": 0.36875941325738265, + "grad_norm": 0.7418240904808044, + "learning_rate": 0.0001401141491212347, + "loss": 0.7916, + "step": 15731 + }, + { + "epoch": 0.36878285483218765, + "grad_norm": 0.08015631884336472, + "learning_rate": 0.00014010740149320722, + "loss": 0.0141, + "step": 15732 + }, + { + "epoch": 0.36880629640699264, + "grad_norm": 0.24502895772457123, + "learning_rate": 0.00014010065364755583, + "loss": 0.072, + "step": 15733 + }, + { + "epoch": 0.36882973798179763, + "grad_norm": 0.43550097942352295, + "learning_rate": 0.0001400939055843172, + "loss": 0.0467, + "step": 15734 + }, + { + "epoch": 0.36885317955660263, + "grad_norm": 0.49391186237335205, + "learning_rate": 0.00014008715730352793, + "loss": 0.1596, + "step": 15735 + }, + { + "epoch": 0.3688766211314076, + "grad_norm": 0.3948333263397217, + "learning_rate": 0.0001400804088052246, + "loss": 0.0727, + "step": 15736 + }, + { + "epoch": 0.3689000627062126, + "grad_norm": 0.4813016355037689, + "learning_rate": 0.00014007366008944385, + "loss": 0.1141, + "step": 15737 + }, + { + "epoch": 0.3689235042810176, + "grad_norm": 0.3098224997520447, + "learning_rate": 0.00014006691115622227, + "loss": 0.0893, + "step": 15738 + }, + { + "epoch": 0.3689469458558226, + "grad_norm": 0.5839622020721436, + "learning_rate": 0.00014006016200559653, + "loss": 0.1178, + "step": 15739 + }, + { + "epoch": 0.3689703874306276, + "grad_norm": 0.5472758412361145, + "learning_rate": 0.00014005341263760319, + "loss": 0.1095, + "step": 15740 + }, + { + "epoch": 0.3689938290054326, + "grad_norm": 0.07441138476133347, + "learning_rate": 0.00014004666305227894, + "loss": 0.0132, + "step": 15741 + }, + { + "epoch": 0.3690172705802376, + "grad_norm": 0.16399739682674408, + "learning_rate": 0.00014003991324966036, + "loss": 0.025, + "step": 15742 + }, + { + "epoch": 0.3690407121550426, + "grad_norm": 0.5877649188041687, + "learning_rate": 0.0001400331632297841, + "loss": 0.1345, + "step": 15743 + }, + { + "epoch": 0.3690641537298476, + "grad_norm": 0.6421027779579163, + "learning_rate": 0.00014002641299268675, + "loss": 0.1423, + "step": 15744 + }, + { + "epoch": 0.36908759530465257, + "grad_norm": 0.46737560629844666, + "learning_rate": 0.00014001966253840496, + "loss": 0.1055, + "step": 15745 + }, + { + "epoch": 0.36911103687945757, + "grad_norm": 1.4370903968811035, + "learning_rate": 0.00014001291186697536, + "loss": 0.1728, + "step": 15746 + }, + { + "epoch": 0.36913447845426256, + "grad_norm": 0.7164363265037537, + "learning_rate": 0.00014000616097843458, + "loss": 0.1688, + "step": 15747 + }, + { + "epoch": 0.36915792002906755, + "grad_norm": 0.5661167502403259, + "learning_rate": 0.00013999940987281925, + "loss": 0.1193, + "step": 15748 + }, + { + "epoch": 0.36918136160387255, + "grad_norm": 0.49718061089515686, + "learning_rate": 0.000139992658550166, + "loss": 0.0545, + "step": 15749 + }, + { + "epoch": 0.36920480317867754, + "grad_norm": 0.17041155695915222, + "learning_rate": 0.00013998590701051145, + "loss": 0.0461, + "step": 15750 + }, + { + "epoch": 0.36922824475348254, + "grad_norm": 0.5700314044952393, + "learning_rate": 0.00013997915525389224, + "loss": 0.1453, + "step": 15751 + }, + { + "epoch": 0.36925168632828753, + "grad_norm": 0.50826495885849, + "learning_rate": 0.00013997240328034497, + "loss": 0.1252, + "step": 15752 + }, + { + "epoch": 0.3692751279030925, + "grad_norm": 0.05047234892845154, + "learning_rate": 0.0001399656510899064, + "loss": 0.0133, + "step": 15753 + }, + { + "epoch": 0.3692985694778975, + "grad_norm": 0.8948851227760315, + "learning_rate": 0.00013995889868261303, + "loss": 0.2133, + "step": 15754 + }, + { + "epoch": 0.3693220110527025, + "grad_norm": 0.140593022108078, + "learning_rate": 0.00013995214605850155, + "loss": 0.0238, + "step": 15755 + }, + { + "epoch": 0.3693454526275075, + "grad_norm": 0.16334550082683563, + "learning_rate": 0.00013994539321760864, + "loss": 0.0385, + "step": 15756 + }, + { + "epoch": 0.3693688942023125, + "grad_norm": 0.5622140169143677, + "learning_rate": 0.0001399386401599709, + "loss": 0.1031, + "step": 15757 + }, + { + "epoch": 0.3693923357771175, + "grad_norm": 0.7819709777832031, + "learning_rate": 0.00013993188688562493, + "loss": 0.1367, + "step": 15758 + }, + { + "epoch": 0.3694157773519225, + "grad_norm": 0.5711849927902222, + "learning_rate": 0.00013992513339460745, + "loss": 0.0997, + "step": 15759 + }, + { + "epoch": 0.3694392189267275, + "grad_norm": 0.39883342385292053, + "learning_rate": 0.00013991837968695508, + "loss": 0.0642, + "step": 15760 + }, + { + "epoch": 0.3694626605015325, + "grad_norm": 0.2920113503932953, + "learning_rate": 0.0001399116257627044, + "loss": 0.0426, + "step": 15761 + }, + { + "epoch": 0.3694861020763375, + "grad_norm": 0.36987587809562683, + "learning_rate": 0.0001399048716218922, + "loss": 0.0794, + "step": 15762 + }, + { + "epoch": 0.36950954365114247, + "grad_norm": 0.4925435185432434, + "learning_rate": 0.000139898117264555, + "loss": 0.0641, + "step": 15763 + }, + { + "epoch": 0.36953298522594746, + "grad_norm": 0.33620500564575195, + "learning_rate": 0.00013989136269072948, + "loss": 0.0754, + "step": 15764 + }, + { + "epoch": 0.36955642680075246, + "grad_norm": 0.7374610304832458, + "learning_rate": 0.0001398846079004523, + "loss": 0.1229, + "step": 15765 + }, + { + "epoch": 0.36957986837555745, + "grad_norm": 0.20942173898220062, + "learning_rate": 0.00013987785289376018, + "loss": 0.0388, + "step": 15766 + }, + { + "epoch": 0.36960330995036245, + "grad_norm": 0.1445106565952301, + "learning_rate": 0.0001398710976706897, + "loss": 0.0331, + "step": 15767 + }, + { + "epoch": 0.36962675152516744, + "grad_norm": 0.22528387606143951, + "learning_rate": 0.00013986434223127746, + "loss": 0.0672, + "step": 15768 + }, + { + "epoch": 0.36965019309997244, + "grad_norm": 0.20863159000873566, + "learning_rate": 0.00013985758657556022, + "loss": 0.0238, + "step": 15769 + }, + { + "epoch": 0.36967363467477743, + "grad_norm": 0.7259862422943115, + "learning_rate": 0.0001398508307035746, + "loss": 0.1643, + "step": 15770 + }, + { + "epoch": 0.3696970762495824, + "grad_norm": 0.5282078981399536, + "learning_rate": 0.0001398440746153572, + "loss": 0.0866, + "step": 15771 + }, + { + "epoch": 0.3697205178243874, + "grad_norm": 0.5220518708229065, + "learning_rate": 0.0001398373183109448, + "loss": 0.094, + "step": 15772 + }, + { + "epoch": 0.3697439593991924, + "grad_norm": 0.272866815328598, + "learning_rate": 0.00013983056179037398, + "loss": 0.0497, + "step": 15773 + }, + { + "epoch": 0.3697674009739974, + "grad_norm": 0.22268635034561157, + "learning_rate": 0.00013982380505368134, + "loss": 0.0483, + "step": 15774 + }, + { + "epoch": 0.3697908425488024, + "grad_norm": 0.667029857635498, + "learning_rate": 0.0001398170481009037, + "loss": 0.1482, + "step": 15775 + }, + { + "epoch": 0.36981428412360745, + "grad_norm": 0.597487211227417, + "learning_rate": 0.00013981029093207761, + "loss": 0.1218, + "step": 15776 + }, + { + "epoch": 0.36983772569841245, + "grad_norm": 0.253761351108551, + "learning_rate": 0.00013980353354723976, + "loss": 0.0281, + "step": 15777 + }, + { + "epoch": 0.36986116727321744, + "grad_norm": 0.6095085144042969, + "learning_rate": 0.0001397967759464268, + "loss": 0.1727, + "step": 15778 + }, + { + "epoch": 0.36988460884802243, + "grad_norm": 0.37292957305908203, + "learning_rate": 0.00013979001812967543, + "loss": 0.0753, + "step": 15779 + }, + { + "epoch": 0.36990805042282743, + "grad_norm": 0.7254984974861145, + "learning_rate": 0.0001397832600970223, + "loss": 0.0955, + "step": 15780 + }, + { + "epoch": 0.3699314919976324, + "grad_norm": 0.6463655829429626, + "learning_rate": 0.0001397765018485041, + "loss": 0.1342, + "step": 15781 + }, + { + "epoch": 0.3699549335724374, + "grad_norm": 0.12205230444669724, + "learning_rate": 0.00013976974338415744, + "loss": 0.0188, + "step": 15782 + }, + { + "epoch": 0.3699783751472424, + "grad_norm": 0.32817748188972473, + "learning_rate": 0.00013976298470401907, + "loss": 0.3161, + "step": 15783 + }, + { + "epoch": 0.3700018167220474, + "grad_norm": 0.27028360962867737, + "learning_rate": 0.0001397562258081256, + "loss": 0.0543, + "step": 15784 + }, + { + "epoch": 0.3700252582968524, + "grad_norm": 0.5839598178863525, + "learning_rate": 0.00013974946669651376, + "loss": 0.131, + "step": 15785 + }, + { + "epoch": 0.3700486998716574, + "grad_norm": 0.3714865744113922, + "learning_rate": 0.00013974270736922018, + "loss": 0.0489, + "step": 15786 + }, + { + "epoch": 0.3700721414464624, + "grad_norm": 0.4927649199962616, + "learning_rate": 0.00013973594782628156, + "loss": 0.1387, + "step": 15787 + }, + { + "epoch": 0.3700955830212674, + "grad_norm": 0.6689778566360474, + "learning_rate": 0.00013972918806773457, + "loss": 0.6717, + "step": 15788 + }, + { + "epoch": 0.3701190245960724, + "grad_norm": 0.549563467502594, + "learning_rate": 0.0001397224280936159, + "loss": 0.1048, + "step": 15789 + }, + { + "epoch": 0.37014246617087737, + "grad_norm": 0.5104742050170898, + "learning_rate": 0.00013971566790396217, + "loss": 0.1116, + "step": 15790 + }, + { + "epoch": 0.37016590774568237, + "grad_norm": 0.23229192197322845, + "learning_rate": 0.00013970890749881017, + "loss": 0.0623, + "step": 15791 + }, + { + "epoch": 0.37018934932048736, + "grad_norm": 0.4353477954864502, + "learning_rate": 0.0001397021468781965, + "loss": 0.1378, + "step": 15792 + }, + { + "epoch": 0.37021279089529235, + "grad_norm": 0.4461476504802704, + "learning_rate": 0.00013969538604215783, + "loss": 0.0897, + "step": 15793 + }, + { + "epoch": 0.37023623247009735, + "grad_norm": 0.6090304851531982, + "learning_rate": 0.0001396886249907309, + "loss": 0.0821, + "step": 15794 + }, + { + "epoch": 0.37025967404490234, + "grad_norm": 0.5100122690200806, + "learning_rate": 0.0001396818637239524, + "loss": 0.0725, + "step": 15795 + }, + { + "epoch": 0.37028311561970734, + "grad_norm": 0.5611236691474915, + "learning_rate": 0.000139675102241859, + "loss": 0.7706, + "step": 15796 + }, + { + "epoch": 0.37030655719451233, + "grad_norm": 0.12949787080287933, + "learning_rate": 0.00013966834054448732, + "loss": 0.0167, + "step": 15797 + }, + { + "epoch": 0.3703299987693173, + "grad_norm": 0.368595153093338, + "learning_rate": 0.00013966157863187415, + "loss": 0.0716, + "step": 15798 + }, + { + "epoch": 0.3703534403441223, + "grad_norm": 0.5461193323135376, + "learning_rate": 0.00013965481650405614, + "loss": 0.1344, + "step": 15799 + }, + { + "epoch": 0.3703768819189273, + "grad_norm": 0.34654274582862854, + "learning_rate": 0.00013964805416106995, + "loss": 0.075, + "step": 15800 + }, + { + "epoch": 0.3704003234937323, + "grad_norm": 0.2349996566772461, + "learning_rate": 0.00013964129160295234, + "loss": 0.0502, + "step": 15801 + }, + { + "epoch": 0.3704237650685373, + "grad_norm": 0.6299496293067932, + "learning_rate": 0.00013963452882973995, + "loss": 0.503, + "step": 15802 + }, + { + "epoch": 0.3704472066433423, + "grad_norm": 0.5710389018058777, + "learning_rate": 0.00013962776584146947, + "loss": 0.0529, + "step": 15803 + }, + { + "epoch": 0.3704706482181473, + "grad_norm": 0.33670374751091003, + "learning_rate": 0.0001396210026381777, + "loss": 0.0714, + "step": 15804 + }, + { + "epoch": 0.3704940897929523, + "grad_norm": 0.33326295018196106, + "learning_rate": 0.00013961423921990117, + "loss": 0.0584, + "step": 15805 + }, + { + "epoch": 0.3705175313677573, + "grad_norm": 0.5865218043327332, + "learning_rate": 0.00013960747558667666, + "loss": 0.1409, + "step": 15806 + }, + { + "epoch": 0.3705409729425623, + "grad_norm": 0.7002782225608826, + "learning_rate": 0.0001396007117385409, + "loss": 0.1806, + "step": 15807 + }, + { + "epoch": 0.37056441451736727, + "grad_norm": 0.38382235169410706, + "learning_rate": 0.00013959394767553064, + "loss": 0.0439, + "step": 15808 + }, + { + "epoch": 0.37058785609217226, + "grad_norm": 0.383134126663208, + "learning_rate": 0.0001395871833976824, + "loss": 0.0771, + "step": 15809 + }, + { + "epoch": 0.37061129766697726, + "grad_norm": 0.6474619507789612, + "learning_rate": 0.00013958041890503305, + "loss": 0.1781, + "step": 15810 + }, + { + "epoch": 0.37063473924178225, + "grad_norm": 0.31945788860321045, + "learning_rate": 0.0001395736541976192, + "loss": 0.0634, + "step": 15811 + }, + { + "epoch": 0.37065818081658725, + "grad_norm": 0.590683102607727, + "learning_rate": 0.0001395668892754776, + "loss": 0.3178, + "step": 15812 + }, + { + "epoch": 0.37068162239139224, + "grad_norm": 0.3477768301963806, + "learning_rate": 0.00013956012413864495, + "loss": 0.0728, + "step": 15813 + }, + { + "epoch": 0.37070506396619723, + "grad_norm": 0.7404635548591614, + "learning_rate": 0.00013955335878715794, + "loss": 0.1058, + "step": 15814 + }, + { + "epoch": 0.37072850554100223, + "grad_norm": 0.5672017931938171, + "learning_rate": 0.00013954659322105333, + "loss": 0.5726, + "step": 15815 + }, + { + "epoch": 0.3707519471158072, + "grad_norm": 0.5141189098358154, + "learning_rate": 0.00013953982744036776, + "loss": 0.1092, + "step": 15816 + }, + { + "epoch": 0.3707753886906122, + "grad_norm": 0.4358260929584503, + "learning_rate": 0.000139533061445138, + "loss": 0.0745, + "step": 15817 + }, + { + "epoch": 0.3707988302654172, + "grad_norm": 0.5823654532432556, + "learning_rate": 0.0001395262952354007, + "loss": 0.123, + "step": 15818 + }, + { + "epoch": 0.3708222718402222, + "grad_norm": 0.4808945655822754, + "learning_rate": 0.00013951952881119262, + "loss": 0.0867, + "step": 15819 + }, + { + "epoch": 0.3708457134150272, + "grad_norm": 0.37890222668647766, + "learning_rate": 0.00013951276217255047, + "loss": 0.1188, + "step": 15820 + }, + { + "epoch": 0.3708691549898322, + "grad_norm": 0.2504867911338806, + "learning_rate": 0.00013950599531951097, + "loss": 0.0632, + "step": 15821 + }, + { + "epoch": 0.3708925965646372, + "grad_norm": 0.608142614364624, + "learning_rate": 0.00013949922825211083, + "loss": 0.0666, + "step": 15822 + }, + { + "epoch": 0.3709160381394422, + "grad_norm": 0.37863409519195557, + "learning_rate": 0.00013949246097038677, + "loss": 0.0784, + "step": 15823 + }, + { + "epoch": 0.3709394797142472, + "grad_norm": 0.7369007468223572, + "learning_rate": 0.00013948569347437554, + "loss": 0.1483, + "step": 15824 + }, + { + "epoch": 0.37096292128905217, + "grad_norm": 0.7980160713195801, + "learning_rate": 0.00013947892576411373, + "loss": 0.1503, + "step": 15825 + }, + { + "epoch": 0.37098636286385717, + "grad_norm": 0.45920121669769287, + "learning_rate": 0.00013947215783963822, + "loss": 0.099, + "step": 15826 + }, + { + "epoch": 0.3710098044386622, + "grad_norm": 0.7063412666320801, + "learning_rate": 0.0001394653897009857, + "loss": 0.1642, + "step": 15827 + }, + { + "epoch": 0.3710332460134672, + "grad_norm": 0.13320721685886383, + "learning_rate": 0.0001394586213481928, + "loss": 0.033, + "step": 15828 + }, + { + "epoch": 0.3710566875882722, + "grad_norm": 0.3989570438861847, + "learning_rate": 0.00013945185278129636, + "loss": 0.0827, + "step": 15829 + }, + { + "epoch": 0.3710801291630772, + "grad_norm": 0.20642413198947906, + "learning_rate": 0.00013944508400033306, + "loss": 0.0306, + "step": 15830 + }, + { + "epoch": 0.3711035707378822, + "grad_norm": 0.25377121567726135, + "learning_rate": 0.0001394383150053396, + "loss": 0.0571, + "step": 15831 + }, + { + "epoch": 0.3711270123126872, + "grad_norm": 0.5692471265792847, + "learning_rate": 0.00013943154579635272, + "loss": 0.1657, + "step": 15832 + }, + { + "epoch": 0.3711504538874922, + "grad_norm": 0.8655399084091187, + "learning_rate": 0.0001394247763734092, + "loss": 0.1538, + "step": 15833 + }, + { + "epoch": 0.3711738954622972, + "grad_norm": 0.46033626794815063, + "learning_rate": 0.00013941800673654571, + "loss": 0.0568, + "step": 15834 + }, + { + "epoch": 0.37119733703710217, + "grad_norm": 0.432630330324173, + "learning_rate": 0.000139411236885799, + "loss": 0.0962, + "step": 15835 + }, + { + "epoch": 0.37122077861190717, + "grad_norm": 0.667506992816925, + "learning_rate": 0.00013940446682120586, + "loss": 0.1176, + "step": 15836 + }, + { + "epoch": 0.37124422018671216, + "grad_norm": 0.7121483683586121, + "learning_rate": 0.0001393976965428029, + "loss": 0.1339, + "step": 15837 + }, + { + "epoch": 0.37126766176151715, + "grad_norm": 0.41680002212524414, + "learning_rate": 0.00013939092605062697, + "loss": 0.3748, + "step": 15838 + }, + { + "epoch": 0.37129110333632215, + "grad_norm": 0.6372932195663452, + "learning_rate": 0.00013938415534471479, + "loss": 0.1106, + "step": 15839 + }, + { + "epoch": 0.37131454491112714, + "grad_norm": 0.21694731712341309, + "learning_rate": 0.00013937738442510307, + "loss": 0.0543, + "step": 15840 + }, + { + "epoch": 0.37133798648593214, + "grad_norm": 0.37222763895988464, + "learning_rate": 0.00013937061329182852, + "loss": 0.109, + "step": 15841 + }, + { + "epoch": 0.37136142806073713, + "grad_norm": 0.19532670080661774, + "learning_rate": 0.00013936384194492796, + "loss": 0.0429, + "step": 15842 + }, + { + "epoch": 0.3713848696355421, + "grad_norm": 0.5586705207824707, + "learning_rate": 0.00013935707038443802, + "loss": 0.1657, + "step": 15843 + }, + { + "epoch": 0.3714083112103471, + "grad_norm": 0.33759114146232605, + "learning_rate": 0.00013935029861039558, + "loss": 0.025, + "step": 15844 + }, + { + "epoch": 0.3714317527851521, + "grad_norm": 0.667149007320404, + "learning_rate": 0.00013934352662283727, + "loss": 0.1276, + "step": 15845 + }, + { + "epoch": 0.3714551943599571, + "grad_norm": 0.8305057883262634, + "learning_rate": 0.00013933675442179993, + "loss": 0.2871, + "step": 15846 + }, + { + "epoch": 0.3714786359347621, + "grad_norm": 0.40814027190208435, + "learning_rate": 0.00013932998200732018, + "loss": 0.1018, + "step": 15847 + }, + { + "epoch": 0.3715020775095671, + "grad_norm": 0.48442384600639343, + "learning_rate": 0.0001393232093794349, + "loss": 0.1365, + "step": 15848 + }, + { + "epoch": 0.3715255190843721, + "grad_norm": 1.215712547302246, + "learning_rate": 0.00013931643653818075, + "loss": 0.0996, + "step": 15849 + }, + { + "epoch": 0.3715489606591771, + "grad_norm": 0.4121159017086029, + "learning_rate": 0.0001393096634835945, + "loss": 0.0955, + "step": 15850 + }, + { + "epoch": 0.3715724022339821, + "grad_norm": 0.5827898979187012, + "learning_rate": 0.00013930289021571294, + "loss": 0.1377, + "step": 15851 + }, + { + "epoch": 0.3715958438087871, + "grad_norm": 0.17210052907466888, + "learning_rate": 0.0001392961167345728, + "loss": 0.0405, + "step": 15852 + }, + { + "epoch": 0.37161928538359207, + "grad_norm": 0.4097435176372528, + "learning_rate": 0.0001392893430402108, + "loss": 0.0995, + "step": 15853 + }, + { + "epoch": 0.37164272695839706, + "grad_norm": 0.5811240077018738, + "learning_rate": 0.00013928256913266373, + "loss": 0.1223, + "step": 15854 + }, + { + "epoch": 0.37166616853320206, + "grad_norm": 0.7198411822319031, + "learning_rate": 0.00013927579501196832, + "loss": 0.1313, + "step": 15855 + }, + { + "epoch": 0.37168961010800705, + "grad_norm": 1.0160585641860962, + "learning_rate": 0.00013926902067816136, + "loss": 0.1326, + "step": 15856 + }, + { + "epoch": 0.37171305168281205, + "grad_norm": 0.4453596770763397, + "learning_rate": 0.00013926224613127958, + "loss": 0.0459, + "step": 15857 + }, + { + "epoch": 0.37173649325761704, + "grad_norm": 0.34094759821891785, + "learning_rate": 0.00013925547137135977, + "loss": 0.113, + "step": 15858 + }, + { + "epoch": 0.37175993483242203, + "grad_norm": 0.38238492608070374, + "learning_rate": 0.00013924869639843866, + "loss": 0.0734, + "step": 15859 + }, + { + "epoch": 0.37178337640722703, + "grad_norm": 0.7460346817970276, + "learning_rate": 0.000139241921212553, + "loss": 0.1604, + "step": 15860 + }, + { + "epoch": 0.371806817982032, + "grad_norm": 0.11029336601495743, + "learning_rate": 0.00013923514581373959, + "loss": 0.0144, + "step": 15861 + }, + { + "epoch": 0.371830259556837, + "grad_norm": 3.9506075382232666, + "learning_rate": 0.00013922837020203515, + "loss": 0.1001, + "step": 15862 + }, + { + "epoch": 0.371853701131642, + "grad_norm": 0.15498857200145721, + "learning_rate": 0.00013922159437747652, + "loss": 0.0423, + "step": 15863 + }, + { + "epoch": 0.371877142706447, + "grad_norm": 0.5065082311630249, + "learning_rate": 0.00013921481834010038, + "loss": 0.1116, + "step": 15864 + }, + { + "epoch": 0.371900584281252, + "grad_norm": 0.4000837802886963, + "learning_rate": 0.00013920804208994352, + "loss": 0.0801, + "step": 15865 + }, + { + "epoch": 0.371924025856057, + "grad_norm": 0.8430947065353394, + "learning_rate": 0.00013920126562704275, + "loss": 0.0901, + "step": 15866 + }, + { + "epoch": 0.371947467430862, + "grad_norm": 0.6523247361183167, + "learning_rate": 0.0001391944889514348, + "loss": 0.8522, + "step": 15867 + }, + { + "epoch": 0.371970909005667, + "grad_norm": 0.22340910136699677, + "learning_rate": 0.00013918771206315647, + "loss": 0.0405, + "step": 15868 + }, + { + "epoch": 0.371994350580472, + "grad_norm": 0.5835574269294739, + "learning_rate": 0.00013918093496224449, + "loss": 0.0796, + "step": 15869 + }, + { + "epoch": 0.37201779215527697, + "grad_norm": 0.219388946890831, + "learning_rate": 0.00013917415764873564, + "loss": 0.0426, + "step": 15870 + }, + { + "epoch": 0.37204123373008197, + "grad_norm": 0.6852166056632996, + "learning_rate": 0.00013916738012266676, + "loss": 0.1564, + "step": 15871 + }, + { + "epoch": 0.37206467530488696, + "grad_norm": 0.2327336072921753, + "learning_rate": 0.00013916060238407452, + "loss": 0.0291, + "step": 15872 + }, + { + "epoch": 0.37208811687969195, + "grad_norm": 0.5896862149238586, + "learning_rate": 0.00013915382443299576, + "loss": 0.1111, + "step": 15873 + }, + { + "epoch": 0.37211155845449695, + "grad_norm": 0.47808000445365906, + "learning_rate": 0.0001391470462694673, + "loss": 0.0857, + "step": 15874 + }, + { + "epoch": 0.37213500002930194, + "grad_norm": 0.5563243627548218, + "learning_rate": 0.0001391402678935258, + "loss": 0.1398, + "step": 15875 + }, + { + "epoch": 0.37215844160410694, + "grad_norm": 0.7407429814338684, + "learning_rate": 0.00013913348930520814, + "loss": 0.1411, + "step": 15876 + }, + { + "epoch": 0.37218188317891193, + "grad_norm": 0.5179618000984192, + "learning_rate": 0.00013912671050455105, + "loss": 0.1178, + "step": 15877 + }, + { + "epoch": 0.372205324753717, + "grad_norm": 0.6211474537849426, + "learning_rate": 0.00013911993149159137, + "loss": 0.0631, + "step": 15878 + }, + { + "epoch": 0.372228766328522, + "grad_norm": 0.35351845622062683, + "learning_rate": 0.00013911315226636578, + "loss": 0.2203, + "step": 15879 + }, + { + "epoch": 0.37225220790332697, + "grad_norm": 0.4130392074584961, + "learning_rate": 0.00013910637282891117, + "loss": 0.0611, + "step": 15880 + }, + { + "epoch": 0.37227564947813196, + "grad_norm": 0.22816374897956848, + "learning_rate": 0.00013909959317926426, + "loss": 0.0171, + "step": 15881 + }, + { + "epoch": 0.37229909105293696, + "grad_norm": 0.5494468808174133, + "learning_rate": 0.00013909281331746186, + "loss": 0.9384, + "step": 15882 + }, + { + "epoch": 0.37232253262774195, + "grad_norm": 0.2970712184906006, + "learning_rate": 0.00013908603324354078, + "loss": 0.0661, + "step": 15883 + }, + { + "epoch": 0.37234597420254695, + "grad_norm": 0.12679775059223175, + "learning_rate": 0.00013907925295753777, + "loss": 0.0317, + "step": 15884 + }, + { + "epoch": 0.37236941577735194, + "grad_norm": 0.26874154806137085, + "learning_rate": 0.00013907247245948962, + "loss": 0.0492, + "step": 15885 + }, + { + "epoch": 0.37239285735215694, + "grad_norm": 0.35653603076934814, + "learning_rate": 0.00013906569174943314, + "loss": 0.0746, + "step": 15886 + }, + { + "epoch": 0.37241629892696193, + "grad_norm": 0.5561107993125916, + "learning_rate": 0.00013905891082740515, + "loss": 0.1327, + "step": 15887 + }, + { + "epoch": 0.3724397405017669, + "grad_norm": 0.11827604472637177, + "learning_rate": 0.0001390521296934424, + "loss": 0.0252, + "step": 15888 + }, + { + "epoch": 0.3724631820765719, + "grad_norm": 0.3079414367675781, + "learning_rate": 0.00013904534834758165, + "loss": 0.0852, + "step": 15889 + }, + { + "epoch": 0.3724866236513769, + "grad_norm": 0.3865712285041809, + "learning_rate": 0.0001390385667898598, + "loss": 0.0306, + "step": 15890 + }, + { + "epoch": 0.3725100652261819, + "grad_norm": 0.2732374370098114, + "learning_rate": 0.00013903178502031356, + "loss": 0.036, + "step": 15891 + }, + { + "epoch": 0.3725335068009869, + "grad_norm": 0.6154868006706238, + "learning_rate": 0.00013902500303897974, + "loss": 0.1423, + "step": 15892 + }, + { + "epoch": 0.3725569483757919, + "grad_norm": 0.08863206952810287, + "learning_rate": 0.00013901822084589518, + "loss": 0.0262, + "step": 15893 + }, + { + "epoch": 0.3725803899505969, + "grad_norm": 0.20032449066638947, + "learning_rate": 0.00013901143844109666, + "loss": 0.0373, + "step": 15894 + }, + { + "epoch": 0.3726038315254019, + "grad_norm": 0.4730222821235657, + "learning_rate": 0.00013900465582462095, + "loss": 0.0692, + "step": 15895 + }, + { + "epoch": 0.3726272731002069, + "grad_norm": 0.8058201670646667, + "learning_rate": 0.00013899787299650492, + "loss": 0.2187, + "step": 15896 + }, + { + "epoch": 0.3726507146750119, + "grad_norm": 0.13401365280151367, + "learning_rate": 0.0001389910899567853, + "loss": 0.0266, + "step": 15897 + }, + { + "epoch": 0.37267415624981687, + "grad_norm": 0.32819265127182007, + "learning_rate": 0.00013898430670549895, + "loss": 0.0514, + "step": 15898 + }, + { + "epoch": 0.37269759782462186, + "grad_norm": 0.22562140226364136, + "learning_rate": 0.00013897752324268264, + "loss": 0.0568, + "step": 15899 + }, + { + "epoch": 0.37272103939942686, + "grad_norm": 0.3798229992389679, + "learning_rate": 0.00013897073956837324, + "loss": 0.0273, + "step": 15900 + }, + { + "epoch": 0.37274448097423185, + "grad_norm": 0.48552459478378296, + "learning_rate": 0.00013896395568260746, + "loss": 0.0997, + "step": 15901 + }, + { + "epoch": 0.37276792254903685, + "grad_norm": 0.49889928102493286, + "learning_rate": 0.00013895717158542219, + "loss": 0.1028, + "step": 15902 + }, + { + "epoch": 0.37279136412384184, + "grad_norm": 0.30575263500213623, + "learning_rate": 0.0001389503872768542, + "loss": 0.0577, + "step": 15903 + }, + { + "epoch": 0.37281480569864683, + "grad_norm": 0.6125985980033875, + "learning_rate": 0.00013894360275694032, + "loss": 0.4824, + "step": 15904 + }, + { + "epoch": 0.37283824727345183, + "grad_norm": 0.9565631747245789, + "learning_rate": 0.00013893681802571735, + "loss": 0.1154, + "step": 15905 + }, + { + "epoch": 0.3728616888482568, + "grad_norm": 0.5546180009841919, + "learning_rate": 0.0001389300330832221, + "loss": 0.0826, + "step": 15906 + }, + { + "epoch": 0.3728851304230618, + "grad_norm": 1.437711238861084, + "learning_rate": 0.00013892324792949144, + "loss": 0.2332, + "step": 15907 + }, + { + "epoch": 0.3729085719978668, + "grad_norm": 0.5224947929382324, + "learning_rate": 0.0001389164625645621, + "loss": 0.0969, + "step": 15908 + }, + { + "epoch": 0.3729320135726718, + "grad_norm": 0.7158041000366211, + "learning_rate": 0.00013890967698847092, + "loss": 0.1472, + "step": 15909 + }, + { + "epoch": 0.3729554551474768, + "grad_norm": 0.6195759773254395, + "learning_rate": 0.00013890289120125477, + "loss": 0.1415, + "step": 15910 + }, + { + "epoch": 0.3729788967222818, + "grad_norm": 0.6435240507125854, + "learning_rate": 0.00013889610520295042, + "loss": 0.1158, + "step": 15911 + }, + { + "epoch": 0.3730023382970868, + "grad_norm": 0.2895351052284241, + "learning_rate": 0.00013888931899359475, + "loss": 0.0718, + "step": 15912 + }, + { + "epoch": 0.3730257798718918, + "grad_norm": 0.8724422454833984, + "learning_rate": 0.0001388825325732245, + "loss": 0.7782, + "step": 15913 + }, + { + "epoch": 0.3730492214466968, + "grad_norm": 0.4915631413459778, + "learning_rate": 0.00013887574594187655, + "loss": 0.0941, + "step": 15914 + }, + { + "epoch": 0.37307266302150177, + "grad_norm": 0.379751980304718, + "learning_rate": 0.00013886895909958773, + "loss": 0.0942, + "step": 15915 + }, + { + "epoch": 0.37309610459630677, + "grad_norm": 0.7039495706558228, + "learning_rate": 0.0001388621720463948, + "loss": 0.1564, + "step": 15916 + }, + { + "epoch": 0.37311954617111176, + "grad_norm": 0.44738221168518066, + "learning_rate": 0.00013885538478233467, + "loss": 0.0798, + "step": 15917 + }, + { + "epoch": 0.37314298774591675, + "grad_norm": 0.5198834538459778, + "learning_rate": 0.0001388485973074441, + "loss": 0.2028, + "step": 15918 + }, + { + "epoch": 0.37316642932072175, + "grad_norm": 0.5363789796829224, + "learning_rate": 0.00013884180962176, + "loss": 0.7087, + "step": 15919 + }, + { + "epoch": 0.37318987089552674, + "grad_norm": 0.15077020227909088, + "learning_rate": 0.0001388350217253191, + "loss": 0.0296, + "step": 15920 + }, + { + "epoch": 0.37321331247033174, + "grad_norm": 0.47837817668914795, + "learning_rate": 0.00013882823361815829, + "loss": 0.0772, + "step": 15921 + }, + { + "epoch": 0.37323675404513673, + "grad_norm": 0.20275089144706726, + "learning_rate": 0.0001388214453003144, + "loss": 0.0252, + "step": 15922 + }, + { + "epoch": 0.3732601956199417, + "grad_norm": 0.37360066175460815, + "learning_rate": 0.00013881465677182425, + "loss": 0.1081, + "step": 15923 + }, + { + "epoch": 0.3732836371947467, + "grad_norm": 0.09009994566440582, + "learning_rate": 0.00013880786803272468, + "loss": 0.0168, + "step": 15924 + }, + { + "epoch": 0.3733070787695517, + "grad_norm": 0.4631577134132385, + "learning_rate": 0.00013880107908305253, + "loss": 0.0486, + "step": 15925 + }, + { + "epoch": 0.3733305203443567, + "grad_norm": 0.3073942959308624, + "learning_rate": 0.00013879428992284466, + "loss": 0.1049, + "step": 15926 + }, + { + "epoch": 0.3733539619191617, + "grad_norm": 0.4085155129432678, + "learning_rate": 0.00013878750055213782, + "loss": 0.1156, + "step": 15927 + }, + { + "epoch": 0.3733774034939667, + "grad_norm": 0.44668644666671753, + "learning_rate": 0.00013878071097096897, + "loss": 0.1326, + "step": 15928 + }, + { + "epoch": 0.37340084506877175, + "grad_norm": 0.22628505527973175, + "learning_rate": 0.00013877392117937485, + "loss": 0.0737, + "step": 15929 + }, + { + "epoch": 0.37342428664357674, + "grad_norm": 0.4085805416107178, + "learning_rate": 0.00013876713117739236, + "loss": 0.0797, + "step": 15930 + }, + { + "epoch": 0.37344772821838174, + "grad_norm": 0.40621501207351685, + "learning_rate": 0.00013876034096505833, + "loss": 0.0996, + "step": 15931 + }, + { + "epoch": 0.37347116979318673, + "grad_norm": 0.4844948649406433, + "learning_rate": 0.00013875355054240963, + "loss": 0.1675, + "step": 15932 + }, + { + "epoch": 0.3734946113679917, + "grad_norm": 0.39149394631385803, + "learning_rate": 0.00013874675990948303, + "loss": 0.1304, + "step": 15933 + }, + { + "epoch": 0.3735180529427967, + "grad_norm": 0.2975917458534241, + "learning_rate": 0.00013873996906631542, + "loss": 0.1067, + "step": 15934 + }, + { + "epoch": 0.3735414945176017, + "grad_norm": 0.13682197034358978, + "learning_rate": 0.00013873317801294367, + "loss": 0.0161, + "step": 15935 + }, + { + "epoch": 0.3735649360924067, + "grad_norm": 0.4138859808444977, + "learning_rate": 0.00013872638674940462, + "loss": 0.0354, + "step": 15936 + }, + { + "epoch": 0.3735883776672117, + "grad_norm": 0.5016474723815918, + "learning_rate": 0.00013871959527573507, + "loss": 0.1091, + "step": 15937 + }, + { + "epoch": 0.3736118192420167, + "grad_norm": 0.14509697258472443, + "learning_rate": 0.00013871280359197197, + "loss": 0.022, + "step": 15938 + }, + { + "epoch": 0.3736352608168217, + "grad_norm": 0.5437679290771484, + "learning_rate": 0.00013870601169815205, + "loss": 0.1754, + "step": 15939 + }, + { + "epoch": 0.3736587023916267, + "grad_norm": 0.48410892486572266, + "learning_rate": 0.0001386992195943122, + "loss": 0.1372, + "step": 15940 + }, + { + "epoch": 0.3736821439664317, + "grad_norm": 0.35171034932136536, + "learning_rate": 0.00013869242728048936, + "loss": 0.0581, + "step": 15941 + }, + { + "epoch": 0.3737055855412367, + "grad_norm": 0.2369382679462433, + "learning_rate": 0.0001386856347567203, + "loss": 0.0386, + "step": 15942 + }, + { + "epoch": 0.37372902711604167, + "grad_norm": 0.31368011236190796, + "learning_rate": 0.00013867884202304188, + "loss": 0.047, + "step": 15943 + }, + { + "epoch": 0.37375246869084666, + "grad_norm": 0.5421419739723206, + "learning_rate": 0.00013867204907949102, + "loss": 0.0641, + "step": 15944 + }, + { + "epoch": 0.37377591026565166, + "grad_norm": 0.6441825032234192, + "learning_rate": 0.0001386652559261045, + "loss": 0.1313, + "step": 15945 + }, + { + "epoch": 0.37379935184045665, + "grad_norm": 0.3008389174938202, + "learning_rate": 0.0001386584625629192, + "loss": 0.044, + "step": 15946 + }, + { + "epoch": 0.37382279341526164, + "grad_norm": 0.2976015508174896, + "learning_rate": 0.000138651668989972, + "loss": 0.0322, + "step": 15947 + }, + { + "epoch": 0.37384623499006664, + "grad_norm": 0.7308828830718994, + "learning_rate": 0.00013864487520729978, + "loss": 0.212, + "step": 15948 + }, + { + "epoch": 0.37386967656487163, + "grad_norm": 0.5505029559135437, + "learning_rate": 0.00013863808121493938, + "loss": 0.126, + "step": 15949 + }, + { + "epoch": 0.37389311813967663, + "grad_norm": 0.4837331771850586, + "learning_rate": 0.00013863128701292765, + "loss": 0.1289, + "step": 15950 + }, + { + "epoch": 0.3739165597144816, + "grad_norm": 0.6548144817352295, + "learning_rate": 0.00013862449260130147, + "loss": 0.1526, + "step": 15951 + }, + { + "epoch": 0.3739400012892866, + "grad_norm": 0.28156983852386475, + "learning_rate": 0.00013861769798009771, + "loss": 0.034, + "step": 15952 + }, + { + "epoch": 0.3739634428640916, + "grad_norm": 0.4262019395828247, + "learning_rate": 0.00013861090314935327, + "loss": 0.0927, + "step": 15953 + }, + { + "epoch": 0.3739868844388966, + "grad_norm": 0.8899794816970825, + "learning_rate": 0.00013860410810910494, + "loss": 0.1459, + "step": 15954 + }, + { + "epoch": 0.3740103260137016, + "grad_norm": 0.286601722240448, + "learning_rate": 0.00013859731285938964, + "loss": 0.0688, + "step": 15955 + }, + { + "epoch": 0.3740337675885066, + "grad_norm": 0.4919109642505646, + "learning_rate": 0.00013859051740024424, + "loss": 0.1116, + "step": 15956 + }, + { + "epoch": 0.3740572091633116, + "grad_norm": 0.2669203579425812, + "learning_rate": 0.0001385837217317056, + "loss": 0.0499, + "step": 15957 + }, + { + "epoch": 0.3740806507381166, + "grad_norm": 0.5643711090087891, + "learning_rate": 0.00013857692585381062, + "loss": 0.1055, + "step": 15958 + }, + { + "epoch": 0.3741040923129216, + "grad_norm": 0.3206412196159363, + "learning_rate": 0.00013857012976659615, + "loss": 0.0772, + "step": 15959 + }, + { + "epoch": 0.37412753388772657, + "grad_norm": 0.39458510279655457, + "learning_rate": 0.00013856333347009908, + "loss": 0.084, + "step": 15960 + }, + { + "epoch": 0.37415097546253157, + "grad_norm": 0.26339101791381836, + "learning_rate": 0.00013855653696435627, + "loss": 0.0349, + "step": 15961 + }, + { + "epoch": 0.37417441703733656, + "grad_norm": 0.4319285750389099, + "learning_rate": 0.00013854974024940457, + "loss": 0.1262, + "step": 15962 + }, + { + "epoch": 0.37419785861214155, + "grad_norm": 0.263409823179245, + "learning_rate": 0.00013854294332528094, + "loss": 0.0626, + "step": 15963 + }, + { + "epoch": 0.37422130018694655, + "grad_norm": 0.7729456424713135, + "learning_rate": 0.0001385361461920222, + "loss": 0.1965, + "step": 15964 + }, + { + "epoch": 0.37424474176175154, + "grad_norm": 0.6120654344558716, + "learning_rate": 0.00013852934884966525, + "loss": 0.1082, + "step": 15965 + }, + { + "epoch": 0.37426818333655654, + "grad_norm": 0.37070533633232117, + "learning_rate": 0.00013852255129824696, + "loss": 0.0439, + "step": 15966 + }, + { + "epoch": 0.37429162491136153, + "grad_norm": 0.06673254817724228, + "learning_rate": 0.00013851575353780423, + "loss": 0.0104, + "step": 15967 + }, + { + "epoch": 0.3743150664861665, + "grad_norm": 0.5020653009414673, + "learning_rate": 0.00013850895556837394, + "loss": 0.1098, + "step": 15968 + }, + { + "epoch": 0.3743385080609715, + "grad_norm": 0.4315912127494812, + "learning_rate": 0.000138502157389993, + "loss": 0.1047, + "step": 15969 + }, + { + "epoch": 0.3743619496357765, + "grad_norm": 0.8873246312141418, + "learning_rate": 0.00013849535900269824, + "loss": 0.2091, + "step": 15970 + }, + { + "epoch": 0.3743853912105815, + "grad_norm": 0.15748928487300873, + "learning_rate": 0.00013848856040652658, + "loss": 0.0195, + "step": 15971 + }, + { + "epoch": 0.3744088327853865, + "grad_norm": 0.5638012886047363, + "learning_rate": 0.0001384817616015149, + "loss": 0.7506, + "step": 15972 + }, + { + "epoch": 0.3744322743601915, + "grad_norm": 0.4390996992588043, + "learning_rate": 0.00013847496258770014, + "loss": 0.1068, + "step": 15973 + }, + { + "epoch": 0.3744557159349965, + "grad_norm": 0.4967864155769348, + "learning_rate": 0.0001384681633651191, + "loss": 0.1444, + "step": 15974 + }, + { + "epoch": 0.3744791575098015, + "grad_norm": 0.1452856808900833, + "learning_rate": 0.00013846136393380878, + "loss": 0.0272, + "step": 15975 + }, + { + "epoch": 0.3745025990846065, + "grad_norm": 0.2369944304227829, + "learning_rate": 0.00013845456429380598, + "loss": 0.0443, + "step": 15976 + }, + { + "epoch": 0.3745260406594115, + "grad_norm": 0.5304374098777771, + "learning_rate": 0.00013844776444514766, + "loss": 0.1315, + "step": 15977 + }, + { + "epoch": 0.37454948223421647, + "grad_norm": 0.46349528431892395, + "learning_rate": 0.00013844096438787065, + "loss": 0.1667, + "step": 15978 + }, + { + "epoch": 0.37457292380902146, + "grad_norm": 0.3009314239025116, + "learning_rate": 0.0001384341641220119, + "loss": 0.0659, + "step": 15979 + }, + { + "epoch": 0.3745963653838265, + "grad_norm": 0.4633219242095947, + "learning_rate": 0.0001384273636476083, + "loss": 0.0877, + "step": 15980 + }, + { + "epoch": 0.3746198069586315, + "grad_norm": 0.13741153478622437, + "learning_rate": 0.00013842056296469677, + "loss": 0.0302, + "step": 15981 + }, + { + "epoch": 0.3746432485334365, + "grad_norm": 0.37889114022254944, + "learning_rate": 0.00013841376207331417, + "loss": 0.0282, + "step": 15982 + }, + { + "epoch": 0.3746666901082415, + "grad_norm": 0.2446453720331192, + "learning_rate": 0.00013840696097349738, + "loss": 0.0451, + "step": 15983 + }, + { + "epoch": 0.3746901316830465, + "grad_norm": 0.10737645626068115, + "learning_rate": 0.00013840015966528336, + "loss": 0.0206, + "step": 15984 + }, + { + "epoch": 0.3747135732578515, + "grad_norm": 0.5746923685073853, + "learning_rate": 0.000138393358148709, + "loss": 0.1021, + "step": 15985 + }, + { + "epoch": 0.3747370148326565, + "grad_norm": 0.8907926082611084, + "learning_rate": 0.00013838655642381121, + "loss": 0.0916, + "step": 15986 + }, + { + "epoch": 0.3747604564074615, + "grad_norm": 0.21047480404376984, + "learning_rate": 0.00013837975449062687, + "loss": 0.0497, + "step": 15987 + }, + { + "epoch": 0.37478389798226647, + "grad_norm": 0.23306891322135925, + "learning_rate": 0.0001383729523491929, + "loss": 0.0329, + "step": 15988 + }, + { + "epoch": 0.37480733955707146, + "grad_norm": 0.14884385466575623, + "learning_rate": 0.00013836614999954624, + "loss": 0.0145, + "step": 15989 + }, + { + "epoch": 0.37483078113187646, + "grad_norm": 0.6767454147338867, + "learning_rate": 0.00013835934744172376, + "loss": 0.1422, + "step": 15990 + }, + { + "epoch": 0.37485422270668145, + "grad_norm": 0.4812409579753876, + "learning_rate": 0.00013835254467576234, + "loss": 0.143, + "step": 15991 + }, + { + "epoch": 0.37487766428148644, + "grad_norm": 0.2451023906469345, + "learning_rate": 0.000138345741701699, + "loss": 0.0501, + "step": 15992 + }, + { + "epoch": 0.37490110585629144, + "grad_norm": 0.4315865635871887, + "learning_rate": 0.00013833893851957055, + "loss": 0.1154, + "step": 15993 + }, + { + "epoch": 0.37492454743109643, + "grad_norm": 0.12226980924606323, + "learning_rate": 0.00013833213512941397, + "loss": 0.0243, + "step": 15994 + }, + { + "epoch": 0.3749479890059014, + "grad_norm": 0.6451293230056763, + "learning_rate": 0.0001383253315312661, + "loss": 0.0485, + "step": 15995 + }, + { + "epoch": 0.3749714305807064, + "grad_norm": 0.5909914970397949, + "learning_rate": 0.00013831852772516393, + "loss": 0.1855, + "step": 15996 + }, + { + "epoch": 0.3749948721555114, + "grad_norm": 0.24529021978378296, + "learning_rate": 0.00013831172371114433, + "loss": 0.0313, + "step": 15997 + }, + { + "epoch": 0.3750183137303164, + "grad_norm": 0.560431718826294, + "learning_rate": 0.0001383049194892443, + "loss": 0.0771, + "step": 15998 + }, + { + "epoch": 0.3750417553051214, + "grad_norm": 0.4784002900123596, + "learning_rate": 0.00013829811505950066, + "loss": 0.1036, + "step": 15999 + }, + { + "epoch": 0.3750651968799264, + "grad_norm": 0.23673208057880402, + "learning_rate": 0.00013829131042195035, + "loss": 0.0336, + "step": 16000 + }, + { + "epoch": 0.3750886384547314, + "grad_norm": 0.8466134667396545, + "learning_rate": 0.00013828450557663033, + "loss": 0.1355, + "step": 16001 + }, + { + "epoch": 0.3751120800295364, + "grad_norm": 0.23497527837753296, + "learning_rate": 0.0001382777005235775, + "loss": 0.0335, + "step": 16002 + }, + { + "epoch": 0.3751355216043414, + "grad_norm": 1.567134976387024, + "learning_rate": 0.00013827089526282878, + "loss": 0.2119, + "step": 16003 + }, + { + "epoch": 0.3751589631791464, + "grad_norm": 0.4207671880722046, + "learning_rate": 0.00013826408979442117, + "loss": 0.0878, + "step": 16004 + }, + { + "epoch": 0.37518240475395137, + "grad_norm": 0.2760700583457947, + "learning_rate": 0.00013825728411839147, + "loss": 0.0709, + "step": 16005 + }, + { + "epoch": 0.37520584632875636, + "grad_norm": 1.1269276142120361, + "learning_rate": 0.0001382504782347767, + "loss": 0.1023, + "step": 16006 + }, + { + "epoch": 0.37522928790356136, + "grad_norm": 0.9151327013969421, + "learning_rate": 0.00013824367214361374, + "loss": 0.5014, + "step": 16007 + }, + { + "epoch": 0.37525272947836635, + "grad_norm": 0.3126753270626068, + "learning_rate": 0.00013823686584493955, + "loss": 0.0543, + "step": 16008 + }, + { + "epoch": 0.37527617105317135, + "grad_norm": 0.4375140368938446, + "learning_rate": 0.00013823005933879102, + "loss": 0.1084, + "step": 16009 + }, + { + "epoch": 0.37529961262797634, + "grad_norm": 0.6605613827705383, + "learning_rate": 0.00013822325262520516, + "loss": 0.0969, + "step": 16010 + }, + { + "epoch": 0.37532305420278134, + "grad_norm": 0.24745023250579834, + "learning_rate": 0.00013821644570421885, + "loss": 0.0416, + "step": 16011 + }, + { + "epoch": 0.37534649577758633, + "grad_norm": 0.47953933477401733, + "learning_rate": 0.00013820963857586903, + "loss": 0.1075, + "step": 16012 + }, + { + "epoch": 0.3753699373523913, + "grad_norm": 0.11355425417423248, + "learning_rate": 0.0001382028312401926, + "loss": 0.0329, + "step": 16013 + }, + { + "epoch": 0.3753933789271963, + "grad_norm": 0.3490316569805145, + "learning_rate": 0.00013819602369722657, + "loss": 0.0488, + "step": 16014 + }, + { + "epoch": 0.3754168205020013, + "grad_norm": 0.300137460231781, + "learning_rate": 0.00013818921594700785, + "loss": 0.0659, + "step": 16015 + }, + { + "epoch": 0.3754402620768063, + "grad_norm": 0.17436571419239044, + "learning_rate": 0.00013818240798957335, + "loss": 0.0283, + "step": 16016 + }, + { + "epoch": 0.3754637036516113, + "grad_norm": 0.4196115732192993, + "learning_rate": 0.00013817559982496006, + "loss": 0.0569, + "step": 16017 + }, + { + "epoch": 0.3754871452264163, + "grad_norm": 0.508630096912384, + "learning_rate": 0.00013816879145320485, + "loss": 0.0806, + "step": 16018 + }, + { + "epoch": 0.3755105868012213, + "grad_norm": 0.22354117035865784, + "learning_rate": 0.0001381619828743447, + "loss": 0.0301, + "step": 16019 + }, + { + "epoch": 0.3755340283760263, + "grad_norm": 0.5976172089576721, + "learning_rate": 0.00013815517408841656, + "loss": 0.0819, + "step": 16020 + }, + { + "epoch": 0.3755574699508313, + "grad_norm": 0.3574950397014618, + "learning_rate": 0.0001381483650954574, + "loss": 0.0636, + "step": 16021 + }, + { + "epoch": 0.3755809115256363, + "grad_norm": 0.4346403479576111, + "learning_rate": 0.00013814155589550413, + "loss": 0.0724, + "step": 16022 + }, + { + "epoch": 0.37560435310044127, + "grad_norm": 0.4398997128009796, + "learning_rate": 0.00013813474648859372, + "loss": 0.0932, + "step": 16023 + }, + { + "epoch": 0.37562779467524626, + "grad_norm": 0.8597443103790283, + "learning_rate": 0.00013812793687476309, + "loss": 0.2178, + "step": 16024 + }, + { + "epoch": 0.37565123625005126, + "grad_norm": 0.5050858855247498, + "learning_rate": 0.0001381211270540492, + "loss": 0.1246, + "step": 16025 + }, + { + "epoch": 0.37567467782485625, + "grad_norm": 0.2484539896249771, + "learning_rate": 0.000138114317026489, + "loss": 0.0417, + "step": 16026 + }, + { + "epoch": 0.37569811939966125, + "grad_norm": 0.46268710494041443, + "learning_rate": 0.00013810750679211944, + "loss": 0.0745, + "step": 16027 + }, + { + "epoch": 0.37572156097446624, + "grad_norm": 0.1671043038368225, + "learning_rate": 0.00013810069635097747, + "loss": 0.0408, + "step": 16028 + }, + { + "epoch": 0.37574500254927123, + "grad_norm": 0.3904089629650116, + "learning_rate": 0.00013809388570310008, + "loss": 0.086, + "step": 16029 + }, + { + "epoch": 0.37576844412407623, + "grad_norm": 0.30005010962486267, + "learning_rate": 0.00013808707484852417, + "loss": 0.0589, + "step": 16030 + }, + { + "epoch": 0.3757918856988813, + "grad_norm": 0.6330426931381226, + "learning_rate": 0.00013808026378728672, + "loss": 0.1425, + "step": 16031 + }, + { + "epoch": 0.3758153272736863, + "grad_norm": 0.2255847007036209, + "learning_rate": 0.0001380734525194247, + "loss": 0.0499, + "step": 16032 + }, + { + "epoch": 0.37583876884849127, + "grad_norm": 0.8329455852508545, + "learning_rate": 0.00013806664104497505, + "loss": 0.0942, + "step": 16033 + }, + { + "epoch": 0.37586221042329626, + "grad_norm": 0.40475359559059143, + "learning_rate": 0.00013805982936397474, + "loss": 0.0756, + "step": 16034 + }, + { + "epoch": 0.37588565199810126, + "grad_norm": 0.846129298210144, + "learning_rate": 0.00013805301747646072, + "loss": 0.1052, + "step": 16035 + }, + { + "epoch": 0.37590909357290625, + "grad_norm": 0.7856776714324951, + "learning_rate": 0.00013804620538246997, + "loss": 0.6822, + "step": 16036 + }, + { + "epoch": 0.37593253514771124, + "grad_norm": 0.4983454644680023, + "learning_rate": 0.00013803939308203945, + "loss": 0.0987, + "step": 16037 + }, + { + "epoch": 0.37595597672251624, + "grad_norm": 0.5987983345985413, + "learning_rate": 0.00013803258057520608, + "loss": 0.0842, + "step": 16038 + }, + { + "epoch": 0.37597941829732123, + "grad_norm": 0.6615681648254395, + "learning_rate": 0.00013802576786200687, + "loss": 0.5043, + "step": 16039 + }, + { + "epoch": 0.3760028598721262, + "grad_norm": 0.36518383026123047, + "learning_rate": 0.00013801895494247878, + "loss": 0.1022, + "step": 16040 + }, + { + "epoch": 0.3760263014469312, + "grad_norm": 0.9051723480224609, + "learning_rate": 0.00013801214181665875, + "loss": 0.6658, + "step": 16041 + }, + { + "epoch": 0.3760497430217362, + "grad_norm": 0.43012964725494385, + "learning_rate": 0.00013800532848458382, + "loss": 0.0807, + "step": 16042 + }, + { + "epoch": 0.3760731845965412, + "grad_norm": 0.31772908568382263, + "learning_rate": 0.00013799851494629087, + "loss": 0.0321, + "step": 16043 + }, + { + "epoch": 0.3760966261713462, + "grad_norm": 0.44403716921806335, + "learning_rate": 0.0001379917012018169, + "loss": 0.0706, + "step": 16044 + }, + { + "epoch": 0.3761200677461512, + "grad_norm": 0.5070023536682129, + "learning_rate": 0.0001379848872511989, + "loss": 0.1079, + "step": 16045 + }, + { + "epoch": 0.3761435093209562, + "grad_norm": 0.6182812452316284, + "learning_rate": 0.00013797807309447386, + "loss": 0.0993, + "step": 16046 + }, + { + "epoch": 0.3761669508957612, + "grad_norm": 0.2251642793416977, + "learning_rate": 0.0001379712587316787, + "loss": 0.0286, + "step": 16047 + }, + { + "epoch": 0.3761903924705662, + "grad_norm": 0.438546359539032, + "learning_rate": 0.00013796444416285043, + "loss": 0.1118, + "step": 16048 + }, + { + "epoch": 0.3762138340453712, + "grad_norm": 0.39570352435112, + "learning_rate": 0.00013795762938802604, + "loss": 0.0353, + "step": 16049 + }, + { + "epoch": 0.37623727562017617, + "grad_norm": 0.13304965198040009, + "learning_rate": 0.00013795081440724246, + "loss": 0.0389, + "step": 16050 + }, + { + "epoch": 0.37626071719498116, + "grad_norm": 0.6600266695022583, + "learning_rate": 0.00013794399922053668, + "loss": 0.1761, + "step": 16051 + }, + { + "epoch": 0.37628415876978616, + "grad_norm": 0.508874773979187, + "learning_rate": 0.00013793718382794574, + "loss": 0.1535, + "step": 16052 + }, + { + "epoch": 0.37630760034459115, + "grad_norm": 0.3003019094467163, + "learning_rate": 0.00013793036822950657, + "loss": 0.0696, + "step": 16053 + }, + { + "epoch": 0.37633104191939615, + "grad_norm": 0.3689631521701813, + "learning_rate": 0.0001379235524252561, + "loss": 0.1074, + "step": 16054 + }, + { + "epoch": 0.37635448349420114, + "grad_norm": 0.24954038858413696, + "learning_rate": 0.00013791673641523142, + "loss": 0.0602, + "step": 16055 + }, + { + "epoch": 0.37637792506900614, + "grad_norm": 0.5252622365951538, + "learning_rate": 0.00013790992019946944, + "loss": 0.1537, + "step": 16056 + }, + { + "epoch": 0.37640136664381113, + "grad_norm": 1.0452210903167725, + "learning_rate": 0.00013790310377800714, + "loss": 0.7166, + "step": 16057 + }, + { + "epoch": 0.3764248082186161, + "grad_norm": 0.5858443379402161, + "learning_rate": 0.0001378962871508816, + "loss": 0.0614, + "step": 16058 + }, + { + "epoch": 0.3764482497934211, + "grad_norm": 0.5921033024787903, + "learning_rate": 0.0001378894703181297, + "loss": 0.1611, + "step": 16059 + }, + { + "epoch": 0.3764716913682261, + "grad_norm": 0.42378470301628113, + "learning_rate": 0.00013788265327978845, + "loss": 0.0927, + "step": 16060 + }, + { + "epoch": 0.3764951329430311, + "grad_norm": 0.4453268051147461, + "learning_rate": 0.0001378758360358949, + "loss": 0.0773, + "step": 16061 + }, + { + "epoch": 0.3765185745178361, + "grad_norm": 0.4226485788822174, + "learning_rate": 0.00013786901858648597, + "loss": 0.1048, + "step": 16062 + }, + { + "epoch": 0.3765420160926411, + "grad_norm": 0.48031750321388245, + "learning_rate": 0.0001378622009315987, + "loss": 0.111, + "step": 16063 + }, + { + "epoch": 0.3765654576674461, + "grad_norm": 0.5241885781288147, + "learning_rate": 0.00013785538307127003, + "loss": 0.0959, + "step": 16064 + }, + { + "epoch": 0.3765888992422511, + "grad_norm": 0.26689010858535767, + "learning_rate": 0.00013784856500553702, + "loss": 0.0561, + "step": 16065 + }, + { + "epoch": 0.3766123408170561, + "grad_norm": 0.09131436049938202, + "learning_rate": 0.0001378417467344366, + "loss": 0.0101, + "step": 16066 + }, + { + "epoch": 0.3766357823918611, + "grad_norm": 0.5530514717102051, + "learning_rate": 0.00013783492825800587, + "loss": 0.1248, + "step": 16067 + }, + { + "epoch": 0.37665922396666607, + "grad_norm": 0.0898144319653511, + "learning_rate": 0.00013782810957628166, + "loss": 0.015, + "step": 16068 + }, + { + "epoch": 0.37668266554147106, + "grad_norm": 0.4983738362789154, + "learning_rate": 0.00013782129068930113, + "loss": 0.7347, + "step": 16069 + }, + { + "epoch": 0.37670610711627606, + "grad_norm": 0.2636258602142334, + "learning_rate": 0.00013781447159710118, + "loss": 0.0853, + "step": 16070 + }, + { + "epoch": 0.37672954869108105, + "grad_norm": 0.6304008364677429, + "learning_rate": 0.00013780765229971887, + "loss": 0.1188, + "step": 16071 + }, + { + "epoch": 0.37675299026588605, + "grad_norm": 0.39351651072502136, + "learning_rate": 0.00013780083279719117, + "loss": 0.085, + "step": 16072 + }, + { + "epoch": 0.37677643184069104, + "grad_norm": 0.8550036549568176, + "learning_rate": 0.00013779401308955505, + "loss": 0.1449, + "step": 16073 + }, + { + "epoch": 0.37679987341549603, + "grad_norm": 0.14138583838939667, + "learning_rate": 0.0001377871931768476, + "loss": 0.0115, + "step": 16074 + }, + { + "epoch": 0.37682331499030103, + "grad_norm": 0.1629418134689331, + "learning_rate": 0.00013778037305910577, + "loss": 0.0131, + "step": 16075 + }, + { + "epoch": 0.376846756565106, + "grad_norm": 0.5999675393104553, + "learning_rate": 0.00013777355273636655, + "loss": 0.6761, + "step": 16076 + }, + { + "epoch": 0.376870198139911, + "grad_norm": 0.39500394463539124, + "learning_rate": 0.00013776673220866698, + "loss": 0.1084, + "step": 16077 + }, + { + "epoch": 0.376893639714716, + "grad_norm": 0.3251785635948181, + "learning_rate": 0.00013775991147604408, + "loss": 0.1266, + "step": 16078 + }, + { + "epoch": 0.376917081289521, + "grad_norm": 0.6888794898986816, + "learning_rate": 0.00013775309053853482, + "loss": 0.1274, + "step": 16079 + }, + { + "epoch": 0.376940522864326, + "grad_norm": 0.4354352056980133, + "learning_rate": 0.00013774626939617622, + "loss": 0.0849, + "step": 16080 + }, + { + "epoch": 0.376963964439131, + "grad_norm": 0.6100835204124451, + "learning_rate": 0.00013773944804900532, + "loss": 0.5742, + "step": 16081 + }, + { + "epoch": 0.37698740601393604, + "grad_norm": 0.5091122388839722, + "learning_rate": 0.00013773262649705912, + "loss": 0.1349, + "step": 16082 + }, + { + "epoch": 0.37701084758874104, + "grad_norm": 0.5444717407226562, + "learning_rate": 0.0001377258047403746, + "loss": 0.6965, + "step": 16083 + }, + { + "epoch": 0.37703428916354603, + "grad_norm": 0.47023171186447144, + "learning_rate": 0.00013771898277898883, + "loss": 0.0477, + "step": 16084 + }, + { + "epoch": 0.377057730738351, + "grad_norm": 1.0520447492599487, + "learning_rate": 0.00013771216061293876, + "loss": 0.1675, + "step": 16085 + }, + { + "epoch": 0.377081172313156, + "grad_norm": 0.2058144062757492, + "learning_rate": 0.00013770533824226147, + "loss": 0.0195, + "step": 16086 + }, + { + "epoch": 0.377104613887961, + "grad_norm": 0.6884499788284302, + "learning_rate": 0.00013769851566699398, + "loss": 0.1014, + "step": 16087 + }, + { + "epoch": 0.377128055462766, + "grad_norm": 0.6005582809448242, + "learning_rate": 0.00013769169288717326, + "loss": 0.0965, + "step": 16088 + }, + { + "epoch": 0.377151497037571, + "grad_norm": 0.5404656529426575, + "learning_rate": 0.00013768486990283632, + "loss": 0.1333, + "step": 16089 + }, + { + "epoch": 0.377174938612376, + "grad_norm": 0.4314992427825928, + "learning_rate": 0.00013767804671402026, + "loss": 0.1532, + "step": 16090 + }, + { + "epoch": 0.377198380187181, + "grad_norm": 0.5187815427780151, + "learning_rate": 0.00013767122332076207, + "loss": 0.1083, + "step": 16091 + }, + { + "epoch": 0.377221821761986, + "grad_norm": 0.5427232980728149, + "learning_rate": 0.0001376643997230987, + "loss": 0.117, + "step": 16092 + }, + { + "epoch": 0.377245263336791, + "grad_norm": 0.14854696393013, + "learning_rate": 0.0001376575759210673, + "loss": 0.0296, + "step": 16093 + }, + { + "epoch": 0.377268704911596, + "grad_norm": 0.6054933071136475, + "learning_rate": 0.00013765075191470483, + "loss": 0.1475, + "step": 16094 + }, + { + "epoch": 0.37729214648640097, + "grad_norm": 0.41277503967285156, + "learning_rate": 0.0001376439277040483, + "loss": 0.0649, + "step": 16095 + }, + { + "epoch": 0.37731558806120596, + "grad_norm": 0.5768762230873108, + "learning_rate": 0.00013763710328913478, + "loss": 0.0822, + "step": 16096 + }, + { + "epoch": 0.37733902963601096, + "grad_norm": 0.6602848768234253, + "learning_rate": 0.00013763027867000127, + "loss": 0.8103, + "step": 16097 + }, + { + "epoch": 0.37736247121081595, + "grad_norm": 0.7690914869308472, + "learning_rate": 0.0001376234538466848, + "loss": 0.1604, + "step": 16098 + }, + { + "epoch": 0.37738591278562095, + "grad_norm": 0.6464647054672241, + "learning_rate": 0.0001376166288192224, + "loss": 0.8293, + "step": 16099 + }, + { + "epoch": 0.37740935436042594, + "grad_norm": 0.4512357711791992, + "learning_rate": 0.00013760980358765114, + "loss": 0.065, + "step": 16100 + }, + { + "epoch": 0.37743279593523094, + "grad_norm": 0.5131099820137024, + "learning_rate": 0.00013760297815200803, + "loss": 0.082, + "step": 16101 + }, + { + "epoch": 0.37745623751003593, + "grad_norm": 0.6964015364646912, + "learning_rate": 0.00013759615251233008, + "loss": 0.192, + "step": 16102 + }, + { + "epoch": 0.3774796790848409, + "grad_norm": 0.2206878811120987, + "learning_rate": 0.0001375893266686544, + "loss": 0.0373, + "step": 16103 + }, + { + "epoch": 0.3775031206596459, + "grad_norm": 0.2044772058725357, + "learning_rate": 0.0001375825006210179, + "loss": 0.0306, + "step": 16104 + }, + { + "epoch": 0.3775265622344509, + "grad_norm": 0.5640440583229065, + "learning_rate": 0.00013757567436945773, + "loss": 0.051, + "step": 16105 + }, + { + "epoch": 0.3775500038092559, + "grad_norm": 0.1245105043053627, + "learning_rate": 0.0001375688479140109, + "loss": 0.0213, + "step": 16106 + }, + { + "epoch": 0.3775734453840609, + "grad_norm": 1.0061476230621338, + "learning_rate": 0.00013756202125471447, + "loss": 0.3051, + "step": 16107 + }, + { + "epoch": 0.3775968869588659, + "grad_norm": 0.13252118229866028, + "learning_rate": 0.0001375551943916054, + "loss": 0.0296, + "step": 16108 + }, + { + "epoch": 0.3776203285336709, + "grad_norm": 0.39050573110580444, + "learning_rate": 0.00013754836732472083, + "loss": 0.0617, + "step": 16109 + }, + { + "epoch": 0.3776437701084759, + "grad_norm": 0.42767560482025146, + "learning_rate": 0.00013754154005409777, + "loss": 0.0581, + "step": 16110 + }, + { + "epoch": 0.3776672116832809, + "grad_norm": 0.6528472304344177, + "learning_rate": 0.0001375347125797732, + "loss": 0.8102, + "step": 16111 + }, + { + "epoch": 0.3776906532580859, + "grad_norm": 0.3280456066131592, + "learning_rate": 0.00013752788490178428, + "loss": 0.1048, + "step": 16112 + }, + { + "epoch": 0.37771409483289087, + "grad_norm": 0.4509669244289398, + "learning_rate": 0.000137521057020168, + "loss": 0.0953, + "step": 16113 + }, + { + "epoch": 0.37773753640769586, + "grad_norm": 0.2841416597366333, + "learning_rate": 0.00013751422893496139, + "loss": 0.0906, + "step": 16114 + }, + { + "epoch": 0.37776097798250086, + "grad_norm": 0.31344568729400635, + "learning_rate": 0.00013750740064620153, + "loss": 0.077, + "step": 16115 + }, + { + "epoch": 0.37778441955730585, + "grad_norm": 0.1380784511566162, + "learning_rate": 0.00013750057215392546, + "loss": 0.0232, + "step": 16116 + }, + { + "epoch": 0.37780786113211084, + "grad_norm": 0.4046826660633087, + "learning_rate": 0.00013749374345817022, + "loss": 0.0727, + "step": 16117 + }, + { + "epoch": 0.37783130270691584, + "grad_norm": 0.6511013507843018, + "learning_rate": 0.00013748691455897286, + "loss": 0.1235, + "step": 16118 + }, + { + "epoch": 0.37785474428172083, + "grad_norm": 0.3910863697528839, + "learning_rate": 0.0001374800854563705, + "loss": 0.1102, + "step": 16119 + }, + { + "epoch": 0.37787818585652583, + "grad_norm": 0.1949615329504013, + "learning_rate": 0.0001374732561504001, + "loss": 0.0475, + "step": 16120 + }, + { + "epoch": 0.3779016274313308, + "grad_norm": 0.3548511266708374, + "learning_rate": 0.00013746642664109877, + "loss": 0.0753, + "step": 16121 + }, + { + "epoch": 0.3779250690061358, + "grad_norm": 0.5353676676750183, + "learning_rate": 0.00013745959692850352, + "loss": 0.1031, + "step": 16122 + }, + { + "epoch": 0.3779485105809408, + "grad_norm": 0.5242133140563965, + "learning_rate": 0.00013745276701265148, + "loss": 0.1389, + "step": 16123 + }, + { + "epoch": 0.3779719521557458, + "grad_norm": 0.31395652890205383, + "learning_rate": 0.00013744593689357968, + "loss": 0.0353, + "step": 16124 + }, + { + "epoch": 0.3779953937305508, + "grad_norm": 0.4379134178161621, + "learning_rate": 0.00013743910657132517, + "loss": 0.0475, + "step": 16125 + }, + { + "epoch": 0.3780188353053558, + "grad_norm": 0.5120961666107178, + "learning_rate": 0.00013743227604592502, + "loss": 0.0545, + "step": 16126 + }, + { + "epoch": 0.3780422768801608, + "grad_norm": 0.5886784791946411, + "learning_rate": 0.00013742544531741626, + "loss": 0.152, + "step": 16127 + }, + { + "epoch": 0.3780657184549658, + "grad_norm": 0.1274913251399994, + "learning_rate": 0.00013741861438583598, + "loss": 0.0284, + "step": 16128 + }, + { + "epoch": 0.3780891600297708, + "grad_norm": 0.11554847657680511, + "learning_rate": 0.00013741178325122128, + "loss": 0.0149, + "step": 16129 + }, + { + "epoch": 0.37811260160457577, + "grad_norm": 0.4742116630077362, + "learning_rate": 0.00013740495191360915, + "loss": 0.101, + "step": 16130 + }, + { + "epoch": 0.37813604317938077, + "grad_norm": 0.4179658889770508, + "learning_rate": 0.00013739812037303674, + "loss": 0.0957, + "step": 16131 + }, + { + "epoch": 0.37815948475418576, + "grad_norm": 0.5589919090270996, + "learning_rate": 0.00013739128862954106, + "loss": 0.5467, + "step": 16132 + }, + { + "epoch": 0.37818292632899075, + "grad_norm": 0.36429548263549805, + "learning_rate": 0.0001373844566831592, + "loss": 0.0868, + "step": 16133 + }, + { + "epoch": 0.3782063679037958, + "grad_norm": 0.7261270880699158, + "learning_rate": 0.00013737762453392823, + "loss": 0.0855, + "step": 16134 + }, + { + "epoch": 0.3782298094786008, + "grad_norm": 0.15757864713668823, + "learning_rate": 0.00013737079218188522, + "loss": 0.0221, + "step": 16135 + }, + { + "epoch": 0.3782532510534058, + "grad_norm": 0.7327487468719482, + "learning_rate": 0.00013736395962706723, + "loss": 0.2113, + "step": 16136 + }, + { + "epoch": 0.3782766926282108, + "grad_norm": 0.20954711735248566, + "learning_rate": 0.00013735712686951135, + "loss": 0.0496, + "step": 16137 + }, + { + "epoch": 0.3783001342030158, + "grad_norm": 0.1201181709766388, + "learning_rate": 0.00013735029390925466, + "loss": 0.0223, + "step": 16138 + }, + { + "epoch": 0.3783235757778208, + "grad_norm": 0.5699028372764587, + "learning_rate": 0.0001373434607463342, + "loss": 0.1529, + "step": 16139 + }, + { + "epoch": 0.37834701735262577, + "grad_norm": 0.11668189615011215, + "learning_rate": 0.0001373366273807871, + "loss": 0.02, + "step": 16140 + }, + { + "epoch": 0.37837045892743076, + "grad_norm": 0.6792649626731873, + "learning_rate": 0.00013732979381265038, + "loss": 0.1139, + "step": 16141 + }, + { + "epoch": 0.37839390050223576, + "grad_norm": 0.6339843273162842, + "learning_rate": 0.00013732296004196116, + "loss": 0.0907, + "step": 16142 + }, + { + "epoch": 0.37841734207704075, + "grad_norm": 0.23605960607528687, + "learning_rate": 0.00013731612606875652, + "loss": 0.0245, + "step": 16143 + }, + { + "epoch": 0.37844078365184575, + "grad_norm": 0.42649802565574646, + "learning_rate": 0.00013730929189307352, + "loss": 0.0795, + "step": 16144 + }, + { + "epoch": 0.37846422522665074, + "grad_norm": 0.5088986158370972, + "learning_rate": 0.00013730245751494927, + "loss": 0.0898, + "step": 16145 + }, + { + "epoch": 0.37848766680145574, + "grad_norm": 0.6122829914093018, + "learning_rate": 0.00013729562293442082, + "loss": 0.1553, + "step": 16146 + }, + { + "epoch": 0.37851110837626073, + "grad_norm": 0.664018452167511, + "learning_rate": 0.00013728878815152525, + "loss": 0.0953, + "step": 16147 + }, + { + "epoch": 0.3785345499510657, + "grad_norm": 0.3236607611179352, + "learning_rate": 0.00013728195316629972, + "loss": 0.0807, + "step": 16148 + }, + { + "epoch": 0.3785579915258707, + "grad_norm": 0.37243255972862244, + "learning_rate": 0.00013727511797878122, + "loss": 0.0923, + "step": 16149 + }, + { + "epoch": 0.3785814331006757, + "grad_norm": 0.48262539505958557, + "learning_rate": 0.0001372682825890069, + "loss": 0.1287, + "step": 16150 + }, + { + "epoch": 0.3786048746754807, + "grad_norm": 0.4510411024093628, + "learning_rate": 0.00013726144699701385, + "loss": 0.0899, + "step": 16151 + }, + { + "epoch": 0.3786283162502857, + "grad_norm": 0.6121153235435486, + "learning_rate": 0.0001372546112028391, + "loss": 0.1221, + "step": 16152 + }, + { + "epoch": 0.3786517578250907, + "grad_norm": 0.47435620427131653, + "learning_rate": 0.0001372477752065198, + "loss": 0.097, + "step": 16153 + }, + { + "epoch": 0.3786751993998957, + "grad_norm": 0.13531093299388885, + "learning_rate": 0.00013724093900809305, + "loss": 0.0143, + "step": 16154 + }, + { + "epoch": 0.3786986409747007, + "grad_norm": 0.2544219493865967, + "learning_rate": 0.0001372341026075959, + "loss": 0.0653, + "step": 16155 + }, + { + "epoch": 0.3787220825495057, + "grad_norm": 0.6278785467147827, + "learning_rate": 0.00013722726600506543, + "loss": 0.1916, + "step": 16156 + }, + { + "epoch": 0.3787455241243107, + "grad_norm": 0.13559582829475403, + "learning_rate": 0.00013722042920053882, + "loss": 0.0374, + "step": 16157 + }, + { + "epoch": 0.37876896569911567, + "grad_norm": 0.6535849571228027, + "learning_rate": 0.0001372135921940531, + "loss": 0.1171, + "step": 16158 + }, + { + "epoch": 0.37879240727392066, + "grad_norm": 0.39778047800064087, + "learning_rate": 0.00013720675498564533, + "loss": 0.0794, + "step": 16159 + }, + { + "epoch": 0.37881584884872566, + "grad_norm": 0.4002685248851776, + "learning_rate": 0.0001371999175753527, + "loss": 0.1255, + "step": 16160 + }, + { + "epoch": 0.37883929042353065, + "grad_norm": 0.5340115427970886, + "learning_rate": 0.00013719307996321232, + "loss": 0.5624, + "step": 16161 + }, + { + "epoch": 0.37886273199833564, + "grad_norm": 0.46359527111053467, + "learning_rate": 0.00013718624214926117, + "loss": 0.1138, + "step": 16162 + }, + { + "epoch": 0.37888617357314064, + "grad_norm": 0.21759697794914246, + "learning_rate": 0.00013717940413353645, + "loss": 0.0248, + "step": 16163 + }, + { + "epoch": 0.37890961514794563, + "grad_norm": 0.49559542536735535, + "learning_rate": 0.00013717256591607525, + "loss": 0.0771, + "step": 16164 + }, + { + "epoch": 0.3789330567227506, + "grad_norm": 0.4553191363811493, + "learning_rate": 0.00013716572749691464, + "loss": 0.0567, + "step": 16165 + }, + { + "epoch": 0.3789564982975556, + "grad_norm": 0.40915369987487793, + "learning_rate": 0.00013715888887609175, + "loss": 0.1229, + "step": 16166 + }, + { + "epoch": 0.3789799398723606, + "grad_norm": 0.6288142204284668, + "learning_rate": 0.00013715205005364372, + "loss": 0.0941, + "step": 16167 + }, + { + "epoch": 0.3790033814471656, + "grad_norm": 0.32046496868133545, + "learning_rate": 0.00013714521102960758, + "loss": 0.0668, + "step": 16168 + }, + { + "epoch": 0.3790268230219706, + "grad_norm": 0.3465205132961273, + "learning_rate": 0.0001371383718040205, + "loss": 0.109, + "step": 16169 + }, + { + "epoch": 0.3790502645967756, + "grad_norm": 0.11142513900995255, + "learning_rate": 0.0001371315323769196, + "loss": 0.0148, + "step": 16170 + }, + { + "epoch": 0.3790737061715806, + "grad_norm": 0.3307263255119324, + "learning_rate": 0.00013712469274834187, + "loss": 0.0711, + "step": 16171 + }, + { + "epoch": 0.3790971477463856, + "grad_norm": 0.05550109222531319, + "learning_rate": 0.00013711785291832459, + "loss": 0.0072, + "step": 16172 + }, + { + "epoch": 0.3791205893211906, + "grad_norm": 0.3253559172153473, + "learning_rate": 0.00013711101288690477, + "loss": 0.0605, + "step": 16173 + }, + { + "epoch": 0.3791440308959956, + "grad_norm": 0.41216620802879333, + "learning_rate": 0.00013710417265411958, + "loss": 0.083, + "step": 16174 + }, + { + "epoch": 0.37916747247080057, + "grad_norm": 0.12182269245386124, + "learning_rate": 0.00013709733222000607, + "loss": 0.0234, + "step": 16175 + }, + { + "epoch": 0.37919091404560556, + "grad_norm": 0.4654836058616638, + "learning_rate": 0.0001370904915846014, + "loss": 0.0646, + "step": 16176 + }, + { + "epoch": 0.37921435562041056, + "grad_norm": 0.516002893447876, + "learning_rate": 0.00013708365074794268, + "loss": 0.1086, + "step": 16177 + }, + { + "epoch": 0.37923779719521555, + "grad_norm": 0.8623549938201904, + "learning_rate": 0.00013707680971006704, + "loss": 0.2305, + "step": 16178 + }, + { + "epoch": 0.37926123877002055, + "grad_norm": 0.4950330853462219, + "learning_rate": 0.00013706996847101158, + "loss": 0.1398, + "step": 16179 + }, + { + "epoch": 0.37928468034482554, + "grad_norm": 1.3129515647888184, + "learning_rate": 0.00013706312703081343, + "loss": 0.1288, + "step": 16180 + }, + { + "epoch": 0.37930812191963054, + "grad_norm": 0.43080854415893555, + "learning_rate": 0.0001370562853895097, + "loss": 0.0685, + "step": 16181 + }, + { + "epoch": 0.37933156349443553, + "grad_norm": 0.104234479367733, + "learning_rate": 0.00013704944354713756, + "loss": 0.0211, + "step": 16182 + }, + { + "epoch": 0.3793550050692405, + "grad_norm": 0.7395557761192322, + "learning_rate": 0.00013704260150373404, + "loss": 0.1459, + "step": 16183 + }, + { + "epoch": 0.3793784466440455, + "grad_norm": 0.14800287783145905, + "learning_rate": 0.00013703575925933637, + "loss": 0.0413, + "step": 16184 + }, + { + "epoch": 0.37940188821885057, + "grad_norm": 0.2571972906589508, + "learning_rate": 0.0001370289168139816, + "loss": 0.0575, + "step": 16185 + }, + { + "epoch": 0.37942532979365556, + "grad_norm": 0.5609398484230042, + "learning_rate": 0.0001370220741677069, + "loss": 0.074, + "step": 16186 + }, + { + "epoch": 0.37944877136846056, + "grad_norm": 0.37906113266944885, + "learning_rate": 0.00013701523132054937, + "loss": 0.0547, + "step": 16187 + }, + { + "epoch": 0.37947221294326555, + "grad_norm": 0.4104067385196686, + "learning_rate": 0.0001370083882725462, + "loss": 0.1029, + "step": 16188 + }, + { + "epoch": 0.37949565451807055, + "grad_norm": 0.15876244008541107, + "learning_rate": 0.00013700154502373443, + "loss": 0.0373, + "step": 16189 + }, + { + "epoch": 0.37951909609287554, + "grad_norm": 0.2920099198818207, + "learning_rate": 0.00013699470157415127, + "loss": 0.0765, + "step": 16190 + }, + { + "epoch": 0.37954253766768054, + "grad_norm": 0.6118242144584656, + "learning_rate": 0.00013698785792383376, + "loss": 0.1343, + "step": 16191 + }, + { + "epoch": 0.37956597924248553, + "grad_norm": 0.6542237401008606, + "learning_rate": 0.00013698101407281916, + "loss": 0.1061, + "step": 16192 + }, + { + "epoch": 0.3795894208172905, + "grad_norm": 0.18690478801727295, + "learning_rate": 0.0001369741700211445, + "loss": 0.0174, + "step": 16193 + }, + { + "epoch": 0.3796128623920955, + "grad_norm": 0.3737681806087494, + "learning_rate": 0.00013696732576884698, + "loss": 0.0506, + "step": 16194 + }, + { + "epoch": 0.3796363039669005, + "grad_norm": 1.2308772802352905, + "learning_rate": 0.0001369604813159637, + "loss": 0.082, + "step": 16195 + }, + { + "epoch": 0.3796597455417055, + "grad_norm": 0.21635586023330688, + "learning_rate": 0.0001369536366625318, + "loss": 0.0338, + "step": 16196 + }, + { + "epoch": 0.3796831871165105, + "grad_norm": 0.44849300384521484, + "learning_rate": 0.0001369467918085884, + "loss": 0.0575, + "step": 16197 + }, + { + "epoch": 0.3797066286913155, + "grad_norm": 0.7035813927650452, + "learning_rate": 0.00013693994675417073, + "loss": 0.1656, + "step": 16198 + }, + { + "epoch": 0.3797300702661205, + "grad_norm": 0.1202281191945076, + "learning_rate": 0.00013693310149931584, + "loss": 0.0225, + "step": 16199 + }, + { + "epoch": 0.3797535118409255, + "grad_norm": 0.41303551197052, + "learning_rate": 0.0001369262560440609, + "loss": 0.072, + "step": 16200 + }, + { + "epoch": 0.3797769534157305, + "grad_norm": 0.6427424550056458, + "learning_rate": 0.00013691941038844303, + "loss": 0.7392, + "step": 16201 + }, + { + "epoch": 0.3798003949905355, + "grad_norm": 0.24583972990512848, + "learning_rate": 0.00013691256453249946, + "loss": 0.0626, + "step": 16202 + }, + { + "epoch": 0.37982383656534047, + "grad_norm": 0.17530478537082672, + "learning_rate": 0.00013690571847626724, + "loss": 0.0394, + "step": 16203 + }, + { + "epoch": 0.37984727814014546, + "grad_norm": 0.6127455830574036, + "learning_rate": 0.00013689887221978356, + "loss": 0.1589, + "step": 16204 + }, + { + "epoch": 0.37987071971495046, + "grad_norm": 0.4508674442768097, + "learning_rate": 0.00013689202576308557, + "loss": 0.0606, + "step": 16205 + }, + { + "epoch": 0.37989416128975545, + "grad_norm": 0.5101771950721741, + "learning_rate": 0.0001368851791062104, + "loss": 0.6357, + "step": 16206 + }, + { + "epoch": 0.37991760286456044, + "grad_norm": 0.2718082070350647, + "learning_rate": 0.00013687833224919522, + "loss": 0.0461, + "step": 16207 + }, + { + "epoch": 0.37994104443936544, + "grad_norm": 0.21530301868915558, + "learning_rate": 0.00013687148519207716, + "loss": 0.0274, + "step": 16208 + }, + { + "epoch": 0.37996448601417043, + "grad_norm": 0.057116203010082245, + "learning_rate": 0.00013686463793489338, + "loss": 0.0103, + "step": 16209 + }, + { + "epoch": 0.3799879275889754, + "grad_norm": 0.18986649811267853, + "learning_rate": 0.00013685779047768105, + "loss": 0.0406, + "step": 16210 + }, + { + "epoch": 0.3800113691637804, + "grad_norm": 0.4703514575958252, + "learning_rate": 0.00013685094282047734, + "loss": 0.1067, + "step": 16211 + }, + { + "epoch": 0.3800348107385854, + "grad_norm": 0.4921557903289795, + "learning_rate": 0.00013684409496331936, + "loss": 0.3811, + "step": 16212 + }, + { + "epoch": 0.3800582523133904, + "grad_norm": 0.3504595160484314, + "learning_rate": 0.00013683724690624426, + "loss": 0.0712, + "step": 16213 + }, + { + "epoch": 0.3800816938881954, + "grad_norm": 0.4775773584842682, + "learning_rate": 0.00013683039864928923, + "loss": 0.0733, + "step": 16214 + }, + { + "epoch": 0.3801051354630004, + "grad_norm": 0.5460594296455383, + "learning_rate": 0.00013682355019249145, + "loss": 0.0845, + "step": 16215 + }, + { + "epoch": 0.3801285770378054, + "grad_norm": 0.6324111223220825, + "learning_rate": 0.000136816701535888, + "loss": 0.1333, + "step": 16216 + }, + { + "epoch": 0.3801520186126104, + "grad_norm": 0.3807811439037323, + "learning_rate": 0.00013680985267951613, + "loss": 0.0841, + "step": 16217 + }, + { + "epoch": 0.3801754601874154, + "grad_norm": 0.18128027021884918, + "learning_rate": 0.00013680300362341297, + "loss": 0.0423, + "step": 16218 + }, + { + "epoch": 0.3801989017622204, + "grad_norm": 0.4223232567310333, + "learning_rate": 0.00013679615436761565, + "loss": 0.4434, + "step": 16219 + }, + { + "epoch": 0.38022234333702537, + "grad_norm": 0.521538257598877, + "learning_rate": 0.00013678930491216136, + "loss": 0.1164, + "step": 16220 + }, + { + "epoch": 0.38024578491183036, + "grad_norm": 0.3762122690677643, + "learning_rate": 0.0001367824552570873, + "loss": 0.356, + "step": 16221 + }, + { + "epoch": 0.38026922648663536, + "grad_norm": 0.5683572888374329, + "learning_rate": 0.00013677560540243058, + "loss": 0.6623, + "step": 16222 + }, + { + "epoch": 0.38029266806144035, + "grad_norm": 0.25807124376296997, + "learning_rate": 0.0001367687553482284, + "loss": 0.0534, + "step": 16223 + }, + { + "epoch": 0.38031610963624535, + "grad_norm": 0.43938684463500977, + "learning_rate": 0.00013676190509451794, + "loss": 0.07, + "step": 16224 + }, + { + "epoch": 0.38033955121105034, + "grad_norm": 0.21505628526210785, + "learning_rate": 0.0001367550546413363, + "loss": 0.0251, + "step": 16225 + }, + { + "epoch": 0.38036299278585534, + "grad_norm": 0.6106878519058228, + "learning_rate": 0.00013674820398872073, + "loss": 0.1166, + "step": 16226 + }, + { + "epoch": 0.38038643436066033, + "grad_norm": 0.128469780087471, + "learning_rate": 0.00013674135313670836, + "loss": 0.0275, + "step": 16227 + }, + { + "epoch": 0.3804098759354653, + "grad_norm": 0.4963012635707855, + "learning_rate": 0.00013673450208533642, + "loss": 0.0985, + "step": 16228 + }, + { + "epoch": 0.3804333175102703, + "grad_norm": 0.4201367497444153, + "learning_rate": 0.000136727650834642, + "loss": 0.0405, + "step": 16229 + }, + { + "epoch": 0.3804567590850753, + "grad_norm": 0.6647369265556335, + "learning_rate": 0.0001367207993846623, + "loss": 0.0867, + "step": 16230 + }, + { + "epoch": 0.3804802006598803, + "grad_norm": 0.13970965147018433, + "learning_rate": 0.00013671394773543457, + "loss": 0.0345, + "step": 16231 + }, + { + "epoch": 0.3805036422346853, + "grad_norm": 0.5023396015167236, + "learning_rate": 0.00013670709588699584, + "loss": 0.1324, + "step": 16232 + }, + { + "epoch": 0.3805270838094903, + "grad_norm": 0.5528430938720703, + "learning_rate": 0.00013670024383938343, + "loss": 0.109, + "step": 16233 + }, + { + "epoch": 0.3805505253842953, + "grad_norm": 1.3373184204101562, + "learning_rate": 0.00013669339159263445, + "loss": 0.1547, + "step": 16234 + }, + { + "epoch": 0.3805739669591003, + "grad_norm": 0.5985179543495178, + "learning_rate": 0.00013668653914678612, + "loss": 0.0891, + "step": 16235 + }, + { + "epoch": 0.38059740853390533, + "grad_norm": 0.18814252316951752, + "learning_rate": 0.00013667968650187557, + "loss": 0.0391, + "step": 16236 + }, + { + "epoch": 0.38062085010871033, + "grad_norm": 0.5530514121055603, + "learning_rate": 0.00013667283365794003, + "loss": 0.0945, + "step": 16237 + }, + { + "epoch": 0.3806442916835153, + "grad_norm": 0.35459408164024353, + "learning_rate": 0.00013666598061501663, + "loss": 0.0696, + "step": 16238 + }, + { + "epoch": 0.3806677332583203, + "grad_norm": 1.0476444959640503, + "learning_rate": 0.0001366591273731426, + "loss": 0.1357, + "step": 16239 + }, + { + "epoch": 0.3806911748331253, + "grad_norm": 0.1279098242521286, + "learning_rate": 0.00013665227393235514, + "loss": 0.0221, + "step": 16240 + }, + { + "epoch": 0.3807146164079303, + "grad_norm": 0.22099992632865906, + "learning_rate": 0.00013664542029269137, + "loss": 0.0618, + "step": 16241 + }, + { + "epoch": 0.3807380579827353, + "grad_norm": 0.16764158010482788, + "learning_rate": 0.00013663856645418856, + "loss": 0.0407, + "step": 16242 + }, + { + "epoch": 0.3807614995575403, + "grad_norm": 0.09556210041046143, + "learning_rate": 0.00013663171241688383, + "loss": 0.0152, + "step": 16243 + }, + { + "epoch": 0.3807849411323453, + "grad_norm": 0.4576391577720642, + "learning_rate": 0.0001366248581808144, + "loss": 0.0783, + "step": 16244 + }, + { + "epoch": 0.3808083827071503, + "grad_norm": 0.3489619791507721, + "learning_rate": 0.00013661800374601747, + "loss": 0.0751, + "step": 16245 + }, + { + "epoch": 0.3808318242819553, + "grad_norm": 0.15710395574569702, + "learning_rate": 0.00013661114911253024, + "loss": 0.0362, + "step": 16246 + }, + { + "epoch": 0.38085526585676027, + "grad_norm": 0.4505813419818878, + "learning_rate": 0.00013660429428038987, + "loss": 0.0675, + "step": 16247 + }, + { + "epoch": 0.38087870743156527, + "grad_norm": 0.6916835308074951, + "learning_rate": 0.00013659743924963354, + "loss": 0.1555, + "step": 16248 + }, + { + "epoch": 0.38090214900637026, + "grad_norm": 0.1340218037366867, + "learning_rate": 0.00013659058402029853, + "loss": 0.0106, + "step": 16249 + }, + { + "epoch": 0.38092559058117526, + "grad_norm": 0.4370891749858856, + "learning_rate": 0.00013658372859242195, + "loss": 0.1009, + "step": 16250 + }, + { + "epoch": 0.38094903215598025, + "grad_norm": 0.5890706777572632, + "learning_rate": 0.00013657687296604102, + "loss": 0.101, + "step": 16251 + }, + { + "epoch": 0.38097247373078524, + "grad_norm": 0.5305570960044861, + "learning_rate": 0.00013657001714119298, + "loss": 0.1157, + "step": 16252 + }, + { + "epoch": 0.38099591530559024, + "grad_norm": 0.1722954660654068, + "learning_rate": 0.000136563161117915, + "loss": 0.0352, + "step": 16253 + }, + { + "epoch": 0.38101935688039523, + "grad_norm": 0.09619231522083282, + "learning_rate": 0.00013655630489624424, + "loss": 0.0135, + "step": 16254 + }, + { + "epoch": 0.3810427984552002, + "grad_norm": 1.0419588088989258, + "learning_rate": 0.000136549448476218, + "loss": 0.1373, + "step": 16255 + }, + { + "epoch": 0.3810662400300052, + "grad_norm": 0.8399791121482849, + "learning_rate": 0.00013654259185787336, + "loss": 0.6752, + "step": 16256 + }, + { + "epoch": 0.3810896816048102, + "grad_norm": 0.38907718658447266, + "learning_rate": 0.00013653573504124761, + "loss": 0.0301, + "step": 16257 + }, + { + "epoch": 0.3811131231796152, + "grad_norm": 0.8693209886550903, + "learning_rate": 0.00013652887802637794, + "loss": 0.1959, + "step": 16258 + }, + { + "epoch": 0.3811365647544202, + "grad_norm": 0.5884665250778198, + "learning_rate": 0.00013652202081330155, + "loss": 0.0896, + "step": 16259 + }, + { + "epoch": 0.3811600063292252, + "grad_norm": 0.5037979483604431, + "learning_rate": 0.00013651516340205567, + "loss": 0.0858, + "step": 16260 + }, + { + "epoch": 0.3811834479040302, + "grad_norm": 0.25831857323646545, + "learning_rate": 0.00013650830579267745, + "loss": 0.0453, + "step": 16261 + }, + { + "epoch": 0.3812068894788352, + "grad_norm": 0.13388818502426147, + "learning_rate": 0.00013650144798520416, + "loss": 0.0201, + "step": 16262 + }, + { + "epoch": 0.3812303310536402, + "grad_norm": 0.2045864313840866, + "learning_rate": 0.00013649458997967297, + "loss": 0.0456, + "step": 16263 + }, + { + "epoch": 0.3812537726284452, + "grad_norm": 0.3119399845600128, + "learning_rate": 0.00013648773177612114, + "loss": 0.0398, + "step": 16264 + }, + { + "epoch": 0.38127721420325017, + "grad_norm": 0.21763285994529724, + "learning_rate": 0.0001364808733745858, + "loss": 0.0341, + "step": 16265 + }, + { + "epoch": 0.38130065577805516, + "grad_norm": 0.5693939328193665, + "learning_rate": 0.0001364740147751043, + "loss": 0.0705, + "step": 16266 + }, + { + "epoch": 0.38132409735286016, + "grad_norm": 0.39160236716270447, + "learning_rate": 0.0001364671559777137, + "loss": 0.0366, + "step": 16267 + }, + { + "epoch": 0.38134753892766515, + "grad_norm": 0.6364375948905945, + "learning_rate": 0.00013646029698245128, + "loss": 0.2009, + "step": 16268 + }, + { + "epoch": 0.38137098050247015, + "grad_norm": 0.5041446685791016, + "learning_rate": 0.00013645343778935428, + "loss": 0.0861, + "step": 16269 + }, + { + "epoch": 0.38139442207727514, + "grad_norm": 0.6197994351387024, + "learning_rate": 0.0001364465783984599, + "loss": 0.118, + "step": 16270 + }, + { + "epoch": 0.38141786365208014, + "grad_norm": 0.44976356625556946, + "learning_rate": 0.0001364397188098054, + "loss": 0.0834, + "step": 16271 + }, + { + "epoch": 0.38144130522688513, + "grad_norm": 0.5954004526138306, + "learning_rate": 0.00013643285902342794, + "loss": 0.1766, + "step": 16272 + }, + { + "epoch": 0.3814647468016901, + "grad_norm": 0.5479387640953064, + "learning_rate": 0.00013642599903936474, + "loss": 0.1273, + "step": 16273 + }, + { + "epoch": 0.3814881883764951, + "grad_norm": 0.4830654561519623, + "learning_rate": 0.00013641913885765306, + "loss": 0.0771, + "step": 16274 + }, + { + "epoch": 0.3815116299513001, + "grad_norm": 0.19040872156620026, + "learning_rate": 0.00013641227847833008, + "loss": 0.0434, + "step": 16275 + }, + { + "epoch": 0.3815350715261051, + "grad_norm": 0.4302874505519867, + "learning_rate": 0.0001364054179014331, + "loss": 0.1134, + "step": 16276 + }, + { + "epoch": 0.3815585131009101, + "grad_norm": 0.6166797876358032, + "learning_rate": 0.00013639855712699927, + "loss": 0.0891, + "step": 16277 + }, + { + "epoch": 0.3815819546757151, + "grad_norm": 0.6078586578369141, + "learning_rate": 0.00013639169615506586, + "loss": 0.0712, + "step": 16278 + }, + { + "epoch": 0.3816053962505201, + "grad_norm": 0.9558401703834534, + "learning_rate": 0.00013638483498567007, + "loss": 0.2001, + "step": 16279 + }, + { + "epoch": 0.3816288378253251, + "grad_norm": 0.1806439757347107, + "learning_rate": 0.00013637797361884914, + "loss": 0.0237, + "step": 16280 + }, + { + "epoch": 0.3816522794001301, + "grad_norm": 0.4448198974132538, + "learning_rate": 0.00013637111205464032, + "loss": 0.0916, + "step": 16281 + }, + { + "epoch": 0.3816757209749351, + "grad_norm": 0.6129941344261169, + "learning_rate": 0.00013636425029308081, + "loss": 0.5521, + "step": 16282 + }, + { + "epoch": 0.38169916254974007, + "grad_norm": 0.5069112777709961, + "learning_rate": 0.00013635738833420787, + "loss": 0.0761, + "step": 16283 + }, + { + "epoch": 0.38172260412454506, + "grad_norm": 0.31045061349868774, + "learning_rate": 0.0001363505261780587, + "loss": 0.0939, + "step": 16284 + }, + { + "epoch": 0.38174604569935006, + "grad_norm": 0.45191752910614014, + "learning_rate": 0.00013634366382467059, + "loss": 0.1154, + "step": 16285 + }, + { + "epoch": 0.38176948727415505, + "grad_norm": 0.090671107172966, + "learning_rate": 0.0001363368012740807, + "loss": 0.0224, + "step": 16286 + }, + { + "epoch": 0.3817929288489601, + "grad_norm": 0.48962846398353577, + "learning_rate": 0.0001363299385263263, + "loss": 0.096, + "step": 16287 + }, + { + "epoch": 0.3818163704237651, + "grad_norm": 0.4061151146888733, + "learning_rate": 0.00013632307558144466, + "loss": 0.0667, + "step": 16288 + }, + { + "epoch": 0.3818398119985701, + "grad_norm": 0.22915786504745483, + "learning_rate": 0.00013631621243947296, + "loss": 0.068, + "step": 16289 + }, + { + "epoch": 0.3818632535733751, + "grad_norm": 0.47657448053359985, + "learning_rate": 0.0001363093491004485, + "loss": 0.1296, + "step": 16290 + }, + { + "epoch": 0.3818866951481801, + "grad_norm": 0.887479841709137, + "learning_rate": 0.0001363024855644085, + "loss": 0.2839, + "step": 16291 + }, + { + "epoch": 0.38191013672298507, + "grad_norm": 0.22482267022132874, + "learning_rate": 0.00013629562183139015, + "loss": 0.0485, + "step": 16292 + }, + { + "epoch": 0.38193357829779007, + "grad_norm": 0.2268168330192566, + "learning_rate": 0.00013628875790143075, + "loss": 0.0487, + "step": 16293 + }, + { + "epoch": 0.38195701987259506, + "grad_norm": 0.653035044670105, + "learning_rate": 0.00013628189377456752, + "loss": 0.672, + "step": 16294 + }, + { + "epoch": 0.38198046144740005, + "grad_norm": 0.21721187233924866, + "learning_rate": 0.00013627502945083776, + "loss": 0.0566, + "step": 16295 + }, + { + "epoch": 0.38200390302220505, + "grad_norm": 0.5712680816650391, + "learning_rate": 0.00013626816493027863, + "loss": 0.1385, + "step": 16296 + }, + { + "epoch": 0.38202734459701004, + "grad_norm": 0.6101430654525757, + "learning_rate": 0.00013626130021292742, + "loss": 0.1192, + "step": 16297 + }, + { + "epoch": 0.38205078617181504, + "grad_norm": 0.35214313864707947, + "learning_rate": 0.0001362544352988214, + "loss": 0.0755, + "step": 16298 + }, + { + "epoch": 0.38207422774662003, + "grad_norm": 0.5066328644752502, + "learning_rate": 0.00013624757018799774, + "loss": 0.1337, + "step": 16299 + }, + { + "epoch": 0.382097669321425, + "grad_norm": 0.4594016969203949, + "learning_rate": 0.0001362407048804938, + "loss": 0.0758, + "step": 16300 + }, + { + "epoch": 0.38212111089623, + "grad_norm": 0.5150830745697021, + "learning_rate": 0.00013623383937634678, + "loss": 0.1345, + "step": 16301 + }, + { + "epoch": 0.382144552471035, + "grad_norm": 0.6220816373825073, + "learning_rate": 0.0001362269736755939, + "loss": 0.1118, + "step": 16302 + }, + { + "epoch": 0.38216799404584, + "grad_norm": 0.36153271794319153, + "learning_rate": 0.00013622010777827247, + "loss": 0.054, + "step": 16303 + }, + { + "epoch": 0.382191435620645, + "grad_norm": 0.12681537866592407, + "learning_rate": 0.0001362132416844197, + "loss": 0.0265, + "step": 16304 + }, + { + "epoch": 0.38221487719545, + "grad_norm": 0.47642770409584045, + "learning_rate": 0.0001362063753940728, + "loss": 0.1419, + "step": 16305 + }, + { + "epoch": 0.382238318770255, + "grad_norm": 0.4092963635921478, + "learning_rate": 0.00013619950890726918, + "loss": 0.0692, + "step": 16306 + }, + { + "epoch": 0.38226176034506, + "grad_norm": 0.6085245609283447, + "learning_rate": 0.00013619264222404598, + "loss": 0.0753, + "step": 16307 + }, + { + "epoch": 0.382285201919865, + "grad_norm": 0.35933494567871094, + "learning_rate": 0.00013618577534444046, + "loss": 0.0548, + "step": 16308 + }, + { + "epoch": 0.38230864349467, + "grad_norm": 0.4516613185405731, + "learning_rate": 0.00013617890826848994, + "loss": 0.0797, + "step": 16309 + }, + { + "epoch": 0.38233208506947497, + "grad_norm": 0.5014689564704895, + "learning_rate": 0.00013617204099623163, + "loss": 0.0488, + "step": 16310 + }, + { + "epoch": 0.38235552664427996, + "grad_norm": 0.18569816648960114, + "learning_rate": 0.0001361651735277028, + "loss": 0.041, + "step": 16311 + }, + { + "epoch": 0.38237896821908496, + "grad_norm": 0.44549760222435, + "learning_rate": 0.00013615830586294074, + "loss": 0.0906, + "step": 16312 + }, + { + "epoch": 0.38240240979388995, + "grad_norm": 0.3538360893726349, + "learning_rate": 0.00013615143800198268, + "loss": 0.0544, + "step": 16313 + }, + { + "epoch": 0.38242585136869495, + "grad_norm": 0.19230026006698608, + "learning_rate": 0.00013614456994486592, + "loss": 0.0455, + "step": 16314 + }, + { + "epoch": 0.38244929294349994, + "grad_norm": 0.3164096176624298, + "learning_rate": 0.0001361377016916277, + "loss": 0.056, + "step": 16315 + }, + { + "epoch": 0.38247273451830494, + "grad_norm": 0.12249341607093811, + "learning_rate": 0.0001361308332423053, + "loss": 0.0115, + "step": 16316 + }, + { + "epoch": 0.38249617609310993, + "grad_norm": 0.11306086927652359, + "learning_rate": 0.00013612396459693595, + "loss": 0.0238, + "step": 16317 + }, + { + "epoch": 0.3825196176679149, + "grad_norm": 0.3444438576698303, + "learning_rate": 0.00013611709575555696, + "loss": 0.0552, + "step": 16318 + }, + { + "epoch": 0.3825430592427199, + "grad_norm": 0.30611652135849, + "learning_rate": 0.00013611022671820565, + "loss": 0.0862, + "step": 16319 + }, + { + "epoch": 0.3825665008175249, + "grad_norm": 0.535873532295227, + "learning_rate": 0.00013610335748491918, + "loss": 0.7319, + "step": 16320 + }, + { + "epoch": 0.3825899423923299, + "grad_norm": 0.4548987150192261, + "learning_rate": 0.0001360964880557349, + "loss": 0.0779, + "step": 16321 + }, + { + "epoch": 0.3826133839671349, + "grad_norm": 0.41781002283096313, + "learning_rate": 0.00013608961843069003, + "loss": 0.0502, + "step": 16322 + }, + { + "epoch": 0.3826368255419399, + "grad_norm": 0.5834082365036011, + "learning_rate": 0.0001360827486098219, + "loss": 0.1111, + "step": 16323 + }, + { + "epoch": 0.3826602671167449, + "grad_norm": 0.35306793451309204, + "learning_rate": 0.00013607587859316774, + "loss": 0.1, + "step": 16324 + }, + { + "epoch": 0.3826837086915499, + "grad_norm": 0.26367422938346863, + "learning_rate": 0.0001360690083807649, + "loss": 0.0252, + "step": 16325 + }, + { + "epoch": 0.3827071502663549, + "grad_norm": 0.24047155678272247, + "learning_rate": 0.00013606213797265058, + "loss": 0.0265, + "step": 16326 + }, + { + "epoch": 0.3827305918411599, + "grad_norm": 0.3867723047733307, + "learning_rate": 0.00013605526736886207, + "loss": 0.0591, + "step": 16327 + }, + { + "epoch": 0.38275403341596487, + "grad_norm": 0.23555032908916473, + "learning_rate": 0.00013604839656943665, + "loss": 0.0266, + "step": 16328 + }, + { + "epoch": 0.38277747499076986, + "grad_norm": 0.32224059104919434, + "learning_rate": 0.00013604152557441165, + "loss": 0.0482, + "step": 16329 + }, + { + "epoch": 0.38280091656557486, + "grad_norm": 0.4801110327243805, + "learning_rate": 0.00013603465438382433, + "loss": 0.0871, + "step": 16330 + }, + { + "epoch": 0.38282435814037985, + "grad_norm": 0.402549147605896, + "learning_rate": 0.00013602778299771193, + "loss": 0.0847, + "step": 16331 + }, + { + "epoch": 0.38284779971518484, + "grad_norm": 0.7119883298873901, + "learning_rate": 0.0001360209114161118, + "loss": 0.1337, + "step": 16332 + }, + { + "epoch": 0.38287124128998984, + "grad_norm": 0.40773525834083557, + "learning_rate": 0.00013601403963906113, + "loss": 0.4621, + "step": 16333 + }, + { + "epoch": 0.38289468286479483, + "grad_norm": 0.5897722244262695, + "learning_rate": 0.00013600716766659732, + "loss": 0.182, + "step": 16334 + }, + { + "epoch": 0.3829181244395998, + "grad_norm": 0.1649540513753891, + "learning_rate": 0.0001360002954987576, + "loss": 0.0337, + "step": 16335 + }, + { + "epoch": 0.3829415660144048, + "grad_norm": 0.10253754258155823, + "learning_rate": 0.00013599342313557925, + "loss": 0.0233, + "step": 16336 + }, + { + "epoch": 0.3829650075892098, + "grad_norm": 0.17003707587718964, + "learning_rate": 0.00013598655057709959, + "loss": 0.0436, + "step": 16337 + }, + { + "epoch": 0.38298844916401487, + "grad_norm": 0.6309219598770142, + "learning_rate": 0.00013597967782335587, + "loss": 0.1314, + "step": 16338 + }, + { + "epoch": 0.38301189073881986, + "grad_norm": 0.3162772059440613, + "learning_rate": 0.00013597280487438545, + "loss": 0.3329, + "step": 16339 + }, + { + "epoch": 0.38303533231362485, + "grad_norm": 0.4339786171913147, + "learning_rate": 0.00013596593173022554, + "loss": 0.0858, + "step": 16340 + }, + { + "epoch": 0.38305877388842985, + "grad_norm": 0.33244162797927856, + "learning_rate": 0.00013595905839091347, + "loss": 0.0807, + "step": 16341 + }, + { + "epoch": 0.38308221546323484, + "grad_norm": 0.23022809624671936, + "learning_rate": 0.00013595218485648657, + "loss": 0.0353, + "step": 16342 + }, + { + "epoch": 0.38310565703803984, + "grad_norm": 0.32398319244384766, + "learning_rate": 0.0001359453111269821, + "loss": 0.0359, + "step": 16343 + }, + { + "epoch": 0.38312909861284483, + "grad_norm": 0.3130492866039276, + "learning_rate": 0.00013593843720243736, + "loss": 0.0443, + "step": 16344 + }, + { + "epoch": 0.3831525401876498, + "grad_norm": 0.8050702214241028, + "learning_rate": 0.00013593156308288964, + "loss": 0.2289, + "step": 16345 + }, + { + "epoch": 0.3831759817624548, + "grad_norm": 0.5165969133377075, + "learning_rate": 0.00013592468876837623, + "loss": 0.162, + "step": 16346 + }, + { + "epoch": 0.3831994233372598, + "grad_norm": 0.5903590321540833, + "learning_rate": 0.00013591781425893444, + "loss": 0.1108, + "step": 16347 + }, + { + "epoch": 0.3832228649120648, + "grad_norm": 0.14830400049686432, + "learning_rate": 0.00013591093955460162, + "loss": 0.0273, + "step": 16348 + }, + { + "epoch": 0.3832463064868698, + "grad_norm": 0.837374746799469, + "learning_rate": 0.00013590406465541503, + "loss": 0.1746, + "step": 16349 + }, + { + "epoch": 0.3832697480616748, + "grad_norm": 0.3019678294658661, + "learning_rate": 0.00013589718956141195, + "loss": 0.0724, + "step": 16350 + }, + { + "epoch": 0.3832931896364798, + "grad_norm": 0.4040251076221466, + "learning_rate": 0.00013589031427262973, + "loss": 0.0382, + "step": 16351 + }, + { + "epoch": 0.3833166312112848, + "grad_norm": 0.13835501670837402, + "learning_rate": 0.00013588343878910566, + "loss": 0.0162, + "step": 16352 + }, + { + "epoch": 0.3833400727860898, + "grad_norm": 0.24013131856918335, + "learning_rate": 0.00013587656311087702, + "loss": 0.0349, + "step": 16353 + }, + { + "epoch": 0.3833635143608948, + "grad_norm": 0.6529499292373657, + "learning_rate": 0.00013586968723798115, + "loss": 0.1277, + "step": 16354 + }, + { + "epoch": 0.38338695593569977, + "grad_norm": 0.6662614345550537, + "learning_rate": 0.00013586281117045537, + "loss": 0.1342, + "step": 16355 + }, + { + "epoch": 0.38341039751050476, + "grad_norm": 0.5416436791419983, + "learning_rate": 0.0001358559349083369, + "loss": 0.1419, + "step": 16356 + }, + { + "epoch": 0.38343383908530976, + "grad_norm": 0.3463529944419861, + "learning_rate": 0.0001358490584516632, + "loss": 0.0923, + "step": 16357 + }, + { + "epoch": 0.38345728066011475, + "grad_norm": 0.49287334084510803, + "learning_rate": 0.0001358421818004715, + "loss": 0.1092, + "step": 16358 + }, + { + "epoch": 0.38348072223491975, + "grad_norm": 0.3316982388496399, + "learning_rate": 0.00013583530495479905, + "loss": 0.0785, + "step": 16359 + }, + { + "epoch": 0.38350416380972474, + "grad_norm": 0.4533955752849579, + "learning_rate": 0.00013582842791468324, + "loss": 0.581, + "step": 16360 + }, + { + "epoch": 0.38352760538452974, + "grad_norm": 0.21185433864593506, + "learning_rate": 0.0001358215506801614, + "loss": 0.0489, + "step": 16361 + }, + { + "epoch": 0.38355104695933473, + "grad_norm": 0.11938168853521347, + "learning_rate": 0.0001358146732512708, + "loss": 0.0203, + "step": 16362 + }, + { + "epoch": 0.3835744885341397, + "grad_norm": 0.6847689151763916, + "learning_rate": 0.00013580779562804882, + "loss": 0.6384, + "step": 16363 + }, + { + "epoch": 0.3835979301089447, + "grad_norm": 0.6582154631614685, + "learning_rate": 0.0001358009178105327, + "loss": 0.1626, + "step": 16364 + }, + { + "epoch": 0.3836213716837497, + "grad_norm": 0.0859277993440628, + "learning_rate": 0.00013579403979875978, + "loss": 0.0143, + "step": 16365 + }, + { + "epoch": 0.3836448132585547, + "grad_norm": 0.45821067690849304, + "learning_rate": 0.0001357871615927674, + "loss": 0.1216, + "step": 16366 + }, + { + "epoch": 0.3836682548333597, + "grad_norm": 0.6600438952445984, + "learning_rate": 0.00013578028319259292, + "loss": 0.2124, + "step": 16367 + }, + { + "epoch": 0.3836916964081647, + "grad_norm": 0.10060546547174454, + "learning_rate": 0.00013577340459827356, + "loss": 0.0121, + "step": 16368 + }, + { + "epoch": 0.3837151379829697, + "grad_norm": 0.5226331353187561, + "learning_rate": 0.0001357665258098467, + "loss": 0.1005, + "step": 16369 + }, + { + "epoch": 0.3837385795577747, + "grad_norm": 0.6973390579223633, + "learning_rate": 0.00013575964682734968, + "loss": 0.1354, + "step": 16370 + }, + { + "epoch": 0.3837620211325797, + "grad_norm": 0.11871732771396637, + "learning_rate": 0.00013575276765081983, + "loss": 0.0248, + "step": 16371 + }, + { + "epoch": 0.3837854627073847, + "grad_norm": 0.5346866250038147, + "learning_rate": 0.00013574588828029445, + "loss": 0.1024, + "step": 16372 + }, + { + "epoch": 0.38380890428218967, + "grad_norm": 0.7720367908477783, + "learning_rate": 0.00013573900871581088, + "loss": 0.199, + "step": 16373 + }, + { + "epoch": 0.38383234585699466, + "grad_norm": 0.38898810744285583, + "learning_rate": 0.00013573212895740645, + "loss": 0.0397, + "step": 16374 + }, + { + "epoch": 0.38385578743179966, + "grad_norm": 0.6852725148200989, + "learning_rate": 0.00013572524900511845, + "loss": 0.2112, + "step": 16375 + }, + { + "epoch": 0.38387922900660465, + "grad_norm": 0.24123363196849823, + "learning_rate": 0.00013571836885898427, + "loss": 0.0549, + "step": 16376 + }, + { + "epoch": 0.38390267058140964, + "grad_norm": 0.15850219130516052, + "learning_rate": 0.00013571148851904122, + "loss": 0.0266, + "step": 16377 + }, + { + "epoch": 0.38392611215621464, + "grad_norm": 1.0093497037887573, + "learning_rate": 0.00013570460798532658, + "loss": 0.0863, + "step": 16378 + }, + { + "epoch": 0.38394955373101963, + "grad_norm": 0.23607879877090454, + "learning_rate": 0.00013569772725787778, + "loss": 0.0301, + "step": 16379 + }, + { + "epoch": 0.3839729953058246, + "grad_norm": 0.48838987946510315, + "learning_rate": 0.0001356908463367321, + "loss": 0.097, + "step": 16380 + }, + { + "epoch": 0.3839964368806296, + "grad_norm": 0.5496447682380676, + "learning_rate": 0.00013568396522192688, + "loss": 0.6798, + "step": 16381 + }, + { + "epoch": 0.3840198784554346, + "grad_norm": 0.8406142592430115, + "learning_rate": 0.00013567708391349948, + "loss": 0.1028, + "step": 16382 + }, + { + "epoch": 0.3840433200302396, + "grad_norm": 0.7888852953910828, + "learning_rate": 0.0001356702024114872, + "loss": 0.1479, + "step": 16383 + }, + { + "epoch": 0.3840667616050446, + "grad_norm": 0.17666319012641907, + "learning_rate": 0.00013566332071592737, + "loss": 0.043, + "step": 16384 + }, + { + "epoch": 0.3840902031798496, + "grad_norm": 0.3594735860824585, + "learning_rate": 0.0001356564388268574, + "loss": 0.073, + "step": 16385 + }, + { + "epoch": 0.3841136447546546, + "grad_norm": 0.576278805732727, + "learning_rate": 0.00013564955674431458, + "loss": 0.1036, + "step": 16386 + }, + { + "epoch": 0.3841370863294596, + "grad_norm": 0.37985944747924805, + "learning_rate": 0.0001356426744683363, + "loss": 0.034, + "step": 16387 + }, + { + "epoch": 0.3841605279042646, + "grad_norm": 0.13588422536849976, + "learning_rate": 0.0001356357919989598, + "loss": 0.0193, + "step": 16388 + }, + { + "epoch": 0.38418396947906963, + "grad_norm": 0.25100862979888916, + "learning_rate": 0.00013562890933622256, + "loss": 0.0366, + "step": 16389 + }, + { + "epoch": 0.3842074110538746, + "grad_norm": 0.4244205951690674, + "learning_rate": 0.00013562202648016177, + "loss": 0.1136, + "step": 16390 + }, + { + "epoch": 0.3842308526286796, + "grad_norm": 0.4879317581653595, + "learning_rate": 0.00013561514343081492, + "loss": 0.088, + "step": 16391 + }, + { + "epoch": 0.3842542942034846, + "grad_norm": 0.8460612893104553, + "learning_rate": 0.00013560826018821928, + "loss": 0.1587, + "step": 16392 + }, + { + "epoch": 0.3842777357782896, + "grad_norm": 0.3173975646495819, + "learning_rate": 0.00013560137675241226, + "loss": 0.4266, + "step": 16393 + }, + { + "epoch": 0.3843011773530946, + "grad_norm": 0.6638728976249695, + "learning_rate": 0.00013559449312343112, + "loss": 0.0655, + "step": 16394 + }, + { + "epoch": 0.3843246189278996, + "grad_norm": 0.4757522642612457, + "learning_rate": 0.0001355876093013133, + "loss": 0.0826, + "step": 16395 + }, + { + "epoch": 0.3843480605027046, + "grad_norm": 0.4862249493598938, + "learning_rate": 0.00013558072528609607, + "loss": 0.0932, + "step": 16396 + }, + { + "epoch": 0.3843715020775096, + "grad_norm": 0.2605648636817932, + "learning_rate": 0.00013557384107781682, + "loss": 0.0313, + "step": 16397 + }, + { + "epoch": 0.3843949436523146, + "grad_norm": 0.22733210027217865, + "learning_rate": 0.00013556695667651293, + "loss": 0.0849, + "step": 16398 + }, + { + "epoch": 0.3844183852271196, + "grad_norm": 0.6518276333808899, + "learning_rate": 0.00013556007208222174, + "loss": 0.5339, + "step": 16399 + }, + { + "epoch": 0.38444182680192457, + "grad_norm": 0.6244145631790161, + "learning_rate": 0.00013555318729498059, + "loss": 0.1708, + "step": 16400 + }, + { + "epoch": 0.38446526837672956, + "grad_norm": 0.21772949397563934, + "learning_rate": 0.00013554630231482685, + "loss": 0.0397, + "step": 16401 + }, + { + "epoch": 0.38448870995153456, + "grad_norm": 0.27683040499687195, + "learning_rate": 0.0001355394171417979, + "loss": 0.0647, + "step": 16402 + }, + { + "epoch": 0.38451215152633955, + "grad_norm": 0.27533459663391113, + "learning_rate": 0.00013553253177593105, + "loss": 0.0758, + "step": 16403 + }, + { + "epoch": 0.38453559310114455, + "grad_norm": 0.3953309655189514, + "learning_rate": 0.00013552564621726365, + "loss": 0.0555, + "step": 16404 + }, + { + "epoch": 0.38455903467594954, + "grad_norm": 0.5918149352073669, + "learning_rate": 0.00013551876046583313, + "loss": 0.1371, + "step": 16405 + }, + { + "epoch": 0.38458247625075453, + "grad_norm": 0.6512346267700195, + "learning_rate": 0.00013551187452167683, + "loss": 0.5624, + "step": 16406 + }, + { + "epoch": 0.38460591782555953, + "grad_norm": 0.14697472751140594, + "learning_rate": 0.00013550498838483206, + "loss": 0.0254, + "step": 16407 + }, + { + "epoch": 0.3846293594003645, + "grad_norm": 0.6033806204795837, + "learning_rate": 0.00013549810205533625, + "loss": 0.0998, + "step": 16408 + }, + { + "epoch": 0.3846528009751695, + "grad_norm": 0.7891350984573364, + "learning_rate": 0.00013549121553322674, + "loss": 0.1204, + "step": 16409 + }, + { + "epoch": 0.3846762425499745, + "grad_norm": 0.7245762944221497, + "learning_rate": 0.00013548432881854087, + "loss": 0.7629, + "step": 16410 + }, + { + "epoch": 0.3846996841247795, + "grad_norm": 0.34265512228012085, + "learning_rate": 0.00013547744191131608, + "loss": 0.1025, + "step": 16411 + }, + { + "epoch": 0.3847231256995845, + "grad_norm": 0.4442783296108246, + "learning_rate": 0.00013547055481158968, + "loss": 0.1079, + "step": 16412 + }, + { + "epoch": 0.3847465672743895, + "grad_norm": 0.1932850182056427, + "learning_rate": 0.00013546366751939902, + "loss": 0.046, + "step": 16413 + }, + { + "epoch": 0.3847700088491945, + "grad_norm": 0.43020108342170715, + "learning_rate": 0.00013545678003478153, + "loss": 0.0749, + "step": 16414 + }, + { + "epoch": 0.3847934504239995, + "grad_norm": 0.9301680326461792, + "learning_rate": 0.00013544989235777456, + "loss": 0.157, + "step": 16415 + }, + { + "epoch": 0.3848168919988045, + "grad_norm": 0.3845131993293762, + "learning_rate": 0.00013544300448841546, + "loss": 0.0601, + "step": 16416 + }, + { + "epoch": 0.38484033357360947, + "grad_norm": 0.45432519912719727, + "learning_rate": 0.00013543611642674163, + "loss": 0.1216, + "step": 16417 + }, + { + "epoch": 0.38486377514841447, + "grad_norm": 0.5485075116157532, + "learning_rate": 0.00013542922817279046, + "loss": 0.1356, + "step": 16418 + }, + { + "epoch": 0.38488721672321946, + "grad_norm": 0.47327035665512085, + "learning_rate": 0.00013542233972659926, + "loss": 0.0872, + "step": 16419 + }, + { + "epoch": 0.38491065829802446, + "grad_norm": 0.6309629678726196, + "learning_rate": 0.00013541545108820545, + "loss": 0.5265, + "step": 16420 + }, + { + "epoch": 0.38493409987282945, + "grad_norm": 0.25302234292030334, + "learning_rate": 0.00013540856225764642, + "loss": 0.0362, + "step": 16421 + }, + { + "epoch": 0.38495754144763444, + "grad_norm": 1.7244408130645752, + "learning_rate": 0.00013540167323495956, + "loss": 0.115, + "step": 16422 + }, + { + "epoch": 0.38498098302243944, + "grad_norm": 0.12782353162765503, + "learning_rate": 0.00013539478402018216, + "loss": 0.0294, + "step": 16423 + }, + { + "epoch": 0.38500442459724443, + "grad_norm": 0.7313823103904724, + "learning_rate": 0.00013538789461335174, + "loss": 0.0637, + "step": 16424 + }, + { + "epoch": 0.3850278661720494, + "grad_norm": 0.3188621699810028, + "learning_rate": 0.00013538100501450559, + "loss": 0.0341, + "step": 16425 + }, + { + "epoch": 0.3850513077468544, + "grad_norm": 0.123659148812294, + "learning_rate": 0.00013537411522368106, + "loss": 0.0259, + "step": 16426 + }, + { + "epoch": 0.3850747493216594, + "grad_norm": 0.3806248605251312, + "learning_rate": 0.00013536722524091566, + "loss": 0.1164, + "step": 16427 + }, + { + "epoch": 0.3850981908964644, + "grad_norm": 0.2906377911567688, + "learning_rate": 0.00013536033506624664, + "loss": 0.0481, + "step": 16428 + }, + { + "epoch": 0.3851216324712694, + "grad_norm": 0.40893322229385376, + "learning_rate": 0.00013535344469971147, + "loss": 0.071, + "step": 16429 + }, + { + "epoch": 0.3851450740460744, + "grad_norm": 0.46458935737609863, + "learning_rate": 0.00013534655414134753, + "loss": 0.1061, + "step": 16430 + }, + { + "epoch": 0.3851685156208794, + "grad_norm": 0.5515010952949524, + "learning_rate": 0.0001353396633911922, + "loss": 0.1, + "step": 16431 + }, + { + "epoch": 0.3851919571956844, + "grad_norm": 0.4211265444755554, + "learning_rate": 0.00013533277244928284, + "loss": 0.1144, + "step": 16432 + }, + { + "epoch": 0.3852153987704894, + "grad_norm": 0.4394645094871521, + "learning_rate": 0.00013532588131565687, + "loss": 0.1021, + "step": 16433 + }, + { + "epoch": 0.3852388403452944, + "grad_norm": 0.6815153360366821, + "learning_rate": 0.00013531898999035168, + "loss": 0.1376, + "step": 16434 + }, + { + "epoch": 0.38526228192009937, + "grad_norm": 0.7316232919692993, + "learning_rate": 0.00013531209847340465, + "loss": 0.1647, + "step": 16435 + }, + { + "epoch": 0.38528572349490436, + "grad_norm": 0.10354836285114288, + "learning_rate": 0.00013530520676485318, + "loss": 0.0181, + "step": 16436 + }, + { + "epoch": 0.38530916506970936, + "grad_norm": 0.5553536415100098, + "learning_rate": 0.00013529831486473469, + "loss": 0.0994, + "step": 16437 + }, + { + "epoch": 0.38533260664451435, + "grad_norm": 0.13495692610740662, + "learning_rate": 0.0001352914227730865, + "loss": 0.0206, + "step": 16438 + }, + { + "epoch": 0.38535604821931935, + "grad_norm": 0.35550424456596375, + "learning_rate": 0.00013528453048994608, + "loss": 0.0513, + "step": 16439 + }, + { + "epoch": 0.3853794897941244, + "grad_norm": 0.7400761246681213, + "learning_rate": 0.00013527763801535084, + "loss": 0.0824, + "step": 16440 + }, + { + "epoch": 0.3854029313689294, + "grad_norm": 0.5638307929039001, + "learning_rate": 0.0001352707453493381, + "loss": 0.1081, + "step": 16441 + }, + { + "epoch": 0.3854263729437344, + "grad_norm": 0.14612813293933868, + "learning_rate": 0.0001352638524919453, + "loss": 0.0256, + "step": 16442 + }, + { + "epoch": 0.3854498145185394, + "grad_norm": 0.483125239610672, + "learning_rate": 0.00013525695944320988, + "loss": 0.4612, + "step": 16443 + }, + { + "epoch": 0.3854732560933444, + "grad_norm": 0.4631843864917755, + "learning_rate": 0.00013525006620316923, + "loss": 0.1014, + "step": 16444 + }, + { + "epoch": 0.38549669766814937, + "grad_norm": 0.4007243514060974, + "learning_rate": 0.00013524317277186065, + "loss": 0.0618, + "step": 16445 + }, + { + "epoch": 0.38552013924295436, + "grad_norm": 0.5168702602386475, + "learning_rate": 0.00013523627914932166, + "loss": 0.067, + "step": 16446 + }, + { + "epoch": 0.38554358081775936, + "grad_norm": 0.15837818384170532, + "learning_rate": 0.00013522938533558967, + "loss": 0.0343, + "step": 16447 + }, + { + "epoch": 0.38556702239256435, + "grad_norm": 0.599742591381073, + "learning_rate": 0.00013522249133070198, + "loss": 0.0713, + "step": 16448 + }, + { + "epoch": 0.38559046396736935, + "grad_norm": 0.24657340347766876, + "learning_rate": 0.0001352155971346961, + "loss": 0.0425, + "step": 16449 + }, + { + "epoch": 0.38561390554217434, + "grad_norm": 0.5181114673614502, + "learning_rate": 0.0001352087027476094, + "loss": 0.0818, + "step": 16450 + }, + { + "epoch": 0.38563734711697933, + "grad_norm": 0.18031315505504608, + "learning_rate": 0.00013520180816947925, + "loss": 0.0245, + "step": 16451 + }, + { + "epoch": 0.38566078869178433, + "grad_norm": 0.8601247668266296, + "learning_rate": 0.00013519491340034315, + "loss": 0.1122, + "step": 16452 + }, + { + "epoch": 0.3856842302665893, + "grad_norm": 0.49645596742630005, + "learning_rate": 0.00013518801844023844, + "loss": 0.1074, + "step": 16453 + }, + { + "epoch": 0.3857076718413943, + "grad_norm": 0.4519799053668976, + "learning_rate": 0.00013518112328920253, + "loss": 0.1149, + "step": 16454 + }, + { + "epoch": 0.3857311134161993, + "grad_norm": 0.507409930229187, + "learning_rate": 0.00013517422794727288, + "loss": 0.1204, + "step": 16455 + }, + { + "epoch": 0.3857545549910043, + "grad_norm": 0.49246108531951904, + "learning_rate": 0.0001351673324144869, + "loss": 0.1451, + "step": 16456 + }, + { + "epoch": 0.3857779965658093, + "grad_norm": 0.4484005272388458, + "learning_rate": 0.00013516043669088194, + "loss": 0.1152, + "step": 16457 + }, + { + "epoch": 0.3858014381406143, + "grad_norm": 0.21880248188972473, + "learning_rate": 0.00013515354077649547, + "loss": 0.0394, + "step": 16458 + }, + { + "epoch": 0.3858248797154193, + "grad_norm": 0.40466243028640747, + "learning_rate": 0.00013514664467136493, + "loss": 0.5653, + "step": 16459 + }, + { + "epoch": 0.3858483212902243, + "grad_norm": 0.17769792675971985, + "learning_rate": 0.0001351397483755277, + "loss": 0.0515, + "step": 16460 + }, + { + "epoch": 0.3858717628650293, + "grad_norm": 0.2704939842224121, + "learning_rate": 0.00013513285188902118, + "loss": 0.067, + "step": 16461 + }, + { + "epoch": 0.38589520443983427, + "grad_norm": 0.9017627239227295, + "learning_rate": 0.00013512595521188284, + "loss": 0.1309, + "step": 16462 + }, + { + "epoch": 0.38591864601463927, + "grad_norm": 0.6292144060134888, + "learning_rate": 0.00013511905834415007, + "loss": 0.7118, + "step": 16463 + }, + { + "epoch": 0.38594208758944426, + "grad_norm": 0.6838327646255493, + "learning_rate": 0.0001351121612858603, + "loss": 0.1109, + "step": 16464 + }, + { + "epoch": 0.38596552916424925, + "grad_norm": 0.49598363041877747, + "learning_rate": 0.00013510526403705096, + "loss": 0.1043, + "step": 16465 + }, + { + "epoch": 0.38598897073905425, + "grad_norm": 0.5744310021400452, + "learning_rate": 0.0001350983665977595, + "loss": 0.1396, + "step": 16466 + }, + { + "epoch": 0.38601241231385924, + "grad_norm": 0.19516393542289734, + "learning_rate": 0.00013509146896802327, + "loss": 0.0649, + "step": 16467 + }, + { + "epoch": 0.38603585388866424, + "grad_norm": 0.5434988141059875, + "learning_rate": 0.00013508457114787977, + "loss": 0.1581, + "step": 16468 + }, + { + "epoch": 0.38605929546346923, + "grad_norm": 0.1940571367740631, + "learning_rate": 0.0001350776731373664, + "loss": 0.0364, + "step": 16469 + }, + { + "epoch": 0.3860827370382742, + "grad_norm": 0.41104161739349365, + "learning_rate": 0.00013507077493652057, + "loss": 0.0573, + "step": 16470 + }, + { + "epoch": 0.3861061786130792, + "grad_norm": 0.8014063239097595, + "learning_rate": 0.00013506387654537973, + "loss": 0.1085, + "step": 16471 + }, + { + "epoch": 0.3861296201878842, + "grad_norm": 0.6144019365310669, + "learning_rate": 0.00013505697796398132, + "loss": 0.1211, + "step": 16472 + }, + { + "epoch": 0.3861530617626892, + "grad_norm": 0.5756582617759705, + "learning_rate": 0.00013505007919236277, + "loss": 0.1525, + "step": 16473 + }, + { + "epoch": 0.3861765033374942, + "grad_norm": 0.30346351861953735, + "learning_rate": 0.00013504318023056151, + "loss": 0.0601, + "step": 16474 + }, + { + "epoch": 0.3861999449122992, + "grad_norm": 0.297690212726593, + "learning_rate": 0.00013503628107861495, + "loss": 0.0384, + "step": 16475 + }, + { + "epoch": 0.3862233864871042, + "grad_norm": 0.21295423805713654, + "learning_rate": 0.00013502938173656055, + "loss": 0.046, + "step": 16476 + }, + { + "epoch": 0.3862468280619092, + "grad_norm": 0.4421113431453705, + "learning_rate": 0.00013502248220443575, + "loss": 0.0602, + "step": 16477 + }, + { + "epoch": 0.3862702696367142, + "grad_norm": 0.342732310295105, + "learning_rate": 0.00013501558248227795, + "loss": 0.0488, + "step": 16478 + }, + { + "epoch": 0.3862937112115192, + "grad_norm": 0.8376913666725159, + "learning_rate": 0.00013500868257012465, + "loss": 0.1552, + "step": 16479 + }, + { + "epoch": 0.38631715278632417, + "grad_norm": 0.6902309060096741, + "learning_rate": 0.00013500178246801322, + "loss": 0.1381, + "step": 16480 + }, + { + "epoch": 0.38634059436112916, + "grad_norm": 0.3498360216617584, + "learning_rate": 0.00013499488217598113, + "loss": 0.0445, + "step": 16481 + }, + { + "epoch": 0.38636403593593416, + "grad_norm": 1.0183829069137573, + "learning_rate": 0.00013498798169406587, + "loss": 0.0739, + "step": 16482 + }, + { + "epoch": 0.38638747751073915, + "grad_norm": 0.6785997152328491, + "learning_rate": 0.0001349810810223048, + "loss": 0.0737, + "step": 16483 + }, + { + "epoch": 0.38641091908554415, + "grad_norm": 0.5422806739807129, + "learning_rate": 0.00013497418016073542, + "loss": 0.1161, + "step": 16484 + }, + { + "epoch": 0.38643436066034914, + "grad_norm": 0.8782721757888794, + "learning_rate": 0.00013496727910939515, + "loss": 0.1813, + "step": 16485 + }, + { + "epoch": 0.38645780223515414, + "grad_norm": 0.21938320994377136, + "learning_rate": 0.00013496037786832143, + "loss": 0.0301, + "step": 16486 + }, + { + "epoch": 0.38648124380995913, + "grad_norm": 0.19223998486995697, + "learning_rate": 0.00013495347643755173, + "loss": 0.0428, + "step": 16487 + }, + { + "epoch": 0.3865046853847641, + "grad_norm": 0.27978143095970154, + "learning_rate": 0.00013494657481712347, + "loss": 0.0521, + "step": 16488 + }, + { + "epoch": 0.3865281269595691, + "grad_norm": 0.5119637250900269, + "learning_rate": 0.00013493967300707413, + "loss": 0.0316, + "step": 16489 + }, + { + "epoch": 0.3865515685343741, + "grad_norm": 0.1564233899116516, + "learning_rate": 0.00013493277100744112, + "loss": 0.0295, + "step": 16490 + }, + { + "epoch": 0.38657501010917916, + "grad_norm": 0.25262781977653503, + "learning_rate": 0.00013492586881826195, + "loss": 0.051, + "step": 16491 + }, + { + "epoch": 0.38659845168398416, + "grad_norm": 0.3699367642402649, + "learning_rate": 0.000134918966439574, + "loss": 0.0762, + "step": 16492 + }, + { + "epoch": 0.38662189325878915, + "grad_norm": 0.5846380591392517, + "learning_rate": 0.00013491206387141475, + "loss": 0.0858, + "step": 16493 + }, + { + "epoch": 0.38664533483359415, + "grad_norm": 0.9213314056396484, + "learning_rate": 0.0001349051611138217, + "loss": 0.2847, + "step": 16494 + }, + { + "epoch": 0.38666877640839914, + "grad_norm": 0.49337053298950195, + "learning_rate": 0.00013489825816683223, + "loss": 0.0718, + "step": 16495 + }, + { + "epoch": 0.38669221798320413, + "grad_norm": 0.2791057229042053, + "learning_rate": 0.00013489135503048382, + "loss": 0.0404, + "step": 16496 + }, + { + "epoch": 0.38671565955800913, + "grad_norm": 0.1560867875814438, + "learning_rate": 0.00013488445170481398, + "loss": 0.0538, + "step": 16497 + }, + { + "epoch": 0.3867391011328141, + "grad_norm": 0.596610426902771, + "learning_rate": 0.0001348775481898601, + "loss": 0.119, + "step": 16498 + }, + { + "epoch": 0.3867625427076191, + "grad_norm": 0.5283240079879761, + "learning_rate": 0.00013487064448565965, + "loss": 0.6615, + "step": 16499 + }, + { + "epoch": 0.3867859842824241, + "grad_norm": 0.510630190372467, + "learning_rate": 0.0001348637405922501, + "loss": 0.143, + "step": 16500 + }, + { + "epoch": 0.3868094258572291, + "grad_norm": 0.24327635765075684, + "learning_rate": 0.00013485683650966897, + "loss": 0.0209, + "step": 16501 + }, + { + "epoch": 0.3868328674320341, + "grad_norm": 0.5154907703399658, + "learning_rate": 0.0001348499322379536, + "loss": 0.1047, + "step": 16502 + }, + { + "epoch": 0.3868563090068391, + "grad_norm": 0.11989247798919678, + "learning_rate": 0.00013484302777714156, + "loss": 0.0311, + "step": 16503 + }, + { + "epoch": 0.3868797505816441, + "grad_norm": 0.12713971734046936, + "learning_rate": 0.00013483612312727026, + "loss": 0.0283, + "step": 16504 + }, + { + "epoch": 0.3869031921564491, + "grad_norm": 0.5972927808761597, + "learning_rate": 0.00013482921828837717, + "loss": 0.1142, + "step": 16505 + }, + { + "epoch": 0.3869266337312541, + "grad_norm": 0.485166996717453, + "learning_rate": 0.00013482231326049973, + "loss": 0.0783, + "step": 16506 + }, + { + "epoch": 0.38695007530605907, + "grad_norm": 0.6248679757118225, + "learning_rate": 0.00013481540804367547, + "loss": 0.0805, + "step": 16507 + }, + { + "epoch": 0.38697351688086407, + "grad_norm": 0.34837546944618225, + "learning_rate": 0.00013480850263794183, + "loss": 0.4156, + "step": 16508 + }, + { + "epoch": 0.38699695845566906, + "grad_norm": 0.11475686728954315, + "learning_rate": 0.00013480159704333628, + "loss": 0.0303, + "step": 16509 + }, + { + "epoch": 0.38702040003047405, + "grad_norm": 0.3915967047214508, + "learning_rate": 0.0001347946912598963, + "loss": 0.071, + "step": 16510 + }, + { + "epoch": 0.38704384160527905, + "grad_norm": 0.4254071116447449, + "learning_rate": 0.00013478778528765932, + "loss": 0.7402, + "step": 16511 + }, + { + "epoch": 0.38706728318008404, + "grad_norm": 0.45971739292144775, + "learning_rate": 0.00013478087912666283, + "loss": 0.0665, + "step": 16512 + }, + { + "epoch": 0.38709072475488904, + "grad_norm": 0.15922722220420837, + "learning_rate": 0.00013477397277694432, + "loss": 0.0212, + "step": 16513 + }, + { + "epoch": 0.38711416632969403, + "grad_norm": 0.15469205379486084, + "learning_rate": 0.00013476706623854128, + "loss": 0.0244, + "step": 16514 + }, + { + "epoch": 0.387137607904499, + "grad_norm": 0.5466366410255432, + "learning_rate": 0.00013476015951149114, + "loss": 0.1297, + "step": 16515 + }, + { + "epoch": 0.387161049479304, + "grad_norm": 0.2830270826816559, + "learning_rate": 0.0001347532525958314, + "loss": 0.0433, + "step": 16516 + }, + { + "epoch": 0.387184491054109, + "grad_norm": 0.5273315906524658, + "learning_rate": 0.00013474634549159956, + "loss": 0.0683, + "step": 16517 + }, + { + "epoch": 0.387207932628914, + "grad_norm": 0.20126210153102875, + "learning_rate": 0.00013473943819883303, + "loss": 0.0374, + "step": 16518 + }, + { + "epoch": 0.387231374203719, + "grad_norm": 0.734315037727356, + "learning_rate": 0.00013473253071756935, + "loss": 0.1732, + "step": 16519 + }, + { + "epoch": 0.387254815778524, + "grad_norm": 0.49270007014274597, + "learning_rate": 0.000134725623047846, + "loss": 0.1036, + "step": 16520 + }, + { + "epoch": 0.387278257353329, + "grad_norm": 1.1614022254943848, + "learning_rate": 0.0001347187151897004, + "loss": 0.1881, + "step": 16521 + }, + { + "epoch": 0.387301698928134, + "grad_norm": 0.7045661211013794, + "learning_rate": 0.00013471180714317012, + "loss": 0.8597, + "step": 16522 + }, + { + "epoch": 0.387325140502939, + "grad_norm": 0.3378550112247467, + "learning_rate": 0.00013470489890829259, + "loss": 0.072, + "step": 16523 + }, + { + "epoch": 0.387348582077744, + "grad_norm": 0.5383263826370239, + "learning_rate": 0.00013469799048510529, + "loss": 0.0887, + "step": 16524 + }, + { + "epoch": 0.38737202365254897, + "grad_norm": 1.0075291395187378, + "learning_rate": 0.0001346910818736457, + "loss": 0.2725, + "step": 16525 + }, + { + "epoch": 0.38739546522735396, + "grad_norm": 0.44340983033180237, + "learning_rate": 0.00013468417307395134, + "loss": 0.0902, + "step": 16526 + }, + { + "epoch": 0.38741890680215896, + "grad_norm": 0.19722433388233185, + "learning_rate": 0.00013467726408605967, + "loss": 0.045, + "step": 16527 + }, + { + "epoch": 0.38744234837696395, + "grad_norm": 0.6641840934753418, + "learning_rate": 0.0001346703549100082, + "loss": 0.1382, + "step": 16528 + }, + { + "epoch": 0.38746578995176895, + "grad_norm": 0.46792325377464294, + "learning_rate": 0.00013466344554583442, + "loss": 0.0929, + "step": 16529 + }, + { + "epoch": 0.38748923152657394, + "grad_norm": 0.3636719584465027, + "learning_rate": 0.0001346565359935758, + "loss": 0.0799, + "step": 16530 + }, + { + "epoch": 0.38751267310137893, + "grad_norm": 0.387751966714859, + "learning_rate": 0.00013464962625326984, + "loss": 0.0521, + "step": 16531 + }, + { + "epoch": 0.38753611467618393, + "grad_norm": 0.5395868420600891, + "learning_rate": 0.00013464271632495405, + "loss": 0.1315, + "step": 16532 + }, + { + "epoch": 0.3875595562509889, + "grad_norm": 0.15964645147323608, + "learning_rate": 0.0001346358062086659, + "loss": 0.0374, + "step": 16533 + }, + { + "epoch": 0.3875829978257939, + "grad_norm": 0.16776630282402039, + "learning_rate": 0.0001346288959044429, + "loss": 0.03, + "step": 16534 + }, + { + "epoch": 0.3876064394005989, + "grad_norm": 0.4725794196128845, + "learning_rate": 0.00013462198541232254, + "loss": 0.1133, + "step": 16535 + }, + { + "epoch": 0.3876298809754039, + "grad_norm": 0.22387713193893433, + "learning_rate": 0.00013461507473234226, + "loss": 0.0402, + "step": 16536 + }, + { + "epoch": 0.3876533225502089, + "grad_norm": 0.5569259524345398, + "learning_rate": 0.00013460816386453965, + "loss": 0.1191, + "step": 16537 + }, + { + "epoch": 0.3876767641250139, + "grad_norm": 0.5787074565887451, + "learning_rate": 0.0001346012528089522, + "loss": 0.6669, + "step": 16538 + }, + { + "epoch": 0.3877002056998189, + "grad_norm": 0.763884425163269, + "learning_rate": 0.00013459434156561735, + "loss": 0.2369, + "step": 16539 + }, + { + "epoch": 0.3877236472746239, + "grad_norm": 0.4988492429256439, + "learning_rate": 0.00013458743013457263, + "loss": 0.0822, + "step": 16540 + }, + { + "epoch": 0.3877470888494289, + "grad_norm": 0.3879390060901642, + "learning_rate": 0.00013458051851585553, + "loss": 0.0916, + "step": 16541 + }, + { + "epoch": 0.3877705304242339, + "grad_norm": 0.1320389360189438, + "learning_rate": 0.00013457360670950357, + "loss": 0.0148, + "step": 16542 + }, + { + "epoch": 0.3877939719990389, + "grad_norm": 0.5361616611480713, + "learning_rate": 0.00013456669471555426, + "loss": 0.1002, + "step": 16543 + }, + { + "epoch": 0.3878174135738439, + "grad_norm": 0.23791712522506714, + "learning_rate": 0.00013455978253404507, + "loss": 0.0483, + "step": 16544 + }, + { + "epoch": 0.3878408551486489, + "grad_norm": 0.369855672121048, + "learning_rate": 0.00013455287016501358, + "loss": 0.0528, + "step": 16545 + }, + { + "epoch": 0.3878642967234539, + "grad_norm": 0.5819802284240723, + "learning_rate": 0.00013454595760849718, + "loss": 0.7622, + "step": 16546 + }, + { + "epoch": 0.3878877382982589, + "grad_norm": 0.46562016010284424, + "learning_rate": 0.00013453904486453348, + "loss": 0.0916, + "step": 16547 + }, + { + "epoch": 0.3879111798730639, + "grad_norm": 0.5329784154891968, + "learning_rate": 0.00013453213193315993, + "loss": 0.4501, + "step": 16548 + }, + { + "epoch": 0.3879346214478689, + "grad_norm": 0.23198503255844116, + "learning_rate": 0.0001345252188144141, + "loss": 0.0696, + "step": 16549 + }, + { + "epoch": 0.3879580630226739, + "grad_norm": 0.49878713488578796, + "learning_rate": 0.0001345183055083334, + "loss": 0.063, + "step": 16550 + }, + { + "epoch": 0.3879815045974789, + "grad_norm": 0.16264715790748596, + "learning_rate": 0.00013451139201495546, + "loss": 0.0275, + "step": 16551 + }, + { + "epoch": 0.38800494617228387, + "grad_norm": 0.4714091420173645, + "learning_rate": 0.0001345044783343177, + "loss": 0.0904, + "step": 16552 + }, + { + "epoch": 0.38802838774708887, + "grad_norm": 0.6999625563621521, + "learning_rate": 0.0001344975644664577, + "loss": 0.9112, + "step": 16553 + }, + { + "epoch": 0.38805182932189386, + "grad_norm": 0.8613266944885254, + "learning_rate": 0.0001344906504114129, + "loss": 0.1734, + "step": 16554 + }, + { + "epoch": 0.38807527089669885, + "grad_norm": 0.6107471585273743, + "learning_rate": 0.00013448373616922093, + "loss": 0.6592, + "step": 16555 + }, + { + "epoch": 0.38809871247150385, + "grad_norm": 0.4788925051689148, + "learning_rate": 0.00013447682173991918, + "loss": 0.3965, + "step": 16556 + }, + { + "epoch": 0.38812215404630884, + "grad_norm": 0.5930721163749695, + "learning_rate": 0.00013446990712354527, + "loss": 0.1028, + "step": 16557 + }, + { + "epoch": 0.38814559562111384, + "grad_norm": 0.3445964753627777, + "learning_rate": 0.00013446299232013667, + "loss": 0.0411, + "step": 16558 + }, + { + "epoch": 0.38816903719591883, + "grad_norm": 0.5266459584236145, + "learning_rate": 0.00013445607732973084, + "loss": 0.079, + "step": 16559 + }, + { + "epoch": 0.3881924787707238, + "grad_norm": 0.15039992332458496, + "learning_rate": 0.00013444916215236543, + "loss": 0.0285, + "step": 16560 + }, + { + "epoch": 0.3882159203455288, + "grad_norm": 0.5467488169670105, + "learning_rate": 0.00013444224678807787, + "loss": 0.1247, + "step": 16561 + }, + { + "epoch": 0.3882393619203338, + "grad_norm": 0.8139004111289978, + "learning_rate": 0.00013443533123690574, + "loss": 0.2031, + "step": 16562 + }, + { + "epoch": 0.3882628034951388, + "grad_norm": 0.33846762776374817, + "learning_rate": 0.00013442841549888648, + "loss": 0.0621, + "step": 16563 + }, + { + "epoch": 0.3882862450699438, + "grad_norm": 0.5687698721885681, + "learning_rate": 0.00013442149957405773, + "loss": 0.1264, + "step": 16564 + }, + { + "epoch": 0.3883096866447488, + "grad_norm": 0.22361236810684204, + "learning_rate": 0.00013441458346245692, + "loss": 0.0301, + "step": 16565 + }, + { + "epoch": 0.3883331282195538, + "grad_norm": 0.5252013206481934, + "learning_rate": 0.0001344076671641216, + "loss": 0.1082, + "step": 16566 + }, + { + "epoch": 0.3883565697943588, + "grad_norm": 0.5643205046653748, + "learning_rate": 0.00013440075067908935, + "loss": 0.0881, + "step": 16567 + }, + { + "epoch": 0.3883800113691638, + "grad_norm": 0.20551292598247528, + "learning_rate": 0.00013439383400739766, + "loss": 0.0254, + "step": 16568 + }, + { + "epoch": 0.3884034529439688, + "grad_norm": 0.1537168323993683, + "learning_rate": 0.000134386917149084, + "loss": 0.0425, + "step": 16569 + }, + { + "epoch": 0.38842689451877377, + "grad_norm": 0.4028581976890564, + "learning_rate": 0.00013438000010418603, + "loss": 0.0774, + "step": 16570 + }, + { + "epoch": 0.38845033609357876, + "grad_norm": 0.35510507225990295, + "learning_rate": 0.0001343730828727412, + "loss": 0.0901, + "step": 16571 + }, + { + "epoch": 0.38847377766838376, + "grad_norm": 0.6499537229537964, + "learning_rate": 0.00013436616545478702, + "loss": 0.6504, + "step": 16572 + }, + { + "epoch": 0.38849721924318875, + "grad_norm": 0.5839397311210632, + "learning_rate": 0.00013435924785036108, + "loss": 0.6148, + "step": 16573 + }, + { + "epoch": 0.38852066081799375, + "grad_norm": 0.4693005681037903, + "learning_rate": 0.0001343523300595009, + "loss": 0.1112, + "step": 16574 + }, + { + "epoch": 0.38854410239279874, + "grad_norm": 0.4757440388202667, + "learning_rate": 0.000134345412082244, + "loss": 0.5771, + "step": 16575 + }, + { + "epoch": 0.38856754396760373, + "grad_norm": 0.47669336199760437, + "learning_rate": 0.00013433849391862794, + "loss": 0.1522, + "step": 16576 + }, + { + "epoch": 0.38859098554240873, + "grad_norm": 0.44482991099357605, + "learning_rate": 0.00013433157556869024, + "loss": 0.0947, + "step": 16577 + }, + { + "epoch": 0.3886144271172137, + "grad_norm": 0.3538377583026886, + "learning_rate": 0.00013432465703246845, + "loss": 0.0431, + "step": 16578 + }, + { + "epoch": 0.3886378686920187, + "grad_norm": 0.2561487555503845, + "learning_rate": 0.0001343177383100001, + "loss": 0.0714, + "step": 16579 + }, + { + "epoch": 0.3886613102668237, + "grad_norm": 0.36052563786506653, + "learning_rate": 0.00013431081940132276, + "loss": 0.0738, + "step": 16580 + }, + { + "epoch": 0.3886847518416287, + "grad_norm": 0.5374552607536316, + "learning_rate": 0.00013430390030647394, + "loss": 0.7851, + "step": 16581 + }, + { + "epoch": 0.3887081934164337, + "grad_norm": 0.15094690024852753, + "learning_rate": 0.00013429698102549115, + "loss": 0.0292, + "step": 16582 + }, + { + "epoch": 0.3887316349912387, + "grad_norm": 0.40724751353263855, + "learning_rate": 0.000134290061558412, + "loss": 0.0411, + "step": 16583 + }, + { + "epoch": 0.3887550765660437, + "grad_norm": 0.6247356534004211, + "learning_rate": 0.00013428314190527403, + "loss": 0.1539, + "step": 16584 + }, + { + "epoch": 0.3887785181408487, + "grad_norm": 0.619898796081543, + "learning_rate": 0.00013427622206611475, + "loss": 0.1233, + "step": 16585 + }, + { + "epoch": 0.3888019597156537, + "grad_norm": 0.25735747814178467, + "learning_rate": 0.00013426930204097175, + "loss": 0.0683, + "step": 16586 + }, + { + "epoch": 0.38882540129045867, + "grad_norm": 0.5166913270950317, + "learning_rate": 0.00013426238182988253, + "loss": 0.1296, + "step": 16587 + }, + { + "epoch": 0.38884884286526367, + "grad_norm": 0.37822169065475464, + "learning_rate": 0.00013425546143288467, + "loss": 0.0607, + "step": 16588 + }, + { + "epoch": 0.38887228444006866, + "grad_norm": 0.17544135451316833, + "learning_rate": 0.0001342485408500157, + "loss": 0.0401, + "step": 16589 + }, + { + "epoch": 0.38889572601487365, + "grad_norm": 0.49200722575187683, + "learning_rate": 0.0001342416200813132, + "loss": 0.0789, + "step": 16590 + }, + { + "epoch": 0.38891916758967865, + "grad_norm": 0.2330198884010315, + "learning_rate": 0.0001342346991268147, + "loss": 0.0509, + "step": 16591 + }, + { + "epoch": 0.38894260916448364, + "grad_norm": 0.9631819128990173, + "learning_rate": 0.00013422777798655777, + "loss": 0.2144, + "step": 16592 + }, + { + "epoch": 0.38896605073928864, + "grad_norm": 0.9130639433860779, + "learning_rate": 0.00013422085666057996, + "loss": 0.1464, + "step": 16593 + }, + { + "epoch": 0.3889894923140937, + "grad_norm": 0.43153443932533264, + "learning_rate": 0.0001342139351489188, + "loss": 0.537, + "step": 16594 + }, + { + "epoch": 0.3890129338888987, + "grad_norm": 0.5199502110481262, + "learning_rate": 0.00013420701345161188, + "loss": 0.0708, + "step": 16595 + }, + { + "epoch": 0.3890363754637037, + "grad_norm": 0.21436254680156708, + "learning_rate": 0.00013420009156869675, + "loss": 0.034, + "step": 16596 + }, + { + "epoch": 0.38905981703850867, + "grad_norm": 0.6045315861701965, + "learning_rate": 0.00013419316950021094, + "loss": 0.0869, + "step": 16597 + }, + { + "epoch": 0.38908325861331367, + "grad_norm": 0.42786064743995667, + "learning_rate": 0.00013418624724619203, + "loss": 0.0703, + "step": 16598 + }, + { + "epoch": 0.38910670018811866, + "grad_norm": 0.16124270856380463, + "learning_rate": 0.0001341793248066776, + "loss": 0.0327, + "step": 16599 + }, + { + "epoch": 0.38913014176292365, + "grad_norm": 0.6462946534156799, + "learning_rate": 0.0001341724021817052, + "loss": 0.6204, + "step": 16600 + }, + { + "epoch": 0.38915358333772865, + "grad_norm": 0.5136579871177673, + "learning_rate": 0.00013416547937131234, + "loss": 0.1147, + "step": 16601 + }, + { + "epoch": 0.38917702491253364, + "grad_norm": 0.2606164515018463, + "learning_rate": 0.00013415855637553668, + "loss": 0.0617, + "step": 16602 + }, + { + "epoch": 0.38920046648733864, + "grad_norm": 0.09324970841407776, + "learning_rate": 0.00013415163319441568, + "loss": 0.0111, + "step": 16603 + }, + { + "epoch": 0.38922390806214363, + "grad_norm": 0.12911953032016754, + "learning_rate": 0.00013414470982798697, + "loss": 0.017, + "step": 16604 + }, + { + "epoch": 0.3892473496369486, + "grad_norm": 0.8486754894256592, + "learning_rate": 0.00013413778627628816, + "loss": 0.1374, + "step": 16605 + }, + { + "epoch": 0.3892707912117536, + "grad_norm": 0.1502382606267929, + "learning_rate": 0.00013413086253935672, + "loss": 0.0206, + "step": 16606 + }, + { + "epoch": 0.3892942327865586, + "grad_norm": 0.4228001832962036, + "learning_rate": 0.00013412393861723026, + "loss": 0.0953, + "step": 16607 + }, + { + "epoch": 0.3893176743613636, + "grad_norm": 0.1327735334634781, + "learning_rate": 0.00013411701450994636, + "loss": 0.0176, + "step": 16608 + }, + { + "epoch": 0.3893411159361686, + "grad_norm": 0.21300047636032104, + "learning_rate": 0.00013411009021754256, + "loss": 0.0413, + "step": 16609 + }, + { + "epoch": 0.3893645575109736, + "grad_norm": 0.8670732378959656, + "learning_rate": 0.00013410316574005644, + "loss": 0.1151, + "step": 16610 + }, + { + "epoch": 0.3893879990857786, + "grad_norm": 0.508267879486084, + "learning_rate": 0.00013409624107752564, + "loss": 0.1244, + "step": 16611 + }, + { + "epoch": 0.3894114406605836, + "grad_norm": 0.1449059098958969, + "learning_rate": 0.00013408931622998764, + "loss": 0.0212, + "step": 16612 + }, + { + "epoch": 0.3894348822353886, + "grad_norm": 0.45502644777297974, + "learning_rate": 0.00013408239119748006, + "loss": 0.1365, + "step": 16613 + }, + { + "epoch": 0.3894583238101936, + "grad_norm": 0.2178552895784378, + "learning_rate": 0.00013407546598004046, + "loss": 0.0703, + "step": 16614 + }, + { + "epoch": 0.38948176538499857, + "grad_norm": 0.3170715570449829, + "learning_rate": 0.00013406854057770642, + "loss": 0.0291, + "step": 16615 + }, + { + "epoch": 0.38950520695980356, + "grad_norm": 0.43520551919937134, + "learning_rate": 0.00013406161499051553, + "loss": 0.1207, + "step": 16616 + }, + { + "epoch": 0.38952864853460856, + "grad_norm": 0.11911958456039429, + "learning_rate": 0.00013405468921850533, + "loss": 0.0299, + "step": 16617 + }, + { + "epoch": 0.38955209010941355, + "grad_norm": 0.20938384532928467, + "learning_rate": 0.00013404776326171348, + "loss": 0.0554, + "step": 16618 + }, + { + "epoch": 0.38957553168421855, + "grad_norm": 0.22845831513404846, + "learning_rate": 0.0001340408371201775, + "loss": 0.0645, + "step": 16619 + }, + { + "epoch": 0.38959897325902354, + "grad_norm": 0.6658967733383179, + "learning_rate": 0.00013403391079393494, + "loss": 0.0859, + "step": 16620 + }, + { + "epoch": 0.38962241483382853, + "grad_norm": 0.40289729833602905, + "learning_rate": 0.00013402698428302345, + "loss": 0.4182, + "step": 16621 + }, + { + "epoch": 0.38964585640863353, + "grad_norm": 0.8430724740028381, + "learning_rate": 0.0001340200575874806, + "loss": 0.7604, + "step": 16622 + }, + { + "epoch": 0.3896692979834385, + "grad_norm": 0.7718609571456909, + "learning_rate": 0.00013401313070734395, + "loss": 0.1229, + "step": 16623 + }, + { + "epoch": 0.3896927395582435, + "grad_norm": 0.28404250741004944, + "learning_rate": 0.00013400620364265107, + "loss": 0.051, + "step": 16624 + }, + { + "epoch": 0.3897161811330485, + "grad_norm": 0.41791966557502747, + "learning_rate": 0.00013399927639343965, + "loss": 0.0395, + "step": 16625 + }, + { + "epoch": 0.3897396227078535, + "grad_norm": 0.8094483017921448, + "learning_rate": 0.0001339923489597471, + "loss": 0.239, + "step": 16626 + }, + { + "epoch": 0.3897630642826585, + "grad_norm": 0.5548243522644043, + "learning_rate": 0.00013398542134161114, + "loss": 0.6746, + "step": 16627 + }, + { + "epoch": 0.3897865058574635, + "grad_norm": 0.6931949257850647, + "learning_rate": 0.00013397849353906938, + "loss": 0.1256, + "step": 16628 + }, + { + "epoch": 0.3898099474322685, + "grad_norm": 0.5065948963165283, + "learning_rate": 0.0001339715655521593, + "loss": 0.1045, + "step": 16629 + }, + { + "epoch": 0.3898333890070735, + "grad_norm": 0.42153215408325195, + "learning_rate": 0.00013396463738091859, + "loss": 0.0747, + "step": 16630 + }, + { + "epoch": 0.3898568305818785, + "grad_norm": 0.709678590297699, + "learning_rate": 0.00013395770902538477, + "loss": 0.1531, + "step": 16631 + }, + { + "epoch": 0.38988027215668347, + "grad_norm": 0.42278316617012024, + "learning_rate": 0.00013395078048559548, + "loss": 0.0785, + "step": 16632 + }, + { + "epoch": 0.38990371373148847, + "grad_norm": 0.5500323176383972, + "learning_rate": 0.0001339438517615883, + "loss": 0.1631, + "step": 16633 + }, + { + "epoch": 0.38992715530629346, + "grad_norm": 0.4807800352573395, + "learning_rate": 0.00013393692285340082, + "loss": 0.1324, + "step": 16634 + }, + { + "epoch": 0.38995059688109845, + "grad_norm": 0.42815643548965454, + "learning_rate": 0.00013392999376107066, + "loss": 0.189, + "step": 16635 + }, + { + "epoch": 0.38997403845590345, + "grad_norm": 0.36719560623168945, + "learning_rate": 0.0001339230644846354, + "loss": 0.0601, + "step": 16636 + }, + { + "epoch": 0.38999748003070844, + "grad_norm": 0.7670131325721741, + "learning_rate": 0.00013391613502413262, + "loss": 0.1548, + "step": 16637 + }, + { + "epoch": 0.39002092160551344, + "grad_norm": 0.24475987255573273, + "learning_rate": 0.00013390920537959997, + "loss": 0.0559, + "step": 16638 + }, + { + "epoch": 0.39004436318031843, + "grad_norm": 0.7515246868133545, + "learning_rate": 0.00013390227555107494, + "loss": 0.1747, + "step": 16639 + }, + { + "epoch": 0.3900678047551234, + "grad_norm": 0.44466373324394226, + "learning_rate": 0.00013389534553859527, + "loss": 0.0809, + "step": 16640 + }, + { + "epoch": 0.3900912463299284, + "grad_norm": 0.8112598061561584, + "learning_rate": 0.00013388841534219852, + "loss": 0.1213, + "step": 16641 + }, + { + "epoch": 0.3901146879047334, + "grad_norm": 0.8840324282646179, + "learning_rate": 0.00013388148496192224, + "loss": 0.122, + "step": 16642 + }, + { + "epoch": 0.3901381294795384, + "grad_norm": 0.7324212789535522, + "learning_rate": 0.00013387455439780409, + "loss": 0.2374, + "step": 16643 + }, + { + "epoch": 0.3901615710543434, + "grad_norm": 0.5789515376091003, + "learning_rate": 0.00013386762364988165, + "loss": 0.1681, + "step": 16644 + }, + { + "epoch": 0.39018501262914845, + "grad_norm": 0.48264405131340027, + "learning_rate": 0.0001338606927181925, + "loss": 0.061, + "step": 16645 + }, + { + "epoch": 0.39020845420395345, + "grad_norm": 0.40587544441223145, + "learning_rate": 0.00013385376160277433, + "loss": 0.1157, + "step": 16646 + }, + { + "epoch": 0.39023189577875844, + "grad_norm": 0.4978112578392029, + "learning_rate": 0.00013384683030366468, + "loss": 0.1256, + "step": 16647 + }, + { + "epoch": 0.39025533735356344, + "grad_norm": 0.7831295728683472, + "learning_rate": 0.00013383989882090113, + "loss": 0.1293, + "step": 16648 + }, + { + "epoch": 0.39027877892836843, + "grad_norm": 0.13050702214241028, + "learning_rate": 0.00013383296715452142, + "loss": 0.0286, + "step": 16649 + }, + { + "epoch": 0.3903022205031734, + "grad_norm": 0.046615052968263626, + "learning_rate": 0.00013382603530456304, + "loss": 0.0044, + "step": 16650 + }, + { + "epoch": 0.3903256620779784, + "grad_norm": 0.44168829917907715, + "learning_rate": 0.00013381910327106363, + "loss": 0.103, + "step": 16651 + }, + { + "epoch": 0.3903491036527834, + "grad_norm": 0.39527493715286255, + "learning_rate": 0.0001338121710540608, + "loss": 0.1024, + "step": 16652 + }, + { + "epoch": 0.3903725452275884, + "grad_norm": 0.7304633259773254, + "learning_rate": 0.00013380523865359222, + "loss": 0.208, + "step": 16653 + }, + { + "epoch": 0.3903959868023934, + "grad_norm": 0.391464501619339, + "learning_rate": 0.00013379830606969546, + "loss": 0.0996, + "step": 16654 + }, + { + "epoch": 0.3904194283771984, + "grad_norm": 0.4376797676086426, + "learning_rate": 0.00013379137330240813, + "loss": 0.0392, + "step": 16655 + }, + { + "epoch": 0.3904428699520034, + "grad_norm": 0.43801000714302063, + "learning_rate": 0.00013378444035176785, + "loss": 0.0717, + "step": 16656 + }, + { + "epoch": 0.3904663115268084, + "grad_norm": 0.2243155539035797, + "learning_rate": 0.00013377750721781223, + "loss": 0.0523, + "step": 16657 + }, + { + "epoch": 0.3904897531016134, + "grad_norm": 0.37026917934417725, + "learning_rate": 0.00013377057390057893, + "loss": 0.0773, + "step": 16658 + }, + { + "epoch": 0.3905131946764184, + "grad_norm": 0.32923656702041626, + "learning_rate": 0.00013376364040010557, + "loss": 0.0746, + "step": 16659 + }, + { + "epoch": 0.39053663625122337, + "grad_norm": 0.5226410627365112, + "learning_rate": 0.00013375670671642972, + "loss": 0.1237, + "step": 16660 + }, + { + "epoch": 0.39056007782602836, + "grad_norm": 0.3883775472640991, + "learning_rate": 0.00013374977284958902, + "loss": 0.0981, + "step": 16661 + }, + { + "epoch": 0.39058351940083336, + "grad_norm": 0.41059601306915283, + "learning_rate": 0.00013374283879962113, + "loss": 0.122, + "step": 16662 + }, + { + "epoch": 0.39060696097563835, + "grad_norm": 0.8769006729125977, + "learning_rate": 0.00013373590456656364, + "loss": 0.141, + "step": 16663 + }, + { + "epoch": 0.39063040255044335, + "grad_norm": 0.7006500959396362, + "learning_rate": 0.00013372897015045417, + "loss": 0.5643, + "step": 16664 + }, + { + "epoch": 0.39065384412524834, + "grad_norm": 0.21657581627368927, + "learning_rate": 0.00013372203555133037, + "loss": 0.0393, + "step": 16665 + }, + { + "epoch": 0.39067728570005333, + "grad_norm": 0.6539408564567566, + "learning_rate": 0.00013371510076922986, + "loss": 0.1252, + "step": 16666 + }, + { + "epoch": 0.39070072727485833, + "grad_norm": 0.14446988701820374, + "learning_rate": 0.00013370816580419024, + "loss": 0.0311, + "step": 16667 + }, + { + "epoch": 0.3907241688496633, + "grad_norm": 0.6508687138557434, + "learning_rate": 0.00013370123065624919, + "loss": 0.2363, + "step": 16668 + }, + { + "epoch": 0.3907476104244683, + "grad_norm": 0.4323688745498657, + "learning_rate": 0.00013369429532544434, + "loss": 0.1195, + "step": 16669 + }, + { + "epoch": 0.3907710519992733, + "grad_norm": 0.4709200859069824, + "learning_rate": 0.00013368735981181327, + "loss": 0.6674, + "step": 16670 + }, + { + "epoch": 0.3907944935740783, + "grad_norm": 0.15491101145744324, + "learning_rate": 0.0001336804241153936, + "loss": 0.0319, + "step": 16671 + }, + { + "epoch": 0.3908179351488833, + "grad_norm": 0.3132411539554596, + "learning_rate": 0.00013367348823622305, + "loss": 0.0686, + "step": 16672 + }, + { + "epoch": 0.3908413767236883, + "grad_norm": 0.1509804129600525, + "learning_rate": 0.00013366655217433922, + "loss": 0.0317, + "step": 16673 + }, + { + "epoch": 0.3908648182984933, + "grad_norm": 0.3968122601509094, + "learning_rate": 0.00013365961592977967, + "loss": 0.0569, + "step": 16674 + }, + { + "epoch": 0.3908882598732983, + "grad_norm": 0.29004448652267456, + "learning_rate": 0.00013365267950258213, + "loss": 0.0502, + "step": 16675 + }, + { + "epoch": 0.3909117014481033, + "grad_norm": 0.5278860330581665, + "learning_rate": 0.0001336457428927842, + "loss": 0.0769, + "step": 16676 + }, + { + "epoch": 0.39093514302290827, + "grad_norm": 0.5362539291381836, + "learning_rate": 0.00013363880610042355, + "loss": 0.0739, + "step": 16677 + }, + { + "epoch": 0.39095858459771327, + "grad_norm": 0.30354487895965576, + "learning_rate": 0.00013363186912553776, + "loss": 0.0606, + "step": 16678 + }, + { + "epoch": 0.39098202617251826, + "grad_norm": 0.5682803392410278, + "learning_rate": 0.00013362493196816453, + "loss": 0.1184, + "step": 16679 + }, + { + "epoch": 0.39100546774732325, + "grad_norm": 0.25917869806289673, + "learning_rate": 0.00013361799462834143, + "loss": 0.0385, + "step": 16680 + }, + { + "epoch": 0.39102890932212825, + "grad_norm": 0.5827093720436096, + "learning_rate": 0.0001336110571061062, + "loss": 0.6554, + "step": 16681 + }, + { + "epoch": 0.39105235089693324, + "grad_norm": 0.441326379776001, + "learning_rate": 0.0001336041194014964, + "loss": 0.0787, + "step": 16682 + }, + { + "epoch": 0.39107579247173824, + "grad_norm": 0.6202601790428162, + "learning_rate": 0.00013359718151454968, + "loss": 0.201, + "step": 16683 + }, + { + "epoch": 0.39109923404654323, + "grad_norm": 0.7305067777633667, + "learning_rate": 0.00013359024344530377, + "loss": 0.1759, + "step": 16684 + }, + { + "epoch": 0.3911226756213482, + "grad_norm": 1.070704698562622, + "learning_rate": 0.0001335833051937962, + "loss": 0.1945, + "step": 16685 + }, + { + "epoch": 0.3911461171961532, + "grad_norm": 0.14408785104751587, + "learning_rate": 0.0001335763667600647, + "loss": 0.0236, + "step": 16686 + }, + { + "epoch": 0.3911695587709582, + "grad_norm": 1.0628116130828857, + "learning_rate": 0.00013356942814414687, + "loss": 0.2152, + "step": 16687 + }, + { + "epoch": 0.3911930003457632, + "grad_norm": 0.18929129838943481, + "learning_rate": 0.0001335624893460804, + "loss": 0.0492, + "step": 16688 + }, + { + "epoch": 0.3912164419205682, + "grad_norm": 0.26639237999916077, + "learning_rate": 0.0001335555503659029, + "loss": 0.051, + "step": 16689 + }, + { + "epoch": 0.3912398834953732, + "grad_norm": 0.4318213164806366, + "learning_rate": 0.00013354861120365205, + "loss": 0.0809, + "step": 16690 + }, + { + "epoch": 0.3912633250701782, + "grad_norm": 0.14030039310455322, + "learning_rate": 0.0001335416718593655, + "loss": 0.0368, + "step": 16691 + }, + { + "epoch": 0.3912867666449832, + "grad_norm": 0.4743475914001465, + "learning_rate": 0.00013353473233308088, + "loss": 0.1056, + "step": 16692 + }, + { + "epoch": 0.3913102082197882, + "grad_norm": 0.3917073607444763, + "learning_rate": 0.00013352779262483587, + "loss": 0.0716, + "step": 16693 + }, + { + "epoch": 0.3913336497945932, + "grad_norm": 0.25372427701950073, + "learning_rate": 0.0001335208527346681, + "loss": 0.0863, + "step": 16694 + }, + { + "epoch": 0.39135709136939817, + "grad_norm": 0.5308656096458435, + "learning_rate": 0.00013351391266261525, + "loss": 0.0726, + "step": 16695 + }, + { + "epoch": 0.3913805329442032, + "grad_norm": 0.5577232241630554, + "learning_rate": 0.00013350697240871497, + "loss": 0.1031, + "step": 16696 + }, + { + "epoch": 0.3914039745190082, + "grad_norm": 0.13715726137161255, + "learning_rate": 0.0001335000319730049, + "loss": 0.0184, + "step": 16697 + }, + { + "epoch": 0.3914274160938132, + "grad_norm": 1.0737441778182983, + "learning_rate": 0.00013349309135552275, + "loss": 0.1064, + "step": 16698 + }, + { + "epoch": 0.3914508576686182, + "grad_norm": 0.40407422184944153, + "learning_rate": 0.0001334861505563061, + "loss": 0.0686, + "step": 16699 + }, + { + "epoch": 0.3914742992434232, + "grad_norm": 0.5388701558113098, + "learning_rate": 0.00013347920957539267, + "loss": 0.1076, + "step": 16700 + }, + { + "epoch": 0.3914977408182282, + "grad_norm": 0.2329699993133545, + "learning_rate": 0.00013347226841282014, + "loss": 0.0724, + "step": 16701 + }, + { + "epoch": 0.3915211823930332, + "grad_norm": 0.4106421172618866, + "learning_rate": 0.0001334653270686261, + "loss": 0.0829, + "step": 16702 + }, + { + "epoch": 0.3915446239678382, + "grad_norm": 0.6796783208847046, + "learning_rate": 0.00013345838554284828, + "loss": 0.122, + "step": 16703 + }, + { + "epoch": 0.3915680655426432, + "grad_norm": 0.2954498529434204, + "learning_rate": 0.00013345144383552433, + "loss": 0.0867, + "step": 16704 + }, + { + "epoch": 0.39159150711744817, + "grad_norm": 0.35258975625038147, + "learning_rate": 0.00013344450194669187, + "loss": 0.0478, + "step": 16705 + }, + { + "epoch": 0.39161494869225316, + "grad_norm": 0.6857312321662903, + "learning_rate": 0.0001334375598763886, + "loss": 0.1178, + "step": 16706 + }, + { + "epoch": 0.39163839026705816, + "grad_norm": 0.5421512722969055, + "learning_rate": 0.00013343061762465224, + "loss": 0.1525, + "step": 16707 + }, + { + "epoch": 0.39166183184186315, + "grad_norm": 0.6381113529205322, + "learning_rate": 0.00013342367519152037, + "loss": 0.1281, + "step": 16708 + }, + { + "epoch": 0.39168527341666814, + "grad_norm": 0.5142504572868347, + "learning_rate": 0.0001334167325770307, + "loss": 0.1268, + "step": 16709 + }, + { + "epoch": 0.39170871499147314, + "grad_norm": 1.067799687385559, + "learning_rate": 0.00013340978978122091, + "loss": 0.1295, + "step": 16710 + }, + { + "epoch": 0.39173215656627813, + "grad_norm": 0.15313729643821716, + "learning_rate": 0.00013340284680412866, + "loss": 0.0299, + "step": 16711 + }, + { + "epoch": 0.39175559814108313, + "grad_norm": 0.5059077143669128, + "learning_rate": 0.0001333959036457916, + "loss": 0.0802, + "step": 16712 + }, + { + "epoch": 0.3917790397158881, + "grad_norm": 0.6255230903625488, + "learning_rate": 0.00013338896030624746, + "loss": 0.1477, + "step": 16713 + }, + { + "epoch": 0.3918024812906931, + "grad_norm": 0.47189682722091675, + "learning_rate": 0.00013338201678553388, + "loss": 0.63, + "step": 16714 + }, + { + "epoch": 0.3918259228654981, + "grad_norm": 0.22034288942813873, + "learning_rate": 0.00013337507308368852, + "loss": 0.0448, + "step": 16715 + }, + { + "epoch": 0.3918493644403031, + "grad_norm": 0.7162479758262634, + "learning_rate": 0.00013336812920074909, + "loss": 0.1877, + "step": 16716 + }, + { + "epoch": 0.3918728060151081, + "grad_norm": 0.4657999277114868, + "learning_rate": 0.00013336118513675323, + "loss": 0.1543, + "step": 16717 + }, + { + "epoch": 0.3918962475899131, + "grad_norm": 0.07530631870031357, + "learning_rate": 0.00013335424089173866, + "loss": 0.015, + "step": 16718 + }, + { + "epoch": 0.3919196891647181, + "grad_norm": 0.6758183836936951, + "learning_rate": 0.00013334729646574304, + "loss": 0.1362, + "step": 16719 + }, + { + "epoch": 0.3919431307395231, + "grad_norm": 0.44739943742752075, + "learning_rate": 0.00013334035185880405, + "loss": 0.0932, + "step": 16720 + }, + { + "epoch": 0.3919665723143281, + "grad_norm": 0.5156533718109131, + "learning_rate": 0.00013333340707095937, + "loss": 0.1059, + "step": 16721 + }, + { + "epoch": 0.39199001388913307, + "grad_norm": 0.2697605788707733, + "learning_rate": 0.0001333264621022467, + "loss": 0.0616, + "step": 16722 + }, + { + "epoch": 0.39201345546393807, + "grad_norm": 0.15865325927734375, + "learning_rate": 0.0001333195169527037, + "loss": 0.0225, + "step": 16723 + }, + { + "epoch": 0.39203689703874306, + "grad_norm": 0.6001664400100708, + "learning_rate": 0.00013331257162236805, + "loss": 0.5215, + "step": 16724 + }, + { + "epoch": 0.39206033861354805, + "grad_norm": 0.5681214928627014, + "learning_rate": 0.00013330562611127745, + "loss": 0.1918, + "step": 16725 + }, + { + "epoch": 0.39208378018835305, + "grad_norm": 0.5622208118438721, + "learning_rate": 0.0001332986804194696, + "loss": 0.0637, + "step": 16726 + }, + { + "epoch": 0.39210722176315804, + "grad_norm": 0.42726999521255493, + "learning_rate": 0.0001332917345469822, + "loss": 0.1258, + "step": 16727 + }, + { + "epoch": 0.39213066333796304, + "grad_norm": 0.09728769212961197, + "learning_rate": 0.0001332847884938529, + "loss": 0.019, + "step": 16728 + }, + { + "epoch": 0.39215410491276803, + "grad_norm": 0.42327576875686646, + "learning_rate": 0.00013327784226011937, + "loss": 0.5925, + "step": 16729 + }, + { + "epoch": 0.392177546487573, + "grad_norm": 0.7454590797424316, + "learning_rate": 0.00013327089584581935, + "loss": 0.1773, + "step": 16730 + }, + { + "epoch": 0.392200988062378, + "grad_norm": 0.18775217235088348, + "learning_rate": 0.00013326394925099048, + "loss": 0.0426, + "step": 16731 + }, + { + "epoch": 0.392224429637183, + "grad_norm": 0.2827722728252411, + "learning_rate": 0.00013325700247567054, + "loss": 0.0255, + "step": 16732 + }, + { + "epoch": 0.392247871211988, + "grad_norm": 0.41965457797050476, + "learning_rate": 0.00013325005551989715, + "loss": 0.1256, + "step": 16733 + }, + { + "epoch": 0.392271312786793, + "grad_norm": 0.1064617931842804, + "learning_rate": 0.00013324310838370803, + "loss": 0.0172, + "step": 16734 + }, + { + "epoch": 0.392294754361598, + "grad_norm": 0.5318948030471802, + "learning_rate": 0.00013323616106714086, + "loss": 0.0798, + "step": 16735 + }, + { + "epoch": 0.392318195936403, + "grad_norm": 0.7040808200836182, + "learning_rate": 0.00013322921357023334, + "loss": 0.2496, + "step": 16736 + }, + { + "epoch": 0.392341637511208, + "grad_norm": 0.15177834033966064, + "learning_rate": 0.0001332222658930232, + "loss": 0.0289, + "step": 16737 + }, + { + "epoch": 0.392365079086013, + "grad_norm": 0.14068660140037537, + "learning_rate": 0.00013321531803554807, + "loss": 0.0353, + "step": 16738 + }, + { + "epoch": 0.392388520660818, + "grad_norm": 0.5227041244506836, + "learning_rate": 0.0001332083699978457, + "loss": 0.2855, + "step": 16739 + }, + { + "epoch": 0.39241196223562297, + "grad_norm": 0.5781932473182678, + "learning_rate": 0.00013320142177995382, + "loss": 0.1755, + "step": 16740 + }, + { + "epoch": 0.39243540381042796, + "grad_norm": 0.5006409287452698, + "learning_rate": 0.00013319447338191006, + "loss": 0.1373, + "step": 16741 + }, + { + "epoch": 0.39245884538523296, + "grad_norm": 0.4798523485660553, + "learning_rate": 0.00013318752480375213, + "loss": 0.0911, + "step": 16742 + }, + { + "epoch": 0.39248228696003795, + "grad_norm": 0.18963243067264557, + "learning_rate": 0.00013318057604551781, + "loss": 0.0491, + "step": 16743 + }, + { + "epoch": 0.39250572853484295, + "grad_norm": 0.4074784219264984, + "learning_rate": 0.0001331736271072447, + "loss": 0.1043, + "step": 16744 + }, + { + "epoch": 0.39252917010964794, + "grad_norm": 0.20674490928649902, + "learning_rate": 0.0001331666779889706, + "loss": 0.0274, + "step": 16745 + }, + { + "epoch": 0.39255261168445293, + "grad_norm": 0.5042455792427063, + "learning_rate": 0.00013315972869073318, + "loss": 0.5817, + "step": 16746 + }, + { + "epoch": 0.392576053259258, + "grad_norm": 0.12337421625852585, + "learning_rate": 0.0001331527792125701, + "loss": 0.0193, + "step": 16747 + }, + { + "epoch": 0.392599494834063, + "grad_norm": 0.41924574971199036, + "learning_rate": 0.00013314582955451914, + "loss": 0.1314, + "step": 16748 + }, + { + "epoch": 0.392622936408868, + "grad_norm": 0.4035458564758301, + "learning_rate": 0.00013313887971661795, + "loss": 0.483, + "step": 16749 + }, + { + "epoch": 0.39264637798367297, + "grad_norm": 0.48532408475875854, + "learning_rate": 0.00013313192969890424, + "loss": 0.0843, + "step": 16750 + }, + { + "epoch": 0.39266981955847796, + "grad_norm": 0.41168928146362305, + "learning_rate": 0.00013312497950141583, + "loss": 0.467, + "step": 16751 + }, + { + "epoch": 0.39269326113328296, + "grad_norm": 0.25306904315948486, + "learning_rate": 0.00013311802912419028, + "loss": 0.0435, + "step": 16752 + }, + { + "epoch": 0.39271670270808795, + "grad_norm": 0.5521862506866455, + "learning_rate": 0.00013311107856726543, + "loss": 0.091, + "step": 16753 + }, + { + "epoch": 0.39274014428289294, + "grad_norm": 0.5184511542320251, + "learning_rate": 0.0001331041278306789, + "loss": 0.1514, + "step": 16754 + }, + { + "epoch": 0.39276358585769794, + "grad_norm": 0.39007893204689026, + "learning_rate": 0.00013309717691446844, + "loss": 0.1142, + "step": 16755 + }, + { + "epoch": 0.39278702743250293, + "grad_norm": 0.6526408791542053, + "learning_rate": 0.00013309022581867177, + "loss": 0.1805, + "step": 16756 + }, + { + "epoch": 0.3928104690073079, + "grad_norm": 0.4067758321762085, + "learning_rate": 0.0001330832745433266, + "loss": 0.0469, + "step": 16757 + }, + { + "epoch": 0.3928339105821129, + "grad_norm": 0.6907901167869568, + "learning_rate": 0.00013307632308847068, + "loss": 0.6875, + "step": 16758 + }, + { + "epoch": 0.3928573521569179, + "grad_norm": 0.40523645281791687, + "learning_rate": 0.00013306937145414165, + "loss": 0.0796, + "step": 16759 + }, + { + "epoch": 0.3928807937317229, + "grad_norm": 0.6161928772926331, + "learning_rate": 0.00013306241964037733, + "loss": 0.687, + "step": 16760 + }, + { + "epoch": 0.3929042353065279, + "grad_norm": 0.21066069602966309, + "learning_rate": 0.00013305546764721536, + "loss": 0.012, + "step": 16761 + }, + { + "epoch": 0.3929276768813329, + "grad_norm": 0.32095545530319214, + "learning_rate": 0.00013304851547469353, + "loss": 0.0453, + "step": 16762 + }, + { + "epoch": 0.3929511184561379, + "grad_norm": 0.4496789574623108, + "learning_rate": 0.0001330415631228495, + "loss": 0.0775, + "step": 16763 + }, + { + "epoch": 0.3929745600309429, + "grad_norm": 0.4578332006931305, + "learning_rate": 0.000133034610591721, + "loss": 0.0847, + "step": 16764 + }, + { + "epoch": 0.3929980016057479, + "grad_norm": 0.12719525396823883, + "learning_rate": 0.0001330276578813458, + "loss": 0.0224, + "step": 16765 + }, + { + "epoch": 0.3930214431805529, + "grad_norm": 0.25506097078323364, + "learning_rate": 0.0001330207049917616, + "loss": 0.0365, + "step": 16766 + }, + { + "epoch": 0.39304488475535787, + "grad_norm": 0.691006064414978, + "learning_rate": 0.00013301375192300609, + "loss": 0.5994, + "step": 16767 + }, + { + "epoch": 0.39306832633016286, + "grad_norm": 0.3824203908443451, + "learning_rate": 0.00013300679867511708, + "loss": 0.0646, + "step": 16768 + }, + { + "epoch": 0.39309176790496786, + "grad_norm": 0.2918822765350342, + "learning_rate": 0.00013299984524813225, + "loss": 0.05, + "step": 16769 + }, + { + "epoch": 0.39311520947977285, + "grad_norm": 0.5345154404640198, + "learning_rate": 0.00013299289164208934, + "loss": 0.5143, + "step": 16770 + }, + { + "epoch": 0.39313865105457785, + "grad_norm": 0.07629945874214172, + "learning_rate": 0.00013298593785702605, + "loss": 0.0098, + "step": 16771 + }, + { + "epoch": 0.39316209262938284, + "grad_norm": 0.16382727026939392, + "learning_rate": 0.00013297898389298012, + "loss": 0.0249, + "step": 16772 + }, + { + "epoch": 0.39318553420418784, + "grad_norm": 0.6378414630889893, + "learning_rate": 0.00013297202974998928, + "loss": 0.1281, + "step": 16773 + }, + { + "epoch": 0.39320897577899283, + "grad_norm": 0.168849378824234, + "learning_rate": 0.00013296507542809134, + "loss": 0.0319, + "step": 16774 + }, + { + "epoch": 0.3932324173537978, + "grad_norm": 0.5109586119651794, + "learning_rate": 0.00013295812092732392, + "loss": 0.1009, + "step": 16775 + }, + { + "epoch": 0.3932558589286028, + "grad_norm": 0.3227660655975342, + "learning_rate": 0.00013295116624772486, + "loss": 0.334, + "step": 16776 + }, + { + "epoch": 0.3932793005034078, + "grad_norm": 0.6151506900787354, + "learning_rate": 0.0001329442113893318, + "loss": 0.1936, + "step": 16777 + }, + { + "epoch": 0.3933027420782128, + "grad_norm": 0.39463719725608826, + "learning_rate": 0.00013293725635218256, + "loss": 0.0903, + "step": 16778 + }, + { + "epoch": 0.3933261836530178, + "grad_norm": 0.15652725100517273, + "learning_rate": 0.0001329303011363148, + "loss": 0.0317, + "step": 16779 + }, + { + "epoch": 0.3933496252278228, + "grad_norm": 0.31755372881889343, + "learning_rate": 0.0001329233457417663, + "loss": 0.0343, + "step": 16780 + }, + { + "epoch": 0.3933730668026278, + "grad_norm": 0.385273277759552, + "learning_rate": 0.00013291639016857487, + "loss": 0.0789, + "step": 16781 + }, + { + "epoch": 0.3933965083774328, + "grad_norm": 0.13208675384521484, + "learning_rate": 0.00013290943441677811, + "loss": 0.0268, + "step": 16782 + }, + { + "epoch": 0.3934199499522378, + "grad_norm": 0.2909526526927948, + "learning_rate": 0.00013290247848641388, + "loss": 0.0291, + "step": 16783 + }, + { + "epoch": 0.3934433915270428, + "grad_norm": 0.13745595514774323, + "learning_rate": 0.00013289552237751986, + "loss": 0.0236, + "step": 16784 + }, + { + "epoch": 0.39346683310184777, + "grad_norm": 0.4922310411930084, + "learning_rate": 0.00013288856609013379, + "loss": 0.0939, + "step": 16785 + }, + { + "epoch": 0.39349027467665276, + "grad_norm": 0.6823568940162659, + "learning_rate": 0.00013288160962429343, + "loss": 0.1041, + "step": 16786 + }, + { + "epoch": 0.39351371625145776, + "grad_norm": 0.8103374242782593, + "learning_rate": 0.00013287465298003658, + "loss": 0.0525, + "step": 16787 + }, + { + "epoch": 0.39353715782626275, + "grad_norm": 0.4456450939178467, + "learning_rate": 0.0001328676961574009, + "loss": 0.1133, + "step": 16788 + }, + { + "epoch": 0.39356059940106775, + "grad_norm": 0.19390568137168884, + "learning_rate": 0.0001328607391564242, + "loss": 0.0496, + "step": 16789 + }, + { + "epoch": 0.39358404097587274, + "grad_norm": 1.1135114431381226, + "learning_rate": 0.0001328537819771442, + "loss": 0.2708, + "step": 16790 + }, + { + "epoch": 0.39360748255067773, + "grad_norm": 0.9127975106239319, + "learning_rate": 0.00013284682461959862, + "loss": 0.1382, + "step": 16791 + }, + { + "epoch": 0.39363092412548273, + "grad_norm": 0.5653930306434631, + "learning_rate": 0.00013283986708382528, + "loss": 0.0938, + "step": 16792 + }, + { + "epoch": 0.3936543657002877, + "grad_norm": 0.3969185948371887, + "learning_rate": 0.0001328329093698619, + "loss": 0.078, + "step": 16793 + }, + { + "epoch": 0.3936778072750927, + "grad_norm": 0.17117412388324738, + "learning_rate": 0.00013282595147774622, + "loss": 0.0312, + "step": 16794 + }, + { + "epoch": 0.3937012488498977, + "grad_norm": 0.8369539380073547, + "learning_rate": 0.000132818993407516, + "loss": 0.3132, + "step": 16795 + }, + { + "epoch": 0.3937246904247027, + "grad_norm": 0.4267025589942932, + "learning_rate": 0.00013281203515920904, + "loss": 0.0771, + "step": 16796 + }, + { + "epoch": 0.3937481319995077, + "grad_norm": 0.5070844888687134, + "learning_rate": 0.000132805076732863, + "loss": 0.0776, + "step": 16797 + }, + { + "epoch": 0.39377157357431275, + "grad_norm": 0.7114118933677673, + "learning_rate": 0.0001327981181285157, + "loss": 0.1062, + "step": 16798 + }, + { + "epoch": 0.39379501514911774, + "grad_norm": 0.6177061796188354, + "learning_rate": 0.00013279115934620493, + "loss": 0.2493, + "step": 16799 + }, + { + "epoch": 0.39381845672392274, + "grad_norm": 0.416763037443161, + "learning_rate": 0.00013278420038596837, + "loss": 0.1045, + "step": 16800 + }, + { + "epoch": 0.39384189829872773, + "grad_norm": 0.9107728004455566, + "learning_rate": 0.0001327772412478438, + "loss": 0.3531, + "step": 16801 + }, + { + "epoch": 0.3938653398735327, + "grad_norm": 0.7009713649749756, + "learning_rate": 0.00013277028193186904, + "loss": 0.5907, + "step": 16802 + }, + { + "epoch": 0.3938887814483377, + "grad_norm": 0.2637082636356354, + "learning_rate": 0.00013276332243808178, + "loss": 0.0426, + "step": 16803 + }, + { + "epoch": 0.3939122230231427, + "grad_norm": 0.4194445312023163, + "learning_rate": 0.00013275636276651983, + "loss": 0.0603, + "step": 16804 + }, + { + "epoch": 0.3939356645979477, + "grad_norm": 0.47253867983818054, + "learning_rate": 0.00013274940291722092, + "loss": 0.0708, + "step": 16805 + }, + { + "epoch": 0.3939591061727527, + "grad_norm": 0.7215438485145569, + "learning_rate": 0.00013274244289022285, + "loss": 0.0994, + "step": 16806 + }, + { + "epoch": 0.3939825477475577, + "grad_norm": 0.22522622346878052, + "learning_rate": 0.00013273548268556335, + "loss": 0.0263, + "step": 16807 + }, + { + "epoch": 0.3940059893223627, + "grad_norm": 0.6232947111129761, + "learning_rate": 0.0001327285223032802, + "loss": 0.1356, + "step": 16808 + }, + { + "epoch": 0.3940294308971677, + "grad_norm": 0.35962870717048645, + "learning_rate": 0.00013272156174341115, + "loss": 0.0675, + "step": 16809 + }, + { + "epoch": 0.3940528724719727, + "grad_norm": 0.3560912013053894, + "learning_rate": 0.00013271460100599403, + "loss": 0.029, + "step": 16810 + }, + { + "epoch": 0.3940763140467777, + "grad_norm": 0.43681395053863525, + "learning_rate": 0.0001327076400910665, + "loss": 0.1181, + "step": 16811 + }, + { + "epoch": 0.39409975562158267, + "grad_norm": 0.5878908038139343, + "learning_rate": 0.00013270067899866646, + "loss": 0.1364, + "step": 16812 + }, + { + "epoch": 0.39412319719638766, + "grad_norm": 0.7481391429901123, + "learning_rate": 0.00013269371772883158, + "loss": 0.2487, + "step": 16813 + }, + { + "epoch": 0.39414663877119266, + "grad_norm": 0.5275336503982544, + "learning_rate": 0.0001326867562815997, + "loss": 0.1221, + "step": 16814 + }, + { + "epoch": 0.39417008034599765, + "grad_norm": 0.454520583152771, + "learning_rate": 0.00013267979465700854, + "loss": 0.5876, + "step": 16815 + }, + { + "epoch": 0.39419352192080265, + "grad_norm": 0.4440452456474304, + "learning_rate": 0.0001326728328550959, + "loss": 0.0495, + "step": 16816 + }, + { + "epoch": 0.39421696349560764, + "grad_norm": 0.4148804545402527, + "learning_rate": 0.00013266587087589952, + "loss": 0.0676, + "step": 16817 + }, + { + "epoch": 0.39424040507041264, + "grad_norm": 1.0364876985549927, + "learning_rate": 0.00013265890871945726, + "loss": 0.0665, + "step": 16818 + }, + { + "epoch": 0.39426384664521763, + "grad_norm": 0.4012022018432617, + "learning_rate": 0.0001326519463858068, + "loss": 0.0999, + "step": 16819 + }, + { + "epoch": 0.3942872882200226, + "grad_norm": 0.1668652594089508, + "learning_rate": 0.00013264498387498597, + "loss": 0.0536, + "step": 16820 + }, + { + "epoch": 0.3943107297948276, + "grad_norm": 0.446648508310318, + "learning_rate": 0.00013263802118703254, + "loss": 0.0773, + "step": 16821 + }, + { + "epoch": 0.3943341713696326, + "grad_norm": 0.12265249341726303, + "learning_rate": 0.0001326310583219843, + "loss": 0.0152, + "step": 16822 + }, + { + "epoch": 0.3943576129444376, + "grad_norm": 0.4164581596851349, + "learning_rate": 0.000132624095279879, + "loss": 0.0905, + "step": 16823 + }, + { + "epoch": 0.3943810545192426, + "grad_norm": 0.15064752101898193, + "learning_rate": 0.00013261713206075447, + "loss": 0.0271, + "step": 16824 + }, + { + "epoch": 0.3944044960940476, + "grad_norm": 0.4297712445259094, + "learning_rate": 0.00013261016866464846, + "loss": 0.0851, + "step": 16825 + }, + { + "epoch": 0.3944279376688526, + "grad_norm": 0.5579659938812256, + "learning_rate": 0.00013260320509159875, + "loss": 0.1557, + "step": 16826 + }, + { + "epoch": 0.3944513792436576, + "grad_norm": 0.4419550597667694, + "learning_rate": 0.00013259624134164314, + "loss": 0.0949, + "step": 16827 + }, + { + "epoch": 0.3944748208184626, + "grad_norm": 0.19799481332302094, + "learning_rate": 0.0001325892774148194, + "loss": 0.0389, + "step": 16828 + }, + { + "epoch": 0.3944982623932676, + "grad_norm": 0.4068719744682312, + "learning_rate": 0.0001325823133111653, + "loss": 0.1066, + "step": 16829 + }, + { + "epoch": 0.39452170396807257, + "grad_norm": 0.32791027426719666, + "learning_rate": 0.0001325753490307187, + "loss": 0.0841, + "step": 16830 + }, + { + "epoch": 0.39454514554287756, + "grad_norm": 0.3869100511074066, + "learning_rate": 0.00013256838457351733, + "loss": 0.0953, + "step": 16831 + }, + { + "epoch": 0.39456858711768256, + "grad_norm": 0.6090846061706543, + "learning_rate": 0.00013256141993959894, + "loss": 0.0952, + "step": 16832 + }, + { + "epoch": 0.39459202869248755, + "grad_norm": 0.5765509009361267, + "learning_rate": 0.0001325544551290014, + "loss": 0.1112, + "step": 16833 + }, + { + "epoch": 0.39461547026729255, + "grad_norm": 0.4314897954463959, + "learning_rate": 0.0001325474901417625, + "loss": 0.0327, + "step": 16834 + }, + { + "epoch": 0.39463891184209754, + "grad_norm": 0.4838068187236786, + "learning_rate": 0.00013254052497791997, + "loss": 0.0827, + "step": 16835 + }, + { + "epoch": 0.39466235341690253, + "grad_norm": 0.238782599568367, + "learning_rate": 0.00013253355963751165, + "loss": 0.0537, + "step": 16836 + }, + { + "epoch": 0.39468579499170753, + "grad_norm": 0.12481699883937836, + "learning_rate": 0.00013252659412057533, + "loss": 0.0138, + "step": 16837 + }, + { + "epoch": 0.3947092365665125, + "grad_norm": 0.42199480533599854, + "learning_rate": 0.00013251962842714878, + "loss": 0.1017, + "step": 16838 + }, + { + "epoch": 0.3947326781413175, + "grad_norm": 0.10615339130163193, + "learning_rate": 0.0001325126625572698, + "loss": 0.0179, + "step": 16839 + }, + { + "epoch": 0.3947561197161225, + "grad_norm": 0.2004469782114029, + "learning_rate": 0.0001325056965109762, + "loss": 0.0456, + "step": 16840 + }, + { + "epoch": 0.3947795612909275, + "grad_norm": 0.2703371047973633, + "learning_rate": 0.00013249873028830578, + "loss": 0.0495, + "step": 16841 + }, + { + "epoch": 0.3948030028657325, + "grad_norm": 0.6076574325561523, + "learning_rate": 0.00013249176388929637, + "loss": 0.1288, + "step": 16842 + }, + { + "epoch": 0.3948264444405375, + "grad_norm": 0.5677744746208191, + "learning_rate": 0.0001324847973139857, + "loss": 0.0662, + "step": 16843 + }, + { + "epoch": 0.3948498860153425, + "grad_norm": 0.5019338130950928, + "learning_rate": 0.00013247783056241167, + "loss": 0.0897, + "step": 16844 + }, + { + "epoch": 0.3948733275901475, + "grad_norm": 0.29679951071739197, + "learning_rate": 0.00013247086363461192, + "loss": 0.0376, + "step": 16845 + }, + { + "epoch": 0.3948967691649525, + "grad_norm": 0.2737821936607361, + "learning_rate": 0.00013246389653062438, + "loss": 0.0642, + "step": 16846 + }, + { + "epoch": 0.39492021073975747, + "grad_norm": 0.5994061827659607, + "learning_rate": 0.00013245692925048687, + "loss": 0.1871, + "step": 16847 + }, + { + "epoch": 0.39494365231456247, + "grad_norm": 0.6448187828063965, + "learning_rate": 0.0001324499617942371, + "loss": 0.1303, + "step": 16848 + }, + { + "epoch": 0.3949670938893675, + "grad_norm": 0.6210426688194275, + "learning_rate": 0.00013244299416191298, + "loss": 0.0961, + "step": 16849 + }, + { + "epoch": 0.3949905354641725, + "grad_norm": 0.5314897298812866, + "learning_rate": 0.00013243602635355222, + "loss": 0.7159, + "step": 16850 + }, + { + "epoch": 0.3950139770389775, + "grad_norm": 0.486867755651474, + "learning_rate": 0.00013242905836919266, + "loss": 0.1614, + "step": 16851 + }, + { + "epoch": 0.3950374186137825, + "grad_norm": 0.46793511509895325, + "learning_rate": 0.00013242209020887212, + "loss": 0.0584, + "step": 16852 + }, + { + "epoch": 0.3950608601885875, + "grad_norm": 0.2474619299173355, + "learning_rate": 0.00013241512187262843, + "loss": 0.0214, + "step": 16853 + }, + { + "epoch": 0.3950843017633925, + "grad_norm": 0.6654395461082458, + "learning_rate": 0.00013240815336049939, + "loss": 0.2063, + "step": 16854 + }, + { + "epoch": 0.3951077433381975, + "grad_norm": 0.41065356135368347, + "learning_rate": 0.00013240118467252275, + "loss": 0.0521, + "step": 16855 + }, + { + "epoch": 0.3951311849130025, + "grad_norm": 0.38393598794937134, + "learning_rate": 0.00013239421580873642, + "loss": 0.7761, + "step": 16856 + }, + { + "epoch": 0.39515462648780747, + "grad_norm": 0.7488507628440857, + "learning_rate": 0.00013238724676917815, + "loss": 0.0938, + "step": 16857 + }, + { + "epoch": 0.39517806806261246, + "grad_norm": 0.339620441198349, + "learning_rate": 0.00013238027755388572, + "loss": 0.1046, + "step": 16858 + }, + { + "epoch": 0.39520150963741746, + "grad_norm": 0.4389500617980957, + "learning_rate": 0.00013237330816289706, + "loss": 0.12, + "step": 16859 + }, + { + "epoch": 0.39522495121222245, + "grad_norm": 0.390291690826416, + "learning_rate": 0.0001323663385962499, + "loss": 0.1229, + "step": 16860 + }, + { + "epoch": 0.39524839278702745, + "grad_norm": 0.4991559684276581, + "learning_rate": 0.00013235936885398205, + "loss": 0.0816, + "step": 16861 + }, + { + "epoch": 0.39527183436183244, + "grad_norm": 0.15820395946502686, + "learning_rate": 0.00013235239893613139, + "loss": 0.0334, + "step": 16862 + }, + { + "epoch": 0.39529527593663744, + "grad_norm": 0.43925487995147705, + "learning_rate": 0.00013234542884273568, + "loss": 0.1081, + "step": 16863 + }, + { + "epoch": 0.39531871751144243, + "grad_norm": 0.5842031240463257, + "learning_rate": 0.00013233845857383276, + "loss": 0.0957, + "step": 16864 + }, + { + "epoch": 0.3953421590862474, + "grad_norm": 0.22962695360183716, + "learning_rate": 0.00013233148812946048, + "loss": 0.045, + "step": 16865 + }, + { + "epoch": 0.3953656006610524, + "grad_norm": 0.18459822237491608, + "learning_rate": 0.00013232451750965662, + "loss": 0.0364, + "step": 16866 + }, + { + "epoch": 0.3953890422358574, + "grad_norm": 0.39270809292793274, + "learning_rate": 0.000132317546714459, + "loss": 0.0472, + "step": 16867 + }, + { + "epoch": 0.3954124838106624, + "grad_norm": 0.11497170478105545, + "learning_rate": 0.0001323105757439055, + "loss": 0.0237, + "step": 16868 + }, + { + "epoch": 0.3954359253854674, + "grad_norm": 0.36099010705947876, + "learning_rate": 0.0001323036045980339, + "loss": 0.0994, + "step": 16869 + }, + { + "epoch": 0.3954593669602724, + "grad_norm": 0.37859413027763367, + "learning_rate": 0.00013229663327688203, + "loss": 0.0877, + "step": 16870 + }, + { + "epoch": 0.3954828085350774, + "grad_norm": 0.3512002229690552, + "learning_rate": 0.0001322896617804877, + "loss": 0.0893, + "step": 16871 + }, + { + "epoch": 0.3955062501098824, + "grad_norm": 0.5624083280563354, + "learning_rate": 0.00013228269010888877, + "loss": 0.1017, + "step": 16872 + }, + { + "epoch": 0.3955296916846874, + "grad_norm": 0.6319852471351624, + "learning_rate": 0.0001322757182621231, + "loss": 0.0975, + "step": 16873 + }, + { + "epoch": 0.3955531332594924, + "grad_norm": 0.8377213478088379, + "learning_rate": 0.0001322687462402284, + "loss": 0.0938, + "step": 16874 + }, + { + "epoch": 0.39557657483429737, + "grad_norm": 0.1880151778459549, + "learning_rate": 0.00013226177404324261, + "loss": 0.0365, + "step": 16875 + }, + { + "epoch": 0.39560001640910236, + "grad_norm": 0.262037456035614, + "learning_rate": 0.0001322548016712035, + "loss": 0.0609, + "step": 16876 + }, + { + "epoch": 0.39562345798390736, + "grad_norm": 0.612358033657074, + "learning_rate": 0.00013224782912414897, + "loss": 0.1291, + "step": 16877 + }, + { + "epoch": 0.39564689955871235, + "grad_norm": 0.46088919043540955, + "learning_rate": 0.0001322408564021168, + "loss": 0.0701, + "step": 16878 + }, + { + "epoch": 0.39567034113351734, + "grad_norm": 0.19850963354110718, + "learning_rate": 0.00013223388350514485, + "loss": 0.0382, + "step": 16879 + }, + { + "epoch": 0.39569378270832234, + "grad_norm": 0.3488193452358246, + "learning_rate": 0.00013222691043327092, + "loss": 0.0616, + "step": 16880 + }, + { + "epoch": 0.39571722428312733, + "grad_norm": 0.3585197925567627, + "learning_rate": 0.00013221993718653288, + "loss": 0.1087, + "step": 16881 + }, + { + "epoch": 0.3957406658579323, + "grad_norm": 0.5025853514671326, + "learning_rate": 0.00013221296376496853, + "loss": 0.17, + "step": 16882 + }, + { + "epoch": 0.3957641074327373, + "grad_norm": 0.1766495406627655, + "learning_rate": 0.00013220599016861576, + "loss": 0.0317, + "step": 16883 + }, + { + "epoch": 0.3957875490075423, + "grad_norm": 0.48378726840019226, + "learning_rate": 0.0001321990163975124, + "loss": 0.4167, + "step": 16884 + }, + { + "epoch": 0.3958109905823473, + "grad_norm": 0.32830482721328735, + "learning_rate": 0.0001321920424516962, + "loss": 0.1047, + "step": 16885 + }, + { + "epoch": 0.3958344321571523, + "grad_norm": 0.12505438923835754, + "learning_rate": 0.00013218506833120514, + "loss": 0.0248, + "step": 16886 + }, + { + "epoch": 0.3958578737319573, + "grad_norm": 0.40154772996902466, + "learning_rate": 0.00013217809403607696, + "loss": 0.0898, + "step": 16887 + }, + { + "epoch": 0.3958813153067623, + "grad_norm": 0.4433303773403168, + "learning_rate": 0.00013217111956634954, + "loss": 0.0722, + "step": 16888 + }, + { + "epoch": 0.3959047568815673, + "grad_norm": 0.44065046310424805, + "learning_rate": 0.00013216414492206075, + "loss": 0.4782, + "step": 16889 + }, + { + "epoch": 0.3959281984563723, + "grad_norm": 0.4446621835231781, + "learning_rate": 0.00013215717010324837, + "loss": 0.0767, + "step": 16890 + }, + { + "epoch": 0.3959516400311773, + "grad_norm": 0.16485224664211273, + "learning_rate": 0.0001321501951099503, + "loss": 0.0396, + "step": 16891 + }, + { + "epoch": 0.39597508160598227, + "grad_norm": 0.52656489610672, + "learning_rate": 0.00013214321994220438, + "loss": 0.1022, + "step": 16892 + }, + { + "epoch": 0.39599852318078727, + "grad_norm": 0.7286596894264221, + "learning_rate": 0.0001321362446000484, + "loss": 0.0952, + "step": 16893 + }, + { + "epoch": 0.39602196475559226, + "grad_norm": 0.5242775082588196, + "learning_rate": 0.00013212926908352029, + "loss": 0.0707, + "step": 16894 + }, + { + "epoch": 0.39604540633039725, + "grad_norm": 0.5210809111595154, + "learning_rate": 0.00013212229339265786, + "loss": 0.0839, + "step": 16895 + }, + { + "epoch": 0.39606884790520225, + "grad_norm": 2.0544023513793945, + "learning_rate": 0.00013211531752749896, + "loss": 0.0854, + "step": 16896 + }, + { + "epoch": 0.39609228948000724, + "grad_norm": 0.9883602857589722, + "learning_rate": 0.0001321083414880814, + "loss": 0.1063, + "step": 16897 + }, + { + "epoch": 0.39611573105481224, + "grad_norm": 0.7242961525917053, + "learning_rate": 0.00013210136527444318, + "loss": 0.1454, + "step": 16898 + }, + { + "epoch": 0.39613917262961723, + "grad_norm": 0.5295402407646179, + "learning_rate": 0.00013209438888662196, + "loss": 0.5357, + "step": 16899 + }, + { + "epoch": 0.3961626142044222, + "grad_norm": 0.46549567580223083, + "learning_rate": 0.00013208741232465573, + "loss": 0.4191, + "step": 16900 + }, + { + "epoch": 0.3961860557792273, + "grad_norm": 0.34918835759162903, + "learning_rate": 0.00013208043558858226, + "loss": 0.0975, + "step": 16901 + }, + { + "epoch": 0.39620949735403227, + "grad_norm": 0.8698142170906067, + "learning_rate": 0.00013207345867843946, + "loss": 0.1552, + "step": 16902 + }, + { + "epoch": 0.39623293892883726, + "grad_norm": 0.500086784362793, + "learning_rate": 0.00013206648159426516, + "loss": 0.1195, + "step": 16903 + }, + { + "epoch": 0.39625638050364226, + "grad_norm": 0.14088524878025055, + "learning_rate": 0.00013205950433609726, + "loss": 0.0389, + "step": 16904 + }, + { + "epoch": 0.39627982207844725, + "grad_norm": 0.5628912448883057, + "learning_rate": 0.0001320525269039736, + "loss": 0.0619, + "step": 16905 + }, + { + "epoch": 0.39630326365325225, + "grad_norm": 0.5629056096076965, + "learning_rate": 0.00013204554929793197, + "loss": 0.0856, + "step": 16906 + }, + { + "epoch": 0.39632670522805724, + "grad_norm": 0.4619743525981903, + "learning_rate": 0.0001320385715180103, + "loss": 0.0814, + "step": 16907 + }, + { + "epoch": 0.39635014680286224, + "grad_norm": 0.7639310956001282, + "learning_rate": 0.00013203159356424646, + "loss": 0.1883, + "step": 16908 + }, + { + "epoch": 0.39637358837766723, + "grad_norm": 0.23565468192100525, + "learning_rate": 0.00013202461543667827, + "loss": 0.0475, + "step": 16909 + }, + { + "epoch": 0.3963970299524722, + "grad_norm": 0.6181635856628418, + "learning_rate": 0.00013201763713534365, + "loss": 0.5935, + "step": 16910 + }, + { + "epoch": 0.3964204715272772, + "grad_norm": 0.5515952706336975, + "learning_rate": 0.00013201065866028043, + "loss": 0.0912, + "step": 16911 + }, + { + "epoch": 0.3964439131020822, + "grad_norm": 0.5726709365844727, + "learning_rate": 0.00013200368001152645, + "loss": 0.6345, + "step": 16912 + }, + { + "epoch": 0.3964673546768872, + "grad_norm": 0.20504967868328094, + "learning_rate": 0.00013199670118911963, + "loss": 0.0407, + "step": 16913 + }, + { + "epoch": 0.3964907962516922, + "grad_norm": 0.20706883072853088, + "learning_rate": 0.0001319897221930978, + "loss": 0.0253, + "step": 16914 + }, + { + "epoch": 0.3965142378264972, + "grad_norm": 0.3524876832962036, + "learning_rate": 0.00013198274302349882, + "loss": 0.0859, + "step": 16915 + }, + { + "epoch": 0.3965376794013022, + "grad_norm": 0.6489385366439819, + "learning_rate": 0.0001319757636803606, + "loss": 0.6878, + "step": 16916 + }, + { + "epoch": 0.3965611209761072, + "grad_norm": 0.7762898802757263, + "learning_rate": 0.00013196878416372098, + "loss": 0.1163, + "step": 16917 + }, + { + "epoch": 0.3965845625509122, + "grad_norm": 0.14881321787834167, + "learning_rate": 0.00013196180447361781, + "loss": 0.0183, + "step": 16918 + }, + { + "epoch": 0.3966080041257172, + "grad_norm": 0.11489390581846237, + "learning_rate": 0.00013195482461008902, + "loss": 0.0262, + "step": 16919 + }, + { + "epoch": 0.39663144570052217, + "grad_norm": 0.342996746301651, + "learning_rate": 0.00013194784457317245, + "loss": 0.1066, + "step": 16920 + }, + { + "epoch": 0.39665488727532716, + "grad_norm": 0.2907664179801941, + "learning_rate": 0.000131940864362906, + "loss": 0.0265, + "step": 16921 + }, + { + "epoch": 0.39667832885013216, + "grad_norm": 0.7563575506210327, + "learning_rate": 0.00013193388397932747, + "loss": 0.1645, + "step": 16922 + }, + { + "epoch": 0.39670177042493715, + "grad_norm": 0.35780513286590576, + "learning_rate": 0.00013192690342247483, + "loss": 0.3304, + "step": 16923 + }, + { + "epoch": 0.39672521199974214, + "grad_norm": 0.7799808382987976, + "learning_rate": 0.0001319199226923859, + "loss": 0.1082, + "step": 16924 + }, + { + "epoch": 0.39674865357454714, + "grad_norm": 0.446979284286499, + "learning_rate": 0.00013191294178909856, + "loss": 0.1574, + "step": 16925 + }, + { + "epoch": 0.39677209514935213, + "grad_norm": 0.36208444833755493, + "learning_rate": 0.00013190596071265071, + "loss": 0.0393, + "step": 16926 + }, + { + "epoch": 0.3967955367241571, + "grad_norm": 0.6596208214759827, + "learning_rate": 0.00013189897946308023, + "loss": 0.1593, + "step": 16927 + }, + { + "epoch": 0.3968189782989621, + "grad_norm": 0.5840670466423035, + "learning_rate": 0.000131891998040425, + "loss": 0.1028, + "step": 16928 + }, + { + "epoch": 0.3968424198737671, + "grad_norm": 0.6966606378555298, + "learning_rate": 0.00013188501644472288, + "loss": 0.1088, + "step": 16929 + }, + { + "epoch": 0.3968658614485721, + "grad_norm": 0.6730362176895142, + "learning_rate": 0.00013187803467601175, + "loss": 0.101, + "step": 16930 + }, + { + "epoch": 0.3968893030233771, + "grad_norm": 0.3575619161128998, + "learning_rate": 0.0001318710527343295, + "loss": 0.0747, + "step": 16931 + }, + { + "epoch": 0.3969127445981821, + "grad_norm": 0.6728456616401672, + "learning_rate": 0.00013186407061971405, + "loss": 0.7091, + "step": 16932 + }, + { + "epoch": 0.3969361861729871, + "grad_norm": 0.5203077793121338, + "learning_rate": 0.00013185708833220324, + "loss": 0.1132, + "step": 16933 + }, + { + "epoch": 0.3969596277477921, + "grad_norm": 0.11781027168035507, + "learning_rate": 0.00013185010587183496, + "loss": 0.0141, + "step": 16934 + }, + { + "epoch": 0.3969830693225971, + "grad_norm": 0.45692741870880127, + "learning_rate": 0.00013184312323864713, + "loss": 0.0872, + "step": 16935 + }, + { + "epoch": 0.3970065108974021, + "grad_norm": 0.4763132929801941, + "learning_rate": 0.00013183614043267764, + "loss": 0.0771, + "step": 16936 + }, + { + "epoch": 0.39702995247220707, + "grad_norm": 0.8402003645896912, + "learning_rate": 0.00013182915745396434, + "loss": 0.0833, + "step": 16937 + }, + { + "epoch": 0.39705339404701206, + "grad_norm": 0.931599497795105, + "learning_rate": 0.0001318221743025451, + "loss": 0.1051, + "step": 16938 + }, + { + "epoch": 0.39707683562181706, + "grad_norm": 0.5166149139404297, + "learning_rate": 0.0001318151909784579, + "loss": 0.1131, + "step": 16939 + }, + { + "epoch": 0.39710027719662205, + "grad_norm": 0.5324820280075073, + "learning_rate": 0.00013180820748174055, + "loss": 0.0899, + "step": 16940 + }, + { + "epoch": 0.39712371877142705, + "grad_norm": 0.16598738729953766, + "learning_rate": 0.00013180122381243096, + "loss": 0.0153, + "step": 16941 + }, + { + "epoch": 0.39714716034623204, + "grad_norm": 0.3525296151638031, + "learning_rate": 0.0001317942399705671, + "loss": 0.0742, + "step": 16942 + }, + { + "epoch": 0.39717060192103704, + "grad_norm": 0.5914868712425232, + "learning_rate": 0.0001317872559561867, + "loss": 0.0963, + "step": 16943 + }, + { + "epoch": 0.39719404349584203, + "grad_norm": 0.4113547205924988, + "learning_rate": 0.00013178027176932782, + "loss": 0.0892, + "step": 16944 + }, + { + "epoch": 0.397217485070647, + "grad_norm": 0.19717949628829956, + "learning_rate": 0.00013177328741002827, + "loss": 0.0389, + "step": 16945 + }, + { + "epoch": 0.397240926645452, + "grad_norm": 0.4806535243988037, + "learning_rate": 0.000131766302878326, + "loss": 0.117, + "step": 16946 + }, + { + "epoch": 0.397264368220257, + "grad_norm": 0.505115270614624, + "learning_rate": 0.00013175931817425887, + "loss": 0.0901, + "step": 16947 + }, + { + "epoch": 0.397287809795062, + "grad_norm": 0.5090712308883667, + "learning_rate": 0.00013175233329786477, + "loss": 0.0933, + "step": 16948 + }, + { + "epoch": 0.397311251369867, + "grad_norm": 0.5744041204452515, + "learning_rate": 0.00013174534824918161, + "loss": 0.092, + "step": 16949 + }, + { + "epoch": 0.397334692944672, + "grad_norm": 1.0076420307159424, + "learning_rate": 0.00013173836302824728, + "loss": 0.1606, + "step": 16950 + }, + { + "epoch": 0.397358134519477, + "grad_norm": 0.5928823351860046, + "learning_rate": 0.00013173137763509976, + "loss": 0.114, + "step": 16951 + }, + { + "epoch": 0.39738157609428204, + "grad_norm": 0.757466197013855, + "learning_rate": 0.00013172439206977687, + "loss": 0.0823, + "step": 16952 + }, + { + "epoch": 0.39740501766908704, + "grad_norm": 0.5523897409439087, + "learning_rate": 0.0001317174063323165, + "loss": 0.1238, + "step": 16953 + }, + { + "epoch": 0.39742845924389203, + "grad_norm": 0.6497388482093811, + "learning_rate": 0.00013171042042275664, + "loss": 0.1269, + "step": 16954 + }, + { + "epoch": 0.397451900818697, + "grad_norm": 0.46748030185699463, + "learning_rate": 0.0001317034343411351, + "loss": 0.0959, + "step": 16955 + }, + { + "epoch": 0.397475342393502, + "grad_norm": 0.5068477392196655, + "learning_rate": 0.00013169644808748988, + "loss": 0.058, + "step": 16956 + }, + { + "epoch": 0.397498783968307, + "grad_norm": 0.6000596284866333, + "learning_rate": 0.0001316894616618588, + "loss": 0.1498, + "step": 16957 + }, + { + "epoch": 0.397522225543112, + "grad_norm": 1.0345121622085571, + "learning_rate": 0.00013168247506427987, + "loss": 0.0932, + "step": 16958 + }, + { + "epoch": 0.397545667117917, + "grad_norm": 0.35929298400878906, + "learning_rate": 0.0001316754882947909, + "loss": 0.1083, + "step": 16959 + }, + { + "epoch": 0.397569108692722, + "grad_norm": 0.538547694683075, + "learning_rate": 0.00013166850135342984, + "loss": 0.1405, + "step": 16960 + }, + { + "epoch": 0.397592550267527, + "grad_norm": 0.629082202911377, + "learning_rate": 0.00013166151424023455, + "loss": 0.0832, + "step": 16961 + }, + { + "epoch": 0.397615991842332, + "grad_norm": 0.37573492527008057, + "learning_rate": 0.00013165452695524307, + "loss": 0.0287, + "step": 16962 + }, + { + "epoch": 0.397639433417137, + "grad_norm": 0.4739755094051361, + "learning_rate": 0.0001316475394984932, + "loss": 0.1348, + "step": 16963 + }, + { + "epoch": 0.397662874991942, + "grad_norm": 0.2476218342781067, + "learning_rate": 0.0001316405518700229, + "loss": 0.0603, + "step": 16964 + }, + { + "epoch": 0.39768631656674697, + "grad_norm": 0.4196135103702545, + "learning_rate": 0.0001316335640698701, + "loss": 0.1048, + "step": 16965 + }, + { + "epoch": 0.39770975814155196, + "grad_norm": 0.2841714918613434, + "learning_rate": 0.00013162657609807268, + "loss": 0.0362, + "step": 16966 + }, + { + "epoch": 0.39773319971635696, + "grad_norm": 0.8099117875099182, + "learning_rate": 0.00013161958795466854, + "loss": 0.176, + "step": 16967 + }, + { + "epoch": 0.39775664129116195, + "grad_norm": 0.5556643009185791, + "learning_rate": 0.00013161259963969564, + "loss": 0.1637, + "step": 16968 + }, + { + "epoch": 0.39778008286596694, + "grad_norm": 0.36567723751068115, + "learning_rate": 0.0001316056111531919, + "loss": 0.0468, + "step": 16969 + }, + { + "epoch": 0.39780352444077194, + "grad_norm": 0.5165386199951172, + "learning_rate": 0.00013159862249519522, + "loss": 0.1319, + "step": 16970 + }, + { + "epoch": 0.39782696601557693, + "grad_norm": 0.29546162486076355, + "learning_rate": 0.00013159163366574354, + "loss": 0.0703, + "step": 16971 + }, + { + "epoch": 0.3978504075903819, + "grad_norm": 0.09143421798944473, + "learning_rate": 0.00013158464466487475, + "loss": 0.0174, + "step": 16972 + }, + { + "epoch": 0.3978738491651869, + "grad_norm": 0.6870927214622498, + "learning_rate": 0.00013157765549262678, + "loss": 0.7283, + "step": 16973 + }, + { + "epoch": 0.3978972907399919, + "grad_norm": 0.22777973115444183, + "learning_rate": 0.00013157066614903756, + "loss": 0.018, + "step": 16974 + }, + { + "epoch": 0.3979207323147969, + "grad_norm": 0.446524441242218, + "learning_rate": 0.00013156367663414505, + "loss": 0.035, + "step": 16975 + }, + { + "epoch": 0.3979441738896019, + "grad_norm": 0.988972544670105, + "learning_rate": 0.0001315566869479871, + "loss": 0.1235, + "step": 16976 + }, + { + "epoch": 0.3979676154644069, + "grad_norm": 0.3385685980319977, + "learning_rate": 0.00013154969709060172, + "loss": 0.0786, + "step": 16977 + }, + { + "epoch": 0.3979910570392119, + "grad_norm": 0.5274783968925476, + "learning_rate": 0.0001315427070620268, + "loss": 0.0741, + "step": 16978 + }, + { + "epoch": 0.3980144986140169, + "grad_norm": 0.21005284786224365, + "learning_rate": 0.00013153571686230024, + "loss": 0.0176, + "step": 16979 + }, + { + "epoch": 0.3980379401888219, + "grad_norm": 0.3564129173755646, + "learning_rate": 0.00013152872649145999, + "loss": 0.0888, + "step": 16980 + }, + { + "epoch": 0.3980613817636269, + "grad_norm": 0.6981661319732666, + "learning_rate": 0.00013152173594954399, + "loss": 0.1227, + "step": 16981 + }, + { + "epoch": 0.39808482333843187, + "grad_norm": 0.8803348541259766, + "learning_rate": 0.00013151474523659016, + "loss": 0.1683, + "step": 16982 + }, + { + "epoch": 0.39810826491323686, + "grad_norm": 0.12940320372581482, + "learning_rate": 0.00013150775435263645, + "loss": 0.0138, + "step": 16983 + }, + { + "epoch": 0.39813170648804186, + "grad_norm": 0.5971940755844116, + "learning_rate": 0.0001315007632977208, + "loss": 0.1329, + "step": 16984 + }, + { + "epoch": 0.39815514806284685, + "grad_norm": 0.1933118849992752, + "learning_rate": 0.00013149377207188106, + "loss": 0.0331, + "step": 16985 + }, + { + "epoch": 0.39817858963765185, + "grad_norm": 0.5924621820449829, + "learning_rate": 0.00013148678067515524, + "loss": 0.1437, + "step": 16986 + }, + { + "epoch": 0.39820203121245684, + "grad_norm": 0.48721036314964294, + "learning_rate": 0.00013147978910758133, + "loss": 0.0878, + "step": 16987 + }, + { + "epoch": 0.39822547278726184, + "grad_norm": 0.5480307340621948, + "learning_rate": 0.0001314727973691971, + "loss": 0.0767, + "step": 16988 + }, + { + "epoch": 0.39824891436206683, + "grad_norm": 0.5606898665428162, + "learning_rate": 0.00013146580546004067, + "loss": 0.0714, + "step": 16989 + }, + { + "epoch": 0.3982723559368718, + "grad_norm": 0.5716010928153992, + "learning_rate": 0.00013145881338014987, + "loss": 0.1385, + "step": 16990 + }, + { + "epoch": 0.3982957975116768, + "grad_norm": 0.6367118954658508, + "learning_rate": 0.00013145182112956266, + "loss": 0.1157, + "step": 16991 + }, + { + "epoch": 0.3983192390864818, + "grad_norm": 0.5923522710800171, + "learning_rate": 0.00013144482870831696, + "loss": 0.1262, + "step": 16992 + }, + { + "epoch": 0.3983426806612868, + "grad_norm": 0.382228821516037, + "learning_rate": 0.00013143783611645077, + "loss": 0.0646, + "step": 16993 + }, + { + "epoch": 0.3983661222360918, + "grad_norm": 0.39631447196006775, + "learning_rate": 0.000131430843354002, + "loss": 0.0864, + "step": 16994 + }, + { + "epoch": 0.3983895638108968, + "grad_norm": 0.47327327728271484, + "learning_rate": 0.00013142385042100855, + "loss": 0.09, + "step": 16995 + }, + { + "epoch": 0.3984130053857018, + "grad_norm": 0.6373918652534485, + "learning_rate": 0.00013141685731750847, + "loss": 0.1075, + "step": 16996 + }, + { + "epoch": 0.3984364469605068, + "grad_norm": 0.21803916990756989, + "learning_rate": 0.0001314098640435396, + "loss": 0.0374, + "step": 16997 + }, + { + "epoch": 0.3984598885353118, + "grad_norm": 0.3157896399497986, + "learning_rate": 0.00013140287059913987, + "loss": 0.0954, + "step": 16998 + }, + { + "epoch": 0.3984833301101168, + "grad_norm": 0.3725642263889313, + "learning_rate": 0.00013139587698434734, + "loss": 0.0469, + "step": 16999 + }, + { + "epoch": 0.39850677168492177, + "grad_norm": 0.42142346501350403, + "learning_rate": 0.00013138888319919994, + "loss": 0.0891, + "step": 17000 + }, + { + "epoch": 0.39853021325972676, + "grad_norm": 0.3342832624912262, + "learning_rate": 0.0001313818892437355, + "loss": 0.1067, + "step": 17001 + }, + { + "epoch": 0.39855365483453176, + "grad_norm": 0.12547306716442108, + "learning_rate": 0.0001313748951179921, + "loss": 0.0223, + "step": 17002 + }, + { + "epoch": 0.3985770964093368, + "grad_norm": 0.9220834970474243, + "learning_rate": 0.0001313679008220076, + "loss": 0.1408, + "step": 17003 + }, + { + "epoch": 0.3986005379841418, + "grad_norm": 0.1373245120048523, + "learning_rate": 0.00013136090635582, + "loss": 0.0253, + "step": 17004 + }, + { + "epoch": 0.3986239795589468, + "grad_norm": 0.5084431767463684, + "learning_rate": 0.00013135391171946722, + "loss": 0.0721, + "step": 17005 + }, + { + "epoch": 0.3986474211337518, + "grad_norm": 0.47085919976234436, + "learning_rate": 0.00013134691691298728, + "loss": 0.1301, + "step": 17006 + }, + { + "epoch": 0.3986708627085568, + "grad_norm": 0.5009521842002869, + "learning_rate": 0.00013133992193641803, + "loss": 0.1697, + "step": 17007 + }, + { + "epoch": 0.3986943042833618, + "grad_norm": 0.5142925977706909, + "learning_rate": 0.00013133292678979753, + "loss": 0.1303, + "step": 17008 + }, + { + "epoch": 0.39871774585816677, + "grad_norm": 0.21345824003219604, + "learning_rate": 0.0001313259314731637, + "loss": 0.0469, + "step": 17009 + }, + { + "epoch": 0.39874118743297177, + "grad_norm": 0.09597723186016083, + "learning_rate": 0.00013131893598655444, + "loss": 0.03, + "step": 17010 + }, + { + "epoch": 0.39876462900777676, + "grad_norm": 0.07349653542041779, + "learning_rate": 0.00013131194033000776, + "loss": 0.0117, + "step": 17011 + }, + { + "epoch": 0.39878807058258176, + "grad_norm": 0.5257062315940857, + "learning_rate": 0.00013130494450356163, + "loss": 0.1035, + "step": 17012 + }, + { + "epoch": 0.39881151215738675, + "grad_norm": 0.5910170674324036, + "learning_rate": 0.000131297948507254, + "loss": 0.1581, + "step": 17013 + }, + { + "epoch": 0.39883495373219174, + "grad_norm": 0.44921401143074036, + "learning_rate": 0.0001312909523411228, + "loss": 0.1252, + "step": 17014 + }, + { + "epoch": 0.39885839530699674, + "grad_norm": 0.5313850045204163, + "learning_rate": 0.000131283956005206, + "loss": 0.1021, + "step": 17015 + }, + { + "epoch": 0.39888183688180173, + "grad_norm": 0.6943158507347107, + "learning_rate": 0.0001312769594995416, + "loss": 0.147, + "step": 17016 + }, + { + "epoch": 0.3989052784566067, + "grad_norm": 0.48922058939933777, + "learning_rate": 0.00013126996282416753, + "loss": 0.1818, + "step": 17017 + }, + { + "epoch": 0.3989287200314117, + "grad_norm": 0.3987615704536438, + "learning_rate": 0.00013126296597912174, + "loss": 0.1221, + "step": 17018 + }, + { + "epoch": 0.3989521616062167, + "grad_norm": 0.23984485864639282, + "learning_rate": 0.00013125596896444228, + "loss": 0.0668, + "step": 17019 + }, + { + "epoch": 0.3989756031810217, + "grad_norm": 0.6858865022659302, + "learning_rate": 0.00013124897178016698, + "loss": 0.1792, + "step": 17020 + }, + { + "epoch": 0.3989990447558267, + "grad_norm": 0.4399375915527344, + "learning_rate": 0.00013124197442633396, + "loss": 0.1047, + "step": 17021 + }, + { + "epoch": 0.3990224863306317, + "grad_norm": 0.2969346344470978, + "learning_rate": 0.00013123497690298105, + "loss": 0.0456, + "step": 17022 + }, + { + "epoch": 0.3990459279054367, + "grad_norm": 0.4217011034488678, + "learning_rate": 0.0001312279792101463, + "loss": 0.4934, + "step": 17023 + }, + { + "epoch": 0.3990693694802417, + "grad_norm": 0.600462794303894, + "learning_rate": 0.0001312209813478676, + "loss": 0.5927, + "step": 17024 + }, + { + "epoch": 0.3990928110550467, + "grad_norm": 0.2776821255683899, + "learning_rate": 0.00013121398331618307, + "loss": 0.0999, + "step": 17025 + }, + { + "epoch": 0.3991162526298517, + "grad_norm": 0.12329195439815521, + "learning_rate": 0.00013120698511513053, + "loss": 0.0285, + "step": 17026 + }, + { + "epoch": 0.39913969420465667, + "grad_norm": 0.6709075570106506, + "learning_rate": 0.00013119998674474803, + "loss": 0.1248, + "step": 17027 + }, + { + "epoch": 0.39916313577946166, + "grad_norm": 0.46770232915878296, + "learning_rate": 0.00013119298820507352, + "loss": 0.0973, + "step": 17028 + }, + { + "epoch": 0.39918657735426666, + "grad_norm": 0.31007465720176697, + "learning_rate": 0.00013118598949614499, + "loss": 0.0641, + "step": 17029 + }, + { + "epoch": 0.39921001892907165, + "grad_norm": 0.5123015642166138, + "learning_rate": 0.00013117899061800036, + "loss": 0.0941, + "step": 17030 + }, + { + "epoch": 0.39923346050387665, + "grad_norm": 0.22553543746471405, + "learning_rate": 0.0001311719915706777, + "loss": 0.0468, + "step": 17031 + }, + { + "epoch": 0.39925690207868164, + "grad_norm": 0.13877001404762268, + "learning_rate": 0.00013116499235421493, + "loss": 0.0306, + "step": 17032 + }, + { + "epoch": 0.39928034365348664, + "grad_norm": 0.33196181058883667, + "learning_rate": 0.00013115799296865, + "loss": 0.0651, + "step": 17033 + }, + { + "epoch": 0.39930378522829163, + "grad_norm": 0.21815502643585205, + "learning_rate": 0.00013115099341402096, + "loss": 0.0289, + "step": 17034 + }, + { + "epoch": 0.3993272268030966, + "grad_norm": 0.35808026790618896, + "learning_rate": 0.00013114399369036574, + "loss": 0.0633, + "step": 17035 + }, + { + "epoch": 0.3993506683779016, + "grad_norm": 0.37518155574798584, + "learning_rate": 0.00013113699379772232, + "loss": 0.0339, + "step": 17036 + }, + { + "epoch": 0.3993741099527066, + "grad_norm": 0.5548306703567505, + "learning_rate": 0.00013112999373612874, + "loss": 0.1117, + "step": 17037 + }, + { + "epoch": 0.3993975515275116, + "grad_norm": 0.392483115196228, + "learning_rate": 0.0001311229935056229, + "loss": 0.0726, + "step": 17038 + }, + { + "epoch": 0.3994209931023166, + "grad_norm": 0.5383638739585876, + "learning_rate": 0.00013111599310624284, + "loss": 0.4466, + "step": 17039 + }, + { + "epoch": 0.3994444346771216, + "grad_norm": 0.8013273477554321, + "learning_rate": 0.00013110899253802652, + "loss": 0.1716, + "step": 17040 + }, + { + "epoch": 0.3994678762519266, + "grad_norm": 0.5695348978042603, + "learning_rate": 0.00013110199180101195, + "loss": 0.166, + "step": 17041 + }, + { + "epoch": 0.3994913178267316, + "grad_norm": 0.39730384945869446, + "learning_rate": 0.00013109499089523707, + "loss": 0.057, + "step": 17042 + }, + { + "epoch": 0.3995147594015366, + "grad_norm": 0.2879921495914459, + "learning_rate": 0.0001310879898207399, + "loss": 0.0413, + "step": 17043 + }, + { + "epoch": 0.3995382009763416, + "grad_norm": 0.5909773707389832, + "learning_rate": 0.00013108098857755844, + "loss": 0.116, + "step": 17044 + }, + { + "epoch": 0.39956164255114657, + "grad_norm": 0.4136868715286255, + "learning_rate": 0.00013107398716573064, + "loss": 0.0764, + "step": 17045 + }, + { + "epoch": 0.39958508412595156, + "grad_norm": 0.8340376615524292, + "learning_rate": 0.0001310669855852945, + "loss": 0.1694, + "step": 17046 + }, + { + "epoch": 0.39960852570075656, + "grad_norm": 0.3179488778114319, + "learning_rate": 0.0001310599838362881, + "loss": 0.0564, + "step": 17047 + }, + { + "epoch": 0.39963196727556155, + "grad_norm": 0.39942190051078796, + "learning_rate": 0.0001310529819187493, + "loss": 0.1069, + "step": 17048 + }, + { + "epoch": 0.39965540885036654, + "grad_norm": 0.6384974122047424, + "learning_rate": 0.00013104597983271612, + "loss": 0.6478, + "step": 17049 + }, + { + "epoch": 0.39967885042517154, + "grad_norm": 0.42658549547195435, + "learning_rate": 0.0001310389775782266, + "loss": 0.1056, + "step": 17050 + }, + { + "epoch": 0.39970229199997653, + "grad_norm": 0.8193593621253967, + "learning_rate": 0.00013103197515531876, + "loss": 0.1529, + "step": 17051 + }, + { + "epoch": 0.3997257335747815, + "grad_norm": 0.2646681070327759, + "learning_rate": 0.00013102497256403046, + "loss": 0.0341, + "step": 17052 + }, + { + "epoch": 0.3997491751495865, + "grad_norm": 0.47310832142829895, + "learning_rate": 0.00013101796980439985, + "loss": 0.0484, + "step": 17053 + }, + { + "epoch": 0.39977261672439157, + "grad_norm": 0.4077851176261902, + "learning_rate": 0.00013101096687646489, + "loss": 0.0894, + "step": 17054 + }, + { + "epoch": 0.39979605829919657, + "grad_norm": 0.3297501802444458, + "learning_rate": 0.0001310039637802635, + "loss": 0.0884, + "step": 17055 + }, + { + "epoch": 0.39981949987400156, + "grad_norm": 0.22626756131649017, + "learning_rate": 0.00013099696051583376, + "loss": 0.0292, + "step": 17056 + }, + { + "epoch": 0.39984294144880655, + "grad_norm": 0.6393819451332092, + "learning_rate": 0.00013098995708321363, + "loss": 0.1967, + "step": 17057 + }, + { + "epoch": 0.39986638302361155, + "grad_norm": 0.2852841317653656, + "learning_rate": 0.0001309829534824411, + "loss": 0.0651, + "step": 17058 + }, + { + "epoch": 0.39988982459841654, + "grad_norm": 0.6591274738311768, + "learning_rate": 0.0001309759497135542, + "loss": 0.1164, + "step": 17059 + }, + { + "epoch": 0.39991326617322154, + "grad_norm": 0.32841557264328003, + "learning_rate": 0.00013096894577659094, + "loss": 0.0435, + "step": 17060 + }, + { + "epoch": 0.39993670774802653, + "grad_norm": 0.16758771240711212, + "learning_rate": 0.0001309619416715893, + "loss": 0.0291, + "step": 17061 + }, + { + "epoch": 0.3999601493228315, + "grad_norm": 0.6580414175987244, + "learning_rate": 0.0001309549373985873, + "loss": 0.1012, + "step": 17062 + }, + { + "epoch": 0.3999835908976365, + "grad_norm": 0.6399386525154114, + "learning_rate": 0.00013094793295762296, + "loss": 0.1411, + "step": 17063 + }, + { + "epoch": 0.4000070324724415, + "grad_norm": 0.3743577003479004, + "learning_rate": 0.00013094092834873425, + "loss": 0.0485, + "step": 17064 + }, + { + "epoch": 0.4000304740472465, + "grad_norm": 0.5568577647209167, + "learning_rate": 0.00013093392357195916, + "loss": 0.1278, + "step": 17065 + }, + { + "epoch": 0.4000539156220515, + "grad_norm": 0.6583091616630554, + "learning_rate": 0.00013092691862733578, + "loss": 0.1264, + "step": 17066 + }, + { + "epoch": 0.4000773571968565, + "grad_norm": 0.14136667549610138, + "learning_rate": 0.00013091991351490205, + "loss": 0.028, + "step": 17067 + }, + { + "epoch": 0.4001007987716615, + "grad_norm": 0.07418894022703171, + "learning_rate": 0.00013091290823469598, + "loss": 0.013, + "step": 17068 + }, + { + "epoch": 0.4001242403464665, + "grad_norm": 0.6722462177276611, + "learning_rate": 0.00013090590278675562, + "loss": 0.0798, + "step": 17069 + }, + { + "epoch": 0.4001476819212715, + "grad_norm": 0.21566174924373627, + "learning_rate": 0.00013089889717111898, + "loss": 0.0371, + "step": 17070 + }, + { + "epoch": 0.4001711234960765, + "grad_norm": 0.37214362621307373, + "learning_rate": 0.00013089189138782402, + "loss": 0.0631, + "step": 17071 + }, + { + "epoch": 0.40019456507088147, + "grad_norm": 0.569297730922699, + "learning_rate": 0.0001308848854369088, + "loss": 0.1853, + "step": 17072 + }, + { + "epoch": 0.40021800664568646, + "grad_norm": 0.22605270147323608, + "learning_rate": 0.00013087787931841133, + "loss": 0.0339, + "step": 17073 + }, + { + "epoch": 0.40024144822049146, + "grad_norm": 0.2514129877090454, + "learning_rate": 0.00013087087303236962, + "loss": 0.046, + "step": 17074 + }, + { + "epoch": 0.40026488979529645, + "grad_norm": 0.1555464267730713, + "learning_rate": 0.0001308638665788217, + "loss": 0.0203, + "step": 17075 + }, + { + "epoch": 0.40028833137010145, + "grad_norm": 0.22633010149002075, + "learning_rate": 0.00013085685995780554, + "loss": 0.0195, + "step": 17076 + }, + { + "epoch": 0.40031177294490644, + "grad_norm": 0.3712191581726074, + "learning_rate": 0.0001308498531693592, + "loss": 0.0593, + "step": 17077 + }, + { + "epoch": 0.40033521451971144, + "grad_norm": 0.3253539502620697, + "learning_rate": 0.0001308428462135207, + "loss": 0.0527, + "step": 17078 + }, + { + "epoch": 0.40035865609451643, + "grad_norm": 0.32086920738220215, + "learning_rate": 0.00013083583909032805, + "loss": 0.0261, + "step": 17079 + }, + { + "epoch": 0.4003820976693214, + "grad_norm": 0.6171104311943054, + "learning_rate": 0.00013082883179981928, + "loss": 0.6852, + "step": 17080 + }, + { + "epoch": 0.4004055392441264, + "grad_norm": 0.7894812226295471, + "learning_rate": 0.00013082182434203237, + "loss": 0.6678, + "step": 17081 + }, + { + "epoch": 0.4004289808189314, + "grad_norm": 0.2870391309261322, + "learning_rate": 0.0001308148167170054, + "loss": 0.0344, + "step": 17082 + }, + { + "epoch": 0.4004524223937364, + "grad_norm": 0.5210778117179871, + "learning_rate": 0.00013080780892477635, + "loss": 0.1557, + "step": 17083 + }, + { + "epoch": 0.4004758639685414, + "grad_norm": 0.1407012939453125, + "learning_rate": 0.00013080080096538325, + "loss": 0.0281, + "step": 17084 + }, + { + "epoch": 0.4004993055433464, + "grad_norm": 0.23954758048057556, + "learning_rate": 0.00013079379283886417, + "loss": 0.0356, + "step": 17085 + }, + { + "epoch": 0.4005227471181514, + "grad_norm": 0.2887277603149414, + "learning_rate": 0.0001307867845452571, + "loss": 0.0698, + "step": 17086 + }, + { + "epoch": 0.4005461886929564, + "grad_norm": 0.4337660074234009, + "learning_rate": 0.00013077977608460007, + "loss": 0.1084, + "step": 17087 + }, + { + "epoch": 0.4005696302677614, + "grad_norm": 0.570869505405426, + "learning_rate": 0.0001307727674569311, + "loss": 0.1189, + "step": 17088 + }, + { + "epoch": 0.4005930718425664, + "grad_norm": 0.5555287599563599, + "learning_rate": 0.0001307657586622882, + "loss": 0.1021, + "step": 17089 + }, + { + "epoch": 0.40061651341737137, + "grad_norm": 0.39334240555763245, + "learning_rate": 0.00013075874970070947, + "loss": 0.0774, + "step": 17090 + }, + { + "epoch": 0.40063995499217636, + "grad_norm": 0.1744498610496521, + "learning_rate": 0.0001307517405722329, + "loss": 0.0257, + "step": 17091 + }, + { + "epoch": 0.40066339656698136, + "grad_norm": 0.6684921383857727, + "learning_rate": 0.00013074473127689653, + "loss": 0.1628, + "step": 17092 + }, + { + "epoch": 0.40068683814178635, + "grad_norm": 0.21216942369937897, + "learning_rate": 0.00013073772181473834, + "loss": 0.0435, + "step": 17093 + }, + { + "epoch": 0.40071027971659134, + "grad_norm": 0.5180004239082336, + "learning_rate": 0.00013073071218579644, + "loss": 0.0538, + "step": 17094 + }, + { + "epoch": 0.40073372129139634, + "grad_norm": 0.4541141986846924, + "learning_rate": 0.0001307237023901088, + "loss": 0.122, + "step": 17095 + }, + { + "epoch": 0.40075716286620133, + "grad_norm": 0.7050163149833679, + "learning_rate": 0.00013071669242771354, + "loss": 0.1131, + "step": 17096 + }, + { + "epoch": 0.4007806044410063, + "grad_norm": 1.275110125541687, + "learning_rate": 0.00013070968229864858, + "loss": 0.244, + "step": 17097 + }, + { + "epoch": 0.4008040460158113, + "grad_norm": 0.15869973599910736, + "learning_rate": 0.00013070267200295207, + "loss": 0.0457, + "step": 17098 + }, + { + "epoch": 0.4008274875906163, + "grad_norm": 0.5992064476013184, + "learning_rate": 0.000130695661540662, + "loss": 0.1103, + "step": 17099 + }, + { + "epoch": 0.4008509291654213, + "grad_norm": 0.5364821553230286, + "learning_rate": 0.0001306886509118164, + "loss": 0.6373, + "step": 17100 + }, + { + "epoch": 0.4008743707402263, + "grad_norm": 0.3803003430366516, + "learning_rate": 0.00013068164011645328, + "loss": 0.0653, + "step": 17101 + }, + { + "epoch": 0.4008978123150313, + "grad_norm": 0.6799290776252747, + "learning_rate": 0.00013067462915461076, + "loss": 0.1836, + "step": 17102 + }, + { + "epoch": 0.4009212538898363, + "grad_norm": 0.27570441365242004, + "learning_rate": 0.0001306676180263268, + "loss": 0.0371, + "step": 17103 + }, + { + "epoch": 0.4009446954646413, + "grad_norm": 0.7947724461555481, + "learning_rate": 0.0001306606067316395, + "loss": 0.1688, + "step": 17104 + }, + { + "epoch": 0.40096813703944634, + "grad_norm": 0.5301347374916077, + "learning_rate": 0.00013065359527058692, + "loss": 0.1368, + "step": 17105 + }, + { + "epoch": 0.40099157861425133, + "grad_norm": 0.39507919549942017, + "learning_rate": 0.00013064658364320703, + "loss": 0.0515, + "step": 17106 + }, + { + "epoch": 0.4010150201890563, + "grad_norm": 0.1646309196949005, + "learning_rate": 0.00013063957184953793, + "loss": 0.0343, + "step": 17107 + }, + { + "epoch": 0.4010384617638613, + "grad_norm": 0.5362452268600464, + "learning_rate": 0.00013063255988961765, + "loss": 0.0683, + "step": 17108 + }, + { + "epoch": 0.4010619033386663, + "grad_norm": 0.7336210608482361, + "learning_rate": 0.00013062554776348425, + "loss": 0.1129, + "step": 17109 + }, + { + "epoch": 0.4010853449134713, + "grad_norm": 0.3846304416656494, + "learning_rate": 0.00013061853547117574, + "loss": 0.0888, + "step": 17110 + }, + { + "epoch": 0.4011087864882763, + "grad_norm": 0.5665589570999146, + "learning_rate": 0.00013061152301273023, + "loss": 0.1316, + "step": 17111 + }, + { + "epoch": 0.4011322280630813, + "grad_norm": 0.6394941806793213, + "learning_rate": 0.00013060451038818572, + "loss": 0.0722, + "step": 17112 + }, + { + "epoch": 0.4011556696378863, + "grad_norm": 0.24860619008541107, + "learning_rate": 0.00013059749759758023, + "loss": 0.0747, + "step": 17113 + }, + { + "epoch": 0.4011791112126913, + "grad_norm": 0.6463137269020081, + "learning_rate": 0.00013059048464095194, + "loss": 0.6199, + "step": 17114 + }, + { + "epoch": 0.4012025527874963, + "grad_norm": 0.40066975355148315, + "learning_rate": 0.00013058347151833876, + "loss": 0.0469, + "step": 17115 + }, + { + "epoch": 0.4012259943623013, + "grad_norm": 0.46217185258865356, + "learning_rate": 0.00013057645822977883, + "loss": 0.0934, + "step": 17116 + }, + { + "epoch": 0.40124943593710627, + "grad_norm": 0.4837738275527954, + "learning_rate": 0.00013056944477531017, + "loss": 0.0919, + "step": 17117 + }, + { + "epoch": 0.40127287751191126, + "grad_norm": 0.6262598037719727, + "learning_rate": 0.00013056243115497083, + "loss": 0.0788, + "step": 17118 + }, + { + "epoch": 0.40129631908671626, + "grad_norm": 0.6706497073173523, + "learning_rate": 0.00013055541736879892, + "loss": 0.612, + "step": 17119 + }, + { + "epoch": 0.40131976066152125, + "grad_norm": 0.4685865044593811, + "learning_rate": 0.0001305484034168324, + "loss": 0.1374, + "step": 17120 + }, + { + "epoch": 0.40134320223632625, + "grad_norm": 0.058593884110450745, + "learning_rate": 0.00013054138929910944, + "loss": 0.0069, + "step": 17121 + }, + { + "epoch": 0.40136664381113124, + "grad_norm": 0.4446004629135132, + "learning_rate": 0.000130534375015668, + "loss": 0.1357, + "step": 17122 + }, + { + "epoch": 0.40139008538593623, + "grad_norm": 0.37616220116615295, + "learning_rate": 0.00013052736056654623, + "loss": 0.1093, + "step": 17123 + }, + { + "epoch": 0.40141352696074123, + "grad_norm": 0.42198750376701355, + "learning_rate": 0.0001305203459517821, + "loss": 0.0471, + "step": 17124 + }, + { + "epoch": 0.4014369685355462, + "grad_norm": 0.630128800868988, + "learning_rate": 0.0001305133311714137, + "loss": 0.1669, + "step": 17125 + }, + { + "epoch": 0.4014604101103512, + "grad_norm": 0.5360876321792603, + "learning_rate": 0.00013050631622547917, + "loss": 0.1186, + "step": 17126 + }, + { + "epoch": 0.4014838516851562, + "grad_norm": 0.20620496571063995, + "learning_rate": 0.00013049930111401646, + "loss": 0.0284, + "step": 17127 + }, + { + "epoch": 0.4015072932599612, + "grad_norm": 0.5876173973083496, + "learning_rate": 0.0001304922858370637, + "loss": 0.1128, + "step": 17128 + }, + { + "epoch": 0.4015307348347662, + "grad_norm": 0.17463156580924988, + "learning_rate": 0.00013048527039465896, + "loss": 0.0352, + "step": 17129 + }, + { + "epoch": 0.4015541764095712, + "grad_norm": 0.5464040637016296, + "learning_rate": 0.00013047825478684026, + "loss": 0.1568, + "step": 17130 + }, + { + "epoch": 0.4015776179843762, + "grad_norm": 0.2589206099510193, + "learning_rate": 0.00013047123901364569, + "loss": 0.0849, + "step": 17131 + }, + { + "epoch": 0.4016010595591812, + "grad_norm": 0.40571853518486023, + "learning_rate": 0.00013046422307511332, + "loss": 0.1099, + "step": 17132 + }, + { + "epoch": 0.4016245011339862, + "grad_norm": 0.13820193707942963, + "learning_rate": 0.00013045720697128123, + "loss": 0.027, + "step": 17133 + }, + { + "epoch": 0.4016479427087912, + "grad_norm": 0.45674312114715576, + "learning_rate": 0.0001304501907021875, + "loss": 0.1069, + "step": 17134 + }, + { + "epoch": 0.40167138428359617, + "grad_norm": 0.15299999713897705, + "learning_rate": 0.00013044317426787015, + "loss": 0.0407, + "step": 17135 + }, + { + "epoch": 0.40169482585840116, + "grad_norm": 0.666644811630249, + "learning_rate": 0.00013043615766836726, + "loss": 0.5375, + "step": 17136 + }, + { + "epoch": 0.40171826743320616, + "grad_norm": 0.08095648884773254, + "learning_rate": 0.00013042914090371697, + "loss": 0.0167, + "step": 17137 + }, + { + "epoch": 0.40174170900801115, + "grad_norm": 0.5427675843238831, + "learning_rate": 0.00013042212397395726, + "loss": 0.1236, + "step": 17138 + }, + { + "epoch": 0.40176515058281614, + "grad_norm": 0.6217243671417236, + "learning_rate": 0.00013041510687912628, + "loss": 0.1558, + "step": 17139 + }, + { + "epoch": 0.40178859215762114, + "grad_norm": 0.26245447993278503, + "learning_rate": 0.0001304080896192621, + "loss": 0.0486, + "step": 17140 + }, + { + "epoch": 0.40181203373242613, + "grad_norm": 0.3787027597427368, + "learning_rate": 0.0001304010721944027, + "loss": 0.1003, + "step": 17141 + }, + { + "epoch": 0.4018354753072311, + "grad_norm": 0.20467637479305267, + "learning_rate": 0.0001303940546045863, + "loss": 0.0747, + "step": 17142 + }, + { + "epoch": 0.4018589168820361, + "grad_norm": 0.2821284532546997, + "learning_rate": 0.00013038703684985087, + "loss": 0.075, + "step": 17143 + }, + { + "epoch": 0.4018823584568411, + "grad_norm": 0.9844167828559875, + "learning_rate": 0.00013038001893023448, + "loss": 0.2168, + "step": 17144 + }, + { + "epoch": 0.4019058000316461, + "grad_norm": 0.5737846493721008, + "learning_rate": 0.00013037300084577533, + "loss": 0.1304, + "step": 17145 + }, + { + "epoch": 0.4019292416064511, + "grad_norm": 0.2545475363731384, + "learning_rate": 0.0001303659825965114, + "loss": 0.0436, + "step": 17146 + }, + { + "epoch": 0.4019526831812561, + "grad_norm": 0.5156170725822449, + "learning_rate": 0.00013035896418248078, + "loss": 0.1356, + "step": 17147 + }, + { + "epoch": 0.4019761247560611, + "grad_norm": 0.5978088974952698, + "learning_rate": 0.00013035194560372156, + "loss": 0.1523, + "step": 17148 + }, + { + "epoch": 0.4019995663308661, + "grad_norm": 0.5805932283401489, + "learning_rate": 0.00013034492686027187, + "loss": 0.6049, + "step": 17149 + }, + { + "epoch": 0.4020230079056711, + "grad_norm": 0.3655981719493866, + "learning_rate": 0.0001303379079521697, + "loss": 0.1271, + "step": 17150 + }, + { + "epoch": 0.4020464494804761, + "grad_norm": 0.47627419233322144, + "learning_rate": 0.0001303308888794532, + "loss": 0.1322, + "step": 17151 + }, + { + "epoch": 0.40206989105528107, + "grad_norm": 0.3792097568511963, + "learning_rate": 0.00013032386964216048, + "loss": 0.0709, + "step": 17152 + }, + { + "epoch": 0.40209333263008606, + "grad_norm": 0.47885289788246155, + "learning_rate": 0.0001303168502403296, + "loss": 0.067, + "step": 17153 + }, + { + "epoch": 0.40211677420489106, + "grad_norm": 0.09976229071617126, + "learning_rate": 0.00013030983067399858, + "loss": 0.0153, + "step": 17154 + }, + { + "epoch": 0.40214021577969605, + "grad_norm": 0.389239639043808, + "learning_rate": 0.0001303028109432056, + "loss": 0.0698, + "step": 17155 + }, + { + "epoch": 0.4021636573545011, + "grad_norm": 0.8738777041435242, + "learning_rate": 0.0001302957910479887, + "loss": 0.6353, + "step": 17156 + }, + { + "epoch": 0.4021870989293061, + "grad_norm": 0.508081316947937, + "learning_rate": 0.00013028877098838598, + "loss": 0.4955, + "step": 17157 + }, + { + "epoch": 0.4022105405041111, + "grad_norm": 0.31330832839012146, + "learning_rate": 0.00013028175076443559, + "loss": 0.0591, + "step": 17158 + }, + { + "epoch": 0.4022339820789161, + "grad_norm": 0.3898457884788513, + "learning_rate": 0.00013027473037617552, + "loss": 0.0842, + "step": 17159 + }, + { + "epoch": 0.4022574236537211, + "grad_norm": 0.21768279373645782, + "learning_rate": 0.00013026770982364392, + "loss": 0.0563, + "step": 17160 + }, + { + "epoch": 0.4022808652285261, + "grad_norm": 0.5053689479827881, + "learning_rate": 0.0001302606891068789, + "loss": 0.0888, + "step": 17161 + }, + { + "epoch": 0.40230430680333107, + "grad_norm": 0.6176466345787048, + "learning_rate": 0.0001302536682259185, + "loss": 0.0959, + "step": 17162 + }, + { + "epoch": 0.40232774837813606, + "grad_norm": 0.3506530821323395, + "learning_rate": 0.00013024664718080085, + "loss": 0.055, + "step": 17163 + }, + { + "epoch": 0.40235118995294106, + "grad_norm": 0.5936071276664734, + "learning_rate": 0.00013023962597156404, + "loss": 0.0978, + "step": 17164 + }, + { + "epoch": 0.40237463152774605, + "grad_norm": 0.42867210507392883, + "learning_rate": 0.00013023260459824618, + "loss": 0.0869, + "step": 17165 + }, + { + "epoch": 0.40239807310255105, + "grad_norm": 0.6436030864715576, + "learning_rate": 0.00013022558306088536, + "loss": 0.1813, + "step": 17166 + }, + { + "epoch": 0.40242151467735604, + "grad_norm": 0.2686777710914612, + "learning_rate": 0.00013021856135951967, + "loss": 0.0476, + "step": 17167 + }, + { + "epoch": 0.40244495625216103, + "grad_norm": 0.13842658698558807, + "learning_rate": 0.0001302115394941872, + "loss": 0.0304, + "step": 17168 + }, + { + "epoch": 0.40246839782696603, + "grad_norm": 0.509946346282959, + "learning_rate": 0.0001302045174649261, + "loss": 0.1513, + "step": 17169 + }, + { + "epoch": 0.402491839401771, + "grad_norm": 0.6690582036972046, + "learning_rate": 0.0001301974952717744, + "loss": 0.1375, + "step": 17170 + }, + { + "epoch": 0.402515280976576, + "grad_norm": 0.33521682024002075, + "learning_rate": 0.00013019047291477028, + "loss": 0.0953, + "step": 17171 + }, + { + "epoch": 0.402538722551381, + "grad_norm": 0.7303422689437866, + "learning_rate": 0.0001301834503939518, + "loss": 0.2428, + "step": 17172 + }, + { + "epoch": 0.402562164126186, + "grad_norm": 0.2821967899799347, + "learning_rate": 0.00013017642770935705, + "loss": 0.0244, + "step": 17173 + }, + { + "epoch": 0.402585605700991, + "grad_norm": 0.20613332092761993, + "learning_rate": 0.00013016940486102415, + "loss": 0.0404, + "step": 17174 + }, + { + "epoch": 0.402609047275796, + "grad_norm": 0.300475537776947, + "learning_rate": 0.00013016238184899122, + "loss": 0.049, + "step": 17175 + }, + { + "epoch": 0.402632488850601, + "grad_norm": 0.4151492714881897, + "learning_rate": 0.00013015535867329638, + "loss": 0.4938, + "step": 17176 + }, + { + "epoch": 0.402655930425406, + "grad_norm": 0.18191221356391907, + "learning_rate": 0.0001301483353339777, + "loss": 0.0413, + "step": 17177 + }, + { + "epoch": 0.402679372000211, + "grad_norm": 0.5018015503883362, + "learning_rate": 0.00013014131183107332, + "loss": 0.0604, + "step": 17178 + }, + { + "epoch": 0.40270281357501597, + "grad_norm": 0.49409574270248413, + "learning_rate": 0.0001301342881646213, + "loss": 0.1, + "step": 17179 + }, + { + "epoch": 0.40272625514982097, + "grad_norm": 0.7799331545829773, + "learning_rate": 0.00013012726433465977, + "loss": 0.1096, + "step": 17180 + }, + { + "epoch": 0.40274969672462596, + "grad_norm": 0.5718485116958618, + "learning_rate": 0.0001301202403412269, + "loss": 0.104, + "step": 17181 + }, + { + "epoch": 0.40277313829943096, + "grad_norm": 0.4236200451850891, + "learning_rate": 0.00013011321618436076, + "loss": 0.0278, + "step": 17182 + }, + { + "epoch": 0.40279657987423595, + "grad_norm": 0.5898062586784363, + "learning_rate": 0.00013010619186409948, + "loss": 0.0987, + "step": 17183 + }, + { + "epoch": 0.40282002144904094, + "grad_norm": 0.5203379988670349, + "learning_rate": 0.00013009916738048111, + "loss": 0.0621, + "step": 17184 + }, + { + "epoch": 0.40284346302384594, + "grad_norm": 0.2510127127170563, + "learning_rate": 0.00013009214273354382, + "loss": 0.0601, + "step": 17185 + }, + { + "epoch": 0.40286690459865093, + "grad_norm": 0.16011637449264526, + "learning_rate": 0.00013008511792332573, + "loss": 0.0378, + "step": 17186 + }, + { + "epoch": 0.4028903461734559, + "grad_norm": 0.3798019289970398, + "learning_rate": 0.00013007809294986492, + "loss": 0.0348, + "step": 17187 + }, + { + "epoch": 0.4029137877482609, + "grad_norm": 0.15074634552001953, + "learning_rate": 0.00013007106781319958, + "loss": 0.0227, + "step": 17188 + }, + { + "epoch": 0.4029372293230659, + "grad_norm": 0.16953830420970917, + "learning_rate": 0.00013006404251336774, + "loss": 0.0222, + "step": 17189 + }, + { + "epoch": 0.4029606708978709, + "grad_norm": 1.2878130674362183, + "learning_rate": 0.00013005701705040757, + "loss": 0.0638, + "step": 17190 + }, + { + "epoch": 0.4029841124726759, + "grad_norm": 0.210703045129776, + "learning_rate": 0.00013004999142435722, + "loss": 0.0176, + "step": 17191 + }, + { + "epoch": 0.4030075540474809, + "grad_norm": 0.9672706723213196, + "learning_rate": 0.0001300429656352547, + "loss": 0.5214, + "step": 17192 + }, + { + "epoch": 0.4030309956222859, + "grad_norm": 1.525344967842102, + "learning_rate": 0.00013003593968313824, + "loss": 0.1894, + "step": 17193 + }, + { + "epoch": 0.4030544371970909, + "grad_norm": 0.5071808695793152, + "learning_rate": 0.00013002891356804594, + "loss": 0.0731, + "step": 17194 + }, + { + "epoch": 0.4030778787718959, + "grad_norm": 0.5389440655708313, + "learning_rate": 0.00013002188729001587, + "loss": 0.0978, + "step": 17195 + }, + { + "epoch": 0.4031013203467009, + "grad_norm": 0.5655834078788757, + "learning_rate": 0.00013001486084908623, + "loss": 0.113, + "step": 17196 + }, + { + "epoch": 0.40312476192150587, + "grad_norm": 0.8982978463172913, + "learning_rate": 0.00013000783424529512, + "loss": 0.1886, + "step": 17197 + }, + { + "epoch": 0.40314820349631086, + "grad_norm": 0.044278793036937714, + "learning_rate": 0.00013000080747868062, + "loss": 0.0078, + "step": 17198 + }, + { + "epoch": 0.40317164507111586, + "grad_norm": 0.7241149544715881, + "learning_rate": 0.00012999378054928093, + "loss": 0.2481, + "step": 17199 + }, + { + "epoch": 0.40319508664592085, + "grad_norm": 0.34950920939445496, + "learning_rate": 0.00012998675345713413, + "loss": 0.0509, + "step": 17200 + }, + { + "epoch": 0.40321852822072585, + "grad_norm": 0.743185818195343, + "learning_rate": 0.00012997972620227833, + "loss": 0.1302, + "step": 17201 + }, + { + "epoch": 0.40324196979553084, + "grad_norm": 0.8080311417579651, + "learning_rate": 0.00012997269878475175, + "loss": 0.122, + "step": 17202 + }, + { + "epoch": 0.40326541137033584, + "grad_norm": 0.16199329495429993, + "learning_rate": 0.00012996567120459246, + "loss": 0.0383, + "step": 17203 + }, + { + "epoch": 0.40328885294514083, + "grad_norm": 0.3626435399055481, + "learning_rate": 0.00012995864346183856, + "loss": 0.0461, + "step": 17204 + }, + { + "epoch": 0.4033122945199458, + "grad_norm": 0.41651779413223267, + "learning_rate": 0.00012995161555652822, + "loss": 0.0731, + "step": 17205 + }, + { + "epoch": 0.4033357360947508, + "grad_norm": 0.527915894985199, + "learning_rate": 0.0001299445874886996, + "loss": 0.0682, + "step": 17206 + }, + { + "epoch": 0.40335917766955587, + "grad_norm": 0.26251840591430664, + "learning_rate": 0.00012993755925839076, + "loss": 0.0317, + "step": 17207 + }, + { + "epoch": 0.40338261924436086, + "grad_norm": 0.123738594353199, + "learning_rate": 0.0001299305308656399, + "loss": 0.0325, + "step": 17208 + }, + { + "epoch": 0.40340606081916586, + "grad_norm": 0.344868540763855, + "learning_rate": 0.00012992350231048517, + "loss": 0.0798, + "step": 17209 + }, + { + "epoch": 0.40342950239397085, + "grad_norm": 0.4846445620059967, + "learning_rate": 0.00012991647359296467, + "loss": 0.0718, + "step": 17210 + }, + { + "epoch": 0.40345294396877585, + "grad_norm": 0.7585868239402771, + "learning_rate": 0.0001299094447131165, + "loss": 0.1148, + "step": 17211 + }, + { + "epoch": 0.40347638554358084, + "grad_norm": 0.5599116086959839, + "learning_rate": 0.00012990241567097888, + "loss": 0.5002, + "step": 17212 + }, + { + "epoch": 0.40349982711838583, + "grad_norm": 0.7207152843475342, + "learning_rate": 0.0001298953864665899, + "loss": 0.1356, + "step": 17213 + }, + { + "epoch": 0.40352326869319083, + "grad_norm": 0.6205438375473022, + "learning_rate": 0.00012988835709998772, + "loss": 0.7732, + "step": 17214 + }, + { + "epoch": 0.4035467102679958, + "grad_norm": 0.527754545211792, + "learning_rate": 0.00012988132757121048, + "loss": 0.1322, + "step": 17215 + }, + { + "epoch": 0.4035701518428008, + "grad_norm": 0.2041957974433899, + "learning_rate": 0.0001298742978802963, + "loss": 0.042, + "step": 17216 + }, + { + "epoch": 0.4035935934176058, + "grad_norm": 0.7608086466789246, + "learning_rate": 0.00012986726802728332, + "loss": 0.8127, + "step": 17217 + }, + { + "epoch": 0.4036170349924108, + "grad_norm": 0.27576500177383423, + "learning_rate": 0.00012986023801220972, + "loss": 0.088, + "step": 17218 + }, + { + "epoch": 0.4036404765672158, + "grad_norm": 0.5716224312782288, + "learning_rate": 0.00012985320783511365, + "loss": 0.0772, + "step": 17219 + }, + { + "epoch": 0.4036639181420208, + "grad_norm": 0.7881506085395813, + "learning_rate": 0.0001298461774960332, + "loss": 0.0687, + "step": 17220 + }, + { + "epoch": 0.4036873597168258, + "grad_norm": 0.19270582497119904, + "learning_rate": 0.0001298391469950066, + "loss": 0.0638, + "step": 17221 + }, + { + "epoch": 0.4037108012916308, + "grad_norm": 0.19871515035629272, + "learning_rate": 0.00012983211633207193, + "loss": 0.0365, + "step": 17222 + }, + { + "epoch": 0.4037342428664358, + "grad_norm": 0.38846296072006226, + "learning_rate": 0.00012982508550726736, + "loss": 0.545, + "step": 17223 + }, + { + "epoch": 0.40375768444124077, + "grad_norm": 0.3056071996688843, + "learning_rate": 0.000129818054520631, + "loss": 0.0833, + "step": 17224 + }, + { + "epoch": 0.40378112601604577, + "grad_norm": 0.4459742605686188, + "learning_rate": 0.0001298110233722011, + "loss": 0.0793, + "step": 17225 + }, + { + "epoch": 0.40380456759085076, + "grad_norm": 0.34678271412849426, + "learning_rate": 0.0001298039920620157, + "loss": 0.0904, + "step": 17226 + }, + { + "epoch": 0.40382800916565575, + "grad_norm": 0.24441160261631012, + "learning_rate": 0.000129796960590113, + "loss": 0.0389, + "step": 17227 + }, + { + "epoch": 0.40385145074046075, + "grad_norm": 0.16381146013736725, + "learning_rate": 0.0001297899289565312, + "loss": 0.0148, + "step": 17228 + }, + { + "epoch": 0.40387489231526574, + "grad_norm": 0.3363240659236908, + "learning_rate": 0.00012978289716130832, + "loss": 0.0576, + "step": 17229 + }, + { + "epoch": 0.40389833389007074, + "grad_norm": 0.10368172079324722, + "learning_rate": 0.00012977586520448268, + "loss": 0.0155, + "step": 17230 + }, + { + "epoch": 0.40392177546487573, + "grad_norm": 0.46303898096084595, + "learning_rate": 0.00012976883308609235, + "loss": 0.0967, + "step": 17231 + }, + { + "epoch": 0.4039452170396807, + "grad_norm": 0.39666563272476196, + "learning_rate": 0.0001297618008061755, + "loss": 0.0625, + "step": 17232 + }, + { + "epoch": 0.4039686586144857, + "grad_norm": 0.3670358955860138, + "learning_rate": 0.00012975476836477024, + "loss": 0.1063, + "step": 17233 + }, + { + "epoch": 0.4039921001892907, + "grad_norm": 0.4610249996185303, + "learning_rate": 0.00012974773576191482, + "loss": 0.1222, + "step": 17234 + }, + { + "epoch": 0.4040155417640957, + "grad_norm": 0.5753495097160339, + "learning_rate": 0.0001297407029976473, + "loss": 0.135, + "step": 17235 + }, + { + "epoch": 0.4040389833389007, + "grad_norm": 0.8126017451286316, + "learning_rate": 0.0001297336700720059, + "loss": 0.2683, + "step": 17236 + }, + { + "epoch": 0.4040624249137057, + "grad_norm": 0.2862735390663147, + "learning_rate": 0.00012972663698502878, + "loss": 0.2936, + "step": 17237 + }, + { + "epoch": 0.4040858664885107, + "grad_norm": 0.13549743592739105, + "learning_rate": 0.00012971960373675412, + "loss": 0.03, + "step": 17238 + }, + { + "epoch": 0.4041093080633157, + "grad_norm": 0.6089333891868591, + "learning_rate": 0.00012971257032722003, + "loss": 0.448, + "step": 17239 + }, + { + "epoch": 0.4041327496381207, + "grad_norm": 0.37769562005996704, + "learning_rate": 0.00012970553675646472, + "loss": 0.091, + "step": 17240 + }, + { + "epoch": 0.4041561912129257, + "grad_norm": 0.3531971871852875, + "learning_rate": 0.00012969850302452625, + "loss": 0.1312, + "step": 17241 + }, + { + "epoch": 0.40417963278773067, + "grad_norm": 0.7227229475975037, + "learning_rate": 0.00012969146913144298, + "loss": 0.6954, + "step": 17242 + }, + { + "epoch": 0.40420307436253566, + "grad_norm": 0.17435337603092194, + "learning_rate": 0.0001296844350772529, + "loss": 0.0371, + "step": 17243 + }, + { + "epoch": 0.40422651593734066, + "grad_norm": 0.7843717336654663, + "learning_rate": 0.00012967740086199426, + "loss": 0.185, + "step": 17244 + }, + { + "epoch": 0.40424995751214565, + "grad_norm": 0.34096968173980713, + "learning_rate": 0.0001296703664857052, + "loss": 0.0379, + "step": 17245 + }, + { + "epoch": 0.40427339908695065, + "grad_norm": 0.27457284927368164, + "learning_rate": 0.00012966333194842388, + "loss": 0.0322, + "step": 17246 + }, + { + "epoch": 0.40429684066175564, + "grad_norm": 0.4019319713115692, + "learning_rate": 0.00012965629725018851, + "loss": 0.0793, + "step": 17247 + }, + { + "epoch": 0.40432028223656064, + "grad_norm": 0.23273183405399323, + "learning_rate": 0.00012964926239103722, + "loss": 0.0467, + "step": 17248 + }, + { + "epoch": 0.40434372381136563, + "grad_norm": 0.574114978313446, + "learning_rate": 0.0001296422273710082, + "loss": 0.1655, + "step": 17249 + }, + { + "epoch": 0.4043671653861706, + "grad_norm": 0.30811604857444763, + "learning_rate": 0.00012963519219013963, + "loss": 0.059, + "step": 17250 + }, + { + "epoch": 0.4043906069609756, + "grad_norm": 0.7150289416313171, + "learning_rate": 0.00012962815684846972, + "loss": 0.0917, + "step": 17251 + }, + { + "epoch": 0.4044140485357806, + "grad_norm": 0.3778713345527649, + "learning_rate": 0.00012962112134603653, + "loss": 0.1007, + "step": 17252 + }, + { + "epoch": 0.4044374901105856, + "grad_norm": 0.6647605299949646, + "learning_rate": 0.0001296140856828783, + "loss": 0.1149, + "step": 17253 + }, + { + "epoch": 0.4044609316853906, + "grad_norm": 0.5054290890693665, + "learning_rate": 0.00012960704985903325, + "loss": 0.1048, + "step": 17254 + }, + { + "epoch": 0.4044843732601956, + "grad_norm": 0.6602954864501953, + "learning_rate": 0.00012960001387453952, + "loss": 0.1087, + "step": 17255 + }, + { + "epoch": 0.4045078148350006, + "grad_norm": 0.45517203211784363, + "learning_rate": 0.00012959297772943524, + "loss": 0.0879, + "step": 17256 + }, + { + "epoch": 0.4045312564098056, + "grad_norm": 0.5497642755508423, + "learning_rate": 0.00012958594142375865, + "loss": 0.1436, + "step": 17257 + }, + { + "epoch": 0.4045546979846106, + "grad_norm": 0.5310180187225342, + "learning_rate": 0.00012957890495754793, + "loss": 0.1034, + "step": 17258 + }, + { + "epoch": 0.40457813955941563, + "grad_norm": 0.21975769102573395, + "learning_rate": 0.00012957186833084121, + "loss": 0.0441, + "step": 17259 + }, + { + "epoch": 0.4046015811342206, + "grad_norm": 0.11345525085926056, + "learning_rate": 0.00012956483154367672, + "loss": 0.0157, + "step": 17260 + }, + { + "epoch": 0.4046250227090256, + "grad_norm": 0.7727428078651428, + "learning_rate": 0.00012955779459609262, + "loss": 0.175, + "step": 17261 + }, + { + "epoch": 0.4046484642838306, + "grad_norm": 0.46049460768699646, + "learning_rate": 0.00012955075748812708, + "loss": 0.1426, + "step": 17262 + }, + { + "epoch": 0.4046719058586356, + "grad_norm": 0.39420536160469055, + "learning_rate": 0.00012954372021981833, + "loss": 0.1009, + "step": 17263 + }, + { + "epoch": 0.4046953474334406, + "grad_norm": 0.2334183007478714, + "learning_rate": 0.0001295366827912045, + "loss": 0.0495, + "step": 17264 + }, + { + "epoch": 0.4047187890082456, + "grad_norm": 0.2475539594888687, + "learning_rate": 0.00012952964520232378, + "loss": 0.0464, + "step": 17265 + }, + { + "epoch": 0.4047422305830506, + "grad_norm": 0.563612699508667, + "learning_rate": 0.0001295226074532144, + "loss": 0.1876, + "step": 17266 + }, + { + "epoch": 0.4047656721578556, + "grad_norm": 0.6098539233207703, + "learning_rate": 0.00012951556954391452, + "loss": 0.1394, + "step": 17267 + }, + { + "epoch": 0.4047891137326606, + "grad_norm": 0.1704530268907547, + "learning_rate": 0.00012950853147446233, + "loss": 0.0545, + "step": 17268 + }, + { + "epoch": 0.40481255530746557, + "grad_norm": 0.16920439898967743, + "learning_rate": 0.000129501493244896, + "loss": 0.0442, + "step": 17269 + }, + { + "epoch": 0.40483599688227057, + "grad_norm": 0.30296528339385986, + "learning_rate": 0.0001294944548552538, + "loss": 0.0959, + "step": 17270 + }, + { + "epoch": 0.40485943845707556, + "grad_norm": 0.6299143433570862, + "learning_rate": 0.0001294874163055738, + "loss": 0.5865, + "step": 17271 + }, + { + "epoch": 0.40488288003188055, + "grad_norm": 0.9034143686294556, + "learning_rate": 0.00012948037759589424, + "loss": 0.2035, + "step": 17272 + }, + { + "epoch": 0.40490632160668555, + "grad_norm": 0.6056340336799622, + "learning_rate": 0.00012947333872625336, + "loss": 0.2086, + "step": 17273 + }, + { + "epoch": 0.40492976318149054, + "grad_norm": 0.4079430103302002, + "learning_rate": 0.0001294662996966893, + "loss": 0.0441, + "step": 17274 + }, + { + "epoch": 0.40495320475629554, + "grad_norm": 0.16755233705043793, + "learning_rate": 0.00012945926050724027, + "loss": 0.0337, + "step": 17275 + }, + { + "epoch": 0.40497664633110053, + "grad_norm": 0.5047956109046936, + "learning_rate": 0.00012945222115794448, + "loss": 0.1144, + "step": 17276 + }, + { + "epoch": 0.4050000879059055, + "grad_norm": 0.4300673007965088, + "learning_rate": 0.0001294451816488401, + "loss": 0.0688, + "step": 17277 + }, + { + "epoch": 0.4050235294807105, + "grad_norm": 0.23691526055335999, + "learning_rate": 0.0001294381419799653, + "loss": 0.0444, + "step": 17278 + }, + { + "epoch": 0.4050469710555155, + "grad_norm": 0.2970563769340515, + "learning_rate": 0.00012943110215135836, + "loss": 0.0713, + "step": 17279 + }, + { + "epoch": 0.4050704126303205, + "grad_norm": 0.4008517861366272, + "learning_rate": 0.00012942406216305742, + "loss": 0.091, + "step": 17280 + }, + { + "epoch": 0.4050938542051255, + "grad_norm": 0.6342095732688904, + "learning_rate": 0.00012941702201510068, + "loss": 0.0772, + "step": 17281 + }, + { + "epoch": 0.4051172957799305, + "grad_norm": 0.44248735904693604, + "learning_rate": 0.0001294099817075264, + "loss": 0.1022, + "step": 17282 + }, + { + "epoch": 0.4051407373547355, + "grad_norm": 0.6981725692749023, + "learning_rate": 0.00012940294124037268, + "loss": 0.2358, + "step": 17283 + }, + { + "epoch": 0.4051641789295405, + "grad_norm": 0.7152397632598877, + "learning_rate": 0.00012939590061367774, + "loss": 0.2, + "step": 17284 + }, + { + "epoch": 0.4051876205043455, + "grad_norm": 0.1922595351934433, + "learning_rate": 0.00012938885982747992, + "loss": 0.032, + "step": 17285 + }, + { + "epoch": 0.4052110620791505, + "grad_norm": 0.40378811955451965, + "learning_rate": 0.00012938181888181723, + "loss": 0.0808, + "step": 17286 + }, + { + "epoch": 0.40523450365395547, + "grad_norm": 0.6018376350402832, + "learning_rate": 0.000129374777776728, + "loss": 0.1061, + "step": 17287 + }, + { + "epoch": 0.40525794522876046, + "grad_norm": 0.6078627109527588, + "learning_rate": 0.0001293677365122504, + "loss": 0.1408, + "step": 17288 + }, + { + "epoch": 0.40528138680356546, + "grad_norm": 0.38852810859680176, + "learning_rate": 0.00012936069508842266, + "loss": 0.0782, + "step": 17289 + }, + { + "epoch": 0.40530482837837045, + "grad_norm": 0.46729549765586853, + "learning_rate": 0.00012935365350528292, + "loss": 0.1244, + "step": 17290 + }, + { + "epoch": 0.40532826995317545, + "grad_norm": 0.2350389063358307, + "learning_rate": 0.00012934661176286944, + "loss": 0.0362, + "step": 17291 + }, + { + "epoch": 0.40535171152798044, + "grad_norm": 0.6100584268569946, + "learning_rate": 0.00012933956986122046, + "loss": 0.1153, + "step": 17292 + }, + { + "epoch": 0.40537515310278543, + "grad_norm": 0.23971541225910187, + "learning_rate": 0.00012933252780037409, + "loss": 0.0531, + "step": 17293 + }, + { + "epoch": 0.40539859467759043, + "grad_norm": 0.8184467554092407, + "learning_rate": 0.00012932548558036862, + "loss": 0.0921, + "step": 17294 + }, + { + "epoch": 0.4054220362523954, + "grad_norm": 0.32233455777168274, + "learning_rate": 0.00012931844320124226, + "loss": 0.0592, + "step": 17295 + }, + { + "epoch": 0.4054454778272004, + "grad_norm": 0.3932591676712036, + "learning_rate": 0.00012931140066303317, + "loss": 0.0944, + "step": 17296 + }, + { + "epoch": 0.4054689194020054, + "grad_norm": 0.21991737186908722, + "learning_rate": 0.0001293043579657796, + "loss": 0.0408, + "step": 17297 + }, + { + "epoch": 0.4054923609768104, + "grad_norm": 0.32361406087875366, + "learning_rate": 0.0001292973151095198, + "loss": 0.063, + "step": 17298 + }, + { + "epoch": 0.4055158025516154, + "grad_norm": 0.7124322056770325, + "learning_rate": 0.00012929027209429195, + "loss": 0.1664, + "step": 17299 + }, + { + "epoch": 0.4055392441264204, + "grad_norm": 0.44360291957855225, + "learning_rate": 0.0001292832289201342, + "loss": 0.0566, + "step": 17300 + }, + { + "epoch": 0.4055626857012254, + "grad_norm": 0.15096886456012726, + "learning_rate": 0.00012927618558708484, + "loss": 0.0232, + "step": 17301 + }, + { + "epoch": 0.4055861272760304, + "grad_norm": 0.4375406801700592, + "learning_rate": 0.0001292691420951821, + "loss": 0.0895, + "step": 17302 + }, + { + "epoch": 0.4056095688508354, + "grad_norm": 0.4758208096027374, + "learning_rate": 0.00012926209844446417, + "loss": 0.0927, + "step": 17303 + }, + { + "epoch": 0.4056330104256404, + "grad_norm": 0.4425889253616333, + "learning_rate": 0.00012925505463496926, + "loss": 0.0585, + "step": 17304 + }, + { + "epoch": 0.40565645200044537, + "grad_norm": 0.5555486679077148, + "learning_rate": 0.0001292480106667356, + "loss": 0.1181, + "step": 17305 + }, + { + "epoch": 0.40567989357525036, + "grad_norm": 0.5075753331184387, + "learning_rate": 0.00012924096653980143, + "loss": 0.0935, + "step": 17306 + }, + { + "epoch": 0.40570333515005536, + "grad_norm": 0.13614590466022491, + "learning_rate": 0.00012923392225420494, + "loss": 0.0169, + "step": 17307 + }, + { + "epoch": 0.40572677672486035, + "grad_norm": 0.44764500856399536, + "learning_rate": 0.00012922687780998434, + "loss": 0.2429, + "step": 17308 + }, + { + "epoch": 0.40575021829966534, + "grad_norm": 0.1702161729335785, + "learning_rate": 0.00012921983320717792, + "loss": 0.0505, + "step": 17309 + }, + { + "epoch": 0.4057736598744704, + "grad_norm": 0.410938024520874, + "learning_rate": 0.00012921278844582382, + "loss": 0.0658, + "step": 17310 + }, + { + "epoch": 0.4057971014492754, + "grad_norm": 0.17533865571022034, + "learning_rate": 0.00012920574352596036, + "loss": 0.0363, + "step": 17311 + }, + { + "epoch": 0.4058205430240804, + "grad_norm": 0.1142522469162941, + "learning_rate": 0.00012919869844762568, + "loss": 0.0109, + "step": 17312 + }, + { + "epoch": 0.4058439845988854, + "grad_norm": 0.4127403497695923, + "learning_rate": 0.00012919165321085806, + "loss": 0.0697, + "step": 17313 + }, + { + "epoch": 0.40586742617369037, + "grad_norm": 0.08317318558692932, + "learning_rate": 0.0001291846078156957, + "loss": 0.0158, + "step": 17314 + }, + { + "epoch": 0.40589086774849537, + "grad_norm": 0.34802743792533875, + "learning_rate": 0.00012917756226217683, + "loss": 0.0878, + "step": 17315 + }, + { + "epoch": 0.40591430932330036, + "grad_norm": 0.0707402229309082, + "learning_rate": 0.0001291705165503397, + "loss": 0.0097, + "step": 17316 + }, + { + "epoch": 0.40593775089810535, + "grad_norm": 0.6088640093803406, + "learning_rate": 0.0001291634706802225, + "loss": 0.077, + "step": 17317 + }, + { + "epoch": 0.40596119247291035, + "grad_norm": 0.3317272961139679, + "learning_rate": 0.00012915642465186354, + "loss": 0.1036, + "step": 17318 + }, + { + "epoch": 0.40598463404771534, + "grad_norm": 0.45331957936286926, + "learning_rate": 0.00012914937846530096, + "loss": 0.0874, + "step": 17319 + }, + { + "epoch": 0.40600807562252034, + "grad_norm": 0.6895178556442261, + "learning_rate": 0.00012914233212057302, + "loss": 0.1832, + "step": 17320 + }, + { + "epoch": 0.40603151719732533, + "grad_norm": 0.14072398841381073, + "learning_rate": 0.000129135285617718, + "loss": 0.0305, + "step": 17321 + }, + { + "epoch": 0.4060549587721303, + "grad_norm": 0.5273860692977905, + "learning_rate": 0.0001291282389567741, + "loss": 0.1323, + "step": 17322 + }, + { + "epoch": 0.4060784003469353, + "grad_norm": 0.5313221216201782, + "learning_rate": 0.00012912119213777956, + "loss": 0.092, + "step": 17323 + }, + { + "epoch": 0.4061018419217403, + "grad_norm": 0.6398321986198425, + "learning_rate": 0.00012911414516077261, + "loss": 0.7637, + "step": 17324 + }, + { + "epoch": 0.4061252834965453, + "grad_norm": 0.4612998366355896, + "learning_rate": 0.00012910709802579147, + "loss": 0.1319, + "step": 17325 + }, + { + "epoch": 0.4061487250713503, + "grad_norm": 0.3589491546154022, + "learning_rate": 0.00012910005073287438, + "loss": 0.0713, + "step": 17326 + }, + { + "epoch": 0.4061721666461553, + "grad_norm": 0.6897099018096924, + "learning_rate": 0.00012909300328205965, + "loss": 0.1761, + "step": 17327 + }, + { + "epoch": 0.4061956082209603, + "grad_norm": 0.10946814715862274, + "learning_rate": 0.00012908595567338543, + "loss": 0.019, + "step": 17328 + }, + { + "epoch": 0.4062190497957653, + "grad_norm": 0.3114737868309021, + "learning_rate": 0.00012907890790689, + "loss": 0.0626, + "step": 17329 + }, + { + "epoch": 0.4062424913705703, + "grad_norm": 0.38787978887557983, + "learning_rate": 0.0001290718599826116, + "loss": 0.1147, + "step": 17330 + }, + { + "epoch": 0.4062659329453753, + "grad_norm": 0.7255786061286926, + "learning_rate": 0.00012906481190058848, + "loss": 0.0863, + "step": 17331 + }, + { + "epoch": 0.40628937452018027, + "grad_norm": 0.5394948720932007, + "learning_rate": 0.00012905776366085885, + "loss": 0.1563, + "step": 17332 + }, + { + "epoch": 0.40631281609498526, + "grad_norm": 0.3042374849319458, + "learning_rate": 0.00012905071526346103, + "loss": 0.0223, + "step": 17333 + }, + { + "epoch": 0.40633625766979026, + "grad_norm": 0.7363254427909851, + "learning_rate": 0.00012904366670843316, + "loss": 0.1768, + "step": 17334 + }, + { + "epoch": 0.40635969924459525, + "grad_norm": 0.3539288341999054, + "learning_rate": 0.00012903661799581356, + "loss": 0.0963, + "step": 17335 + }, + { + "epoch": 0.40638314081940025, + "grad_norm": 0.4133909046649933, + "learning_rate": 0.00012902956912564046, + "loss": 0.0639, + "step": 17336 + }, + { + "epoch": 0.40640658239420524, + "grad_norm": 0.4780500531196594, + "learning_rate": 0.00012902252009795208, + "loss": 0.103, + "step": 17337 + }, + { + "epoch": 0.40643002396901023, + "grad_norm": 0.2811535596847534, + "learning_rate": 0.00012901547091278667, + "loss": 0.0583, + "step": 17338 + }, + { + "epoch": 0.40645346554381523, + "grad_norm": 0.42182502150535583, + "learning_rate": 0.00012900842157018255, + "loss": 0.0506, + "step": 17339 + }, + { + "epoch": 0.4064769071186202, + "grad_norm": 0.6161218881607056, + "learning_rate": 0.0001290013720701779, + "loss": 0.1166, + "step": 17340 + }, + { + "epoch": 0.4065003486934252, + "grad_norm": 0.33113089203834534, + "learning_rate": 0.00012899432241281097, + "loss": 0.0452, + "step": 17341 + }, + { + "epoch": 0.4065237902682302, + "grad_norm": 0.47618988156318665, + "learning_rate": 0.00012898727259812008, + "loss": 0.0621, + "step": 17342 + }, + { + "epoch": 0.4065472318430352, + "grad_norm": 0.524762749671936, + "learning_rate": 0.0001289802226261434, + "loss": 0.1274, + "step": 17343 + }, + { + "epoch": 0.4065706734178402, + "grad_norm": 0.1709185093641281, + "learning_rate": 0.0001289731724969192, + "loss": 0.0242, + "step": 17344 + }, + { + "epoch": 0.4065941149926452, + "grad_norm": 0.24674272537231445, + "learning_rate": 0.0001289661222104858, + "loss": 0.0612, + "step": 17345 + }, + { + "epoch": 0.4066175565674502, + "grad_norm": 0.29095280170440674, + "learning_rate": 0.00012895907176688136, + "loss": 0.0883, + "step": 17346 + }, + { + "epoch": 0.4066409981422552, + "grad_norm": 0.43306347727775574, + "learning_rate": 0.00012895202116614422, + "loss": 0.09, + "step": 17347 + }, + { + "epoch": 0.4066644397170602, + "grad_norm": 0.4741191267967224, + "learning_rate": 0.00012894497040831258, + "loss": 0.4348, + "step": 17348 + }, + { + "epoch": 0.40668788129186517, + "grad_norm": 0.2790962755680084, + "learning_rate": 0.0001289379194934247, + "loss": 0.0403, + "step": 17349 + }, + { + "epoch": 0.40671132286667017, + "grad_norm": 0.2664991617202759, + "learning_rate": 0.00012893086842151886, + "loss": 0.0426, + "step": 17350 + }, + { + "epoch": 0.40673476444147516, + "grad_norm": 0.6418903470039368, + "learning_rate": 0.00012892381719263333, + "loss": 0.1268, + "step": 17351 + }, + { + "epoch": 0.40675820601628015, + "grad_norm": 0.5728424787521362, + "learning_rate": 0.00012891676580680637, + "loss": 0.1247, + "step": 17352 + }, + { + "epoch": 0.40678164759108515, + "grad_norm": 0.4807819128036499, + "learning_rate": 0.0001289097142640762, + "loss": 0.1569, + "step": 17353 + }, + { + "epoch": 0.40680508916589014, + "grad_norm": 0.31820687651634216, + "learning_rate": 0.0001289026625644811, + "loss": 0.064, + "step": 17354 + }, + { + "epoch": 0.40682853074069514, + "grad_norm": 0.14373046159744263, + "learning_rate": 0.00012889561070805936, + "loss": 0.0155, + "step": 17355 + }, + { + "epoch": 0.40685197231550013, + "grad_norm": 0.3854498863220215, + "learning_rate": 0.00012888855869484922, + "loss": 0.0874, + "step": 17356 + }, + { + "epoch": 0.4068754138903051, + "grad_norm": 0.44034501910209656, + "learning_rate": 0.00012888150652488894, + "loss": 0.115, + "step": 17357 + }, + { + "epoch": 0.4068988554651101, + "grad_norm": 0.16129666566848755, + "learning_rate": 0.0001288744541982168, + "loss": 0.0417, + "step": 17358 + }, + { + "epoch": 0.4069222970399151, + "grad_norm": 0.5424202680587769, + "learning_rate": 0.0001288674017148711, + "loss": 0.6337, + "step": 17359 + }, + { + "epoch": 0.4069457386147201, + "grad_norm": 0.49351465702056885, + "learning_rate": 0.00012886034907489, + "loss": 0.0709, + "step": 17360 + }, + { + "epoch": 0.40696918018952516, + "grad_norm": 0.4886469542980194, + "learning_rate": 0.00012885329627831187, + "loss": 0.0597, + "step": 17361 + }, + { + "epoch": 0.40699262176433015, + "grad_norm": 0.2950018346309662, + "learning_rate": 0.00012884624332517493, + "loss": 0.0615, + "step": 17362 + }, + { + "epoch": 0.40701606333913515, + "grad_norm": 0.885370135307312, + "learning_rate": 0.00012883919021551744, + "loss": 0.1019, + "step": 17363 + }, + { + "epoch": 0.40703950491394014, + "grad_norm": 0.3686838448047638, + "learning_rate": 0.00012883213694937773, + "loss": 0.0328, + "step": 17364 + }, + { + "epoch": 0.40706294648874514, + "grad_norm": 0.15637922286987305, + "learning_rate": 0.00012882508352679404, + "loss": 0.0391, + "step": 17365 + }, + { + "epoch": 0.40708638806355013, + "grad_norm": 0.4031120240688324, + "learning_rate": 0.0001288180299478046, + "loss": 0.11, + "step": 17366 + }, + { + "epoch": 0.4071098296383551, + "grad_norm": 0.15750935673713684, + "learning_rate": 0.00012881097621244775, + "loss": 0.0313, + "step": 17367 + }, + { + "epoch": 0.4071332712131601, + "grad_norm": 0.4628004729747772, + "learning_rate": 0.00012880392232076174, + "loss": 0.128, + "step": 17368 + }, + { + "epoch": 0.4071567127879651, + "grad_norm": 0.2180127203464508, + "learning_rate": 0.00012879686827278478, + "loss": 0.0311, + "step": 17369 + }, + { + "epoch": 0.4071801543627701, + "grad_norm": 0.1860944777727127, + "learning_rate": 0.00012878981406855527, + "loss": 0.029, + "step": 17370 + }, + { + "epoch": 0.4072035959375751, + "grad_norm": 0.32346686720848083, + "learning_rate": 0.0001287827597081114, + "loss": 0.065, + "step": 17371 + }, + { + "epoch": 0.4072270375123801, + "grad_norm": 0.44090887904167175, + "learning_rate": 0.00012877570519149144, + "loss": 0.4001, + "step": 17372 + }, + { + "epoch": 0.4072504790871851, + "grad_norm": 0.40233781933784485, + "learning_rate": 0.0001287686505187337, + "loss": 0.4804, + "step": 17373 + }, + { + "epoch": 0.4072739206619901, + "grad_norm": 0.6963333487510681, + "learning_rate": 0.0001287615956898765, + "loss": 0.121, + "step": 17374 + }, + { + "epoch": 0.4072973622367951, + "grad_norm": 0.6796548962593079, + "learning_rate": 0.00012875454070495802, + "loss": 0.1124, + "step": 17375 + }, + { + "epoch": 0.4073208038116001, + "grad_norm": 0.9009605050086975, + "learning_rate": 0.00012874748556401659, + "loss": 0.1961, + "step": 17376 + }, + { + "epoch": 0.40734424538640507, + "grad_norm": 0.43848752975463867, + "learning_rate": 0.00012874043026709053, + "loss": 0.0962, + "step": 17377 + }, + { + "epoch": 0.40736768696121006, + "grad_norm": 0.5183226466178894, + "learning_rate": 0.0001287333748142181, + "loss": 0.0878, + "step": 17378 + }, + { + "epoch": 0.40739112853601506, + "grad_norm": 0.6463416814804077, + "learning_rate": 0.00012872631920543754, + "loss": 0.1159, + "step": 17379 + }, + { + "epoch": 0.40741457011082005, + "grad_norm": 0.3517349362373352, + "learning_rate": 0.00012871926344078718, + "loss": 0.4295, + "step": 17380 + }, + { + "epoch": 0.40743801168562505, + "grad_norm": 0.15784984827041626, + "learning_rate": 0.00012871220752030526, + "loss": 0.0202, + "step": 17381 + }, + { + "epoch": 0.40746145326043004, + "grad_norm": 0.46051549911499023, + "learning_rate": 0.00012870515144403012, + "loss": 0.1255, + "step": 17382 + }, + { + "epoch": 0.40748489483523503, + "grad_norm": 0.46748316287994385, + "learning_rate": 0.000128698095212, + "loss": 0.0536, + "step": 17383 + }, + { + "epoch": 0.40750833641004003, + "grad_norm": 0.5881630778312683, + "learning_rate": 0.00012869103882425325, + "loss": 0.1183, + "step": 17384 + }, + { + "epoch": 0.407531777984845, + "grad_norm": 0.520068883895874, + "learning_rate": 0.0001286839822808281, + "loss": 0.5008, + "step": 17385 + }, + { + "epoch": 0.40755521955965, + "grad_norm": 0.06524686515331268, + "learning_rate": 0.00012867692558176283, + "loss": 0.0111, + "step": 17386 + }, + { + "epoch": 0.407578661134455, + "grad_norm": 0.7107399702072144, + "learning_rate": 0.00012866986872709578, + "loss": 0.1551, + "step": 17387 + }, + { + "epoch": 0.40760210270926, + "grad_norm": 0.45665568113327026, + "learning_rate": 0.0001286628117168652, + "loss": 0.1416, + "step": 17388 + }, + { + "epoch": 0.407625544284065, + "grad_norm": 0.2559674382209778, + "learning_rate": 0.00012865575455110937, + "loss": 0.0613, + "step": 17389 + }, + { + "epoch": 0.40764898585887, + "grad_norm": 0.7874266505241394, + "learning_rate": 0.00012864869722986668, + "loss": 0.1821, + "step": 17390 + }, + { + "epoch": 0.407672427433675, + "grad_norm": 0.5781041979789734, + "learning_rate": 0.00012864163975317533, + "loss": 0.823, + "step": 17391 + }, + { + "epoch": 0.40769586900848, + "grad_norm": 0.5001242756843567, + "learning_rate": 0.00012863458212107362, + "loss": 0.115, + "step": 17392 + }, + { + "epoch": 0.407719310583285, + "grad_norm": 0.21128682792186737, + "learning_rate": 0.00012862752433359984, + "loss": 0.0571, + "step": 17393 + }, + { + "epoch": 0.40774275215808997, + "grad_norm": 0.325927197933197, + "learning_rate": 0.00012862046639079236, + "loss": 0.0919, + "step": 17394 + }, + { + "epoch": 0.40776619373289497, + "grad_norm": 0.627676784992218, + "learning_rate": 0.00012861340829268937, + "loss": 0.1355, + "step": 17395 + }, + { + "epoch": 0.40778963530769996, + "grad_norm": 0.12859167158603668, + "learning_rate": 0.00012860635003932928, + "loss": 0.0184, + "step": 17396 + }, + { + "epoch": 0.40781307688250495, + "grad_norm": 0.23149511218070984, + "learning_rate": 0.0001285992916307503, + "loss": 0.047, + "step": 17397 + }, + { + "epoch": 0.40783651845730995, + "grad_norm": 0.14943577349185944, + "learning_rate": 0.00012859223306699075, + "loss": 0.0385, + "step": 17398 + }, + { + "epoch": 0.40785996003211494, + "grad_norm": 0.4630756378173828, + "learning_rate": 0.00012858517434808893, + "loss": 0.0765, + "step": 17399 + }, + { + "epoch": 0.40788340160691994, + "grad_norm": 0.3301449418067932, + "learning_rate": 0.00012857811547408322, + "loss": 0.0901, + "step": 17400 + }, + { + "epoch": 0.40790684318172493, + "grad_norm": 0.23514656722545624, + "learning_rate": 0.0001285710564450118, + "loss": 0.0486, + "step": 17401 + }, + { + "epoch": 0.4079302847565299, + "grad_norm": 0.592130720615387, + "learning_rate": 0.000128563997260913, + "loss": 0.1283, + "step": 17402 + }, + { + "epoch": 0.4079537263313349, + "grad_norm": 0.3073213994503021, + "learning_rate": 0.0001285569379218252, + "loss": 0.0698, + "step": 17403 + }, + { + "epoch": 0.4079771679061399, + "grad_norm": 0.6490917205810547, + "learning_rate": 0.00012854987842778664, + "loss": 0.0968, + "step": 17404 + }, + { + "epoch": 0.4080006094809449, + "grad_norm": 0.8353047370910645, + "learning_rate": 0.0001285428187788356, + "loss": 0.1826, + "step": 17405 + }, + { + "epoch": 0.4080240510557499, + "grad_norm": 0.45551109313964844, + "learning_rate": 0.00012853575897501043, + "loss": 0.0807, + "step": 17406 + }, + { + "epoch": 0.4080474926305549, + "grad_norm": 0.5757564902305603, + "learning_rate": 0.00012852869901634948, + "loss": 0.0968, + "step": 17407 + }, + { + "epoch": 0.4080709342053599, + "grad_norm": 0.23397605121135712, + "learning_rate": 0.00012852163890289097, + "loss": 0.0449, + "step": 17408 + }, + { + "epoch": 0.4080943757801649, + "grad_norm": 0.7980169653892517, + "learning_rate": 0.00012851457863467325, + "loss": 0.1037, + "step": 17409 + }, + { + "epoch": 0.4081178173549699, + "grad_norm": 0.6401036381721497, + "learning_rate": 0.00012850751821173463, + "loss": 0.1247, + "step": 17410 + }, + { + "epoch": 0.4081412589297749, + "grad_norm": 0.27312055230140686, + "learning_rate": 0.0001285004576341134, + "loss": 0.0232, + "step": 17411 + }, + { + "epoch": 0.4081647005045799, + "grad_norm": 0.4353926479816437, + "learning_rate": 0.0001284933969018479, + "loss": 0.1199, + "step": 17412 + }, + { + "epoch": 0.4081881420793849, + "grad_norm": 0.48947301506996155, + "learning_rate": 0.00012848633601497643, + "loss": 0.6295, + "step": 17413 + }, + { + "epoch": 0.4082115836541899, + "grad_norm": 0.5907922983169556, + "learning_rate": 0.00012847927497353727, + "loss": 0.0943, + "step": 17414 + }, + { + "epoch": 0.4082350252289949, + "grad_norm": 0.5806772112846375, + "learning_rate": 0.0001284722137775688, + "loss": 0.1045, + "step": 17415 + }, + { + "epoch": 0.4082584668037999, + "grad_norm": 0.35521048307418823, + "learning_rate": 0.0001284651524271093, + "loss": 0.0343, + "step": 17416 + }, + { + "epoch": 0.4082819083786049, + "grad_norm": 0.6247165203094482, + "learning_rate": 0.00012845809092219704, + "loss": 0.2012, + "step": 17417 + }, + { + "epoch": 0.4083053499534099, + "grad_norm": 0.3769995868206024, + "learning_rate": 0.00012845102926287041, + "loss": 0.0697, + "step": 17418 + }, + { + "epoch": 0.4083287915282149, + "grad_norm": 0.17561469972133636, + "learning_rate": 0.00012844396744916766, + "loss": 0.0265, + "step": 17419 + }, + { + "epoch": 0.4083522331030199, + "grad_norm": 0.2044462114572525, + "learning_rate": 0.0001284369054811272, + "loss": 0.0312, + "step": 17420 + }, + { + "epoch": 0.4083756746778249, + "grad_norm": 0.1419542282819748, + "learning_rate": 0.00012842984335878724, + "loss": 0.0216, + "step": 17421 + }, + { + "epoch": 0.40839911625262987, + "grad_norm": 0.4241708219051361, + "learning_rate": 0.0001284227810821862, + "loss": 0.0909, + "step": 17422 + }, + { + "epoch": 0.40842255782743486, + "grad_norm": 0.3428323268890381, + "learning_rate": 0.0001284157186513623, + "loss": 0.0733, + "step": 17423 + }, + { + "epoch": 0.40844599940223986, + "grad_norm": 0.2738063335418701, + "learning_rate": 0.00012840865606635393, + "loss": 0.0418, + "step": 17424 + }, + { + "epoch": 0.40846944097704485, + "grad_norm": 0.5405604839324951, + "learning_rate": 0.0001284015933271994, + "loss": 0.0657, + "step": 17425 + }, + { + "epoch": 0.40849288255184985, + "grad_norm": 0.36236998438835144, + "learning_rate": 0.00012839453043393705, + "loss": 0.0849, + "step": 17426 + }, + { + "epoch": 0.40851632412665484, + "grad_norm": 0.11333684623241425, + "learning_rate": 0.00012838746738660513, + "loss": 0.0313, + "step": 17427 + }, + { + "epoch": 0.40853976570145983, + "grad_norm": 0.07744666934013367, + "learning_rate": 0.00012838040418524207, + "loss": 0.007, + "step": 17428 + }, + { + "epoch": 0.40856320727626483, + "grad_norm": 0.18863613903522491, + "learning_rate": 0.00012837334082988608, + "loss": 0.0387, + "step": 17429 + }, + { + "epoch": 0.4085866488510698, + "grad_norm": 0.4992409348487854, + "learning_rate": 0.00012836627732057556, + "loss": 0.0363, + "step": 17430 + }, + { + "epoch": 0.4086100904258748, + "grad_norm": 0.8907987475395203, + "learning_rate": 0.00012835921365734885, + "loss": 0.2294, + "step": 17431 + }, + { + "epoch": 0.4086335320006798, + "grad_norm": 0.19882549345493317, + "learning_rate": 0.00012835214984024425, + "loss": 0.0485, + "step": 17432 + }, + { + "epoch": 0.4086569735754848, + "grad_norm": 0.6732211112976074, + "learning_rate": 0.00012834508586930007, + "loss": 0.7543, + "step": 17433 + }, + { + "epoch": 0.4086804151502898, + "grad_norm": 0.7981846928596497, + "learning_rate": 0.00012833802174455465, + "loss": 0.0955, + "step": 17434 + }, + { + "epoch": 0.4087038567250948, + "grad_norm": 0.6243313550949097, + "learning_rate": 0.00012833095746604632, + "loss": 0.1876, + "step": 17435 + }, + { + "epoch": 0.4087272982998998, + "grad_norm": 0.17884299159049988, + "learning_rate": 0.00012832389303381344, + "loss": 0.0293, + "step": 17436 + }, + { + "epoch": 0.4087507398747048, + "grad_norm": 0.4887430667877197, + "learning_rate": 0.0001283168284478943, + "loss": 0.1157, + "step": 17437 + }, + { + "epoch": 0.4087741814495098, + "grad_norm": 0.9142494797706604, + "learning_rate": 0.00012830976370832727, + "loss": 0.1007, + "step": 17438 + }, + { + "epoch": 0.40879762302431477, + "grad_norm": 0.5360110402107239, + "learning_rate": 0.0001283026988151507, + "loss": 0.1086, + "step": 17439 + }, + { + "epoch": 0.40882106459911977, + "grad_norm": 0.8037793040275574, + "learning_rate": 0.00012829563376840283, + "loss": 0.09, + "step": 17440 + }, + { + "epoch": 0.40884450617392476, + "grad_norm": 0.5270575284957886, + "learning_rate": 0.00012828856856812208, + "loss": 0.0992, + "step": 17441 + }, + { + "epoch": 0.40886794774872975, + "grad_norm": 0.26385197043418884, + "learning_rate": 0.00012828150321434677, + "loss": 0.0707, + "step": 17442 + }, + { + "epoch": 0.40889138932353475, + "grad_norm": 0.6776712536811829, + "learning_rate": 0.00012827443770711518, + "loss": 0.1917, + "step": 17443 + }, + { + "epoch": 0.40891483089833974, + "grad_norm": 0.4851464629173279, + "learning_rate": 0.00012826737204646576, + "loss": 0.107, + "step": 17444 + }, + { + "epoch": 0.40893827247314474, + "grad_norm": 0.10839390009641647, + "learning_rate": 0.00012826030623243676, + "loss": 0.0208, + "step": 17445 + }, + { + "epoch": 0.40896171404794973, + "grad_norm": 0.4812001585960388, + "learning_rate": 0.00012825324026506654, + "loss": 0.0755, + "step": 17446 + }, + { + "epoch": 0.4089851556227547, + "grad_norm": 0.21969780325889587, + "learning_rate": 0.00012824617414439345, + "loss": 0.0505, + "step": 17447 + }, + { + "epoch": 0.4090085971975597, + "grad_norm": 0.41640329360961914, + "learning_rate": 0.0001282391078704558, + "loss": 0.107, + "step": 17448 + }, + { + "epoch": 0.4090320387723647, + "grad_norm": 0.5805526971817017, + "learning_rate": 0.00012823204144329198, + "loss": 0.1047, + "step": 17449 + }, + { + "epoch": 0.4090554803471697, + "grad_norm": 0.38992470502853394, + "learning_rate": 0.00012822497486294032, + "loss": 0.0781, + "step": 17450 + }, + { + "epoch": 0.4090789219219747, + "grad_norm": 0.6665995717048645, + "learning_rate": 0.00012821790812943916, + "loss": 0.131, + "step": 17451 + }, + { + "epoch": 0.4091023634967797, + "grad_norm": 0.753831684589386, + "learning_rate": 0.00012821084124282678, + "loss": 0.7107, + "step": 17452 + }, + { + "epoch": 0.4091258050715847, + "grad_norm": 0.539993941783905, + "learning_rate": 0.00012820377420314164, + "loss": 0.1155, + "step": 17453 + }, + { + "epoch": 0.4091492466463897, + "grad_norm": 0.5027894973754883, + "learning_rate": 0.00012819670701042199, + "loss": 0.1163, + "step": 17454 + }, + { + "epoch": 0.4091726882211947, + "grad_norm": 0.2983097732067108, + "learning_rate": 0.00012818963966470626, + "loss": 0.072, + "step": 17455 + }, + { + "epoch": 0.4091961297959997, + "grad_norm": 0.4024394750595093, + "learning_rate": 0.00012818257216603272, + "loss": 0.1041, + "step": 17456 + }, + { + "epoch": 0.40921957137080467, + "grad_norm": 0.4751189947128296, + "learning_rate": 0.00012817550451443973, + "loss": 0.0928, + "step": 17457 + }, + { + "epoch": 0.40924301294560966, + "grad_norm": 0.6819523572921753, + "learning_rate": 0.00012816843670996573, + "loss": 0.1301, + "step": 17458 + }, + { + "epoch": 0.40926645452041466, + "grad_norm": 0.19162097573280334, + "learning_rate": 0.00012816136875264892, + "loss": 0.0385, + "step": 17459 + }, + { + "epoch": 0.40928989609521965, + "grad_norm": 0.4159654676914215, + "learning_rate": 0.00012815430064252775, + "loss": 0.0605, + "step": 17460 + }, + { + "epoch": 0.40931333767002465, + "grad_norm": 0.5862693786621094, + "learning_rate": 0.0001281472323796406, + "loss": 0.1119, + "step": 17461 + }, + { + "epoch": 0.40933677924482964, + "grad_norm": 0.558858335018158, + "learning_rate": 0.00012814016396402574, + "loss": 0.0988, + "step": 17462 + }, + { + "epoch": 0.4093602208196347, + "grad_norm": 0.4943818151950836, + "learning_rate": 0.00012813309539572157, + "loss": 0.1342, + "step": 17463 + }, + { + "epoch": 0.4093836623944397, + "grad_norm": 0.39450475573539734, + "learning_rate": 0.00012812602667476643, + "loss": 0.0503, + "step": 17464 + }, + { + "epoch": 0.4094071039692447, + "grad_norm": 0.3664763569831848, + "learning_rate": 0.00012811895780119866, + "loss": 0.0707, + "step": 17465 + }, + { + "epoch": 0.4094305455440497, + "grad_norm": 0.4890187382698059, + "learning_rate": 0.00012811188877505665, + "loss": 0.0538, + "step": 17466 + }, + { + "epoch": 0.40945398711885467, + "grad_norm": 0.45596614480018616, + "learning_rate": 0.00012810481959637875, + "loss": 0.1244, + "step": 17467 + }, + { + "epoch": 0.40947742869365966, + "grad_norm": 0.47375020384788513, + "learning_rate": 0.0001280977502652033, + "loss": 0.0704, + "step": 17468 + }, + { + "epoch": 0.40950087026846466, + "grad_norm": 0.5619401335716248, + "learning_rate": 0.00012809068078156868, + "loss": 0.1721, + "step": 17469 + }, + { + "epoch": 0.40952431184326965, + "grad_norm": 0.2291768491268158, + "learning_rate": 0.0001280836111455132, + "loss": 0.0604, + "step": 17470 + }, + { + "epoch": 0.40954775341807464, + "grad_norm": 0.19221945106983185, + "learning_rate": 0.0001280765413570753, + "loss": 0.0333, + "step": 17471 + }, + { + "epoch": 0.40957119499287964, + "grad_norm": 0.401974618434906, + "learning_rate": 0.00012806947141629326, + "loss": 0.0594, + "step": 17472 + }, + { + "epoch": 0.40959463656768463, + "grad_norm": 0.3991016149520874, + "learning_rate": 0.0001280624013232055, + "loss": 0.0868, + "step": 17473 + }, + { + "epoch": 0.40961807814248963, + "grad_norm": 0.19233404099941254, + "learning_rate": 0.00012805533107785038, + "loss": 0.0392, + "step": 17474 + }, + { + "epoch": 0.4096415197172946, + "grad_norm": 0.39951807260513306, + "learning_rate": 0.0001280482606802662, + "loss": 0.0472, + "step": 17475 + }, + { + "epoch": 0.4096649612920996, + "grad_norm": 0.44441062211990356, + "learning_rate": 0.00012804119013049142, + "loss": 0.0882, + "step": 17476 + }, + { + "epoch": 0.4096884028669046, + "grad_norm": 0.5321190357208252, + "learning_rate": 0.0001280341194285643, + "loss": 0.148, + "step": 17477 + }, + { + "epoch": 0.4097118444417096, + "grad_norm": 0.14126887917518616, + "learning_rate": 0.00012802704857452327, + "loss": 0.0358, + "step": 17478 + }, + { + "epoch": 0.4097352860165146, + "grad_norm": 0.26370546221733093, + "learning_rate": 0.00012801997756840673, + "loss": 0.056, + "step": 17479 + }, + { + "epoch": 0.4097587275913196, + "grad_norm": 0.44003456830978394, + "learning_rate": 0.00012801290641025297, + "loss": 0.0603, + "step": 17480 + }, + { + "epoch": 0.4097821691661246, + "grad_norm": 0.32301437854766846, + "learning_rate": 0.0001280058351001004, + "loss": 0.0728, + "step": 17481 + }, + { + "epoch": 0.4098056107409296, + "grad_norm": 0.34035056829452515, + "learning_rate": 0.00012799876363798736, + "loss": 0.0859, + "step": 17482 + }, + { + "epoch": 0.4098290523157346, + "grad_norm": 0.4405897259712219, + "learning_rate": 0.0001279916920239523, + "loss": 0.0766, + "step": 17483 + }, + { + "epoch": 0.40985249389053957, + "grad_norm": 0.5232386589050293, + "learning_rate": 0.00012798462025803348, + "loss": 0.073, + "step": 17484 + }, + { + "epoch": 0.40987593546534457, + "grad_norm": 0.9206066131591797, + "learning_rate": 0.0001279775483402693, + "loss": 0.181, + "step": 17485 + }, + { + "epoch": 0.40989937704014956, + "grad_norm": 0.49038639664649963, + "learning_rate": 0.00012797047627069824, + "loss": 0.1151, + "step": 17486 + }, + { + "epoch": 0.40992281861495455, + "grad_norm": 0.3641539216041565, + "learning_rate": 0.00012796340404935851, + "loss": 0.0452, + "step": 17487 + }, + { + "epoch": 0.40994626018975955, + "grad_norm": 0.4427322745323181, + "learning_rate": 0.0001279563316762886, + "loss": 0.0651, + "step": 17488 + }, + { + "epoch": 0.40996970176456454, + "grad_norm": 0.45952585339546204, + "learning_rate": 0.00012794925915152688, + "loss": 0.0874, + "step": 17489 + }, + { + "epoch": 0.40999314333936954, + "grad_norm": 0.5495655536651611, + "learning_rate": 0.0001279421864751116, + "loss": 0.1426, + "step": 17490 + }, + { + "epoch": 0.41001658491417453, + "grad_norm": 0.5052682757377625, + "learning_rate": 0.0001279351136470813, + "loss": 0.0973, + "step": 17491 + }, + { + "epoch": 0.4100400264889795, + "grad_norm": 0.13023601472377777, + "learning_rate": 0.00012792804066747427, + "loss": 0.0176, + "step": 17492 + }, + { + "epoch": 0.4100634680637845, + "grad_norm": 0.3476754128932953, + "learning_rate": 0.00012792096753632895, + "loss": 0.073, + "step": 17493 + }, + { + "epoch": 0.4100869096385895, + "grad_norm": 0.3483503460884094, + "learning_rate": 0.00012791389425368362, + "loss": 0.2746, + "step": 17494 + }, + { + "epoch": 0.4101103512133945, + "grad_norm": 0.6388274431228638, + "learning_rate": 0.00012790682081957676, + "loss": 0.0776, + "step": 17495 + }, + { + "epoch": 0.4101337927881995, + "grad_norm": 0.22350351512432098, + "learning_rate": 0.00012789974723404664, + "loss": 0.0505, + "step": 17496 + }, + { + "epoch": 0.4101572343630045, + "grad_norm": 0.3556686341762543, + "learning_rate": 0.00012789267349713174, + "loss": 0.1268, + "step": 17497 + }, + { + "epoch": 0.4101806759378095, + "grad_norm": 0.16813723742961884, + "learning_rate": 0.00012788559960887042, + "loss": 0.0147, + "step": 17498 + }, + { + "epoch": 0.4102041175126145, + "grad_norm": 0.7669781446456909, + "learning_rate": 0.00012787852556930105, + "loss": 0.753, + "step": 17499 + }, + { + "epoch": 0.4102275590874195, + "grad_norm": 0.5087222456932068, + "learning_rate": 0.000127871451378462, + "loss": 0.0758, + "step": 17500 + }, + { + "epoch": 0.4102510006622245, + "grad_norm": 0.844919741153717, + "learning_rate": 0.00012786437703639169, + "loss": 0.146, + "step": 17501 + }, + { + "epoch": 0.41027444223702947, + "grad_norm": 0.19317062199115753, + "learning_rate": 0.00012785730254312845, + "loss": 0.0178, + "step": 17502 + }, + { + "epoch": 0.41029788381183446, + "grad_norm": 0.4281427562236786, + "learning_rate": 0.00012785022789871077, + "loss": 0.0848, + "step": 17503 + }, + { + "epoch": 0.41032132538663946, + "grad_norm": 0.08393967151641846, + "learning_rate": 0.0001278431531031769, + "loss": 0.0171, + "step": 17504 + }, + { + "epoch": 0.41034476696144445, + "grad_norm": 0.5281568765640259, + "learning_rate": 0.00012783607815656535, + "loss": 0.1323, + "step": 17505 + }, + { + "epoch": 0.41036820853624945, + "grad_norm": 0.14741535484790802, + "learning_rate": 0.00012782900305891444, + "loss": 0.0239, + "step": 17506 + }, + { + "epoch": 0.41039165011105444, + "grad_norm": 0.46128544211387634, + "learning_rate": 0.0001278219278102626, + "loss": 0.0637, + "step": 17507 + }, + { + "epoch": 0.41041509168585943, + "grad_norm": 0.60612952709198, + "learning_rate": 0.00012781485241064814, + "loss": 0.1452, + "step": 17508 + }, + { + "epoch": 0.41043853326066443, + "grad_norm": 0.4844747483730316, + "learning_rate": 0.00012780777686010957, + "loss": 0.0956, + "step": 17509 + }, + { + "epoch": 0.4104619748354694, + "grad_norm": 0.20060832798480988, + "learning_rate": 0.00012780070115868517, + "loss": 0.024, + "step": 17510 + }, + { + "epoch": 0.4104854164102744, + "grad_norm": 0.2748340368270874, + "learning_rate": 0.0001277936253064134, + "loss": 0.0608, + "step": 17511 + }, + { + "epoch": 0.4105088579850794, + "grad_norm": 0.6416604518890381, + "learning_rate": 0.00012778654930333268, + "loss": 0.0879, + "step": 17512 + }, + { + "epoch": 0.4105322995598844, + "grad_norm": 0.14080704748630524, + "learning_rate": 0.0001277794731494813, + "loss": 0.0284, + "step": 17513 + }, + { + "epoch": 0.41055574113468946, + "grad_norm": 0.16354867815971375, + "learning_rate": 0.00012777239684489775, + "loss": 0.0318, + "step": 17514 + }, + { + "epoch": 0.41057918270949445, + "grad_norm": 0.19866226613521576, + "learning_rate": 0.00012776532038962039, + "loss": 0.0431, + "step": 17515 + }, + { + "epoch": 0.41060262428429944, + "grad_norm": 0.15452346205711365, + "learning_rate": 0.0001277582437836876, + "loss": 0.0204, + "step": 17516 + }, + { + "epoch": 0.41062606585910444, + "grad_norm": 0.7435904741287231, + "learning_rate": 0.00012775116702713784, + "loss": 0.6931, + "step": 17517 + }, + { + "epoch": 0.41064950743390943, + "grad_norm": 0.6622195839881897, + "learning_rate": 0.00012774409012000944, + "loss": 0.1139, + "step": 17518 + }, + { + "epoch": 0.4106729490087144, + "grad_norm": 0.11747987568378448, + "learning_rate": 0.00012773701306234081, + "loss": 0.0194, + "step": 17519 + }, + { + "epoch": 0.4106963905835194, + "grad_norm": 0.5093726515769958, + "learning_rate": 0.0001277299358541704, + "loss": 0.1209, + "step": 17520 + }, + { + "epoch": 0.4107198321583244, + "grad_norm": 0.4038519561290741, + "learning_rate": 0.00012772285849553658, + "loss": 0.0938, + "step": 17521 + }, + { + "epoch": 0.4107432737331294, + "grad_norm": 0.4293972849845886, + "learning_rate": 0.00012771578098647775, + "loss": 0.0727, + "step": 17522 + }, + { + "epoch": 0.4107667153079344, + "grad_norm": 0.13009512424468994, + "learning_rate": 0.00012770870332703228, + "loss": 0.0221, + "step": 17523 + }, + { + "epoch": 0.4107901568827394, + "grad_norm": 0.5791256427764893, + "learning_rate": 0.00012770162551723865, + "loss": 0.1067, + "step": 17524 + }, + { + "epoch": 0.4108135984575444, + "grad_norm": 0.6815416812896729, + "learning_rate": 0.00012769454755713516, + "loss": 0.1341, + "step": 17525 + }, + { + "epoch": 0.4108370400323494, + "grad_norm": 0.7398602366447449, + "learning_rate": 0.00012768746944676033, + "loss": 0.1934, + "step": 17526 + }, + { + "epoch": 0.4108604816071544, + "grad_norm": 0.22602954506874084, + "learning_rate": 0.0001276803911861525, + "loss": 0.0859, + "step": 17527 + }, + { + "epoch": 0.4108839231819594, + "grad_norm": 0.8783400654792786, + "learning_rate": 0.00012767331277535008, + "loss": 0.0758, + "step": 17528 + }, + { + "epoch": 0.41090736475676437, + "grad_norm": 0.5123146772384644, + "learning_rate": 0.0001276662342143915, + "loss": 0.076, + "step": 17529 + }, + { + "epoch": 0.41093080633156936, + "grad_norm": 0.19757062196731567, + "learning_rate": 0.00012765915550331516, + "loss": 0.0357, + "step": 17530 + }, + { + "epoch": 0.41095424790637436, + "grad_norm": 0.7413903474807739, + "learning_rate": 0.00012765207664215946, + "loss": 0.1163, + "step": 17531 + }, + { + "epoch": 0.41097768948117935, + "grad_norm": 0.83043372631073, + "learning_rate": 0.0001276449976309628, + "loss": 0.1436, + "step": 17532 + }, + { + "epoch": 0.41100113105598435, + "grad_norm": 0.17051100730895996, + "learning_rate": 0.0001276379184697636, + "loss": 0.0167, + "step": 17533 + }, + { + "epoch": 0.41102457263078934, + "grad_norm": 0.48030856251716614, + "learning_rate": 0.0001276308391586003, + "loss": 0.0979, + "step": 17534 + }, + { + "epoch": 0.41104801420559434, + "grad_norm": 0.561970055103302, + "learning_rate": 0.00012762375969751125, + "loss": 0.0984, + "step": 17535 + }, + { + "epoch": 0.41107145578039933, + "grad_norm": 0.5265380144119263, + "learning_rate": 0.00012761668008653496, + "loss": 0.1378, + "step": 17536 + }, + { + "epoch": 0.4110948973552043, + "grad_norm": 0.4374142289161682, + "learning_rate": 0.00012760960032570976, + "loss": 0.0756, + "step": 17537 + }, + { + "epoch": 0.4111183389300093, + "grad_norm": 0.10085159540176392, + "learning_rate": 0.0001276025204150741, + "loss": 0.0211, + "step": 17538 + }, + { + "epoch": 0.4111417805048143, + "grad_norm": 0.40687546133995056, + "learning_rate": 0.00012759544035466637, + "loss": 0.084, + "step": 17539 + }, + { + "epoch": 0.4111652220796193, + "grad_norm": 0.3061887323856354, + "learning_rate": 0.00012758836014452503, + "loss": 0.0466, + "step": 17540 + }, + { + "epoch": 0.4111886636544243, + "grad_norm": 0.4596894085407257, + "learning_rate": 0.00012758127978468846, + "loss": 0.0972, + "step": 17541 + }, + { + "epoch": 0.4112121052292293, + "grad_norm": 0.4765591621398926, + "learning_rate": 0.00012757419927519507, + "loss": 0.088, + "step": 17542 + }, + { + "epoch": 0.4112355468040343, + "grad_norm": 1.0031123161315918, + "learning_rate": 0.00012756711861608335, + "loss": 0.1246, + "step": 17543 + }, + { + "epoch": 0.4112589883788393, + "grad_norm": 0.6503029465675354, + "learning_rate": 0.0001275600378073916, + "loss": 0.1097, + "step": 17544 + }, + { + "epoch": 0.4112824299536443, + "grad_norm": 0.4670427441596985, + "learning_rate": 0.00012755295684915833, + "loss": 0.0661, + "step": 17545 + }, + { + "epoch": 0.4113058715284493, + "grad_norm": 0.5620805025100708, + "learning_rate": 0.000127545875741422, + "loss": 0.5777, + "step": 17546 + }, + { + "epoch": 0.41132931310325427, + "grad_norm": 0.4771442711353302, + "learning_rate": 0.00012753879448422094, + "loss": 0.1329, + "step": 17547 + }, + { + "epoch": 0.41135275467805926, + "grad_norm": 0.670247495174408, + "learning_rate": 0.00012753171307759356, + "loss": 0.1123, + "step": 17548 + }, + { + "epoch": 0.41137619625286426, + "grad_norm": 0.6680316925048828, + "learning_rate": 0.0001275246315215784, + "loss": 0.1399, + "step": 17549 + }, + { + "epoch": 0.41139963782766925, + "grad_norm": 0.37985959649086, + "learning_rate": 0.00012751754981621378, + "loss": 0.0671, + "step": 17550 + }, + { + "epoch": 0.41142307940247425, + "grad_norm": 0.3051097095012665, + "learning_rate": 0.00012751046796153815, + "loss": 0.0718, + "step": 17551 + }, + { + "epoch": 0.41144652097727924, + "grad_norm": 0.33780959248542786, + "learning_rate": 0.00012750338595758997, + "loss": 0.0556, + "step": 17552 + }, + { + "epoch": 0.41146996255208423, + "grad_norm": 0.6661794185638428, + "learning_rate": 0.00012749630380440766, + "loss": 0.169, + "step": 17553 + }, + { + "epoch": 0.41149340412688923, + "grad_norm": 0.40906158089637756, + "learning_rate": 0.00012748922150202958, + "loss": 0.1013, + "step": 17554 + }, + { + "epoch": 0.4115168457016942, + "grad_norm": 0.642262876033783, + "learning_rate": 0.00012748213905049423, + "loss": 0.1561, + "step": 17555 + }, + { + "epoch": 0.4115402872764992, + "grad_norm": 0.2578464448451996, + "learning_rate": 0.00012747505644984005, + "loss": 0.0706, + "step": 17556 + }, + { + "epoch": 0.4115637288513042, + "grad_norm": 0.6640239953994751, + "learning_rate": 0.00012746797370010544, + "loss": 0.1954, + "step": 17557 + }, + { + "epoch": 0.4115871704261092, + "grad_norm": 0.2201426923274994, + "learning_rate": 0.00012746089080132878, + "loss": 0.0631, + "step": 17558 + }, + { + "epoch": 0.4116106120009142, + "grad_norm": 0.19520220160484314, + "learning_rate": 0.00012745380775354862, + "loss": 0.0229, + "step": 17559 + }, + { + "epoch": 0.4116340535757192, + "grad_norm": 0.18038231134414673, + "learning_rate": 0.00012744672455680328, + "loss": 0.016, + "step": 17560 + }, + { + "epoch": 0.4116574951505242, + "grad_norm": 0.29775330424308777, + "learning_rate": 0.00012743964121113125, + "loss": 0.0568, + "step": 17561 + }, + { + "epoch": 0.4116809367253292, + "grad_norm": 0.3868085443973541, + "learning_rate": 0.000127432557716571, + "loss": 0.1084, + "step": 17562 + }, + { + "epoch": 0.4117043783001342, + "grad_norm": 0.2663519084453583, + "learning_rate": 0.00012742547407316085, + "loss": 0.0538, + "step": 17563 + }, + { + "epoch": 0.41172781987493917, + "grad_norm": 0.25590309500694275, + "learning_rate": 0.00012741839028093934, + "loss": 0.0585, + "step": 17564 + }, + { + "epoch": 0.4117512614497442, + "grad_norm": 0.39440906047821045, + "learning_rate": 0.00012741130633994488, + "loss": 0.0795, + "step": 17565 + }, + { + "epoch": 0.4117747030245492, + "grad_norm": 0.4996785819530487, + "learning_rate": 0.00012740422225021591, + "loss": 0.0849, + "step": 17566 + }, + { + "epoch": 0.4117981445993542, + "grad_norm": 0.363445520401001, + "learning_rate": 0.00012739713801179084, + "loss": 0.0656, + "step": 17567 + }, + { + "epoch": 0.4118215861741592, + "grad_norm": 0.3726573586463928, + "learning_rate": 0.00012739005362470812, + "loss": 0.0622, + "step": 17568 + }, + { + "epoch": 0.4118450277489642, + "grad_norm": 0.5611435770988464, + "learning_rate": 0.00012738296908900622, + "loss": 0.6423, + "step": 17569 + }, + { + "epoch": 0.4118684693237692, + "grad_norm": 0.38617080450057983, + "learning_rate": 0.00012737588440472355, + "loss": 0.0911, + "step": 17570 + }, + { + "epoch": 0.4118919108985742, + "grad_norm": 0.5997442603111267, + "learning_rate": 0.00012736879957189856, + "loss": 0.6152, + "step": 17571 + }, + { + "epoch": 0.4119153524733792, + "grad_norm": 0.42778506875038147, + "learning_rate": 0.0001273617145905697, + "loss": 0.0935, + "step": 17572 + }, + { + "epoch": 0.4119387940481842, + "grad_norm": 0.26022863388061523, + "learning_rate": 0.0001273546294607754, + "loss": 0.0361, + "step": 17573 + }, + { + "epoch": 0.41196223562298917, + "grad_norm": 0.24662572145462036, + "learning_rate": 0.00012734754418255412, + "loss": 0.0293, + "step": 17574 + }, + { + "epoch": 0.41198567719779416, + "grad_norm": 0.4399275779724121, + "learning_rate": 0.0001273404587559443, + "loss": 0.087, + "step": 17575 + }, + { + "epoch": 0.41200911877259916, + "grad_norm": 0.9210866689682007, + "learning_rate": 0.00012733337318098436, + "loss": 0.0841, + "step": 17576 + }, + { + "epoch": 0.41203256034740415, + "grad_norm": 0.5498445630073547, + "learning_rate": 0.0001273262874577128, + "loss": 0.1281, + "step": 17577 + }, + { + "epoch": 0.41205600192220915, + "grad_norm": 0.6110089421272278, + "learning_rate": 0.000127319201586168, + "loss": 0.1644, + "step": 17578 + }, + { + "epoch": 0.41207944349701414, + "grad_norm": 0.5231370329856873, + "learning_rate": 0.00012731211556638847, + "loss": 0.1055, + "step": 17579 + }, + { + "epoch": 0.41210288507181914, + "grad_norm": 0.6847103834152222, + "learning_rate": 0.00012730502939841263, + "loss": 0.0738, + "step": 17580 + }, + { + "epoch": 0.41212632664662413, + "grad_norm": 0.3741935193538666, + "learning_rate": 0.00012729794308227895, + "loss": 0.0771, + "step": 17581 + }, + { + "epoch": 0.4121497682214291, + "grad_norm": 0.3349286615848541, + "learning_rate": 0.00012729085661802582, + "loss": 0.0596, + "step": 17582 + }, + { + "epoch": 0.4121732097962341, + "grad_norm": 0.4717310965061188, + "learning_rate": 0.00012728377000569173, + "loss": 0.113, + "step": 17583 + }, + { + "epoch": 0.4121966513710391, + "grad_norm": 0.12615591287612915, + "learning_rate": 0.00012727668324531518, + "loss": 0.0229, + "step": 17584 + }, + { + "epoch": 0.4122200929458441, + "grad_norm": 0.545849621295929, + "learning_rate": 0.0001272695963369346, + "loss": 0.1138, + "step": 17585 + }, + { + "epoch": 0.4122435345206491, + "grad_norm": 0.8003215789794922, + "learning_rate": 0.00012726250928058835, + "loss": 0.1125, + "step": 17586 + }, + { + "epoch": 0.4122669760954541, + "grad_norm": 0.3229649066925049, + "learning_rate": 0.000127255422076315, + "loss": 0.0599, + "step": 17587 + }, + { + "epoch": 0.4122904176702591, + "grad_norm": 0.13257846236228943, + "learning_rate": 0.00012724833472415296, + "loss": 0.0165, + "step": 17588 + }, + { + "epoch": 0.4123138592450641, + "grad_norm": 0.5253573656082153, + "learning_rate": 0.00012724124722414068, + "loss": 0.569, + "step": 17589 + }, + { + "epoch": 0.4123373008198691, + "grad_norm": 0.5408599972724915, + "learning_rate": 0.00012723415957631665, + "loss": 0.5043, + "step": 17590 + }, + { + "epoch": 0.4123607423946741, + "grad_norm": 0.5814654231071472, + "learning_rate": 0.0001272270717807193, + "loss": 0.1266, + "step": 17591 + }, + { + "epoch": 0.41238418396947907, + "grad_norm": 1.015707015991211, + "learning_rate": 0.00012721998383738708, + "loss": 0.112, + "step": 17592 + }, + { + "epoch": 0.41240762554428406, + "grad_norm": 0.21777483820915222, + "learning_rate": 0.00012721289574635844, + "loss": 0.0471, + "step": 17593 + }, + { + "epoch": 0.41243106711908906, + "grad_norm": 0.5563521385192871, + "learning_rate": 0.0001272058075076719, + "loss": 0.7294, + "step": 17594 + }, + { + "epoch": 0.41245450869389405, + "grad_norm": 0.6772028207778931, + "learning_rate": 0.00012719871912136586, + "loss": 0.0883, + "step": 17595 + }, + { + "epoch": 0.41247795026869905, + "grad_norm": 0.5405786633491516, + "learning_rate": 0.00012719163058747882, + "loss": 0.1481, + "step": 17596 + }, + { + "epoch": 0.41250139184350404, + "grad_norm": 0.3732033967971802, + "learning_rate": 0.00012718454190604922, + "loss": 0.0371, + "step": 17597 + }, + { + "epoch": 0.41252483341830903, + "grad_norm": 0.8413582444190979, + "learning_rate": 0.00012717745307711553, + "loss": 0.0903, + "step": 17598 + }, + { + "epoch": 0.41254827499311403, + "grad_norm": 0.40190279483795166, + "learning_rate": 0.00012717036410071623, + "loss": 0.0841, + "step": 17599 + }, + { + "epoch": 0.412571716567919, + "grad_norm": 0.5346001982688904, + "learning_rate": 0.00012716327497688974, + "loss": 0.0935, + "step": 17600 + }, + { + "epoch": 0.412595158142724, + "grad_norm": 0.12585587799549103, + "learning_rate": 0.0001271561857056746, + "loss": 0.0134, + "step": 17601 + }, + { + "epoch": 0.412618599717529, + "grad_norm": 0.4266885221004486, + "learning_rate": 0.0001271490962871092, + "loss": 0.0592, + "step": 17602 + }, + { + "epoch": 0.412642041292334, + "grad_norm": 0.28007829189300537, + "learning_rate": 0.00012714200672123206, + "loss": 0.0595, + "step": 17603 + }, + { + "epoch": 0.412665482867139, + "grad_norm": 0.6626378297805786, + "learning_rate": 0.00012713491700808163, + "loss": 0.1418, + "step": 17604 + }, + { + "epoch": 0.412688924441944, + "grad_norm": 0.44891512393951416, + "learning_rate": 0.00012712782714769636, + "loss": 0.089, + "step": 17605 + }, + { + "epoch": 0.412712366016749, + "grad_norm": 0.18466821312904358, + "learning_rate": 0.00012712073714011473, + "loss": 0.0498, + "step": 17606 + }, + { + "epoch": 0.412735807591554, + "grad_norm": 0.7148712277412415, + "learning_rate": 0.00012711364698537523, + "loss": 0.0686, + "step": 17607 + }, + { + "epoch": 0.412759249166359, + "grad_norm": 0.2589268088340759, + "learning_rate": 0.00012710655668351633, + "loss": 0.0693, + "step": 17608 + }, + { + "epoch": 0.41278269074116397, + "grad_norm": 0.7499342560768127, + "learning_rate": 0.00012709946623457648, + "loss": 0.1127, + "step": 17609 + }, + { + "epoch": 0.41280613231596897, + "grad_norm": 0.7462445497512817, + "learning_rate": 0.0001270923756385942, + "loss": 0.0999, + "step": 17610 + }, + { + "epoch": 0.41282957389077396, + "grad_norm": 0.11596372723579407, + "learning_rate": 0.00012708528489560784, + "loss": 0.0219, + "step": 17611 + }, + { + "epoch": 0.41285301546557895, + "grad_norm": 0.6336319446563721, + "learning_rate": 0.00012707819400565603, + "loss": 0.0529, + "step": 17612 + }, + { + "epoch": 0.41287645704038395, + "grad_norm": 0.8342041969299316, + "learning_rate": 0.00012707110296877714, + "loss": 0.1378, + "step": 17613 + }, + { + "epoch": 0.41289989861518894, + "grad_norm": 0.3616504371166229, + "learning_rate": 0.00012706401178500974, + "loss": 0.1044, + "step": 17614 + }, + { + "epoch": 0.41292334018999394, + "grad_norm": 0.3671642541885376, + "learning_rate": 0.0001270569204543922, + "loss": 0.0779, + "step": 17615 + }, + { + "epoch": 0.412946781764799, + "grad_norm": 0.3351990580558777, + "learning_rate": 0.0001270498289769631, + "loss": 0.062, + "step": 17616 + }, + { + "epoch": 0.412970223339604, + "grad_norm": 0.33478280901908875, + "learning_rate": 0.0001270427373527608, + "loss": 0.0397, + "step": 17617 + }, + { + "epoch": 0.412993664914409, + "grad_norm": 0.5575921535491943, + "learning_rate": 0.00012703564558182387, + "loss": 0.1027, + "step": 17618 + }, + { + "epoch": 0.41301710648921397, + "grad_norm": 0.8689039349555969, + "learning_rate": 0.0001270285536641908, + "loss": 0.2549, + "step": 17619 + }, + { + "epoch": 0.41304054806401896, + "grad_norm": 0.2518700361251831, + "learning_rate": 0.00012702146159990002, + "loss": 0.0497, + "step": 17620 + }, + { + "epoch": 0.41306398963882396, + "grad_norm": 0.2904057204723358, + "learning_rate": 0.00012701436938899002, + "loss": 0.0308, + "step": 17621 + }, + { + "epoch": 0.41308743121362895, + "grad_norm": 0.5299798846244812, + "learning_rate": 0.0001270072770314993, + "loss": 0.0863, + "step": 17622 + }, + { + "epoch": 0.41311087278843395, + "grad_norm": 0.6052252650260925, + "learning_rate": 0.00012700018452746633, + "loss": 0.6783, + "step": 17623 + }, + { + "epoch": 0.41313431436323894, + "grad_norm": 0.11981169134378433, + "learning_rate": 0.0001269930918769296, + "loss": 0.0081, + "step": 17624 + }, + { + "epoch": 0.41315775593804394, + "grad_norm": 0.4538418650627136, + "learning_rate": 0.0001269859990799276, + "loss": 0.1442, + "step": 17625 + }, + { + "epoch": 0.41318119751284893, + "grad_norm": 0.08285674452781677, + "learning_rate": 0.00012697890613649882, + "loss": 0.0143, + "step": 17626 + }, + { + "epoch": 0.4132046390876539, + "grad_norm": 0.3101785480976105, + "learning_rate": 0.0001269718130466817, + "loss": 0.0795, + "step": 17627 + }, + { + "epoch": 0.4132280806624589, + "grad_norm": 0.7638940215110779, + "learning_rate": 0.00012696471981051482, + "loss": 0.1238, + "step": 17628 + }, + { + "epoch": 0.4132515222372639, + "grad_norm": 0.36772406101226807, + "learning_rate": 0.00012695762642803656, + "loss": 0.064, + "step": 17629 + }, + { + "epoch": 0.4132749638120689, + "grad_norm": 0.25127899646759033, + "learning_rate": 0.00012695053289928549, + "loss": 0.0393, + "step": 17630 + }, + { + "epoch": 0.4132984053868739, + "grad_norm": 0.16676737368106842, + "learning_rate": 0.00012694343922430005, + "loss": 0.0242, + "step": 17631 + }, + { + "epoch": 0.4133218469616789, + "grad_norm": 0.6736819744110107, + "learning_rate": 0.00012693634540311875, + "loss": 0.1097, + "step": 17632 + }, + { + "epoch": 0.4133452885364839, + "grad_norm": 0.19421035051345825, + "learning_rate": 0.00012692925143578015, + "loss": 0.0452, + "step": 17633 + }, + { + "epoch": 0.4133687301112889, + "grad_norm": 0.39838650822639465, + "learning_rate": 0.00012692215732232258, + "loss": 0.0634, + "step": 17634 + }, + { + "epoch": 0.4133921716860939, + "grad_norm": 0.3642065227031708, + "learning_rate": 0.0001269150630627847, + "loss": 0.0644, + "step": 17635 + }, + { + "epoch": 0.4134156132608989, + "grad_norm": 0.9876407980918884, + "learning_rate": 0.00012690796865720487, + "loss": 0.161, + "step": 17636 + }, + { + "epoch": 0.41343905483570387, + "grad_norm": 0.4624843895435333, + "learning_rate": 0.00012690087410562165, + "loss": 0.0949, + "step": 17637 + }, + { + "epoch": 0.41346249641050886, + "grad_norm": 0.25351834297180176, + "learning_rate": 0.00012689377940807356, + "loss": 0.0474, + "step": 17638 + }, + { + "epoch": 0.41348593798531386, + "grad_norm": 0.5937238335609436, + "learning_rate": 0.0001268866845645991, + "loss": 0.1527, + "step": 17639 + }, + { + "epoch": 0.41350937956011885, + "grad_norm": 0.5791309475898743, + "learning_rate": 0.00012687958957523666, + "loss": 0.0998, + "step": 17640 + }, + { + "epoch": 0.41353282113492384, + "grad_norm": 0.6378076672554016, + "learning_rate": 0.00012687249444002488, + "loss": 0.1639, + "step": 17641 + }, + { + "epoch": 0.41355626270972884, + "grad_norm": 0.4583111107349396, + "learning_rate": 0.00012686539915900214, + "loss": 0.5589, + "step": 17642 + }, + { + "epoch": 0.41357970428453383, + "grad_norm": 0.4103587567806244, + "learning_rate": 0.00012685830373220699, + "loss": 0.1122, + "step": 17643 + }, + { + "epoch": 0.4136031458593388, + "grad_norm": 0.23473933339118958, + "learning_rate": 0.00012685120815967795, + "loss": 0.0717, + "step": 17644 + }, + { + "epoch": 0.4136265874341438, + "grad_norm": 0.21141324937343597, + "learning_rate": 0.0001268441124414535, + "loss": 0.0564, + "step": 17645 + }, + { + "epoch": 0.4136500290089488, + "grad_norm": 0.2214125245809555, + "learning_rate": 0.0001268370165775721, + "loss": 0.0458, + "step": 17646 + }, + { + "epoch": 0.4136734705837538, + "grad_norm": 1.7169780731201172, + "learning_rate": 0.00012682992056807236, + "loss": 0.1358, + "step": 17647 + }, + { + "epoch": 0.4136969121585588, + "grad_norm": 0.13623230159282684, + "learning_rate": 0.00012682282441299267, + "loss": 0.0181, + "step": 17648 + }, + { + "epoch": 0.4137203537333638, + "grad_norm": 0.47331002354621887, + "learning_rate": 0.00012681572811237162, + "loss": 0.0708, + "step": 17649 + }, + { + "epoch": 0.4137437953081688, + "grad_norm": 0.4570799767971039, + "learning_rate": 0.00012680863166624764, + "loss": 0.1107, + "step": 17650 + }, + { + "epoch": 0.4137672368829738, + "grad_norm": 0.571949303150177, + "learning_rate": 0.00012680153507465928, + "loss": 0.5667, + "step": 17651 + }, + { + "epoch": 0.4137906784577788, + "grad_norm": 0.6220473647117615, + "learning_rate": 0.00012679443833764505, + "loss": 0.1202, + "step": 17652 + }, + { + "epoch": 0.4138141200325838, + "grad_norm": 0.47003811597824097, + "learning_rate": 0.00012678734145524342, + "loss": 0.0613, + "step": 17653 + }, + { + "epoch": 0.41383756160738877, + "grad_norm": 0.5067164897918701, + "learning_rate": 0.00012678024442749292, + "loss": 0.1281, + "step": 17654 + }, + { + "epoch": 0.41386100318219377, + "grad_norm": 0.16048811376094818, + "learning_rate": 0.0001267731472544321, + "loss": 0.0233, + "step": 17655 + }, + { + "epoch": 0.41388444475699876, + "grad_norm": 0.4557516872882843, + "learning_rate": 0.0001267660499360994, + "loss": 0.0424, + "step": 17656 + }, + { + "epoch": 0.41390788633180375, + "grad_norm": 0.49607786536216736, + "learning_rate": 0.00012675895247253336, + "loss": 0.0918, + "step": 17657 + }, + { + "epoch": 0.41393132790660875, + "grad_norm": 0.5399413108825684, + "learning_rate": 0.0001267518548637725, + "loss": 0.5258, + "step": 17658 + }, + { + "epoch": 0.41395476948141374, + "grad_norm": 0.31178995966911316, + "learning_rate": 0.0001267447571098553, + "loss": 0.0647, + "step": 17659 + }, + { + "epoch": 0.41397821105621874, + "grad_norm": 0.375895231962204, + "learning_rate": 0.00012673765921082032, + "loss": 0.0633, + "step": 17660 + }, + { + "epoch": 0.41400165263102373, + "grad_norm": 0.3850369453430176, + "learning_rate": 0.00012673056116670604, + "loss": 0.066, + "step": 17661 + }, + { + "epoch": 0.4140250942058287, + "grad_norm": 0.24434149265289307, + "learning_rate": 0.000126723462977551, + "loss": 0.0454, + "step": 17662 + }, + { + "epoch": 0.4140485357806337, + "grad_norm": 0.4443809986114502, + "learning_rate": 0.00012671636464339368, + "loss": 0.0889, + "step": 17663 + }, + { + "epoch": 0.4140719773554387, + "grad_norm": 0.49359816312789917, + "learning_rate": 0.00012670926616427267, + "loss": 0.0878, + "step": 17664 + }, + { + "epoch": 0.4140954189302437, + "grad_norm": 0.5658742785453796, + "learning_rate": 0.00012670216754022634, + "loss": 0.0659, + "step": 17665 + }, + { + "epoch": 0.4141188605050487, + "grad_norm": 0.4053058922290802, + "learning_rate": 0.00012669506877129334, + "loss": 0.0639, + "step": 17666 + }, + { + "epoch": 0.4141423020798537, + "grad_norm": 0.29614973068237305, + "learning_rate": 0.00012668796985751217, + "loss": 0.0506, + "step": 17667 + }, + { + "epoch": 0.41416574365465875, + "grad_norm": 0.5336498618125916, + "learning_rate": 0.0001266808707989213, + "loss": 0.0598, + "step": 17668 + }, + { + "epoch": 0.41418918522946374, + "grad_norm": 0.610877275466919, + "learning_rate": 0.00012667377159555925, + "loss": 0.0765, + "step": 17669 + }, + { + "epoch": 0.41421262680426874, + "grad_norm": 0.3606583774089813, + "learning_rate": 0.00012666667224746461, + "loss": 0.0814, + "step": 17670 + }, + { + "epoch": 0.41423606837907373, + "grad_norm": 0.8723753690719604, + "learning_rate": 0.00012665957275467585, + "loss": 0.1804, + "step": 17671 + }, + { + "epoch": 0.4142595099538787, + "grad_norm": 0.8053492903709412, + "learning_rate": 0.00012665247311723147, + "loss": 0.1367, + "step": 17672 + }, + { + "epoch": 0.4142829515286837, + "grad_norm": 0.4397830665111542, + "learning_rate": 0.00012664537333517006, + "loss": 0.089, + "step": 17673 + }, + { + "epoch": 0.4143063931034887, + "grad_norm": 0.633624792098999, + "learning_rate": 0.00012663827340853007, + "loss": 0.1224, + "step": 17674 + }, + { + "epoch": 0.4143298346782937, + "grad_norm": 0.41091790795326233, + "learning_rate": 0.00012663117333735005, + "loss": 0.0684, + "step": 17675 + }, + { + "epoch": 0.4143532762530987, + "grad_norm": 0.5494083166122437, + "learning_rate": 0.0001266240731216686, + "loss": 0.0935, + "step": 17676 + }, + { + "epoch": 0.4143767178279037, + "grad_norm": 0.332810640335083, + "learning_rate": 0.00012661697276152413, + "loss": 0.0666, + "step": 17677 + }, + { + "epoch": 0.4144001594027087, + "grad_norm": 0.45426544547080994, + "learning_rate": 0.00012660987225695522, + "loss": 0.084, + "step": 17678 + }, + { + "epoch": 0.4144236009775137, + "grad_norm": 0.7192975282669067, + "learning_rate": 0.0001266027716080004, + "loss": 0.5831, + "step": 17679 + }, + { + "epoch": 0.4144470425523187, + "grad_norm": 0.534134566783905, + "learning_rate": 0.0001265956708146982, + "loss": 0.0628, + "step": 17680 + }, + { + "epoch": 0.4144704841271237, + "grad_norm": 0.6050258874893188, + "learning_rate": 0.00012658856987708713, + "loss": 0.6614, + "step": 17681 + }, + { + "epoch": 0.41449392570192867, + "grad_norm": 0.31917697191238403, + "learning_rate": 0.00012658146879520574, + "loss": 0.0724, + "step": 17682 + }, + { + "epoch": 0.41451736727673366, + "grad_norm": 0.41029757261276245, + "learning_rate": 0.00012657436756909255, + "loss": 0.1001, + "step": 17683 + }, + { + "epoch": 0.41454080885153866, + "grad_norm": 0.4128530025482178, + "learning_rate": 0.0001265672661987861, + "loss": 0.0855, + "step": 17684 + }, + { + "epoch": 0.41456425042634365, + "grad_norm": 0.2013026475906372, + "learning_rate": 0.0001265601646843249, + "loss": 0.0349, + "step": 17685 + }, + { + "epoch": 0.41458769200114864, + "grad_norm": 0.6050139665603638, + "learning_rate": 0.00012655306302574752, + "loss": 0.1829, + "step": 17686 + }, + { + "epoch": 0.41461113357595364, + "grad_norm": 0.6503193974494934, + "learning_rate": 0.00012654596122309248, + "loss": 0.1436, + "step": 17687 + }, + { + "epoch": 0.41463457515075863, + "grad_norm": 0.22645509243011475, + "learning_rate": 0.00012653885927639829, + "loss": 0.043, + "step": 17688 + }, + { + "epoch": 0.4146580167255636, + "grad_norm": 0.35215267539024353, + "learning_rate": 0.00012653175718570353, + "loss": 0.0766, + "step": 17689 + }, + { + "epoch": 0.4146814583003686, + "grad_norm": 0.20765964686870575, + "learning_rate": 0.0001265246549510467, + "loss": 0.0491, + "step": 17690 + }, + { + "epoch": 0.4147048998751736, + "grad_norm": 0.15359093248844147, + "learning_rate": 0.00012651755257246628, + "loss": 0.0232, + "step": 17691 + }, + { + "epoch": 0.4147283414499786, + "grad_norm": 0.38061171770095825, + "learning_rate": 0.00012651045005000096, + "loss": 0.1213, + "step": 17692 + }, + { + "epoch": 0.4147517830247836, + "grad_norm": 0.6693227291107178, + "learning_rate": 0.00012650334738368915, + "loss": 0.136, + "step": 17693 + }, + { + "epoch": 0.4147752245995886, + "grad_norm": 0.31487932801246643, + "learning_rate": 0.00012649624457356943, + "loss": 0.0491, + "step": 17694 + }, + { + "epoch": 0.4147986661743936, + "grad_norm": 0.5899526476860046, + "learning_rate": 0.00012648914161968037, + "loss": 0.1578, + "step": 17695 + }, + { + "epoch": 0.4148221077491986, + "grad_norm": 0.17340397834777832, + "learning_rate": 0.00012648203852206046, + "loss": 0.0422, + "step": 17696 + }, + { + "epoch": 0.4148455493240036, + "grad_norm": 0.6805101633071899, + "learning_rate": 0.00012647493528074826, + "loss": 0.6257, + "step": 17697 + }, + { + "epoch": 0.4148689908988086, + "grad_norm": 0.5888572931289673, + "learning_rate": 0.00012646783189578233, + "loss": 0.1044, + "step": 17698 + }, + { + "epoch": 0.41489243247361357, + "grad_norm": 1.435436725616455, + "learning_rate": 0.0001264607283672012, + "loss": 0.0899, + "step": 17699 + }, + { + "epoch": 0.41491587404841856, + "grad_norm": 0.7078174948692322, + "learning_rate": 0.0001264536246950434, + "loss": 0.0798, + "step": 17700 + }, + { + "epoch": 0.41493931562322356, + "grad_norm": 0.2540208399295807, + "learning_rate": 0.00012644652087934752, + "loss": 0.0437, + "step": 17701 + }, + { + "epoch": 0.41496275719802855, + "grad_norm": 0.4132293462753296, + "learning_rate": 0.00012643941692015208, + "loss": 0.0578, + "step": 17702 + }, + { + "epoch": 0.41498619877283355, + "grad_norm": 0.4793316423892975, + "learning_rate": 0.00012643231281749557, + "loss": 0.0886, + "step": 17703 + }, + { + "epoch": 0.41500964034763854, + "grad_norm": 0.6090397238731384, + "learning_rate": 0.0001264252085714166, + "loss": 0.182, + "step": 17704 + }, + { + "epoch": 0.41503308192244354, + "grad_norm": 0.6882587671279907, + "learning_rate": 0.00012641810418195372, + "loss": 0.0775, + "step": 17705 + }, + { + "epoch": 0.41505652349724853, + "grad_norm": 0.7127084136009216, + "learning_rate": 0.00012641099964914548, + "loss": 0.096, + "step": 17706 + }, + { + "epoch": 0.4150799650720535, + "grad_norm": 0.6145640015602112, + "learning_rate": 0.00012640389497303037, + "loss": 0.615, + "step": 17707 + }, + { + "epoch": 0.4151034066468585, + "grad_norm": 0.03388654440641403, + "learning_rate": 0.00012639679015364702, + "loss": 0.0058, + "step": 17708 + }, + { + "epoch": 0.4151268482216635, + "grad_norm": 0.20414461195468903, + "learning_rate": 0.00012638968519103393, + "loss": 0.0348, + "step": 17709 + }, + { + "epoch": 0.4151502897964685, + "grad_norm": 0.46794819831848145, + "learning_rate": 0.00012638258008522966, + "loss": 0.0641, + "step": 17710 + }, + { + "epoch": 0.4151737313712735, + "grad_norm": 0.5024659037590027, + "learning_rate": 0.00012637547483627278, + "loss": 0.1078, + "step": 17711 + }, + { + "epoch": 0.4151971729460785, + "grad_norm": 0.5857095122337341, + "learning_rate": 0.00012636836944420184, + "loss": 0.6034, + "step": 17712 + }, + { + "epoch": 0.4152206145208835, + "grad_norm": 0.5403420925140381, + "learning_rate": 0.00012636126390905535, + "loss": 0.6751, + "step": 17713 + }, + { + "epoch": 0.4152440560956885, + "grad_norm": 0.3876228928565979, + "learning_rate": 0.00012635415823087194, + "loss": 0.891, + "step": 17714 + }, + { + "epoch": 0.4152674976704935, + "grad_norm": 0.1619645655155182, + "learning_rate": 0.0001263470524096901, + "loss": 0.0407, + "step": 17715 + }, + { + "epoch": 0.4152909392452985, + "grad_norm": 0.24739013612270355, + "learning_rate": 0.00012633994644554842, + "loss": 0.0363, + "step": 17716 + }, + { + "epoch": 0.41531438082010347, + "grad_norm": 0.3054184913635254, + "learning_rate": 0.00012633284033848548, + "loss": 0.0788, + "step": 17717 + }, + { + "epoch": 0.41533782239490846, + "grad_norm": 0.8022623658180237, + "learning_rate": 0.0001263257340885398, + "loss": 0.1655, + "step": 17718 + }, + { + "epoch": 0.4153612639697135, + "grad_norm": 0.463748037815094, + "learning_rate": 0.00012631862769574992, + "loss": 0.0755, + "step": 17719 + }, + { + "epoch": 0.4153847055445185, + "grad_norm": 0.6374783515930176, + "learning_rate": 0.00012631152116015444, + "loss": 0.1137, + "step": 17720 + }, + { + "epoch": 0.4154081471193235, + "grad_norm": 0.6623714566230774, + "learning_rate": 0.0001263044144817919, + "loss": 0.6622, + "step": 17721 + }, + { + "epoch": 0.4154315886941285, + "grad_norm": 1.0873931646347046, + "learning_rate": 0.00012629730766070086, + "loss": 0.2386, + "step": 17722 + }, + { + "epoch": 0.4154550302689335, + "grad_norm": 0.18610070645809174, + "learning_rate": 0.0001262902006969199, + "loss": 0.0407, + "step": 17723 + }, + { + "epoch": 0.4154784718437385, + "grad_norm": 0.6006312966346741, + "learning_rate": 0.00012628309359048758, + "loss": 0.1014, + "step": 17724 + }, + { + "epoch": 0.4155019134185435, + "grad_norm": 0.5806153416633606, + "learning_rate": 0.00012627598634144243, + "loss": 0.1212, + "step": 17725 + }, + { + "epoch": 0.4155253549933485, + "grad_norm": 0.5050919651985168, + "learning_rate": 0.00012626887894982307, + "loss": 0.0781, + "step": 17726 + }, + { + "epoch": 0.41554879656815347, + "grad_norm": 0.36245325207710266, + "learning_rate": 0.000126261771415668, + "loss": 0.3946, + "step": 17727 + }, + { + "epoch": 0.41557223814295846, + "grad_norm": 0.299697607755661, + "learning_rate": 0.00012625466373901583, + "loss": 0.0667, + "step": 17728 + }, + { + "epoch": 0.41559567971776346, + "grad_norm": 0.34101417660713196, + "learning_rate": 0.00012624755591990515, + "loss": 0.034, + "step": 17729 + }, + { + "epoch": 0.41561912129256845, + "grad_norm": 0.40795373916625977, + "learning_rate": 0.00012624044795837445, + "loss": 0.0698, + "step": 17730 + }, + { + "epoch": 0.41564256286737344, + "grad_norm": 0.8129117488861084, + "learning_rate": 0.0001262333398544624, + "loss": 0.1292, + "step": 17731 + }, + { + "epoch": 0.41566600444217844, + "grad_norm": 0.12468592077493668, + "learning_rate": 0.00012622623160820745, + "loss": 0.0241, + "step": 17732 + }, + { + "epoch": 0.41568944601698343, + "grad_norm": 0.6516845226287842, + "learning_rate": 0.00012621912321964825, + "loss": 0.1081, + "step": 17733 + }, + { + "epoch": 0.4157128875917884, + "grad_norm": 0.2959563434123993, + "learning_rate": 0.00012621201468882335, + "loss": 0.0669, + "step": 17734 + }, + { + "epoch": 0.4157363291665934, + "grad_norm": 0.4239794611930847, + "learning_rate": 0.00012620490601577132, + "loss": 0.1086, + "step": 17735 + }, + { + "epoch": 0.4157597707413984, + "grad_norm": 0.11062194406986237, + "learning_rate": 0.00012619779720053073, + "loss": 0.0202, + "step": 17736 + }, + { + "epoch": 0.4157832123162034, + "grad_norm": 0.1595527082681656, + "learning_rate": 0.00012619068824314017, + "loss": 0.0355, + "step": 17737 + }, + { + "epoch": 0.4158066538910084, + "grad_norm": 0.644401490688324, + "learning_rate": 0.0001261835791436382, + "loss": 0.1886, + "step": 17738 + }, + { + "epoch": 0.4158300954658134, + "grad_norm": 1.1370041370391846, + "learning_rate": 0.00012617646990206337, + "loss": 0.2111, + "step": 17739 + }, + { + "epoch": 0.4158535370406184, + "grad_norm": 0.45367881655693054, + "learning_rate": 0.0001261693605184543, + "loss": 0.0891, + "step": 17740 + }, + { + "epoch": 0.4158769786154234, + "grad_norm": 0.11070843786001205, + "learning_rate": 0.00012616225099284953, + "loss": 0.023, + "step": 17741 + }, + { + "epoch": 0.4159004201902284, + "grad_norm": 0.08891312777996063, + "learning_rate": 0.00012615514132528766, + "loss": 0.0127, + "step": 17742 + }, + { + "epoch": 0.4159238617650334, + "grad_norm": 0.24724599719047546, + "learning_rate": 0.00012614803151580724, + "loss": 0.0589, + "step": 17743 + }, + { + "epoch": 0.41594730333983837, + "grad_norm": 0.29330381751060486, + "learning_rate": 0.00012614092156444691, + "loss": 0.0511, + "step": 17744 + }, + { + "epoch": 0.41597074491464336, + "grad_norm": 0.4880194067955017, + "learning_rate": 0.00012613381147124513, + "loss": 0.0735, + "step": 17745 + }, + { + "epoch": 0.41599418648944836, + "grad_norm": 0.4045422077178955, + "learning_rate": 0.00012612670123624063, + "loss": 0.1086, + "step": 17746 + }, + { + "epoch": 0.41601762806425335, + "grad_norm": 0.2965463101863861, + "learning_rate": 0.0001261195908594719, + "loss": 0.0826, + "step": 17747 + }, + { + "epoch": 0.41604106963905835, + "grad_norm": 0.6848034858703613, + "learning_rate": 0.0001261124803409775, + "loss": 0.183, + "step": 17748 + }, + { + "epoch": 0.41606451121386334, + "grad_norm": 0.8820486664772034, + "learning_rate": 0.00012610536968079607, + "loss": 0.6273, + "step": 17749 + }, + { + "epoch": 0.41608795278866834, + "grad_norm": 0.5260997414588928, + "learning_rate": 0.00012609825887896617, + "loss": 0.1427, + "step": 17750 + }, + { + "epoch": 0.41611139436347333, + "grad_norm": 0.39579468965530396, + "learning_rate": 0.00012609114793552637, + "loss": 0.0623, + "step": 17751 + }, + { + "epoch": 0.4161348359382783, + "grad_norm": 0.6255387663841248, + "learning_rate": 0.00012608403685051526, + "loss": 0.2029, + "step": 17752 + }, + { + "epoch": 0.4161582775130833, + "grad_norm": 0.6685929298400879, + "learning_rate": 0.00012607692562397146, + "loss": 0.3282, + "step": 17753 + }, + { + "epoch": 0.4161817190878883, + "grad_norm": 0.5073458552360535, + "learning_rate": 0.00012606981425593351, + "loss": 0.1027, + "step": 17754 + }, + { + "epoch": 0.4162051606626933, + "grad_norm": 0.46488332748413086, + "learning_rate": 0.00012606270274644006, + "loss": 0.1047, + "step": 17755 + }, + { + "epoch": 0.4162286022374983, + "grad_norm": 0.5481857061386108, + "learning_rate": 0.0001260555910955296, + "loss": 0.1623, + "step": 17756 + }, + { + "epoch": 0.4162520438123033, + "grad_norm": 0.43872347474098206, + "learning_rate": 0.00012604847930324078, + "loss": 0.0881, + "step": 17757 + }, + { + "epoch": 0.4162754853871083, + "grad_norm": 0.07668989896774292, + "learning_rate": 0.00012604136736961218, + "loss": 0.0114, + "step": 17758 + }, + { + "epoch": 0.4162989269619133, + "grad_norm": 0.2088543176651001, + "learning_rate": 0.00012603425529468241, + "loss": 0.0397, + "step": 17759 + }, + { + "epoch": 0.4163223685367183, + "grad_norm": 0.22223563492298126, + "learning_rate": 0.00012602714307849002, + "loss": 0.046, + "step": 17760 + }, + { + "epoch": 0.4163458101115233, + "grad_norm": 0.5680258274078369, + "learning_rate": 0.00012602003072107362, + "loss": 0.6587, + "step": 17761 + }, + { + "epoch": 0.41636925168632827, + "grad_norm": 0.27797839045524597, + "learning_rate": 0.0001260129182224718, + "loss": 0.0595, + "step": 17762 + }, + { + "epoch": 0.41639269326113326, + "grad_norm": 0.6094444394111633, + "learning_rate": 0.00012600580558272317, + "loss": 0.6736, + "step": 17763 + }, + { + "epoch": 0.41641613483593826, + "grad_norm": 0.3546416759490967, + "learning_rate": 0.00012599869280186627, + "loss": 0.0354, + "step": 17764 + }, + { + "epoch": 0.41643957641074325, + "grad_norm": 0.5866180658340454, + "learning_rate": 0.00012599157987993977, + "loss": 0.0923, + "step": 17765 + }, + { + "epoch": 0.41646301798554825, + "grad_norm": 0.24072052538394928, + "learning_rate": 0.00012598446681698222, + "loss": 0.0385, + "step": 17766 + }, + { + "epoch": 0.41648645956035324, + "grad_norm": 0.4476650655269623, + "learning_rate": 0.0001259773536130322, + "loss": 0.6925, + "step": 17767 + }, + { + "epoch": 0.41650990113515823, + "grad_norm": 0.5203543305397034, + "learning_rate": 0.00012597024026812835, + "loss": 0.0907, + "step": 17768 + }, + { + "epoch": 0.41653334270996323, + "grad_norm": 0.5455637574195862, + "learning_rate": 0.00012596312678230925, + "loss": 0.1242, + "step": 17769 + }, + { + "epoch": 0.4165567842847683, + "grad_norm": 0.5036846995353699, + "learning_rate": 0.00012595601315561346, + "loss": 0.0779, + "step": 17770 + }, + { + "epoch": 0.41658022585957327, + "grad_norm": 0.0938621386885643, + "learning_rate": 0.00012594889938807966, + "loss": 0.0157, + "step": 17771 + }, + { + "epoch": 0.41660366743437827, + "grad_norm": 0.1859806925058365, + "learning_rate": 0.00012594178547974637, + "loss": 0.0235, + "step": 17772 + }, + { + "epoch": 0.41662710900918326, + "grad_norm": 0.5032287240028381, + "learning_rate": 0.00012593467143065222, + "loss": 0.1115, + "step": 17773 + }, + { + "epoch": 0.41665055058398826, + "grad_norm": 0.4629936218261719, + "learning_rate": 0.00012592755724083585, + "loss": 0.073, + "step": 17774 + }, + { + "epoch": 0.41667399215879325, + "grad_norm": 0.5299308896064758, + "learning_rate": 0.0001259204429103358, + "loss": 0.1349, + "step": 17775 + }, + { + "epoch": 0.41669743373359824, + "grad_norm": 0.8101091980934143, + "learning_rate": 0.0001259133284391907, + "loss": 0.1301, + "step": 17776 + }, + { + "epoch": 0.41672087530840324, + "grad_norm": 0.6266899108886719, + "learning_rate": 0.00012590621382743914, + "loss": 0.1205, + "step": 17777 + }, + { + "epoch": 0.41674431688320823, + "grad_norm": 0.5339972972869873, + "learning_rate": 0.00012589909907511975, + "loss": 0.1158, + "step": 17778 + }, + { + "epoch": 0.4167677584580132, + "grad_norm": 0.3294038474559784, + "learning_rate": 0.00012589198418227113, + "loss": 0.0517, + "step": 17779 + }, + { + "epoch": 0.4167912000328182, + "grad_norm": 0.3666056990623474, + "learning_rate": 0.00012588486914893185, + "loss": 0.0688, + "step": 17780 + }, + { + "epoch": 0.4168146416076232, + "grad_norm": 0.7213447690010071, + "learning_rate": 0.00012587775397514056, + "loss": 0.7438, + "step": 17781 + }, + { + "epoch": 0.4168380831824282, + "grad_norm": 0.48676997423171997, + "learning_rate": 0.00012587063866093584, + "loss": 0.6483, + "step": 17782 + }, + { + "epoch": 0.4168615247572332, + "grad_norm": 0.18755929172039032, + "learning_rate": 0.00012586352320635632, + "loss": 0.038, + "step": 17783 + }, + { + "epoch": 0.4168849663320382, + "grad_norm": 0.6353346109390259, + "learning_rate": 0.00012585640761144057, + "loss": 0.0759, + "step": 17784 + }, + { + "epoch": 0.4169084079068432, + "grad_norm": 0.5773447751998901, + "learning_rate": 0.00012584929187622724, + "loss": 0.1756, + "step": 17785 + }, + { + "epoch": 0.4169318494816482, + "grad_norm": 0.5038147568702698, + "learning_rate": 0.00012584217600075494, + "loss": 0.4697, + "step": 17786 + }, + { + "epoch": 0.4169552910564532, + "grad_norm": 0.6169010996818542, + "learning_rate": 0.00012583505998506228, + "loss": 0.0825, + "step": 17787 + }, + { + "epoch": 0.4169787326312582, + "grad_norm": 0.3399032950401306, + "learning_rate": 0.00012582794382918778, + "loss": 0.0931, + "step": 17788 + }, + { + "epoch": 0.41700217420606317, + "grad_norm": 0.7414070963859558, + "learning_rate": 0.00012582082753317023, + "loss": 0.9126, + "step": 17789 + }, + { + "epoch": 0.41702561578086816, + "grad_norm": 0.6378469467163086, + "learning_rate": 0.00012581371109704806, + "loss": 0.1199, + "step": 17790 + }, + { + "epoch": 0.41704905735567316, + "grad_norm": 0.22655728459358215, + "learning_rate": 0.00012580659452086004, + "loss": 0.0419, + "step": 17791 + }, + { + "epoch": 0.41707249893047815, + "grad_norm": 0.1566033810377121, + "learning_rate": 0.00012579947780464465, + "loss": 0.0247, + "step": 17792 + }, + { + "epoch": 0.41709594050528315, + "grad_norm": 0.35463231801986694, + "learning_rate": 0.0001257923609484406, + "loss": 0.0744, + "step": 17793 + }, + { + "epoch": 0.41711938208008814, + "grad_norm": 0.4390411376953125, + "learning_rate": 0.00012578524395228646, + "loss": 0.0602, + "step": 17794 + }, + { + "epoch": 0.41714282365489314, + "grad_norm": 0.44926509261131287, + "learning_rate": 0.0001257781268162209, + "loss": 0.1102, + "step": 17795 + }, + { + "epoch": 0.41716626522969813, + "grad_norm": 0.582764208316803, + "learning_rate": 0.00012577100954028244, + "loss": 0.1439, + "step": 17796 + }, + { + "epoch": 0.4171897068045031, + "grad_norm": 0.13953731954097748, + "learning_rate": 0.0001257638921245098, + "loss": 0.0329, + "step": 17797 + }, + { + "epoch": 0.4172131483793081, + "grad_norm": 0.649236261844635, + "learning_rate": 0.00012575677456894155, + "loss": 0.0823, + "step": 17798 + }, + { + "epoch": 0.4172365899541131, + "grad_norm": 0.6255751848220825, + "learning_rate": 0.0001257496568736163, + "loss": 0.4735, + "step": 17799 + }, + { + "epoch": 0.4172600315289181, + "grad_norm": 0.14063963294029236, + "learning_rate": 0.0001257425390385727, + "loss": 0.022, + "step": 17800 + }, + { + "epoch": 0.4172834731037231, + "grad_norm": 0.6531091332435608, + "learning_rate": 0.00012573542106384937, + "loss": 0.2167, + "step": 17801 + }, + { + "epoch": 0.4173069146785281, + "grad_norm": 0.5160853862762451, + "learning_rate": 0.0001257283029494849, + "loss": 0.8457, + "step": 17802 + }, + { + "epoch": 0.4173303562533331, + "grad_norm": 0.4037550091743469, + "learning_rate": 0.00012572118469551794, + "loss": 0.0484, + "step": 17803 + }, + { + "epoch": 0.4173537978281381, + "grad_norm": 0.5075722932815552, + "learning_rate": 0.00012571406630198712, + "loss": 0.103, + "step": 17804 + }, + { + "epoch": 0.4173772394029431, + "grad_norm": 0.5962439179420471, + "learning_rate": 0.00012570694776893105, + "loss": 0.0966, + "step": 17805 + }, + { + "epoch": 0.4174006809777481, + "grad_norm": 0.14507977664470673, + "learning_rate": 0.00012569982909638834, + "loss": 0.0297, + "step": 17806 + }, + { + "epoch": 0.41742412255255307, + "grad_norm": 0.438349187374115, + "learning_rate": 0.00012569271028439765, + "loss": 0.0801, + "step": 17807 + }, + { + "epoch": 0.41744756412735806, + "grad_norm": 0.48477301001548767, + "learning_rate": 0.0001256855913329976, + "loss": 0.693, + "step": 17808 + }, + { + "epoch": 0.41747100570216306, + "grad_norm": 0.30039721727371216, + "learning_rate": 0.0001256784722422268, + "loss": 0.0753, + "step": 17809 + }, + { + "epoch": 0.41749444727696805, + "grad_norm": 0.25917288661003113, + "learning_rate": 0.0001256713530121239, + "loss": 0.0422, + "step": 17810 + }, + { + "epoch": 0.41751788885177304, + "grad_norm": 0.09165243804454803, + "learning_rate": 0.00012566423364272747, + "loss": 0.0069, + "step": 17811 + }, + { + "epoch": 0.41754133042657804, + "grad_norm": 0.4966448247432709, + "learning_rate": 0.00012565711413407623, + "loss": 0.0799, + "step": 17812 + }, + { + "epoch": 0.41756477200138303, + "grad_norm": 0.3805251717567444, + "learning_rate": 0.00012564999448620877, + "loss": 0.0753, + "step": 17813 + }, + { + "epoch": 0.417588213576188, + "grad_norm": 0.3437667787075043, + "learning_rate": 0.00012564287469916371, + "loss": 0.0659, + "step": 17814 + }, + { + "epoch": 0.417611655150993, + "grad_norm": 0.4833795130252838, + "learning_rate": 0.00012563575477297968, + "loss": 0.1192, + "step": 17815 + }, + { + "epoch": 0.417635096725798, + "grad_norm": 0.49492040276527405, + "learning_rate": 0.00012562863470769533, + "loss": 0.1885, + "step": 17816 + }, + { + "epoch": 0.417658538300603, + "grad_norm": 0.29721343517303467, + "learning_rate": 0.00012562151450334932, + "loss": 0.2476, + "step": 17817 + }, + { + "epoch": 0.417681979875408, + "grad_norm": 0.6398082971572876, + "learning_rate": 0.0001256143941599802, + "loss": 0.1787, + "step": 17818 + }, + { + "epoch": 0.417705421450213, + "grad_norm": 0.19649149477481842, + "learning_rate": 0.00012560727367762667, + "loss": 0.0473, + "step": 17819 + }, + { + "epoch": 0.417728863025018, + "grad_norm": 0.6620813608169556, + "learning_rate": 0.00012560015305632737, + "loss": 0.6051, + "step": 17820 + }, + { + "epoch": 0.41775230459982304, + "grad_norm": 0.6570800542831421, + "learning_rate": 0.00012559303229612093, + "loss": 0.1314, + "step": 17821 + }, + { + "epoch": 0.41777574617462804, + "grad_norm": 0.28361618518829346, + "learning_rate": 0.00012558591139704594, + "loss": 0.0574, + "step": 17822 + }, + { + "epoch": 0.41779918774943303, + "grad_norm": 0.4339616000652313, + "learning_rate": 0.0001255787903591411, + "loss": 0.1039, + "step": 17823 + }, + { + "epoch": 0.417822629324238, + "grad_norm": 0.7347432971000671, + "learning_rate": 0.00012557166918244502, + "loss": 0.1225, + "step": 17824 + }, + { + "epoch": 0.417846070899043, + "grad_norm": 0.24098563194274902, + "learning_rate": 0.0001255645478669963, + "loss": 0.0243, + "step": 17825 + }, + { + "epoch": 0.417869512473848, + "grad_norm": 0.17962931096553802, + "learning_rate": 0.0001255574264128337, + "loss": 0.0243, + "step": 17826 + }, + { + "epoch": 0.417892954048653, + "grad_norm": 0.5771443843841553, + "learning_rate": 0.00012555030481999576, + "loss": 0.0953, + "step": 17827 + }, + { + "epoch": 0.417916395623458, + "grad_norm": 0.5428943037986755, + "learning_rate": 0.00012554318308852114, + "loss": 0.0854, + "step": 17828 + }, + { + "epoch": 0.417939837198263, + "grad_norm": 0.5940494537353516, + "learning_rate": 0.0001255360612184485, + "loss": 0.0684, + "step": 17829 + }, + { + "epoch": 0.417963278773068, + "grad_norm": 0.6141021251678467, + "learning_rate": 0.00012552893920981644, + "loss": 0.087, + "step": 17830 + }, + { + "epoch": 0.417986720347873, + "grad_norm": 0.6207919120788574, + "learning_rate": 0.00012552181706266367, + "loss": 0.1328, + "step": 17831 + }, + { + "epoch": 0.418010161922678, + "grad_norm": 0.25734809041023254, + "learning_rate": 0.0001255146947770288, + "loss": 0.06, + "step": 17832 + }, + { + "epoch": 0.418033603497483, + "grad_norm": 0.7601072192192078, + "learning_rate": 0.0001255075723529505, + "loss": 0.1571, + "step": 17833 + }, + { + "epoch": 0.41805704507228797, + "grad_norm": 0.7101529836654663, + "learning_rate": 0.00012550044979046736, + "loss": 0.1503, + "step": 17834 + }, + { + "epoch": 0.41808048664709296, + "grad_norm": 0.40771591663360596, + "learning_rate": 0.00012549332708961804, + "loss": 0.0836, + "step": 17835 + }, + { + "epoch": 0.41810392822189796, + "grad_norm": 0.5172186493873596, + "learning_rate": 0.00012548620425044125, + "loss": 0.0985, + "step": 17836 + }, + { + "epoch": 0.41812736979670295, + "grad_norm": 0.4147930443286896, + "learning_rate": 0.00012547908127297558, + "loss": 0.0375, + "step": 17837 + }, + { + "epoch": 0.41815081137150795, + "grad_norm": 0.5080838799476624, + "learning_rate": 0.00012547195815725974, + "loss": 0.0998, + "step": 17838 + }, + { + "epoch": 0.41817425294631294, + "grad_norm": 0.5555517077445984, + "learning_rate": 0.0001254648349033323, + "loss": 0.0912, + "step": 17839 + }, + { + "epoch": 0.41819769452111794, + "grad_norm": 0.4987277388572693, + "learning_rate": 0.00012545771151123195, + "loss": 0.0741, + "step": 17840 + }, + { + "epoch": 0.41822113609592293, + "grad_norm": 0.6992177963256836, + "learning_rate": 0.00012545058798099735, + "loss": 0.1591, + "step": 17841 + }, + { + "epoch": 0.4182445776707279, + "grad_norm": 0.3450790345668793, + "learning_rate": 0.00012544346431266716, + "loss": 0.0771, + "step": 17842 + }, + { + "epoch": 0.4182680192455329, + "grad_norm": 0.33115872740745544, + "learning_rate": 0.00012543634050628, + "loss": 0.0504, + "step": 17843 + }, + { + "epoch": 0.4182914608203379, + "grad_norm": 0.5691986083984375, + "learning_rate": 0.0001254292165618745, + "loss": 0.1118, + "step": 17844 + }, + { + "epoch": 0.4183149023951429, + "grad_norm": 0.14599810540676117, + "learning_rate": 0.00012542209247948943, + "loss": 0.0189, + "step": 17845 + }, + { + "epoch": 0.4183383439699479, + "grad_norm": 0.4041896164417267, + "learning_rate": 0.00012541496825916335, + "loss": 0.0696, + "step": 17846 + }, + { + "epoch": 0.4183617855447529, + "grad_norm": 0.6168891787528992, + "learning_rate": 0.0001254078439009349, + "loss": 0.1238, + "step": 17847 + }, + { + "epoch": 0.4183852271195579, + "grad_norm": 0.3438777029514313, + "learning_rate": 0.00012540071940484284, + "loss": 0.1076, + "step": 17848 + }, + { + "epoch": 0.4184086686943629, + "grad_norm": 0.5691259503364563, + "learning_rate": 0.00012539359477092572, + "loss": 0.1408, + "step": 17849 + }, + { + "epoch": 0.4184321102691679, + "grad_norm": 0.8843030333518982, + "learning_rate": 0.00012538646999922222, + "loss": 0.18, + "step": 17850 + }, + { + "epoch": 0.4184555518439729, + "grad_norm": 0.21494179964065552, + "learning_rate": 0.00012537934508977106, + "loss": 0.0143, + "step": 17851 + }, + { + "epoch": 0.41847899341877787, + "grad_norm": 0.4510642886161804, + "learning_rate": 0.00012537222004261085, + "loss": 0.079, + "step": 17852 + }, + { + "epoch": 0.41850243499358286, + "grad_norm": 0.3585941195487976, + "learning_rate": 0.00012536509485778024, + "loss": 0.0916, + "step": 17853 + }, + { + "epoch": 0.41852587656838786, + "grad_norm": 0.8323943614959717, + "learning_rate": 0.00012535796953531793, + "loss": 0.1007, + "step": 17854 + }, + { + "epoch": 0.41854931814319285, + "grad_norm": 0.37651529908180237, + "learning_rate": 0.00012535084407526258, + "loss": 0.0671, + "step": 17855 + }, + { + "epoch": 0.41857275971799784, + "grad_norm": 0.5204330682754517, + "learning_rate": 0.0001253437184776528, + "loss": 0.1048, + "step": 17856 + }, + { + "epoch": 0.41859620129280284, + "grad_norm": 0.48836320638656616, + "learning_rate": 0.00012533659274252732, + "loss": 0.0761, + "step": 17857 + }, + { + "epoch": 0.41861964286760783, + "grad_norm": 0.05272923782467842, + "learning_rate": 0.00012532946686992478, + "loss": 0.0069, + "step": 17858 + }, + { + "epoch": 0.4186430844424128, + "grad_norm": 0.5588878393173218, + "learning_rate": 0.0001253223408598838, + "loss": 0.1423, + "step": 17859 + }, + { + "epoch": 0.4186665260172178, + "grad_norm": 0.5243292450904846, + "learning_rate": 0.0001253152147124431, + "loss": 0.1201, + "step": 17860 + }, + { + "epoch": 0.4186899675920228, + "grad_norm": 0.7811484336853027, + "learning_rate": 0.00012530808842764138, + "loss": 0.1544, + "step": 17861 + }, + { + "epoch": 0.4187134091668278, + "grad_norm": 0.6071288585662842, + "learning_rate": 0.00012530096200551722, + "loss": 0.0962, + "step": 17862 + }, + { + "epoch": 0.4187368507416328, + "grad_norm": 0.07760170847177505, + "learning_rate": 0.00012529383544610933, + "loss": 0.0176, + "step": 17863 + }, + { + "epoch": 0.4187602923164378, + "grad_norm": 0.7671490907669067, + "learning_rate": 0.00012528670874945637, + "loss": 0.0718, + "step": 17864 + }, + { + "epoch": 0.4187837338912428, + "grad_norm": 0.42576864361763, + "learning_rate": 0.00012527958191559707, + "loss": 0.0965, + "step": 17865 + }, + { + "epoch": 0.4188071754660478, + "grad_norm": 0.114604651927948, + "learning_rate": 0.00012527245494456997, + "loss": 0.0212, + "step": 17866 + }, + { + "epoch": 0.4188306170408528, + "grad_norm": 0.5239590406417847, + "learning_rate": 0.00012526532783641386, + "loss": 0.096, + "step": 17867 + }, + { + "epoch": 0.4188540586156578, + "grad_norm": 0.525140643119812, + "learning_rate": 0.00012525820059116737, + "loss": 0.1046, + "step": 17868 + }, + { + "epoch": 0.41887750019046277, + "grad_norm": 0.5858599543571472, + "learning_rate": 0.00012525107320886914, + "loss": 0.7576, + "step": 17869 + }, + { + "epoch": 0.41890094176526776, + "grad_norm": 0.4560758173465729, + "learning_rate": 0.00012524394568955795, + "loss": 0.111, + "step": 17870 + }, + { + "epoch": 0.41892438334007276, + "grad_norm": 0.43140909075737, + "learning_rate": 0.00012523681803327236, + "loss": 0.0718, + "step": 17871 + }, + { + "epoch": 0.4189478249148778, + "grad_norm": 0.11538049578666687, + "learning_rate": 0.00012522969024005107, + "loss": 0.0321, + "step": 17872 + }, + { + "epoch": 0.4189712664896828, + "grad_norm": 0.5798034071922302, + "learning_rate": 0.00012522256230993278, + "loss": 0.6026, + "step": 17873 + }, + { + "epoch": 0.4189947080644878, + "grad_norm": 0.13037385046482086, + "learning_rate": 0.00012521543424295615, + "loss": 0.0298, + "step": 17874 + }, + { + "epoch": 0.4190181496392928, + "grad_norm": 0.4175902009010315, + "learning_rate": 0.0001252083060391599, + "loss": 0.0814, + "step": 17875 + }, + { + "epoch": 0.4190415912140978, + "grad_norm": 0.5759587287902832, + "learning_rate": 0.00012520117769858265, + "loss": 0.1019, + "step": 17876 + }, + { + "epoch": 0.4190650327889028, + "grad_norm": 0.41143763065338135, + "learning_rate": 0.0001251940492212631, + "loss": 0.0787, + "step": 17877 + }, + { + "epoch": 0.4190884743637078, + "grad_norm": 0.33133360743522644, + "learning_rate": 0.00012518692060723994, + "loss": 0.047, + "step": 17878 + }, + { + "epoch": 0.41911191593851277, + "grad_norm": 0.7474397420883179, + "learning_rate": 0.00012517979185655182, + "loss": 0.0871, + "step": 17879 + }, + { + "epoch": 0.41913535751331776, + "grad_norm": 0.5222281217575073, + "learning_rate": 0.00012517266296923745, + "loss": 0.1428, + "step": 17880 + }, + { + "epoch": 0.41915879908812276, + "grad_norm": 0.5732590556144714, + "learning_rate": 0.00012516553394533554, + "loss": 0.6337, + "step": 17881 + }, + { + "epoch": 0.41918224066292775, + "grad_norm": 0.4302443861961365, + "learning_rate": 0.00012515840478488467, + "loss": 0.0409, + "step": 17882 + }, + { + "epoch": 0.41920568223773275, + "grad_norm": 0.8153172135353088, + "learning_rate": 0.00012515127548792365, + "loss": 0.1648, + "step": 17883 + }, + { + "epoch": 0.41922912381253774, + "grad_norm": 0.40330931544303894, + "learning_rate": 0.00012514414605449106, + "loss": 0.044, + "step": 17884 + }, + { + "epoch": 0.41925256538734273, + "grad_norm": 0.5014044642448425, + "learning_rate": 0.0001251370164846256, + "loss": 0.0842, + "step": 17885 + }, + { + "epoch": 0.41927600696214773, + "grad_norm": 0.13257108628749847, + "learning_rate": 0.00012512988677836604, + "loss": 0.0318, + "step": 17886 + }, + { + "epoch": 0.4192994485369527, + "grad_norm": 0.41104060411453247, + "learning_rate": 0.00012512275693575103, + "loss": 0.0647, + "step": 17887 + }, + { + "epoch": 0.4193228901117577, + "grad_norm": 0.4128212630748749, + "learning_rate": 0.00012511562695681917, + "loss": 0.094, + "step": 17888 + }, + { + "epoch": 0.4193463316865627, + "grad_norm": 0.24927940964698792, + "learning_rate": 0.00012510849684160925, + "loss": 0.0224, + "step": 17889 + }, + { + "epoch": 0.4193697732613677, + "grad_norm": 0.38238492608070374, + "learning_rate": 0.0001251013665901599, + "loss": 0.0583, + "step": 17890 + }, + { + "epoch": 0.4193932148361727, + "grad_norm": 0.31028735637664795, + "learning_rate": 0.00012509423620250983, + "loss": 0.0535, + "step": 17891 + }, + { + "epoch": 0.4194166564109777, + "grad_norm": 0.6153818964958191, + "learning_rate": 0.00012508710567869772, + "loss": 0.1256, + "step": 17892 + }, + { + "epoch": 0.4194400979857827, + "grad_norm": 0.43827489018440247, + "learning_rate": 0.0001250799750187623, + "loss": 0.1701, + "step": 17893 + }, + { + "epoch": 0.4194635395605877, + "grad_norm": 0.8281877040863037, + "learning_rate": 0.0001250728442227422, + "loss": 0.1002, + "step": 17894 + }, + { + "epoch": 0.4194869811353927, + "grad_norm": 0.38761189579963684, + "learning_rate": 0.00012506571329067617, + "loss": 0.1168, + "step": 17895 + }, + { + "epoch": 0.4195104227101977, + "grad_norm": 0.2193010449409485, + "learning_rate": 0.00012505858222260286, + "loss": 0.0352, + "step": 17896 + }, + { + "epoch": 0.41953386428500267, + "grad_norm": 0.7297746539115906, + "learning_rate": 0.00012505145101856098, + "loss": 0.1413, + "step": 17897 + }, + { + "epoch": 0.41955730585980766, + "grad_norm": 0.30867886543273926, + "learning_rate": 0.00012504431967858917, + "loss": 0.0391, + "step": 17898 + }, + { + "epoch": 0.41958074743461266, + "grad_norm": 0.6008310914039612, + "learning_rate": 0.0001250371882027263, + "loss": 0.053, + "step": 17899 + }, + { + "epoch": 0.41960418900941765, + "grad_norm": 0.4749504327774048, + "learning_rate": 0.00012503005659101084, + "loss": 0.0925, + "step": 17900 + }, + { + "epoch": 0.41962763058422264, + "grad_norm": 0.7304711937904358, + "learning_rate": 0.0001250229248434816, + "loss": 0.2628, + "step": 17901 + }, + { + "epoch": 0.41965107215902764, + "grad_norm": 0.5623766183853149, + "learning_rate": 0.0001250157929601773, + "loss": 0.0866, + "step": 17902 + }, + { + "epoch": 0.41967451373383263, + "grad_norm": 0.13681654632091522, + "learning_rate": 0.00012500866094113659, + "loss": 0.0151, + "step": 17903 + }, + { + "epoch": 0.4196979553086376, + "grad_norm": 0.3594019412994385, + "learning_rate": 0.00012500152878639818, + "loss": 0.0951, + "step": 17904 + }, + { + "epoch": 0.4197213968834426, + "grad_norm": 0.5366530418395996, + "learning_rate": 0.0001249943964960008, + "loss": 0.0527, + "step": 17905 + }, + { + "epoch": 0.4197448384582476, + "grad_norm": 0.7940669059753418, + "learning_rate": 0.0001249872640699831, + "loss": 0.2133, + "step": 17906 + }, + { + "epoch": 0.4197682800330526, + "grad_norm": 0.4322478473186493, + "learning_rate": 0.00012498013150838376, + "loss": 0.0572, + "step": 17907 + }, + { + "epoch": 0.4197917216078576, + "grad_norm": 0.357267826795578, + "learning_rate": 0.00012497299881124158, + "loss": 0.0385, + "step": 17908 + }, + { + "epoch": 0.4198151631826626, + "grad_norm": 0.6180031299591064, + "learning_rate": 0.0001249658659785952, + "loss": 0.0892, + "step": 17909 + }, + { + "epoch": 0.4198386047574676, + "grad_norm": 0.2738879323005676, + "learning_rate": 0.00012495873301048328, + "loss": 0.0578, + "step": 17910 + }, + { + "epoch": 0.4198620463322726, + "grad_norm": 0.41220957040786743, + "learning_rate": 0.00012495159990694464, + "loss": 0.0383, + "step": 17911 + }, + { + "epoch": 0.4198854879070776, + "grad_norm": 0.20179012417793274, + "learning_rate": 0.00012494446666801788, + "loss": 0.0464, + "step": 17912 + }, + { + "epoch": 0.4199089294818826, + "grad_norm": 0.13337790966033936, + "learning_rate": 0.00012493733329374177, + "loss": 0.0272, + "step": 17913 + }, + { + "epoch": 0.41993237105668757, + "grad_norm": 0.7588247060775757, + "learning_rate": 0.00012493019978415497, + "loss": 0.1562, + "step": 17914 + }, + { + "epoch": 0.41995581263149256, + "grad_norm": 0.09823830425739288, + "learning_rate": 0.0001249230661392962, + "loss": 0.0134, + "step": 17915 + }, + { + "epoch": 0.41997925420629756, + "grad_norm": 0.78653883934021, + "learning_rate": 0.00012491593235920418, + "loss": 0.1954, + "step": 17916 + }, + { + "epoch": 0.42000269578110255, + "grad_norm": 0.15990157425403595, + "learning_rate": 0.00012490879844391758, + "loss": 0.0296, + "step": 17917 + }, + { + "epoch": 0.42002613735590755, + "grad_norm": 0.5316932201385498, + "learning_rate": 0.0001249016643934752, + "loss": 0.8021, + "step": 17918 + }, + { + "epoch": 0.42004957893071254, + "grad_norm": 0.3695715665817261, + "learning_rate": 0.00012489453020791565, + "loss": 0.3999, + "step": 17919 + }, + { + "epoch": 0.42007302050551754, + "grad_norm": 0.6387342214584351, + "learning_rate": 0.00012488739588727767, + "loss": 0.1378, + "step": 17920 + }, + { + "epoch": 0.42009646208032253, + "grad_norm": 0.8286920189857483, + "learning_rate": 0.0001248802614316, + "loss": 0.1306, + "step": 17921 + }, + { + "epoch": 0.4201199036551275, + "grad_norm": 0.8876656293869019, + "learning_rate": 0.00012487312684092132, + "loss": 0.1284, + "step": 17922 + }, + { + "epoch": 0.4201433452299326, + "grad_norm": 0.5098271369934082, + "learning_rate": 0.00012486599211528036, + "loss": 0.8028, + "step": 17923 + }, + { + "epoch": 0.42016678680473757, + "grad_norm": 0.5010907053947449, + "learning_rate": 0.00012485885725471585, + "loss": 0.1715, + "step": 17924 + }, + { + "epoch": 0.42019022837954256, + "grad_norm": 0.5085670948028564, + "learning_rate": 0.00012485172225926643, + "loss": 0.0897, + "step": 17925 + }, + { + "epoch": 0.42021366995434756, + "grad_norm": 0.5472214221954346, + "learning_rate": 0.00012484458712897092, + "loss": 0.5838, + "step": 17926 + }, + { + "epoch": 0.42023711152915255, + "grad_norm": 0.5239253044128418, + "learning_rate": 0.00012483745186386795, + "loss": 0.6992, + "step": 17927 + }, + { + "epoch": 0.42026055310395755, + "grad_norm": 0.164164200425148, + "learning_rate": 0.00012483031646399627, + "loss": 0.0375, + "step": 17928 + }, + { + "epoch": 0.42028399467876254, + "grad_norm": 0.4653654396533966, + "learning_rate": 0.00012482318092939458, + "loss": 0.1193, + "step": 17929 + }, + { + "epoch": 0.42030743625356753, + "grad_norm": 0.6526607871055603, + "learning_rate": 0.00012481604526010163, + "loss": 0.1233, + "step": 17930 + }, + { + "epoch": 0.42033087782837253, + "grad_norm": 0.3877002000808716, + "learning_rate": 0.00012480890945615615, + "loss": 0.1201, + "step": 17931 + }, + { + "epoch": 0.4203543194031775, + "grad_norm": 0.1443975567817688, + "learning_rate": 0.0001248017735175968, + "loss": 0.0242, + "step": 17932 + }, + { + "epoch": 0.4203777609779825, + "grad_norm": 0.13059285283088684, + "learning_rate": 0.00012479463744446233, + "loss": 0.0314, + "step": 17933 + }, + { + "epoch": 0.4204012025527875, + "grad_norm": 0.746931254863739, + "learning_rate": 0.00012478750123679145, + "loss": 0.207, + "step": 17934 + }, + { + "epoch": 0.4204246441275925, + "grad_norm": 0.40940552949905396, + "learning_rate": 0.00012478036489462293, + "loss": 0.0726, + "step": 17935 + }, + { + "epoch": 0.4204480857023975, + "grad_norm": 0.3305313289165497, + "learning_rate": 0.0001247732284179954, + "loss": 0.0602, + "step": 17936 + }, + { + "epoch": 0.4204715272772025, + "grad_norm": 0.11780773848295212, + "learning_rate": 0.00012476609180694768, + "loss": 0.0238, + "step": 17937 + }, + { + "epoch": 0.4204949688520075, + "grad_norm": 0.19294142723083496, + "learning_rate": 0.00012475895506151844, + "loss": 0.0455, + "step": 17938 + }, + { + "epoch": 0.4205184104268125, + "grad_norm": 0.8619065284729004, + "learning_rate": 0.0001247518181817464, + "loss": 0.1217, + "step": 17939 + }, + { + "epoch": 0.4205418520016175, + "grad_norm": 0.5061706304550171, + "learning_rate": 0.00012474468116767032, + "loss": 0.1317, + "step": 17940 + }, + { + "epoch": 0.42056529357642247, + "grad_norm": 0.5301730036735535, + "learning_rate": 0.0001247375440193289, + "loss": 0.2193, + "step": 17941 + }, + { + "epoch": 0.42058873515122747, + "grad_norm": 0.2456730306148529, + "learning_rate": 0.00012473040673676084, + "loss": 0.0406, + "step": 17942 + }, + { + "epoch": 0.42061217672603246, + "grad_norm": 0.41953352093696594, + "learning_rate": 0.00012472326932000494, + "loss": 0.0753, + "step": 17943 + }, + { + "epoch": 0.42063561830083745, + "grad_norm": 0.17967869341373444, + "learning_rate": 0.00012471613176909987, + "loss": 0.0231, + "step": 17944 + }, + { + "epoch": 0.42065905987564245, + "grad_norm": 0.1500011831521988, + "learning_rate": 0.00012470899408408435, + "loss": 0.0364, + "step": 17945 + }, + { + "epoch": 0.42068250145044744, + "grad_norm": 0.3971990644931793, + "learning_rate": 0.00012470185626499712, + "loss": 0.0852, + "step": 17946 + }, + { + "epoch": 0.42070594302525244, + "grad_norm": 0.3249782621860504, + "learning_rate": 0.00012469471831187698, + "loss": 0.0731, + "step": 17947 + }, + { + "epoch": 0.42072938460005743, + "grad_norm": 0.5994591116905212, + "learning_rate": 0.00012468758022476256, + "loss": 0.1124, + "step": 17948 + }, + { + "epoch": 0.4207528261748624, + "grad_norm": 0.6231318712234497, + "learning_rate": 0.00012468044200369266, + "loss": 0.1319, + "step": 17949 + }, + { + "epoch": 0.4207762677496674, + "grad_norm": 0.38427460193634033, + "learning_rate": 0.000124673303648706, + "loss": 0.0334, + "step": 17950 + }, + { + "epoch": 0.4207997093244724, + "grad_norm": 0.20802967250347137, + "learning_rate": 0.00012466616515984126, + "loss": 0.0547, + "step": 17951 + }, + { + "epoch": 0.4208231508992774, + "grad_norm": 0.20769503712654114, + "learning_rate": 0.00012465902653713723, + "loss": 0.0526, + "step": 17952 + }, + { + "epoch": 0.4208465924740824, + "grad_norm": 0.4576960802078247, + "learning_rate": 0.00012465188778063263, + "loss": 0.4444, + "step": 17953 + }, + { + "epoch": 0.4208700340488874, + "grad_norm": 0.36288490891456604, + "learning_rate": 0.00012464474889036623, + "loss": 0.0379, + "step": 17954 + }, + { + "epoch": 0.4208934756236924, + "grad_norm": 1.2023147344589233, + "learning_rate": 0.0001246376098663767, + "loss": 0.2161, + "step": 17955 + }, + { + "epoch": 0.4209169171984974, + "grad_norm": 0.8568710684776306, + "learning_rate": 0.0001246304707087028, + "loss": 0.1039, + "step": 17956 + }, + { + "epoch": 0.4209403587733024, + "grad_norm": 0.29256585240364075, + "learning_rate": 0.00012462333141738329, + "loss": 0.0635, + "step": 17957 + }, + { + "epoch": 0.4209638003481074, + "grad_norm": 0.9139067530632019, + "learning_rate": 0.00012461619199245685, + "loss": 0.1139, + "step": 17958 + }, + { + "epoch": 0.42098724192291237, + "grad_norm": 0.3023431897163391, + "learning_rate": 0.0001246090524339623, + "loss": 0.0609, + "step": 17959 + }, + { + "epoch": 0.42101068349771736, + "grad_norm": 0.7007490396499634, + "learning_rate": 0.00012460191274193833, + "loss": 0.0885, + "step": 17960 + }, + { + "epoch": 0.42103412507252236, + "grad_norm": 0.3547215163707733, + "learning_rate": 0.0001245947729164237, + "loss": 0.0748, + "step": 17961 + }, + { + "epoch": 0.42105756664732735, + "grad_norm": 0.10818848758935928, + "learning_rate": 0.00012458763295745712, + "loss": 0.012, + "step": 17962 + }, + { + "epoch": 0.42108100822213235, + "grad_norm": 0.6604170203208923, + "learning_rate": 0.00012458049286507737, + "loss": 0.15, + "step": 17963 + }, + { + "epoch": 0.42110444979693734, + "grad_norm": 0.5502769351005554, + "learning_rate": 0.00012457335263932315, + "loss": 0.1698, + "step": 17964 + }, + { + "epoch": 0.42112789137174234, + "grad_norm": 0.44146376848220825, + "learning_rate": 0.00012456621228023324, + "loss": 0.1372, + "step": 17965 + }, + { + "epoch": 0.42115133294654733, + "grad_norm": 0.3703961670398712, + "learning_rate": 0.0001245590717878464, + "loss": 0.057, + "step": 17966 + }, + { + "epoch": 0.4211747745213523, + "grad_norm": 0.6516943573951721, + "learning_rate": 0.0001245519311622013, + "loss": 0.1047, + "step": 17967 + }, + { + "epoch": 0.4211982160961573, + "grad_norm": 0.6383272409439087, + "learning_rate": 0.00012454479040333675, + "loss": 0.097, + "step": 17968 + }, + { + "epoch": 0.4212216576709623, + "grad_norm": 0.33608654141426086, + "learning_rate": 0.0001245376495112915, + "loss": 0.0652, + "step": 17969 + }, + { + "epoch": 0.4212450992457673, + "grad_norm": 0.46762025356292725, + "learning_rate": 0.00012453050848610424, + "loss": 0.0743, + "step": 17970 + }, + { + "epoch": 0.4212685408205723, + "grad_norm": 0.17191191017627716, + "learning_rate": 0.00012452336732781376, + "loss": 0.0198, + "step": 17971 + }, + { + "epoch": 0.4212919823953773, + "grad_norm": 0.5217917561531067, + "learning_rate": 0.0001245162260364588, + "loss": 0.1539, + "step": 17972 + }, + { + "epoch": 0.4213154239701823, + "grad_norm": 0.6216496825218201, + "learning_rate": 0.00012450908461207814, + "loss": 0.1, + "step": 17973 + }, + { + "epoch": 0.42133886554498734, + "grad_norm": 0.4978505074977875, + "learning_rate": 0.00012450194305471045, + "loss": 0.0983, + "step": 17974 + }, + { + "epoch": 0.42136230711979233, + "grad_norm": 0.7180485725402832, + "learning_rate": 0.00012449480136439455, + "loss": 0.2095, + "step": 17975 + }, + { + "epoch": 0.42138574869459733, + "grad_norm": 0.6193528771400452, + "learning_rate": 0.00012448765954116916, + "loss": 0.0799, + "step": 17976 + }, + { + "epoch": 0.4214091902694023, + "grad_norm": 0.5578337907791138, + "learning_rate": 0.00012448051758507305, + "loss": 0.1417, + "step": 17977 + }, + { + "epoch": 0.4214326318442073, + "grad_norm": 0.21260887384414673, + "learning_rate": 0.00012447337549614495, + "loss": 0.0651, + "step": 17978 + }, + { + "epoch": 0.4214560734190123, + "grad_norm": 0.4876859486103058, + "learning_rate": 0.00012446623327442367, + "loss": 0.1162, + "step": 17979 + }, + { + "epoch": 0.4214795149938173, + "grad_norm": 0.29386523365974426, + "learning_rate": 0.00012445909091994786, + "loss": 0.0608, + "step": 17980 + }, + { + "epoch": 0.4215029565686223, + "grad_norm": 0.6919259428977966, + "learning_rate": 0.00012445194843275638, + "loss": 0.1361, + "step": 17981 + }, + { + "epoch": 0.4215263981434273, + "grad_norm": 0.6117023825645447, + "learning_rate": 0.00012444480581288794, + "loss": 0.6219, + "step": 17982 + }, + { + "epoch": 0.4215498397182323, + "grad_norm": 0.2976846396923065, + "learning_rate": 0.00012443766306038126, + "loss": 0.071, + "step": 17983 + }, + { + "epoch": 0.4215732812930373, + "grad_norm": 1.1908133029937744, + "learning_rate": 0.00012443052017527514, + "loss": 0.1168, + "step": 17984 + }, + { + "epoch": 0.4215967228678423, + "grad_norm": 0.33199653029441833, + "learning_rate": 0.00012442337715760834, + "loss": 0.0456, + "step": 17985 + }, + { + "epoch": 0.42162016444264727, + "grad_norm": 0.46635591983795166, + "learning_rate": 0.0001244162340074196, + "loss": 0.0805, + "step": 17986 + }, + { + "epoch": 0.42164360601745227, + "grad_norm": 0.3264494240283966, + "learning_rate": 0.00012440909072474772, + "loss": 0.0534, + "step": 17987 + }, + { + "epoch": 0.42166704759225726, + "grad_norm": 0.3587973415851593, + "learning_rate": 0.0001244019473096314, + "loss": 0.0646, + "step": 17988 + }, + { + "epoch": 0.42169048916706225, + "grad_norm": 0.47460606694221497, + "learning_rate": 0.00012439480376210944, + "loss": 0.1373, + "step": 17989 + }, + { + "epoch": 0.42171393074186725, + "grad_norm": 0.5514890551567078, + "learning_rate": 0.00012438766008222057, + "loss": 0.1456, + "step": 17990 + }, + { + "epoch": 0.42173737231667224, + "grad_norm": 0.2926003932952881, + "learning_rate": 0.0001243805162700036, + "loss": 0.0666, + "step": 17991 + }, + { + "epoch": 0.42176081389147724, + "grad_norm": 1.0002626180648804, + "learning_rate": 0.00012437337232549725, + "loss": 0.1454, + "step": 17992 + }, + { + "epoch": 0.42178425546628223, + "grad_norm": 0.5943312048912048, + "learning_rate": 0.0001243662282487403, + "loss": 0.1327, + "step": 17993 + }, + { + "epoch": 0.4218076970410872, + "grad_norm": 0.3826901912689209, + "learning_rate": 0.0001243590840397715, + "loss": 0.4207, + "step": 17994 + }, + { + "epoch": 0.4218311386158922, + "grad_norm": 0.3442911207675934, + "learning_rate": 0.00012435193969862962, + "loss": 0.0875, + "step": 17995 + }, + { + "epoch": 0.4218545801906972, + "grad_norm": 0.49400973320007324, + "learning_rate": 0.00012434479522535342, + "loss": 0.0742, + "step": 17996 + }, + { + "epoch": 0.4218780217655022, + "grad_norm": 0.08254388719797134, + "learning_rate": 0.00012433765061998172, + "loss": 0.0111, + "step": 17997 + }, + { + "epoch": 0.4219014633403072, + "grad_norm": 0.35497573018074036, + "learning_rate": 0.00012433050588255324, + "loss": 0.0661, + "step": 17998 + }, + { + "epoch": 0.4219249049151122, + "grad_norm": 0.4487396478652954, + "learning_rate": 0.00012432336101310672, + "loss": 0.1159, + "step": 17999 + }, + { + "epoch": 0.4219483464899172, + "grad_norm": 0.39413201808929443, + "learning_rate": 0.00012431621601168098, + "loss": 0.0571, + "step": 18000 + }, + { + "epoch": 0.4219717880647222, + "grad_norm": 0.4969704747200012, + "learning_rate": 0.00012430907087831474, + "loss": 0.133, + "step": 18001 + }, + { + "epoch": 0.4219952296395272, + "grad_norm": 0.44872286915779114, + "learning_rate": 0.00012430192561304682, + "loss": 0.084, + "step": 18002 + }, + { + "epoch": 0.4220186712143322, + "grad_norm": 0.3814332187175751, + "learning_rate": 0.00012429478021591596, + "loss": 0.0617, + "step": 18003 + }, + { + "epoch": 0.42204211278913717, + "grad_norm": 0.5807467699050903, + "learning_rate": 0.00012428763468696094, + "loss": 0.6443, + "step": 18004 + }, + { + "epoch": 0.42206555436394216, + "grad_norm": 0.5756037831306458, + "learning_rate": 0.00012428048902622053, + "loss": 0.1691, + "step": 18005 + }, + { + "epoch": 0.42208899593874716, + "grad_norm": 0.10324284434318542, + "learning_rate": 0.00012427334323373353, + "loss": 0.0169, + "step": 18006 + }, + { + "epoch": 0.42211243751355215, + "grad_norm": 0.5851662158966064, + "learning_rate": 0.00012426619730953865, + "loss": 0.1321, + "step": 18007 + }, + { + "epoch": 0.42213587908835715, + "grad_norm": 0.4094673693180084, + "learning_rate": 0.00012425905125367472, + "loss": 0.0693, + "step": 18008 + }, + { + "epoch": 0.42215932066316214, + "grad_norm": 0.39370423555374146, + "learning_rate": 0.0001242519050661805, + "loss": 0.1029, + "step": 18009 + }, + { + "epoch": 0.42218276223796714, + "grad_norm": 0.3885771334171295, + "learning_rate": 0.00012424475874709475, + "loss": 0.0684, + "step": 18010 + }, + { + "epoch": 0.42220620381277213, + "grad_norm": 0.4743575155735016, + "learning_rate": 0.00012423761229645626, + "loss": 0.0676, + "step": 18011 + }, + { + "epoch": 0.4222296453875771, + "grad_norm": 0.6152377724647522, + "learning_rate": 0.0001242304657143038, + "loss": 0.3733, + "step": 18012 + }, + { + "epoch": 0.4222530869623821, + "grad_norm": 0.7028368711471558, + "learning_rate": 0.00012422331900067615, + "loss": 0.1437, + "step": 18013 + }, + { + "epoch": 0.4222765285371871, + "grad_norm": 0.4990314543247223, + "learning_rate": 0.00012421617215561213, + "loss": 0.1033, + "step": 18014 + }, + { + "epoch": 0.4222999701119921, + "grad_norm": 0.6866616010665894, + "learning_rate": 0.00012420902517915042, + "loss": 0.709, + "step": 18015 + }, + { + "epoch": 0.4223234116867971, + "grad_norm": 0.5932245254516602, + "learning_rate": 0.0001242018780713299, + "loss": 0.1471, + "step": 18016 + }, + { + "epoch": 0.4223468532616021, + "grad_norm": 0.3998543620109558, + "learning_rate": 0.0001241947308321893, + "loss": 0.0709, + "step": 18017 + }, + { + "epoch": 0.4223702948364071, + "grad_norm": 0.22512535750865936, + "learning_rate": 0.0001241875834617674, + "loss": 0.0464, + "step": 18018 + }, + { + "epoch": 0.4223937364112121, + "grad_norm": 0.21890705823898315, + "learning_rate": 0.00012418043596010297, + "loss": 0.0481, + "step": 18019 + }, + { + "epoch": 0.4224171779860171, + "grad_norm": 0.5398525595664978, + "learning_rate": 0.00012417328832723485, + "loss": 0.0554, + "step": 18020 + }, + { + "epoch": 0.4224406195608221, + "grad_norm": 0.5363470911979675, + "learning_rate": 0.00012416614056320177, + "loss": 0.7205, + "step": 18021 + }, + { + "epoch": 0.42246406113562707, + "grad_norm": 0.25952979922294617, + "learning_rate": 0.00012415899266804257, + "loss": 0.0454, + "step": 18022 + }, + { + "epoch": 0.42248750271043206, + "grad_norm": 0.3834105134010315, + "learning_rate": 0.00012415184464179596, + "loss": 0.0559, + "step": 18023 + }, + { + "epoch": 0.42251094428523706, + "grad_norm": 0.4185526967048645, + "learning_rate": 0.00012414469648450077, + "loss": 0.0774, + "step": 18024 + }, + { + "epoch": 0.42253438586004205, + "grad_norm": 0.35968324542045593, + "learning_rate": 0.00012413754819619576, + "loss": 0.0938, + "step": 18025 + }, + { + "epoch": 0.4225578274348471, + "grad_norm": 0.45448073744773865, + "learning_rate": 0.00012413039977691976, + "loss": 0.1008, + "step": 18026 + }, + { + "epoch": 0.4225812690096521, + "grad_norm": 0.6061311960220337, + "learning_rate": 0.00012412325122671154, + "loss": 0.5905, + "step": 18027 + }, + { + "epoch": 0.4226047105844571, + "grad_norm": 0.4806012809276581, + "learning_rate": 0.00012411610254560985, + "loss": 0.1181, + "step": 18028 + }, + { + "epoch": 0.4226281521592621, + "grad_norm": 0.6344555616378784, + "learning_rate": 0.00012410895373365355, + "loss": 0.1457, + "step": 18029 + }, + { + "epoch": 0.4226515937340671, + "grad_norm": 0.5221174955368042, + "learning_rate": 0.00012410180479088137, + "loss": 0.781, + "step": 18030 + }, + { + "epoch": 0.42267503530887207, + "grad_norm": 0.5808958411216736, + "learning_rate": 0.00012409465571733207, + "loss": 0.0889, + "step": 18031 + }, + { + "epoch": 0.42269847688367707, + "grad_norm": 0.8140263557434082, + "learning_rate": 0.00012408750651304455, + "loss": 0.0775, + "step": 18032 + }, + { + "epoch": 0.42272191845848206, + "grad_norm": 0.5430117249488831, + "learning_rate": 0.00012408035717805752, + "loss": 0.1084, + "step": 18033 + }, + { + "epoch": 0.42274536003328705, + "grad_norm": 0.42530450224876404, + "learning_rate": 0.0001240732077124098, + "loss": 0.0616, + "step": 18034 + }, + { + "epoch": 0.42276880160809205, + "grad_norm": 0.23118391633033752, + "learning_rate": 0.00012406605811614017, + "loss": 0.0339, + "step": 18035 + }, + { + "epoch": 0.42279224318289704, + "grad_norm": 0.3448765277862549, + "learning_rate": 0.00012405890838928745, + "loss": 0.0349, + "step": 18036 + }, + { + "epoch": 0.42281568475770204, + "grad_norm": 0.5214492082595825, + "learning_rate": 0.00012405175853189042, + "loss": 0.0774, + "step": 18037 + }, + { + "epoch": 0.42283912633250703, + "grad_norm": 0.7326116561889648, + "learning_rate": 0.00012404460854398785, + "loss": 0.1773, + "step": 18038 + }, + { + "epoch": 0.422862567907312, + "grad_norm": 0.4858463406562805, + "learning_rate": 0.00012403745842561858, + "loss": 0.6573, + "step": 18039 + }, + { + "epoch": 0.422886009482117, + "grad_norm": 0.34548690915107727, + "learning_rate": 0.00012403030817682135, + "loss": 0.0681, + "step": 18040 + }, + { + "epoch": 0.422909451056922, + "grad_norm": 0.07528246194124222, + "learning_rate": 0.00012402315779763503, + "loss": 0.0248, + "step": 18041 + }, + { + "epoch": 0.422932892631727, + "grad_norm": 0.6182510256767273, + "learning_rate": 0.00012401600728809836, + "loss": 0.0994, + "step": 18042 + }, + { + "epoch": 0.422956334206532, + "grad_norm": 0.6953322291374207, + "learning_rate": 0.00012400885664825016, + "loss": 0.164, + "step": 18043 + }, + { + "epoch": 0.422979775781337, + "grad_norm": 0.35706549882888794, + "learning_rate": 0.00012400170587812922, + "loss": 0.0525, + "step": 18044 + }, + { + "epoch": 0.423003217356142, + "grad_norm": 0.4661093056201935, + "learning_rate": 0.00012399455497777437, + "loss": 0.6142, + "step": 18045 + }, + { + "epoch": 0.423026658930947, + "grad_norm": 0.3169422447681427, + "learning_rate": 0.0001239874039472244, + "loss": 0.0535, + "step": 18046 + }, + { + "epoch": 0.423050100505752, + "grad_norm": 0.4007335603237152, + "learning_rate": 0.00012398025278651805, + "loss": 0.062, + "step": 18047 + }, + { + "epoch": 0.423073542080557, + "grad_norm": 0.7628768086433411, + "learning_rate": 0.00012397310149569424, + "loss": 0.1955, + "step": 18048 + }, + { + "epoch": 0.42309698365536197, + "grad_norm": 0.43845731019973755, + "learning_rate": 0.00012396595007479165, + "loss": 0.0791, + "step": 18049 + }, + { + "epoch": 0.42312042523016696, + "grad_norm": 0.35599979758262634, + "learning_rate": 0.00012395879852384917, + "loss": 0.0637, + "step": 18050 + }, + { + "epoch": 0.42314386680497196, + "grad_norm": 0.26923826336860657, + "learning_rate": 0.00012395164684290557, + "loss": 0.0212, + "step": 18051 + }, + { + "epoch": 0.42316730837977695, + "grad_norm": 0.3939668834209442, + "learning_rate": 0.00012394449503199968, + "loss": 0.0677, + "step": 18052 + }, + { + "epoch": 0.42319074995458195, + "grad_norm": 0.516233503818512, + "learning_rate": 0.00012393734309117024, + "loss": 0.1012, + "step": 18053 + }, + { + "epoch": 0.42321419152938694, + "grad_norm": 0.41692090034484863, + "learning_rate": 0.00012393019102045617, + "loss": 0.0599, + "step": 18054 + }, + { + "epoch": 0.42323763310419193, + "grad_norm": 0.8643631935119629, + "learning_rate": 0.00012392303881989615, + "loss": 0.1833, + "step": 18055 + }, + { + "epoch": 0.42326107467899693, + "grad_norm": 0.9205330610275269, + "learning_rate": 0.00012391588648952908, + "loss": 0.2118, + "step": 18056 + }, + { + "epoch": 0.4232845162538019, + "grad_norm": 0.2500476837158203, + "learning_rate": 0.0001239087340293937, + "loss": 0.039, + "step": 18057 + }, + { + "epoch": 0.4233079578286069, + "grad_norm": 0.4302493929862976, + "learning_rate": 0.00012390158143952892, + "loss": 0.5251, + "step": 18058 + }, + { + "epoch": 0.4233313994034119, + "grad_norm": 0.27465710043907166, + "learning_rate": 0.00012389442871997344, + "loss": 0.0425, + "step": 18059 + }, + { + "epoch": 0.4233548409782169, + "grad_norm": 0.580486536026001, + "learning_rate": 0.00012388727587076613, + "loss": 0.0516, + "step": 18060 + }, + { + "epoch": 0.4233782825530219, + "grad_norm": 0.2550821304321289, + "learning_rate": 0.00012388012289194582, + "loss": 0.066, + "step": 18061 + }, + { + "epoch": 0.4234017241278269, + "grad_norm": 0.46103689074516296, + "learning_rate": 0.00012387296978355123, + "loss": 0.156, + "step": 18062 + }, + { + "epoch": 0.4234251657026319, + "grad_norm": 1.1666804552078247, + "learning_rate": 0.00012386581654562126, + "loss": 0.1021, + "step": 18063 + }, + { + "epoch": 0.4234486072774369, + "grad_norm": 0.32371652126312256, + "learning_rate": 0.00012385866317819472, + "loss": 0.0678, + "step": 18064 + }, + { + "epoch": 0.4234720488522419, + "grad_norm": 0.7019479274749756, + "learning_rate": 0.0001238515096813104, + "loss": 0.0707, + "step": 18065 + }, + { + "epoch": 0.4234954904270469, + "grad_norm": 0.43539997935295105, + "learning_rate": 0.00012384435605500708, + "loss": 0.1216, + "step": 18066 + }, + { + "epoch": 0.42351893200185187, + "grad_norm": 0.5551956295967102, + "learning_rate": 0.00012383720229932368, + "loss": 0.1677, + "step": 18067 + }, + { + "epoch": 0.42354237357665686, + "grad_norm": 0.3195284307003021, + "learning_rate": 0.0001238300484142989, + "loss": 0.0609, + "step": 18068 + }, + { + "epoch": 0.42356581515146186, + "grad_norm": 0.29384586215019226, + "learning_rate": 0.0001238228943999716, + "loss": 0.0858, + "step": 18069 + }, + { + "epoch": 0.42358925672626685, + "grad_norm": 0.7251036763191223, + "learning_rate": 0.0001238157402563806, + "loss": 0.1321, + "step": 18070 + }, + { + "epoch": 0.42361269830107184, + "grad_norm": 0.27748343348503113, + "learning_rate": 0.00012380858598356478, + "loss": 0.0415, + "step": 18071 + }, + { + "epoch": 0.42363613987587684, + "grad_norm": 0.4709922969341278, + "learning_rate": 0.00012380143158156285, + "loss": 0.1402, + "step": 18072 + }, + { + "epoch": 0.42365958145068183, + "grad_norm": 0.38541409373283386, + "learning_rate": 0.0001237942770504137, + "loss": 0.0548, + "step": 18073 + }, + { + "epoch": 0.4236830230254868, + "grad_norm": 0.2752937376499176, + "learning_rate": 0.00012378712239015613, + "loss": 0.0347, + "step": 18074 + }, + { + "epoch": 0.4237064646002918, + "grad_norm": 0.5215861201286316, + "learning_rate": 0.000123779967600829, + "loss": 0.0887, + "step": 18075 + }, + { + "epoch": 0.4237299061750968, + "grad_norm": 0.26002269983291626, + "learning_rate": 0.00012377281268247106, + "loss": 0.0473, + "step": 18076 + }, + { + "epoch": 0.42375334774990187, + "grad_norm": 0.4983392357826233, + "learning_rate": 0.00012376565763512116, + "loss": 0.1209, + "step": 18077 + }, + { + "epoch": 0.42377678932470686, + "grad_norm": 0.2559681832790375, + "learning_rate": 0.00012375850245881814, + "loss": 0.0244, + "step": 18078 + }, + { + "epoch": 0.42380023089951185, + "grad_norm": 0.7600650191307068, + "learning_rate": 0.00012375134715360083, + "loss": 0.1878, + "step": 18079 + }, + { + "epoch": 0.42382367247431685, + "grad_norm": 0.44285067915916443, + "learning_rate": 0.00012374419171950804, + "loss": 0.1315, + "step": 18080 + }, + { + "epoch": 0.42384711404912184, + "grad_norm": 0.3812047839164734, + "learning_rate": 0.0001237370361565786, + "loss": 0.1093, + "step": 18081 + }, + { + "epoch": 0.42387055562392684, + "grad_norm": 0.5286680459976196, + "learning_rate": 0.00012372988046485136, + "loss": 0.0763, + "step": 18082 + }, + { + "epoch": 0.42389399719873183, + "grad_norm": 0.4184293746948242, + "learning_rate": 0.00012372272464436507, + "loss": 0.0991, + "step": 18083 + }, + { + "epoch": 0.4239174387735368, + "grad_norm": 0.5169493556022644, + "learning_rate": 0.00012371556869515866, + "loss": 0.1207, + "step": 18084 + }, + { + "epoch": 0.4239408803483418, + "grad_norm": 0.6354016661643982, + "learning_rate": 0.00012370841261727085, + "loss": 0.1522, + "step": 18085 + }, + { + "epoch": 0.4239643219231468, + "grad_norm": 0.2565118074417114, + "learning_rate": 0.00012370125641074058, + "loss": 0.0234, + "step": 18086 + }, + { + "epoch": 0.4239877634979518, + "grad_norm": 0.3945787250995636, + "learning_rate": 0.00012369410007560661, + "loss": 0.0741, + "step": 18087 + }, + { + "epoch": 0.4240112050727568, + "grad_norm": 0.15952008962631226, + "learning_rate": 0.00012368694361190776, + "loss": 0.0451, + "step": 18088 + }, + { + "epoch": 0.4240346466475618, + "grad_norm": 0.3925718665122986, + "learning_rate": 0.00012367978701968292, + "loss": 0.0893, + "step": 18089 + }, + { + "epoch": 0.4240580882223668, + "grad_norm": 0.5369745492935181, + "learning_rate": 0.00012367263029897088, + "loss": 0.1145, + "step": 18090 + }, + { + "epoch": 0.4240815297971718, + "grad_norm": 1.0604605674743652, + "learning_rate": 0.0001236654734498105, + "loss": 0.287, + "step": 18091 + }, + { + "epoch": 0.4241049713719768, + "grad_norm": 0.40092530846595764, + "learning_rate": 0.00012365831647224054, + "loss": 0.0901, + "step": 18092 + }, + { + "epoch": 0.4241284129467818, + "grad_norm": 1.1326911449432373, + "learning_rate": 0.00012365115936629993, + "loss": 0.175, + "step": 18093 + }, + { + "epoch": 0.42415185452158677, + "grad_norm": 0.6747488975524902, + "learning_rate": 0.00012364400213202748, + "loss": 0.2308, + "step": 18094 + }, + { + "epoch": 0.42417529609639176, + "grad_norm": 0.29555416107177734, + "learning_rate": 0.00012363684476946198, + "loss": 0.0594, + "step": 18095 + }, + { + "epoch": 0.42419873767119676, + "grad_norm": 0.5136518478393555, + "learning_rate": 0.0001236296872786423, + "loss": 0.0885, + "step": 18096 + }, + { + "epoch": 0.42422217924600175, + "grad_norm": 0.6149302124977112, + "learning_rate": 0.0001236225296596073, + "loss": 0.1338, + "step": 18097 + }, + { + "epoch": 0.42424562082080675, + "grad_norm": 0.2780423164367676, + "learning_rate": 0.00012361537191239575, + "loss": 0.0715, + "step": 18098 + }, + { + "epoch": 0.42426906239561174, + "grad_norm": 0.37405499815940857, + "learning_rate": 0.00012360821403704654, + "loss": 0.0561, + "step": 18099 + }, + { + "epoch": 0.42429250397041673, + "grad_norm": 0.4374551773071289, + "learning_rate": 0.0001236010560335985, + "loss": 0.5204, + "step": 18100 + }, + { + "epoch": 0.42431594554522173, + "grad_norm": 0.268763929605484, + "learning_rate": 0.00012359389790209048, + "loss": 0.0367, + "step": 18101 + }, + { + "epoch": 0.4243393871200267, + "grad_norm": 0.1584588885307312, + "learning_rate": 0.0001235867396425613, + "loss": 0.0451, + "step": 18102 + }, + { + "epoch": 0.4243628286948317, + "grad_norm": 0.6492366194725037, + "learning_rate": 0.0001235795812550498, + "loss": 0.09, + "step": 18103 + }, + { + "epoch": 0.4243862702696367, + "grad_norm": 0.1790637969970703, + "learning_rate": 0.0001235724227395948, + "loss": 0.0346, + "step": 18104 + }, + { + "epoch": 0.4244097118444417, + "grad_norm": 0.340910941362381, + "learning_rate": 0.00012356526409623524, + "loss": 0.0639, + "step": 18105 + }, + { + "epoch": 0.4244331534192467, + "grad_norm": 0.49181023240089417, + "learning_rate": 0.00012355810532500987, + "loss": 0.1103, + "step": 18106 + }, + { + "epoch": 0.4244565949940517, + "grad_norm": 0.3698181211948395, + "learning_rate": 0.00012355094642595754, + "loss": 0.1083, + "step": 18107 + }, + { + "epoch": 0.4244800365688567, + "grad_norm": 0.4592745304107666, + "learning_rate": 0.00012354378739911712, + "loss": 0.0994, + "step": 18108 + }, + { + "epoch": 0.4245034781436617, + "grad_norm": 0.639081597328186, + "learning_rate": 0.00012353662824452746, + "loss": 0.1446, + "step": 18109 + }, + { + "epoch": 0.4245269197184667, + "grad_norm": 0.3632713258266449, + "learning_rate": 0.00012352946896222736, + "loss": 0.0738, + "step": 18110 + }, + { + "epoch": 0.42455036129327167, + "grad_norm": 0.38334524631500244, + "learning_rate": 0.0001235223095522557, + "loss": 0.0588, + "step": 18111 + }, + { + "epoch": 0.42457380286807667, + "grad_norm": 0.621841311454773, + "learning_rate": 0.00012351515001465138, + "loss": 0.114, + "step": 18112 + }, + { + "epoch": 0.42459724444288166, + "grad_norm": 0.48082685470581055, + "learning_rate": 0.00012350799034945316, + "loss": 0.064, + "step": 18113 + }, + { + "epoch": 0.42462068601768665, + "grad_norm": 0.5574867725372314, + "learning_rate": 0.0001235008305566999, + "loss": 0.0957, + "step": 18114 + }, + { + "epoch": 0.42464412759249165, + "grad_norm": 0.6211841702461243, + "learning_rate": 0.00012349367063643052, + "loss": 0.1409, + "step": 18115 + }, + { + "epoch": 0.42466756916729664, + "grad_norm": 0.1894981414079666, + "learning_rate": 0.0001234865105886838, + "loss": 0.0284, + "step": 18116 + }, + { + "epoch": 0.42469101074210164, + "grad_norm": 0.4803718328475952, + "learning_rate": 0.00012347935041349858, + "loss": 0.1166, + "step": 18117 + }, + { + "epoch": 0.42471445231690663, + "grad_norm": 0.4190163314342499, + "learning_rate": 0.0001234721901109138, + "loss": 0.5494, + "step": 18118 + }, + { + "epoch": 0.4247378938917116, + "grad_norm": 0.27137497067451477, + "learning_rate": 0.00012346502968096825, + "loss": 0.0284, + "step": 18119 + }, + { + "epoch": 0.4247613354665166, + "grad_norm": 0.22009404003620148, + "learning_rate": 0.00012345786912370074, + "loss": 0.0498, + "step": 18120 + }, + { + "epoch": 0.4247847770413216, + "grad_norm": 0.15654900670051575, + "learning_rate": 0.00012345070843915024, + "loss": 0.0262, + "step": 18121 + }, + { + "epoch": 0.4248082186161266, + "grad_norm": 0.668846607208252, + "learning_rate": 0.00012344354762735548, + "loss": 0.0602, + "step": 18122 + }, + { + "epoch": 0.4248316601909316, + "grad_norm": 0.8227798938751221, + "learning_rate": 0.00012343638668835537, + "loss": 0.1115, + "step": 18123 + }, + { + "epoch": 0.4248551017657366, + "grad_norm": 0.5356855988502502, + "learning_rate": 0.0001234292256221888, + "loss": 0.1301, + "step": 18124 + }, + { + "epoch": 0.4248785433405416, + "grad_norm": 0.3611714541912079, + "learning_rate": 0.0001234220644288946, + "loss": 0.1043, + "step": 18125 + }, + { + "epoch": 0.4249019849153466, + "grad_norm": 0.6467259526252747, + "learning_rate": 0.0001234149031085116, + "loss": 0.1522, + "step": 18126 + }, + { + "epoch": 0.4249254264901516, + "grad_norm": 0.6218408942222595, + "learning_rate": 0.00012340774166107868, + "loss": 0.1782, + "step": 18127 + }, + { + "epoch": 0.42494886806495663, + "grad_norm": 0.5428221225738525, + "learning_rate": 0.0001234005800866347, + "loss": 0.0576, + "step": 18128 + }, + { + "epoch": 0.4249723096397616, + "grad_norm": 0.7201728224754333, + "learning_rate": 0.00012339341838521852, + "loss": 0.1171, + "step": 18129 + }, + { + "epoch": 0.4249957512145666, + "grad_norm": 0.7970850467681885, + "learning_rate": 0.00012338625655686897, + "loss": 0.127, + "step": 18130 + }, + { + "epoch": 0.4250191927893716, + "grad_norm": 0.48456940054893494, + "learning_rate": 0.00012337909460162497, + "loss": 0.1511, + "step": 18131 + }, + { + "epoch": 0.4250426343641766, + "grad_norm": 0.9443095326423645, + "learning_rate": 0.00012337193251952533, + "loss": 0.4054, + "step": 18132 + }, + { + "epoch": 0.4250660759389816, + "grad_norm": 0.6771392226219177, + "learning_rate": 0.00012336477031060889, + "loss": 0.0719, + "step": 18133 + }, + { + "epoch": 0.4250895175137866, + "grad_norm": 0.5666278004646301, + "learning_rate": 0.0001233576079749146, + "loss": 0.0728, + "step": 18134 + }, + { + "epoch": 0.4251129590885916, + "grad_norm": 0.46733933687210083, + "learning_rate": 0.00012335044551248125, + "loss": 0.0713, + "step": 18135 + }, + { + "epoch": 0.4251364006633966, + "grad_norm": 0.6203305125236511, + "learning_rate": 0.00012334328292334771, + "loss": 0.5953, + "step": 18136 + }, + { + "epoch": 0.4251598422382016, + "grad_norm": 0.4151611030101776, + "learning_rate": 0.0001233361202075529, + "loss": 0.6239, + "step": 18137 + }, + { + "epoch": 0.4251832838130066, + "grad_norm": 0.5097776651382446, + "learning_rate": 0.00012332895736513562, + "loss": 0.0541, + "step": 18138 + }, + { + "epoch": 0.42520672538781157, + "grad_norm": 1.8789583444595337, + "learning_rate": 0.00012332179439613475, + "loss": 0.1674, + "step": 18139 + }, + { + "epoch": 0.42523016696261656, + "grad_norm": 0.44616493582725525, + "learning_rate": 0.0001233146313005892, + "loss": 0.0447, + "step": 18140 + }, + { + "epoch": 0.42525360853742156, + "grad_norm": 0.5024029612541199, + "learning_rate": 0.00012330746807853778, + "loss": 0.0998, + "step": 18141 + }, + { + "epoch": 0.42527705011222655, + "grad_norm": 1.0133064985275269, + "learning_rate": 0.0001233003047300194, + "loss": 0.122, + "step": 18142 + }, + { + "epoch": 0.42530049168703155, + "grad_norm": 0.3978859782218933, + "learning_rate": 0.0001232931412550729, + "loss": 0.0804, + "step": 18143 + }, + { + "epoch": 0.42532393326183654, + "grad_norm": 0.44627150893211365, + "learning_rate": 0.00012328597765373718, + "loss": 0.0918, + "step": 18144 + }, + { + "epoch": 0.42534737483664153, + "grad_norm": 0.18445232510566711, + "learning_rate": 0.00012327881392605106, + "loss": 0.0143, + "step": 18145 + }, + { + "epoch": 0.42537081641144653, + "grad_norm": 0.16002947092056274, + "learning_rate": 0.00012327165007205347, + "loss": 0.0242, + "step": 18146 + }, + { + "epoch": 0.4253942579862515, + "grad_norm": 0.7382593154907227, + "learning_rate": 0.00012326448609178322, + "loss": 0.1861, + "step": 18147 + }, + { + "epoch": 0.4254176995610565, + "grad_norm": 0.31842881441116333, + "learning_rate": 0.00012325732198527925, + "loss": 0.0338, + "step": 18148 + }, + { + "epoch": 0.4254411411358615, + "grad_norm": 0.6179298758506775, + "learning_rate": 0.00012325015775258038, + "loss": 0.1623, + "step": 18149 + }, + { + "epoch": 0.4254645827106665, + "grad_norm": 0.6516647934913635, + "learning_rate": 0.00012324299339372548, + "loss": 0.0814, + "step": 18150 + }, + { + "epoch": 0.4254880242854715, + "grad_norm": 1.1188111305236816, + "learning_rate": 0.00012323582890875348, + "loss": 0.2014, + "step": 18151 + }, + { + "epoch": 0.4255114658602765, + "grad_norm": 0.22645050287246704, + "learning_rate": 0.0001232286642977032, + "loss": 0.0376, + "step": 18152 + }, + { + "epoch": 0.4255349074350815, + "grad_norm": 0.5449807047843933, + "learning_rate": 0.00012322149956061352, + "loss": 0.1049, + "step": 18153 + }, + { + "epoch": 0.4255583490098865, + "grad_norm": 0.6036306619644165, + "learning_rate": 0.00012321433469752338, + "loss": 0.1338, + "step": 18154 + }, + { + "epoch": 0.4255817905846915, + "grad_norm": 0.40898382663726807, + "learning_rate": 0.00012320716970847157, + "loss": 0.0715, + "step": 18155 + }, + { + "epoch": 0.42560523215949647, + "grad_norm": 0.22732506692409515, + "learning_rate": 0.000123200004593497, + "loss": 0.0401, + "step": 18156 + }, + { + "epoch": 0.42562867373430147, + "grad_norm": 0.4000405967235565, + "learning_rate": 0.00012319283935263858, + "loss": 0.3295, + "step": 18157 + }, + { + "epoch": 0.42565211530910646, + "grad_norm": 0.3946205675601959, + "learning_rate": 0.00012318567398593512, + "loss": 0.0895, + "step": 18158 + }, + { + "epoch": 0.42567555688391145, + "grad_norm": 0.21853843331336975, + "learning_rate": 0.00012317850849342555, + "loss": 0.0461, + "step": 18159 + }, + { + "epoch": 0.42569899845871645, + "grad_norm": 0.28588131070137024, + "learning_rate": 0.00012317134287514878, + "loss": 0.061, + "step": 18160 + }, + { + "epoch": 0.42572244003352144, + "grad_norm": 0.543054461479187, + "learning_rate": 0.00012316417713114364, + "loss": 0.1154, + "step": 18161 + }, + { + "epoch": 0.42574588160832644, + "grad_norm": 0.5721607208251953, + "learning_rate": 0.000123157011261449, + "loss": 0.1304, + "step": 18162 + }, + { + "epoch": 0.42576932318313143, + "grad_norm": 0.7505148649215698, + "learning_rate": 0.0001231498452661038, + "loss": 0.151, + "step": 18163 + }, + { + "epoch": 0.4257927647579364, + "grad_norm": 0.4239693582057953, + "learning_rate": 0.00012314267914514685, + "loss": 0.0604, + "step": 18164 + }, + { + "epoch": 0.4258162063327414, + "grad_norm": 0.7742342948913574, + "learning_rate": 0.0001231355128986171, + "loss": 0.1486, + "step": 18165 + }, + { + "epoch": 0.4258396479075464, + "grad_norm": 0.3932700753211975, + "learning_rate": 0.00012312834652655337, + "loss": 0.0547, + "step": 18166 + }, + { + "epoch": 0.4258630894823514, + "grad_norm": 0.7166780233383179, + "learning_rate": 0.00012312118002899463, + "loss": 0.1442, + "step": 18167 + }, + { + "epoch": 0.4258865310571564, + "grad_norm": 0.2870137393474579, + "learning_rate": 0.00012311401340597968, + "loss": 0.0329, + "step": 18168 + }, + { + "epoch": 0.4259099726319614, + "grad_norm": 0.7749749422073364, + "learning_rate": 0.00012310684665754748, + "loss": 0.0849, + "step": 18169 + }, + { + "epoch": 0.4259334142067664, + "grad_norm": 0.28879448771476746, + "learning_rate": 0.00012309967978373686, + "loss": 0.0444, + "step": 18170 + }, + { + "epoch": 0.4259568557815714, + "grad_norm": 0.5183791518211365, + "learning_rate": 0.00012309251278458668, + "loss": 0.1296, + "step": 18171 + }, + { + "epoch": 0.4259802973563764, + "grad_norm": 0.357839435338974, + "learning_rate": 0.00012308534566013595, + "loss": 0.0753, + "step": 18172 + }, + { + "epoch": 0.4260037389311814, + "grad_norm": 0.3966471552848816, + "learning_rate": 0.00012307817841042347, + "loss": 0.0908, + "step": 18173 + }, + { + "epoch": 0.42602718050598637, + "grad_norm": 0.4758198857307434, + "learning_rate": 0.0001230710110354881, + "loss": 0.1078, + "step": 18174 + }, + { + "epoch": 0.42605062208079136, + "grad_norm": 0.35749825835227966, + "learning_rate": 0.00012306384353536883, + "loss": 0.1834, + "step": 18175 + }, + { + "epoch": 0.42607406365559636, + "grad_norm": 0.5545543432235718, + "learning_rate": 0.0001230566759101045, + "loss": 0.1318, + "step": 18176 + }, + { + "epoch": 0.42609750523040135, + "grad_norm": 0.6011847257614136, + "learning_rate": 0.00012304950815973393, + "loss": 0.1187, + "step": 18177 + }, + { + "epoch": 0.42612094680520635, + "grad_norm": 0.40875208377838135, + "learning_rate": 0.00012304234028429612, + "loss": 0.0734, + "step": 18178 + }, + { + "epoch": 0.4261443883800114, + "grad_norm": 0.37325721979141235, + "learning_rate": 0.00012303517228382996, + "loss": 0.0512, + "step": 18179 + }, + { + "epoch": 0.4261678299548164, + "grad_norm": 0.5803213119506836, + "learning_rate": 0.00012302800415837426, + "loss": 0.1256, + "step": 18180 + }, + { + "epoch": 0.4261912715296214, + "grad_norm": 0.2873462438583374, + "learning_rate": 0.00012302083590796798, + "loss": 0.0458, + "step": 18181 + }, + { + "epoch": 0.4262147131044264, + "grad_norm": 0.3939838707447052, + "learning_rate": 0.00012301366753265, + "loss": 0.0919, + "step": 18182 + }, + { + "epoch": 0.4262381546792314, + "grad_norm": 0.31936249136924744, + "learning_rate": 0.0001230064990324592, + "loss": 0.0463, + "step": 18183 + }, + { + "epoch": 0.42626159625403637, + "grad_norm": 0.5757301449775696, + "learning_rate": 0.00012299933040743448, + "loss": 0.0817, + "step": 18184 + }, + { + "epoch": 0.42628503782884136, + "grad_norm": 0.23048007488250732, + "learning_rate": 0.00012299216165761478, + "loss": 0.0482, + "step": 18185 + }, + { + "epoch": 0.42630847940364636, + "grad_norm": 0.6280977129936218, + "learning_rate": 0.00012298499278303894, + "loss": 0.5361, + "step": 18186 + }, + { + "epoch": 0.42633192097845135, + "grad_norm": 0.43029096722602844, + "learning_rate": 0.0001229778237837459, + "loss": 0.0649, + "step": 18187 + }, + { + "epoch": 0.42635536255325635, + "grad_norm": 0.5509220957756042, + "learning_rate": 0.00012297065465977452, + "loss": 0.7804, + "step": 18188 + }, + { + "epoch": 0.42637880412806134, + "grad_norm": 0.09114977717399597, + "learning_rate": 0.00012296348541116373, + "loss": 0.0158, + "step": 18189 + }, + { + "epoch": 0.42640224570286633, + "grad_norm": 0.4210982918739319, + "learning_rate": 0.00012295631603795242, + "loss": 0.1139, + "step": 18190 + }, + { + "epoch": 0.42642568727767133, + "grad_norm": 0.6419405937194824, + "learning_rate": 0.0001229491465401795, + "loss": 0.1373, + "step": 18191 + }, + { + "epoch": 0.4264491288524763, + "grad_norm": 0.3183995485305786, + "learning_rate": 0.00012294197691788388, + "loss": 0.0643, + "step": 18192 + }, + { + "epoch": 0.4264725704272813, + "grad_norm": 0.36224615573883057, + "learning_rate": 0.00012293480717110442, + "loss": 0.3132, + "step": 18193 + }, + { + "epoch": 0.4264960120020863, + "grad_norm": 0.5311828851699829, + "learning_rate": 0.00012292763729988007, + "loss": 0.1627, + "step": 18194 + }, + { + "epoch": 0.4265194535768913, + "grad_norm": 0.5579147338867188, + "learning_rate": 0.0001229204673042497, + "loss": 0.6594, + "step": 18195 + }, + { + "epoch": 0.4265428951516963, + "grad_norm": 0.1588459461927414, + "learning_rate": 0.00012291329718425224, + "loss": 0.033, + "step": 18196 + }, + { + "epoch": 0.4265663367265013, + "grad_norm": 0.48566094040870667, + "learning_rate": 0.0001229061269399266, + "loss": 0.0961, + "step": 18197 + }, + { + "epoch": 0.4265897783013063, + "grad_norm": 0.420253187417984, + "learning_rate": 0.00012289895657131166, + "loss": 0.0716, + "step": 18198 + }, + { + "epoch": 0.4266132198761113, + "grad_norm": 0.1391037404537201, + "learning_rate": 0.00012289178607844634, + "loss": 0.0216, + "step": 18199 + }, + { + "epoch": 0.4266366614509163, + "grad_norm": 0.5256561040878296, + "learning_rate": 0.00012288461546136953, + "loss": 0.1253, + "step": 18200 + }, + { + "epoch": 0.42666010302572127, + "grad_norm": 0.14222094416618347, + "learning_rate": 0.00012287744472012015, + "loss": 0.0233, + "step": 18201 + }, + { + "epoch": 0.42668354460052627, + "grad_norm": 0.2005978524684906, + "learning_rate": 0.00012287027385473714, + "loss": 0.0446, + "step": 18202 + }, + { + "epoch": 0.42670698617533126, + "grad_norm": 0.35919997096061707, + "learning_rate": 0.00012286310286525935, + "loss": 0.0923, + "step": 18203 + }, + { + "epoch": 0.42673042775013625, + "grad_norm": 0.12897011637687683, + "learning_rate": 0.00012285593175172577, + "loss": 0.0384, + "step": 18204 + }, + { + "epoch": 0.42675386932494125, + "grad_norm": 0.2771797776222229, + "learning_rate": 0.0001228487605141752, + "loss": 0.0561, + "step": 18205 + }, + { + "epoch": 0.42677731089974624, + "grad_norm": 0.5464357137680054, + "learning_rate": 0.00012284158915264664, + "loss": 0.2231, + "step": 18206 + }, + { + "epoch": 0.42680075247455124, + "grad_norm": 0.4446119964122772, + "learning_rate": 0.00012283441766717897, + "loss": 0.0908, + "step": 18207 + }, + { + "epoch": 0.42682419404935623, + "grad_norm": 0.5432897806167603, + "learning_rate": 0.00012282724605781114, + "loss": 0.1429, + "step": 18208 + }, + { + "epoch": 0.4268476356241612, + "grad_norm": 0.30687201023101807, + "learning_rate": 0.000122820074324582, + "loss": 0.0401, + "step": 18209 + }, + { + "epoch": 0.4268710771989662, + "grad_norm": 0.2105770707130432, + "learning_rate": 0.0001228129024675305, + "loss": 0.0397, + "step": 18210 + }, + { + "epoch": 0.4268945187737712, + "grad_norm": 0.46051129698753357, + "learning_rate": 0.00012280573048669555, + "loss": 0.0753, + "step": 18211 + }, + { + "epoch": 0.4269179603485762, + "grad_norm": 0.43281102180480957, + "learning_rate": 0.00012279855838211604, + "loss": 0.062, + "step": 18212 + }, + { + "epoch": 0.4269414019233812, + "grad_norm": 0.23203478753566742, + "learning_rate": 0.00012279138615383097, + "loss": 0.0279, + "step": 18213 + }, + { + "epoch": 0.4269648434981862, + "grad_norm": 0.7918080687522888, + "learning_rate": 0.00012278421380187914, + "loss": 0.156, + "step": 18214 + }, + { + "epoch": 0.4269882850729912, + "grad_norm": 0.34224677085876465, + "learning_rate": 0.00012277704132629955, + "loss": 0.0607, + "step": 18215 + }, + { + "epoch": 0.4270117266477962, + "grad_norm": 0.853503942489624, + "learning_rate": 0.0001227698687271311, + "loss": 0.1624, + "step": 18216 + }, + { + "epoch": 0.4270351682226012, + "grad_norm": 0.14410711824893951, + "learning_rate": 0.00012276269600441271, + "loss": 0.0254, + "step": 18217 + }, + { + "epoch": 0.4270586097974062, + "grad_norm": 0.549951434135437, + "learning_rate": 0.00012275552315818325, + "loss": 0.0828, + "step": 18218 + }, + { + "epoch": 0.42708205137221117, + "grad_norm": 0.4042661488056183, + "learning_rate": 0.00012274835018848173, + "loss": 0.0799, + "step": 18219 + }, + { + "epoch": 0.42710549294701616, + "grad_norm": 0.4342494606971741, + "learning_rate": 0.00012274117709534698, + "loss": 0.0892, + "step": 18220 + }, + { + "epoch": 0.42712893452182116, + "grad_norm": 0.6262431740760803, + "learning_rate": 0.000122734003878818, + "loss": 0.1095, + "step": 18221 + }, + { + "epoch": 0.42715237609662615, + "grad_norm": 0.055164169520139694, + "learning_rate": 0.00012272683053893364, + "loss": 0.008, + "step": 18222 + }, + { + "epoch": 0.42717581767143115, + "grad_norm": 0.4540445804595947, + "learning_rate": 0.00012271965707573288, + "loss": 0.0509, + "step": 18223 + }, + { + "epoch": 0.42719925924623614, + "grad_norm": 0.4447457790374756, + "learning_rate": 0.00012271248348925463, + "loss": 0.0524, + "step": 18224 + }, + { + "epoch": 0.42722270082104113, + "grad_norm": 0.48901134729385376, + "learning_rate": 0.00012270530977953777, + "loss": 0.0608, + "step": 18225 + }, + { + "epoch": 0.42724614239584613, + "grad_norm": 0.5143091678619385, + "learning_rate": 0.0001226981359466213, + "loss": 0.1288, + "step": 18226 + }, + { + "epoch": 0.4272695839706511, + "grad_norm": 0.6391741633415222, + "learning_rate": 0.0001226909619905441, + "loss": 0.6127, + "step": 18227 + }, + { + "epoch": 0.4272930255454561, + "grad_norm": 0.28029969334602356, + "learning_rate": 0.00012268378791134506, + "loss": 0.0381, + "step": 18228 + }, + { + "epoch": 0.4273164671202611, + "grad_norm": 0.5533029437065125, + "learning_rate": 0.0001226766137090632, + "loss": 0.4988, + "step": 18229 + }, + { + "epoch": 0.42733990869506616, + "grad_norm": 0.39183568954467773, + "learning_rate": 0.0001226694393837374, + "loss": 0.0696, + "step": 18230 + }, + { + "epoch": 0.42736335026987116, + "grad_norm": 0.512783408164978, + "learning_rate": 0.00012266226493540657, + "loss": 0.099, + "step": 18231 + }, + { + "epoch": 0.42738679184467615, + "grad_norm": 0.06786199659109116, + "learning_rate": 0.00012265509036410964, + "loss": 0.0046, + "step": 18232 + }, + { + "epoch": 0.42741023341948114, + "grad_norm": 0.1797865778207779, + "learning_rate": 0.00012264791566988557, + "loss": 0.0329, + "step": 18233 + }, + { + "epoch": 0.42743367499428614, + "grad_norm": 0.38343262672424316, + "learning_rate": 0.00012264074085277323, + "loss": 0.0568, + "step": 18234 + }, + { + "epoch": 0.42745711656909113, + "grad_norm": 0.39472949504852295, + "learning_rate": 0.00012263356591281162, + "loss": 0.0637, + "step": 18235 + }, + { + "epoch": 0.4274805581438961, + "grad_norm": 0.47631335258483887, + "learning_rate": 0.00012262639085003967, + "loss": 0.0775, + "step": 18236 + }, + { + "epoch": 0.4275039997187011, + "grad_norm": 0.7963035702705383, + "learning_rate": 0.00012261921566449624, + "loss": 0.1645, + "step": 18237 + }, + { + "epoch": 0.4275274412935061, + "grad_norm": 0.32703495025634766, + "learning_rate": 0.00012261204035622037, + "loss": 0.0618, + "step": 18238 + }, + { + "epoch": 0.4275508828683111, + "grad_norm": 0.1888350248336792, + "learning_rate": 0.0001226048649252509, + "loss": 0.0411, + "step": 18239 + }, + { + "epoch": 0.4275743244431161, + "grad_norm": 0.5931689739227295, + "learning_rate": 0.00012259768937162682, + "loss": 0.6446, + "step": 18240 + }, + { + "epoch": 0.4275977660179211, + "grad_norm": 0.6914189457893372, + "learning_rate": 0.00012259051369538699, + "loss": 0.1713, + "step": 18241 + }, + { + "epoch": 0.4276212075927261, + "grad_norm": 0.6582628488540649, + "learning_rate": 0.00012258333789657043, + "loss": 0.5361, + "step": 18242 + }, + { + "epoch": 0.4276446491675311, + "grad_norm": 0.47928619384765625, + "learning_rate": 0.00012257616197521606, + "loss": 0.0711, + "step": 18243 + }, + { + "epoch": 0.4276680907423361, + "grad_norm": 0.14004915952682495, + "learning_rate": 0.00012256898593136274, + "loss": 0.0188, + "step": 18244 + }, + { + "epoch": 0.4276915323171411, + "grad_norm": 0.36035212874412537, + "learning_rate": 0.00012256180976504954, + "loss": 0.0847, + "step": 18245 + }, + { + "epoch": 0.42771497389194607, + "grad_norm": 0.49660027027130127, + "learning_rate": 0.0001225546334763153, + "loss": 0.0767, + "step": 18246 + }, + { + "epoch": 0.42773841546675107, + "grad_norm": 0.3529573678970337, + "learning_rate": 0.00012254745706519898, + "loss": 0.0854, + "step": 18247 + }, + { + "epoch": 0.42776185704155606, + "grad_norm": 0.6059090495109558, + "learning_rate": 0.00012254028053173952, + "loss": 0.1289, + "step": 18248 + }, + { + "epoch": 0.42778529861636105, + "grad_norm": 0.6931180953979492, + "learning_rate": 0.00012253310387597588, + "loss": 0.7341, + "step": 18249 + }, + { + "epoch": 0.42780874019116605, + "grad_norm": 0.5747976899147034, + "learning_rate": 0.00012252592709794698, + "loss": 0.1795, + "step": 18250 + }, + { + "epoch": 0.42783218176597104, + "grad_norm": 0.6044572591781616, + "learning_rate": 0.00012251875019769174, + "loss": 0.7912, + "step": 18251 + }, + { + "epoch": 0.42785562334077604, + "grad_norm": 0.20756807923316956, + "learning_rate": 0.00012251157317524916, + "loss": 0.0236, + "step": 18252 + }, + { + "epoch": 0.42787906491558103, + "grad_norm": 0.7496609687805176, + "learning_rate": 0.00012250439603065814, + "loss": 0.1253, + "step": 18253 + }, + { + "epoch": 0.427902506490386, + "grad_norm": 0.8504759073257446, + "learning_rate": 0.0001224972187639576, + "loss": 0.0771, + "step": 18254 + }, + { + "epoch": 0.427925948065191, + "grad_norm": 0.29412585496902466, + "learning_rate": 0.0001224900413751866, + "loss": 0.0735, + "step": 18255 + }, + { + "epoch": 0.427949389639996, + "grad_norm": 0.4560866355895996, + "learning_rate": 0.00012248286386438393, + "loss": 0.0652, + "step": 18256 + }, + { + "epoch": 0.427972831214801, + "grad_norm": 0.4704948365688324, + "learning_rate": 0.00012247568623158863, + "loss": 0.1198, + "step": 18257 + }, + { + "epoch": 0.427996272789606, + "grad_norm": 0.492005318403244, + "learning_rate": 0.00012246850847683963, + "loss": 0.1046, + "step": 18258 + }, + { + "epoch": 0.428019714364411, + "grad_norm": 0.6043349504470825, + "learning_rate": 0.00012246133060017588, + "loss": 0.1068, + "step": 18259 + }, + { + "epoch": 0.428043155939216, + "grad_norm": 0.5525323748588562, + "learning_rate": 0.00012245415260163627, + "loss": 0.1268, + "step": 18260 + }, + { + "epoch": 0.428066597514021, + "grad_norm": 0.5580487847328186, + "learning_rate": 0.00012244697448125986, + "loss": 0.1354, + "step": 18261 + }, + { + "epoch": 0.428090039088826, + "grad_norm": 0.24272453784942627, + "learning_rate": 0.0001224397962390855, + "loss": 0.0404, + "step": 18262 + }, + { + "epoch": 0.428113480663631, + "grad_norm": 0.22352071106433868, + "learning_rate": 0.00012243261787515217, + "loss": 0.0272, + "step": 18263 + }, + { + "epoch": 0.42813692223843597, + "grad_norm": 0.5518500208854675, + "learning_rate": 0.00012242543938949886, + "loss": 0.118, + "step": 18264 + }, + { + "epoch": 0.42816036381324096, + "grad_norm": 0.6422665119171143, + "learning_rate": 0.00012241826078216448, + "loss": 0.1306, + "step": 18265 + }, + { + "epoch": 0.42818380538804596, + "grad_norm": 0.5476027727127075, + "learning_rate": 0.00012241108205318795, + "loss": 0.1856, + "step": 18266 + }, + { + "epoch": 0.42820724696285095, + "grad_norm": 0.5798920392990112, + "learning_rate": 0.00012240390320260829, + "loss": 0.1237, + "step": 18267 + }, + { + "epoch": 0.42823068853765595, + "grad_norm": 0.38181787729263306, + "learning_rate": 0.0001223967242304644, + "loss": 0.05, + "step": 18268 + }, + { + "epoch": 0.42825413011246094, + "grad_norm": 0.31559744477272034, + "learning_rate": 0.00012238954513679528, + "loss": 0.2889, + "step": 18269 + }, + { + "epoch": 0.42827757168726593, + "grad_norm": 0.6280022263526917, + "learning_rate": 0.00012238236592163984, + "loss": 0.1904, + "step": 18270 + }, + { + "epoch": 0.42830101326207093, + "grad_norm": 0.09311484545469284, + "learning_rate": 0.00012237518658503706, + "loss": 0.0184, + "step": 18271 + }, + { + "epoch": 0.4283244548368759, + "grad_norm": 0.6305034756660461, + "learning_rate": 0.00012236800712702586, + "loss": 0.1282, + "step": 18272 + }, + { + "epoch": 0.4283478964116809, + "grad_norm": 0.9066306948661804, + "learning_rate": 0.00012236082754764525, + "loss": 0.1474, + "step": 18273 + }, + { + "epoch": 0.4283713379864859, + "grad_norm": 0.4460716247558594, + "learning_rate": 0.00012235364784693418, + "loss": 0.108, + "step": 18274 + }, + { + "epoch": 0.4283947795612909, + "grad_norm": 0.3469406068325043, + "learning_rate": 0.00012234646802493157, + "loss": 0.0939, + "step": 18275 + }, + { + "epoch": 0.4284182211360959, + "grad_norm": 0.7411766648292542, + "learning_rate": 0.0001223392880816764, + "loss": 0.1054, + "step": 18276 + }, + { + "epoch": 0.4284416627109009, + "grad_norm": 0.25831136107444763, + "learning_rate": 0.0001223321080172076, + "loss": 0.0408, + "step": 18277 + }, + { + "epoch": 0.4284651042857059, + "grad_norm": 0.4018542170524597, + "learning_rate": 0.0001223249278315642, + "loss": 0.0744, + "step": 18278 + }, + { + "epoch": 0.4284885458605109, + "grad_norm": 0.1801736205816269, + "learning_rate": 0.00012231774752478506, + "loss": 0.0291, + "step": 18279 + }, + { + "epoch": 0.4285119874353159, + "grad_norm": 0.3147275447845459, + "learning_rate": 0.0001223105670969092, + "loss": 0.0768, + "step": 18280 + }, + { + "epoch": 0.4285354290101209, + "grad_norm": 0.40413567423820496, + "learning_rate": 0.00012230338654797563, + "loss": 0.0672, + "step": 18281 + }, + { + "epoch": 0.4285588705849259, + "grad_norm": 0.34767237305641174, + "learning_rate": 0.00012229620587802322, + "loss": 0.0324, + "step": 18282 + }, + { + "epoch": 0.4285823121597309, + "grad_norm": 0.694636344909668, + "learning_rate": 0.000122289025087091, + "loss": 0.1247, + "step": 18283 + }, + { + "epoch": 0.4286057537345359, + "grad_norm": 0.20096078515052795, + "learning_rate": 0.00012228184417521786, + "loss": 0.0703, + "step": 18284 + }, + { + "epoch": 0.4286291953093409, + "grad_norm": 0.3821902871131897, + "learning_rate": 0.00012227466314244283, + "loss": 0.0622, + "step": 18285 + }, + { + "epoch": 0.4286526368841459, + "grad_norm": 0.6757070422172546, + "learning_rate": 0.0001222674819888048, + "loss": 0.1757, + "step": 18286 + }, + { + "epoch": 0.4286760784589509, + "grad_norm": 0.46236279606819153, + "learning_rate": 0.00012226030071434287, + "loss": 0.1186, + "step": 18287 + }, + { + "epoch": 0.4286995200337559, + "grad_norm": 0.45013442635536194, + "learning_rate": 0.0001222531193190959, + "loss": 0.076, + "step": 18288 + }, + { + "epoch": 0.4287229616085609, + "grad_norm": 0.25233155488967896, + "learning_rate": 0.00012224593780310285, + "loss": 0.0269, + "step": 18289 + }, + { + "epoch": 0.4287464031833659, + "grad_norm": 0.4207859933376312, + "learning_rate": 0.00012223875616640276, + "loss": 0.1569, + "step": 18290 + }, + { + "epoch": 0.42876984475817087, + "grad_norm": 0.5283132791519165, + "learning_rate": 0.00012223157440903449, + "loss": 0.1082, + "step": 18291 + }, + { + "epoch": 0.42879328633297586, + "grad_norm": 0.9554046988487244, + "learning_rate": 0.00012222439253103712, + "loss": 0.2346, + "step": 18292 + }, + { + "epoch": 0.42881672790778086, + "grad_norm": 0.3012985587120056, + "learning_rate": 0.00012221721053244953, + "loss": 0.3077, + "step": 18293 + }, + { + "epoch": 0.42884016948258585, + "grad_norm": 0.419210821390152, + "learning_rate": 0.0001222100284133108, + "loss": 0.1254, + "step": 18294 + }, + { + "epoch": 0.42886361105739085, + "grad_norm": 0.7000311613082886, + "learning_rate": 0.00012220284617365978, + "loss": 0.0553, + "step": 18295 + }, + { + "epoch": 0.42888705263219584, + "grad_norm": 0.6091331839561462, + "learning_rate": 0.00012219566381353552, + "loss": 0.0987, + "step": 18296 + }, + { + "epoch": 0.42891049420700084, + "grad_norm": 0.5284620523452759, + "learning_rate": 0.00012218848133297694, + "loss": 0.1086, + "step": 18297 + }, + { + "epoch": 0.42893393578180583, + "grad_norm": 0.29441970586776733, + "learning_rate": 0.00012218129873202305, + "loss": 0.1034, + "step": 18298 + }, + { + "epoch": 0.4289573773566108, + "grad_norm": 0.5497800707817078, + "learning_rate": 0.00012217411601071277, + "loss": 0.0887, + "step": 18299 + }, + { + "epoch": 0.4289808189314158, + "grad_norm": 0.4657178819179535, + "learning_rate": 0.00012216693316908516, + "loss": 0.0995, + "step": 18300 + }, + { + "epoch": 0.4290042605062208, + "grad_norm": 0.3581220507621765, + "learning_rate": 0.00012215975020717911, + "loss": 0.0847, + "step": 18301 + }, + { + "epoch": 0.4290277020810258, + "grad_norm": 0.21094867587089539, + "learning_rate": 0.00012215256712503367, + "loss": 0.041, + "step": 18302 + }, + { + "epoch": 0.4290511436558308, + "grad_norm": 0.5333085060119629, + "learning_rate": 0.00012214538392268777, + "loss": 0.4934, + "step": 18303 + }, + { + "epoch": 0.4290745852306358, + "grad_norm": 0.28275689482688904, + "learning_rate": 0.00012213820060018036, + "loss": 0.0224, + "step": 18304 + }, + { + "epoch": 0.4290980268054408, + "grad_norm": 0.354768842458725, + "learning_rate": 0.00012213101715755049, + "loss": 0.0816, + "step": 18305 + }, + { + "epoch": 0.4291214683802458, + "grad_norm": 0.2658538222312927, + "learning_rate": 0.0001221238335948371, + "loss": 0.0792, + "step": 18306 + }, + { + "epoch": 0.4291449099550508, + "grad_norm": 0.385406494140625, + "learning_rate": 0.00012211664991207917, + "loss": 0.0955, + "step": 18307 + }, + { + "epoch": 0.4291683515298558, + "grad_norm": 0.3976958096027374, + "learning_rate": 0.00012210946610931563, + "loss": 0.0705, + "step": 18308 + }, + { + "epoch": 0.42919179310466077, + "grad_norm": 0.11473571509122849, + "learning_rate": 0.00012210228218658556, + "loss": 0.0142, + "step": 18309 + }, + { + "epoch": 0.42921523467946576, + "grad_norm": 0.3530401885509491, + "learning_rate": 0.00012209509814392784, + "loss": 0.1043, + "step": 18310 + }, + { + "epoch": 0.42923867625427076, + "grad_norm": 0.163717120885849, + "learning_rate": 0.00012208791398138152, + "loss": 0.0409, + "step": 18311 + }, + { + "epoch": 0.42926211782907575, + "grad_norm": 1.2702873945236206, + "learning_rate": 0.00012208072969898556, + "loss": 0.104, + "step": 18312 + }, + { + "epoch": 0.42928555940388075, + "grad_norm": 0.7431073784828186, + "learning_rate": 0.00012207354529677894, + "loss": 0.1629, + "step": 18313 + }, + { + "epoch": 0.42930900097868574, + "grad_norm": 0.7401965856552124, + "learning_rate": 0.00012206636077480062, + "loss": 0.0537, + "step": 18314 + }, + { + "epoch": 0.42933244255349073, + "grad_norm": 0.4020996391773224, + "learning_rate": 0.00012205917613308963, + "loss": 0.0991, + "step": 18315 + }, + { + "epoch": 0.42935588412829573, + "grad_norm": 0.7185876369476318, + "learning_rate": 0.00012205199137168495, + "loss": 0.1978, + "step": 18316 + }, + { + "epoch": 0.4293793257031007, + "grad_norm": 0.3950897455215454, + "learning_rate": 0.00012204480649062549, + "loss": 0.0735, + "step": 18317 + }, + { + "epoch": 0.4294027672779057, + "grad_norm": 0.4838024973869324, + "learning_rate": 0.00012203762148995031, + "loss": 0.104, + "step": 18318 + }, + { + "epoch": 0.4294262088527107, + "grad_norm": 0.4532347619533539, + "learning_rate": 0.0001220304363696984, + "loss": 0.1045, + "step": 18319 + }, + { + "epoch": 0.4294496504275157, + "grad_norm": 0.3006364703178406, + "learning_rate": 0.00012202325112990869, + "loss": 0.0365, + "step": 18320 + }, + { + "epoch": 0.4294730920023207, + "grad_norm": 0.5977652668952942, + "learning_rate": 0.00012201606577062024, + "loss": 0.6547, + "step": 18321 + }, + { + "epoch": 0.4294965335771257, + "grad_norm": 0.5292467474937439, + "learning_rate": 0.00012200888029187196, + "loss": 0.1399, + "step": 18322 + }, + { + "epoch": 0.4295199751519307, + "grad_norm": 0.549814760684967, + "learning_rate": 0.00012200169469370289, + "loss": 0.165, + "step": 18323 + }, + { + "epoch": 0.4295434167267357, + "grad_norm": 0.16093672811985016, + "learning_rate": 0.00012199450897615201, + "loss": 0.0224, + "step": 18324 + }, + { + "epoch": 0.4295668583015407, + "grad_norm": 0.8803834915161133, + "learning_rate": 0.00012198732313925831, + "loss": 0.0749, + "step": 18325 + }, + { + "epoch": 0.42959029987634567, + "grad_norm": 0.17394159734249115, + "learning_rate": 0.00012198013718306076, + "loss": 0.0429, + "step": 18326 + }, + { + "epoch": 0.42961374145115067, + "grad_norm": 0.430606871843338, + "learning_rate": 0.00012197295110759839, + "loss": 0.0985, + "step": 18327 + }, + { + "epoch": 0.42963718302595566, + "grad_norm": 0.739337146282196, + "learning_rate": 0.00012196576491291017, + "loss": 0.1117, + "step": 18328 + }, + { + "epoch": 0.42966062460076065, + "grad_norm": 0.4227277934551239, + "learning_rate": 0.00012195857859903506, + "loss": 0.078, + "step": 18329 + }, + { + "epoch": 0.42968406617556565, + "grad_norm": 0.41365480422973633, + "learning_rate": 0.00012195139216601208, + "loss": 0.0885, + "step": 18330 + }, + { + "epoch": 0.42970750775037064, + "grad_norm": 0.5068076252937317, + "learning_rate": 0.00012194420561388028, + "loss": 0.043, + "step": 18331 + }, + { + "epoch": 0.4297309493251757, + "grad_norm": 0.1648019403219223, + "learning_rate": 0.00012193701894267857, + "loss": 0.0201, + "step": 18332 + }, + { + "epoch": 0.4297543908999807, + "grad_norm": 0.43115609884262085, + "learning_rate": 0.00012192983215244598, + "loss": 0.0523, + "step": 18333 + }, + { + "epoch": 0.4297778324747857, + "grad_norm": 0.5747120976448059, + "learning_rate": 0.00012192264524322153, + "loss": 0.1444, + "step": 18334 + }, + { + "epoch": 0.4298012740495907, + "grad_norm": 0.3471578359603882, + "learning_rate": 0.00012191545821504413, + "loss": 0.083, + "step": 18335 + }, + { + "epoch": 0.42982471562439567, + "grad_norm": 0.21516162157058716, + "learning_rate": 0.00012190827106795289, + "loss": 0.0382, + "step": 18336 + }, + { + "epoch": 0.42984815719920066, + "grad_norm": 0.4145839512348175, + "learning_rate": 0.00012190108380198675, + "loss": 0.1012, + "step": 18337 + }, + { + "epoch": 0.42987159877400566, + "grad_norm": 0.19642187654972076, + "learning_rate": 0.0001218938964171847, + "loss": 0.0196, + "step": 18338 + }, + { + "epoch": 0.42989504034881065, + "grad_norm": 0.43394285440444946, + "learning_rate": 0.00012188670891358576, + "loss": 0.0638, + "step": 18339 + }, + { + "epoch": 0.42991848192361565, + "grad_norm": 0.8311657309532166, + "learning_rate": 0.00012187952129122892, + "loss": 0.201, + "step": 18340 + }, + { + "epoch": 0.42994192349842064, + "grad_norm": 0.47958627343177795, + "learning_rate": 0.00012187233355015317, + "loss": 0.6268, + "step": 18341 + }, + { + "epoch": 0.42996536507322564, + "grad_norm": 0.12372738122940063, + "learning_rate": 0.00012186514569039756, + "loss": 0.0309, + "step": 18342 + }, + { + "epoch": 0.42998880664803063, + "grad_norm": 0.34562447667121887, + "learning_rate": 0.00012185795771200101, + "loss": 0.0574, + "step": 18343 + }, + { + "epoch": 0.4300122482228356, + "grad_norm": 0.44623950123786926, + "learning_rate": 0.0001218507696150026, + "loss": 0.0722, + "step": 18344 + }, + { + "epoch": 0.4300356897976406, + "grad_norm": 0.6334419846534729, + "learning_rate": 0.00012184358139944128, + "loss": 0.1005, + "step": 18345 + }, + { + "epoch": 0.4300591313724456, + "grad_norm": 0.35225003957748413, + "learning_rate": 0.00012183639306535608, + "loss": 0.06, + "step": 18346 + }, + { + "epoch": 0.4300825729472506, + "grad_norm": 0.3517150580883026, + "learning_rate": 0.00012182920461278597, + "loss": 0.0392, + "step": 18347 + }, + { + "epoch": 0.4301060145220556, + "grad_norm": 0.15977945923805237, + "learning_rate": 0.00012182201604177003, + "loss": 0.0281, + "step": 18348 + }, + { + "epoch": 0.4301294560968606, + "grad_norm": 0.20922960340976715, + "learning_rate": 0.00012181482735234719, + "loss": 0.0288, + "step": 18349 + }, + { + "epoch": 0.4301528976716656, + "grad_norm": 0.39167651534080505, + "learning_rate": 0.0001218076385445565, + "loss": 0.0473, + "step": 18350 + }, + { + "epoch": 0.4301763392464706, + "grad_norm": 0.18791690468788147, + "learning_rate": 0.00012180044961843696, + "loss": 0.0207, + "step": 18351 + }, + { + "epoch": 0.4301997808212756, + "grad_norm": 0.09691961109638214, + "learning_rate": 0.00012179326057402752, + "loss": 0.0125, + "step": 18352 + }, + { + "epoch": 0.4302232223960806, + "grad_norm": 0.7928451895713806, + "learning_rate": 0.00012178607141136725, + "loss": 0.155, + "step": 18353 + }, + { + "epoch": 0.43024666397088557, + "grad_norm": 0.6694973707199097, + "learning_rate": 0.00012177888213049518, + "loss": 0.1769, + "step": 18354 + }, + { + "epoch": 0.43027010554569056, + "grad_norm": 0.05606300011277199, + "learning_rate": 0.00012177169273145024, + "loss": 0.0093, + "step": 18355 + }, + { + "epoch": 0.43029354712049556, + "grad_norm": 0.755928099155426, + "learning_rate": 0.0001217645032142715, + "loss": 0.1389, + "step": 18356 + }, + { + "epoch": 0.43031698869530055, + "grad_norm": 0.4600154161453247, + "learning_rate": 0.00012175731357899798, + "loss": 0.0935, + "step": 18357 + }, + { + "epoch": 0.43034043027010555, + "grad_norm": 0.5117748975753784, + "learning_rate": 0.00012175012382566862, + "loss": 0.1331, + "step": 18358 + }, + { + "epoch": 0.43036387184491054, + "grad_norm": 0.21911051869392395, + "learning_rate": 0.00012174293395432248, + "loss": 0.0346, + "step": 18359 + }, + { + "epoch": 0.43038731341971553, + "grad_norm": 0.30939674377441406, + "learning_rate": 0.0001217357439649986, + "loss": 0.0703, + "step": 18360 + }, + { + "epoch": 0.43041075499452053, + "grad_norm": 0.3744431138038635, + "learning_rate": 0.00012172855385773596, + "loss": 0.0851, + "step": 18361 + }, + { + "epoch": 0.4304341965693255, + "grad_norm": 0.32469886541366577, + "learning_rate": 0.00012172136363257356, + "loss": 0.0426, + "step": 18362 + }, + { + "epoch": 0.4304576381441305, + "grad_norm": 0.7760801911354065, + "learning_rate": 0.00012171417328955044, + "loss": 0.104, + "step": 18363 + }, + { + "epoch": 0.4304810797189355, + "grad_norm": 0.36103060841560364, + "learning_rate": 0.00012170698282870559, + "loss": 0.0379, + "step": 18364 + }, + { + "epoch": 0.4305045212937405, + "grad_norm": 0.6263872385025024, + "learning_rate": 0.00012169979225007804, + "loss": 0.0869, + "step": 18365 + }, + { + "epoch": 0.4305279628685455, + "grad_norm": 0.2619091868400574, + "learning_rate": 0.00012169260155370683, + "loss": 0.0368, + "step": 18366 + }, + { + "epoch": 0.4305514044433505, + "grad_norm": 0.22038638591766357, + "learning_rate": 0.00012168541073963096, + "loss": 0.0481, + "step": 18367 + }, + { + "epoch": 0.4305748460181555, + "grad_norm": 0.1311264932155609, + "learning_rate": 0.0001216782198078894, + "loss": 0.0241, + "step": 18368 + }, + { + "epoch": 0.4305982875929605, + "grad_norm": 0.2203410118818283, + "learning_rate": 0.00012167102875852126, + "loss": 0.0405, + "step": 18369 + }, + { + "epoch": 0.4306217291677655, + "grad_norm": 0.4006845951080322, + "learning_rate": 0.00012166383759156547, + "loss": 0.0778, + "step": 18370 + }, + { + "epoch": 0.43064517074257047, + "grad_norm": 0.3269847631454468, + "learning_rate": 0.00012165664630706111, + "loss": 0.0554, + "step": 18371 + }, + { + "epoch": 0.43066861231737547, + "grad_norm": 0.5402309894561768, + "learning_rate": 0.00012164945490504713, + "loss": 0.0807, + "step": 18372 + }, + { + "epoch": 0.43069205389218046, + "grad_norm": 0.45714157819747925, + "learning_rate": 0.00012164226338556267, + "loss": 0.1135, + "step": 18373 + }, + { + "epoch": 0.43071549546698545, + "grad_norm": 0.48861438035964966, + "learning_rate": 0.00012163507174864664, + "loss": 0.0839, + "step": 18374 + }, + { + "epoch": 0.43073893704179045, + "grad_norm": 0.6509950160980225, + "learning_rate": 0.0001216278799943381, + "loss": 0.1255, + "step": 18375 + }, + { + "epoch": 0.43076237861659544, + "grad_norm": 0.44081151485443115, + "learning_rate": 0.00012162068812267611, + "loss": 0.105, + "step": 18376 + }, + { + "epoch": 0.43078582019140044, + "grad_norm": 0.2959405779838562, + "learning_rate": 0.00012161349613369961, + "loss": 0.0765, + "step": 18377 + }, + { + "epoch": 0.43080926176620543, + "grad_norm": 0.47316884994506836, + "learning_rate": 0.00012160630402744768, + "loss": 0.5349, + "step": 18378 + }, + { + "epoch": 0.4308327033410104, + "grad_norm": 0.2757807970046997, + "learning_rate": 0.00012159911180395937, + "loss": 0.0427, + "step": 18379 + }, + { + "epoch": 0.4308561449158154, + "grad_norm": 0.7531532049179077, + "learning_rate": 0.00012159191946327366, + "loss": 0.077, + "step": 18380 + }, + { + "epoch": 0.4308795864906204, + "grad_norm": 0.36346709728240967, + "learning_rate": 0.00012158472700542958, + "loss": 0.0602, + "step": 18381 + }, + { + "epoch": 0.4309030280654254, + "grad_norm": 0.8342222571372986, + "learning_rate": 0.00012157753443046618, + "loss": 0.223, + "step": 18382 + }, + { + "epoch": 0.43092646964023046, + "grad_norm": 0.8023208975791931, + "learning_rate": 0.00012157034173842243, + "loss": 0.2169, + "step": 18383 + }, + { + "epoch": 0.43094991121503545, + "grad_norm": 0.3868570029735565, + "learning_rate": 0.00012156314892933742, + "loss": 0.025, + "step": 18384 + }, + { + "epoch": 0.43097335278984045, + "grad_norm": 0.5169109106063843, + "learning_rate": 0.00012155595600325018, + "loss": 0.0655, + "step": 18385 + }, + { + "epoch": 0.43099679436464544, + "grad_norm": 0.373820960521698, + "learning_rate": 0.00012154876296019971, + "loss": 0.4561, + "step": 18386 + }, + { + "epoch": 0.43102023593945044, + "grad_norm": 0.17498333752155304, + "learning_rate": 0.00012154156980022502, + "loss": 0.0364, + "step": 18387 + }, + { + "epoch": 0.43104367751425543, + "grad_norm": 0.3972924053668976, + "learning_rate": 0.00012153437652336519, + "loss": 0.082, + "step": 18388 + }, + { + "epoch": 0.4310671190890604, + "grad_norm": 0.6001479625701904, + "learning_rate": 0.00012152718312965925, + "loss": 0.587, + "step": 18389 + }, + { + "epoch": 0.4310905606638654, + "grad_norm": 0.25855135917663574, + "learning_rate": 0.00012151998961914612, + "loss": 0.0278, + "step": 18390 + }, + { + "epoch": 0.4311140022386704, + "grad_norm": 0.5999819040298462, + "learning_rate": 0.000121512795991865, + "loss": 0.2177, + "step": 18391 + }, + { + "epoch": 0.4311374438134754, + "grad_norm": 0.3990262746810913, + "learning_rate": 0.00012150560224785485, + "loss": 0.1005, + "step": 18392 + }, + { + "epoch": 0.4311608853882804, + "grad_norm": 0.6034466028213501, + "learning_rate": 0.00012149840838715465, + "loss": 0.1305, + "step": 18393 + }, + { + "epoch": 0.4311843269630854, + "grad_norm": 0.5135246515274048, + "learning_rate": 0.00012149121440980353, + "loss": 0.1237, + "step": 18394 + }, + { + "epoch": 0.4312077685378904, + "grad_norm": 0.21754354238510132, + "learning_rate": 0.00012148402031584046, + "loss": 0.0522, + "step": 18395 + }, + { + "epoch": 0.4312312101126954, + "grad_norm": 0.29803624749183655, + "learning_rate": 0.00012147682610530447, + "loss": 0.0251, + "step": 18396 + }, + { + "epoch": 0.4312546516875004, + "grad_norm": 0.42364075779914856, + "learning_rate": 0.00012146963177823463, + "loss": 0.0738, + "step": 18397 + }, + { + "epoch": 0.4312780932623054, + "grad_norm": 0.4097188413143158, + "learning_rate": 0.00012146243733466995, + "loss": 0.0739, + "step": 18398 + }, + { + "epoch": 0.43130153483711037, + "grad_norm": 0.5480936169624329, + "learning_rate": 0.00012145524277464953, + "loss": 0.0638, + "step": 18399 + }, + { + "epoch": 0.43132497641191536, + "grad_norm": 0.29795077443122864, + "learning_rate": 0.00012144804809821231, + "loss": 0.0659, + "step": 18400 + }, + { + "epoch": 0.43134841798672036, + "grad_norm": 0.5507327318191528, + "learning_rate": 0.0001214408533053974, + "loss": 0.1147, + "step": 18401 + }, + { + "epoch": 0.43137185956152535, + "grad_norm": 0.24808159470558167, + "learning_rate": 0.00012143365839624379, + "loss": 0.0596, + "step": 18402 + }, + { + "epoch": 0.43139530113633034, + "grad_norm": 0.7152336835861206, + "learning_rate": 0.00012142646337079055, + "loss": 0.0792, + "step": 18403 + }, + { + "epoch": 0.43141874271113534, + "grad_norm": 0.39026519656181335, + "learning_rate": 0.00012141926822907676, + "loss": 0.0772, + "step": 18404 + }, + { + "epoch": 0.43144218428594033, + "grad_norm": 0.5037596821784973, + "learning_rate": 0.00012141207297114141, + "loss": 0.1228, + "step": 18405 + }, + { + "epoch": 0.4314656258607453, + "grad_norm": 0.42099133133888245, + "learning_rate": 0.00012140487759702353, + "loss": 0.0706, + "step": 18406 + }, + { + "epoch": 0.4314890674355503, + "grad_norm": 0.1693289428949356, + "learning_rate": 0.00012139768210676219, + "loss": 0.0292, + "step": 18407 + }, + { + "epoch": 0.4315125090103553, + "grad_norm": 0.5082495212554932, + "learning_rate": 0.0001213904865003964, + "loss": 0.0642, + "step": 18408 + }, + { + "epoch": 0.4315359505851603, + "grad_norm": 0.3073553740978241, + "learning_rate": 0.00012138329077796524, + "loss": 0.0413, + "step": 18409 + }, + { + "epoch": 0.4315593921599653, + "grad_norm": 0.7429680228233337, + "learning_rate": 0.00012137609493950776, + "loss": 0.639, + "step": 18410 + }, + { + "epoch": 0.4315828337347703, + "grad_norm": 0.6095271706581116, + "learning_rate": 0.00012136889898506298, + "loss": 0.1278, + "step": 18411 + }, + { + "epoch": 0.4316062753095753, + "grad_norm": 0.17439304292201996, + "learning_rate": 0.00012136170291466994, + "loss": 0.03, + "step": 18412 + }, + { + "epoch": 0.4316297168843803, + "grad_norm": 0.5049945712089539, + "learning_rate": 0.00012135450672836774, + "loss": 0.0633, + "step": 18413 + }, + { + "epoch": 0.4316531584591853, + "grad_norm": 0.2768179774284363, + "learning_rate": 0.00012134731042619534, + "loss": 0.0449, + "step": 18414 + }, + { + "epoch": 0.4316766000339903, + "grad_norm": 0.5908000469207764, + "learning_rate": 0.00012134011400819187, + "loss": 0.1226, + "step": 18415 + }, + { + "epoch": 0.43170004160879527, + "grad_norm": 0.5305533409118652, + "learning_rate": 0.00012133291747439632, + "loss": 0.0582, + "step": 18416 + }, + { + "epoch": 0.43172348318360027, + "grad_norm": 0.44850486516952515, + "learning_rate": 0.00012132572082484776, + "loss": 0.1016, + "step": 18417 + }, + { + "epoch": 0.43174692475840526, + "grad_norm": 0.341921329498291, + "learning_rate": 0.00012131852405958526, + "loss": 0.0445, + "step": 18418 + }, + { + "epoch": 0.43177036633321025, + "grad_norm": 0.5411810278892517, + "learning_rate": 0.00012131132717864781, + "loss": 0.1088, + "step": 18419 + }, + { + "epoch": 0.43179380790801525, + "grad_norm": 0.43923965096473694, + "learning_rate": 0.00012130413018207451, + "loss": 0.0693, + "step": 18420 + }, + { + "epoch": 0.43181724948282024, + "grad_norm": 0.8088317513465881, + "learning_rate": 0.00012129693306990443, + "loss": 0.1367, + "step": 18421 + }, + { + "epoch": 0.43184069105762524, + "grad_norm": 0.6284001469612122, + "learning_rate": 0.00012128973584217658, + "loss": 0.1434, + "step": 18422 + }, + { + "epoch": 0.43186413263243023, + "grad_norm": 0.21049511432647705, + "learning_rate": 0.00012128253849893001, + "loss": 0.0341, + "step": 18423 + }, + { + "epoch": 0.4318875742072352, + "grad_norm": 0.1880580633878708, + "learning_rate": 0.00012127534104020382, + "loss": 0.0227, + "step": 18424 + }, + { + "epoch": 0.4319110157820402, + "grad_norm": 0.27240657806396484, + "learning_rate": 0.00012126814346603699, + "loss": 0.0348, + "step": 18425 + }, + { + "epoch": 0.4319344573568452, + "grad_norm": 0.3909108340740204, + "learning_rate": 0.00012126094577646863, + "loss": 0.0467, + "step": 18426 + }, + { + "epoch": 0.4319578989316502, + "grad_norm": 0.19333158433437347, + "learning_rate": 0.00012125374797153779, + "loss": 0.0348, + "step": 18427 + }, + { + "epoch": 0.4319813405064552, + "grad_norm": 0.11938606947660446, + "learning_rate": 0.00012124655005128349, + "loss": 0.021, + "step": 18428 + }, + { + "epoch": 0.4320047820812602, + "grad_norm": 0.17580096423625946, + "learning_rate": 0.00012123935201574485, + "loss": 0.036, + "step": 18429 + }, + { + "epoch": 0.4320282236560652, + "grad_norm": 0.5326219797134399, + "learning_rate": 0.00012123215386496087, + "loss": 0.1226, + "step": 18430 + }, + { + "epoch": 0.4320516652308702, + "grad_norm": 0.6232505440711975, + "learning_rate": 0.00012122495559897063, + "loss": 0.1066, + "step": 18431 + }, + { + "epoch": 0.4320751068056752, + "grad_norm": 0.49075978994369507, + "learning_rate": 0.00012121775721781317, + "loss": 0.1245, + "step": 18432 + }, + { + "epoch": 0.4320985483804802, + "grad_norm": 0.39959731698036194, + "learning_rate": 0.00012121055872152758, + "loss": 0.0787, + "step": 18433 + }, + { + "epoch": 0.43212198995528517, + "grad_norm": 0.33466458320617676, + "learning_rate": 0.00012120336011015289, + "loss": 0.0819, + "step": 18434 + }, + { + "epoch": 0.4321454315300902, + "grad_norm": 0.34046971797943115, + "learning_rate": 0.00012119616138372818, + "loss": 0.0765, + "step": 18435 + }, + { + "epoch": 0.4321688731048952, + "grad_norm": 0.5559329986572266, + "learning_rate": 0.00012118896254229253, + "loss": 0.0852, + "step": 18436 + }, + { + "epoch": 0.4321923146797002, + "grad_norm": 0.6242532134056091, + "learning_rate": 0.00012118176358588493, + "loss": 0.0959, + "step": 18437 + }, + { + "epoch": 0.4322157562545052, + "grad_norm": 0.28886064887046814, + "learning_rate": 0.00012117456451454447, + "loss": 0.063, + "step": 18438 + }, + { + "epoch": 0.4322391978293102, + "grad_norm": 0.5859123468399048, + "learning_rate": 0.00012116736532831027, + "loss": 0.7277, + "step": 18439 + }, + { + "epoch": 0.4322626394041152, + "grad_norm": 0.15531370043754578, + "learning_rate": 0.00012116016602722136, + "loss": 0.0264, + "step": 18440 + }, + { + "epoch": 0.4322860809789202, + "grad_norm": 0.25168266892433167, + "learning_rate": 0.00012115296661131674, + "loss": 0.046, + "step": 18441 + }, + { + "epoch": 0.4323095225537252, + "grad_norm": 0.621652364730835, + "learning_rate": 0.00012114576708063559, + "loss": 0.0821, + "step": 18442 + }, + { + "epoch": 0.4323329641285302, + "grad_norm": 0.637734591960907, + "learning_rate": 0.00012113856743521687, + "loss": 0.4622, + "step": 18443 + }, + { + "epoch": 0.43235640570333517, + "grad_norm": 0.36357855796813965, + "learning_rate": 0.00012113136767509972, + "loss": 0.0627, + "step": 18444 + }, + { + "epoch": 0.43237984727814016, + "grad_norm": 0.22141028940677643, + "learning_rate": 0.00012112416780032312, + "loss": 0.0394, + "step": 18445 + }, + { + "epoch": 0.43240328885294516, + "grad_norm": 0.13798120617866516, + "learning_rate": 0.00012111696781092625, + "loss": 0.0279, + "step": 18446 + }, + { + "epoch": 0.43242673042775015, + "grad_norm": 0.18053317070007324, + "learning_rate": 0.00012110976770694808, + "loss": 0.0452, + "step": 18447 + }, + { + "epoch": 0.43245017200255514, + "grad_norm": 0.4031915068626404, + "learning_rate": 0.00012110256748842774, + "loss": 0.1082, + "step": 18448 + }, + { + "epoch": 0.43247361357736014, + "grad_norm": 0.19622932374477386, + "learning_rate": 0.00012109536715540428, + "loss": 0.033, + "step": 18449 + }, + { + "epoch": 0.43249705515216513, + "grad_norm": 0.5861278176307678, + "learning_rate": 0.00012108816670791676, + "loss": 0.0898, + "step": 18450 + }, + { + "epoch": 0.4325204967269701, + "grad_norm": 0.5036653876304626, + "learning_rate": 0.00012108096614600421, + "loss": 0.7263, + "step": 18451 + }, + { + "epoch": 0.4325439383017751, + "grad_norm": 0.523786187171936, + "learning_rate": 0.00012107376546970581, + "loss": 0.1346, + "step": 18452 + }, + { + "epoch": 0.4325673798765801, + "grad_norm": 0.4010632038116455, + "learning_rate": 0.00012106656467906055, + "loss": 0.08, + "step": 18453 + }, + { + "epoch": 0.4325908214513851, + "grad_norm": 0.5806302428245544, + "learning_rate": 0.0001210593637741075, + "loss": 0.0913, + "step": 18454 + }, + { + "epoch": 0.4326142630261901, + "grad_norm": 0.5248380303382874, + "learning_rate": 0.00012105216275488576, + "loss": 0.6282, + "step": 18455 + }, + { + "epoch": 0.4326377046009951, + "grad_norm": 0.619812548160553, + "learning_rate": 0.00012104496162143442, + "loss": 0.1391, + "step": 18456 + }, + { + "epoch": 0.4326611461758001, + "grad_norm": 0.6575249433517456, + "learning_rate": 0.00012103776037379244, + "loss": 0.1485, + "step": 18457 + }, + { + "epoch": 0.4326845877506051, + "grad_norm": 0.4664037823677063, + "learning_rate": 0.00012103055901199907, + "loss": 0.0485, + "step": 18458 + }, + { + "epoch": 0.4327080293254101, + "grad_norm": 0.4285522997379303, + "learning_rate": 0.00012102335753609326, + "loss": 0.0604, + "step": 18459 + }, + { + "epoch": 0.4327314709002151, + "grad_norm": 0.45485904812812805, + "learning_rate": 0.0001210161559461141, + "loss": 0.0554, + "step": 18460 + }, + { + "epoch": 0.43275491247502007, + "grad_norm": 0.6725266575813293, + "learning_rate": 0.00012100895424210072, + "loss": 0.8647, + "step": 18461 + }, + { + "epoch": 0.43277835404982506, + "grad_norm": 0.3740650415420532, + "learning_rate": 0.00012100175242409218, + "loss": 0.0686, + "step": 18462 + }, + { + "epoch": 0.43280179562463006, + "grad_norm": 0.15236493945121765, + "learning_rate": 0.0001209945504921275, + "loss": 0.0282, + "step": 18463 + }, + { + "epoch": 0.43282523719943505, + "grad_norm": 0.38417738676071167, + "learning_rate": 0.0001209873484462458, + "loss": 0.0845, + "step": 18464 + }, + { + "epoch": 0.43284867877424005, + "grad_norm": 0.3508607745170593, + "learning_rate": 0.00012098014628648618, + "loss": 0.0515, + "step": 18465 + }, + { + "epoch": 0.43287212034904504, + "grad_norm": 0.38718339800834656, + "learning_rate": 0.00012097294401288768, + "loss": 0.0658, + "step": 18466 + }, + { + "epoch": 0.43289556192385004, + "grad_norm": 0.43678009510040283, + "learning_rate": 0.00012096574162548942, + "loss": 0.0736, + "step": 18467 + }, + { + "epoch": 0.43291900349865503, + "grad_norm": 0.3894180357456207, + "learning_rate": 0.00012095853912433044, + "loss": 0.0398, + "step": 18468 + }, + { + "epoch": 0.43294244507346, + "grad_norm": 0.21037788689136505, + "learning_rate": 0.00012095133650944984, + "loss": 0.0245, + "step": 18469 + }, + { + "epoch": 0.432965886648265, + "grad_norm": 1.0525847673416138, + "learning_rate": 0.00012094413378088666, + "loss": 0.0885, + "step": 18470 + }, + { + "epoch": 0.43298932822307, + "grad_norm": 0.6645625233650208, + "learning_rate": 0.00012093693093868008, + "loss": 0.113, + "step": 18471 + }, + { + "epoch": 0.433012769797875, + "grad_norm": 0.7336456179618835, + "learning_rate": 0.00012092972798286913, + "loss": 0.6995, + "step": 18472 + }, + { + "epoch": 0.43303621137268, + "grad_norm": 0.722369909286499, + "learning_rate": 0.00012092252491349283, + "loss": 0.5479, + "step": 18473 + }, + { + "epoch": 0.433059652947485, + "grad_norm": 0.15314382314682007, + "learning_rate": 0.00012091532173059037, + "loss": 0.0215, + "step": 18474 + }, + { + "epoch": 0.43308309452229, + "grad_norm": 0.5889340043067932, + "learning_rate": 0.00012090811843420074, + "loss": 0.7383, + "step": 18475 + }, + { + "epoch": 0.433106536097095, + "grad_norm": 0.9780542254447937, + "learning_rate": 0.0001209009150243631, + "loss": 0.1077, + "step": 18476 + }, + { + "epoch": 0.4331299776719, + "grad_norm": 1.0164462327957153, + "learning_rate": 0.0001208937115011165, + "loss": 0.1046, + "step": 18477 + }, + { + "epoch": 0.433153419246705, + "grad_norm": 0.19079934060573578, + "learning_rate": 0.00012088650786450007, + "loss": 0.0349, + "step": 18478 + }, + { + "epoch": 0.43317686082150997, + "grad_norm": 0.32329729199409485, + "learning_rate": 0.00012087930411455281, + "loss": 0.1082, + "step": 18479 + }, + { + "epoch": 0.43320030239631496, + "grad_norm": 0.25937148928642273, + "learning_rate": 0.00012087210025131387, + "loss": 0.03, + "step": 18480 + }, + { + "epoch": 0.43322374397111996, + "grad_norm": 0.2379029244184494, + "learning_rate": 0.00012086489627482233, + "loss": 0.0371, + "step": 18481 + }, + { + "epoch": 0.43324718554592495, + "grad_norm": 0.0676611140370369, + "learning_rate": 0.00012085769218511729, + "loss": 0.0074, + "step": 18482 + }, + { + "epoch": 0.43327062712072995, + "grad_norm": 0.13327385485172272, + "learning_rate": 0.00012085048798223781, + "loss": 0.0241, + "step": 18483 + }, + { + "epoch": 0.43329406869553494, + "grad_norm": 0.3880396783351898, + "learning_rate": 0.00012084328366622302, + "loss": 0.0648, + "step": 18484 + }, + { + "epoch": 0.43331751027033993, + "grad_norm": 0.3722897469997406, + "learning_rate": 0.00012083607923711198, + "loss": 0.079, + "step": 18485 + }, + { + "epoch": 0.433340951845145, + "grad_norm": 0.676798403263092, + "learning_rate": 0.00012082887469494377, + "loss": 0.0631, + "step": 18486 + }, + { + "epoch": 0.43336439341995, + "grad_norm": 0.5401264429092407, + "learning_rate": 0.00012082167003975751, + "loss": 0.0806, + "step": 18487 + }, + { + "epoch": 0.433387834994755, + "grad_norm": 0.755571186542511, + "learning_rate": 0.00012081446527159228, + "loss": 0.1373, + "step": 18488 + }, + { + "epoch": 0.43341127656955997, + "grad_norm": 0.1497010886669159, + "learning_rate": 0.00012080726039048718, + "loss": 0.0139, + "step": 18489 + }, + { + "epoch": 0.43343471814436496, + "grad_norm": 0.4095161259174347, + "learning_rate": 0.00012080005539648129, + "loss": 0.0826, + "step": 18490 + }, + { + "epoch": 0.43345815971916996, + "grad_norm": 0.5331777930259705, + "learning_rate": 0.00012079285028961376, + "loss": 0.0647, + "step": 18491 + }, + { + "epoch": 0.43348160129397495, + "grad_norm": 0.16047517955303192, + "learning_rate": 0.00012078564506992359, + "loss": 0.0294, + "step": 18492 + }, + { + "epoch": 0.43350504286877994, + "grad_norm": 0.4305850863456726, + "learning_rate": 0.0001207784397374499, + "loss": 0.0555, + "step": 18493 + }, + { + "epoch": 0.43352848444358494, + "grad_norm": 0.4973353445529938, + "learning_rate": 0.00012077123429223187, + "loss": 0.0961, + "step": 18494 + }, + { + "epoch": 0.43355192601838993, + "grad_norm": 0.27126631140708923, + "learning_rate": 0.00012076402873430852, + "loss": 0.052, + "step": 18495 + }, + { + "epoch": 0.4335753675931949, + "grad_norm": 0.14675965905189514, + "learning_rate": 0.00012075682306371897, + "loss": 0.0344, + "step": 18496 + }, + { + "epoch": 0.4335988091679999, + "grad_norm": 0.3873271644115448, + "learning_rate": 0.00012074961728050231, + "loss": 0.0899, + "step": 18497 + }, + { + "epoch": 0.4336222507428049, + "grad_norm": 0.17883066833019257, + "learning_rate": 0.00012074241138469765, + "loss": 0.0259, + "step": 18498 + }, + { + "epoch": 0.4336456923176099, + "grad_norm": 0.48254022002220154, + "learning_rate": 0.00012073520537634407, + "loss": 0.1284, + "step": 18499 + }, + { + "epoch": 0.4336691338924149, + "grad_norm": 0.5425277948379517, + "learning_rate": 0.00012072799925548069, + "loss": 0.16, + "step": 18500 + }, + { + "epoch": 0.4336925754672199, + "grad_norm": 0.3846063017845154, + "learning_rate": 0.00012072079302214659, + "loss": 0.1101, + "step": 18501 + }, + { + "epoch": 0.4337160170420249, + "grad_norm": 0.5146716833114624, + "learning_rate": 0.00012071358667638089, + "loss": 0.1147, + "step": 18502 + }, + { + "epoch": 0.4337394586168299, + "grad_norm": 0.35753193497657776, + "learning_rate": 0.00012070638021822272, + "loss": 0.085, + "step": 18503 + }, + { + "epoch": 0.4337629001916349, + "grad_norm": 0.5221149325370789, + "learning_rate": 0.00012069917364771111, + "loss": 0.0703, + "step": 18504 + }, + { + "epoch": 0.4337863417664399, + "grad_norm": 0.464459627866745, + "learning_rate": 0.00012069196696488518, + "loss": 0.1028, + "step": 18505 + }, + { + "epoch": 0.43380978334124487, + "grad_norm": 0.5017939209938049, + "learning_rate": 0.00012068476016978412, + "loss": 0.1053, + "step": 18506 + }, + { + "epoch": 0.43383322491604986, + "grad_norm": 0.44193556904792786, + "learning_rate": 0.00012067755326244693, + "loss": 0.1003, + "step": 18507 + }, + { + "epoch": 0.43385666649085486, + "grad_norm": 0.13723888993263245, + "learning_rate": 0.00012067034624291275, + "loss": 0.0139, + "step": 18508 + }, + { + "epoch": 0.43388010806565985, + "grad_norm": 0.27972811460494995, + "learning_rate": 0.00012066313911122072, + "loss": 0.0507, + "step": 18509 + }, + { + "epoch": 0.43390354964046485, + "grad_norm": 0.5713235139846802, + "learning_rate": 0.00012065593186740991, + "loss": 0.1378, + "step": 18510 + }, + { + "epoch": 0.43392699121526984, + "grad_norm": 0.15753555297851562, + "learning_rate": 0.00012064872451151938, + "loss": 0.0327, + "step": 18511 + }, + { + "epoch": 0.43395043279007484, + "grad_norm": 0.45589953660964966, + "learning_rate": 0.00012064151704358835, + "loss": 0.0994, + "step": 18512 + }, + { + "epoch": 0.43397387436487983, + "grad_norm": 0.11564910411834717, + "learning_rate": 0.00012063430946365586, + "loss": 0.0227, + "step": 18513 + }, + { + "epoch": 0.4339973159396848, + "grad_norm": 0.3767254054546356, + "learning_rate": 0.000120627101771761, + "loss": 0.0861, + "step": 18514 + }, + { + "epoch": 0.4340207575144898, + "grad_norm": 0.6901561617851257, + "learning_rate": 0.00012061989396794293, + "loss": 0.1232, + "step": 18515 + }, + { + "epoch": 0.4340441990892948, + "grad_norm": 0.47593942284584045, + "learning_rate": 0.00012061268605224074, + "loss": 0.1111, + "step": 18516 + }, + { + "epoch": 0.4340676406640998, + "grad_norm": 0.5610255002975464, + "learning_rate": 0.0001206054780246935, + "loss": 0.1039, + "step": 18517 + }, + { + "epoch": 0.4340910822389048, + "grad_norm": 0.3312905430793762, + "learning_rate": 0.00012059826988534038, + "loss": 0.0542, + "step": 18518 + }, + { + "epoch": 0.4341145238137098, + "grad_norm": 0.9389671087265015, + "learning_rate": 0.00012059106163422047, + "loss": 0.3012, + "step": 18519 + }, + { + "epoch": 0.4341379653885148, + "grad_norm": 0.6320354342460632, + "learning_rate": 0.00012058385327137288, + "loss": 0.1828, + "step": 18520 + }, + { + "epoch": 0.4341614069633198, + "grad_norm": 0.5643121004104614, + "learning_rate": 0.00012057664479683673, + "loss": 0.0852, + "step": 18521 + }, + { + "epoch": 0.4341848485381248, + "grad_norm": 0.2588857114315033, + "learning_rate": 0.00012056943621065113, + "loss": 0.0449, + "step": 18522 + }, + { + "epoch": 0.4342082901129298, + "grad_norm": 0.27743691205978394, + "learning_rate": 0.00012056222751285517, + "loss": 0.0538, + "step": 18523 + }, + { + "epoch": 0.43423173168773477, + "grad_norm": 0.3305028975009918, + "learning_rate": 0.00012055501870348798, + "loss": 0.0648, + "step": 18524 + }, + { + "epoch": 0.43425517326253976, + "grad_norm": 0.5272247791290283, + "learning_rate": 0.0001205478097825887, + "loss": 0.1134, + "step": 18525 + }, + { + "epoch": 0.43427861483734476, + "grad_norm": 0.3158084452152252, + "learning_rate": 0.00012054060075019644, + "loss": 0.0679, + "step": 18526 + }, + { + "epoch": 0.43430205641214975, + "grad_norm": 0.4792177081108093, + "learning_rate": 0.00012053339160635029, + "loss": 0.0756, + "step": 18527 + }, + { + "epoch": 0.43432549798695474, + "grad_norm": 0.5880590677261353, + "learning_rate": 0.00012052618235108937, + "loss": 0.0743, + "step": 18528 + }, + { + "epoch": 0.43434893956175974, + "grad_norm": 0.37643110752105713, + "learning_rate": 0.00012051897298445283, + "loss": 0.0708, + "step": 18529 + }, + { + "epoch": 0.43437238113656473, + "grad_norm": 0.31465744972229004, + "learning_rate": 0.0001205117635064797, + "loss": 0.0585, + "step": 18530 + }, + { + "epoch": 0.43439582271136973, + "grad_norm": 0.219486802816391, + "learning_rate": 0.00012050455391720923, + "loss": 0.0537, + "step": 18531 + }, + { + "epoch": 0.4344192642861747, + "grad_norm": 0.702614963054657, + "learning_rate": 0.00012049734421668046, + "loss": 0.168, + "step": 18532 + }, + { + "epoch": 0.4344427058609797, + "grad_norm": 0.17432866990566254, + "learning_rate": 0.0001204901344049325, + "loss": 0.0284, + "step": 18533 + }, + { + "epoch": 0.4344661474357847, + "grad_norm": 0.6896733045578003, + "learning_rate": 0.00012048292448200453, + "loss": 0.1541, + "step": 18534 + }, + { + "epoch": 0.4344895890105897, + "grad_norm": 0.5894426703453064, + "learning_rate": 0.00012047571444793564, + "loss": 0.1359, + "step": 18535 + }, + { + "epoch": 0.4345130305853947, + "grad_norm": 0.40998220443725586, + "learning_rate": 0.00012046850430276493, + "loss": 0.0691, + "step": 18536 + }, + { + "epoch": 0.43453647216019975, + "grad_norm": 0.6602746248245239, + "learning_rate": 0.00012046129404653152, + "loss": 0.1951, + "step": 18537 + }, + { + "epoch": 0.43455991373500474, + "grad_norm": 0.5794771313667297, + "learning_rate": 0.00012045408367927458, + "loss": 0.0837, + "step": 18538 + }, + { + "epoch": 0.43458335530980974, + "grad_norm": 0.5723437070846558, + "learning_rate": 0.00012044687320103319, + "loss": 0.6402, + "step": 18539 + }, + { + "epoch": 0.43460679688461473, + "grad_norm": 0.07694362103939056, + "learning_rate": 0.00012043966261184651, + "loss": 0.0125, + "step": 18540 + }, + { + "epoch": 0.4346302384594197, + "grad_norm": 0.769047200679779, + "learning_rate": 0.00012043245191175365, + "loss": 0.1617, + "step": 18541 + }, + { + "epoch": 0.4346536800342247, + "grad_norm": 0.3016709089279175, + "learning_rate": 0.0001204252411007937, + "loss": 0.0489, + "step": 18542 + }, + { + "epoch": 0.4346771216090297, + "grad_norm": 0.31563296914100647, + "learning_rate": 0.00012041803017900583, + "loss": 0.0645, + "step": 18543 + }, + { + "epoch": 0.4347005631838347, + "grad_norm": 0.8494781255722046, + "learning_rate": 0.00012041081914642918, + "loss": 0.1379, + "step": 18544 + }, + { + "epoch": 0.4347240047586397, + "grad_norm": 0.1789165437221527, + "learning_rate": 0.00012040360800310281, + "loss": 0.0377, + "step": 18545 + }, + { + "epoch": 0.4347474463334447, + "grad_norm": 0.27175283432006836, + "learning_rate": 0.00012039639674906591, + "loss": 0.0514, + "step": 18546 + }, + { + "epoch": 0.4347708879082497, + "grad_norm": 0.3266373574733734, + "learning_rate": 0.00012038918538435759, + "loss": 0.0411, + "step": 18547 + }, + { + "epoch": 0.4347943294830547, + "grad_norm": 0.36330392956733704, + "learning_rate": 0.00012038197390901698, + "loss": 0.0608, + "step": 18548 + }, + { + "epoch": 0.4348177710578597, + "grad_norm": 0.7374926805496216, + "learning_rate": 0.00012037476232308318, + "loss": 0.1639, + "step": 18549 + }, + { + "epoch": 0.4348412126326647, + "grad_norm": 0.7990385293960571, + "learning_rate": 0.00012036755062659537, + "loss": 0.1244, + "step": 18550 + }, + { + "epoch": 0.43486465420746967, + "grad_norm": 0.23026031255722046, + "learning_rate": 0.00012036033881959268, + "loss": 0.0397, + "step": 18551 + }, + { + "epoch": 0.43488809578227466, + "grad_norm": 0.479414701461792, + "learning_rate": 0.00012035312690211417, + "loss": 0.0932, + "step": 18552 + }, + { + "epoch": 0.43491153735707966, + "grad_norm": 0.25350421667099, + "learning_rate": 0.00012034591487419906, + "loss": 0.0738, + "step": 18553 + }, + { + "epoch": 0.43493497893188465, + "grad_norm": 0.4551681578159332, + "learning_rate": 0.00012033870273588643, + "loss": 0.0826, + "step": 18554 + }, + { + "epoch": 0.43495842050668965, + "grad_norm": 0.3318312466144562, + "learning_rate": 0.00012033149048721544, + "loss": 0.0739, + "step": 18555 + }, + { + "epoch": 0.43498186208149464, + "grad_norm": 0.37817850708961487, + "learning_rate": 0.00012032427812822517, + "loss": 0.1005, + "step": 18556 + }, + { + "epoch": 0.43500530365629964, + "grad_norm": 0.12082022428512573, + "learning_rate": 0.00012031706565895484, + "loss": 0.0231, + "step": 18557 + }, + { + "epoch": 0.43502874523110463, + "grad_norm": 0.5484866499900818, + "learning_rate": 0.00012030985307944351, + "loss": 0.5686, + "step": 18558 + }, + { + "epoch": 0.4350521868059096, + "grad_norm": 0.6902454495429993, + "learning_rate": 0.00012030264038973035, + "loss": 0.0937, + "step": 18559 + }, + { + "epoch": 0.4350756283807146, + "grad_norm": 0.38798025250434875, + "learning_rate": 0.0001202954275898545, + "loss": 0.0766, + "step": 18560 + }, + { + "epoch": 0.4350990699555196, + "grad_norm": 0.5322076082229614, + "learning_rate": 0.00012028821467985512, + "loss": 0.6795, + "step": 18561 + }, + { + "epoch": 0.4351225115303246, + "grad_norm": 0.44724205136299133, + "learning_rate": 0.00012028100165977125, + "loss": 0.1156, + "step": 18562 + }, + { + "epoch": 0.4351459531051296, + "grad_norm": 0.39303410053253174, + "learning_rate": 0.00012027378852964215, + "loss": 0.0774, + "step": 18563 + }, + { + "epoch": 0.4351693946799346, + "grad_norm": 0.55629962682724, + "learning_rate": 0.00012026657528950691, + "loss": 0.656, + "step": 18564 + }, + { + "epoch": 0.4351928362547396, + "grad_norm": 0.2838899791240692, + "learning_rate": 0.00012025936193940461, + "loss": 0.0514, + "step": 18565 + }, + { + "epoch": 0.4352162778295446, + "grad_norm": 0.540873110294342, + "learning_rate": 0.00012025214847937447, + "loss": 0.1373, + "step": 18566 + }, + { + "epoch": 0.4352397194043496, + "grad_norm": 0.5683231353759766, + "learning_rate": 0.00012024493490945561, + "loss": 0.2228, + "step": 18567 + }, + { + "epoch": 0.4352631609791546, + "grad_norm": 0.10692895203828812, + "learning_rate": 0.00012023772122968715, + "loss": 0.0122, + "step": 18568 + }, + { + "epoch": 0.43528660255395957, + "grad_norm": 0.7517273426055908, + "learning_rate": 0.00012023050744010826, + "loss": 0.2553, + "step": 18569 + }, + { + "epoch": 0.43531004412876456, + "grad_norm": 0.17035694420337677, + "learning_rate": 0.00012022329354075807, + "loss": 0.0249, + "step": 18570 + }, + { + "epoch": 0.43533348570356956, + "grad_norm": 0.5523582100868225, + "learning_rate": 0.00012021607953167571, + "loss": 0.1079, + "step": 18571 + }, + { + "epoch": 0.43535692727837455, + "grad_norm": 0.21588295698165894, + "learning_rate": 0.0001202088654129003, + "loss": 0.0501, + "step": 18572 + }, + { + "epoch": 0.43538036885317954, + "grad_norm": 0.444560706615448, + "learning_rate": 0.00012020165118447109, + "loss": 0.1021, + "step": 18573 + }, + { + "epoch": 0.43540381042798454, + "grad_norm": 0.6771506071090698, + "learning_rate": 0.00012019443684642712, + "loss": 0.1333, + "step": 18574 + }, + { + "epoch": 0.43542725200278953, + "grad_norm": 0.47258052229881287, + "learning_rate": 0.00012018722239880755, + "loss": 0.0426, + "step": 18575 + }, + { + "epoch": 0.4354506935775945, + "grad_norm": 0.5627678036689758, + "learning_rate": 0.00012018000784165158, + "loss": 0.1273, + "step": 18576 + }, + { + "epoch": 0.4354741351523995, + "grad_norm": 0.34461551904678345, + "learning_rate": 0.00012017279317499827, + "loss": 0.0557, + "step": 18577 + }, + { + "epoch": 0.4354975767272045, + "grad_norm": 0.08940549194812775, + "learning_rate": 0.00012016557839888682, + "loss": 0.0193, + "step": 18578 + }, + { + "epoch": 0.4355210183020095, + "grad_norm": 0.271157830953598, + "learning_rate": 0.00012015836351335643, + "loss": 0.029, + "step": 18579 + }, + { + "epoch": 0.4355444598768145, + "grad_norm": 0.1681888997554779, + "learning_rate": 0.00012015114851844618, + "loss": 0.0232, + "step": 18580 + }, + { + "epoch": 0.4355679014516195, + "grad_norm": 0.427414208650589, + "learning_rate": 0.00012014393341419518, + "loss": 0.04, + "step": 18581 + }, + { + "epoch": 0.4355913430264245, + "grad_norm": 0.4944758117198944, + "learning_rate": 0.00012013671820064269, + "loss": 0.0637, + "step": 18582 + }, + { + "epoch": 0.4356147846012295, + "grad_norm": 0.4009266495704651, + "learning_rate": 0.00012012950287782779, + "loss": 0.0496, + "step": 18583 + }, + { + "epoch": 0.4356382261760345, + "grad_norm": 0.4122588336467743, + "learning_rate": 0.0001201222874457896, + "loss": 0.1086, + "step": 18584 + }, + { + "epoch": 0.4356616677508395, + "grad_norm": 0.5075170993804932, + "learning_rate": 0.00012011507190456734, + "loss": 0.1315, + "step": 18585 + }, + { + "epoch": 0.43568510932564447, + "grad_norm": 0.5125580430030823, + "learning_rate": 0.00012010785625420013, + "loss": 0.08, + "step": 18586 + }, + { + "epoch": 0.43570855090044947, + "grad_norm": 0.4477315843105316, + "learning_rate": 0.00012010064049472714, + "loss": 0.1016, + "step": 18587 + }, + { + "epoch": 0.4357319924752545, + "grad_norm": 0.18676458299160004, + "learning_rate": 0.00012009342462618749, + "loss": 0.0335, + "step": 18588 + }, + { + "epoch": 0.4357554340500595, + "grad_norm": 0.17404130101203918, + "learning_rate": 0.0001200862086486204, + "loss": 0.0186, + "step": 18589 + }, + { + "epoch": 0.4357788756248645, + "grad_norm": 0.48017868399620056, + "learning_rate": 0.00012007899256206491, + "loss": 0.1753, + "step": 18590 + }, + { + "epoch": 0.4358023171996695, + "grad_norm": 0.6506650447845459, + "learning_rate": 0.00012007177636656026, + "loss": 0.1264, + "step": 18591 + }, + { + "epoch": 0.4358257587744745, + "grad_norm": 0.7991200089454651, + "learning_rate": 0.00012006456006214562, + "loss": 0.1511, + "step": 18592 + }, + { + "epoch": 0.4358492003492795, + "grad_norm": 0.5033959746360779, + "learning_rate": 0.00012005734364886009, + "loss": 0.0913, + "step": 18593 + }, + { + "epoch": 0.4358726419240845, + "grad_norm": 0.4179837107658386, + "learning_rate": 0.00012005012712674284, + "loss": 0.0612, + "step": 18594 + }, + { + "epoch": 0.4358960834988895, + "grad_norm": 0.5896621942520142, + "learning_rate": 0.00012004291049583305, + "loss": 0.0773, + "step": 18595 + }, + { + "epoch": 0.43591952507369447, + "grad_norm": 0.3194873034954071, + "learning_rate": 0.00012003569375616985, + "loss": 0.0524, + "step": 18596 + }, + { + "epoch": 0.43594296664849946, + "grad_norm": 0.29664257168769836, + "learning_rate": 0.00012002847690779242, + "loss": 0.0424, + "step": 18597 + }, + { + "epoch": 0.43596640822330446, + "grad_norm": 0.5230752229690552, + "learning_rate": 0.00012002125995073991, + "loss": 0.0914, + "step": 18598 + }, + { + "epoch": 0.43598984979810945, + "grad_norm": 0.334584504365921, + "learning_rate": 0.00012001404288505151, + "loss": 0.063, + "step": 18599 + }, + { + "epoch": 0.43601329137291445, + "grad_norm": 0.10446123778820038, + "learning_rate": 0.00012000682571076631, + "loss": 0.0169, + "step": 18600 + }, + { + "epoch": 0.43603673294771944, + "grad_norm": 0.538956344127655, + "learning_rate": 0.00011999960842792355, + "loss": 0.7836, + "step": 18601 + }, + { + "epoch": 0.43606017452252444, + "grad_norm": 0.2914605140686035, + "learning_rate": 0.00011999239103656233, + "loss": 0.051, + "step": 18602 + }, + { + "epoch": 0.43608361609732943, + "grad_norm": 0.5015743374824524, + "learning_rate": 0.0001199851735367218, + "loss": 0.1035, + "step": 18603 + }, + { + "epoch": 0.4361070576721344, + "grad_norm": 0.5638622641563416, + "learning_rate": 0.00011997795592844122, + "loss": 0.1056, + "step": 18604 + }, + { + "epoch": 0.4361304992469394, + "grad_norm": 0.3236839175224304, + "learning_rate": 0.00011997073821175966, + "loss": 0.0451, + "step": 18605 + }, + { + "epoch": 0.4361539408217444, + "grad_norm": 0.08380353450775146, + "learning_rate": 0.00011996352038671629, + "loss": 0.0086, + "step": 18606 + }, + { + "epoch": 0.4361773823965494, + "grad_norm": 0.48601698875427246, + "learning_rate": 0.00011995630245335032, + "loss": 0.1203, + "step": 18607 + }, + { + "epoch": 0.4362008239713544, + "grad_norm": 0.27334779500961304, + "learning_rate": 0.00011994908441170092, + "loss": 0.0572, + "step": 18608 + }, + { + "epoch": 0.4362242655461594, + "grad_norm": 0.7708706855773926, + "learning_rate": 0.0001199418662618072, + "loss": 0.1273, + "step": 18609 + }, + { + "epoch": 0.4362477071209644, + "grad_norm": 0.46697038412094116, + "learning_rate": 0.00011993464800370833, + "loss": 0.0919, + "step": 18610 + }, + { + "epoch": 0.4362711486957694, + "grad_norm": 0.377560555934906, + "learning_rate": 0.00011992742963744352, + "loss": 0.464, + "step": 18611 + }, + { + "epoch": 0.4362945902705744, + "grad_norm": 0.5926446914672852, + "learning_rate": 0.00011992021116305192, + "loss": 0.1282, + "step": 18612 + }, + { + "epoch": 0.4363180318453794, + "grad_norm": 0.4230823814868927, + "learning_rate": 0.00011991299258057269, + "loss": 0.1239, + "step": 18613 + }, + { + "epoch": 0.43634147342018437, + "grad_norm": 0.39257144927978516, + "learning_rate": 0.00011990577389004501, + "loss": 0.0372, + "step": 18614 + }, + { + "epoch": 0.43636491499498936, + "grad_norm": 0.5053682923316956, + "learning_rate": 0.00011989855509150802, + "loss": 0.6781, + "step": 18615 + }, + { + "epoch": 0.43638835656979436, + "grad_norm": 0.23856709897518158, + "learning_rate": 0.00011989133618500092, + "loss": 0.0335, + "step": 18616 + }, + { + "epoch": 0.43641179814459935, + "grad_norm": 0.26663345098495483, + "learning_rate": 0.00011988411717056289, + "loss": 0.0321, + "step": 18617 + }, + { + "epoch": 0.43643523971940434, + "grad_norm": 0.40198272466659546, + "learning_rate": 0.00011987689804823308, + "loss": 0.0749, + "step": 18618 + }, + { + "epoch": 0.43645868129420934, + "grad_norm": 1.172118902206421, + "learning_rate": 0.00011986967881805063, + "loss": 0.0982, + "step": 18619 + }, + { + "epoch": 0.43648212286901433, + "grad_norm": 0.6836522817611694, + "learning_rate": 0.00011986245948005477, + "loss": 0.1218, + "step": 18620 + }, + { + "epoch": 0.4365055644438193, + "grad_norm": 0.5596814155578613, + "learning_rate": 0.00011985524003428463, + "loss": 0.0888, + "step": 18621 + }, + { + "epoch": 0.4365290060186243, + "grad_norm": 0.47307148575782776, + "learning_rate": 0.00011984802048077941, + "loss": 0.0572, + "step": 18622 + }, + { + "epoch": 0.4365524475934293, + "grad_norm": 0.53468257188797, + "learning_rate": 0.00011984080081957827, + "loss": 0.0585, + "step": 18623 + }, + { + "epoch": 0.4365758891682343, + "grad_norm": 0.35222646594047546, + "learning_rate": 0.0001198335810507204, + "loss": 0.092, + "step": 18624 + }, + { + "epoch": 0.4365993307430393, + "grad_norm": 0.6594772934913635, + "learning_rate": 0.00011982636117424495, + "loss": 0.157, + "step": 18625 + }, + { + "epoch": 0.4366227723178443, + "grad_norm": 0.5589398741722107, + "learning_rate": 0.0001198191411901911, + "loss": 0.0895, + "step": 18626 + }, + { + "epoch": 0.4366462138926493, + "grad_norm": 0.5654404163360596, + "learning_rate": 0.00011981192109859805, + "loss": 0.1397, + "step": 18627 + }, + { + "epoch": 0.4366696554674543, + "grad_norm": 0.3986319899559021, + "learning_rate": 0.00011980470089950495, + "loss": 0.0639, + "step": 18628 + }, + { + "epoch": 0.4366930970422593, + "grad_norm": 0.3540768325328827, + "learning_rate": 0.00011979748059295097, + "loss": 0.0783, + "step": 18629 + }, + { + "epoch": 0.4367165386170643, + "grad_norm": 0.5162176489830017, + "learning_rate": 0.00011979026017897532, + "loss": 0.5001, + "step": 18630 + }, + { + "epoch": 0.43673998019186927, + "grad_norm": 0.5482103228569031, + "learning_rate": 0.00011978303965761718, + "loss": 0.1243, + "step": 18631 + }, + { + "epoch": 0.43676342176667426, + "grad_norm": 0.7831047773361206, + "learning_rate": 0.00011977581902891568, + "loss": 0.1492, + "step": 18632 + }, + { + "epoch": 0.43678686334147926, + "grad_norm": 0.6033049821853638, + "learning_rate": 0.00011976859829291003, + "loss": 0.1049, + "step": 18633 + }, + { + "epoch": 0.43681030491628425, + "grad_norm": 0.4664370119571686, + "learning_rate": 0.00011976137744963942, + "loss": 0.078, + "step": 18634 + }, + { + "epoch": 0.43683374649108925, + "grad_norm": 0.7918118238449097, + "learning_rate": 0.00011975415649914304, + "loss": 0.0926, + "step": 18635 + }, + { + "epoch": 0.43685718806589424, + "grad_norm": 0.606850802898407, + "learning_rate": 0.00011974693544146003, + "loss": 0.0881, + "step": 18636 + }, + { + "epoch": 0.43688062964069924, + "grad_norm": 0.6721004247665405, + "learning_rate": 0.0001197397142766296, + "loss": 0.0881, + "step": 18637 + }, + { + "epoch": 0.43690407121550423, + "grad_norm": 0.3955049514770508, + "learning_rate": 0.00011973249300469092, + "loss": 0.0855, + "step": 18638 + }, + { + "epoch": 0.4369275127903093, + "grad_norm": 0.3121303617954254, + "learning_rate": 0.00011972527162568316, + "loss": 0.1295, + "step": 18639 + }, + { + "epoch": 0.4369509543651143, + "grad_norm": 0.5567218661308289, + "learning_rate": 0.00011971805013964555, + "loss": 0.1349, + "step": 18640 + }, + { + "epoch": 0.43697439593991927, + "grad_norm": 0.20950432121753693, + "learning_rate": 0.00011971082854661721, + "loss": 0.032, + "step": 18641 + }, + { + "epoch": 0.43699783751472426, + "grad_norm": 0.29568204283714294, + "learning_rate": 0.00011970360684663742, + "loss": 0.0554, + "step": 18642 + }, + { + "epoch": 0.43702127908952926, + "grad_norm": 0.5103851556777954, + "learning_rate": 0.00011969638503974527, + "loss": 0.1674, + "step": 18643 + }, + { + "epoch": 0.43704472066433425, + "grad_norm": 0.35431522130966187, + "learning_rate": 0.00011968916312597998, + "loss": 0.0398, + "step": 18644 + }, + { + "epoch": 0.43706816223913925, + "grad_norm": 0.1609257608652115, + "learning_rate": 0.00011968194110538071, + "loss": 0.0297, + "step": 18645 + }, + { + "epoch": 0.43709160381394424, + "grad_norm": 0.7607953548431396, + "learning_rate": 0.00011967471897798672, + "loss": 0.1349, + "step": 18646 + }, + { + "epoch": 0.43711504538874923, + "grad_norm": 0.23383192718029022, + "learning_rate": 0.00011966749674383715, + "loss": 0.0656, + "step": 18647 + }, + { + "epoch": 0.43713848696355423, + "grad_norm": 0.6739006638526917, + "learning_rate": 0.00011966027440297116, + "loss": 0.1194, + "step": 18648 + }, + { + "epoch": 0.4371619285383592, + "grad_norm": 0.48175591230392456, + "learning_rate": 0.000119653051955428, + "loss": 0.1427, + "step": 18649 + }, + { + "epoch": 0.4371853701131642, + "grad_norm": 0.3476828336715698, + "learning_rate": 0.00011964582940124682, + "loss": 0.1352, + "step": 18650 + }, + { + "epoch": 0.4372088116879692, + "grad_norm": 0.11821237206459045, + "learning_rate": 0.00011963860674046677, + "loss": 0.0222, + "step": 18651 + }, + { + "epoch": 0.4372322532627742, + "grad_norm": 0.3900964856147766, + "learning_rate": 0.00011963138397312715, + "loss": 0.1043, + "step": 18652 + }, + { + "epoch": 0.4372556948375792, + "grad_norm": 0.18107005953788757, + "learning_rate": 0.00011962416109926706, + "loss": 0.0264, + "step": 18653 + }, + { + "epoch": 0.4372791364123842, + "grad_norm": 0.4423693120479584, + "learning_rate": 0.00011961693811892571, + "loss": 0.1039, + "step": 18654 + }, + { + "epoch": 0.4373025779871892, + "grad_norm": 0.6965243220329285, + "learning_rate": 0.00011960971503214233, + "loss": 0.0956, + "step": 18655 + }, + { + "epoch": 0.4373260195619942, + "grad_norm": 0.7826374173164368, + "learning_rate": 0.00011960249183895609, + "loss": 0.1302, + "step": 18656 + }, + { + "epoch": 0.4373494611367992, + "grad_norm": 0.9355246424674988, + "learning_rate": 0.00011959526853940612, + "loss": 0.1565, + "step": 18657 + }, + { + "epoch": 0.4373729027116042, + "grad_norm": 0.19802843034267426, + "learning_rate": 0.00011958804513353173, + "loss": 0.0335, + "step": 18658 + }, + { + "epoch": 0.43739634428640917, + "grad_norm": 0.7574199438095093, + "learning_rate": 0.00011958082162137203, + "loss": 0.0839, + "step": 18659 + }, + { + "epoch": 0.43741978586121416, + "grad_norm": 0.518308162689209, + "learning_rate": 0.00011957359800296624, + "loss": 0.0785, + "step": 18660 + }, + { + "epoch": 0.43744322743601916, + "grad_norm": 0.436896413564682, + "learning_rate": 0.00011956637427835358, + "loss": 0.0721, + "step": 18661 + }, + { + "epoch": 0.43746666901082415, + "grad_norm": 0.23708078265190125, + "learning_rate": 0.00011955915044757322, + "loss": 0.0496, + "step": 18662 + }, + { + "epoch": 0.43749011058562914, + "grad_norm": 0.24671891331672668, + "learning_rate": 0.00011955192651066434, + "loss": 0.0328, + "step": 18663 + }, + { + "epoch": 0.43751355216043414, + "grad_norm": 0.3857559263706207, + "learning_rate": 0.00011954470246766615, + "loss": 0.0644, + "step": 18664 + }, + { + "epoch": 0.43753699373523913, + "grad_norm": 0.436366468667984, + "learning_rate": 0.00011953747831861789, + "loss": 0.0973, + "step": 18665 + }, + { + "epoch": 0.4375604353100441, + "grad_norm": 0.42273011803627014, + "learning_rate": 0.00011953025406355871, + "loss": 0.1577, + "step": 18666 + }, + { + "epoch": 0.4375838768848491, + "grad_norm": 0.2677377164363861, + "learning_rate": 0.0001195230297025278, + "loss": 0.0379, + "step": 18667 + }, + { + "epoch": 0.4376073184596541, + "grad_norm": 0.20815807580947876, + "learning_rate": 0.00011951580523556441, + "loss": 0.0438, + "step": 18668 + }, + { + "epoch": 0.4376307600344591, + "grad_norm": 0.09939385205507278, + "learning_rate": 0.00011950858066270769, + "loss": 0.0115, + "step": 18669 + }, + { + "epoch": 0.4376542016092641, + "grad_norm": 0.6266546845436096, + "learning_rate": 0.00011950135598399686, + "loss": 0.1179, + "step": 18670 + }, + { + "epoch": 0.4376776431840691, + "grad_norm": 0.39878347516059875, + "learning_rate": 0.00011949413119947115, + "loss": 0.0635, + "step": 18671 + }, + { + "epoch": 0.4377010847588741, + "grad_norm": 0.48299291729927063, + "learning_rate": 0.00011948690630916975, + "loss": 0.0777, + "step": 18672 + }, + { + "epoch": 0.4377245263336791, + "grad_norm": 0.6086690425872803, + "learning_rate": 0.00011947968131313179, + "loss": 0.1428, + "step": 18673 + }, + { + "epoch": 0.4377479679084841, + "grad_norm": 0.44009512662887573, + "learning_rate": 0.00011947245621139658, + "loss": 0.1148, + "step": 18674 + }, + { + "epoch": 0.4377714094832891, + "grad_norm": 0.1461200714111328, + "learning_rate": 0.00011946523100400326, + "loss": 0.0372, + "step": 18675 + }, + { + "epoch": 0.43779485105809407, + "grad_norm": 0.43553832173347473, + "learning_rate": 0.00011945800569099102, + "loss": 0.0727, + "step": 18676 + }, + { + "epoch": 0.43781829263289906, + "grad_norm": 0.16781537234783173, + "learning_rate": 0.00011945078027239913, + "loss": 0.0264, + "step": 18677 + }, + { + "epoch": 0.43784173420770406, + "grad_norm": 0.17950855195522308, + "learning_rate": 0.00011944355474826676, + "loss": 0.0286, + "step": 18678 + }, + { + "epoch": 0.43786517578250905, + "grad_norm": 0.45056772232055664, + "learning_rate": 0.00011943632911863309, + "loss": 0.0872, + "step": 18679 + }, + { + "epoch": 0.43788861735731405, + "grad_norm": 0.6841634511947632, + "learning_rate": 0.00011942910338353739, + "loss": 0.6047, + "step": 18680 + }, + { + "epoch": 0.43791205893211904, + "grad_norm": 0.48241910338401794, + "learning_rate": 0.00011942187754301881, + "loss": 0.1123, + "step": 18681 + }, + { + "epoch": 0.43793550050692404, + "grad_norm": 0.5285466313362122, + "learning_rate": 0.00011941465159711656, + "loss": 0.1049, + "step": 18682 + }, + { + "epoch": 0.43795894208172903, + "grad_norm": 0.32962000370025635, + "learning_rate": 0.00011940742554586985, + "loss": 0.0559, + "step": 18683 + }, + { + "epoch": 0.437982383656534, + "grad_norm": 0.2472154051065445, + "learning_rate": 0.00011940019938931795, + "loss": 0.0576, + "step": 18684 + }, + { + "epoch": 0.438005825231339, + "grad_norm": 0.050710469484329224, + "learning_rate": 0.00011939297312750002, + "loss": 0.0098, + "step": 18685 + }, + { + "epoch": 0.438029266806144, + "grad_norm": 0.48606637120246887, + "learning_rate": 0.00011938574676045524, + "loss": 0.1186, + "step": 18686 + }, + { + "epoch": 0.438052708380949, + "grad_norm": 0.5427677631378174, + "learning_rate": 0.00011937852028822287, + "loss": 0.1112, + "step": 18687 + }, + { + "epoch": 0.438076149955754, + "grad_norm": 0.2978229820728302, + "learning_rate": 0.0001193712937108421, + "loss": 0.0749, + "step": 18688 + }, + { + "epoch": 0.438099591530559, + "grad_norm": 0.4520573914051056, + "learning_rate": 0.00011936406702835214, + "loss": 0.1424, + "step": 18689 + }, + { + "epoch": 0.43812303310536405, + "grad_norm": 0.7442111968994141, + "learning_rate": 0.00011935684024079223, + "loss": 0.1142, + "step": 18690 + }, + { + "epoch": 0.43814647468016904, + "grad_norm": 0.6294404864311218, + "learning_rate": 0.00011934961334820156, + "loss": 0.0503, + "step": 18691 + }, + { + "epoch": 0.43816991625497403, + "grad_norm": 0.3543986976146698, + "learning_rate": 0.00011934238635061931, + "loss": 0.0443, + "step": 18692 + }, + { + "epoch": 0.43819335782977903, + "grad_norm": 0.5578708648681641, + "learning_rate": 0.00011933515924808477, + "loss": 0.1281, + "step": 18693 + }, + { + "epoch": 0.438216799404584, + "grad_norm": 0.3193129301071167, + "learning_rate": 0.00011932793204063708, + "loss": 0.0548, + "step": 18694 + }, + { + "epoch": 0.438240240979389, + "grad_norm": 0.39082518219947815, + "learning_rate": 0.00011932070472831547, + "loss": 0.0926, + "step": 18695 + }, + { + "epoch": 0.438263682554194, + "grad_norm": 0.5173609256744385, + "learning_rate": 0.00011931347731115923, + "loss": 0.1353, + "step": 18696 + }, + { + "epoch": 0.438287124128999, + "grad_norm": 0.32320436835289, + "learning_rate": 0.00011930624978920748, + "loss": 0.2611, + "step": 18697 + }, + { + "epoch": 0.438310565703804, + "grad_norm": 0.3331701457500458, + "learning_rate": 0.00011929902216249948, + "loss": 0.0725, + "step": 18698 + }, + { + "epoch": 0.438334007278609, + "grad_norm": 0.16860809922218323, + "learning_rate": 0.00011929179443107445, + "loss": 0.0337, + "step": 18699 + }, + { + "epoch": 0.438357448853414, + "grad_norm": 0.33277878165245056, + "learning_rate": 0.0001192845665949716, + "loss": 0.042, + "step": 18700 + }, + { + "epoch": 0.438380890428219, + "grad_norm": 0.3151845335960388, + "learning_rate": 0.00011927733865423016, + "loss": 0.0805, + "step": 18701 + }, + { + "epoch": 0.438404332003024, + "grad_norm": 0.6553575396537781, + "learning_rate": 0.0001192701106088893, + "loss": 0.1759, + "step": 18702 + }, + { + "epoch": 0.43842777357782897, + "grad_norm": 0.5781931281089783, + "learning_rate": 0.00011926288245898834, + "loss": 0.648, + "step": 18703 + }, + { + "epoch": 0.43845121515263397, + "grad_norm": 0.5175037384033203, + "learning_rate": 0.00011925565420456641, + "loss": 0.0988, + "step": 18704 + }, + { + "epoch": 0.43847465672743896, + "grad_norm": 0.547186017036438, + "learning_rate": 0.00011924842584566274, + "loss": 0.1562, + "step": 18705 + }, + { + "epoch": 0.43849809830224395, + "grad_norm": 0.3987734019756317, + "learning_rate": 0.00011924119738231657, + "loss": 0.1091, + "step": 18706 + }, + { + "epoch": 0.43852153987704895, + "grad_norm": 0.17536015808582306, + "learning_rate": 0.00011923396881456715, + "loss": 0.0173, + "step": 18707 + }, + { + "epoch": 0.43854498145185394, + "grad_norm": 0.3060256540775299, + "learning_rate": 0.00011922674014245365, + "loss": 0.1014, + "step": 18708 + }, + { + "epoch": 0.43856842302665894, + "grad_norm": 0.6180941462516785, + "learning_rate": 0.00011921951136601533, + "loss": 0.0845, + "step": 18709 + }, + { + "epoch": 0.43859186460146393, + "grad_norm": 0.7526373863220215, + "learning_rate": 0.00011921228248529142, + "loss": 0.1801, + "step": 18710 + }, + { + "epoch": 0.4386153061762689, + "grad_norm": 0.28934136033058167, + "learning_rate": 0.00011920505350032107, + "loss": 0.051, + "step": 18711 + }, + { + "epoch": 0.4386387477510739, + "grad_norm": 0.15328539907932281, + "learning_rate": 0.00011919782441114358, + "loss": 0.0311, + "step": 18712 + }, + { + "epoch": 0.4386621893258789, + "grad_norm": 0.12682199478149414, + "learning_rate": 0.00011919059521779819, + "loss": 0.0196, + "step": 18713 + }, + { + "epoch": 0.4386856309006839, + "grad_norm": 0.2316531091928482, + "learning_rate": 0.00011918336592032404, + "loss": 0.0433, + "step": 18714 + }, + { + "epoch": 0.4387090724754889, + "grad_norm": 0.6077195405960083, + "learning_rate": 0.00011917613651876045, + "loss": 0.1402, + "step": 18715 + }, + { + "epoch": 0.4387325140502939, + "grad_norm": 0.32673755288124084, + "learning_rate": 0.0001191689070131466, + "loss": 0.043, + "step": 18716 + }, + { + "epoch": 0.4387559556250989, + "grad_norm": 1.0108240842819214, + "learning_rate": 0.00011916167740352171, + "loss": 0.2618, + "step": 18717 + }, + { + "epoch": 0.4387793971999039, + "grad_norm": 0.6539466381072998, + "learning_rate": 0.00011915444768992501, + "loss": 0.102, + "step": 18718 + }, + { + "epoch": 0.4388028387747089, + "grad_norm": 0.21894720196723938, + "learning_rate": 0.00011914721787239576, + "loss": 0.0386, + "step": 18719 + }, + { + "epoch": 0.4388262803495139, + "grad_norm": 0.806237518787384, + "learning_rate": 0.00011913998795097316, + "loss": 0.1507, + "step": 18720 + }, + { + "epoch": 0.43884972192431887, + "grad_norm": 0.4015085697174072, + "learning_rate": 0.00011913275792569643, + "loss": 0.2695, + "step": 18721 + }, + { + "epoch": 0.43887316349912386, + "grad_norm": 0.3450060486793518, + "learning_rate": 0.00011912552779660484, + "loss": 0.0542, + "step": 18722 + }, + { + "epoch": 0.43889660507392886, + "grad_norm": 0.4307279884815216, + "learning_rate": 0.0001191182975637376, + "loss": 0.0977, + "step": 18723 + }, + { + "epoch": 0.43892004664873385, + "grad_norm": 0.6714643239974976, + "learning_rate": 0.0001191110672271339, + "loss": 0.6955, + "step": 18724 + }, + { + "epoch": 0.43894348822353885, + "grad_norm": 0.5477673411369324, + "learning_rate": 0.00011910383678683305, + "loss": 0.083, + "step": 18725 + }, + { + "epoch": 0.43896692979834384, + "grad_norm": 0.23582158982753754, + "learning_rate": 0.00011909660624287425, + "loss": 0.0407, + "step": 18726 + }, + { + "epoch": 0.43899037137314884, + "grad_norm": 0.15812858939170837, + "learning_rate": 0.0001190893755952967, + "loss": 0.0219, + "step": 18727 + }, + { + "epoch": 0.43901381294795383, + "grad_norm": 0.7496152520179749, + "learning_rate": 0.00011908214484413969, + "loss": 0.1102, + "step": 18728 + }, + { + "epoch": 0.4390372545227588, + "grad_norm": 0.2979866862297058, + "learning_rate": 0.00011907491398944242, + "loss": 0.0495, + "step": 18729 + }, + { + "epoch": 0.4390606960975638, + "grad_norm": 0.5894970893859863, + "learning_rate": 0.00011906768303124409, + "loss": 0.1038, + "step": 18730 + }, + { + "epoch": 0.4390841376723688, + "grad_norm": 0.4622367322444916, + "learning_rate": 0.000119060451969584, + "loss": 0.0991, + "step": 18731 + }, + { + "epoch": 0.4391075792471738, + "grad_norm": 0.3925388753414154, + "learning_rate": 0.00011905322080450139, + "loss": 0.0822, + "step": 18732 + }, + { + "epoch": 0.4391310208219788, + "grad_norm": 0.5438684821128845, + "learning_rate": 0.00011904598953603542, + "loss": 0.1066, + "step": 18733 + }, + { + "epoch": 0.4391544623967838, + "grad_norm": 0.22306916117668152, + "learning_rate": 0.00011903875816422543, + "loss": 0.0415, + "step": 18734 + }, + { + "epoch": 0.4391779039715888, + "grad_norm": 0.2638497054576874, + "learning_rate": 0.00011903152668911057, + "loss": 0.02, + "step": 18735 + }, + { + "epoch": 0.4392013455463938, + "grad_norm": 0.1127275824546814, + "learning_rate": 0.0001190242951107301, + "loss": 0.0122, + "step": 18736 + }, + { + "epoch": 0.4392247871211988, + "grad_norm": 0.10554006695747375, + "learning_rate": 0.00011901706342912328, + "loss": 0.014, + "step": 18737 + }, + { + "epoch": 0.4392482286960038, + "grad_norm": 0.4871062636375427, + "learning_rate": 0.00011900983164432936, + "loss": 0.1012, + "step": 18738 + }, + { + "epoch": 0.43927167027080877, + "grad_norm": 0.13015778362751007, + "learning_rate": 0.00011900259975638755, + "loss": 0.0148, + "step": 18739 + }, + { + "epoch": 0.43929511184561376, + "grad_norm": 0.108600914478302, + "learning_rate": 0.00011899536776533709, + "loss": 0.0233, + "step": 18740 + }, + { + "epoch": 0.4393185534204188, + "grad_norm": 0.08007248491048813, + "learning_rate": 0.00011898813567121723, + "loss": 0.0142, + "step": 18741 + }, + { + "epoch": 0.4393419949952238, + "grad_norm": 0.46693092584609985, + "learning_rate": 0.00011898090347406721, + "loss": 0.6817, + "step": 18742 + }, + { + "epoch": 0.4393654365700288, + "grad_norm": 0.6301419138908386, + "learning_rate": 0.00011897367117392627, + "loss": 0.1357, + "step": 18743 + }, + { + "epoch": 0.4393888781448338, + "grad_norm": 0.11421965807676315, + "learning_rate": 0.00011896643877083367, + "loss": 0.0144, + "step": 18744 + }, + { + "epoch": 0.4394123197196388, + "grad_norm": 0.46331673860549927, + "learning_rate": 0.00011895920626482863, + "loss": 0.0427, + "step": 18745 + }, + { + "epoch": 0.4394357612944438, + "grad_norm": 0.6673301458358765, + "learning_rate": 0.00011895197365595041, + "loss": 0.0656, + "step": 18746 + }, + { + "epoch": 0.4394592028692488, + "grad_norm": 0.5264275670051575, + "learning_rate": 0.00011894474094423826, + "loss": 0.0952, + "step": 18747 + }, + { + "epoch": 0.43948264444405377, + "grad_norm": 0.12291093915700912, + "learning_rate": 0.0001189375081297314, + "loss": 0.0129, + "step": 18748 + }, + { + "epoch": 0.43950608601885877, + "grad_norm": 1.0840357542037964, + "learning_rate": 0.0001189302752124691, + "loss": 0.1561, + "step": 18749 + }, + { + "epoch": 0.43952952759366376, + "grad_norm": 0.3535124957561493, + "learning_rate": 0.00011892304219249055, + "loss": 0.0771, + "step": 18750 + }, + { + "epoch": 0.43955296916846875, + "grad_norm": 0.5270281434059143, + "learning_rate": 0.00011891580906983508, + "loss": 0.0964, + "step": 18751 + }, + { + "epoch": 0.43957641074327375, + "grad_norm": 0.7278668284416199, + "learning_rate": 0.0001189085758445419, + "loss": 0.1102, + "step": 18752 + }, + { + "epoch": 0.43959985231807874, + "grad_norm": 0.8059981465339661, + "learning_rate": 0.00011890134251665026, + "loss": 0.0911, + "step": 18753 + }, + { + "epoch": 0.43962329389288374, + "grad_norm": 0.23286423087120056, + "learning_rate": 0.00011889410908619937, + "loss": 0.0162, + "step": 18754 + }, + { + "epoch": 0.43964673546768873, + "grad_norm": 0.4719640910625458, + "learning_rate": 0.00011888687555322855, + "loss": 0.1143, + "step": 18755 + }, + { + "epoch": 0.4396701770424937, + "grad_norm": 0.6069931387901306, + "learning_rate": 0.00011887964191777699, + "loss": 0.071, + "step": 18756 + }, + { + "epoch": 0.4396936186172987, + "grad_norm": 0.3511047065258026, + "learning_rate": 0.00011887240817988396, + "loss": 0.0642, + "step": 18757 + }, + { + "epoch": 0.4397170601921037, + "grad_norm": 0.8964203596115112, + "learning_rate": 0.00011886517433958875, + "loss": 0.1534, + "step": 18758 + }, + { + "epoch": 0.4397405017669087, + "grad_norm": 0.43364405632019043, + "learning_rate": 0.00011885794039693052, + "loss": 0.0692, + "step": 18759 + }, + { + "epoch": 0.4397639433417137, + "grad_norm": 0.918651819229126, + "learning_rate": 0.00011885070635194861, + "loss": 0.3338, + "step": 18760 + }, + { + "epoch": 0.4397873849165187, + "grad_norm": 0.14760780334472656, + "learning_rate": 0.00011884347220468222, + "loss": 0.0271, + "step": 18761 + }, + { + "epoch": 0.4398108264913237, + "grad_norm": 0.13870403170585632, + "learning_rate": 0.00011883623795517062, + "loss": 0.0233, + "step": 18762 + }, + { + "epoch": 0.4398342680661287, + "grad_norm": 0.42213672399520874, + "learning_rate": 0.00011882900360345308, + "loss": 0.068, + "step": 18763 + }, + { + "epoch": 0.4398577096409337, + "grad_norm": 0.6454298496246338, + "learning_rate": 0.00011882176914956884, + "loss": 0.1212, + "step": 18764 + }, + { + "epoch": 0.4398811512157387, + "grad_norm": 0.4532947242259979, + "learning_rate": 0.00011881453459355713, + "loss": 0.0866, + "step": 18765 + }, + { + "epoch": 0.43990459279054367, + "grad_norm": 0.3629102110862732, + "learning_rate": 0.00011880729993545725, + "loss": 0.0593, + "step": 18766 + }, + { + "epoch": 0.43992803436534866, + "grad_norm": 0.5823991894721985, + "learning_rate": 0.00011880006517530839, + "loss": 0.1528, + "step": 18767 + }, + { + "epoch": 0.43995147594015366, + "grad_norm": 0.4413352608680725, + "learning_rate": 0.0001187928303131499, + "loss": 0.5802, + "step": 18768 + }, + { + "epoch": 0.43997491751495865, + "grad_norm": 0.2623179852962494, + "learning_rate": 0.00011878559534902095, + "loss": 0.0434, + "step": 18769 + }, + { + "epoch": 0.43999835908976365, + "grad_norm": 0.16763174533843994, + "learning_rate": 0.00011877836028296083, + "loss": 0.0218, + "step": 18770 + }, + { + "epoch": 0.44002180066456864, + "grad_norm": 0.5552038550376892, + "learning_rate": 0.00011877112511500881, + "loss": 0.0546, + "step": 18771 + }, + { + "epoch": 0.44004524223937364, + "grad_norm": 0.561551034450531, + "learning_rate": 0.00011876388984520414, + "loss": 0.1327, + "step": 18772 + }, + { + "epoch": 0.44006868381417863, + "grad_norm": 0.38916119933128357, + "learning_rate": 0.00011875665447358606, + "loss": 0.074, + "step": 18773 + }, + { + "epoch": 0.4400921253889836, + "grad_norm": 0.4972449541091919, + "learning_rate": 0.00011874941900019387, + "loss": 0.1271, + "step": 18774 + }, + { + "epoch": 0.4401155669637886, + "grad_norm": 0.45905882120132446, + "learning_rate": 0.00011874218342506678, + "loss": 0.1014, + "step": 18775 + }, + { + "epoch": 0.4401390085385936, + "grad_norm": 0.15518692135810852, + "learning_rate": 0.00011873494774824412, + "loss": 0.0384, + "step": 18776 + }, + { + "epoch": 0.4401624501133986, + "grad_norm": 0.44641321897506714, + "learning_rate": 0.00011872771196976509, + "loss": 0.0688, + "step": 18777 + }, + { + "epoch": 0.4401858916882036, + "grad_norm": 0.28708919882774353, + "learning_rate": 0.00011872047608966892, + "loss": 0.0723, + "step": 18778 + }, + { + "epoch": 0.4402093332630086, + "grad_norm": 0.380013644695282, + "learning_rate": 0.00011871324010799496, + "loss": 0.0587, + "step": 18779 + }, + { + "epoch": 0.4402327748378136, + "grad_norm": 0.12320293486118317, + "learning_rate": 0.00011870600402478242, + "loss": 0.024, + "step": 18780 + }, + { + "epoch": 0.4402562164126186, + "grad_norm": 0.19537724554538727, + "learning_rate": 0.00011869876784007059, + "loss": 0.0375, + "step": 18781 + }, + { + "epoch": 0.4402796579874236, + "grad_norm": 0.1685011088848114, + "learning_rate": 0.00011869153155389872, + "loss": 0.0323, + "step": 18782 + }, + { + "epoch": 0.4403030995622286, + "grad_norm": 0.49608302116394043, + "learning_rate": 0.00011868429516630609, + "loss": 0.1428, + "step": 18783 + }, + { + "epoch": 0.44032654113703357, + "grad_norm": 0.29104578495025635, + "learning_rate": 0.0001186770586773319, + "loss": 0.0612, + "step": 18784 + }, + { + "epoch": 0.44034998271183856, + "grad_norm": 0.4616275429725647, + "learning_rate": 0.0001186698220870155, + "loss": 0.08, + "step": 18785 + }, + { + "epoch": 0.44037342428664356, + "grad_norm": 0.7637836933135986, + "learning_rate": 0.00011866258539539612, + "loss": 0.1446, + "step": 18786 + }, + { + "epoch": 0.44039686586144855, + "grad_norm": 0.34855061769485474, + "learning_rate": 0.000118655348602513, + "loss": 0.0588, + "step": 18787 + }, + { + "epoch": 0.44042030743625354, + "grad_norm": 0.20277929306030273, + "learning_rate": 0.00011864811170840546, + "loss": 0.0355, + "step": 18788 + }, + { + "epoch": 0.44044374901105854, + "grad_norm": 0.7554590106010437, + "learning_rate": 0.00011864087471311274, + "loss": 0.1318, + "step": 18789 + }, + { + "epoch": 0.44046719058586353, + "grad_norm": 0.20725469291210175, + "learning_rate": 0.0001186336376166741, + "loss": 0.0287, + "step": 18790 + }, + { + "epoch": 0.4404906321606685, + "grad_norm": 0.24743717908859253, + "learning_rate": 0.0001186264004191288, + "loss": 0.0755, + "step": 18791 + }, + { + "epoch": 0.4405140737354735, + "grad_norm": 0.47281065583229065, + "learning_rate": 0.00011861916312051617, + "loss": 0.0997, + "step": 18792 + }, + { + "epoch": 0.44053751531027857, + "grad_norm": 0.3707830309867859, + "learning_rate": 0.00011861192572087542, + "loss": 0.0546, + "step": 18793 + }, + { + "epoch": 0.44056095688508357, + "grad_norm": 0.1646915078163147, + "learning_rate": 0.00011860468822024584, + "loss": 0.027, + "step": 18794 + }, + { + "epoch": 0.44058439845988856, + "grad_norm": 0.3855610191822052, + "learning_rate": 0.00011859745061866669, + "loss": 0.0673, + "step": 18795 + }, + { + "epoch": 0.44060784003469355, + "grad_norm": 0.2431011199951172, + "learning_rate": 0.00011859021291617727, + "loss": 0.0249, + "step": 18796 + }, + { + "epoch": 0.44063128160949855, + "grad_norm": 0.7607524991035461, + "learning_rate": 0.00011858297511281677, + "loss": 0.0974, + "step": 18797 + }, + { + "epoch": 0.44065472318430354, + "grad_norm": 0.15816034376621246, + "learning_rate": 0.0001185757372086246, + "loss": 0.0283, + "step": 18798 + }, + { + "epoch": 0.44067816475910854, + "grad_norm": 0.3421500027179718, + "learning_rate": 0.00011856849920363993, + "loss": 0.107, + "step": 18799 + }, + { + "epoch": 0.44070160633391353, + "grad_norm": 0.40424540638923645, + "learning_rate": 0.00011856126109790204, + "loss": 0.0949, + "step": 18800 + }, + { + "epoch": 0.4407250479087185, + "grad_norm": 0.48091137409210205, + "learning_rate": 0.00011855402289145025, + "loss": 0.0921, + "step": 18801 + }, + { + "epoch": 0.4407484894835235, + "grad_norm": 0.5473427176475525, + "learning_rate": 0.00011854678458432382, + "loss": 0.1661, + "step": 18802 + }, + { + "epoch": 0.4407719310583285, + "grad_norm": 0.322723925113678, + "learning_rate": 0.000118539546176562, + "loss": 0.0231, + "step": 18803 + }, + { + "epoch": 0.4407953726331335, + "grad_norm": 0.5400674343109131, + "learning_rate": 0.00011853230766820407, + "loss": 0.092, + "step": 18804 + }, + { + "epoch": 0.4408188142079385, + "grad_norm": 0.13568054139614105, + "learning_rate": 0.00011852506905928936, + "loss": 0.0314, + "step": 18805 + }, + { + "epoch": 0.4408422557827435, + "grad_norm": 0.37812739610671997, + "learning_rate": 0.00011851783034985705, + "loss": 0.0974, + "step": 18806 + }, + { + "epoch": 0.4408656973575485, + "grad_norm": 0.6819260120391846, + "learning_rate": 0.0001185105915399465, + "loss": 0.1329, + "step": 18807 + }, + { + "epoch": 0.4408891389323535, + "grad_norm": 0.20117171108722687, + "learning_rate": 0.00011850335262959698, + "loss": 0.021, + "step": 18808 + }, + { + "epoch": 0.4409125805071585, + "grad_norm": 0.6268067955970764, + "learning_rate": 0.00011849611361884769, + "loss": 0.0619, + "step": 18809 + }, + { + "epoch": 0.4409360220819635, + "grad_norm": 0.42457282543182373, + "learning_rate": 0.000118488874507738, + "loss": 0.0845, + "step": 18810 + }, + { + "epoch": 0.44095946365676847, + "grad_norm": 0.637723445892334, + "learning_rate": 0.00011848163529630716, + "loss": 0.0988, + "step": 18811 + }, + { + "epoch": 0.44098290523157346, + "grad_norm": 0.41967156529426575, + "learning_rate": 0.00011847439598459447, + "loss": 0.0585, + "step": 18812 + }, + { + "epoch": 0.44100634680637846, + "grad_norm": 0.47637277841567993, + "learning_rate": 0.00011846715657263913, + "loss": 0.1112, + "step": 18813 + }, + { + "epoch": 0.44102978838118345, + "grad_norm": 0.5410817265510559, + "learning_rate": 0.00011845991706048054, + "loss": 0.1522, + "step": 18814 + }, + { + "epoch": 0.44105322995598845, + "grad_norm": 0.5324115753173828, + "learning_rate": 0.00011845267744815787, + "loss": 0.6842, + "step": 18815 + }, + { + "epoch": 0.44107667153079344, + "grad_norm": 0.2622149884700775, + "learning_rate": 0.00011844543773571047, + "loss": 0.0391, + "step": 18816 + }, + { + "epoch": 0.44110011310559843, + "grad_norm": 0.7622635960578918, + "learning_rate": 0.00011843819792317762, + "loss": 0.1814, + "step": 18817 + }, + { + "epoch": 0.44112355468040343, + "grad_norm": 0.40698158740997314, + "learning_rate": 0.0001184309580105986, + "loss": 0.0398, + "step": 18818 + }, + { + "epoch": 0.4411469962552084, + "grad_norm": 0.26359128952026367, + "learning_rate": 0.00011842371799801264, + "loss": 0.0621, + "step": 18819 + }, + { + "epoch": 0.4411704378300134, + "grad_norm": 0.4443817436695099, + "learning_rate": 0.00011841647788545911, + "loss": 0.0776, + "step": 18820 + }, + { + "epoch": 0.4411938794048184, + "grad_norm": 0.369758665561676, + "learning_rate": 0.00011840923767297723, + "loss": 0.5135, + "step": 18821 + }, + { + "epoch": 0.4412173209796234, + "grad_norm": 0.5637450218200684, + "learning_rate": 0.00011840199736060633, + "loss": 0.056, + "step": 18822 + }, + { + "epoch": 0.4412407625544284, + "grad_norm": 0.5384041666984558, + "learning_rate": 0.00011839475694838564, + "loss": 0.1041, + "step": 18823 + }, + { + "epoch": 0.4412642041292334, + "grad_norm": 0.24565215408802032, + "learning_rate": 0.00011838751643635453, + "loss": 0.035, + "step": 18824 + }, + { + "epoch": 0.4412876457040384, + "grad_norm": 0.181771919131279, + "learning_rate": 0.0001183802758245522, + "loss": 0.0255, + "step": 18825 + }, + { + "epoch": 0.4413110872788434, + "grad_norm": 0.7565917372703552, + "learning_rate": 0.00011837303511301799, + "loss": 0.1041, + "step": 18826 + }, + { + "epoch": 0.4413345288536484, + "grad_norm": 0.784144937992096, + "learning_rate": 0.00011836579430179115, + "loss": 0.1258, + "step": 18827 + }, + { + "epoch": 0.44135797042845337, + "grad_norm": 0.24834276735782623, + "learning_rate": 0.00011835855339091104, + "loss": 0.0525, + "step": 18828 + }, + { + "epoch": 0.44138141200325837, + "grad_norm": 0.24830235540866852, + "learning_rate": 0.00011835131238041685, + "loss": 0.0387, + "step": 18829 + }, + { + "epoch": 0.44140485357806336, + "grad_norm": 0.3259541690349579, + "learning_rate": 0.00011834407127034799, + "loss": 0.0741, + "step": 18830 + }, + { + "epoch": 0.44142829515286836, + "grad_norm": 0.3030549883842468, + "learning_rate": 0.00011833683006074365, + "loss": 0.0344, + "step": 18831 + }, + { + "epoch": 0.44145173672767335, + "grad_norm": 0.5055719614028931, + "learning_rate": 0.00011832958875164313, + "loss": 0.1101, + "step": 18832 + }, + { + "epoch": 0.44147517830247834, + "grad_norm": 0.3949491083621979, + "learning_rate": 0.00011832234734308578, + "loss": 0.0361, + "step": 18833 + }, + { + "epoch": 0.44149861987728334, + "grad_norm": 0.6571527123451233, + "learning_rate": 0.00011831510583511082, + "loss": 0.1249, + "step": 18834 + }, + { + "epoch": 0.44152206145208833, + "grad_norm": 0.5287962555885315, + "learning_rate": 0.00011830786422775761, + "loss": 0.1373, + "step": 18835 + }, + { + "epoch": 0.4415455030268933, + "grad_norm": 0.42554759979248047, + "learning_rate": 0.00011830062252106542, + "loss": 0.0969, + "step": 18836 + }, + { + "epoch": 0.4415689446016983, + "grad_norm": 0.19722995162010193, + "learning_rate": 0.00011829338071507355, + "loss": 0.0371, + "step": 18837 + }, + { + "epoch": 0.4415923861765033, + "grad_norm": 0.5529041886329651, + "learning_rate": 0.00011828613880982122, + "loss": 0.0643, + "step": 18838 + }, + { + "epoch": 0.4416158277513083, + "grad_norm": 0.48631879687309265, + "learning_rate": 0.00011827889680534785, + "loss": 0.0958, + "step": 18839 + }, + { + "epoch": 0.4416392693261133, + "grad_norm": 0.43774330615997314, + "learning_rate": 0.00011827165470169263, + "loss": 0.0953, + "step": 18840 + }, + { + "epoch": 0.4416627109009183, + "grad_norm": 0.8925555944442749, + "learning_rate": 0.00011826441249889492, + "loss": 0.7623, + "step": 18841 + }, + { + "epoch": 0.4416861524757233, + "grad_norm": 0.28410735726356506, + "learning_rate": 0.00011825717019699398, + "loss": 0.0496, + "step": 18842 + }, + { + "epoch": 0.4417095940505283, + "grad_norm": 0.5351518988609314, + "learning_rate": 0.00011824992779602914, + "loss": 0.0937, + "step": 18843 + }, + { + "epoch": 0.44173303562533334, + "grad_norm": 0.1912076324224472, + "learning_rate": 0.00011824268529603968, + "loss": 0.0324, + "step": 18844 + }, + { + "epoch": 0.44175647720013833, + "grad_norm": 0.3458230197429657, + "learning_rate": 0.00011823544269706487, + "loss": 0.0383, + "step": 18845 + }, + { + "epoch": 0.4417799187749433, + "grad_norm": 0.13350333273410797, + "learning_rate": 0.00011822819999914406, + "loss": 0.0218, + "step": 18846 + }, + { + "epoch": 0.4418033603497483, + "grad_norm": 0.4031766653060913, + "learning_rate": 0.00011822095720231652, + "loss": 0.0781, + "step": 18847 + }, + { + "epoch": 0.4418268019245533, + "grad_norm": 0.18749774992465973, + "learning_rate": 0.00011821371430662153, + "loss": 0.028, + "step": 18848 + }, + { + "epoch": 0.4418502434993583, + "grad_norm": 0.17916882038116455, + "learning_rate": 0.00011820647131209845, + "loss": 0.0211, + "step": 18849 + }, + { + "epoch": 0.4418736850741633, + "grad_norm": 0.8515681028366089, + "learning_rate": 0.00011819922821878655, + "loss": 0.1585, + "step": 18850 + }, + { + "epoch": 0.4418971266489683, + "grad_norm": 0.47625425457954407, + "learning_rate": 0.00011819198502672508, + "loss": 0.0897, + "step": 18851 + }, + { + "epoch": 0.4419205682237733, + "grad_norm": 0.731265664100647, + "learning_rate": 0.0001181847417359534, + "loss": 0.2135, + "step": 18852 + }, + { + "epoch": 0.4419440097985783, + "grad_norm": 0.50315260887146, + "learning_rate": 0.00011817749834651084, + "loss": 0.0888, + "step": 18853 + }, + { + "epoch": 0.4419674513733833, + "grad_norm": 0.4643884003162384, + "learning_rate": 0.0001181702548584366, + "loss": 0.056, + "step": 18854 + }, + { + "epoch": 0.4419908929481883, + "grad_norm": 0.5716490149497986, + "learning_rate": 0.0001181630112717701, + "loss": 0.1507, + "step": 18855 + }, + { + "epoch": 0.44201433452299327, + "grad_norm": 0.4701029062271118, + "learning_rate": 0.00011815576758655058, + "loss": 0.1007, + "step": 18856 + }, + { + "epoch": 0.44203777609779826, + "grad_norm": 0.3888786733150482, + "learning_rate": 0.00011814852380281733, + "loss": 0.1113, + "step": 18857 + }, + { + "epoch": 0.44206121767260326, + "grad_norm": 0.2688297927379608, + "learning_rate": 0.0001181412799206097, + "loss": 0.0534, + "step": 18858 + }, + { + "epoch": 0.44208465924740825, + "grad_norm": 0.3677898049354553, + "learning_rate": 0.00011813403593996696, + "loss": 0.0287, + "step": 18859 + }, + { + "epoch": 0.44210810082221325, + "grad_norm": 0.6184038519859314, + "learning_rate": 0.00011812679186092847, + "loss": 0.1439, + "step": 18860 + }, + { + "epoch": 0.44213154239701824, + "grad_norm": 0.638100266456604, + "learning_rate": 0.00011811954768353347, + "loss": 0.1248, + "step": 18861 + }, + { + "epoch": 0.44215498397182323, + "grad_norm": 0.42064768075942993, + "learning_rate": 0.00011811230340782128, + "loss": 0.0453, + "step": 18862 + }, + { + "epoch": 0.44217842554662823, + "grad_norm": 0.38547179102897644, + "learning_rate": 0.00011810505903383124, + "loss": 0.1047, + "step": 18863 + }, + { + "epoch": 0.4422018671214332, + "grad_norm": 0.2534622550010681, + "learning_rate": 0.00011809781456160262, + "loss": 0.0237, + "step": 18864 + }, + { + "epoch": 0.4422253086962382, + "grad_norm": 0.21562904119491577, + "learning_rate": 0.00011809056999117479, + "loss": 0.0464, + "step": 18865 + }, + { + "epoch": 0.4422487502710432, + "grad_norm": 0.9610432982444763, + "learning_rate": 0.00011808332532258699, + "loss": 0.2047, + "step": 18866 + }, + { + "epoch": 0.4422721918458482, + "grad_norm": 0.5378356575965881, + "learning_rate": 0.00011807608055587855, + "loss": 0.0951, + "step": 18867 + }, + { + "epoch": 0.4422956334206532, + "grad_norm": 0.7434307932853699, + "learning_rate": 0.00011806883569108882, + "loss": 0.0705, + "step": 18868 + }, + { + "epoch": 0.4423190749954582, + "grad_norm": 0.7585975527763367, + "learning_rate": 0.00011806159072825706, + "loss": 0.1479, + "step": 18869 + }, + { + "epoch": 0.4423425165702632, + "grad_norm": 0.15130209922790527, + "learning_rate": 0.00011805434566742257, + "loss": 0.0308, + "step": 18870 + }, + { + "epoch": 0.4423659581450682, + "grad_norm": 0.18953686952590942, + "learning_rate": 0.00011804710050862476, + "loss": 0.0345, + "step": 18871 + }, + { + "epoch": 0.4423893997198732, + "grad_norm": 0.27436843514442444, + "learning_rate": 0.00011803985525190284, + "loss": 0.0337, + "step": 18872 + }, + { + "epoch": 0.44241284129467817, + "grad_norm": 0.33246487379074097, + "learning_rate": 0.00011803260989729614, + "loss": 0.0951, + "step": 18873 + }, + { + "epoch": 0.44243628286948317, + "grad_norm": 0.481986939907074, + "learning_rate": 0.00011802536444484402, + "loss": 0.0965, + "step": 18874 + }, + { + "epoch": 0.44245972444428816, + "grad_norm": 0.32105550169944763, + "learning_rate": 0.00011801811889458578, + "loss": 0.0425, + "step": 18875 + }, + { + "epoch": 0.44248316601909315, + "grad_norm": 0.5560633540153503, + "learning_rate": 0.00011801087324656068, + "loss": 0.1045, + "step": 18876 + }, + { + "epoch": 0.44250660759389815, + "grad_norm": 0.10176104307174683, + "learning_rate": 0.00011800362750080807, + "loss": 0.0138, + "step": 18877 + }, + { + "epoch": 0.44253004916870314, + "grad_norm": 0.31695201992988586, + "learning_rate": 0.0001179963816573673, + "loss": 0.0674, + "step": 18878 + }, + { + "epoch": 0.44255349074350814, + "grad_norm": 0.45236214995384216, + "learning_rate": 0.00011798913571627768, + "loss": 0.0912, + "step": 18879 + }, + { + "epoch": 0.44257693231831313, + "grad_norm": 0.7550972104072571, + "learning_rate": 0.00011798188967757844, + "loss": 0.1768, + "step": 18880 + }, + { + "epoch": 0.4426003738931181, + "grad_norm": 0.628800094127655, + "learning_rate": 0.00011797464354130902, + "loss": 0.0999, + "step": 18881 + }, + { + "epoch": 0.4426238154679231, + "grad_norm": 0.23889446258544922, + "learning_rate": 0.00011796739730750863, + "loss": 0.0646, + "step": 18882 + }, + { + "epoch": 0.4426472570427281, + "grad_norm": 0.3101552724838257, + "learning_rate": 0.00011796015097621667, + "loss": 0.0494, + "step": 18883 + }, + { + "epoch": 0.4426706986175331, + "grad_norm": 0.6147527098655701, + "learning_rate": 0.0001179529045474724, + "loss": 0.0818, + "step": 18884 + }, + { + "epoch": 0.4426941401923381, + "grad_norm": 1.055607557296753, + "learning_rate": 0.00011794565802131521, + "loss": 0.2042, + "step": 18885 + }, + { + "epoch": 0.4427175817671431, + "grad_norm": 0.351171612739563, + "learning_rate": 0.00011793841139778432, + "loss": 0.0681, + "step": 18886 + }, + { + "epoch": 0.4427410233419481, + "grad_norm": 0.6291891932487488, + "learning_rate": 0.00011793116467691913, + "loss": 0.0904, + "step": 18887 + }, + { + "epoch": 0.4427644649167531, + "grad_norm": 0.6008216142654419, + "learning_rate": 0.00011792391785875893, + "loss": 0.1127, + "step": 18888 + }, + { + "epoch": 0.4427879064915581, + "grad_norm": 0.5426152944564819, + "learning_rate": 0.00011791667094334307, + "loss": 0.1128, + "step": 18889 + }, + { + "epoch": 0.4428113480663631, + "grad_norm": 0.18344323337078094, + "learning_rate": 0.00011790942393071083, + "loss": 0.0292, + "step": 18890 + }, + { + "epoch": 0.44283478964116807, + "grad_norm": 0.5137671232223511, + "learning_rate": 0.00011790217682090157, + "loss": 0.1108, + "step": 18891 + }, + { + "epoch": 0.44285823121597306, + "grad_norm": 0.22621434926986694, + "learning_rate": 0.00011789492961395456, + "loss": 0.0257, + "step": 18892 + }, + { + "epoch": 0.44288167279077806, + "grad_norm": 0.35932907462120056, + "learning_rate": 0.0001178876823099092, + "loss": 0.0817, + "step": 18893 + }, + { + "epoch": 0.44290511436558305, + "grad_norm": 0.6050002574920654, + "learning_rate": 0.00011788043490880473, + "loss": 0.1266, + "step": 18894 + }, + { + "epoch": 0.4429285559403881, + "grad_norm": 0.308356374502182, + "learning_rate": 0.00011787318741068056, + "loss": 0.0367, + "step": 18895 + }, + { + "epoch": 0.4429519975151931, + "grad_norm": 0.43486881256103516, + "learning_rate": 0.00011786593981557595, + "loss": 0.0445, + "step": 18896 + }, + { + "epoch": 0.4429754390899981, + "grad_norm": 0.5807939171791077, + "learning_rate": 0.00011785869212353026, + "loss": 0.0957, + "step": 18897 + }, + { + "epoch": 0.4429988806648031, + "grad_norm": 0.7290796041488647, + "learning_rate": 0.00011785144433458282, + "loss": 0.095, + "step": 18898 + }, + { + "epoch": 0.4430223222396081, + "grad_norm": 0.09192673861980438, + "learning_rate": 0.00011784419644877289, + "loss": 0.0114, + "step": 18899 + }, + { + "epoch": 0.4430457638144131, + "grad_norm": 0.5756540298461914, + "learning_rate": 0.00011783694846613986, + "loss": 0.1273, + "step": 18900 + }, + { + "epoch": 0.44306920538921807, + "grad_norm": 0.3725152313709259, + "learning_rate": 0.00011782970038672308, + "loss": 0.0513, + "step": 18901 + }, + { + "epoch": 0.44309264696402306, + "grad_norm": 0.11966247856616974, + "learning_rate": 0.00011782245221056183, + "loss": 0.0363, + "step": 18902 + }, + { + "epoch": 0.44311608853882806, + "grad_norm": 0.11261773109436035, + "learning_rate": 0.00011781520393769545, + "loss": 0.0095, + "step": 18903 + }, + { + "epoch": 0.44313953011363305, + "grad_norm": 0.47004467248916626, + "learning_rate": 0.00011780795556816328, + "loss": 0.0763, + "step": 18904 + }, + { + "epoch": 0.44316297168843805, + "grad_norm": 0.6354259848594666, + "learning_rate": 0.00011780070710200464, + "loss": 0.1087, + "step": 18905 + }, + { + "epoch": 0.44318641326324304, + "grad_norm": 0.5639665126800537, + "learning_rate": 0.00011779345853925885, + "loss": 0.0597, + "step": 18906 + }, + { + "epoch": 0.44320985483804803, + "grad_norm": 0.5289936661720276, + "learning_rate": 0.00011778620987996525, + "loss": 0.0503, + "step": 18907 + }, + { + "epoch": 0.44323329641285303, + "grad_norm": 0.6349722743034363, + "learning_rate": 0.00011777896112416318, + "loss": 0.1128, + "step": 18908 + }, + { + "epoch": 0.443256737987658, + "grad_norm": 0.7432114481925964, + "learning_rate": 0.00011777171227189199, + "loss": 0.0729, + "step": 18909 + }, + { + "epoch": 0.443280179562463, + "grad_norm": 0.2096230685710907, + "learning_rate": 0.00011776446332319098, + "loss": 0.0266, + "step": 18910 + }, + { + "epoch": 0.443303621137268, + "grad_norm": 0.11319633573293686, + "learning_rate": 0.00011775721427809947, + "loss": 0.0138, + "step": 18911 + }, + { + "epoch": 0.443327062712073, + "grad_norm": 0.2484886795282364, + "learning_rate": 0.00011774996513665683, + "loss": 0.0276, + "step": 18912 + }, + { + "epoch": 0.443350504286878, + "grad_norm": 0.4998299777507782, + "learning_rate": 0.00011774271589890237, + "loss": 0.1305, + "step": 18913 + }, + { + "epoch": 0.443373945861683, + "grad_norm": 0.5164887309074402, + "learning_rate": 0.00011773546656487546, + "loss": 0.1006, + "step": 18914 + }, + { + "epoch": 0.443397387436488, + "grad_norm": 0.38368088006973267, + "learning_rate": 0.00011772821713461539, + "loss": 0.0745, + "step": 18915 + }, + { + "epoch": 0.443420829011293, + "grad_norm": 0.191741481423378, + "learning_rate": 0.00011772096760816154, + "loss": 0.036, + "step": 18916 + }, + { + "epoch": 0.443444270586098, + "grad_norm": 0.5260637402534485, + "learning_rate": 0.00011771371798555319, + "loss": 0.0765, + "step": 18917 + }, + { + "epoch": 0.44346771216090297, + "grad_norm": 0.6082981824874878, + "learning_rate": 0.00011770646826682972, + "loss": 0.0533, + "step": 18918 + }, + { + "epoch": 0.44349115373570797, + "grad_norm": 0.4466967284679413, + "learning_rate": 0.00011769921845203044, + "loss": 0.0392, + "step": 18919 + }, + { + "epoch": 0.44351459531051296, + "grad_norm": 0.5669583082199097, + "learning_rate": 0.00011769196854119474, + "loss": 0.4207, + "step": 18920 + }, + { + "epoch": 0.44353803688531795, + "grad_norm": 0.7254787087440491, + "learning_rate": 0.0001176847185343619, + "loss": 0.1465, + "step": 18921 + }, + { + "epoch": 0.44356147846012295, + "grad_norm": 0.6406053900718689, + "learning_rate": 0.00011767746843157127, + "loss": 0.0615, + "step": 18922 + }, + { + "epoch": 0.44358492003492794, + "grad_norm": 0.6244761943817139, + "learning_rate": 0.00011767021823286222, + "loss": 0.1798, + "step": 18923 + }, + { + "epoch": 0.44360836160973294, + "grad_norm": 0.29978129267692566, + "learning_rate": 0.00011766296793827405, + "loss": 0.179, + "step": 18924 + }, + { + "epoch": 0.44363180318453793, + "grad_norm": 0.46710658073425293, + "learning_rate": 0.00011765571754784612, + "loss": 0.083, + "step": 18925 + }, + { + "epoch": 0.4436552447593429, + "grad_norm": 0.4996388852596283, + "learning_rate": 0.00011764846706161778, + "loss": 0.1055, + "step": 18926 + }, + { + "epoch": 0.4436786863341479, + "grad_norm": 0.6681941747665405, + "learning_rate": 0.00011764121647962835, + "loss": 0.0997, + "step": 18927 + }, + { + "epoch": 0.4437021279089529, + "grad_norm": 0.927168607711792, + "learning_rate": 0.00011763396580191722, + "loss": 0.1763, + "step": 18928 + }, + { + "epoch": 0.4437255694837579, + "grad_norm": 0.8069921135902405, + "learning_rate": 0.00011762671502852367, + "loss": 0.0598, + "step": 18929 + }, + { + "epoch": 0.4437490110585629, + "grad_norm": 0.1878271996974945, + "learning_rate": 0.00011761946415948706, + "loss": 0.0198, + "step": 18930 + }, + { + "epoch": 0.4437724526333679, + "grad_norm": 0.8084039688110352, + "learning_rate": 0.00011761221319484674, + "loss": 0.0742, + "step": 18931 + }, + { + "epoch": 0.4437958942081729, + "grad_norm": 0.6816291809082031, + "learning_rate": 0.00011760496213464207, + "loss": 0.1551, + "step": 18932 + }, + { + "epoch": 0.4438193357829779, + "grad_norm": 0.22848068177700043, + "learning_rate": 0.00011759771097891239, + "loss": 0.0326, + "step": 18933 + }, + { + "epoch": 0.4438427773577829, + "grad_norm": 0.67310631275177, + "learning_rate": 0.000117590459727697, + "loss": 0.1192, + "step": 18934 + }, + { + "epoch": 0.4438662189325879, + "grad_norm": 0.36073678731918335, + "learning_rate": 0.00011758320838103532, + "loss": 0.0708, + "step": 18935 + }, + { + "epoch": 0.44388966050739287, + "grad_norm": 0.6232055425643921, + "learning_rate": 0.00011757595693896666, + "loss": 0.4832, + "step": 18936 + }, + { + "epoch": 0.44391310208219786, + "grad_norm": 0.5385510921478271, + "learning_rate": 0.00011756870540153029, + "loss": 0.0741, + "step": 18937 + }, + { + "epoch": 0.44393654365700286, + "grad_norm": 0.7261967658996582, + "learning_rate": 0.00011756145376876571, + "loss": 0.0564, + "step": 18938 + }, + { + "epoch": 0.44395998523180785, + "grad_norm": 0.48236462473869324, + "learning_rate": 0.00011755420204071217, + "loss": 0.0704, + "step": 18939 + }, + { + "epoch": 0.44398342680661285, + "grad_norm": 0.34367263317108154, + "learning_rate": 0.000117546950217409, + "loss": 0.0537, + "step": 18940 + }, + { + "epoch": 0.44400686838141784, + "grad_norm": 0.6979508996009827, + "learning_rate": 0.00011753969829889564, + "loss": 0.1149, + "step": 18941 + }, + { + "epoch": 0.44403030995622284, + "grad_norm": 0.6676976680755615, + "learning_rate": 0.00011753244628521137, + "loss": 0.1132, + "step": 18942 + }, + { + "epoch": 0.44405375153102783, + "grad_norm": 0.45783504843711853, + "learning_rate": 0.00011752519417639549, + "loss": 0.049, + "step": 18943 + }, + { + "epoch": 0.4440771931058328, + "grad_norm": 0.20909681916236877, + "learning_rate": 0.00011751794197248749, + "loss": 0.0461, + "step": 18944 + }, + { + "epoch": 0.4441006346806378, + "grad_norm": 0.5018036365509033, + "learning_rate": 0.00011751068967352661, + "loss": 0.0839, + "step": 18945 + }, + { + "epoch": 0.44412407625544287, + "grad_norm": 0.8339835405349731, + "learning_rate": 0.00011750343727955224, + "loss": 0.1725, + "step": 18946 + }, + { + "epoch": 0.44414751783024786, + "grad_norm": 0.5560197830200195, + "learning_rate": 0.00011749618479060373, + "loss": 0.1253, + "step": 18947 + }, + { + "epoch": 0.44417095940505286, + "grad_norm": 0.1919477880001068, + "learning_rate": 0.00011748893220672044, + "loss": 0.0379, + "step": 18948 + }, + { + "epoch": 0.44419440097985785, + "grad_norm": 0.4007949233055115, + "learning_rate": 0.0001174816795279417, + "loss": 0.0467, + "step": 18949 + }, + { + "epoch": 0.44421784255466285, + "grad_norm": 0.1611466109752655, + "learning_rate": 0.00011747442675430687, + "loss": 0.0353, + "step": 18950 + }, + { + "epoch": 0.44424128412946784, + "grad_norm": 0.33404409885406494, + "learning_rate": 0.0001174671738858553, + "loss": 0.0338, + "step": 18951 + }, + { + "epoch": 0.44426472570427283, + "grad_norm": 0.2775132954120636, + "learning_rate": 0.0001174599209226264, + "loss": 0.0477, + "step": 18952 + }, + { + "epoch": 0.44428816727907783, + "grad_norm": 0.5616459250450134, + "learning_rate": 0.00011745266786465942, + "loss": 0.1191, + "step": 18953 + }, + { + "epoch": 0.4443116088538828, + "grad_norm": 0.4094189703464508, + "learning_rate": 0.00011744541471199382, + "loss": 0.0489, + "step": 18954 + }, + { + "epoch": 0.4443350504286878, + "grad_norm": 0.2920169532299042, + "learning_rate": 0.00011743816146466889, + "loss": 0.0531, + "step": 18955 + }, + { + "epoch": 0.4443584920034928, + "grad_norm": 0.5205414295196533, + "learning_rate": 0.000117430908122724, + "loss": 0.0869, + "step": 18956 + }, + { + "epoch": 0.4443819335782978, + "grad_norm": 0.23865431547164917, + "learning_rate": 0.00011742365468619851, + "loss": 0.025, + "step": 18957 + }, + { + "epoch": 0.4444053751531028, + "grad_norm": 0.4021987020969391, + "learning_rate": 0.00011741640115513179, + "loss": 0.0411, + "step": 18958 + }, + { + "epoch": 0.4444288167279078, + "grad_norm": 0.612495481967926, + "learning_rate": 0.00011740914752956318, + "loss": 0.0755, + "step": 18959 + }, + { + "epoch": 0.4444522583027128, + "grad_norm": 0.5353649854660034, + "learning_rate": 0.00011740189380953205, + "loss": 0.0887, + "step": 18960 + }, + { + "epoch": 0.4444756998775178, + "grad_norm": 0.18038839101791382, + "learning_rate": 0.00011739463999507776, + "loss": 0.034, + "step": 18961 + }, + { + "epoch": 0.4444991414523228, + "grad_norm": 0.8933013081550598, + "learning_rate": 0.00011738738608623965, + "loss": 0.7134, + "step": 18962 + }, + { + "epoch": 0.44452258302712777, + "grad_norm": 0.1214144304394722, + "learning_rate": 0.00011738013208305712, + "loss": 0.0337, + "step": 18963 + }, + { + "epoch": 0.44454602460193277, + "grad_norm": 0.5001159310340881, + "learning_rate": 0.00011737287798556949, + "loss": 0.1472, + "step": 18964 + }, + { + "epoch": 0.44456946617673776, + "grad_norm": 0.6043328642845154, + "learning_rate": 0.00011736562379381612, + "loss": 0.6365, + "step": 18965 + }, + { + "epoch": 0.44459290775154275, + "grad_norm": 0.7919276356697083, + "learning_rate": 0.00011735836950783643, + "loss": 0.2006, + "step": 18966 + }, + { + "epoch": 0.44461634932634775, + "grad_norm": 0.15206608176231384, + "learning_rate": 0.0001173511151276697, + "loss": 0.0191, + "step": 18967 + }, + { + "epoch": 0.44463979090115274, + "grad_norm": 0.47441908717155457, + "learning_rate": 0.00011734386065335534, + "loss": 0.0513, + "step": 18968 + }, + { + "epoch": 0.44466323247595774, + "grad_norm": 0.6920582056045532, + "learning_rate": 0.0001173366060849327, + "loss": 0.698, + "step": 18969 + }, + { + "epoch": 0.44468667405076273, + "grad_norm": 0.4686689078807831, + "learning_rate": 0.00011732935142244117, + "loss": 0.0633, + "step": 18970 + }, + { + "epoch": 0.4447101156255677, + "grad_norm": 1.0399556159973145, + "learning_rate": 0.00011732209666592009, + "loss": 0.2707, + "step": 18971 + }, + { + "epoch": 0.4447335572003727, + "grad_norm": 0.4266577661037445, + "learning_rate": 0.0001173148418154088, + "loss": 0.0574, + "step": 18972 + }, + { + "epoch": 0.4447569987751777, + "grad_norm": 0.567639946937561, + "learning_rate": 0.00011730758687094669, + "loss": 0.0794, + "step": 18973 + }, + { + "epoch": 0.4447804403499827, + "grad_norm": 1.5876729488372803, + "learning_rate": 0.00011730033183257314, + "loss": 0.1373, + "step": 18974 + }, + { + "epoch": 0.4448038819247877, + "grad_norm": 0.7054211497306824, + "learning_rate": 0.0001172930767003275, + "loss": 0.1337, + "step": 18975 + }, + { + "epoch": 0.4448273234995927, + "grad_norm": 0.7651920318603516, + "learning_rate": 0.00011728582147424915, + "loss": 0.1465, + "step": 18976 + }, + { + "epoch": 0.4448507650743977, + "grad_norm": 0.6273882985115051, + "learning_rate": 0.00011727856615437747, + "loss": 0.0842, + "step": 18977 + }, + { + "epoch": 0.4448742066492027, + "grad_norm": 0.22533835470676422, + "learning_rate": 0.00011727131074075176, + "loss": 0.0473, + "step": 18978 + }, + { + "epoch": 0.4448976482240077, + "grad_norm": 0.326172798871994, + "learning_rate": 0.00011726405523341141, + "loss": 0.0833, + "step": 18979 + }, + { + "epoch": 0.4449210897988127, + "grad_norm": 0.3000805675983429, + "learning_rate": 0.00011725679963239589, + "loss": 0.0912, + "step": 18980 + }, + { + "epoch": 0.44494453137361767, + "grad_norm": 0.5901049375534058, + "learning_rate": 0.00011724954393774443, + "loss": 0.0793, + "step": 18981 + }, + { + "epoch": 0.44496797294842266, + "grad_norm": 0.7278786301612854, + "learning_rate": 0.00011724228814949646, + "loss": 0.1807, + "step": 18982 + }, + { + "epoch": 0.44499141452322766, + "grad_norm": 0.5534162521362305, + "learning_rate": 0.00011723503226769139, + "loss": 0.7882, + "step": 18983 + }, + { + "epoch": 0.44501485609803265, + "grad_norm": 0.2648674547672272, + "learning_rate": 0.00011722777629236852, + "loss": 0.0323, + "step": 18984 + }, + { + "epoch": 0.44503829767283765, + "grad_norm": 0.3992499113082886, + "learning_rate": 0.00011722052022356726, + "loss": 0.101, + "step": 18985 + }, + { + "epoch": 0.44506173924764264, + "grad_norm": 0.3662668466567993, + "learning_rate": 0.00011721326406132698, + "loss": 0.0597, + "step": 18986 + }, + { + "epoch": 0.44508518082244763, + "grad_norm": 0.37755268812179565, + "learning_rate": 0.00011720600780568705, + "loss": 0.0959, + "step": 18987 + }, + { + "epoch": 0.44510862239725263, + "grad_norm": 0.5574079155921936, + "learning_rate": 0.00011719875145668679, + "loss": 0.0974, + "step": 18988 + }, + { + "epoch": 0.4451320639720576, + "grad_norm": 0.433860719203949, + "learning_rate": 0.00011719149501436568, + "loss": 0.064, + "step": 18989 + }, + { + "epoch": 0.4451555055468626, + "grad_norm": 1.0782865285873413, + "learning_rate": 0.00011718423847876303, + "loss": 0.1934, + "step": 18990 + }, + { + "epoch": 0.4451789471216676, + "grad_norm": 0.22377747297286987, + "learning_rate": 0.00011717698184991819, + "loss": 0.0419, + "step": 18991 + }, + { + "epoch": 0.4452023886964726, + "grad_norm": 0.7712418437004089, + "learning_rate": 0.00011716972512787054, + "loss": 0.1435, + "step": 18992 + }, + { + "epoch": 0.4452258302712776, + "grad_norm": 0.13601942360401154, + "learning_rate": 0.00011716246831265954, + "loss": 0.0231, + "step": 18993 + }, + { + "epoch": 0.4452492718460826, + "grad_norm": 0.5034643411636353, + "learning_rate": 0.00011715521140432446, + "loss": 0.1067, + "step": 18994 + }, + { + "epoch": 0.4452727134208876, + "grad_norm": 0.5892740488052368, + "learning_rate": 0.00011714795440290475, + "loss": 0.0725, + "step": 18995 + }, + { + "epoch": 0.4452961549956926, + "grad_norm": 0.62813800573349, + "learning_rate": 0.00011714069730843976, + "loss": 0.0263, + "step": 18996 + }, + { + "epoch": 0.44531959657049763, + "grad_norm": 0.47348055243492126, + "learning_rate": 0.00011713344012096884, + "loss": 0.1186, + "step": 18997 + }, + { + "epoch": 0.4453430381453026, + "grad_norm": 0.18118636310100555, + "learning_rate": 0.00011712618284053139, + "loss": 0.017, + "step": 18998 + }, + { + "epoch": 0.4453664797201076, + "grad_norm": 0.4646197259426117, + "learning_rate": 0.00011711892546716681, + "loss": 0.1329, + "step": 18999 + }, + { + "epoch": 0.4453899212949126, + "grad_norm": 0.2331189066171646, + "learning_rate": 0.00011711166800091441, + "loss": 0.0439, + "step": 19000 + }, + { + "epoch": 0.4454133628697176, + "grad_norm": 0.33937138319015503, + "learning_rate": 0.00011710441044181368, + "loss": 0.0409, + "step": 19001 + }, + { + "epoch": 0.4454368044445226, + "grad_norm": 0.5387174487113953, + "learning_rate": 0.00011709715278990395, + "loss": 0.0523, + "step": 19002 + }, + { + "epoch": 0.4454602460193276, + "grad_norm": 0.46712085604667664, + "learning_rate": 0.00011708989504522454, + "loss": 0.1438, + "step": 19003 + }, + { + "epoch": 0.4454836875941326, + "grad_norm": 0.5109812021255493, + "learning_rate": 0.00011708263720781488, + "loss": 0.5274, + "step": 19004 + }, + { + "epoch": 0.4455071291689376, + "grad_norm": 0.45513516664505005, + "learning_rate": 0.00011707537927771436, + "loss": 0.1042, + "step": 19005 + }, + { + "epoch": 0.4455305707437426, + "grad_norm": 0.4837704002857208, + "learning_rate": 0.00011706812125496238, + "loss": 0.1173, + "step": 19006 + }, + { + "epoch": 0.4455540123185476, + "grad_norm": 0.5169152617454529, + "learning_rate": 0.00011706086313959825, + "loss": 0.1365, + "step": 19007 + }, + { + "epoch": 0.44557745389335257, + "grad_norm": 0.43108120560646057, + "learning_rate": 0.00011705360493166142, + "loss": 0.0874, + "step": 19008 + }, + { + "epoch": 0.44560089546815757, + "grad_norm": 0.35394009947776794, + "learning_rate": 0.00011704634663119127, + "loss": 0.0505, + "step": 19009 + }, + { + "epoch": 0.44562433704296256, + "grad_norm": 0.13793091475963593, + "learning_rate": 0.00011703908823822711, + "loss": 0.0274, + "step": 19010 + }, + { + "epoch": 0.44564777861776755, + "grad_norm": 0.18611522018909454, + "learning_rate": 0.00011703182975280843, + "loss": 0.0415, + "step": 19011 + }, + { + "epoch": 0.44567122019257255, + "grad_norm": 0.859254002571106, + "learning_rate": 0.00011702457117497454, + "loss": 0.2673, + "step": 19012 + }, + { + "epoch": 0.44569466176737754, + "grad_norm": 0.45215165615081787, + "learning_rate": 0.00011701731250476484, + "loss": 0.0887, + "step": 19013 + }, + { + "epoch": 0.44571810334218254, + "grad_norm": 0.26706740260124207, + "learning_rate": 0.00011701005374221874, + "loss": 0.0535, + "step": 19014 + }, + { + "epoch": 0.44574154491698753, + "grad_norm": 0.1573217660188675, + "learning_rate": 0.00011700279488737562, + "loss": 0.0332, + "step": 19015 + }, + { + "epoch": 0.4457649864917925, + "grad_norm": 0.5813862681388855, + "learning_rate": 0.00011699553594027481, + "loss": 0.1109, + "step": 19016 + }, + { + "epoch": 0.4457884280665975, + "grad_norm": 0.12234500050544739, + "learning_rate": 0.00011698827690095578, + "loss": 0.0298, + "step": 19017 + }, + { + "epoch": 0.4458118696414025, + "grad_norm": 0.2803437113761902, + "learning_rate": 0.00011698101776945789, + "loss": 0.0781, + "step": 19018 + }, + { + "epoch": 0.4458353112162075, + "grad_norm": 0.3363938629627228, + "learning_rate": 0.00011697375854582048, + "loss": 0.079, + "step": 19019 + }, + { + "epoch": 0.4458587527910125, + "grad_norm": 0.5193291902542114, + "learning_rate": 0.00011696649923008303, + "loss": 0.0904, + "step": 19020 + }, + { + "epoch": 0.4458821943658175, + "grad_norm": 0.3860474228858948, + "learning_rate": 0.00011695923982228488, + "loss": 0.0889, + "step": 19021 + }, + { + "epoch": 0.4459056359406225, + "grad_norm": 0.5199116468429565, + "learning_rate": 0.00011695198032246537, + "loss": 0.1055, + "step": 19022 + }, + { + "epoch": 0.4459290775154275, + "grad_norm": 0.6604631543159485, + "learning_rate": 0.00011694472073066394, + "loss": 0.7556, + "step": 19023 + }, + { + "epoch": 0.4459525190902325, + "grad_norm": 0.11754266172647476, + "learning_rate": 0.00011693746104692001, + "loss": 0.0197, + "step": 19024 + }, + { + "epoch": 0.4459759606650375, + "grad_norm": 0.2288566678762436, + "learning_rate": 0.00011693020127127294, + "loss": 0.0429, + "step": 19025 + }, + { + "epoch": 0.44599940223984247, + "grad_norm": 0.7134778499603271, + "learning_rate": 0.0001169229414037621, + "loss": 0.6993, + "step": 19026 + }, + { + "epoch": 0.44602284381464746, + "grad_norm": 0.5125005841255188, + "learning_rate": 0.00011691568144442693, + "loss": 0.1391, + "step": 19027 + }, + { + "epoch": 0.44604628538945246, + "grad_norm": 0.43396803736686707, + "learning_rate": 0.00011690842139330675, + "loss": 0.0625, + "step": 19028 + }, + { + "epoch": 0.44606972696425745, + "grad_norm": 0.5534859299659729, + "learning_rate": 0.00011690116125044104, + "loss": 0.4912, + "step": 19029 + }, + { + "epoch": 0.44609316853906245, + "grad_norm": 0.42412668466567993, + "learning_rate": 0.00011689390101586914, + "loss": 0.1083, + "step": 19030 + }, + { + "epoch": 0.44611661011386744, + "grad_norm": 0.2805858850479126, + "learning_rate": 0.00011688664068963049, + "loss": 0.0326, + "step": 19031 + }, + { + "epoch": 0.44614005168867243, + "grad_norm": 0.6200242638587952, + "learning_rate": 0.00011687938027176441, + "loss": 0.1231, + "step": 19032 + }, + { + "epoch": 0.44616349326347743, + "grad_norm": 0.607356607913971, + "learning_rate": 0.00011687211976231037, + "loss": 0.1565, + "step": 19033 + }, + { + "epoch": 0.4461869348382824, + "grad_norm": 0.4431946277618408, + "learning_rate": 0.00011686485916130769, + "loss": 0.0594, + "step": 19034 + }, + { + "epoch": 0.4462103764130874, + "grad_norm": 0.47545790672302246, + "learning_rate": 0.00011685759846879585, + "loss": 0.0924, + "step": 19035 + }, + { + "epoch": 0.4462338179878924, + "grad_norm": 0.25524649024009705, + "learning_rate": 0.00011685033768481419, + "loss": 0.0186, + "step": 19036 + }, + { + "epoch": 0.4462572595626974, + "grad_norm": 0.5890184640884399, + "learning_rate": 0.00011684307680940215, + "loss": 0.154, + "step": 19037 + }, + { + "epoch": 0.4462807011375024, + "grad_norm": 0.8317458033561707, + "learning_rate": 0.00011683581584259906, + "loss": 0.0869, + "step": 19038 + }, + { + "epoch": 0.4463041427123074, + "grad_norm": 0.5021626949310303, + "learning_rate": 0.00011682855478444439, + "loss": 0.1015, + "step": 19039 + }, + { + "epoch": 0.4463275842871124, + "grad_norm": 0.507881760597229, + "learning_rate": 0.0001168212936349775, + "loss": 0.1165, + "step": 19040 + }, + { + "epoch": 0.4463510258619174, + "grad_norm": 0.43347474932670593, + "learning_rate": 0.0001168140323942378, + "loss": 0.0993, + "step": 19041 + }, + { + "epoch": 0.4463744674367224, + "grad_norm": 0.4553903639316559, + "learning_rate": 0.00011680677106226468, + "loss": 0.0598, + "step": 19042 + }, + { + "epoch": 0.44639790901152737, + "grad_norm": 0.41858547925949097, + "learning_rate": 0.00011679950963909756, + "loss": 0.0318, + "step": 19043 + }, + { + "epoch": 0.44642135058633237, + "grad_norm": 0.6179815530776978, + "learning_rate": 0.00011679224812477583, + "loss": 0.1084, + "step": 19044 + }, + { + "epoch": 0.44644479216113736, + "grad_norm": 0.6196333765983582, + "learning_rate": 0.00011678498651933887, + "loss": 0.1054, + "step": 19045 + }, + { + "epoch": 0.44646823373594235, + "grad_norm": 1.1336703300476074, + "learning_rate": 0.00011677772482282609, + "loss": 0.1688, + "step": 19046 + }, + { + "epoch": 0.44649167531074735, + "grad_norm": 0.42738866806030273, + "learning_rate": 0.00011677046303527695, + "loss": 0.0653, + "step": 19047 + }, + { + "epoch": 0.4465151168855524, + "grad_norm": 0.2504068911075592, + "learning_rate": 0.00011676320115673076, + "loss": 0.0556, + "step": 19048 + }, + { + "epoch": 0.4465385584603574, + "grad_norm": 0.545324981212616, + "learning_rate": 0.000116755939187227, + "loss": 0.7278, + "step": 19049 + }, + { + "epoch": 0.4465620000351624, + "grad_norm": 0.47028687596321106, + "learning_rate": 0.00011674867712680504, + "loss": 0.0991, + "step": 19050 + }, + { + "epoch": 0.4465854416099674, + "grad_norm": 0.09426885098218918, + "learning_rate": 0.00011674141497550426, + "loss": 0.0149, + "step": 19051 + }, + { + "epoch": 0.4466088831847724, + "grad_norm": 0.5399019122123718, + "learning_rate": 0.0001167341527333641, + "loss": 0.1182, + "step": 19052 + }, + { + "epoch": 0.44663232475957737, + "grad_norm": 0.4044003188610077, + "learning_rate": 0.000116726890400424, + "loss": 0.0716, + "step": 19053 + }, + { + "epoch": 0.44665576633438236, + "grad_norm": 0.5588838458061218, + "learning_rate": 0.0001167196279767233, + "loss": 0.1334, + "step": 19054 + }, + { + "epoch": 0.44667920790918736, + "grad_norm": 0.21009555459022522, + "learning_rate": 0.0001167123654623014, + "loss": 0.0168, + "step": 19055 + }, + { + "epoch": 0.44670264948399235, + "grad_norm": 0.5107800364494324, + "learning_rate": 0.00011670510285719776, + "loss": 0.111, + "step": 19056 + }, + { + "epoch": 0.44672609105879735, + "grad_norm": 0.17324212193489075, + "learning_rate": 0.00011669784016145173, + "loss": 0.0193, + "step": 19057 + }, + { + "epoch": 0.44674953263360234, + "grad_norm": 0.9729820489883423, + "learning_rate": 0.00011669057737510279, + "loss": 0.2222, + "step": 19058 + }, + { + "epoch": 0.44677297420840734, + "grad_norm": 0.5580671429634094, + "learning_rate": 0.0001166833144981903, + "loss": 0.0994, + "step": 19059 + }, + { + "epoch": 0.44679641578321233, + "grad_norm": 0.38448506593704224, + "learning_rate": 0.00011667605153075367, + "loss": 0.0865, + "step": 19060 + }, + { + "epoch": 0.4468198573580173, + "grad_norm": 0.4001024663448334, + "learning_rate": 0.00011666878847283231, + "loss": 0.0927, + "step": 19061 + }, + { + "epoch": 0.4468432989328223, + "grad_norm": 0.35018986463546753, + "learning_rate": 0.00011666152532446564, + "loss": 0.0522, + "step": 19062 + }, + { + "epoch": 0.4468667405076273, + "grad_norm": 0.6538384556770325, + "learning_rate": 0.00011665426208569307, + "loss": 0.1187, + "step": 19063 + }, + { + "epoch": 0.4468901820824323, + "grad_norm": 0.5337197184562683, + "learning_rate": 0.000116646998756554, + "loss": 0.0839, + "step": 19064 + }, + { + "epoch": 0.4469136236572373, + "grad_norm": 0.09585881978273392, + "learning_rate": 0.00011663973533708783, + "loss": 0.0312, + "step": 19065 + }, + { + "epoch": 0.4469370652320423, + "grad_norm": 1.237483263015747, + "learning_rate": 0.00011663247182733403, + "loss": 0.3019, + "step": 19066 + }, + { + "epoch": 0.4469605068068473, + "grad_norm": 0.4673166871070862, + "learning_rate": 0.00011662520822733195, + "loss": 0.0716, + "step": 19067 + }, + { + "epoch": 0.4469839483816523, + "grad_norm": 0.4275135397911072, + "learning_rate": 0.000116617944537121, + "loss": 0.0581, + "step": 19068 + }, + { + "epoch": 0.4470073899564573, + "grad_norm": 0.5044841766357422, + "learning_rate": 0.00011661068075674065, + "loss": 0.1218, + "step": 19069 + }, + { + "epoch": 0.4470308315312623, + "grad_norm": 0.5960826873779297, + "learning_rate": 0.00011660341688623027, + "loss": 0.4035, + "step": 19070 + }, + { + "epoch": 0.44705427310606727, + "grad_norm": 0.5861292481422424, + "learning_rate": 0.00011659615292562924, + "loss": 0.1335, + "step": 19071 + }, + { + "epoch": 0.44707771468087226, + "grad_norm": 0.5868262052536011, + "learning_rate": 0.00011658888887497707, + "loss": 0.1465, + "step": 19072 + }, + { + "epoch": 0.44710115625567726, + "grad_norm": 0.5389832854270935, + "learning_rate": 0.00011658162473431314, + "loss": 0.1383, + "step": 19073 + }, + { + "epoch": 0.44712459783048225, + "grad_norm": 0.43120869994163513, + "learning_rate": 0.00011657436050367678, + "loss": 0.0873, + "step": 19074 + }, + { + "epoch": 0.44714803940528725, + "grad_norm": 0.22385090589523315, + "learning_rate": 0.00011656709618310752, + "loss": 0.0601, + "step": 19075 + }, + { + "epoch": 0.44717148098009224, + "grad_norm": 0.6967511177062988, + "learning_rate": 0.00011655983177264472, + "loss": 0.0969, + "step": 19076 + }, + { + "epoch": 0.44719492255489723, + "grad_norm": 0.2616565525531769, + "learning_rate": 0.00011655256727232779, + "loss": 0.0573, + "step": 19077 + }, + { + "epoch": 0.44721836412970223, + "grad_norm": 0.15445251762866974, + "learning_rate": 0.00011654530268219619, + "loss": 0.0341, + "step": 19078 + }, + { + "epoch": 0.4472418057045072, + "grad_norm": 0.47242745757102966, + "learning_rate": 0.0001165380380022893, + "loss": 0.103, + "step": 19079 + }, + { + "epoch": 0.4472652472793122, + "grad_norm": 0.6749503016471863, + "learning_rate": 0.00011653077323264656, + "loss": 0.0903, + "step": 19080 + }, + { + "epoch": 0.4472886888541172, + "grad_norm": 0.16084057092666626, + "learning_rate": 0.0001165235083733074, + "loss": 0.016, + "step": 19081 + }, + { + "epoch": 0.4473121304289222, + "grad_norm": 0.18932035565376282, + "learning_rate": 0.00011651624342431118, + "loss": 0.0504, + "step": 19082 + }, + { + "epoch": 0.4473355720037272, + "grad_norm": 0.48780015110969543, + "learning_rate": 0.00011650897838569735, + "loss": 0.069, + "step": 19083 + }, + { + "epoch": 0.4473590135785322, + "grad_norm": 0.5077248215675354, + "learning_rate": 0.00011650171325750536, + "loss": 0.0402, + "step": 19084 + }, + { + "epoch": 0.4473824551533372, + "grad_norm": 0.9493318200111389, + "learning_rate": 0.00011649444803977463, + "loss": 0.1514, + "step": 19085 + }, + { + "epoch": 0.4474058967281422, + "grad_norm": 0.6660122871398926, + "learning_rate": 0.00011648718273254453, + "loss": 0.1416, + "step": 19086 + }, + { + "epoch": 0.4474293383029472, + "grad_norm": 1.007999062538147, + "learning_rate": 0.00011647991733585453, + "loss": 0.2621, + "step": 19087 + }, + { + "epoch": 0.44745277987775217, + "grad_norm": 0.5278313159942627, + "learning_rate": 0.00011647265184974406, + "loss": 0.1178, + "step": 19088 + }, + { + "epoch": 0.44747622145255717, + "grad_norm": 0.40175026655197144, + "learning_rate": 0.00011646538627425246, + "loss": 0.1044, + "step": 19089 + }, + { + "epoch": 0.44749966302736216, + "grad_norm": 0.41502541303634644, + "learning_rate": 0.00011645812060941924, + "loss": 0.1091, + "step": 19090 + }, + { + "epoch": 0.44752310460216715, + "grad_norm": 0.26385223865509033, + "learning_rate": 0.0001164508548552838, + "loss": 0.0499, + "step": 19091 + }, + { + "epoch": 0.44754654617697215, + "grad_norm": 0.4997352659702301, + "learning_rate": 0.00011644358901188554, + "loss": 0.0928, + "step": 19092 + }, + { + "epoch": 0.44756998775177714, + "grad_norm": 0.40674591064453125, + "learning_rate": 0.00011643632307926391, + "loss": 0.1235, + "step": 19093 + }, + { + "epoch": 0.44759342932658214, + "grad_norm": 0.2506468892097473, + "learning_rate": 0.00011642905705745835, + "loss": 0.0798, + "step": 19094 + }, + { + "epoch": 0.44761687090138713, + "grad_norm": 0.20055465400218964, + "learning_rate": 0.00011642179094650824, + "loss": 0.0185, + "step": 19095 + }, + { + "epoch": 0.4476403124761921, + "grad_norm": 0.733568012714386, + "learning_rate": 0.00011641452474645302, + "loss": 0.1257, + "step": 19096 + }, + { + "epoch": 0.4476637540509971, + "grad_norm": 0.33085522055625916, + "learning_rate": 0.00011640725845733215, + "loss": 0.0458, + "step": 19097 + }, + { + "epoch": 0.4476871956258021, + "grad_norm": 0.6207969784736633, + "learning_rate": 0.00011639999207918506, + "loss": 0.1414, + "step": 19098 + }, + { + "epoch": 0.44771063720060716, + "grad_norm": 0.12081552296876907, + "learning_rate": 0.00011639272561205112, + "loss": 0.0296, + "step": 19099 + }, + { + "epoch": 0.44773407877541216, + "grad_norm": 0.4619813561439514, + "learning_rate": 0.00011638545905596981, + "loss": 0.1387, + "step": 19100 + }, + { + "epoch": 0.44775752035021715, + "grad_norm": 0.4252200126647949, + "learning_rate": 0.00011637819241098049, + "loss": 0.0735, + "step": 19101 + }, + { + "epoch": 0.44778096192502215, + "grad_norm": 0.47818127274513245, + "learning_rate": 0.00011637092567712267, + "loss": 0.108, + "step": 19102 + }, + { + "epoch": 0.44780440349982714, + "grad_norm": 0.5551939606666565, + "learning_rate": 0.00011636365885443574, + "loss": 0.1384, + "step": 19103 + }, + { + "epoch": 0.44782784507463214, + "grad_norm": 0.3885766863822937, + "learning_rate": 0.00011635639194295915, + "loss": 0.0651, + "step": 19104 + }, + { + "epoch": 0.44785128664943713, + "grad_norm": 0.8608995079994202, + "learning_rate": 0.00011634912494273231, + "loss": 0.1899, + "step": 19105 + }, + { + "epoch": 0.4478747282242421, + "grad_norm": 0.6298999190330505, + "learning_rate": 0.00011634185785379468, + "loss": 0.133, + "step": 19106 + }, + { + "epoch": 0.4478981697990471, + "grad_norm": 0.44720137119293213, + "learning_rate": 0.00011633459067618564, + "loss": 0.0912, + "step": 19107 + }, + { + "epoch": 0.4479216113738521, + "grad_norm": 0.15515318512916565, + "learning_rate": 0.00011632732340994466, + "loss": 0.029, + "step": 19108 + }, + { + "epoch": 0.4479450529486571, + "grad_norm": 0.5245988965034485, + "learning_rate": 0.00011632005605511117, + "loss": 0.5223, + "step": 19109 + }, + { + "epoch": 0.4479684945234621, + "grad_norm": 0.4990474581718445, + "learning_rate": 0.00011631278861172458, + "loss": 0.108, + "step": 19110 + }, + { + "epoch": 0.4479919360982671, + "grad_norm": 0.2315836250782013, + "learning_rate": 0.00011630552107982435, + "loss": 0.0671, + "step": 19111 + }, + { + "epoch": 0.4480153776730721, + "grad_norm": 0.44710269570350647, + "learning_rate": 0.00011629825345944991, + "loss": 0.0841, + "step": 19112 + }, + { + "epoch": 0.4480388192478771, + "grad_norm": 0.19969934225082397, + "learning_rate": 0.00011629098575064067, + "loss": 0.0399, + "step": 19113 + }, + { + "epoch": 0.4480622608226821, + "grad_norm": 0.5156720876693726, + "learning_rate": 0.00011628371795343611, + "loss": 0.1003, + "step": 19114 + }, + { + "epoch": 0.4480857023974871, + "grad_norm": 0.8507466912269592, + "learning_rate": 0.00011627645006787562, + "loss": 0.193, + "step": 19115 + }, + { + "epoch": 0.44810914397229207, + "grad_norm": 0.8429913520812988, + "learning_rate": 0.00011626918209399866, + "loss": 0.0757, + "step": 19116 + }, + { + "epoch": 0.44813258554709706, + "grad_norm": 0.12944120168685913, + "learning_rate": 0.00011626191403184468, + "loss": 0.026, + "step": 19117 + }, + { + "epoch": 0.44815602712190206, + "grad_norm": 0.6474951505661011, + "learning_rate": 0.00011625464588145307, + "loss": 0.1589, + "step": 19118 + }, + { + "epoch": 0.44817946869670705, + "grad_norm": 0.45611876249313354, + "learning_rate": 0.00011624737764286328, + "loss": 0.106, + "step": 19119 + }, + { + "epoch": 0.44820291027151204, + "grad_norm": 0.5795412659645081, + "learning_rate": 0.00011624010931611479, + "loss": 0.0936, + "step": 19120 + }, + { + "epoch": 0.44822635184631704, + "grad_norm": 0.3294137418270111, + "learning_rate": 0.00011623284090124703, + "loss": 0.0527, + "step": 19121 + }, + { + "epoch": 0.44824979342112203, + "grad_norm": 0.27563080191612244, + "learning_rate": 0.00011622557239829939, + "loss": 0.0932, + "step": 19122 + }, + { + "epoch": 0.44827323499592703, + "grad_norm": 0.24909396469593048, + "learning_rate": 0.00011621830380731133, + "loss": 0.0322, + "step": 19123 + }, + { + "epoch": 0.448296676570732, + "grad_norm": 0.4556884467601776, + "learning_rate": 0.00011621103512832232, + "loss": 0.1007, + "step": 19124 + }, + { + "epoch": 0.448320118145537, + "grad_norm": 0.19676578044891357, + "learning_rate": 0.00011620376636137175, + "loss": 0.0342, + "step": 19125 + }, + { + "epoch": 0.448343559720342, + "grad_norm": 0.34882405400276184, + "learning_rate": 0.00011619649750649913, + "loss": 0.0624, + "step": 19126 + }, + { + "epoch": 0.448367001295147, + "grad_norm": 0.2588217854499817, + "learning_rate": 0.00011618922856374384, + "loss": 0.0697, + "step": 19127 + }, + { + "epoch": 0.448390442869952, + "grad_norm": 0.3474788963794708, + "learning_rate": 0.0001161819595331453, + "loss": 0.0559, + "step": 19128 + }, + { + "epoch": 0.448413884444757, + "grad_norm": 0.20391014218330383, + "learning_rate": 0.00011617469041474305, + "loss": 0.0378, + "step": 19129 + }, + { + "epoch": 0.448437326019562, + "grad_norm": 0.6647820472717285, + "learning_rate": 0.00011616742120857644, + "loss": 0.1591, + "step": 19130 + }, + { + "epoch": 0.448460767594367, + "grad_norm": 0.18717125058174133, + "learning_rate": 0.00011616015191468495, + "loss": 0.0488, + "step": 19131 + }, + { + "epoch": 0.448484209169172, + "grad_norm": 0.17565768957138062, + "learning_rate": 0.00011615288253310803, + "loss": 0.0151, + "step": 19132 + }, + { + "epoch": 0.44850765074397697, + "grad_norm": 0.6732317805290222, + "learning_rate": 0.00011614561306388513, + "loss": 0.1129, + "step": 19133 + }, + { + "epoch": 0.44853109231878197, + "grad_norm": 0.3285571336746216, + "learning_rate": 0.00011613834350705566, + "loss": 0.0911, + "step": 19134 + }, + { + "epoch": 0.44855453389358696, + "grad_norm": 0.5535503625869751, + "learning_rate": 0.00011613107386265912, + "loss": 0.7566, + "step": 19135 + }, + { + "epoch": 0.44857797546839195, + "grad_norm": 0.49505478143692017, + "learning_rate": 0.00011612380413073489, + "loss": 0.0881, + "step": 19136 + }, + { + "epoch": 0.44860141704319695, + "grad_norm": 0.2720015048980713, + "learning_rate": 0.00011611653431132244, + "loss": 0.0624, + "step": 19137 + }, + { + "epoch": 0.44862485861800194, + "grad_norm": 0.12748754024505615, + "learning_rate": 0.0001161092644044612, + "loss": 0.0216, + "step": 19138 + }, + { + "epoch": 0.44864830019280694, + "grad_norm": 0.8015116453170776, + "learning_rate": 0.00011610199441019067, + "loss": 0.1493, + "step": 19139 + }, + { + "epoch": 0.44867174176761193, + "grad_norm": 0.5269908905029297, + "learning_rate": 0.00011609472432855027, + "loss": 0.1057, + "step": 19140 + }, + { + "epoch": 0.4486951833424169, + "grad_norm": 0.12901195883750916, + "learning_rate": 0.00011608745415957944, + "loss": 0.0268, + "step": 19141 + }, + { + "epoch": 0.4487186249172219, + "grad_norm": 0.5843356251716614, + "learning_rate": 0.00011608018390331762, + "loss": 0.0845, + "step": 19142 + }, + { + "epoch": 0.4487420664920269, + "grad_norm": 0.6956465244293213, + "learning_rate": 0.00011607291355980427, + "loss": 0.4657, + "step": 19143 + }, + { + "epoch": 0.4487655080668319, + "grad_norm": 0.26738303899765015, + "learning_rate": 0.00011606564312907884, + "loss": 0.0695, + "step": 19144 + }, + { + "epoch": 0.4487889496416369, + "grad_norm": 0.11824965476989746, + "learning_rate": 0.0001160583726111808, + "loss": 0.0157, + "step": 19145 + }, + { + "epoch": 0.4488123912164419, + "grad_norm": 0.9558650255203247, + "learning_rate": 0.00011605110200614957, + "loss": 0.1442, + "step": 19146 + }, + { + "epoch": 0.4488358327912469, + "grad_norm": 0.4282510280609131, + "learning_rate": 0.00011604383131402459, + "loss": 0.0928, + "step": 19147 + }, + { + "epoch": 0.4488592743660519, + "grad_norm": 0.19700811803340912, + "learning_rate": 0.00011603656053484534, + "loss": 0.0294, + "step": 19148 + }, + { + "epoch": 0.4488827159408569, + "grad_norm": 0.40769028663635254, + "learning_rate": 0.00011602928966865127, + "loss": 0.0968, + "step": 19149 + }, + { + "epoch": 0.4489061575156619, + "grad_norm": 0.4302930533885956, + "learning_rate": 0.0001160220187154818, + "loss": 0.1085, + "step": 19150 + }, + { + "epoch": 0.4489295990904669, + "grad_norm": 0.5949162244796753, + "learning_rate": 0.00011601474767537644, + "loss": 0.092, + "step": 19151 + }, + { + "epoch": 0.4489530406652719, + "grad_norm": 0.36742812395095825, + "learning_rate": 0.0001160074765483746, + "loss": 0.0826, + "step": 19152 + }, + { + "epoch": 0.4489764822400769, + "grad_norm": 0.43820104002952576, + "learning_rate": 0.00011600020533451572, + "loss": 0.1018, + "step": 19153 + }, + { + "epoch": 0.4489999238148819, + "grad_norm": 0.4058329164981842, + "learning_rate": 0.00011599293403383931, + "loss": 0.1153, + "step": 19154 + }, + { + "epoch": 0.4490233653896869, + "grad_norm": 0.8547626733779907, + "learning_rate": 0.00011598566264638478, + "loss": 0.7102, + "step": 19155 + }, + { + "epoch": 0.4490468069644919, + "grad_norm": 0.1745922565460205, + "learning_rate": 0.00011597839117219154, + "loss": 0.0196, + "step": 19156 + }, + { + "epoch": 0.4490702485392969, + "grad_norm": 0.35652589797973633, + "learning_rate": 0.00011597111961129916, + "loss": 0.0922, + "step": 19157 + }, + { + "epoch": 0.4490936901141019, + "grad_norm": 0.3794170618057251, + "learning_rate": 0.00011596384796374702, + "loss": 0.0798, + "step": 19158 + }, + { + "epoch": 0.4491171316889069, + "grad_norm": 0.19436822831630707, + "learning_rate": 0.0001159565762295746, + "loss": 0.017, + "step": 19159 + }, + { + "epoch": 0.4491405732637119, + "grad_norm": 0.6350924372673035, + "learning_rate": 0.00011594930440882135, + "loss": 0.1176, + "step": 19160 + }, + { + "epoch": 0.44916401483851687, + "grad_norm": 0.7160879969596863, + "learning_rate": 0.00011594203250152671, + "loss": 0.2107, + "step": 19161 + }, + { + "epoch": 0.44918745641332186, + "grad_norm": 0.10925374180078506, + "learning_rate": 0.00011593476050773016, + "loss": 0.0188, + "step": 19162 + }, + { + "epoch": 0.44921089798812686, + "grad_norm": 0.2748194932937622, + "learning_rate": 0.00011592748842747115, + "loss": 0.056, + "step": 19163 + }, + { + "epoch": 0.44923433956293185, + "grad_norm": 0.6353046894073486, + "learning_rate": 0.00011592021626078914, + "loss": 0.0986, + "step": 19164 + }, + { + "epoch": 0.44925778113773684, + "grad_norm": 0.2604065537452698, + "learning_rate": 0.0001159129440077236, + "loss": 0.0414, + "step": 19165 + }, + { + "epoch": 0.44928122271254184, + "grad_norm": 0.33348405361175537, + "learning_rate": 0.00011590567166831397, + "loss": 0.328, + "step": 19166 + }, + { + "epoch": 0.44930466428734683, + "grad_norm": 0.6746055483818054, + "learning_rate": 0.00011589839924259974, + "loss": 0.1183, + "step": 19167 + }, + { + "epoch": 0.4493281058621518, + "grad_norm": 0.5774089097976685, + "learning_rate": 0.00011589112673062031, + "loss": 0.1329, + "step": 19168 + }, + { + "epoch": 0.4493515474369568, + "grad_norm": 0.3615409731864929, + "learning_rate": 0.00011588385413241519, + "loss": 0.0721, + "step": 19169 + }, + { + "epoch": 0.4493749890117618, + "grad_norm": 0.5086125731468201, + "learning_rate": 0.00011587658144802383, + "loss": 0.4733, + "step": 19170 + }, + { + "epoch": 0.4493984305865668, + "grad_norm": 0.6429566144943237, + "learning_rate": 0.00011586930867748571, + "loss": 0.1879, + "step": 19171 + }, + { + "epoch": 0.4494218721613718, + "grad_norm": 0.35422205924987793, + "learning_rate": 0.00011586203582084025, + "loss": 0.0858, + "step": 19172 + }, + { + "epoch": 0.4494453137361768, + "grad_norm": 0.10447216778993607, + "learning_rate": 0.00011585476287812698, + "loss": 0.0136, + "step": 19173 + }, + { + "epoch": 0.4494687553109818, + "grad_norm": 0.5498565435409546, + "learning_rate": 0.00011584748984938527, + "loss": 0.0954, + "step": 19174 + }, + { + "epoch": 0.4494921968857868, + "grad_norm": 0.39828649163246155, + "learning_rate": 0.00011584021673465467, + "loss": 0.0553, + "step": 19175 + }, + { + "epoch": 0.4495156384605918, + "grad_norm": 0.42195016145706177, + "learning_rate": 0.00011583294353397458, + "loss": 0.0528, + "step": 19176 + }, + { + "epoch": 0.4495390800353968, + "grad_norm": 0.7452142834663391, + "learning_rate": 0.00011582567024738453, + "loss": 0.1705, + "step": 19177 + }, + { + "epoch": 0.44956252161020177, + "grad_norm": 0.6604562401771545, + "learning_rate": 0.00011581839687492391, + "loss": 0.1704, + "step": 19178 + }, + { + "epoch": 0.44958596318500677, + "grad_norm": 0.564057469367981, + "learning_rate": 0.00011581112341663225, + "loss": 0.7784, + "step": 19179 + }, + { + "epoch": 0.44960940475981176, + "grad_norm": 0.5331504940986633, + "learning_rate": 0.00011580384987254898, + "loss": 0.074, + "step": 19180 + }, + { + "epoch": 0.44963284633461675, + "grad_norm": 0.41938379406929016, + "learning_rate": 0.00011579657624271358, + "loss": 0.0897, + "step": 19181 + }, + { + "epoch": 0.44965628790942175, + "grad_norm": 0.43013638257980347, + "learning_rate": 0.00011578930252716551, + "loss": 0.1261, + "step": 19182 + }, + { + "epoch": 0.44967972948422674, + "grad_norm": 0.17293986678123474, + "learning_rate": 0.00011578202872594424, + "loss": 0.0463, + "step": 19183 + }, + { + "epoch": 0.44970317105903174, + "grad_norm": 0.4385128915309906, + "learning_rate": 0.00011577475483908926, + "loss": 0.1029, + "step": 19184 + }, + { + "epoch": 0.44972661263383673, + "grad_norm": 0.532748281955719, + "learning_rate": 0.00011576748086663998, + "loss": 0.1223, + "step": 19185 + }, + { + "epoch": 0.4497500542086417, + "grad_norm": 0.500676691532135, + "learning_rate": 0.0001157602068086359, + "loss": 0.0975, + "step": 19186 + }, + { + "epoch": 0.4497734957834467, + "grad_norm": 0.5452912449836731, + "learning_rate": 0.00011575293266511654, + "loss": 0.6465, + "step": 19187 + }, + { + "epoch": 0.4497969373582517, + "grad_norm": 0.4949006736278534, + "learning_rate": 0.00011574565843612128, + "loss": 0.094, + "step": 19188 + }, + { + "epoch": 0.4498203789330567, + "grad_norm": 0.2675779461860657, + "learning_rate": 0.00011573838412168967, + "loss": 0.028, + "step": 19189 + }, + { + "epoch": 0.4498438205078617, + "grad_norm": 0.35104605555534363, + "learning_rate": 0.00011573110972186115, + "loss": 0.353, + "step": 19190 + }, + { + "epoch": 0.4498672620826667, + "grad_norm": 0.3835105895996094, + "learning_rate": 0.00011572383523667515, + "loss": 0.1205, + "step": 19191 + }, + { + "epoch": 0.4498907036574717, + "grad_norm": 0.42315179109573364, + "learning_rate": 0.00011571656066617118, + "loss": 0.1018, + "step": 19192 + }, + { + "epoch": 0.4499141452322767, + "grad_norm": 0.5163920521736145, + "learning_rate": 0.00011570928601038873, + "loss": 0.1144, + "step": 19193 + }, + { + "epoch": 0.4499375868070817, + "grad_norm": 0.821953296661377, + "learning_rate": 0.00011570201126936723, + "loss": 0.1256, + "step": 19194 + }, + { + "epoch": 0.4499610283818867, + "grad_norm": 0.7242940068244934, + "learning_rate": 0.00011569473644314622, + "loss": 0.7697, + "step": 19195 + }, + { + "epoch": 0.44998446995669167, + "grad_norm": 0.333646684885025, + "learning_rate": 0.00011568746153176511, + "loss": 0.0699, + "step": 19196 + }, + { + "epoch": 0.45000791153149666, + "grad_norm": 0.5618605017662048, + "learning_rate": 0.00011568018653526338, + "loss": 0.104, + "step": 19197 + }, + { + "epoch": 0.45003135310630166, + "grad_norm": 0.12733641266822815, + "learning_rate": 0.00011567291145368051, + "loss": 0.0442, + "step": 19198 + }, + { + "epoch": 0.45005479468110665, + "grad_norm": 0.44820401072502136, + "learning_rate": 0.000115665636287056, + "loss": 0.0812, + "step": 19199 + }, + { + "epoch": 0.45007823625591165, + "grad_norm": 0.32468730211257935, + "learning_rate": 0.0001156583610354293, + "loss": 0.0478, + "step": 19200 + }, + { + "epoch": 0.45010167783071664, + "grad_norm": 0.5329147577285767, + "learning_rate": 0.00011565108569883989, + "loss": 0.0541, + "step": 19201 + }, + { + "epoch": 0.4501251194055217, + "grad_norm": 0.6090916395187378, + "learning_rate": 0.00011564381027732727, + "loss": 0.6295, + "step": 19202 + }, + { + "epoch": 0.4501485609803267, + "grad_norm": 0.45048946142196655, + "learning_rate": 0.00011563653477093088, + "loss": 0.0826, + "step": 19203 + }, + { + "epoch": 0.4501720025551317, + "grad_norm": 1.4789214134216309, + "learning_rate": 0.00011562925917969019, + "loss": 0.2103, + "step": 19204 + }, + { + "epoch": 0.4501954441299367, + "grad_norm": 0.1784764677286148, + "learning_rate": 0.00011562198350364473, + "loss": 0.0431, + "step": 19205 + }, + { + "epoch": 0.45021888570474167, + "grad_norm": 0.3643466830253601, + "learning_rate": 0.00011561470774283394, + "loss": 0.0821, + "step": 19206 + }, + { + "epoch": 0.45024232727954666, + "grad_norm": 0.4737583100795746, + "learning_rate": 0.00011560743189729728, + "loss": 0.1195, + "step": 19207 + }, + { + "epoch": 0.45026576885435166, + "grad_norm": 0.6266655325889587, + "learning_rate": 0.0001156001559670743, + "loss": 0.7015, + "step": 19208 + }, + { + "epoch": 0.45028921042915665, + "grad_norm": 0.3351445496082306, + "learning_rate": 0.00011559287995220442, + "loss": 0.0357, + "step": 19209 + }, + { + "epoch": 0.45031265200396164, + "grad_norm": 0.34446147084236145, + "learning_rate": 0.00011558560385272712, + "loss": 0.0597, + "step": 19210 + }, + { + "epoch": 0.45033609357876664, + "grad_norm": 0.15225179493427277, + "learning_rate": 0.00011557832766868191, + "loss": 0.0314, + "step": 19211 + }, + { + "epoch": 0.45035953515357163, + "grad_norm": 0.18312424421310425, + "learning_rate": 0.00011557105140010827, + "loss": 0.0256, + "step": 19212 + }, + { + "epoch": 0.4503829767283766, + "grad_norm": 0.18076756596565247, + "learning_rate": 0.00011556377504704563, + "loss": 0.032, + "step": 19213 + }, + { + "epoch": 0.4504064183031816, + "grad_norm": 0.6746828556060791, + "learning_rate": 0.00011555649860953352, + "loss": 0.0771, + "step": 19214 + }, + { + "epoch": 0.4504298598779866, + "grad_norm": 0.6621037721633911, + "learning_rate": 0.00011554922208761146, + "loss": 0.0985, + "step": 19215 + }, + { + "epoch": 0.4504533014527916, + "grad_norm": 0.4214770197868347, + "learning_rate": 0.00011554194548131883, + "loss": 0.1123, + "step": 19216 + }, + { + "epoch": 0.4504767430275966, + "grad_norm": 0.24663637578487396, + "learning_rate": 0.00011553466879069514, + "loss": 0.0156, + "step": 19217 + }, + { + "epoch": 0.4505001846024016, + "grad_norm": 0.5133554339408875, + "learning_rate": 0.00011552739201577995, + "loss": 0.0868, + "step": 19218 + }, + { + "epoch": 0.4505236261772066, + "grad_norm": 0.5424937605857849, + "learning_rate": 0.00011552011515661271, + "loss": 0.1223, + "step": 19219 + }, + { + "epoch": 0.4505470677520116, + "grad_norm": 0.11609095335006714, + "learning_rate": 0.00011551283821323285, + "loss": 0.0203, + "step": 19220 + }, + { + "epoch": 0.4505705093268166, + "grad_norm": 0.16402465105056763, + "learning_rate": 0.00011550556118567992, + "loss": 0.0388, + "step": 19221 + }, + { + "epoch": 0.4505939509016216, + "grad_norm": 0.8175992965698242, + "learning_rate": 0.00011549828407399337, + "loss": 0.1486, + "step": 19222 + }, + { + "epoch": 0.45061739247642657, + "grad_norm": 0.22823229432106018, + "learning_rate": 0.00011549100687821266, + "loss": 0.0513, + "step": 19223 + }, + { + "epoch": 0.45064083405123156, + "grad_norm": 0.5684293508529663, + "learning_rate": 0.00011548372959837737, + "loss": 0.1013, + "step": 19224 + }, + { + "epoch": 0.45066427562603656, + "grad_norm": 0.6648968458175659, + "learning_rate": 0.0001154764522345269, + "loss": 0.1575, + "step": 19225 + }, + { + "epoch": 0.45068771720084155, + "grad_norm": 0.20743700861930847, + "learning_rate": 0.00011546917478670077, + "loss": 0.0268, + "step": 19226 + }, + { + "epoch": 0.45071115877564655, + "grad_norm": 0.3750842213630676, + "learning_rate": 0.00011546189725493845, + "loss": 0.102, + "step": 19227 + }, + { + "epoch": 0.45073460035045154, + "grad_norm": 0.3391234874725342, + "learning_rate": 0.00011545461963927946, + "loss": 0.0858, + "step": 19228 + }, + { + "epoch": 0.45075804192525654, + "grad_norm": 0.8023936152458191, + "learning_rate": 0.00011544734193976323, + "loss": 0.1349, + "step": 19229 + }, + { + "epoch": 0.45078148350006153, + "grad_norm": 0.5406041741371155, + "learning_rate": 0.00011544006415642932, + "loss": 0.0785, + "step": 19230 + }, + { + "epoch": 0.4508049250748665, + "grad_norm": 0.24713358283042908, + "learning_rate": 0.0001154327862893172, + "loss": 0.0565, + "step": 19231 + }, + { + "epoch": 0.4508283666496715, + "grad_norm": 0.635608971118927, + "learning_rate": 0.00011542550833846633, + "loss": 0.1296, + "step": 19232 + }, + { + "epoch": 0.4508518082244765, + "grad_norm": 0.5813199877738953, + "learning_rate": 0.00011541823030391621, + "loss": 0.1126, + "step": 19233 + }, + { + "epoch": 0.4508752497992815, + "grad_norm": 0.17300143837928772, + "learning_rate": 0.00011541095218570636, + "loss": 0.037, + "step": 19234 + }, + { + "epoch": 0.4508986913740865, + "grad_norm": 0.808551013469696, + "learning_rate": 0.00011540367398387624, + "loss": 0.0971, + "step": 19235 + }, + { + "epoch": 0.4509221329488915, + "grad_norm": 0.6298969388008118, + "learning_rate": 0.00011539639569846532, + "loss": 0.1757, + "step": 19236 + }, + { + "epoch": 0.4509455745236965, + "grad_norm": 0.5619145035743713, + "learning_rate": 0.00011538911732951319, + "loss": 0.6066, + "step": 19237 + }, + { + "epoch": 0.4509690160985015, + "grad_norm": 0.6139551997184753, + "learning_rate": 0.00011538183887705925, + "loss": 0.064, + "step": 19238 + }, + { + "epoch": 0.4509924576733065, + "grad_norm": 0.0642402395606041, + "learning_rate": 0.00011537456034114298, + "loss": 0.0064, + "step": 19239 + }, + { + "epoch": 0.4510158992481115, + "grad_norm": 0.6166842579841614, + "learning_rate": 0.00011536728172180396, + "loss": 0.1473, + "step": 19240 + }, + { + "epoch": 0.45103934082291647, + "grad_norm": 0.4730003774166107, + "learning_rate": 0.00011536000301908161, + "loss": 0.1155, + "step": 19241 + }, + { + "epoch": 0.45106278239772146, + "grad_norm": 0.5660020709037781, + "learning_rate": 0.00011535272423301547, + "loss": 0.0817, + "step": 19242 + }, + { + "epoch": 0.45108622397252646, + "grad_norm": 0.5456214547157288, + "learning_rate": 0.000115345445363645, + "loss": 0.1177, + "step": 19243 + }, + { + "epoch": 0.45110966554733145, + "grad_norm": 0.49437519907951355, + "learning_rate": 0.00011533816641100974, + "loss": 0.1049, + "step": 19244 + }, + { + "epoch": 0.45113310712213645, + "grad_norm": 0.44243893027305603, + "learning_rate": 0.00011533088737514912, + "loss": 0.0928, + "step": 19245 + }, + { + "epoch": 0.45115654869694144, + "grad_norm": 0.8173104524612427, + "learning_rate": 0.0001153236082561027, + "loss": 0.1148, + "step": 19246 + }, + { + "epoch": 0.45117999027174643, + "grad_norm": 0.17914533615112305, + "learning_rate": 0.00011531632905390994, + "loss": 0.0284, + "step": 19247 + }, + { + "epoch": 0.45120343184655143, + "grad_norm": 0.3290238082408905, + "learning_rate": 0.00011530904976861034, + "loss": 0.0932, + "step": 19248 + }, + { + "epoch": 0.4512268734213564, + "grad_norm": 0.5671761631965637, + "learning_rate": 0.00011530177040024344, + "loss": 0.084, + "step": 19249 + }, + { + "epoch": 0.4512503149961614, + "grad_norm": 0.2904781103134155, + "learning_rate": 0.00011529449094884867, + "loss": 0.0364, + "step": 19250 + }, + { + "epoch": 0.4512737565709664, + "grad_norm": 0.4963240623474121, + "learning_rate": 0.00011528721141446557, + "loss": 0.115, + "step": 19251 + }, + { + "epoch": 0.4512971981457714, + "grad_norm": 0.38977286219596863, + "learning_rate": 0.00011527993179713365, + "loss": 0.0693, + "step": 19252 + }, + { + "epoch": 0.45132063972057646, + "grad_norm": 0.25760677456855774, + "learning_rate": 0.00011527265209689236, + "loss": 0.0354, + "step": 19253 + }, + { + "epoch": 0.45134408129538145, + "grad_norm": 0.12246337532997131, + "learning_rate": 0.00011526537231378125, + "loss": 0.0368, + "step": 19254 + }, + { + "epoch": 0.45136752287018644, + "grad_norm": 0.5157880783081055, + "learning_rate": 0.00011525809244783979, + "loss": 0.1389, + "step": 19255 + }, + { + "epoch": 0.45139096444499144, + "grad_norm": 0.5765251517295837, + "learning_rate": 0.0001152508124991075, + "loss": 0.1628, + "step": 19256 + }, + { + "epoch": 0.45141440601979643, + "grad_norm": 0.21443164348602295, + "learning_rate": 0.00011524353246762389, + "loss": 0.0297, + "step": 19257 + }, + { + "epoch": 0.4514378475946014, + "grad_norm": 0.627859890460968, + "learning_rate": 0.0001152362523534284, + "loss": 0.0823, + "step": 19258 + }, + { + "epoch": 0.4514612891694064, + "grad_norm": 0.32150572538375854, + "learning_rate": 0.00011522897215656062, + "loss": 0.0701, + "step": 19259 + }, + { + "epoch": 0.4514847307442114, + "grad_norm": 0.4683074653148651, + "learning_rate": 0.00011522169187705998, + "loss": 0.0645, + "step": 19260 + }, + { + "epoch": 0.4515081723190164, + "grad_norm": 0.5073751211166382, + "learning_rate": 0.00011521441151496601, + "loss": 0.0695, + "step": 19261 + }, + { + "epoch": 0.4515316138938214, + "grad_norm": 0.39569124579429626, + "learning_rate": 0.00011520713107031824, + "loss": 0.0675, + "step": 19262 + }, + { + "epoch": 0.4515550554686264, + "grad_norm": 0.2849173843860626, + "learning_rate": 0.00011519985054315613, + "loss": 0.0477, + "step": 19263 + }, + { + "epoch": 0.4515784970434314, + "grad_norm": 0.22060735523700714, + "learning_rate": 0.0001151925699335192, + "loss": 0.0272, + "step": 19264 + }, + { + "epoch": 0.4516019386182364, + "grad_norm": 0.5514877438545227, + "learning_rate": 0.00011518528924144696, + "loss": 0.1699, + "step": 19265 + }, + { + "epoch": 0.4516253801930414, + "grad_norm": 0.5450724363327026, + "learning_rate": 0.00011517800846697892, + "loss": 0.8268, + "step": 19266 + }, + { + "epoch": 0.4516488217678464, + "grad_norm": 0.32585492730140686, + "learning_rate": 0.00011517072761015458, + "loss": 0.0679, + "step": 19267 + }, + { + "epoch": 0.45167226334265137, + "grad_norm": 0.4961458444595337, + "learning_rate": 0.00011516344667101346, + "loss": 0.0777, + "step": 19268 + }, + { + "epoch": 0.45169570491745636, + "grad_norm": 0.31750747561454773, + "learning_rate": 0.00011515616564959505, + "loss": 0.0643, + "step": 19269 + }, + { + "epoch": 0.45171914649226136, + "grad_norm": 0.5074719786643982, + "learning_rate": 0.00011514888454593884, + "loss": 0.097, + "step": 19270 + }, + { + "epoch": 0.45174258806706635, + "grad_norm": 0.5634347796440125, + "learning_rate": 0.00011514160336008435, + "loss": 0.0934, + "step": 19271 + }, + { + "epoch": 0.45176602964187135, + "grad_norm": 0.2132744938135147, + "learning_rate": 0.00011513432209207111, + "loss": 0.0307, + "step": 19272 + }, + { + "epoch": 0.45178947121667634, + "grad_norm": 0.7488279342651367, + "learning_rate": 0.00011512704074193863, + "loss": 0.1984, + "step": 19273 + }, + { + "epoch": 0.45181291279148134, + "grad_norm": 0.5770421028137207, + "learning_rate": 0.00011511975930972636, + "loss": 0.0857, + "step": 19274 + }, + { + "epoch": 0.45183635436628633, + "grad_norm": 0.5514312386512756, + "learning_rate": 0.0001151124777954739, + "loss": 0.1393, + "step": 19275 + }, + { + "epoch": 0.4518597959410913, + "grad_norm": 0.27135205268859863, + "learning_rate": 0.00011510519619922069, + "loss": 0.0299, + "step": 19276 + }, + { + "epoch": 0.4518832375158963, + "grad_norm": 0.4731314480304718, + "learning_rate": 0.00011509791452100622, + "loss": 0.0809, + "step": 19277 + }, + { + "epoch": 0.4519066790907013, + "grad_norm": 0.21256685256958008, + "learning_rate": 0.00011509063276087009, + "loss": 0.0431, + "step": 19278 + }, + { + "epoch": 0.4519301206655063, + "grad_norm": 0.2604510188102722, + "learning_rate": 0.00011508335091885174, + "loss": 0.0486, + "step": 19279 + }, + { + "epoch": 0.4519535622403113, + "grad_norm": 0.1034046858549118, + "learning_rate": 0.0001150760689949907, + "loss": 0.0213, + "step": 19280 + }, + { + "epoch": 0.4519770038151163, + "grad_norm": 0.5470448732376099, + "learning_rate": 0.00011506878698932651, + "loss": 0.5878, + "step": 19281 + }, + { + "epoch": 0.4520004453899213, + "grad_norm": 0.6220190525054932, + "learning_rate": 0.00011506150490189867, + "loss": 0.1659, + "step": 19282 + }, + { + "epoch": 0.4520238869647263, + "grad_norm": 0.4703480005264282, + "learning_rate": 0.00011505422273274665, + "loss": 0.1461, + "step": 19283 + }, + { + "epoch": 0.4520473285395313, + "grad_norm": 0.470039427280426, + "learning_rate": 0.00011504694048190998, + "loss": 0.1125, + "step": 19284 + }, + { + "epoch": 0.4520707701143363, + "grad_norm": 0.32982316613197327, + "learning_rate": 0.00011503965814942821, + "loss": 0.0629, + "step": 19285 + }, + { + "epoch": 0.45209421168914127, + "grad_norm": 0.11446043848991394, + "learning_rate": 0.00011503237573534082, + "loss": 0.023, + "step": 19286 + }, + { + "epoch": 0.45211765326394626, + "grad_norm": 0.8582016825675964, + "learning_rate": 0.00011502509323968735, + "loss": 0.2402, + "step": 19287 + }, + { + "epoch": 0.45214109483875126, + "grad_norm": 0.4680531919002533, + "learning_rate": 0.00011501781066250732, + "loss": 0.0996, + "step": 19288 + }, + { + "epoch": 0.45216453641355625, + "grad_norm": 0.5631410479545593, + "learning_rate": 0.0001150105280038402, + "loss": 0.115, + "step": 19289 + }, + { + "epoch": 0.45218797798836124, + "grad_norm": 0.2687508165836334, + "learning_rate": 0.00011500324526372552, + "loss": 0.0579, + "step": 19290 + }, + { + "epoch": 0.45221141956316624, + "grad_norm": 0.4243224859237671, + "learning_rate": 0.00011499596244220283, + "loss": 0.4394, + "step": 19291 + }, + { + "epoch": 0.45223486113797123, + "grad_norm": 0.1021881029009819, + "learning_rate": 0.00011498867953931165, + "loss": 0.0156, + "step": 19292 + }, + { + "epoch": 0.45225830271277623, + "grad_norm": 0.5199275016784668, + "learning_rate": 0.00011498139655509143, + "loss": 0.064, + "step": 19293 + }, + { + "epoch": 0.4522817442875812, + "grad_norm": 0.2006005197763443, + "learning_rate": 0.00011497411348958176, + "loss": 0.0239, + "step": 19294 + }, + { + "epoch": 0.4523051858623862, + "grad_norm": 0.07879212498664856, + "learning_rate": 0.00011496683034282214, + "loss": 0.0203, + "step": 19295 + }, + { + "epoch": 0.4523286274371912, + "grad_norm": 0.39447999000549316, + "learning_rate": 0.00011495954711485202, + "loss": 0.0653, + "step": 19296 + }, + { + "epoch": 0.4523520690119962, + "grad_norm": 0.5738615393638611, + "learning_rate": 0.000114952263805711, + "loss": 0.0711, + "step": 19297 + }, + { + "epoch": 0.4523755105868012, + "grad_norm": 0.5761364698410034, + "learning_rate": 0.00011494498041543862, + "loss": 0.1699, + "step": 19298 + }, + { + "epoch": 0.4523989521616062, + "grad_norm": 0.07857882231473923, + "learning_rate": 0.0001149376969440743, + "loss": 0.0156, + "step": 19299 + }, + { + "epoch": 0.4524223937364112, + "grad_norm": 0.31004518270492554, + "learning_rate": 0.00011493041339165767, + "loss": 0.0737, + "step": 19300 + }, + { + "epoch": 0.4524458353112162, + "grad_norm": 0.6100266575813293, + "learning_rate": 0.00011492312975822817, + "loss": 0.126, + "step": 19301 + }, + { + "epoch": 0.4524692768860212, + "grad_norm": 0.5357660055160522, + "learning_rate": 0.00011491584604382532, + "loss": 0.1265, + "step": 19302 + }, + { + "epoch": 0.45249271846082617, + "grad_norm": 0.372649222612381, + "learning_rate": 0.0001149085622484887, + "loss": 0.0662, + "step": 19303 + }, + { + "epoch": 0.4525161600356312, + "grad_norm": 0.3906787037849426, + "learning_rate": 0.0001149012783722578, + "loss": 0.0949, + "step": 19304 + }, + { + "epoch": 0.4525396016104362, + "grad_norm": 0.553884744644165, + "learning_rate": 0.00011489399441517213, + "loss": 0.1752, + "step": 19305 + }, + { + "epoch": 0.4525630431852412, + "grad_norm": 0.3356323540210724, + "learning_rate": 0.00011488671037727126, + "loss": 0.0308, + "step": 19306 + }, + { + "epoch": 0.4525864847600462, + "grad_norm": 0.233801007270813, + "learning_rate": 0.00011487942625859467, + "loss": 0.0394, + "step": 19307 + }, + { + "epoch": 0.4526099263348512, + "grad_norm": 0.14294689893722534, + "learning_rate": 0.00011487214205918186, + "loss": 0.0295, + "step": 19308 + }, + { + "epoch": 0.4526333679096562, + "grad_norm": 0.1848202645778656, + "learning_rate": 0.00011486485777907241, + "loss": 0.0212, + "step": 19309 + }, + { + "epoch": 0.4526568094844612, + "grad_norm": 0.3358393907546997, + "learning_rate": 0.00011485757341830585, + "loss": 0.0399, + "step": 19310 + }, + { + "epoch": 0.4526802510592662, + "grad_norm": 0.11116503924131393, + "learning_rate": 0.00011485028897692167, + "loss": 0.0157, + "step": 19311 + }, + { + "epoch": 0.4527036926340712, + "grad_norm": 0.34907546639442444, + "learning_rate": 0.00011484300445495936, + "loss": 0.0686, + "step": 19312 + }, + { + "epoch": 0.45272713420887617, + "grad_norm": 0.5830644369125366, + "learning_rate": 0.00011483571985245856, + "loss": 0.1164, + "step": 19313 + }, + { + "epoch": 0.45275057578368116, + "grad_norm": 0.9249502420425415, + "learning_rate": 0.00011482843516945867, + "loss": 0.2264, + "step": 19314 + }, + { + "epoch": 0.45277401735848616, + "grad_norm": 0.6882204413414001, + "learning_rate": 0.00011482115040599928, + "loss": 0.168, + "step": 19315 + }, + { + "epoch": 0.45279745893329115, + "grad_norm": 0.24059359729290009, + "learning_rate": 0.00011481386556211993, + "loss": 0.0324, + "step": 19316 + }, + { + "epoch": 0.45282090050809615, + "grad_norm": 0.5134545564651489, + "learning_rate": 0.00011480658063786014, + "loss": 0.0797, + "step": 19317 + }, + { + "epoch": 0.45284434208290114, + "grad_norm": 0.21411681175231934, + "learning_rate": 0.00011479929563325939, + "loss": 0.0683, + "step": 19318 + }, + { + "epoch": 0.45286778365770614, + "grad_norm": 0.5374053120613098, + "learning_rate": 0.00011479201054835728, + "loss": 0.1069, + "step": 19319 + }, + { + "epoch": 0.45289122523251113, + "grad_norm": 0.15229465067386627, + "learning_rate": 0.0001147847253831933, + "loss": 0.0287, + "step": 19320 + }, + { + "epoch": 0.4529146668073161, + "grad_norm": 0.5342978835105896, + "learning_rate": 0.00011477744013780695, + "loss": 0.076, + "step": 19321 + }, + { + "epoch": 0.4529381083821211, + "grad_norm": 0.5051354169845581, + "learning_rate": 0.00011477015481223783, + "loss": 0.0853, + "step": 19322 + }, + { + "epoch": 0.4529615499569261, + "grad_norm": 0.5096838474273682, + "learning_rate": 0.00011476286940652543, + "loss": 0.096, + "step": 19323 + }, + { + "epoch": 0.4529849915317311, + "grad_norm": 0.17577293515205383, + "learning_rate": 0.00011475558392070927, + "loss": 0.0354, + "step": 19324 + }, + { + "epoch": 0.4530084331065361, + "grad_norm": 0.2756168246269226, + "learning_rate": 0.00011474829835482893, + "loss": 0.0434, + "step": 19325 + }, + { + "epoch": 0.4530318746813411, + "grad_norm": 0.14613814651966095, + "learning_rate": 0.00011474101270892386, + "loss": 0.019, + "step": 19326 + }, + { + "epoch": 0.4530553162561461, + "grad_norm": 0.34556686878204346, + "learning_rate": 0.00011473372698303368, + "loss": 0.0555, + "step": 19327 + }, + { + "epoch": 0.4530787578309511, + "grad_norm": 0.44546443223953247, + "learning_rate": 0.00011472644117719787, + "loss": 0.4038, + "step": 19328 + }, + { + "epoch": 0.4531021994057561, + "grad_norm": 0.61228346824646, + "learning_rate": 0.00011471915529145599, + "loss": 0.0996, + "step": 19329 + }, + { + "epoch": 0.4531256409805611, + "grad_norm": 0.2723667323589325, + "learning_rate": 0.00011471186932584755, + "loss": 0.0548, + "step": 19330 + }, + { + "epoch": 0.45314908255536607, + "grad_norm": 0.19454586505889893, + "learning_rate": 0.00011470458328041207, + "loss": 0.0478, + "step": 19331 + }, + { + "epoch": 0.45317252413017106, + "grad_norm": 0.47742924094200134, + "learning_rate": 0.0001146972971551891, + "loss": 0.0948, + "step": 19332 + }, + { + "epoch": 0.45319596570497606, + "grad_norm": 0.24301601946353912, + "learning_rate": 0.00011469001095021823, + "loss": 0.0607, + "step": 19333 + }, + { + "epoch": 0.45321940727978105, + "grad_norm": 0.41418445110321045, + "learning_rate": 0.0001146827246655389, + "loss": 0.1011, + "step": 19334 + }, + { + "epoch": 0.45324284885458604, + "grad_norm": 1.3202868700027466, + "learning_rate": 0.00011467543830119075, + "loss": 0.0917, + "step": 19335 + }, + { + "epoch": 0.45326629042939104, + "grad_norm": 0.4464036226272583, + "learning_rate": 0.00011466815185721323, + "loss": 0.1069, + "step": 19336 + }, + { + "epoch": 0.45328973200419603, + "grad_norm": 0.48687705397605896, + "learning_rate": 0.00011466086533364589, + "loss": 0.1499, + "step": 19337 + }, + { + "epoch": 0.453313173579001, + "grad_norm": 0.4243782162666321, + "learning_rate": 0.00011465357873052829, + "loss": 0.1093, + "step": 19338 + }, + { + "epoch": 0.453336615153806, + "grad_norm": 0.7265857458114624, + "learning_rate": 0.00011464629204789996, + "loss": 0.116, + "step": 19339 + }, + { + "epoch": 0.453360056728611, + "grad_norm": 0.4591706097126007, + "learning_rate": 0.00011463900528580047, + "loss": 0.1149, + "step": 19340 + }, + { + "epoch": 0.453383498303416, + "grad_norm": 0.5900180339813232, + "learning_rate": 0.00011463171844426928, + "loss": 0.1344, + "step": 19341 + }, + { + "epoch": 0.453406939878221, + "grad_norm": 0.4665236473083496, + "learning_rate": 0.000114624431523346, + "loss": 0.149, + "step": 19342 + }, + { + "epoch": 0.453430381453026, + "grad_norm": 0.43571561574935913, + "learning_rate": 0.00011461714452307012, + "loss": 0.0709, + "step": 19343 + }, + { + "epoch": 0.453453823027831, + "grad_norm": 0.2947794795036316, + "learning_rate": 0.00011460985744348121, + "loss": 0.0471, + "step": 19344 + }, + { + "epoch": 0.453477264602636, + "grad_norm": 0.5229131579399109, + "learning_rate": 0.00011460257028461883, + "loss": 0.6133, + "step": 19345 + }, + { + "epoch": 0.453500706177441, + "grad_norm": 0.15374435484409332, + "learning_rate": 0.0001145952830465225, + "loss": 0.0477, + "step": 19346 + }, + { + "epoch": 0.453524147752246, + "grad_norm": 0.08983080834150314, + "learning_rate": 0.00011458799572923171, + "loss": 0.0108, + "step": 19347 + }, + { + "epoch": 0.45354758932705097, + "grad_norm": 0.13538511097431183, + "learning_rate": 0.00011458070833278607, + "loss": 0.0388, + "step": 19348 + }, + { + "epoch": 0.45357103090185596, + "grad_norm": 0.27558091282844543, + "learning_rate": 0.00011457342085722509, + "loss": 0.0177, + "step": 19349 + }, + { + "epoch": 0.45359447247666096, + "grad_norm": 0.6720229983329773, + "learning_rate": 0.0001145661333025883, + "loss": 0.1243, + "step": 19350 + }, + { + "epoch": 0.45361791405146595, + "grad_norm": 0.29466646909713745, + "learning_rate": 0.00011455884566891528, + "loss": 0.0449, + "step": 19351 + }, + { + "epoch": 0.45364135562627095, + "grad_norm": 0.46037667989730835, + "learning_rate": 0.00011455155795624557, + "loss": 0.1301, + "step": 19352 + }, + { + "epoch": 0.45366479720107594, + "grad_norm": 0.23828543722629547, + "learning_rate": 0.00011454427016461866, + "loss": 0.0414, + "step": 19353 + }, + { + "epoch": 0.45368823877588094, + "grad_norm": 0.22785411775112152, + "learning_rate": 0.00011453698229407417, + "loss": 0.0456, + "step": 19354 + }, + { + "epoch": 0.453711680350686, + "grad_norm": 0.744719386100769, + "learning_rate": 0.0001145296943446516, + "loss": 0.1703, + "step": 19355 + }, + { + "epoch": 0.453735121925491, + "grad_norm": 0.4146398901939392, + "learning_rate": 0.00011452240631639048, + "loss": 0.0993, + "step": 19356 + }, + { + "epoch": 0.453758563500296, + "grad_norm": 0.35647088289260864, + "learning_rate": 0.00011451511820933037, + "loss": 0.0507, + "step": 19357 + }, + { + "epoch": 0.45378200507510097, + "grad_norm": 0.8496389389038086, + "learning_rate": 0.00011450783002351084, + "loss": 0.3262, + "step": 19358 + }, + { + "epoch": 0.45380544664990596, + "grad_norm": 0.5228109955787659, + "learning_rate": 0.00011450054175897143, + "loss": 0.0854, + "step": 19359 + }, + { + "epoch": 0.45382888822471096, + "grad_norm": 0.47761270403862, + "learning_rate": 0.00011449325341575163, + "loss": 0.0768, + "step": 19360 + }, + { + "epoch": 0.45385232979951595, + "grad_norm": 0.08594398945569992, + "learning_rate": 0.00011448596499389105, + "loss": 0.0207, + "step": 19361 + }, + { + "epoch": 0.45387577137432095, + "grad_norm": 0.10180538147687912, + "learning_rate": 0.00011447867649342922, + "loss": 0.021, + "step": 19362 + }, + { + "epoch": 0.45389921294912594, + "grad_norm": 0.19941119849681854, + "learning_rate": 0.00011447138791440568, + "loss": 0.0478, + "step": 19363 + }, + { + "epoch": 0.45392265452393094, + "grad_norm": 0.48536592721939087, + "learning_rate": 0.00011446409925686, + "loss": 0.0898, + "step": 19364 + }, + { + "epoch": 0.45394609609873593, + "grad_norm": 0.5928276777267456, + "learning_rate": 0.0001144568105208317, + "loss": 0.7867, + "step": 19365 + }, + { + "epoch": 0.4539695376735409, + "grad_norm": 0.4265228807926178, + "learning_rate": 0.00011444952170636032, + "loss": 0.0739, + "step": 19366 + }, + { + "epoch": 0.4539929792483459, + "grad_norm": 0.3101101815700531, + "learning_rate": 0.00011444223281348545, + "loss": 0.0408, + "step": 19367 + }, + { + "epoch": 0.4540164208231509, + "grad_norm": 0.9806315302848816, + "learning_rate": 0.00011443494384224665, + "loss": 0.0656, + "step": 19368 + }, + { + "epoch": 0.4540398623979559, + "grad_norm": 0.4826527237892151, + "learning_rate": 0.00011442765479268335, + "loss": 0.0727, + "step": 19369 + }, + { + "epoch": 0.4540633039727609, + "grad_norm": 0.10580020397901535, + "learning_rate": 0.00011442036566483525, + "loss": 0.0146, + "step": 19370 + }, + { + "epoch": 0.4540867455475659, + "grad_norm": 0.41085752844810486, + "learning_rate": 0.00011441307645874183, + "loss": 0.105, + "step": 19371 + }, + { + "epoch": 0.4541101871223709, + "grad_norm": 0.6080525517463684, + "learning_rate": 0.00011440578717444265, + "loss": 0.1649, + "step": 19372 + }, + { + "epoch": 0.4541336286971759, + "grad_norm": 0.10885502398014069, + "learning_rate": 0.0001143984978119773, + "loss": 0.0121, + "step": 19373 + }, + { + "epoch": 0.4541570702719809, + "grad_norm": 0.3767455816268921, + "learning_rate": 0.00011439120837138526, + "loss": 0.0763, + "step": 19374 + }, + { + "epoch": 0.4541805118467859, + "grad_norm": 0.4801114797592163, + "learning_rate": 0.00011438391885270611, + "loss": 0.118, + "step": 19375 + }, + { + "epoch": 0.45420395342159087, + "grad_norm": 0.6060811281204224, + "learning_rate": 0.00011437662925597941, + "loss": 0.5727, + "step": 19376 + }, + { + "epoch": 0.45422739499639586, + "grad_norm": 0.5384500622749329, + "learning_rate": 0.00011436933958124474, + "loss": 0.553, + "step": 19377 + }, + { + "epoch": 0.45425083657120086, + "grad_norm": 0.4160885214805603, + "learning_rate": 0.00011436204982854164, + "loss": 0.0499, + "step": 19378 + }, + { + "epoch": 0.45427427814600585, + "grad_norm": 0.5868976712226868, + "learning_rate": 0.00011435475999790961, + "loss": 0.108, + "step": 19379 + }, + { + "epoch": 0.45429771972081084, + "grad_norm": 0.13158567249774933, + "learning_rate": 0.00011434747008938828, + "loss": 0.0164, + "step": 19380 + }, + { + "epoch": 0.45432116129561584, + "grad_norm": 0.5532781481742859, + "learning_rate": 0.00011434018010301717, + "loss": 0.0952, + "step": 19381 + }, + { + "epoch": 0.45434460287042083, + "grad_norm": 0.48349860310554504, + "learning_rate": 0.00011433289003883581, + "loss": 0.1083, + "step": 19382 + }, + { + "epoch": 0.4543680444452258, + "grad_norm": 0.44593170285224915, + "learning_rate": 0.00011432559989688382, + "loss": 0.1131, + "step": 19383 + }, + { + "epoch": 0.4543914860200308, + "grad_norm": 0.05377292260527611, + "learning_rate": 0.00011431830967720074, + "loss": 0.0064, + "step": 19384 + }, + { + "epoch": 0.4544149275948358, + "grad_norm": 0.22532442212104797, + "learning_rate": 0.00011431101937982607, + "loss": 0.0592, + "step": 19385 + }, + { + "epoch": 0.4544383691696408, + "grad_norm": 0.2498566061258316, + "learning_rate": 0.00011430372900479942, + "loss": 0.0311, + "step": 19386 + }, + { + "epoch": 0.4544618107444458, + "grad_norm": 0.25929322838783264, + "learning_rate": 0.00011429643855216032, + "loss": 0.0405, + "step": 19387 + }, + { + "epoch": 0.4544852523192508, + "grad_norm": 0.7764396667480469, + "learning_rate": 0.00011428914802194834, + "loss": 0.5914, + "step": 19388 + }, + { + "epoch": 0.4545086938940558, + "grad_norm": 0.24353253841400146, + "learning_rate": 0.00011428185741420307, + "loss": 0.0399, + "step": 19389 + }, + { + "epoch": 0.4545321354688608, + "grad_norm": 0.3260209858417511, + "learning_rate": 0.00011427456672896403, + "loss": 0.036, + "step": 19390 + }, + { + "epoch": 0.4545555770436658, + "grad_norm": 0.44648224115371704, + "learning_rate": 0.00011426727596627079, + "loss": 0.0732, + "step": 19391 + }, + { + "epoch": 0.4545790186184708, + "grad_norm": 0.37829700112342834, + "learning_rate": 0.0001142599851261629, + "loss": 0.0996, + "step": 19392 + }, + { + "epoch": 0.45460246019327577, + "grad_norm": 0.6710569858551025, + "learning_rate": 0.00011425269420867992, + "loss": 0.5094, + "step": 19393 + }, + { + "epoch": 0.45462590176808076, + "grad_norm": 0.40005046129226685, + "learning_rate": 0.00011424540321386146, + "loss": 0.0764, + "step": 19394 + }, + { + "epoch": 0.45464934334288576, + "grad_norm": 0.11162742972373962, + "learning_rate": 0.000114238112141747, + "loss": 0.0313, + "step": 19395 + }, + { + "epoch": 0.45467278491769075, + "grad_norm": 0.5475130081176758, + "learning_rate": 0.00011423082099237617, + "loss": 0.1113, + "step": 19396 + }, + { + "epoch": 0.45469622649249575, + "grad_norm": 0.5621642470359802, + "learning_rate": 0.0001142235297657885, + "loss": 0.5827, + "step": 19397 + }, + { + "epoch": 0.45471966806730074, + "grad_norm": 0.51836097240448, + "learning_rate": 0.00011421623846202354, + "loss": 0.1029, + "step": 19398 + }, + { + "epoch": 0.45474310964210574, + "grad_norm": 0.29650723934173584, + "learning_rate": 0.00011420894708112089, + "loss": 0.3172, + "step": 19399 + }, + { + "epoch": 0.45476655121691073, + "grad_norm": 0.11432261019945145, + "learning_rate": 0.00011420165562312008, + "loss": 0.0183, + "step": 19400 + }, + { + "epoch": 0.4547899927917157, + "grad_norm": 0.5583848357200623, + "learning_rate": 0.00011419436408806068, + "loss": 0.1198, + "step": 19401 + }, + { + "epoch": 0.4548134343665207, + "grad_norm": 0.416889488697052, + "learning_rate": 0.0001141870724759823, + "loss": 0.0709, + "step": 19402 + }, + { + "epoch": 0.4548368759413257, + "grad_norm": 0.7438718676567078, + "learning_rate": 0.00011417978078692445, + "loss": 0.169, + "step": 19403 + }, + { + "epoch": 0.4548603175161307, + "grad_norm": 0.6155825257301331, + "learning_rate": 0.0001141724890209267, + "loss": 0.1395, + "step": 19404 + }, + { + "epoch": 0.4548837590909357, + "grad_norm": 0.31664982438087463, + "learning_rate": 0.0001141651971780286, + "loss": 0.0322, + "step": 19405 + }, + { + "epoch": 0.45490720066574075, + "grad_norm": 0.3337644636631012, + "learning_rate": 0.00011415790525826979, + "loss": 0.0698, + "step": 19406 + }, + { + "epoch": 0.45493064224054575, + "grad_norm": 0.1648971140384674, + "learning_rate": 0.00011415061326168976, + "loss": 0.0223, + "step": 19407 + }, + { + "epoch": 0.45495408381535074, + "grad_norm": 0.4140975773334503, + "learning_rate": 0.00011414332118832811, + "loss": 0.0931, + "step": 19408 + }, + { + "epoch": 0.45497752539015573, + "grad_norm": 0.5865005850791931, + "learning_rate": 0.00011413602903822443, + "loss": 0.161, + "step": 19409 + }, + { + "epoch": 0.45500096696496073, + "grad_norm": 0.3068358600139618, + "learning_rate": 0.0001141287368114182, + "loss": 0.0385, + "step": 19410 + }, + { + "epoch": 0.4550244085397657, + "grad_norm": 0.4833405613899231, + "learning_rate": 0.00011412144450794909, + "loss": 0.0631, + "step": 19411 + }, + { + "epoch": 0.4550478501145707, + "grad_norm": 0.22794747352600098, + "learning_rate": 0.00011411415212785662, + "loss": 0.0697, + "step": 19412 + }, + { + "epoch": 0.4550712916893757, + "grad_norm": 0.5718029141426086, + "learning_rate": 0.00011410685967118036, + "loss": 0.0921, + "step": 19413 + }, + { + "epoch": 0.4550947332641807, + "grad_norm": 0.7950491309165955, + "learning_rate": 0.00011409956713795988, + "loss": 0.0475, + "step": 19414 + }, + { + "epoch": 0.4551181748389857, + "grad_norm": 0.6081467270851135, + "learning_rate": 0.00011409227452823477, + "loss": 0.1016, + "step": 19415 + }, + { + "epoch": 0.4551416164137907, + "grad_norm": 0.372320294380188, + "learning_rate": 0.00011408498184204457, + "loss": 0.1089, + "step": 19416 + }, + { + "epoch": 0.4551650579885957, + "grad_norm": 0.6862439513206482, + "learning_rate": 0.00011407768907942882, + "loss": 0.1471, + "step": 19417 + }, + { + "epoch": 0.4551884995634007, + "grad_norm": 0.6628676652908325, + "learning_rate": 0.00011407039624042719, + "loss": 0.1141, + "step": 19418 + }, + { + "epoch": 0.4552119411382057, + "grad_norm": 0.5361999869346619, + "learning_rate": 0.00011406310332507919, + "loss": 0.0787, + "step": 19419 + }, + { + "epoch": 0.4552353827130107, + "grad_norm": 0.28153419494628906, + "learning_rate": 0.00011405581033342436, + "loss": 0.0633, + "step": 19420 + }, + { + "epoch": 0.45525882428781567, + "grad_norm": 0.377432644367218, + "learning_rate": 0.00011404851726550234, + "loss": 0.1143, + "step": 19421 + }, + { + "epoch": 0.45528226586262066, + "grad_norm": 0.1709413379430771, + "learning_rate": 0.00011404122412135268, + "loss": 0.0366, + "step": 19422 + }, + { + "epoch": 0.45530570743742566, + "grad_norm": 0.46102404594421387, + "learning_rate": 0.00011403393090101488, + "loss": 0.0917, + "step": 19423 + }, + { + "epoch": 0.45532914901223065, + "grad_norm": 0.29340022802352905, + "learning_rate": 0.00011402663760452863, + "loss": 0.0648, + "step": 19424 + }, + { + "epoch": 0.45535259058703564, + "grad_norm": 0.4911142587661743, + "learning_rate": 0.00011401934423193345, + "loss": 0.094, + "step": 19425 + }, + { + "epoch": 0.45537603216184064, + "grad_norm": 0.15195120871067047, + "learning_rate": 0.00011401205078326889, + "loss": 0.0311, + "step": 19426 + }, + { + "epoch": 0.45539947373664563, + "grad_norm": 0.571685791015625, + "learning_rate": 0.00011400475725857457, + "loss": 0.7445, + "step": 19427 + }, + { + "epoch": 0.4554229153114506, + "grad_norm": 0.7612290978431702, + "learning_rate": 0.00011399746365789006, + "loss": 0.7833, + "step": 19428 + }, + { + "epoch": 0.4554463568862556, + "grad_norm": 0.6331161856651306, + "learning_rate": 0.00011399016998125487, + "loss": 0.1181, + "step": 19429 + }, + { + "epoch": 0.4554697984610606, + "grad_norm": 0.5450359582901001, + "learning_rate": 0.00011398287622870864, + "loss": 0.0538, + "step": 19430 + }, + { + "epoch": 0.4554932400358656, + "grad_norm": 0.41907790303230286, + "learning_rate": 0.00011397558240029096, + "loss": 0.1086, + "step": 19431 + }, + { + "epoch": 0.4555166816106706, + "grad_norm": 0.7384724617004395, + "learning_rate": 0.00011396828849604135, + "loss": 0.0736, + "step": 19432 + }, + { + "epoch": 0.4555401231854756, + "grad_norm": 0.46811386942863464, + "learning_rate": 0.00011396099451599942, + "loss": 0.0694, + "step": 19433 + }, + { + "epoch": 0.4555635647602806, + "grad_norm": 0.3026329576969147, + "learning_rate": 0.00011395370046020474, + "loss": 0.0417, + "step": 19434 + }, + { + "epoch": 0.4555870063350856, + "grad_norm": 0.26237592101097107, + "learning_rate": 0.00011394640632869688, + "loss": 0.2049, + "step": 19435 + }, + { + "epoch": 0.4556104479098906, + "grad_norm": 0.43890491127967834, + "learning_rate": 0.00011393911212151542, + "loss": 0.1361, + "step": 19436 + }, + { + "epoch": 0.4556338894846956, + "grad_norm": 0.24058187007904053, + "learning_rate": 0.00011393181783869998, + "loss": 0.0398, + "step": 19437 + }, + { + "epoch": 0.45565733105950057, + "grad_norm": 0.5602035522460938, + "learning_rate": 0.0001139245234802901, + "loss": 0.1029, + "step": 19438 + }, + { + "epoch": 0.45568077263430556, + "grad_norm": 0.4980970323085785, + "learning_rate": 0.00011391722904632534, + "loss": 0.0829, + "step": 19439 + }, + { + "epoch": 0.45570421420911056, + "grad_norm": 0.778049111366272, + "learning_rate": 0.00011390993453684533, + "loss": 0.1101, + "step": 19440 + }, + { + "epoch": 0.45572765578391555, + "grad_norm": 0.659074604511261, + "learning_rate": 0.00011390263995188958, + "loss": 0.1415, + "step": 19441 + }, + { + "epoch": 0.45575109735872055, + "grad_norm": 0.4280851483345032, + "learning_rate": 0.00011389534529149775, + "loss": 0.0871, + "step": 19442 + }, + { + "epoch": 0.45577453893352554, + "grad_norm": 0.3890379071235657, + "learning_rate": 0.00011388805055570939, + "loss": 0.0874, + "step": 19443 + }, + { + "epoch": 0.45579798050833054, + "grad_norm": 0.3680047392845154, + "learning_rate": 0.00011388075574456407, + "loss": 0.0922, + "step": 19444 + }, + { + "epoch": 0.45582142208313553, + "grad_norm": 0.7224517464637756, + "learning_rate": 0.00011387346085810135, + "loss": 0.2057, + "step": 19445 + }, + { + "epoch": 0.4558448636579405, + "grad_norm": 0.2291937917470932, + "learning_rate": 0.0001138661658963609, + "loss": 0.0494, + "step": 19446 + }, + { + "epoch": 0.4558683052327455, + "grad_norm": 0.5513250827789307, + "learning_rate": 0.0001138588708593822, + "loss": 0.0964, + "step": 19447 + }, + { + "epoch": 0.4558917468075505, + "grad_norm": 0.5790050625801086, + "learning_rate": 0.00011385157574720488, + "loss": 0.1226, + "step": 19448 + }, + { + "epoch": 0.4559151883823555, + "grad_norm": 0.5708324313163757, + "learning_rate": 0.00011384428055986853, + "loss": 0.63, + "step": 19449 + }, + { + "epoch": 0.4559386299571605, + "grad_norm": 0.4303536117076874, + "learning_rate": 0.00011383698529741274, + "loss": 0.0439, + "step": 19450 + }, + { + "epoch": 0.4559620715319655, + "grad_norm": 0.6182339191436768, + "learning_rate": 0.00011382968995987706, + "loss": 0.068, + "step": 19451 + }, + { + "epoch": 0.4559855131067705, + "grad_norm": 0.23798871040344238, + "learning_rate": 0.00011382239454730108, + "loss": 0.0453, + "step": 19452 + }, + { + "epoch": 0.4560089546815755, + "grad_norm": 0.38775312900543213, + "learning_rate": 0.00011381509905972443, + "loss": 0.0736, + "step": 19453 + }, + { + "epoch": 0.4560323962563805, + "grad_norm": 0.4518750011920929, + "learning_rate": 0.00011380780349718662, + "loss": 0.1026, + "step": 19454 + }, + { + "epoch": 0.4560558378311855, + "grad_norm": 0.1401844322681427, + "learning_rate": 0.00011380050785972731, + "loss": 0.0304, + "step": 19455 + }, + { + "epoch": 0.45607927940599047, + "grad_norm": 0.7161044478416443, + "learning_rate": 0.00011379321214738607, + "loss": 0.2173, + "step": 19456 + }, + { + "epoch": 0.4561027209807955, + "grad_norm": 0.5110958814620972, + "learning_rate": 0.00011378591636020246, + "loss": 0.0861, + "step": 19457 + }, + { + "epoch": 0.4561261625556005, + "grad_norm": 0.2443901002407074, + "learning_rate": 0.00011377862049821608, + "loss": 0.0599, + "step": 19458 + }, + { + "epoch": 0.4561496041304055, + "grad_norm": 0.4250696897506714, + "learning_rate": 0.00011377132456146651, + "loss": 0.0961, + "step": 19459 + }, + { + "epoch": 0.4561730457052105, + "grad_norm": 0.5364399552345276, + "learning_rate": 0.00011376402854999335, + "loss": 0.066, + "step": 19460 + }, + { + "epoch": 0.4561964872800155, + "grad_norm": 0.19963331520557404, + "learning_rate": 0.00011375673246383618, + "loss": 0.0316, + "step": 19461 + }, + { + "epoch": 0.4562199288548205, + "grad_norm": 0.4947296679019928, + "learning_rate": 0.0001137494363030346, + "loss": 0.1043, + "step": 19462 + }, + { + "epoch": 0.4562433704296255, + "grad_norm": 0.4676910936832428, + "learning_rate": 0.0001137421400676282, + "loss": 0.0878, + "step": 19463 + }, + { + "epoch": 0.4562668120044305, + "grad_norm": 0.6706791520118713, + "learning_rate": 0.00011373484375765652, + "loss": 0.0919, + "step": 19464 + }, + { + "epoch": 0.45629025357923547, + "grad_norm": 0.5678118467330933, + "learning_rate": 0.00011372754737315925, + "loss": 0.1974, + "step": 19465 + }, + { + "epoch": 0.45631369515404047, + "grad_norm": 0.2836950123310089, + "learning_rate": 0.00011372025091417589, + "loss": 0.0664, + "step": 19466 + }, + { + "epoch": 0.45633713672884546, + "grad_norm": 0.5735926628112793, + "learning_rate": 0.00011371295438074607, + "loss": 0.081, + "step": 19467 + }, + { + "epoch": 0.45636057830365045, + "grad_norm": 0.179335355758667, + "learning_rate": 0.00011370565777290935, + "loss": 0.0173, + "step": 19468 + }, + { + "epoch": 0.45638401987845545, + "grad_norm": 0.527811586856842, + "learning_rate": 0.0001136983610907054, + "loss": 0.0682, + "step": 19469 + }, + { + "epoch": 0.45640746145326044, + "grad_norm": 0.32728323340415955, + "learning_rate": 0.00011369106433417371, + "loss": 0.0334, + "step": 19470 + }, + { + "epoch": 0.45643090302806544, + "grad_norm": 0.3426095247268677, + "learning_rate": 0.00011368376750335394, + "loss": 0.036, + "step": 19471 + }, + { + "epoch": 0.45645434460287043, + "grad_norm": 0.12868943810462952, + "learning_rate": 0.00011367647059828563, + "loss": 0.0355, + "step": 19472 + }, + { + "epoch": 0.4564777861776754, + "grad_norm": 0.7333024740219116, + "learning_rate": 0.00011366917361900846, + "loss": 0.8846, + "step": 19473 + }, + { + "epoch": 0.4565012277524804, + "grad_norm": 0.2908947467803955, + "learning_rate": 0.00011366187656556192, + "loss": 0.0948, + "step": 19474 + }, + { + "epoch": 0.4565246693272854, + "grad_norm": 0.4210386872291565, + "learning_rate": 0.0001136545794379857, + "loss": 0.0604, + "step": 19475 + }, + { + "epoch": 0.4565481109020904, + "grad_norm": 0.5939552783966064, + "learning_rate": 0.00011364728223631934, + "loss": 0.0973, + "step": 19476 + }, + { + "epoch": 0.4565715524768954, + "grad_norm": 0.4816949963569641, + "learning_rate": 0.00011363998496060242, + "loss": 0.1012, + "step": 19477 + }, + { + "epoch": 0.4565949940517004, + "grad_norm": 0.495505690574646, + "learning_rate": 0.00011363268761087453, + "loss": 0.1118, + "step": 19478 + }, + { + "epoch": 0.4566184356265054, + "grad_norm": 0.689054012298584, + "learning_rate": 0.00011362539018717535, + "loss": 0.1272, + "step": 19479 + }, + { + "epoch": 0.4566418772013104, + "grad_norm": 0.0996941328048706, + "learning_rate": 0.0001136180926895444, + "loss": 0.0197, + "step": 19480 + }, + { + "epoch": 0.4566653187761154, + "grad_norm": 0.16982613503932953, + "learning_rate": 0.00011361079511802129, + "loss": 0.0292, + "step": 19481 + }, + { + "epoch": 0.4566887603509204, + "grad_norm": 0.46552401781082153, + "learning_rate": 0.00011360349747264565, + "loss": 0.1281, + "step": 19482 + }, + { + "epoch": 0.45671220192572537, + "grad_norm": 0.5507971048355103, + "learning_rate": 0.00011359619975345701, + "loss": 0.1126, + "step": 19483 + }, + { + "epoch": 0.45673564350053036, + "grad_norm": 0.6384790539741516, + "learning_rate": 0.000113588901960495, + "loss": 0.1106, + "step": 19484 + }, + { + "epoch": 0.45675908507533536, + "grad_norm": 0.3544723391532898, + "learning_rate": 0.00011358160409379927, + "loss": 0.0603, + "step": 19485 + }, + { + "epoch": 0.45678252665014035, + "grad_norm": 0.42914092540740967, + "learning_rate": 0.00011357430615340938, + "loss": 0.092, + "step": 19486 + }, + { + "epoch": 0.45680596822494535, + "grad_norm": 0.6477246284484863, + "learning_rate": 0.00011356700813936489, + "loss": 0.0742, + "step": 19487 + }, + { + "epoch": 0.45682940979975034, + "grad_norm": 0.342852383852005, + "learning_rate": 0.00011355971005170545, + "loss": 0.0358, + "step": 19488 + }, + { + "epoch": 0.45685285137455534, + "grad_norm": 0.36855173110961914, + "learning_rate": 0.00011355241189047064, + "loss": 0.0633, + "step": 19489 + }, + { + "epoch": 0.45687629294936033, + "grad_norm": 0.15689033269882202, + "learning_rate": 0.00011354511365570002, + "loss": 0.0236, + "step": 19490 + }, + { + "epoch": 0.4568997345241653, + "grad_norm": 0.14100289344787598, + "learning_rate": 0.00011353781534743328, + "loss": 0.0307, + "step": 19491 + }, + { + "epoch": 0.4569231760989703, + "grad_norm": 0.48312076926231384, + "learning_rate": 0.00011353051696570998, + "loss": 0.1185, + "step": 19492 + }, + { + "epoch": 0.4569466176737753, + "grad_norm": 0.5573525428771973, + "learning_rate": 0.00011352321851056968, + "loss": 0.0766, + "step": 19493 + }, + { + "epoch": 0.4569700592485803, + "grad_norm": 0.3618801236152649, + "learning_rate": 0.00011351591998205205, + "loss": 0.0465, + "step": 19494 + }, + { + "epoch": 0.4569935008233853, + "grad_norm": 0.23731957376003265, + "learning_rate": 0.00011350862138019664, + "loss": 0.0536, + "step": 19495 + }, + { + "epoch": 0.4570169423981903, + "grad_norm": 0.5478266477584839, + "learning_rate": 0.00011350132270504305, + "loss": 0.1083, + "step": 19496 + }, + { + "epoch": 0.4570403839729953, + "grad_norm": 0.10259192436933517, + "learning_rate": 0.00011349402395663093, + "loss": 0.0135, + "step": 19497 + }, + { + "epoch": 0.4570638255478003, + "grad_norm": 0.43239983916282654, + "learning_rate": 0.00011348672513499985, + "loss": 0.1065, + "step": 19498 + }, + { + "epoch": 0.4570872671226053, + "grad_norm": 0.6875700354576111, + "learning_rate": 0.00011347942624018941, + "loss": 0.1313, + "step": 19499 + }, + { + "epoch": 0.4571107086974103, + "grad_norm": 0.3962087631225586, + "learning_rate": 0.00011347212727223926, + "loss": 0.0572, + "step": 19500 + }, + { + "epoch": 0.45713415027221527, + "grad_norm": 0.5666725039482117, + "learning_rate": 0.00011346482823118894, + "loss": 0.1131, + "step": 19501 + }, + { + "epoch": 0.45715759184702026, + "grad_norm": 0.43456658720970154, + "learning_rate": 0.00011345752911707808, + "loss": 0.0678, + "step": 19502 + }, + { + "epoch": 0.45718103342182526, + "grad_norm": 0.9313445091247559, + "learning_rate": 0.00011345022992994629, + "loss": 0.0706, + "step": 19503 + }, + { + "epoch": 0.45720447499663025, + "grad_norm": 0.38604581356048584, + "learning_rate": 0.00011344293066983318, + "loss": 0.108, + "step": 19504 + }, + { + "epoch": 0.45722791657143524, + "grad_norm": 0.26721668243408203, + "learning_rate": 0.00011343563133677838, + "loss": 0.0523, + "step": 19505 + }, + { + "epoch": 0.45725135814624024, + "grad_norm": 0.4792135953903198, + "learning_rate": 0.00011342833193082141, + "loss": 0.0784, + "step": 19506 + }, + { + "epoch": 0.45727479972104523, + "grad_norm": 0.8554685115814209, + "learning_rate": 0.00011342103245200198, + "loss": 0.1222, + "step": 19507 + }, + { + "epoch": 0.4572982412958503, + "grad_norm": 0.13001759350299835, + "learning_rate": 0.00011341373290035962, + "loss": 0.0223, + "step": 19508 + }, + { + "epoch": 0.4573216828706553, + "grad_norm": 0.5468817353248596, + "learning_rate": 0.00011340643327593396, + "loss": 0.1188, + "step": 19509 + }, + { + "epoch": 0.45734512444546027, + "grad_norm": 0.48975634574890137, + "learning_rate": 0.00011339913357876464, + "loss": 0.0723, + "step": 19510 + }, + { + "epoch": 0.45736856602026527, + "grad_norm": 0.31413358449935913, + "learning_rate": 0.00011339183380889125, + "loss": 0.0596, + "step": 19511 + }, + { + "epoch": 0.45739200759507026, + "grad_norm": 0.5783222913742065, + "learning_rate": 0.00011338453396635338, + "loss": 0.1436, + "step": 19512 + }, + { + "epoch": 0.45741544916987525, + "grad_norm": 0.3289085030555725, + "learning_rate": 0.00011337723405119067, + "loss": 0.0612, + "step": 19513 + }, + { + "epoch": 0.45743889074468025, + "grad_norm": 0.15881043672561646, + "learning_rate": 0.00011336993406344267, + "loss": 0.0182, + "step": 19514 + }, + { + "epoch": 0.45746233231948524, + "grad_norm": 0.10634510219097137, + "learning_rate": 0.00011336263400314908, + "loss": 0.019, + "step": 19515 + }, + { + "epoch": 0.45748577389429024, + "grad_norm": 0.4215477705001831, + "learning_rate": 0.00011335533387034942, + "loss": 0.0779, + "step": 19516 + }, + { + "epoch": 0.45750921546909523, + "grad_norm": 0.19392022490501404, + "learning_rate": 0.00011334803366508338, + "loss": 0.0242, + "step": 19517 + }, + { + "epoch": 0.4575326570439002, + "grad_norm": 0.4124523103237152, + "learning_rate": 0.00011334073338739051, + "loss": 0.0733, + "step": 19518 + }, + { + "epoch": 0.4575560986187052, + "grad_norm": 0.414124459028244, + "learning_rate": 0.00011333343303731046, + "loss": 0.103, + "step": 19519 + }, + { + "epoch": 0.4575795401935102, + "grad_norm": 1.2097707986831665, + "learning_rate": 0.00011332613261488281, + "loss": 0.133, + "step": 19520 + }, + { + "epoch": 0.4576029817683152, + "grad_norm": 0.21076129376888275, + "learning_rate": 0.00011331883212014722, + "loss": 0.0341, + "step": 19521 + }, + { + "epoch": 0.4576264233431202, + "grad_norm": 0.3642481863498688, + "learning_rate": 0.00011331153155314324, + "loss": 0.0472, + "step": 19522 + }, + { + "epoch": 0.4576498649179252, + "grad_norm": 0.38632503151893616, + "learning_rate": 0.00011330423091391054, + "loss": 0.3449, + "step": 19523 + }, + { + "epoch": 0.4576733064927302, + "grad_norm": 0.34955570101737976, + "learning_rate": 0.0001132969302024887, + "loss": 0.0593, + "step": 19524 + }, + { + "epoch": 0.4576967480675352, + "grad_norm": 0.5111928582191467, + "learning_rate": 0.00011328962941891734, + "loss": 0.1158, + "step": 19525 + }, + { + "epoch": 0.4577201896423402, + "grad_norm": 0.4750450551509857, + "learning_rate": 0.00011328232856323608, + "loss": 0.0854, + "step": 19526 + }, + { + "epoch": 0.4577436312171452, + "grad_norm": 0.26833420991897583, + "learning_rate": 0.00011327502763548452, + "loss": 0.0642, + "step": 19527 + }, + { + "epoch": 0.45776707279195017, + "grad_norm": 0.5141380429267883, + "learning_rate": 0.00011326772663570229, + "loss": 0.0613, + "step": 19528 + }, + { + "epoch": 0.45779051436675516, + "grad_norm": 0.67591792345047, + "learning_rate": 0.00011326042556392902, + "loss": 0.1048, + "step": 19529 + }, + { + "epoch": 0.45781395594156016, + "grad_norm": 0.3082846403121948, + "learning_rate": 0.0001132531244202043, + "loss": 0.0644, + "step": 19530 + }, + { + "epoch": 0.45783739751636515, + "grad_norm": 0.2768884599208832, + "learning_rate": 0.00011324582320456774, + "loss": 0.0321, + "step": 19531 + }, + { + "epoch": 0.45786083909117015, + "grad_norm": 0.5103486180305481, + "learning_rate": 0.000113238521917059, + "loss": 0.4776, + "step": 19532 + }, + { + "epoch": 0.45788428066597514, + "grad_norm": 0.20484423637390137, + "learning_rate": 0.00011323122055771763, + "loss": 0.0418, + "step": 19533 + }, + { + "epoch": 0.45790772224078014, + "grad_norm": 0.43324699997901917, + "learning_rate": 0.00011322391912658332, + "loss": 0.0672, + "step": 19534 + }, + { + "epoch": 0.45793116381558513, + "grad_norm": 0.23515692353248596, + "learning_rate": 0.00011321661762369562, + "loss": 0.0481, + "step": 19535 + }, + { + "epoch": 0.4579546053903901, + "grad_norm": 0.12109826505184174, + "learning_rate": 0.00011320931604909421, + "loss": 0.0237, + "step": 19536 + }, + { + "epoch": 0.4579780469651951, + "grad_norm": 0.4030793607234955, + "learning_rate": 0.00011320201440281865, + "loss": 0.1003, + "step": 19537 + }, + { + "epoch": 0.4580014885400001, + "grad_norm": 0.09107037633657455, + "learning_rate": 0.00011319471268490862, + "loss": 0.0164, + "step": 19538 + }, + { + "epoch": 0.4580249301148051, + "grad_norm": 0.4341185986995697, + "learning_rate": 0.00011318741089540368, + "loss": 0.4897, + "step": 19539 + }, + { + "epoch": 0.4580483716896101, + "grad_norm": 0.47295212745666504, + "learning_rate": 0.00011318010903434351, + "loss": 0.0814, + "step": 19540 + }, + { + "epoch": 0.4580718132644151, + "grad_norm": 0.489892840385437, + "learning_rate": 0.00011317280710176767, + "loss": 0.0855, + "step": 19541 + }, + { + "epoch": 0.4580952548392201, + "grad_norm": 0.5738192796707153, + "learning_rate": 0.00011316550509771584, + "loss": 0.099, + "step": 19542 + }, + { + "epoch": 0.4581186964140251, + "grad_norm": 0.6149821877479553, + "learning_rate": 0.0001131582030222276, + "loss": 0.1194, + "step": 19543 + }, + { + "epoch": 0.4581421379888301, + "grad_norm": 0.3232951760292053, + "learning_rate": 0.00011315090087534254, + "loss": 0.0581, + "step": 19544 + }, + { + "epoch": 0.4581655795636351, + "grad_norm": 0.39900317788124084, + "learning_rate": 0.00011314359865710033, + "loss": 0.0763, + "step": 19545 + }, + { + "epoch": 0.45818902113844007, + "grad_norm": 0.6770327687263489, + "learning_rate": 0.00011313629636754064, + "loss": 0.1649, + "step": 19546 + }, + { + "epoch": 0.45821246271324506, + "grad_norm": 0.21557359397411346, + "learning_rate": 0.00011312899400670298, + "loss": 0.028, + "step": 19547 + }, + { + "epoch": 0.45823590428805006, + "grad_norm": 0.7646905779838562, + "learning_rate": 0.00011312169157462706, + "loss": 0.1157, + "step": 19548 + }, + { + "epoch": 0.45825934586285505, + "grad_norm": 0.3966282606124878, + "learning_rate": 0.00011311438907135248, + "loss": 0.1048, + "step": 19549 + }, + { + "epoch": 0.45828278743766004, + "grad_norm": 0.3425169885158539, + "learning_rate": 0.00011310708649691881, + "loss": 0.0636, + "step": 19550 + }, + { + "epoch": 0.45830622901246504, + "grad_norm": 0.7716038227081299, + "learning_rate": 0.00011309978385136574, + "loss": 0.0909, + "step": 19551 + }, + { + "epoch": 0.45832967058727003, + "grad_norm": 0.15399503707885742, + "learning_rate": 0.00011309248113473289, + "loss": 0.0252, + "step": 19552 + }, + { + "epoch": 0.458353112162075, + "grad_norm": 0.8100013732910156, + "learning_rate": 0.00011308517834705989, + "loss": 0.0395, + "step": 19553 + }, + { + "epoch": 0.45837655373688, + "grad_norm": 0.6717228889465332, + "learning_rate": 0.00011307787548838631, + "loss": 0.6795, + "step": 19554 + }, + { + "epoch": 0.458399995311685, + "grad_norm": 0.3478306829929352, + "learning_rate": 0.00011307057255875182, + "loss": 0.0567, + "step": 19555 + }, + { + "epoch": 0.45842343688649, + "grad_norm": 0.34432452917099, + "learning_rate": 0.00011306326955819602, + "loss": 0.0393, + "step": 19556 + }, + { + "epoch": 0.458446878461295, + "grad_norm": 0.39959847927093506, + "learning_rate": 0.00011305596648675855, + "loss": 0.0833, + "step": 19557 + }, + { + "epoch": 0.4584703200361, + "grad_norm": 0.20379725098609924, + "learning_rate": 0.00011304866334447908, + "loss": 0.0306, + "step": 19558 + }, + { + "epoch": 0.458493761610905, + "grad_norm": 0.6449576616287231, + "learning_rate": 0.00011304136013139716, + "loss": 0.158, + "step": 19559 + }, + { + "epoch": 0.45851720318571004, + "grad_norm": 0.6491860747337341, + "learning_rate": 0.00011303405684755247, + "loss": 0.076, + "step": 19560 + }, + { + "epoch": 0.45854064476051504, + "grad_norm": 0.6397985219955444, + "learning_rate": 0.00011302675349298461, + "loss": 0.6022, + "step": 19561 + }, + { + "epoch": 0.45856408633532003, + "grad_norm": 0.5839025378227234, + "learning_rate": 0.00011301945006773324, + "loss": 0.1108, + "step": 19562 + }, + { + "epoch": 0.458587527910125, + "grad_norm": 0.6306737661361694, + "learning_rate": 0.00011301214657183791, + "loss": 0.0714, + "step": 19563 + }, + { + "epoch": 0.45861096948493, + "grad_norm": 0.5769508481025696, + "learning_rate": 0.00011300484300533835, + "loss": 0.161, + "step": 19564 + }, + { + "epoch": 0.458634411059735, + "grad_norm": 0.6613409519195557, + "learning_rate": 0.00011299753936827418, + "loss": 0.1072, + "step": 19565 + }, + { + "epoch": 0.45865785263454, + "grad_norm": 0.49195900559425354, + "learning_rate": 0.00011299023566068494, + "loss": 0.1054, + "step": 19566 + }, + { + "epoch": 0.458681294209345, + "grad_norm": 0.5025227069854736, + "learning_rate": 0.00011298293188261035, + "loss": 0.0701, + "step": 19567 + }, + { + "epoch": 0.45870473578415, + "grad_norm": 0.5176863074302673, + "learning_rate": 0.00011297562803409, + "loss": 0.0829, + "step": 19568 + }, + { + "epoch": 0.458728177358955, + "grad_norm": 0.4528326392173767, + "learning_rate": 0.00011296832411516352, + "loss": 0.0657, + "step": 19569 + }, + { + "epoch": 0.45875161893376, + "grad_norm": 0.35104605555534363, + "learning_rate": 0.00011296102012587053, + "loss": 0.0385, + "step": 19570 + }, + { + "epoch": 0.458775060508565, + "grad_norm": 0.15178516507148743, + "learning_rate": 0.00011295371606625072, + "loss": 0.0332, + "step": 19571 + }, + { + "epoch": 0.45879850208337, + "grad_norm": 0.8051806688308716, + "learning_rate": 0.00011294641193634364, + "loss": 0.1166, + "step": 19572 + }, + { + "epoch": 0.45882194365817497, + "grad_norm": 0.4783637225627899, + "learning_rate": 0.00011293910773618899, + "loss": 0.0856, + "step": 19573 + }, + { + "epoch": 0.45884538523297996, + "grad_norm": 0.2832334339618683, + "learning_rate": 0.00011293180346582639, + "loss": 0.047, + "step": 19574 + }, + { + "epoch": 0.45886882680778496, + "grad_norm": 0.25798672437667847, + "learning_rate": 0.00011292449912529543, + "loss": 0.0523, + "step": 19575 + }, + { + "epoch": 0.45889226838258995, + "grad_norm": 0.5112977027893066, + "learning_rate": 0.00011291719471463578, + "loss": 0.1074, + "step": 19576 + }, + { + "epoch": 0.45891570995739495, + "grad_norm": 0.46237435936927795, + "learning_rate": 0.0001129098902338871, + "loss": 0.0717, + "step": 19577 + }, + { + "epoch": 0.45893915153219994, + "grad_norm": 0.6301459670066833, + "learning_rate": 0.00011290258568308895, + "loss": 0.1099, + "step": 19578 + }, + { + "epoch": 0.45896259310700493, + "grad_norm": 0.3310921788215637, + "learning_rate": 0.00011289528106228102, + "loss": 0.0381, + "step": 19579 + }, + { + "epoch": 0.45898603468180993, + "grad_norm": 0.27474069595336914, + "learning_rate": 0.00011288797637150297, + "loss": 0.0512, + "step": 19580 + }, + { + "epoch": 0.4590094762566149, + "grad_norm": 0.9632107615470886, + "learning_rate": 0.00011288067161079434, + "loss": 0.1967, + "step": 19581 + }, + { + "epoch": 0.4590329178314199, + "grad_norm": 0.4615296721458435, + "learning_rate": 0.00011287336678019484, + "loss": 0.0937, + "step": 19582 + }, + { + "epoch": 0.4590563594062249, + "grad_norm": 0.515038013458252, + "learning_rate": 0.00011286606187974411, + "loss": 0.0861, + "step": 19583 + }, + { + "epoch": 0.4590798009810299, + "grad_norm": 0.7157445549964905, + "learning_rate": 0.00011285875690948178, + "loss": 0.6205, + "step": 19584 + }, + { + "epoch": 0.4591032425558349, + "grad_norm": 0.9232449531555176, + "learning_rate": 0.00011285145186944744, + "loss": 0.2256, + "step": 19585 + }, + { + "epoch": 0.4591266841306399, + "grad_norm": 0.15491971373558044, + "learning_rate": 0.00011284414675968078, + "loss": 0.0357, + "step": 19586 + }, + { + "epoch": 0.4591501257054449, + "grad_norm": 0.4847120940685272, + "learning_rate": 0.0001128368415802214, + "loss": 0.1067, + "step": 19587 + }, + { + "epoch": 0.4591735672802499, + "grad_norm": 0.3333476185798645, + "learning_rate": 0.00011282953633110897, + "loss": 0.0711, + "step": 19588 + }, + { + "epoch": 0.4591970088550549, + "grad_norm": 0.5546610355377197, + "learning_rate": 0.00011282223101238312, + "loss": 0.1318, + "step": 19589 + }, + { + "epoch": 0.45922045042985987, + "grad_norm": 0.5877282023429871, + "learning_rate": 0.00011281492562408347, + "loss": 0.106, + "step": 19590 + }, + { + "epoch": 0.45924389200466487, + "grad_norm": 0.13581712543964386, + "learning_rate": 0.00011280762016624966, + "loss": 0.0234, + "step": 19591 + }, + { + "epoch": 0.45926733357946986, + "grad_norm": 0.5029058456420898, + "learning_rate": 0.00011280031463892138, + "loss": 0.1144, + "step": 19592 + }, + { + "epoch": 0.45929077515427486, + "grad_norm": 0.8369141817092896, + "learning_rate": 0.00011279300904213822, + "loss": 0.0639, + "step": 19593 + }, + { + "epoch": 0.45931421672907985, + "grad_norm": 0.3785848021507263, + "learning_rate": 0.00011278570337593983, + "loss": 0.0658, + "step": 19594 + }, + { + "epoch": 0.45933765830388484, + "grad_norm": 0.30926448106765747, + "learning_rate": 0.00011277839764036584, + "loss": 0.0405, + "step": 19595 + }, + { + "epoch": 0.45936109987868984, + "grad_norm": 0.6683403849601746, + "learning_rate": 0.00011277109183545593, + "loss": 0.1402, + "step": 19596 + }, + { + "epoch": 0.45938454145349483, + "grad_norm": 0.47164562344551086, + "learning_rate": 0.0001127637859612497, + "loss": 0.1573, + "step": 19597 + }, + { + "epoch": 0.4594079830282998, + "grad_norm": 0.5045403838157654, + "learning_rate": 0.0001127564800177868, + "loss": 0.1242, + "step": 19598 + }, + { + "epoch": 0.4594314246031048, + "grad_norm": 0.6608902812004089, + "learning_rate": 0.0001127491740051069, + "loss": 0.1945, + "step": 19599 + }, + { + "epoch": 0.4594548661779098, + "grad_norm": 0.1263086348772049, + "learning_rate": 0.00011274186792324961, + "loss": 0.0225, + "step": 19600 + }, + { + "epoch": 0.4594783077527148, + "grad_norm": 0.5416107773780823, + "learning_rate": 0.00011273456177225459, + "loss": 0.0884, + "step": 19601 + }, + { + "epoch": 0.4595017493275198, + "grad_norm": 0.3703921139240265, + "learning_rate": 0.0001127272555521615, + "loss": 0.087, + "step": 19602 + }, + { + "epoch": 0.4595251909023248, + "grad_norm": 1.313743233680725, + "learning_rate": 0.00011271994926300996, + "loss": 0.1086, + "step": 19603 + }, + { + "epoch": 0.4595486324771298, + "grad_norm": 0.6132753491401672, + "learning_rate": 0.00011271264290483958, + "loss": 0.6211, + "step": 19604 + }, + { + "epoch": 0.4595720740519348, + "grad_norm": 0.11094996333122253, + "learning_rate": 0.00011270533647769008, + "loss": 0.0172, + "step": 19605 + }, + { + "epoch": 0.4595955156267398, + "grad_norm": 0.7848434448242188, + "learning_rate": 0.00011269802998160103, + "loss": 0.1508, + "step": 19606 + }, + { + "epoch": 0.4596189572015448, + "grad_norm": 0.2577967047691345, + "learning_rate": 0.00011269072341661216, + "loss": 0.0592, + "step": 19607 + }, + { + "epoch": 0.45964239877634977, + "grad_norm": 0.5556939840316772, + "learning_rate": 0.00011268341678276302, + "loss": 0.4536, + "step": 19608 + }, + { + "epoch": 0.45966584035115476, + "grad_norm": 0.2116139978170395, + "learning_rate": 0.00011267611008009336, + "loss": 0.03, + "step": 19609 + }, + { + "epoch": 0.45968928192595976, + "grad_norm": 0.4399493932723999, + "learning_rate": 0.00011266880330864271, + "loss": 0.092, + "step": 19610 + }, + { + "epoch": 0.4597127235007648, + "grad_norm": 0.5721597075462341, + "learning_rate": 0.00011266149646845081, + "loss": 0.1376, + "step": 19611 + }, + { + "epoch": 0.4597361650755698, + "grad_norm": 0.31926649808883667, + "learning_rate": 0.00011265418955955727, + "loss": 0.0777, + "step": 19612 + }, + { + "epoch": 0.4597596066503748, + "grad_norm": 0.41540130972862244, + "learning_rate": 0.00011264688258200175, + "loss": 0.0998, + "step": 19613 + }, + { + "epoch": 0.4597830482251798, + "grad_norm": 0.71217280626297, + "learning_rate": 0.00011263957553582388, + "loss": 0.1263, + "step": 19614 + }, + { + "epoch": 0.4598064897999848, + "grad_norm": 0.4933594763278961, + "learning_rate": 0.00011263226842106332, + "loss": 0.0982, + "step": 19615 + }, + { + "epoch": 0.4598299313747898, + "grad_norm": 0.1874329298734665, + "learning_rate": 0.00011262496123775972, + "loss": 0.0538, + "step": 19616 + }, + { + "epoch": 0.4598533729495948, + "grad_norm": 0.5459165573120117, + "learning_rate": 0.0001126176539859527, + "loss": 0.102, + "step": 19617 + }, + { + "epoch": 0.45987681452439977, + "grad_norm": 0.6992403268814087, + "learning_rate": 0.00011261034666568195, + "loss": 0.1763, + "step": 19618 + }, + { + "epoch": 0.45990025609920476, + "grad_norm": 0.5952387452125549, + "learning_rate": 0.00011260303927698712, + "loss": 0.0971, + "step": 19619 + }, + { + "epoch": 0.45992369767400976, + "grad_norm": 0.5028740763664246, + "learning_rate": 0.00011259573181990782, + "loss": 0.0559, + "step": 19620 + }, + { + "epoch": 0.45994713924881475, + "grad_norm": 0.42800173163414, + "learning_rate": 0.00011258842429448376, + "loss": 0.046, + "step": 19621 + }, + { + "epoch": 0.45997058082361975, + "grad_norm": 0.552784264087677, + "learning_rate": 0.00011258111670075453, + "loss": 0.0847, + "step": 19622 + }, + { + "epoch": 0.45999402239842474, + "grad_norm": 0.3451559841632843, + "learning_rate": 0.00011257380903875981, + "loss": 0.0686, + "step": 19623 + }, + { + "epoch": 0.46001746397322973, + "grad_norm": 0.38941746950149536, + "learning_rate": 0.00011256650130853923, + "loss": 0.048, + "step": 19624 + }, + { + "epoch": 0.46004090554803473, + "grad_norm": 0.5464795827865601, + "learning_rate": 0.00011255919351013248, + "loss": 0.1035, + "step": 19625 + }, + { + "epoch": 0.4600643471228397, + "grad_norm": 0.3523922264575958, + "learning_rate": 0.00011255188564357921, + "loss": 0.0515, + "step": 19626 + }, + { + "epoch": 0.4600877886976447, + "grad_norm": 0.6378428339958191, + "learning_rate": 0.00011254457770891905, + "loss": 0.1225, + "step": 19627 + }, + { + "epoch": 0.4601112302724497, + "grad_norm": 0.5731033682823181, + "learning_rate": 0.00011253726970619162, + "loss": 0.1628, + "step": 19628 + }, + { + "epoch": 0.4601346718472547, + "grad_norm": 0.4451814591884613, + "learning_rate": 0.00011252996163543665, + "loss": 0.0988, + "step": 19629 + }, + { + "epoch": 0.4601581134220597, + "grad_norm": 0.7117102742195129, + "learning_rate": 0.00011252265349669373, + "loss": 0.1414, + "step": 19630 + }, + { + "epoch": 0.4601815549968647, + "grad_norm": 0.8316993713378906, + "learning_rate": 0.00011251534529000257, + "loss": 0.0775, + "step": 19631 + }, + { + "epoch": 0.4602049965716697, + "grad_norm": 0.31157350540161133, + "learning_rate": 0.00011250803701540277, + "loss": 0.0362, + "step": 19632 + }, + { + "epoch": 0.4602284381464747, + "grad_norm": 0.5871089696884155, + "learning_rate": 0.00011250072867293401, + "loss": 0.1053, + "step": 19633 + }, + { + "epoch": 0.4602518797212797, + "grad_norm": 0.4813952147960663, + "learning_rate": 0.00011249342026263598, + "loss": 0.0893, + "step": 19634 + }, + { + "epoch": 0.46027532129608467, + "grad_norm": 0.4245591163635254, + "learning_rate": 0.00011248611178454829, + "loss": 0.1169, + "step": 19635 + }, + { + "epoch": 0.46029876287088967, + "grad_norm": 0.4828568696975708, + "learning_rate": 0.00011247880323871053, + "loss": 0.2986, + "step": 19636 + }, + { + "epoch": 0.46032220444569466, + "grad_norm": 0.4976584315299988, + "learning_rate": 0.00011247149462516251, + "loss": 0.1167, + "step": 19637 + }, + { + "epoch": 0.46034564602049965, + "grad_norm": 0.5835678577423096, + "learning_rate": 0.00011246418594394381, + "loss": 0.1037, + "step": 19638 + }, + { + "epoch": 0.46036908759530465, + "grad_norm": 0.17183087766170502, + "learning_rate": 0.00011245687719509405, + "loss": 0.0179, + "step": 19639 + }, + { + "epoch": 0.46039252917010964, + "grad_norm": 0.3053854703903198, + "learning_rate": 0.00011244956837865297, + "loss": 0.0516, + "step": 19640 + }, + { + "epoch": 0.46041597074491464, + "grad_norm": 0.29974663257598877, + "learning_rate": 0.00011244225949466017, + "loss": 0.2364, + "step": 19641 + }, + { + "epoch": 0.46043941231971963, + "grad_norm": 0.4058040380477905, + "learning_rate": 0.0001124349505431553, + "loss": 0.0921, + "step": 19642 + }, + { + "epoch": 0.4604628538945246, + "grad_norm": 0.15503205358982086, + "learning_rate": 0.00011242764152417802, + "loss": 0.0291, + "step": 19643 + }, + { + "epoch": 0.4604862954693296, + "grad_norm": 0.5144335627555847, + "learning_rate": 0.00011242033243776804, + "loss": 0.0972, + "step": 19644 + }, + { + "epoch": 0.4605097370441346, + "grad_norm": 0.3093080222606659, + "learning_rate": 0.00011241302328396499, + "loss": 0.0429, + "step": 19645 + }, + { + "epoch": 0.4605331786189396, + "grad_norm": 0.5700169205665588, + "learning_rate": 0.00011240571406280851, + "loss": 0.1153, + "step": 19646 + }, + { + "epoch": 0.4605566201937446, + "grad_norm": 0.21138685941696167, + "learning_rate": 0.00011239840477433829, + "loss": 0.049, + "step": 19647 + }, + { + "epoch": 0.4605800617685496, + "grad_norm": 0.1145833283662796, + "learning_rate": 0.00011239109541859397, + "loss": 0.0266, + "step": 19648 + }, + { + "epoch": 0.4606035033433546, + "grad_norm": 0.5218353271484375, + "learning_rate": 0.00011238378599561519, + "loss": 0.068, + "step": 19649 + }, + { + "epoch": 0.4606269449181596, + "grad_norm": 0.4615887403488159, + "learning_rate": 0.00011237647650544167, + "loss": 0.0601, + "step": 19650 + }, + { + "epoch": 0.4606503864929646, + "grad_norm": 0.7016150951385498, + "learning_rate": 0.00011236916694811306, + "loss": 0.2029, + "step": 19651 + }, + { + "epoch": 0.4606738280677696, + "grad_norm": 0.42041561007499695, + "learning_rate": 0.00011236185732366896, + "loss": 0.0781, + "step": 19652 + }, + { + "epoch": 0.46069726964257457, + "grad_norm": 0.440088152885437, + "learning_rate": 0.0001123545476321491, + "loss": 0.0772, + "step": 19653 + }, + { + "epoch": 0.46072071121737956, + "grad_norm": 0.3841603994369507, + "learning_rate": 0.00011234723787359311, + "loss": 0.05, + "step": 19654 + }, + { + "epoch": 0.46074415279218456, + "grad_norm": 0.6887404322624207, + "learning_rate": 0.00011233992804804063, + "loss": 0.0994, + "step": 19655 + }, + { + "epoch": 0.46076759436698955, + "grad_norm": 0.4007423222064972, + "learning_rate": 0.0001123326181555314, + "loss": 0.0636, + "step": 19656 + }, + { + "epoch": 0.46079103594179455, + "grad_norm": 0.5839460492134094, + "learning_rate": 0.00011232530819610502, + "loss": 0.1449, + "step": 19657 + }, + { + "epoch": 0.46081447751659954, + "grad_norm": 0.372008740901947, + "learning_rate": 0.00011231799816980117, + "loss": 0.06, + "step": 19658 + }, + { + "epoch": 0.46083791909140454, + "grad_norm": 0.6665142774581909, + "learning_rate": 0.00011231068807665955, + "loss": 0.1281, + "step": 19659 + }, + { + "epoch": 0.46086136066620953, + "grad_norm": 0.1191333681344986, + "learning_rate": 0.00011230337791671974, + "loss": 0.0192, + "step": 19660 + }, + { + "epoch": 0.4608848022410145, + "grad_norm": 0.662433385848999, + "learning_rate": 0.00011229606769002147, + "loss": 0.1044, + "step": 19661 + }, + { + "epoch": 0.4609082438158196, + "grad_norm": 0.3493807315826416, + "learning_rate": 0.00011228875739660438, + "loss": 0.0272, + "step": 19662 + }, + { + "epoch": 0.46093168539062457, + "grad_norm": 0.23160628974437714, + "learning_rate": 0.00011228144703650818, + "loss": 0.0242, + "step": 19663 + }, + { + "epoch": 0.46095512696542956, + "grad_norm": 0.7010703086853027, + "learning_rate": 0.0001122741366097725, + "loss": 0.6269, + "step": 19664 + }, + { + "epoch": 0.46097856854023456, + "grad_norm": 0.15780766308307648, + "learning_rate": 0.00011226682611643698, + "loss": 0.0265, + "step": 19665 + }, + { + "epoch": 0.46100201011503955, + "grad_norm": 0.5120252370834351, + "learning_rate": 0.00011225951555654133, + "loss": 0.0867, + "step": 19666 + }, + { + "epoch": 0.46102545168984455, + "grad_norm": 0.4342571496963501, + "learning_rate": 0.0001122522049301252, + "loss": 0.1081, + "step": 19667 + }, + { + "epoch": 0.46104889326464954, + "grad_norm": 0.1995570957660675, + "learning_rate": 0.00011224489423722826, + "loss": 0.019, + "step": 19668 + }, + { + "epoch": 0.46107233483945453, + "grad_norm": 0.4827476739883423, + "learning_rate": 0.00011223758347789019, + "loss": 0.1011, + "step": 19669 + }, + { + "epoch": 0.46109577641425953, + "grad_norm": 0.08465450257062912, + "learning_rate": 0.00011223027265215067, + "loss": 0.0156, + "step": 19670 + }, + { + "epoch": 0.4611192179890645, + "grad_norm": 1.3254847526550293, + "learning_rate": 0.0001122229617600493, + "loss": 0.104, + "step": 19671 + }, + { + "epoch": 0.4611426595638695, + "grad_norm": 0.24238555133342743, + "learning_rate": 0.00011221565080162582, + "loss": 0.2272, + "step": 19672 + }, + { + "epoch": 0.4611661011386745, + "grad_norm": 0.7571600079536438, + "learning_rate": 0.00011220833977691986, + "loss": 0.1204, + "step": 19673 + }, + { + "epoch": 0.4611895427134795, + "grad_norm": 0.5753772854804993, + "learning_rate": 0.00011220102868597111, + "loss": 0.1009, + "step": 19674 + }, + { + "epoch": 0.4612129842882845, + "grad_norm": 0.37253785133361816, + "learning_rate": 0.00011219371752881927, + "loss": 0.0559, + "step": 19675 + }, + { + "epoch": 0.4612364258630895, + "grad_norm": 0.7256549596786499, + "learning_rate": 0.00011218640630550394, + "loss": 0.1944, + "step": 19676 + }, + { + "epoch": 0.4612598674378945, + "grad_norm": 0.8691658973693848, + "learning_rate": 0.00011217909501606482, + "loss": 0.1297, + "step": 19677 + }, + { + "epoch": 0.4612833090126995, + "grad_norm": 0.5642176866531372, + "learning_rate": 0.00011217178366054162, + "loss": 0.0803, + "step": 19678 + }, + { + "epoch": 0.4613067505875045, + "grad_norm": 0.1869717687368393, + "learning_rate": 0.00011216447223897395, + "loss": 0.0236, + "step": 19679 + }, + { + "epoch": 0.46133019216230947, + "grad_norm": 0.4654639959335327, + "learning_rate": 0.00011215716075140155, + "loss": 0.0851, + "step": 19680 + }, + { + "epoch": 0.46135363373711447, + "grad_norm": 0.5296827554702759, + "learning_rate": 0.000112149849197864, + "loss": 0.5818, + "step": 19681 + }, + { + "epoch": 0.46137707531191946, + "grad_norm": 0.7225725054740906, + "learning_rate": 0.00011214253757840108, + "loss": 0.11, + "step": 19682 + }, + { + "epoch": 0.46140051688672445, + "grad_norm": 0.38451993465423584, + "learning_rate": 0.00011213522589305239, + "loss": 0.0634, + "step": 19683 + }, + { + "epoch": 0.46142395846152945, + "grad_norm": 0.6880707740783691, + "learning_rate": 0.0001121279141418576, + "loss": 0.0996, + "step": 19684 + }, + { + "epoch": 0.46144740003633444, + "grad_norm": 0.09412284940481186, + "learning_rate": 0.00011212060232485642, + "loss": 0.015, + "step": 19685 + }, + { + "epoch": 0.46147084161113944, + "grad_norm": 0.31398552656173706, + "learning_rate": 0.00011211329044208851, + "loss": 0.0473, + "step": 19686 + }, + { + "epoch": 0.46149428318594443, + "grad_norm": 0.3052680790424347, + "learning_rate": 0.00011210597849359355, + "loss": 0.0788, + "step": 19687 + }, + { + "epoch": 0.4615177247607494, + "grad_norm": 0.5240883827209473, + "learning_rate": 0.00011209866647941121, + "loss": 0.0354, + "step": 19688 + }, + { + "epoch": 0.4615411663355544, + "grad_norm": 0.7180687785148621, + "learning_rate": 0.00011209135439958118, + "loss": 0.201, + "step": 19689 + }, + { + "epoch": 0.4615646079103594, + "grad_norm": 0.604925811290741, + "learning_rate": 0.0001120840422541431, + "loss": 0.1215, + "step": 19690 + }, + { + "epoch": 0.4615880494851644, + "grad_norm": 0.28236398100852966, + "learning_rate": 0.00011207673004313666, + "loss": 0.0271, + "step": 19691 + }, + { + "epoch": 0.4616114910599694, + "grad_norm": 0.1726970374584198, + "learning_rate": 0.00011206941776660159, + "loss": 0.0466, + "step": 19692 + }, + { + "epoch": 0.4616349326347744, + "grad_norm": 0.14076419174671173, + "learning_rate": 0.00011206210542457745, + "loss": 0.0231, + "step": 19693 + }, + { + "epoch": 0.4616583742095794, + "grad_norm": 0.39034780859947205, + "learning_rate": 0.00011205479301710405, + "loss": 0.0589, + "step": 19694 + }, + { + "epoch": 0.4616818157843844, + "grad_norm": 0.2587895095348358, + "learning_rate": 0.00011204748054422099, + "loss": 0.0532, + "step": 19695 + }, + { + "epoch": 0.4617052573591894, + "grad_norm": 0.569381058216095, + "learning_rate": 0.00011204016800596792, + "loss": 0.1467, + "step": 19696 + }, + { + "epoch": 0.4617286989339944, + "grad_norm": 0.6164261698722839, + "learning_rate": 0.0001120328554023846, + "loss": 0.5243, + "step": 19697 + }, + { + "epoch": 0.46175214050879937, + "grad_norm": 0.348213791847229, + "learning_rate": 0.00011202554273351065, + "loss": 0.053, + "step": 19698 + }, + { + "epoch": 0.46177558208360436, + "grad_norm": 0.30403780937194824, + "learning_rate": 0.0001120182299993858, + "loss": 0.0592, + "step": 19699 + }, + { + "epoch": 0.46179902365840936, + "grad_norm": 0.16864068806171417, + "learning_rate": 0.00011201091720004965, + "loss": 0.0261, + "step": 19700 + }, + { + "epoch": 0.46182246523321435, + "grad_norm": 0.710660457611084, + "learning_rate": 0.00011200360433554195, + "loss": 0.1595, + "step": 19701 + }, + { + "epoch": 0.46184590680801935, + "grad_norm": 0.6275258660316467, + "learning_rate": 0.00011199629140590238, + "loss": 0.1126, + "step": 19702 + }, + { + "epoch": 0.46186934838282434, + "grad_norm": 0.25744104385375977, + "learning_rate": 0.00011198897841117054, + "loss": 0.0386, + "step": 19703 + }, + { + "epoch": 0.46189278995762933, + "grad_norm": 0.4324868321418762, + "learning_rate": 0.0001119816653513862, + "loss": 0.1271, + "step": 19704 + }, + { + "epoch": 0.46191623153243433, + "grad_norm": 0.05458454787731171, + "learning_rate": 0.00011197435222658901, + "loss": 0.0034, + "step": 19705 + }, + { + "epoch": 0.4619396731072393, + "grad_norm": 0.8551275730133057, + "learning_rate": 0.00011196703903681863, + "loss": 0.1612, + "step": 19706 + }, + { + "epoch": 0.4619631146820443, + "grad_norm": 0.5047973990440369, + "learning_rate": 0.00011195972578211478, + "loss": 0.1262, + "step": 19707 + }, + { + "epoch": 0.4619865562568493, + "grad_norm": 0.1496276557445526, + "learning_rate": 0.0001119524124625171, + "loss": 0.0265, + "step": 19708 + }, + { + "epoch": 0.4620099978316543, + "grad_norm": 0.3387342691421509, + "learning_rate": 0.00011194509907806528, + "loss": 0.0709, + "step": 19709 + }, + { + "epoch": 0.4620334394064593, + "grad_norm": 0.25369107723236084, + "learning_rate": 0.00011193778562879905, + "loss": 0.027, + "step": 19710 + }, + { + "epoch": 0.4620568809812643, + "grad_norm": 0.33568987250328064, + "learning_rate": 0.00011193047211475806, + "loss": 0.2006, + "step": 19711 + }, + { + "epoch": 0.4620803225560693, + "grad_norm": 0.7805234789848328, + "learning_rate": 0.00011192315853598197, + "loss": 0.1638, + "step": 19712 + }, + { + "epoch": 0.46210376413087434, + "grad_norm": 0.7480425834655762, + "learning_rate": 0.00011191584489251053, + "loss": 0.1317, + "step": 19713 + }, + { + "epoch": 0.46212720570567933, + "grad_norm": 1.4849518537521362, + "learning_rate": 0.00011190853118438333, + "loss": 0.1409, + "step": 19714 + }, + { + "epoch": 0.46215064728048433, + "grad_norm": 0.34906935691833496, + "learning_rate": 0.00011190121741164012, + "loss": 0.0444, + "step": 19715 + }, + { + "epoch": 0.4621740888552893, + "grad_norm": 0.6302060484886169, + "learning_rate": 0.00011189390357432057, + "loss": 0.6821, + "step": 19716 + }, + { + "epoch": 0.4621975304300943, + "grad_norm": 0.1151827797293663, + "learning_rate": 0.00011188658967246439, + "loss": 0.0218, + "step": 19717 + }, + { + "epoch": 0.4622209720048993, + "grad_norm": 0.5141693949699402, + "learning_rate": 0.00011187927570611122, + "loss": 0.0934, + "step": 19718 + }, + { + "epoch": 0.4622444135797043, + "grad_norm": 0.5105534791946411, + "learning_rate": 0.00011187196167530077, + "loss": 0.0891, + "step": 19719 + }, + { + "epoch": 0.4622678551545093, + "grad_norm": 0.37621578574180603, + "learning_rate": 0.00011186464758007271, + "loss": 0.0895, + "step": 19720 + }, + { + "epoch": 0.4622912967293143, + "grad_norm": 0.5562925338745117, + "learning_rate": 0.00011185733342046674, + "loss": 0.1143, + "step": 19721 + }, + { + "epoch": 0.4623147383041193, + "grad_norm": 0.2740859389305115, + "learning_rate": 0.00011185001919652254, + "loss": 0.064, + "step": 19722 + }, + { + "epoch": 0.4623381798789243, + "grad_norm": 0.5259649157524109, + "learning_rate": 0.00011184270490827982, + "loss": 0.0849, + "step": 19723 + }, + { + "epoch": 0.4623616214537293, + "grad_norm": 0.30758705735206604, + "learning_rate": 0.00011183539055577826, + "loss": 0.0433, + "step": 19724 + }, + { + "epoch": 0.46238506302853427, + "grad_norm": 0.117362841963768, + "learning_rate": 0.0001118280761390575, + "loss": 0.0303, + "step": 19725 + }, + { + "epoch": 0.46240850460333927, + "grad_norm": 0.24593548476696014, + "learning_rate": 0.00011182076165815732, + "loss": 0.0358, + "step": 19726 + }, + { + "epoch": 0.46243194617814426, + "grad_norm": 0.2395251840353012, + "learning_rate": 0.0001118134471131173, + "loss": 0.052, + "step": 19727 + }, + { + "epoch": 0.46245538775294925, + "grad_norm": 0.3077397048473358, + "learning_rate": 0.0001118061325039772, + "loss": 0.0135, + "step": 19728 + }, + { + "epoch": 0.46247882932775425, + "grad_norm": 0.2903604805469513, + "learning_rate": 0.00011179881783077671, + "loss": 0.0506, + "step": 19729 + }, + { + "epoch": 0.46250227090255924, + "grad_norm": 0.367341548204422, + "learning_rate": 0.0001117915030935555, + "loss": 0.0564, + "step": 19730 + }, + { + "epoch": 0.46252571247736424, + "grad_norm": 0.6129302978515625, + "learning_rate": 0.00011178418829235324, + "loss": 0.1434, + "step": 19731 + }, + { + "epoch": 0.46254915405216923, + "grad_norm": 0.1746271401643753, + "learning_rate": 0.00011177687342720969, + "loss": 0.0261, + "step": 19732 + }, + { + "epoch": 0.4625725956269742, + "grad_norm": 0.42454037070274353, + "learning_rate": 0.00011176955849816444, + "loss": 0.1003, + "step": 19733 + }, + { + "epoch": 0.4625960372017792, + "grad_norm": 0.634913444519043, + "learning_rate": 0.00011176224350525728, + "loss": 0.5518, + "step": 19734 + }, + { + "epoch": 0.4626194787765842, + "grad_norm": 0.347402960062027, + "learning_rate": 0.00011175492844852782, + "loss": 0.0532, + "step": 19735 + }, + { + "epoch": 0.4626429203513892, + "grad_norm": 0.18201789259910583, + "learning_rate": 0.0001117476133280158, + "loss": 0.027, + "step": 19736 + }, + { + "epoch": 0.4626663619261942, + "grad_norm": 0.3782779574394226, + "learning_rate": 0.00011174029814376092, + "loss": 0.0606, + "step": 19737 + }, + { + "epoch": 0.4626898035009992, + "grad_norm": 0.46106091141700745, + "learning_rate": 0.00011173298289580283, + "loss": 0.0641, + "step": 19738 + }, + { + "epoch": 0.4627132450758042, + "grad_norm": 0.527888834476471, + "learning_rate": 0.00011172566758418122, + "loss": 0.0947, + "step": 19739 + }, + { + "epoch": 0.4627366866506092, + "grad_norm": 0.2592986524105072, + "learning_rate": 0.00011171835220893585, + "loss": 0.031, + "step": 19740 + }, + { + "epoch": 0.4627601282254142, + "grad_norm": 0.3872702419757843, + "learning_rate": 0.00011171103677010634, + "loss": 0.0788, + "step": 19741 + }, + { + "epoch": 0.4627835698002192, + "grad_norm": 0.47570255398750305, + "learning_rate": 0.00011170372126773245, + "loss": 0.0597, + "step": 19742 + }, + { + "epoch": 0.46280701137502417, + "grad_norm": 0.541289746761322, + "learning_rate": 0.00011169640570185382, + "loss": 0.114, + "step": 19743 + }, + { + "epoch": 0.46283045294982916, + "grad_norm": 0.6058759093284607, + "learning_rate": 0.00011168909007251015, + "loss": 0.0903, + "step": 19744 + }, + { + "epoch": 0.46285389452463416, + "grad_norm": 0.2497289478778839, + "learning_rate": 0.00011168177437974116, + "loss": 0.0165, + "step": 19745 + }, + { + "epoch": 0.46287733609943915, + "grad_norm": 0.4785565137863159, + "learning_rate": 0.00011167445862358657, + "loss": 0.0328, + "step": 19746 + }, + { + "epoch": 0.46290077767424415, + "grad_norm": 0.6283652782440186, + "learning_rate": 0.000111667142804086, + "loss": 0.6744, + "step": 19747 + }, + { + "epoch": 0.46292421924904914, + "grad_norm": 0.7914628386497498, + "learning_rate": 0.00011165982692127918, + "loss": 0.0771, + "step": 19748 + }, + { + "epoch": 0.46294766082385413, + "grad_norm": 0.6815500259399414, + "learning_rate": 0.00011165251097520584, + "loss": 0.7504, + "step": 19749 + }, + { + "epoch": 0.46297110239865913, + "grad_norm": 0.8683542013168335, + "learning_rate": 0.00011164519496590562, + "loss": 0.0961, + "step": 19750 + }, + { + "epoch": 0.4629945439734641, + "grad_norm": 0.6162951588630676, + "learning_rate": 0.00011163787889341825, + "loss": 0.1189, + "step": 19751 + }, + { + "epoch": 0.4630179855482691, + "grad_norm": 0.8390434384346008, + "learning_rate": 0.00011163056275778343, + "loss": 0.1632, + "step": 19752 + }, + { + "epoch": 0.4630414271230741, + "grad_norm": 0.38164690136909485, + "learning_rate": 0.00011162324655904086, + "loss": 0.0637, + "step": 19753 + }, + { + "epoch": 0.4630648686978791, + "grad_norm": 0.5852351188659668, + "learning_rate": 0.00011161593029723021, + "loss": 0.1395, + "step": 19754 + }, + { + "epoch": 0.4630883102726841, + "grad_norm": 0.4880880117416382, + "learning_rate": 0.00011160861397239122, + "loss": 0.084, + "step": 19755 + }, + { + "epoch": 0.4631117518474891, + "grad_norm": 0.35837090015411377, + "learning_rate": 0.00011160129758456355, + "loss": 0.0519, + "step": 19756 + }, + { + "epoch": 0.4631351934222941, + "grad_norm": 0.48842716217041016, + "learning_rate": 0.0001115939811337869, + "loss": 0.1189, + "step": 19757 + }, + { + "epoch": 0.4631586349970991, + "grad_norm": 0.385979026556015, + "learning_rate": 0.00011158666462010096, + "loss": 0.0649, + "step": 19758 + }, + { + "epoch": 0.4631820765719041, + "grad_norm": 0.1864292025566101, + "learning_rate": 0.0001115793480435455, + "loss": 0.036, + "step": 19759 + }, + { + "epoch": 0.46320551814670907, + "grad_norm": 0.17766359448432922, + "learning_rate": 0.00011157203140416015, + "loss": 0.0464, + "step": 19760 + }, + { + "epoch": 0.46322895972151407, + "grad_norm": 0.6099346876144409, + "learning_rate": 0.00011156471470198464, + "loss": 0.0858, + "step": 19761 + }, + { + "epoch": 0.46325240129631906, + "grad_norm": 0.06848588585853577, + "learning_rate": 0.00011155739793705869, + "loss": 0.0056, + "step": 19762 + }, + { + "epoch": 0.46327584287112406, + "grad_norm": 0.5501253604888916, + "learning_rate": 0.00011155008110942191, + "loss": 0.0733, + "step": 19763 + }, + { + "epoch": 0.4632992844459291, + "grad_norm": 0.122245192527771, + "learning_rate": 0.0001115427642191141, + "loss": 0.0134, + "step": 19764 + }, + { + "epoch": 0.4633227260207341, + "grad_norm": 0.6765368580818176, + "learning_rate": 0.00011153544726617492, + "loss": 0.0966, + "step": 19765 + }, + { + "epoch": 0.4633461675955391, + "grad_norm": 0.4598994255065918, + "learning_rate": 0.00011152813025064409, + "loss": 0.0892, + "step": 19766 + }, + { + "epoch": 0.4633696091703441, + "grad_norm": 0.557113766670227, + "learning_rate": 0.00011152081317256129, + "loss": 0.1334, + "step": 19767 + }, + { + "epoch": 0.4633930507451491, + "grad_norm": 0.5878928303718567, + "learning_rate": 0.00011151349603196626, + "loss": 0.1158, + "step": 19768 + }, + { + "epoch": 0.4634164923199541, + "grad_norm": 0.22453218698501587, + "learning_rate": 0.00011150617882889864, + "loss": 0.0431, + "step": 19769 + }, + { + "epoch": 0.46343993389475907, + "grad_norm": 0.4358481466770172, + "learning_rate": 0.00011149886156339818, + "loss": 0.0883, + "step": 19770 + }, + { + "epoch": 0.46346337546956407, + "grad_norm": 0.5161450505256653, + "learning_rate": 0.00011149154423550458, + "loss": 0.0769, + "step": 19771 + }, + { + "epoch": 0.46348681704436906, + "grad_norm": 0.6931847333908081, + "learning_rate": 0.00011148422684525755, + "loss": 0.1122, + "step": 19772 + }, + { + "epoch": 0.46351025861917405, + "grad_norm": 0.7367222905158997, + "learning_rate": 0.00011147690939269675, + "loss": 0.0666, + "step": 19773 + }, + { + "epoch": 0.46353370019397905, + "grad_norm": 0.09834793210029602, + "learning_rate": 0.00011146959187786195, + "loss": 0.0178, + "step": 19774 + }, + { + "epoch": 0.46355714176878404, + "grad_norm": 0.8055204749107361, + "learning_rate": 0.00011146227430079284, + "loss": 0.7134, + "step": 19775 + }, + { + "epoch": 0.46358058334358904, + "grad_norm": 0.5910544395446777, + "learning_rate": 0.00011145495666152903, + "loss": 0.1298, + "step": 19776 + }, + { + "epoch": 0.46360402491839403, + "grad_norm": 0.3308093249797821, + "learning_rate": 0.00011144763896011037, + "loss": 0.0754, + "step": 19777 + }, + { + "epoch": 0.463627466493199, + "grad_norm": 0.4088188111782074, + "learning_rate": 0.00011144032119657647, + "loss": 0.0904, + "step": 19778 + }, + { + "epoch": 0.463650908068004, + "grad_norm": 0.4568464457988739, + "learning_rate": 0.00011143300337096707, + "loss": 0.0631, + "step": 19779 + }, + { + "epoch": 0.463674349642809, + "grad_norm": 0.416240930557251, + "learning_rate": 0.00011142568548332189, + "loss": 0.5654, + "step": 19780 + }, + { + "epoch": 0.463697791217614, + "grad_norm": 0.43873727321624756, + "learning_rate": 0.0001114183675336806, + "loss": 0.0623, + "step": 19781 + }, + { + "epoch": 0.463721232792419, + "grad_norm": 0.8654210567474365, + "learning_rate": 0.00011141104952208292, + "loss": 0.159, + "step": 19782 + }, + { + "epoch": 0.463744674367224, + "grad_norm": 0.6611652374267578, + "learning_rate": 0.00011140373144856857, + "loss": 0.5974, + "step": 19783 + }, + { + "epoch": 0.463768115942029, + "grad_norm": 0.5146528482437134, + "learning_rate": 0.00011139641331317728, + "loss": 0.0821, + "step": 19784 + }, + { + "epoch": 0.463791557516834, + "grad_norm": 0.6245187520980835, + "learning_rate": 0.00011138909511594869, + "loss": 0.121, + "step": 19785 + }, + { + "epoch": 0.463814999091639, + "grad_norm": 0.6135607361793518, + "learning_rate": 0.00011138177685692256, + "loss": 0.0842, + "step": 19786 + }, + { + "epoch": 0.463838440666444, + "grad_norm": 0.11993974447250366, + "learning_rate": 0.00011137445853613863, + "loss": 0.0188, + "step": 19787 + }, + { + "epoch": 0.46386188224124897, + "grad_norm": 0.37878134846687317, + "learning_rate": 0.00011136714015363652, + "loss": 0.085, + "step": 19788 + }, + { + "epoch": 0.46388532381605396, + "grad_norm": 0.2947389781475067, + "learning_rate": 0.000111359821709456, + "loss": 0.0549, + "step": 19789 + }, + { + "epoch": 0.46390876539085896, + "grad_norm": 0.20713062584400177, + "learning_rate": 0.00011135250320363677, + "loss": 0.0297, + "step": 19790 + }, + { + "epoch": 0.46393220696566395, + "grad_norm": 0.8841922283172607, + "learning_rate": 0.00011134518463621857, + "loss": 0.609, + "step": 19791 + }, + { + "epoch": 0.46395564854046895, + "grad_norm": 0.27628293633461, + "learning_rate": 0.00011133786600724104, + "loss": 0.055, + "step": 19792 + }, + { + "epoch": 0.46397909011527394, + "grad_norm": 0.455228716135025, + "learning_rate": 0.00011133054731674395, + "loss": 0.0613, + "step": 19793 + }, + { + "epoch": 0.46400253169007893, + "grad_norm": 0.506056547164917, + "learning_rate": 0.00011132322856476696, + "loss": 0.1001, + "step": 19794 + }, + { + "epoch": 0.46402597326488393, + "grad_norm": 0.5537370443344116, + "learning_rate": 0.00011131590975134983, + "loss": 0.0876, + "step": 19795 + }, + { + "epoch": 0.4640494148396889, + "grad_norm": 0.421577513217926, + "learning_rate": 0.00011130859087653227, + "loss": 0.0622, + "step": 19796 + }, + { + "epoch": 0.4640728564144939, + "grad_norm": 0.6258364319801331, + "learning_rate": 0.00011130127194035398, + "loss": 0.0512, + "step": 19797 + }, + { + "epoch": 0.4640962979892989, + "grad_norm": 1.6011254787445068, + "learning_rate": 0.00011129395294285465, + "loss": 0.1903, + "step": 19798 + }, + { + "epoch": 0.4641197395641039, + "grad_norm": 0.46755337715148926, + "learning_rate": 0.00011128663388407403, + "loss": 0.0684, + "step": 19799 + }, + { + "epoch": 0.4641431811389089, + "grad_norm": 0.4011356830596924, + "learning_rate": 0.0001112793147640518, + "loss": 0.0365, + "step": 19800 + }, + { + "epoch": 0.4641666227137139, + "grad_norm": 0.5965684652328491, + "learning_rate": 0.0001112719955828277, + "loss": 0.0661, + "step": 19801 + }, + { + "epoch": 0.4641900642885189, + "grad_norm": 0.44568097591400146, + "learning_rate": 0.00011126467634044143, + "loss": 0.1049, + "step": 19802 + }, + { + "epoch": 0.4642135058633239, + "grad_norm": 0.5855648517608643, + "learning_rate": 0.00011125735703693272, + "loss": 0.0715, + "step": 19803 + }, + { + "epoch": 0.4642369474381289, + "grad_norm": 0.4227292537689209, + "learning_rate": 0.00011125003767234126, + "loss": 0.0787, + "step": 19804 + }, + { + "epoch": 0.46426038901293387, + "grad_norm": 0.15599681437015533, + "learning_rate": 0.0001112427182467068, + "loss": 0.0292, + "step": 19805 + }, + { + "epoch": 0.46428383058773887, + "grad_norm": 0.8978458046913147, + "learning_rate": 0.000111235398760069, + "loss": 0.0872, + "step": 19806 + }, + { + "epoch": 0.46430727216254386, + "grad_norm": 0.09129626303911209, + "learning_rate": 0.00011122807921246767, + "loss": 0.0146, + "step": 19807 + }, + { + "epoch": 0.46433071373734885, + "grad_norm": 0.3352244198322296, + "learning_rate": 0.0001112207596039424, + "loss": 0.0889, + "step": 19808 + }, + { + "epoch": 0.46435415531215385, + "grad_norm": 0.3908972442150116, + "learning_rate": 0.00011121343993453304, + "loss": 0.4957, + "step": 19809 + }, + { + "epoch": 0.46437759688695884, + "grad_norm": 0.631474494934082, + "learning_rate": 0.00011120612020427921, + "loss": 0.1444, + "step": 19810 + }, + { + "epoch": 0.46440103846176384, + "grad_norm": 0.5450695157051086, + "learning_rate": 0.00011119880041322064, + "loss": 0.786, + "step": 19811 + }, + { + "epoch": 0.46442448003656883, + "grad_norm": 0.5094525218009949, + "learning_rate": 0.00011119148056139706, + "loss": 0.0856, + "step": 19812 + }, + { + "epoch": 0.4644479216113738, + "grad_norm": 0.3844059705734253, + "learning_rate": 0.0001111841606488482, + "loss": 0.0331, + "step": 19813 + }, + { + "epoch": 0.4644713631861788, + "grad_norm": 0.1929774135351181, + "learning_rate": 0.00011117684067561378, + "loss": 0.0546, + "step": 19814 + }, + { + "epoch": 0.46449480476098387, + "grad_norm": 0.5697033405303955, + "learning_rate": 0.0001111695206417335, + "loss": 0.1537, + "step": 19815 + }, + { + "epoch": 0.46451824633578886, + "grad_norm": 0.244686096906662, + "learning_rate": 0.00011116220054724711, + "loss": 0.0405, + "step": 19816 + }, + { + "epoch": 0.46454168791059386, + "grad_norm": 0.494617223739624, + "learning_rate": 0.00011115488039219429, + "loss": 0.0847, + "step": 19817 + }, + { + "epoch": 0.46456512948539885, + "grad_norm": 0.1857261210680008, + "learning_rate": 0.00011114756017661474, + "loss": 0.03, + "step": 19818 + }, + { + "epoch": 0.46458857106020385, + "grad_norm": 0.5420560836791992, + "learning_rate": 0.00011114023990054825, + "loss": 0.1608, + "step": 19819 + }, + { + "epoch": 0.46461201263500884, + "grad_norm": 0.11159572005271912, + "learning_rate": 0.00011113291956403453, + "loss": 0.0225, + "step": 19820 + }, + { + "epoch": 0.46463545420981384, + "grad_norm": 0.37686389684677124, + "learning_rate": 0.00011112559916711323, + "loss": 0.0651, + "step": 19821 + }, + { + "epoch": 0.46465889578461883, + "grad_norm": 0.3384087085723877, + "learning_rate": 0.00011111827870982414, + "loss": 0.0884, + "step": 19822 + }, + { + "epoch": 0.4646823373594238, + "grad_norm": 0.1383894979953766, + "learning_rate": 0.00011111095819220693, + "loss": 0.0163, + "step": 19823 + }, + { + "epoch": 0.4647057789342288, + "grad_norm": 0.6718759536743164, + "learning_rate": 0.00011110363761430136, + "loss": 0.1853, + "step": 19824 + }, + { + "epoch": 0.4647292205090338, + "grad_norm": 0.15141016244888306, + "learning_rate": 0.00011109631697614717, + "loss": 0.0178, + "step": 19825 + }, + { + "epoch": 0.4647526620838388, + "grad_norm": 0.37872129678726196, + "learning_rate": 0.00011108899627778404, + "loss": 0.0803, + "step": 19826 + }, + { + "epoch": 0.4647761036586438, + "grad_norm": 0.4696303904056549, + "learning_rate": 0.00011108167551925168, + "loss": 0.7208, + "step": 19827 + }, + { + "epoch": 0.4647995452334488, + "grad_norm": 0.17446269094944, + "learning_rate": 0.00011107435470058986, + "loss": 0.0278, + "step": 19828 + }, + { + "epoch": 0.4648229868082538, + "grad_norm": 0.5887683033943176, + "learning_rate": 0.0001110670338218383, + "loss": 0.1278, + "step": 19829 + }, + { + "epoch": 0.4648464283830588, + "grad_norm": 0.4045616388320923, + "learning_rate": 0.00011105971288303663, + "loss": 0.0473, + "step": 19830 + }, + { + "epoch": 0.4648698699578638, + "grad_norm": 0.4164351224899292, + "learning_rate": 0.00011105239188422472, + "loss": 0.0455, + "step": 19831 + }, + { + "epoch": 0.4648933115326688, + "grad_norm": 1.1216298341751099, + "learning_rate": 0.0001110450708254422, + "loss": 0.1827, + "step": 19832 + }, + { + "epoch": 0.46491675310747377, + "grad_norm": 0.4249085783958435, + "learning_rate": 0.0001110377497067288, + "loss": 0.1248, + "step": 19833 + }, + { + "epoch": 0.46494019468227876, + "grad_norm": 0.597198486328125, + "learning_rate": 0.00011103042852812427, + "loss": 0.1035, + "step": 19834 + }, + { + "epoch": 0.46496363625708376, + "grad_norm": 0.6281474232673645, + "learning_rate": 0.00011102310728966834, + "loss": 0.6306, + "step": 19835 + }, + { + "epoch": 0.46498707783188875, + "grad_norm": 0.5462809801101685, + "learning_rate": 0.00011101578599140067, + "loss": 0.1053, + "step": 19836 + }, + { + "epoch": 0.46501051940669375, + "grad_norm": 0.5779865980148315, + "learning_rate": 0.00011100846463336107, + "loss": 0.0856, + "step": 19837 + }, + { + "epoch": 0.46503396098149874, + "grad_norm": 0.6656907200813293, + "learning_rate": 0.00011100114321558925, + "loss": 0.6233, + "step": 19838 + }, + { + "epoch": 0.46505740255630373, + "grad_norm": 0.47982168197631836, + "learning_rate": 0.00011099382173812492, + "loss": 0.1192, + "step": 19839 + }, + { + "epoch": 0.46508084413110873, + "grad_norm": 0.8557499647140503, + "learning_rate": 0.00011098650020100775, + "loss": 0.1121, + "step": 19840 + }, + { + "epoch": 0.4651042857059137, + "grad_norm": 0.2635098695755005, + "learning_rate": 0.00011097917860427758, + "loss": 0.0539, + "step": 19841 + }, + { + "epoch": 0.4651277272807187, + "grad_norm": 0.44246000051498413, + "learning_rate": 0.00011097185694797403, + "loss": 0.499, + "step": 19842 + }, + { + "epoch": 0.4651511688555237, + "grad_norm": 0.27065858244895935, + "learning_rate": 0.00011096453523213688, + "loss": 0.0342, + "step": 19843 + }, + { + "epoch": 0.4651746104303287, + "grad_norm": 0.6273062825202942, + "learning_rate": 0.00011095721345680588, + "loss": 0.1546, + "step": 19844 + }, + { + "epoch": 0.4651980520051337, + "grad_norm": 0.350779265165329, + "learning_rate": 0.00011094989162202075, + "loss": 0.0427, + "step": 19845 + }, + { + "epoch": 0.4652214935799387, + "grad_norm": 0.37674620747566223, + "learning_rate": 0.00011094256972782115, + "loss": 0.0611, + "step": 19846 + }, + { + "epoch": 0.4652449351547437, + "grad_norm": 0.3510473966598511, + "learning_rate": 0.00011093524777424689, + "loss": 0.0789, + "step": 19847 + }, + { + "epoch": 0.4652683767295487, + "grad_norm": 0.5381286144256592, + "learning_rate": 0.00011092792576133766, + "loss": 0.1227, + "step": 19848 + }, + { + "epoch": 0.4652918183043537, + "grad_norm": 0.6532329320907593, + "learning_rate": 0.00011092060368913318, + "loss": 0.1039, + "step": 19849 + }, + { + "epoch": 0.46531525987915867, + "grad_norm": 0.5138506889343262, + "learning_rate": 0.00011091328155767323, + "loss": 0.1483, + "step": 19850 + }, + { + "epoch": 0.46533870145396367, + "grad_norm": 0.5916589498519897, + "learning_rate": 0.00011090595936699747, + "loss": 0.1554, + "step": 19851 + }, + { + "epoch": 0.46536214302876866, + "grad_norm": 0.19797521829605103, + "learning_rate": 0.0001108986371171457, + "loss": 0.0246, + "step": 19852 + }, + { + "epoch": 0.46538558460357365, + "grad_norm": 0.36232784390449524, + "learning_rate": 0.00011089131480815762, + "loss": 0.0628, + "step": 19853 + }, + { + "epoch": 0.46540902617837865, + "grad_norm": 0.4822314381599426, + "learning_rate": 0.00011088399244007295, + "loss": 0.0744, + "step": 19854 + }, + { + "epoch": 0.46543246775318364, + "grad_norm": 0.17336691915988922, + "learning_rate": 0.00011087667001293142, + "loss": 0.0431, + "step": 19855 + }, + { + "epoch": 0.46545590932798864, + "grad_norm": 0.2171989232301712, + "learning_rate": 0.00011086934752677277, + "loss": 0.0537, + "step": 19856 + }, + { + "epoch": 0.46547935090279363, + "grad_norm": 0.21047809720039368, + "learning_rate": 0.00011086202498163676, + "loss": 0.0248, + "step": 19857 + }, + { + "epoch": 0.4655027924775986, + "grad_norm": 0.08705263584852219, + "learning_rate": 0.0001108547023775631, + "loss": 0.0172, + "step": 19858 + }, + { + "epoch": 0.4655262340524036, + "grad_norm": 0.20299668610095978, + "learning_rate": 0.00011084737971459147, + "loss": 0.0476, + "step": 19859 + }, + { + "epoch": 0.4655496756272086, + "grad_norm": 0.5863640308380127, + "learning_rate": 0.0001108400569927617, + "loss": 0.1681, + "step": 19860 + }, + { + "epoch": 0.4655731172020136, + "grad_norm": 0.4481906294822693, + "learning_rate": 0.00011083273421211347, + "loss": 0.1257, + "step": 19861 + }, + { + "epoch": 0.4655965587768186, + "grad_norm": 0.5501872897148132, + "learning_rate": 0.0001108254113726865, + "loss": 0.3075, + "step": 19862 + }, + { + "epoch": 0.4656200003516236, + "grad_norm": 0.4329565167427063, + "learning_rate": 0.00011081808847452055, + "loss": 0.088, + "step": 19863 + }, + { + "epoch": 0.4656434419264286, + "grad_norm": 0.500991702079773, + "learning_rate": 0.00011081076551765537, + "loss": 0.1102, + "step": 19864 + }, + { + "epoch": 0.4656668835012336, + "grad_norm": 0.6144776940345764, + "learning_rate": 0.00011080344250213064, + "loss": 0.0996, + "step": 19865 + }, + { + "epoch": 0.46569032507603864, + "grad_norm": 0.476589173078537, + "learning_rate": 0.00011079611942798615, + "loss": 0.5983, + "step": 19866 + }, + { + "epoch": 0.46571376665084363, + "grad_norm": 0.45906054973602295, + "learning_rate": 0.00011078879629526159, + "loss": 0.0951, + "step": 19867 + }, + { + "epoch": 0.4657372082256486, + "grad_norm": 0.28828078508377075, + "learning_rate": 0.00011078147310399672, + "loss": 0.0609, + "step": 19868 + }, + { + "epoch": 0.4657606498004536, + "grad_norm": 0.6051353216171265, + "learning_rate": 0.00011077414985423129, + "loss": 0.1773, + "step": 19869 + }, + { + "epoch": 0.4657840913752586, + "grad_norm": 0.31626519560813904, + "learning_rate": 0.00011076682654600504, + "loss": 0.0305, + "step": 19870 + }, + { + "epoch": 0.4658075329500636, + "grad_norm": 0.21596160531044006, + "learning_rate": 0.00011075950317935764, + "loss": 0.0356, + "step": 19871 + }, + { + "epoch": 0.4658309745248686, + "grad_norm": 0.07769156992435455, + "learning_rate": 0.00011075217975432892, + "loss": 0.0107, + "step": 19872 + }, + { + "epoch": 0.4658544160996736, + "grad_norm": 0.5457080602645874, + "learning_rate": 0.00011074485627095853, + "loss": 0.0859, + "step": 19873 + }, + { + "epoch": 0.4658778576744786, + "grad_norm": 0.607674241065979, + "learning_rate": 0.00011073753272928627, + "loss": 0.1182, + "step": 19874 + }, + { + "epoch": 0.4659012992492836, + "grad_norm": 0.6847226619720459, + "learning_rate": 0.00011073020912935183, + "loss": 0.134, + "step": 19875 + }, + { + "epoch": 0.4659247408240886, + "grad_norm": 0.6558052897453308, + "learning_rate": 0.000110722885471195, + "loss": 0.1603, + "step": 19876 + }, + { + "epoch": 0.4659481823988936, + "grad_norm": 0.31145554780960083, + "learning_rate": 0.00011071556175485551, + "loss": 0.0347, + "step": 19877 + }, + { + "epoch": 0.46597162397369857, + "grad_norm": 0.42221513390541077, + "learning_rate": 0.00011070823798037304, + "loss": 0.0558, + "step": 19878 + }, + { + "epoch": 0.46599506554850356, + "grad_norm": 0.3004206717014313, + "learning_rate": 0.00011070091414778738, + "loss": 0.0534, + "step": 19879 + }, + { + "epoch": 0.46601850712330856, + "grad_norm": 0.36848875880241394, + "learning_rate": 0.00011069359025713828, + "loss": 0.0814, + "step": 19880 + }, + { + "epoch": 0.46604194869811355, + "grad_norm": 0.6142743229866028, + "learning_rate": 0.00011068626630846541, + "loss": 0.2166, + "step": 19881 + }, + { + "epoch": 0.46606539027291854, + "grad_norm": 0.5633513927459717, + "learning_rate": 0.00011067894230180862, + "loss": 0.1097, + "step": 19882 + }, + { + "epoch": 0.46608883184772354, + "grad_norm": 0.4055306911468506, + "learning_rate": 0.00011067161823720755, + "loss": 0.0588, + "step": 19883 + }, + { + "epoch": 0.46611227342252853, + "grad_norm": 0.5682446956634521, + "learning_rate": 0.00011066429411470198, + "loss": 0.0742, + "step": 19884 + }, + { + "epoch": 0.46613571499733353, + "grad_norm": 0.30532217025756836, + "learning_rate": 0.00011065696993433166, + "loss": 0.0523, + "step": 19885 + }, + { + "epoch": 0.4661591565721385, + "grad_norm": 0.3682006597518921, + "learning_rate": 0.00011064964569613633, + "loss": 0.0502, + "step": 19886 + }, + { + "epoch": 0.4661825981469435, + "grad_norm": 0.6343429684638977, + "learning_rate": 0.00011064232140015569, + "loss": 0.0841, + "step": 19887 + }, + { + "epoch": 0.4662060397217485, + "grad_norm": 0.2862553596496582, + "learning_rate": 0.00011063499704642956, + "loss": 0.0577, + "step": 19888 + }, + { + "epoch": 0.4662294812965535, + "grad_norm": 0.41882777214050293, + "learning_rate": 0.00011062767263499762, + "loss": 0.0891, + "step": 19889 + }, + { + "epoch": 0.4662529228713585, + "grad_norm": 0.12760552763938904, + "learning_rate": 0.00011062034816589962, + "loss": 0.0173, + "step": 19890 + }, + { + "epoch": 0.4662763644461635, + "grad_norm": 0.9399291276931763, + "learning_rate": 0.00011061302363917531, + "loss": 0.3172, + "step": 19891 + }, + { + "epoch": 0.4662998060209685, + "grad_norm": 0.3636549115180969, + "learning_rate": 0.00011060569905486444, + "loss": 0.1141, + "step": 19892 + }, + { + "epoch": 0.4663232475957735, + "grad_norm": 0.7383451461791992, + "learning_rate": 0.00011059837441300678, + "loss": 0.2175, + "step": 19893 + }, + { + "epoch": 0.4663466891705785, + "grad_norm": 0.4635298252105713, + "learning_rate": 0.000110591049713642, + "loss": 0.0511, + "step": 19894 + }, + { + "epoch": 0.46637013074538347, + "grad_norm": 0.2965179979801178, + "learning_rate": 0.00011058372495680991, + "loss": 0.0663, + "step": 19895 + }, + { + "epoch": 0.46639357232018847, + "grad_norm": 0.38567233085632324, + "learning_rate": 0.00011057640014255019, + "loss": 0.1227, + "step": 19896 + }, + { + "epoch": 0.46641701389499346, + "grad_norm": 0.4891257882118225, + "learning_rate": 0.00011056907527090266, + "loss": 0.0919, + "step": 19897 + }, + { + "epoch": 0.46644045546979845, + "grad_norm": 0.47148966789245605, + "learning_rate": 0.00011056175034190703, + "loss": 0.0578, + "step": 19898 + }, + { + "epoch": 0.46646389704460345, + "grad_norm": 0.5335034728050232, + "learning_rate": 0.00011055442535560305, + "loss": 0.5191, + "step": 19899 + }, + { + "epoch": 0.46648733861940844, + "grad_norm": 0.2546776235103607, + "learning_rate": 0.00011054710031203046, + "loss": 0.0469, + "step": 19900 + }, + { + "epoch": 0.46651078019421344, + "grad_norm": 0.6218728423118591, + "learning_rate": 0.000110539775211229, + "loss": 0.6185, + "step": 19901 + }, + { + "epoch": 0.46653422176901843, + "grad_norm": 0.22877667844295502, + "learning_rate": 0.00011053245005323844, + "loss": 0.0594, + "step": 19902 + }, + { + "epoch": 0.4665576633438234, + "grad_norm": 0.06793178617954254, + "learning_rate": 0.00011052512483809847, + "loss": 0.0047, + "step": 19903 + }, + { + "epoch": 0.4665811049186284, + "grad_norm": 0.3557995855808258, + "learning_rate": 0.00011051779956584891, + "loss": 0.0545, + "step": 19904 + }, + { + "epoch": 0.4666045464934334, + "grad_norm": 0.20483127236366272, + "learning_rate": 0.00011051047423652947, + "loss": 0.0404, + "step": 19905 + }, + { + "epoch": 0.4666279880682384, + "grad_norm": 0.22849221527576447, + "learning_rate": 0.0001105031488501799, + "loss": 0.0658, + "step": 19906 + }, + { + "epoch": 0.4666514296430434, + "grad_norm": 0.5600315928459167, + "learning_rate": 0.00011049582340683992, + "loss": 0.0549, + "step": 19907 + }, + { + "epoch": 0.4666748712178484, + "grad_norm": 0.5651976466178894, + "learning_rate": 0.00011048849790654935, + "loss": 0.0941, + "step": 19908 + }, + { + "epoch": 0.4666983127926534, + "grad_norm": 0.2008616328239441, + "learning_rate": 0.00011048117234934788, + "loss": 0.0274, + "step": 19909 + }, + { + "epoch": 0.4667217543674584, + "grad_norm": 0.13706165552139282, + "learning_rate": 0.00011047384673527525, + "loss": 0.0244, + "step": 19910 + }, + { + "epoch": 0.4667451959422634, + "grad_norm": 0.4935170114040375, + "learning_rate": 0.00011046652106437128, + "loss": 0.0634, + "step": 19911 + }, + { + "epoch": 0.4667686375170684, + "grad_norm": 0.49747008085250854, + "learning_rate": 0.00011045919533667566, + "loss": 0.1114, + "step": 19912 + }, + { + "epoch": 0.46679207909187337, + "grad_norm": 0.4587969481945038, + "learning_rate": 0.00011045186955222813, + "loss": 0.1041, + "step": 19913 + }, + { + "epoch": 0.46681552066667836, + "grad_norm": 0.4526284337043762, + "learning_rate": 0.00011044454371106847, + "loss": 0.1661, + "step": 19914 + }, + { + "epoch": 0.46683896224148336, + "grad_norm": 0.21779100596904755, + "learning_rate": 0.00011043721781323641, + "loss": 0.034, + "step": 19915 + }, + { + "epoch": 0.46686240381628835, + "grad_norm": 0.44894251227378845, + "learning_rate": 0.0001104298918587717, + "loss": 0.0608, + "step": 19916 + }, + { + "epoch": 0.46688584539109335, + "grad_norm": 0.4624857008457184, + "learning_rate": 0.00011042256584771415, + "loss": 0.0847, + "step": 19917 + }, + { + "epoch": 0.4669092869658984, + "grad_norm": 0.5817509889602661, + "learning_rate": 0.00011041523978010347, + "loss": 0.1195, + "step": 19918 + }, + { + "epoch": 0.4669327285407034, + "grad_norm": 0.37461528182029724, + "learning_rate": 0.00011040791365597935, + "loss": 0.0768, + "step": 19919 + }, + { + "epoch": 0.4669561701155084, + "grad_norm": 0.21458259224891663, + "learning_rate": 0.00011040058747538164, + "loss": 0.0284, + "step": 19920 + }, + { + "epoch": 0.4669796116903134, + "grad_norm": 0.3035953938961029, + "learning_rate": 0.00011039326123835006, + "loss": 0.0669, + "step": 19921 + }, + { + "epoch": 0.4670030532651184, + "grad_norm": 0.27663272619247437, + "learning_rate": 0.00011038593494492429, + "loss": 0.0397, + "step": 19922 + }, + { + "epoch": 0.46702649483992337, + "grad_norm": 0.465132474899292, + "learning_rate": 0.00011037860859514419, + "loss": 0.0829, + "step": 19923 + }, + { + "epoch": 0.46704993641472836, + "grad_norm": 0.4269911050796509, + "learning_rate": 0.00011037128218904949, + "loss": 0.0546, + "step": 19924 + }, + { + "epoch": 0.46707337798953336, + "grad_norm": 0.4189812242984772, + "learning_rate": 0.00011036395572667989, + "loss": 0.0797, + "step": 19925 + }, + { + "epoch": 0.46709681956433835, + "grad_norm": 0.9540292620658875, + "learning_rate": 0.00011035662920807517, + "loss": 0.7734, + "step": 19926 + }, + { + "epoch": 0.46712026113914334, + "grad_norm": 0.60362309217453, + "learning_rate": 0.00011034930263327513, + "loss": 0.0775, + "step": 19927 + }, + { + "epoch": 0.46714370271394834, + "grad_norm": 0.4090692102909088, + "learning_rate": 0.00011034197600231943, + "loss": 0.059, + "step": 19928 + }, + { + "epoch": 0.46716714428875333, + "grad_norm": 0.21175093948841095, + "learning_rate": 0.00011033464931524788, + "loss": 0.0358, + "step": 19929 + }, + { + "epoch": 0.4671905858635583, + "grad_norm": 1.8984919786453247, + "learning_rate": 0.00011032732257210026, + "loss": 0.1316, + "step": 19930 + }, + { + "epoch": 0.4672140274383633, + "grad_norm": 0.3063039779663086, + "learning_rate": 0.0001103199957729163, + "loss": 0.0709, + "step": 19931 + }, + { + "epoch": 0.4672374690131683, + "grad_norm": 0.45067188143730164, + "learning_rate": 0.00011031266891773573, + "loss": 0.0926, + "step": 19932 + }, + { + "epoch": 0.4672609105879733, + "grad_norm": 0.423259973526001, + "learning_rate": 0.00011030534200659834, + "loss": 0.0555, + "step": 19933 + }, + { + "epoch": 0.4672843521627783, + "grad_norm": 0.5178632736206055, + "learning_rate": 0.00011029801503954386, + "loss": 0.0708, + "step": 19934 + }, + { + "epoch": 0.4673077937375833, + "grad_norm": 0.3809351623058319, + "learning_rate": 0.00011029068801661208, + "loss": 0.128, + "step": 19935 + }, + { + "epoch": 0.4673312353123883, + "grad_norm": 0.8585474491119385, + "learning_rate": 0.00011028336093784273, + "loss": 0.2026, + "step": 19936 + }, + { + "epoch": 0.4673546768871933, + "grad_norm": 0.41338080167770386, + "learning_rate": 0.00011027603380327561, + "loss": 0.0963, + "step": 19937 + }, + { + "epoch": 0.4673781184619983, + "grad_norm": 0.19708067178726196, + "learning_rate": 0.00011026870661295038, + "loss": 0.018, + "step": 19938 + }, + { + "epoch": 0.4674015600368033, + "grad_norm": 0.4825126528739929, + "learning_rate": 0.0001102613793669069, + "loss": 0.0261, + "step": 19939 + }, + { + "epoch": 0.46742500161160827, + "grad_norm": 0.4868907034397125, + "learning_rate": 0.00011025405206518486, + "loss": 0.0765, + "step": 19940 + }, + { + "epoch": 0.46744844318641326, + "grad_norm": 0.2145536094903946, + "learning_rate": 0.00011024672470782404, + "loss": 0.0349, + "step": 19941 + }, + { + "epoch": 0.46747188476121826, + "grad_norm": 0.8750147819519043, + "learning_rate": 0.00011023939729486423, + "loss": 0.07, + "step": 19942 + }, + { + "epoch": 0.46749532633602325, + "grad_norm": 0.42823052406311035, + "learning_rate": 0.00011023206982634517, + "loss": 0.0969, + "step": 19943 + }, + { + "epoch": 0.46751876791082825, + "grad_norm": 0.5648764371871948, + "learning_rate": 0.0001102247423023066, + "loss": 0.1257, + "step": 19944 + }, + { + "epoch": 0.46754220948563324, + "grad_norm": 0.542639970779419, + "learning_rate": 0.0001102174147227883, + "loss": 0.1248, + "step": 19945 + }, + { + "epoch": 0.46756565106043824, + "grad_norm": 0.25158464908599854, + "learning_rate": 0.00011021008708782998, + "loss": 0.0441, + "step": 19946 + }, + { + "epoch": 0.46758909263524323, + "grad_norm": 0.12290117144584656, + "learning_rate": 0.00011020275939747149, + "loss": 0.0125, + "step": 19947 + }, + { + "epoch": 0.4676125342100482, + "grad_norm": 0.387747585773468, + "learning_rate": 0.0001101954316517525, + "loss": 0.0702, + "step": 19948 + }, + { + "epoch": 0.4676359757848532, + "grad_norm": 0.3492679297924042, + "learning_rate": 0.00011018810385071286, + "loss": 0.0589, + "step": 19949 + }, + { + "epoch": 0.4676594173596582, + "grad_norm": 0.5311342477798462, + "learning_rate": 0.00011018077599439224, + "loss": 0.0792, + "step": 19950 + }, + { + "epoch": 0.4676828589344632, + "grad_norm": 0.30103975534439087, + "learning_rate": 0.00011017344808283046, + "loss": 0.0608, + "step": 19951 + }, + { + "epoch": 0.4677063005092682, + "grad_norm": 1.1129977703094482, + "learning_rate": 0.00011016612011606724, + "loss": 0.149, + "step": 19952 + }, + { + "epoch": 0.4677297420840732, + "grad_norm": 0.3886096179485321, + "learning_rate": 0.00011015879209414239, + "loss": 0.0608, + "step": 19953 + }, + { + "epoch": 0.4677531836588782, + "grad_norm": 0.46575042605400085, + "learning_rate": 0.00011015146401709564, + "loss": 0.0778, + "step": 19954 + }, + { + "epoch": 0.4677766252336832, + "grad_norm": 0.5869630575180054, + "learning_rate": 0.00011014413588496677, + "loss": 0.1268, + "step": 19955 + }, + { + "epoch": 0.4678000668084882, + "grad_norm": 0.7297108173370361, + "learning_rate": 0.00011013680769779556, + "loss": 0.1598, + "step": 19956 + }, + { + "epoch": 0.4678235083832932, + "grad_norm": 0.6092800498008728, + "learning_rate": 0.00011012947945562169, + "loss": 0.1186, + "step": 19957 + }, + { + "epoch": 0.46784694995809817, + "grad_norm": 0.14919763803482056, + "learning_rate": 0.000110122151158485, + "loss": 0.0282, + "step": 19958 + }, + { + "epoch": 0.46787039153290316, + "grad_norm": 0.4795074462890625, + "learning_rate": 0.00011011482280642524, + "loss": 0.0797, + "step": 19959 + }, + { + "epoch": 0.46789383310770816, + "grad_norm": 0.09133979678153992, + "learning_rate": 0.00011010749439948216, + "loss": 0.0141, + "step": 19960 + }, + { + "epoch": 0.46791727468251315, + "grad_norm": 0.5729914903640747, + "learning_rate": 0.00011010016593769554, + "loss": 0.0987, + "step": 19961 + }, + { + "epoch": 0.46794071625731815, + "grad_norm": 0.5702813267707825, + "learning_rate": 0.00011009283742110515, + "loss": 0.0761, + "step": 19962 + }, + { + "epoch": 0.46796415783212314, + "grad_norm": 0.5332027077674866, + "learning_rate": 0.00011008550884975071, + "loss": 0.0937, + "step": 19963 + }, + { + "epoch": 0.46798759940692813, + "grad_norm": 0.482828289270401, + "learning_rate": 0.00011007818022367202, + "loss": 0.0767, + "step": 19964 + }, + { + "epoch": 0.46801104098173313, + "grad_norm": 0.5028046369552612, + "learning_rate": 0.00011007085154290885, + "loss": 0.6691, + "step": 19965 + }, + { + "epoch": 0.4680344825565381, + "grad_norm": 0.6494023203849792, + "learning_rate": 0.00011006352280750098, + "loss": 0.11, + "step": 19966 + }, + { + "epoch": 0.4680579241313431, + "grad_norm": 0.2919808328151703, + "learning_rate": 0.0001100561940174881, + "loss": 0.0641, + "step": 19967 + }, + { + "epoch": 0.4680813657061481, + "grad_norm": 0.3294227719306946, + "learning_rate": 0.00011004886517291008, + "loss": 0.0959, + "step": 19968 + }, + { + "epoch": 0.46810480728095316, + "grad_norm": 0.5823600888252258, + "learning_rate": 0.00011004153627380662, + "loss": 0.4542, + "step": 19969 + }, + { + "epoch": 0.46812824885575816, + "grad_norm": 0.3696732223033905, + "learning_rate": 0.00011003420732021745, + "loss": 0.0928, + "step": 19970 + }, + { + "epoch": 0.46815169043056315, + "grad_norm": 0.6879019141197205, + "learning_rate": 0.00011002687831218246, + "loss": 0.1531, + "step": 19971 + }, + { + "epoch": 0.46817513200536814, + "grad_norm": 0.9078011512756348, + "learning_rate": 0.00011001954924974132, + "loss": 0.2027, + "step": 19972 + }, + { + "epoch": 0.46819857358017314, + "grad_norm": 0.311008095741272, + "learning_rate": 0.00011001222013293382, + "loss": 0.0512, + "step": 19973 + }, + { + "epoch": 0.46822201515497813, + "grad_norm": 0.10131573677062988, + "learning_rate": 0.00011000489096179974, + "loss": 0.0088, + "step": 19974 + }, + { + "epoch": 0.4682454567297831, + "grad_norm": 0.3389565646648407, + "learning_rate": 0.00010999756173637885, + "loss": 0.0718, + "step": 19975 + }, + { + "epoch": 0.4682688983045881, + "grad_norm": 0.09297478944063187, + "learning_rate": 0.00010999023245671087, + "loss": 0.0181, + "step": 19976 + }, + { + "epoch": 0.4682923398793931, + "grad_norm": 0.4097417891025543, + "learning_rate": 0.00010998290312283564, + "loss": 0.0519, + "step": 19977 + }, + { + "epoch": 0.4683157814541981, + "grad_norm": 0.44761380553245544, + "learning_rate": 0.00010997557373479289, + "loss": 0.1245, + "step": 19978 + }, + { + "epoch": 0.4683392230290031, + "grad_norm": 0.145425945520401, + "learning_rate": 0.00010996824429262237, + "loss": 0.0378, + "step": 19979 + }, + { + "epoch": 0.4683626646038081, + "grad_norm": 0.5294145941734314, + "learning_rate": 0.00010996091479636393, + "loss": 0.0382, + "step": 19980 + }, + { + "epoch": 0.4683861061786131, + "grad_norm": 0.5395401120185852, + "learning_rate": 0.00010995358524605726, + "loss": 0.0847, + "step": 19981 + }, + { + "epoch": 0.4684095477534181, + "grad_norm": 0.4730908274650574, + "learning_rate": 0.00010994625564174214, + "loss": 0.092, + "step": 19982 + }, + { + "epoch": 0.4684329893282231, + "grad_norm": 0.2573331296443939, + "learning_rate": 0.00010993892598345836, + "loss": 0.0566, + "step": 19983 + }, + { + "epoch": 0.4684564309030281, + "grad_norm": 1.1798145771026611, + "learning_rate": 0.00010993159627124572, + "loss": 0.1682, + "step": 19984 + }, + { + "epoch": 0.46847987247783307, + "grad_norm": 0.5397258400917053, + "learning_rate": 0.00010992426650514394, + "loss": 0.107, + "step": 19985 + }, + { + "epoch": 0.46850331405263806, + "grad_norm": 1.1699668169021606, + "learning_rate": 0.00010991693668519278, + "loss": 0.2696, + "step": 19986 + }, + { + "epoch": 0.46852675562744306, + "grad_norm": 0.4877455532550812, + "learning_rate": 0.0001099096068114321, + "loss": 0.08, + "step": 19987 + }, + { + "epoch": 0.46855019720224805, + "grad_norm": 0.4494047164916992, + "learning_rate": 0.0001099022768839016, + "loss": 0.1317, + "step": 19988 + }, + { + "epoch": 0.46857363877705305, + "grad_norm": 0.220086008310318, + "learning_rate": 0.00010989494690264102, + "loss": 0.0483, + "step": 19989 + }, + { + "epoch": 0.46859708035185804, + "grad_norm": 0.4097368121147156, + "learning_rate": 0.00010988761686769023, + "loss": 0.0611, + "step": 19990 + }, + { + "epoch": 0.46862052192666304, + "grad_norm": 0.36230361461639404, + "learning_rate": 0.00010988028677908895, + "loss": 0.047, + "step": 19991 + }, + { + "epoch": 0.46864396350146803, + "grad_norm": 0.40020856261253357, + "learning_rate": 0.00010987295663687691, + "loss": 0.0823, + "step": 19992 + }, + { + "epoch": 0.468667405076273, + "grad_norm": 0.31333842873573303, + "learning_rate": 0.00010986562644109398, + "loss": 0.028, + "step": 19993 + }, + { + "epoch": 0.468690846651078, + "grad_norm": 0.3428453803062439, + "learning_rate": 0.00010985829619177987, + "loss": 0.0635, + "step": 19994 + }, + { + "epoch": 0.468714288225883, + "grad_norm": 0.5410399436950684, + "learning_rate": 0.00010985096588897437, + "loss": 0.0856, + "step": 19995 + }, + { + "epoch": 0.468737729800688, + "grad_norm": 0.5140367150306702, + "learning_rate": 0.00010984363553271722, + "loss": 0.1466, + "step": 19996 + }, + { + "epoch": 0.468761171375493, + "grad_norm": 0.20292839407920837, + "learning_rate": 0.00010983630512304828, + "loss": 0.0495, + "step": 19997 + }, + { + "epoch": 0.468784612950298, + "grad_norm": 0.30008646845817566, + "learning_rate": 0.00010982897466000724, + "loss": 0.0362, + "step": 19998 + }, + { + "epoch": 0.468808054525103, + "grad_norm": 0.3759301006793976, + "learning_rate": 0.00010982164414363389, + "loss": 0.0746, + "step": 19999 + }, + { + "epoch": 0.468831496099908, + "grad_norm": 0.6445008516311646, + "learning_rate": 0.00010981431357396808, + "loss": 0.121, + "step": 20000 + }, + { + "epoch": 0.468854937674713, + "grad_norm": 0.1813822239637375, + "learning_rate": 0.00010980698295104948, + "loss": 0.0343, + "step": 20001 + }, + { + "epoch": 0.468878379249518, + "grad_norm": 0.4784215986728668, + "learning_rate": 0.00010979965227491791, + "loss": 0.1178, + "step": 20002 + }, + { + "epoch": 0.46890182082432297, + "grad_norm": 0.24572786688804626, + "learning_rate": 0.00010979232154561318, + "loss": 0.0475, + "step": 20003 + }, + { + "epoch": 0.46892526239912796, + "grad_norm": 0.480196475982666, + "learning_rate": 0.00010978499076317504, + "loss": 0.0767, + "step": 20004 + }, + { + "epoch": 0.46894870397393296, + "grad_norm": 0.2900577783584595, + "learning_rate": 0.00010977765992764324, + "loss": 0.0357, + "step": 20005 + }, + { + "epoch": 0.46897214554873795, + "grad_norm": 1.6215876340866089, + "learning_rate": 0.0001097703290390576, + "loss": 0.0791, + "step": 20006 + }, + { + "epoch": 0.46899558712354295, + "grad_norm": 0.41489145159721375, + "learning_rate": 0.00010976299809745787, + "loss": 0.0899, + "step": 20007 + }, + { + "epoch": 0.46901902869834794, + "grad_norm": 0.4454469382762909, + "learning_rate": 0.00010975566710288383, + "loss": 0.1019, + "step": 20008 + }, + { + "epoch": 0.46904247027315293, + "grad_norm": 0.563855767250061, + "learning_rate": 0.0001097483360553753, + "loss": 0.7645, + "step": 20009 + }, + { + "epoch": 0.46906591184795793, + "grad_norm": 0.48581433296203613, + "learning_rate": 0.000109741004954972, + "loss": 0.119, + "step": 20010 + }, + { + "epoch": 0.4690893534227629, + "grad_norm": 0.4709799885749817, + "learning_rate": 0.00010973367380171371, + "loss": 0.5929, + "step": 20011 + }, + { + "epoch": 0.4691127949975679, + "grad_norm": 0.47642138600349426, + "learning_rate": 0.00010972634259564028, + "loss": 0.1021, + "step": 20012 + }, + { + "epoch": 0.4691362365723729, + "grad_norm": 0.18067492544651031, + "learning_rate": 0.00010971901133679141, + "loss": 0.0275, + "step": 20013 + }, + { + "epoch": 0.4691596781471779, + "grad_norm": 0.35139715671539307, + "learning_rate": 0.00010971168002520692, + "loss": 0.0544, + "step": 20014 + }, + { + "epoch": 0.4691831197219829, + "grad_norm": 0.6609050035476685, + "learning_rate": 0.00010970434866092658, + "loss": 0.1119, + "step": 20015 + }, + { + "epoch": 0.4692065612967879, + "grad_norm": 0.5786214470863342, + "learning_rate": 0.00010969701724399017, + "loss": 0.519, + "step": 20016 + }, + { + "epoch": 0.4692300028715929, + "grad_norm": 0.5435694456100464, + "learning_rate": 0.00010968968577443748, + "loss": 0.107, + "step": 20017 + }, + { + "epoch": 0.4692534444463979, + "grad_norm": 0.3517053723335266, + "learning_rate": 0.00010968235425230829, + "loss": 0.0233, + "step": 20018 + }, + { + "epoch": 0.4692768860212029, + "grad_norm": 0.39737093448638916, + "learning_rate": 0.00010967502267764234, + "loss": 0.1028, + "step": 20019 + }, + { + "epoch": 0.4693003275960079, + "grad_norm": 0.5623407959938049, + "learning_rate": 0.00010966769105047947, + "loss": 0.1326, + "step": 20020 + }, + { + "epoch": 0.4693237691708129, + "grad_norm": 0.23890765011310577, + "learning_rate": 0.00010966035937085944, + "loss": 0.029, + "step": 20021 + }, + { + "epoch": 0.4693472107456179, + "grad_norm": 0.30442535877227783, + "learning_rate": 0.00010965302763882202, + "loss": 0.0457, + "step": 20022 + }, + { + "epoch": 0.4693706523204229, + "grad_norm": 0.6619386076927185, + "learning_rate": 0.000109645695854407, + "loss": 0.0899, + "step": 20023 + }, + { + "epoch": 0.4693940938952279, + "grad_norm": 0.08669286966323853, + "learning_rate": 0.00010963836401765418, + "loss": 0.0142, + "step": 20024 + }, + { + "epoch": 0.4694175354700329, + "grad_norm": 0.6170503497123718, + "learning_rate": 0.00010963103212860328, + "loss": 0.1388, + "step": 20025 + }, + { + "epoch": 0.4694409770448379, + "grad_norm": 0.38391345739364624, + "learning_rate": 0.00010962370018729418, + "loss": 0.0633, + "step": 20026 + }, + { + "epoch": 0.4694644186196429, + "grad_norm": 0.5912513136863708, + "learning_rate": 0.00010961636819376657, + "loss": 0.6774, + "step": 20027 + }, + { + "epoch": 0.4694878601944479, + "grad_norm": 0.6576737761497498, + "learning_rate": 0.00010960903614806031, + "loss": 0.6736, + "step": 20028 + }, + { + "epoch": 0.4695113017692529, + "grad_norm": 0.6724212765693665, + "learning_rate": 0.00010960170405021515, + "loss": 0.8069, + "step": 20029 + }, + { + "epoch": 0.46953474334405787, + "grad_norm": 0.6329348087310791, + "learning_rate": 0.00010959437190027084, + "loss": 0.1355, + "step": 20030 + }, + { + "epoch": 0.46955818491886286, + "grad_norm": 0.3439592719078064, + "learning_rate": 0.00010958703969826722, + "loss": 0.0378, + "step": 20031 + }, + { + "epoch": 0.46958162649366786, + "grad_norm": 0.7172462344169617, + "learning_rate": 0.00010957970744424406, + "loss": 0.168, + "step": 20032 + }, + { + "epoch": 0.46960506806847285, + "grad_norm": 0.5713506937026978, + "learning_rate": 0.00010957237513824112, + "loss": 0.143, + "step": 20033 + }, + { + "epoch": 0.46962850964327785, + "grad_norm": 0.8214192390441895, + "learning_rate": 0.00010956504278029822, + "loss": 0.1477, + "step": 20034 + }, + { + "epoch": 0.46965195121808284, + "grad_norm": 0.5199635624885559, + "learning_rate": 0.00010955771037045515, + "loss": 0.097, + "step": 20035 + }, + { + "epoch": 0.46967539279288784, + "grad_norm": 0.4647965431213379, + "learning_rate": 0.00010955037790875162, + "loss": 0.0508, + "step": 20036 + }, + { + "epoch": 0.46969883436769283, + "grad_norm": 0.5102932453155518, + "learning_rate": 0.00010954304539522749, + "loss": 0.0992, + "step": 20037 + }, + { + "epoch": 0.4697222759424978, + "grad_norm": 0.3575665354728699, + "learning_rate": 0.00010953571282992254, + "loss": 0.0676, + "step": 20038 + }, + { + "epoch": 0.4697457175173028, + "grad_norm": 0.32054954767227173, + "learning_rate": 0.00010952838021287655, + "loss": 0.0834, + "step": 20039 + }, + { + "epoch": 0.4697691590921078, + "grad_norm": 0.7095454931259155, + "learning_rate": 0.00010952104754412929, + "loss": 0.0862, + "step": 20040 + }, + { + "epoch": 0.4697926006669128, + "grad_norm": 0.5506814122200012, + "learning_rate": 0.00010951371482372057, + "loss": 0.1525, + "step": 20041 + }, + { + "epoch": 0.4698160422417178, + "grad_norm": 0.638996422290802, + "learning_rate": 0.00010950638205169018, + "loss": 0.0874, + "step": 20042 + }, + { + "epoch": 0.4698394838165228, + "grad_norm": 0.9866832494735718, + "learning_rate": 0.00010949904922807781, + "loss": 0.6786, + "step": 20043 + }, + { + "epoch": 0.4698629253913278, + "grad_norm": 0.719231128692627, + "learning_rate": 0.00010949171635292342, + "loss": 0.5081, + "step": 20044 + }, + { + "epoch": 0.4698863669661328, + "grad_norm": 0.6221085786819458, + "learning_rate": 0.00010948438342626667, + "loss": 0.1801, + "step": 20045 + }, + { + "epoch": 0.4699098085409378, + "grad_norm": 0.650041937828064, + "learning_rate": 0.0001094770504481474, + "loss": 0.1605, + "step": 20046 + }, + { + "epoch": 0.4699332501157428, + "grad_norm": 0.33097919821739197, + "learning_rate": 0.00010946971741860539, + "loss": 0.0382, + "step": 20047 + }, + { + "epoch": 0.46995669169054777, + "grad_norm": 0.33026206493377686, + "learning_rate": 0.00010946238433768043, + "loss": 0.073, + "step": 20048 + }, + { + "epoch": 0.46998013326535276, + "grad_norm": 0.2847825884819031, + "learning_rate": 0.00010945505120541231, + "loss": 0.0435, + "step": 20049 + }, + { + "epoch": 0.47000357484015776, + "grad_norm": 0.6493288278579712, + "learning_rate": 0.00010944771802184077, + "loss": 0.1742, + "step": 20050 + }, + { + "epoch": 0.47002701641496275, + "grad_norm": 0.4709710478782654, + "learning_rate": 0.00010944038478700571, + "loss": 0.0979, + "step": 20051 + }, + { + "epoch": 0.47005045798976774, + "grad_norm": 0.37869080901145935, + "learning_rate": 0.00010943305150094683, + "loss": 0.0641, + "step": 20052 + }, + { + "epoch": 0.47007389956457274, + "grad_norm": 0.17609862983226776, + "learning_rate": 0.00010942571816370395, + "loss": 0.0206, + "step": 20053 + }, + { + "epoch": 0.47009734113937773, + "grad_norm": 0.3883053660392761, + "learning_rate": 0.00010941838477531685, + "loss": 0.0468, + "step": 20054 + }, + { + "epoch": 0.47012078271418273, + "grad_norm": 0.60393887758255, + "learning_rate": 0.00010941105133582535, + "loss": 0.2011, + "step": 20055 + }, + { + "epoch": 0.4701442242889877, + "grad_norm": 0.6882362961769104, + "learning_rate": 0.00010940371784526917, + "loss": 0.1267, + "step": 20056 + }, + { + "epoch": 0.4701676658637927, + "grad_norm": 0.31332314014434814, + "learning_rate": 0.00010939638430368822, + "loss": 0.0574, + "step": 20057 + }, + { + "epoch": 0.4701911074385977, + "grad_norm": 0.2492150515317917, + "learning_rate": 0.0001093890507111222, + "loss": 0.039, + "step": 20058 + }, + { + "epoch": 0.4702145490134027, + "grad_norm": 0.2679550349712372, + "learning_rate": 0.00010938171706761088, + "loss": 0.0761, + "step": 20059 + }, + { + "epoch": 0.4702379905882077, + "grad_norm": 0.41675621271133423, + "learning_rate": 0.00010937438337319417, + "loss": 0.0951, + "step": 20060 + }, + { + "epoch": 0.4702614321630127, + "grad_norm": 0.3782549500465393, + "learning_rate": 0.00010936704962791175, + "loss": 0.3417, + "step": 20061 + }, + { + "epoch": 0.4702848737378177, + "grad_norm": 0.6284823417663574, + "learning_rate": 0.00010935971583180347, + "loss": 0.0749, + "step": 20062 + }, + { + "epoch": 0.4703083153126227, + "grad_norm": 0.3931965231895447, + "learning_rate": 0.00010935238198490911, + "loss": 0.0929, + "step": 20063 + }, + { + "epoch": 0.4703317568874277, + "grad_norm": 0.9261178374290466, + "learning_rate": 0.00010934504808726847, + "loss": 0.7813, + "step": 20064 + }, + { + "epoch": 0.47035519846223267, + "grad_norm": 0.6066458821296692, + "learning_rate": 0.00010933771413892131, + "loss": 0.1563, + "step": 20065 + }, + { + "epoch": 0.47037864003703767, + "grad_norm": 0.7174713611602783, + "learning_rate": 0.00010933038013990747, + "loss": 0.2253, + "step": 20066 + }, + { + "epoch": 0.47040208161184266, + "grad_norm": 0.7213202714920044, + "learning_rate": 0.00010932304609026673, + "loss": 0.127, + "step": 20067 + }, + { + "epoch": 0.47042552318664765, + "grad_norm": 0.5453382134437561, + "learning_rate": 0.00010931571199003885, + "loss": 0.4923, + "step": 20068 + }, + { + "epoch": 0.47044896476145265, + "grad_norm": 0.6417896747589111, + "learning_rate": 0.00010930837783926367, + "loss": 0.159, + "step": 20069 + }, + { + "epoch": 0.47047240633625764, + "grad_norm": 0.37313202023506165, + "learning_rate": 0.000109301043637981, + "loss": 0.1, + "step": 20070 + }, + { + "epoch": 0.4704958479110627, + "grad_norm": 0.4709102511405945, + "learning_rate": 0.00010929370938623056, + "loss": 0.093, + "step": 20071 + }, + { + "epoch": 0.4705192894858677, + "grad_norm": 0.39038753509521484, + "learning_rate": 0.00010928637508405222, + "loss": 0.0677, + "step": 20072 + }, + { + "epoch": 0.4705427310606727, + "grad_norm": 0.43205738067626953, + "learning_rate": 0.00010927904073148574, + "loss": 0.0836, + "step": 20073 + }, + { + "epoch": 0.4705661726354777, + "grad_norm": 0.4742891490459442, + "learning_rate": 0.00010927170632857091, + "loss": 0.6234, + "step": 20074 + }, + { + "epoch": 0.47058961421028267, + "grad_norm": 0.48086780309677124, + "learning_rate": 0.00010926437187534755, + "loss": 0.0686, + "step": 20075 + }, + { + "epoch": 0.47061305578508766, + "grad_norm": 0.34498876333236694, + "learning_rate": 0.00010925703737185546, + "loss": 0.0528, + "step": 20076 + }, + { + "epoch": 0.47063649735989266, + "grad_norm": 0.29138824343681335, + "learning_rate": 0.00010924970281813443, + "loss": 0.0565, + "step": 20077 + }, + { + "epoch": 0.47065993893469765, + "grad_norm": 0.6907092928886414, + "learning_rate": 0.00010924236821422422, + "loss": 0.1245, + "step": 20078 + }, + { + "epoch": 0.47068338050950265, + "grad_norm": 0.4741314947605133, + "learning_rate": 0.00010923503356016469, + "loss": 0.071, + "step": 20079 + }, + { + "epoch": 0.47070682208430764, + "grad_norm": 0.7453532218933105, + "learning_rate": 0.00010922769885599559, + "loss": 0.1203, + "step": 20080 + }, + { + "epoch": 0.47073026365911264, + "grad_norm": 0.42801567912101746, + "learning_rate": 0.00010922036410175673, + "loss": 0.102, + "step": 20081 + }, + { + "epoch": 0.47075370523391763, + "grad_norm": 0.23661494255065918, + "learning_rate": 0.00010921302929748795, + "loss": 0.0351, + "step": 20082 + }, + { + "epoch": 0.4707771468087226, + "grad_norm": 0.4367506504058838, + "learning_rate": 0.00010920569444322902, + "loss": 0.0946, + "step": 20083 + }, + { + "epoch": 0.4708005883835276, + "grad_norm": 0.3750397264957428, + "learning_rate": 0.00010919835953901967, + "loss": 0.0454, + "step": 20084 + }, + { + "epoch": 0.4708240299583326, + "grad_norm": 0.10950445383787155, + "learning_rate": 0.00010919102458489982, + "loss": 0.0246, + "step": 20085 + }, + { + "epoch": 0.4708474715331376, + "grad_norm": 0.555176854133606, + "learning_rate": 0.00010918368958090918, + "loss": 0.122, + "step": 20086 + }, + { + "epoch": 0.4708709131079426, + "grad_norm": 0.48057791590690613, + "learning_rate": 0.0001091763545270876, + "loss": 0.092, + "step": 20087 + }, + { + "epoch": 0.4708943546827476, + "grad_norm": 0.2325844168663025, + "learning_rate": 0.00010916901942347484, + "loss": 0.0485, + "step": 20088 + }, + { + "epoch": 0.4709177962575526, + "grad_norm": 0.21028809249401093, + "learning_rate": 0.00010916168427011075, + "loss": 0.0575, + "step": 20089 + }, + { + "epoch": 0.4709412378323576, + "grad_norm": 0.5379297733306885, + "learning_rate": 0.00010915434906703508, + "loss": 0.102, + "step": 20090 + }, + { + "epoch": 0.4709646794071626, + "grad_norm": 0.40535131096839905, + "learning_rate": 0.00010914701381428769, + "loss": 0.0885, + "step": 20091 + }, + { + "epoch": 0.4709881209819676, + "grad_norm": 0.5123797059059143, + "learning_rate": 0.00010913967851190829, + "loss": 0.1099, + "step": 20092 + }, + { + "epoch": 0.47101156255677257, + "grad_norm": 0.22418221831321716, + "learning_rate": 0.00010913234315993678, + "loss": 0.0633, + "step": 20093 + }, + { + "epoch": 0.47103500413157756, + "grad_norm": 0.3865492045879364, + "learning_rate": 0.00010912500775841293, + "loss": 0.4448, + "step": 20094 + }, + { + "epoch": 0.47105844570638256, + "grad_norm": 0.3236922025680542, + "learning_rate": 0.00010911767230737651, + "loss": 0.0284, + "step": 20095 + }, + { + "epoch": 0.47108188728118755, + "grad_norm": 0.5670966506004333, + "learning_rate": 0.00010911033680686735, + "loss": 0.1909, + "step": 20096 + }, + { + "epoch": 0.47110532885599254, + "grad_norm": 0.3736276626586914, + "learning_rate": 0.00010910300125692523, + "loss": 0.1115, + "step": 20097 + }, + { + "epoch": 0.47112877043079754, + "grad_norm": 0.4361186921596527, + "learning_rate": 0.00010909566565758997, + "loss": 0.0833, + "step": 20098 + }, + { + "epoch": 0.47115221200560253, + "grad_norm": 0.4966333210468292, + "learning_rate": 0.0001090883300089014, + "loss": 0.525, + "step": 20099 + }, + { + "epoch": 0.4711756535804075, + "grad_norm": 0.42455536127090454, + "learning_rate": 0.00010908099431089926, + "loss": 0.1036, + "step": 20100 + }, + { + "epoch": 0.4711990951552125, + "grad_norm": 0.3784312903881073, + "learning_rate": 0.00010907365856362341, + "loss": 0.0754, + "step": 20101 + }, + { + "epoch": 0.4712225367300175, + "grad_norm": 0.4426412284374237, + "learning_rate": 0.00010906632276711364, + "loss": 0.7262, + "step": 20102 + }, + { + "epoch": 0.4712459783048225, + "grad_norm": 0.6162981390953064, + "learning_rate": 0.00010905898692140972, + "loss": 0.0572, + "step": 20103 + }, + { + "epoch": 0.4712694198796275, + "grad_norm": 0.5335206985473633, + "learning_rate": 0.0001090516510265515, + "loss": 0.1073, + "step": 20104 + }, + { + "epoch": 0.4712928614544325, + "grad_norm": 0.37592560052871704, + "learning_rate": 0.00010904431508257877, + "loss": 0.0987, + "step": 20105 + }, + { + "epoch": 0.4713163030292375, + "grad_norm": 0.4750927984714508, + "learning_rate": 0.00010903697908953134, + "loss": 0.704, + "step": 20106 + }, + { + "epoch": 0.4713397446040425, + "grad_norm": 0.288006067276001, + "learning_rate": 0.000109029643047449, + "loss": 0.029, + "step": 20107 + }, + { + "epoch": 0.4713631861788475, + "grad_norm": 0.40943101048469543, + "learning_rate": 0.00010902230695637156, + "loss": 0.0507, + "step": 20108 + }, + { + "epoch": 0.4713866277536525, + "grad_norm": 0.3223012089729309, + "learning_rate": 0.00010901497081633883, + "loss": 0.0549, + "step": 20109 + }, + { + "epoch": 0.47141006932845747, + "grad_norm": 0.24380981922149658, + "learning_rate": 0.00010900763462739059, + "loss": 0.0261, + "step": 20110 + }, + { + "epoch": 0.47143351090326246, + "grad_norm": 0.6668372750282288, + "learning_rate": 0.00010900029838956671, + "loss": 0.0699, + "step": 20111 + }, + { + "epoch": 0.47145695247806746, + "grad_norm": 0.3924509882926941, + "learning_rate": 0.00010899296210290695, + "loss": 0.077, + "step": 20112 + }, + { + "epoch": 0.47148039405287245, + "grad_norm": 0.709135890007019, + "learning_rate": 0.00010898562576745111, + "loss": 0.8181, + "step": 20113 + }, + { + "epoch": 0.47150383562767745, + "grad_norm": 0.38963064551353455, + "learning_rate": 0.00010897828938323904, + "loss": 0.1027, + "step": 20114 + }, + { + "epoch": 0.47152727720248244, + "grad_norm": 0.5477942228317261, + "learning_rate": 0.0001089709529503105, + "loss": 0.085, + "step": 20115 + }, + { + "epoch": 0.47155071877728744, + "grad_norm": 0.6208884119987488, + "learning_rate": 0.00010896361646870527, + "loss": 0.127, + "step": 20116 + }, + { + "epoch": 0.47157416035209243, + "grad_norm": 0.5597436428070068, + "learning_rate": 0.00010895627993846327, + "loss": 0.1089, + "step": 20117 + }, + { + "epoch": 0.4715976019268974, + "grad_norm": 0.29569125175476074, + "learning_rate": 0.00010894894335962425, + "loss": 0.0466, + "step": 20118 + }, + { + "epoch": 0.4716210435017024, + "grad_norm": 0.48626890778541565, + "learning_rate": 0.00010894160673222798, + "loss": 0.088, + "step": 20119 + }, + { + "epoch": 0.4716444850765074, + "grad_norm": 0.3699604272842407, + "learning_rate": 0.00010893427005631432, + "loss": 0.4023, + "step": 20120 + }, + { + "epoch": 0.4716679266513124, + "grad_norm": 0.46799418330192566, + "learning_rate": 0.00010892693333192303, + "loss": 0.09, + "step": 20121 + }, + { + "epoch": 0.47169136822611746, + "grad_norm": 0.4515278935432434, + "learning_rate": 0.00010891959655909396, + "loss": 0.0697, + "step": 20122 + }, + { + "epoch": 0.47171480980092245, + "grad_norm": 0.5693953633308411, + "learning_rate": 0.00010891225973786689, + "loss": 0.8009, + "step": 20123 + }, + { + "epoch": 0.47173825137572745, + "grad_norm": 0.16705547273159027, + "learning_rate": 0.00010890492286828169, + "loss": 0.0177, + "step": 20124 + }, + { + "epoch": 0.47176169295053244, + "grad_norm": 0.4597031772136688, + "learning_rate": 0.00010889758595037813, + "loss": 0.1129, + "step": 20125 + }, + { + "epoch": 0.47178513452533744, + "grad_norm": 0.9492121934890747, + "learning_rate": 0.00010889024898419597, + "loss": 0.3229, + "step": 20126 + }, + { + "epoch": 0.47180857610014243, + "grad_norm": 0.5024853348731995, + "learning_rate": 0.00010888291196977512, + "loss": 0.6702, + "step": 20127 + }, + { + "epoch": 0.4718320176749474, + "grad_norm": 0.14243270456790924, + "learning_rate": 0.00010887557490715529, + "loss": 0.0264, + "step": 20128 + }, + { + "epoch": 0.4718554592497524, + "grad_norm": 0.1302923560142517, + "learning_rate": 0.00010886823779637638, + "loss": 0.0316, + "step": 20129 + }, + { + "epoch": 0.4718789008245574, + "grad_norm": 0.13416741788387299, + "learning_rate": 0.00010886090063747815, + "loss": 0.0198, + "step": 20130 + }, + { + "epoch": 0.4719023423993624, + "grad_norm": 0.5573978424072266, + "learning_rate": 0.00010885356343050044, + "loss": 0.1204, + "step": 20131 + }, + { + "epoch": 0.4719257839741674, + "grad_norm": 0.6588783264160156, + "learning_rate": 0.00010884622617548301, + "loss": 0.0506, + "step": 20132 + }, + { + "epoch": 0.4719492255489724, + "grad_norm": 0.39227256178855896, + "learning_rate": 0.00010883888887246573, + "loss": 0.066, + "step": 20133 + }, + { + "epoch": 0.4719726671237774, + "grad_norm": 0.3634951710700989, + "learning_rate": 0.00010883155152148838, + "loss": 0.0695, + "step": 20134 + }, + { + "epoch": 0.4719961086985824, + "grad_norm": 0.11594857275485992, + "learning_rate": 0.00010882421412259078, + "loss": 0.0289, + "step": 20135 + }, + { + "epoch": 0.4720195502733874, + "grad_norm": 0.31286075711250305, + "learning_rate": 0.00010881687667581278, + "loss": 0.0679, + "step": 20136 + }, + { + "epoch": 0.4720429918481924, + "grad_norm": 0.36174967885017395, + "learning_rate": 0.00010880953918119414, + "loss": 0.075, + "step": 20137 + }, + { + "epoch": 0.47206643342299737, + "grad_norm": 0.31394076347351074, + "learning_rate": 0.00010880220163877468, + "loss": 0.0437, + "step": 20138 + }, + { + "epoch": 0.47208987499780236, + "grad_norm": 0.4227500557899475, + "learning_rate": 0.00010879486404859425, + "loss": 0.1018, + "step": 20139 + }, + { + "epoch": 0.47211331657260736, + "grad_norm": 0.6915328502655029, + "learning_rate": 0.00010878752641069262, + "loss": 0.1686, + "step": 20140 + }, + { + "epoch": 0.47213675814741235, + "grad_norm": 0.6179440021514893, + "learning_rate": 0.00010878018872510965, + "loss": 0.699, + "step": 20141 + }, + { + "epoch": 0.47216019972221734, + "grad_norm": 0.5907325744628906, + "learning_rate": 0.00010877285099188508, + "loss": 0.1109, + "step": 20142 + }, + { + "epoch": 0.47218364129702234, + "grad_norm": 0.718669056892395, + "learning_rate": 0.00010876551321105881, + "loss": 0.1211, + "step": 20143 + }, + { + "epoch": 0.47220708287182733, + "grad_norm": 2.1933295726776123, + "learning_rate": 0.00010875817538267063, + "loss": 0.0678, + "step": 20144 + }, + { + "epoch": 0.4722305244466323, + "grad_norm": 0.13433897495269775, + "learning_rate": 0.00010875083750676031, + "loss": 0.0279, + "step": 20145 + }, + { + "epoch": 0.4722539660214373, + "grad_norm": 0.2244611531496048, + "learning_rate": 0.00010874349958336774, + "loss": 0.0518, + "step": 20146 + }, + { + "epoch": 0.4722774075962423, + "grad_norm": 0.08258657902479172, + "learning_rate": 0.00010873616161253267, + "loss": 0.0163, + "step": 20147 + }, + { + "epoch": 0.4723008491710473, + "grad_norm": 0.4431115686893463, + "learning_rate": 0.00010872882359429493, + "loss": 0.0802, + "step": 20148 + }, + { + "epoch": 0.4723242907458523, + "grad_norm": 0.41308656334877014, + "learning_rate": 0.00010872148552869436, + "loss": 0.095, + "step": 20149 + }, + { + "epoch": 0.4723477323206573, + "grad_norm": 0.3335592746734619, + "learning_rate": 0.00010871414741577077, + "loss": 0.222, + "step": 20150 + }, + { + "epoch": 0.4723711738954623, + "grad_norm": 0.8661858439445496, + "learning_rate": 0.00010870680925556395, + "loss": 0.2141, + "step": 20151 + }, + { + "epoch": 0.4723946154702673, + "grad_norm": 0.6046480536460876, + "learning_rate": 0.00010869947104811377, + "loss": 0.0904, + "step": 20152 + }, + { + "epoch": 0.4724180570450723, + "grad_norm": 0.4544602930545807, + "learning_rate": 0.00010869213279346, + "loss": 0.0486, + "step": 20153 + }, + { + "epoch": 0.4724414986198773, + "grad_norm": 0.5379098057746887, + "learning_rate": 0.00010868479449164246, + "loss": 0.156, + "step": 20154 + }, + { + "epoch": 0.47246494019468227, + "grad_norm": 0.35415464639663696, + "learning_rate": 0.00010867745614270099, + "loss": 0.0628, + "step": 20155 + }, + { + "epoch": 0.47248838176948726, + "grad_norm": 0.4550473690032959, + "learning_rate": 0.00010867011774667543, + "loss": 0.0988, + "step": 20156 + }, + { + "epoch": 0.47251182334429226, + "grad_norm": 0.1554698646068573, + "learning_rate": 0.00010866277930360553, + "loss": 0.0343, + "step": 20157 + }, + { + "epoch": 0.47253526491909725, + "grad_norm": 0.5115536451339722, + "learning_rate": 0.00010865544081353117, + "loss": 0.1275, + "step": 20158 + }, + { + "epoch": 0.47255870649390225, + "grad_norm": 0.4223329424858093, + "learning_rate": 0.00010864810227649212, + "loss": 0.0941, + "step": 20159 + }, + { + "epoch": 0.47258214806870724, + "grad_norm": 0.4681428372859955, + "learning_rate": 0.00010864076369252825, + "loss": 0.1071, + "step": 20160 + }, + { + "epoch": 0.47260558964351224, + "grad_norm": 0.37776899337768555, + "learning_rate": 0.00010863342506167931, + "loss": 0.1049, + "step": 20161 + }, + { + "epoch": 0.47262903121831723, + "grad_norm": 0.39884379506111145, + "learning_rate": 0.00010862608638398521, + "loss": 0.0883, + "step": 20162 + }, + { + "epoch": 0.4726524727931222, + "grad_norm": 0.35696497559547424, + "learning_rate": 0.00010861874765948572, + "loss": 0.0432, + "step": 20163 + }, + { + "epoch": 0.4726759143679272, + "grad_norm": 0.4781264364719391, + "learning_rate": 0.00010861140888822064, + "loss": 0.0951, + "step": 20164 + }, + { + "epoch": 0.4726993559427322, + "grad_norm": 0.5778598189353943, + "learning_rate": 0.00010860407007022979, + "loss": 0.1274, + "step": 20165 + }, + { + "epoch": 0.4727227975175372, + "grad_norm": 0.3417130708694458, + "learning_rate": 0.00010859673120555304, + "loss": 0.0857, + "step": 20166 + }, + { + "epoch": 0.4727462390923422, + "grad_norm": 0.5028223991394043, + "learning_rate": 0.0001085893922942302, + "loss": 0.1228, + "step": 20167 + }, + { + "epoch": 0.4727696806671472, + "grad_norm": 0.3833504915237427, + "learning_rate": 0.00010858205333630107, + "loss": 0.0878, + "step": 20168 + }, + { + "epoch": 0.4727931222419522, + "grad_norm": 0.42538031935691833, + "learning_rate": 0.00010857471433180548, + "loss": 0.095, + "step": 20169 + }, + { + "epoch": 0.4728165638167572, + "grad_norm": 0.5169487595558167, + "learning_rate": 0.00010856737528078322, + "loss": 0.0906, + "step": 20170 + }, + { + "epoch": 0.4728400053915622, + "grad_norm": 0.5595131516456604, + "learning_rate": 0.00010856003618327415, + "loss": 0.0616, + "step": 20171 + }, + { + "epoch": 0.4728634469663672, + "grad_norm": 0.36516231298446655, + "learning_rate": 0.00010855269703931812, + "loss": 0.0792, + "step": 20172 + }, + { + "epoch": 0.4728868885411722, + "grad_norm": 0.5073685050010681, + "learning_rate": 0.00010854535784895488, + "loss": 0.0844, + "step": 20173 + }, + { + "epoch": 0.4729103301159772, + "grad_norm": 0.4126063585281372, + "learning_rate": 0.00010853801861222431, + "loss": 0.1111, + "step": 20174 + }, + { + "epoch": 0.4729337716907822, + "grad_norm": 0.5142242908477783, + "learning_rate": 0.00010853067932916621, + "loss": 0.1237, + "step": 20175 + }, + { + "epoch": 0.4729572132655872, + "grad_norm": 0.3292841911315918, + "learning_rate": 0.00010852333999982039, + "loss": 0.0483, + "step": 20176 + }, + { + "epoch": 0.4729806548403922, + "grad_norm": 0.3438548743724823, + "learning_rate": 0.00010851600062422667, + "loss": 0.0852, + "step": 20177 + }, + { + "epoch": 0.4730040964151972, + "grad_norm": 0.1796315759420395, + "learning_rate": 0.00010850866120242491, + "loss": 0.0374, + "step": 20178 + }, + { + "epoch": 0.4730275379900022, + "grad_norm": 0.4001437723636627, + "learning_rate": 0.00010850132173445492, + "loss": 0.0748, + "step": 20179 + }, + { + "epoch": 0.4730509795648072, + "grad_norm": 0.6361743807792664, + "learning_rate": 0.00010849398222035651, + "loss": 0.1185, + "step": 20180 + }, + { + "epoch": 0.4730744211396122, + "grad_norm": 0.45117253065109253, + "learning_rate": 0.00010848664266016955, + "loss": 0.1223, + "step": 20181 + }, + { + "epoch": 0.47309786271441717, + "grad_norm": 0.35965919494628906, + "learning_rate": 0.00010847930305393378, + "loss": 0.2932, + "step": 20182 + }, + { + "epoch": 0.47312130428922217, + "grad_norm": 0.5576571226119995, + "learning_rate": 0.00010847196340168908, + "loss": 0.1202, + "step": 20183 + }, + { + "epoch": 0.47314474586402716, + "grad_norm": 0.8207151293754578, + "learning_rate": 0.00010846462370347527, + "loss": 0.198, + "step": 20184 + }, + { + "epoch": 0.47316818743883216, + "grad_norm": 0.8227134943008423, + "learning_rate": 0.00010845728395933221, + "loss": 0.1219, + "step": 20185 + }, + { + "epoch": 0.47319162901363715, + "grad_norm": 0.7539254426956177, + "learning_rate": 0.00010844994416929966, + "loss": 0.1335, + "step": 20186 + }, + { + "epoch": 0.47321507058844214, + "grad_norm": 0.1750892847776413, + "learning_rate": 0.00010844260433341749, + "loss": 0.0427, + "step": 20187 + }, + { + "epoch": 0.47323851216324714, + "grad_norm": 0.29201868176460266, + "learning_rate": 0.00010843526445172549, + "loss": 0.0471, + "step": 20188 + }, + { + "epoch": 0.47326195373805213, + "grad_norm": 0.4854620695114136, + "learning_rate": 0.00010842792452426349, + "loss": 0.1524, + "step": 20189 + }, + { + "epoch": 0.4732853953128571, + "grad_norm": 0.6052225232124329, + "learning_rate": 0.00010842058455107137, + "loss": 0.0853, + "step": 20190 + }, + { + "epoch": 0.4733088368876621, + "grad_norm": 0.628575325012207, + "learning_rate": 0.00010841324453218892, + "loss": 0.7318, + "step": 20191 + }, + { + "epoch": 0.4733322784624671, + "grad_norm": 0.716803252696991, + "learning_rate": 0.00010840590446765595, + "loss": 0.5472, + "step": 20192 + }, + { + "epoch": 0.4733557200372721, + "grad_norm": 0.9957521557807922, + "learning_rate": 0.00010839856435751234, + "loss": 0.091, + "step": 20193 + }, + { + "epoch": 0.4733791616120771, + "grad_norm": 0.7564340829849243, + "learning_rate": 0.00010839122420179789, + "loss": 0.0837, + "step": 20194 + }, + { + "epoch": 0.4734026031868821, + "grad_norm": 0.34614598751068115, + "learning_rate": 0.00010838388400055237, + "loss": 0.0457, + "step": 20195 + }, + { + "epoch": 0.4734260447616871, + "grad_norm": 0.3044460713863373, + "learning_rate": 0.00010837654375381565, + "loss": 0.0728, + "step": 20196 + }, + { + "epoch": 0.4734494863364921, + "grad_norm": 0.262210488319397, + "learning_rate": 0.00010836920346162762, + "loss": 0.0408, + "step": 20197 + }, + { + "epoch": 0.4734729279112971, + "grad_norm": 0.5419490933418274, + "learning_rate": 0.00010836186312402805, + "loss": 0.082, + "step": 20198 + }, + { + "epoch": 0.4734963694861021, + "grad_norm": 0.48275288939476013, + "learning_rate": 0.00010835452274105675, + "loss": 0.1131, + "step": 20199 + }, + { + "epoch": 0.47351981106090707, + "grad_norm": 0.24857288599014282, + "learning_rate": 0.00010834718231275361, + "loss": 0.0303, + "step": 20200 + }, + { + "epoch": 0.47354325263571206, + "grad_norm": 0.4774693250656128, + "learning_rate": 0.00010833984183915837, + "loss": 0.0567, + "step": 20201 + }, + { + "epoch": 0.47356669421051706, + "grad_norm": 0.5983973741531372, + "learning_rate": 0.00010833250132031094, + "loss": 0.7034, + "step": 20202 + }, + { + "epoch": 0.47359013578532205, + "grad_norm": 0.4597349464893341, + "learning_rate": 0.00010832516075625114, + "loss": 0.0907, + "step": 20203 + }, + { + "epoch": 0.47361357736012705, + "grad_norm": 0.26300108432769775, + "learning_rate": 0.0001083178201470188, + "loss": 0.0512, + "step": 20204 + }, + { + "epoch": 0.47363701893493204, + "grad_norm": 0.5379002094268799, + "learning_rate": 0.0001083104794926537, + "loss": 0.1264, + "step": 20205 + }, + { + "epoch": 0.47366046050973704, + "grad_norm": 0.48884913325309753, + "learning_rate": 0.00010830313879319571, + "loss": 0.1618, + "step": 20206 + }, + { + "epoch": 0.47368390208454203, + "grad_norm": 0.4646918475627899, + "learning_rate": 0.00010829579804868464, + "loss": 0.1267, + "step": 20207 + }, + { + "epoch": 0.473707343659347, + "grad_norm": 0.48846668004989624, + "learning_rate": 0.00010828845725916035, + "loss": 0.0918, + "step": 20208 + }, + { + "epoch": 0.473730785234152, + "grad_norm": 0.7196761965751648, + "learning_rate": 0.00010828111642466268, + "loss": 0.0678, + "step": 20209 + }, + { + "epoch": 0.473754226808957, + "grad_norm": 0.5436921119689941, + "learning_rate": 0.00010827377554523141, + "loss": 0.0859, + "step": 20210 + }, + { + "epoch": 0.473777668383762, + "grad_norm": 0.5455275177955627, + "learning_rate": 0.00010826643462090643, + "loss": 0.0461, + "step": 20211 + }, + { + "epoch": 0.473801109958567, + "grad_norm": 0.626663863658905, + "learning_rate": 0.00010825909365172751, + "loss": 0.1287, + "step": 20212 + }, + { + "epoch": 0.473824551533372, + "grad_norm": 0.5195255279541016, + "learning_rate": 0.00010825175263773453, + "loss": 0.1171, + "step": 20213 + }, + { + "epoch": 0.473847993108177, + "grad_norm": 0.6714012622833252, + "learning_rate": 0.00010824441157896732, + "loss": 0.1073, + "step": 20214 + }, + { + "epoch": 0.473871434682982, + "grad_norm": 0.31533172726631165, + "learning_rate": 0.00010823707047546567, + "loss": 0.0535, + "step": 20215 + }, + { + "epoch": 0.473894876257787, + "grad_norm": 0.3208502233028412, + "learning_rate": 0.00010822972932726947, + "loss": 0.0656, + "step": 20216 + }, + { + "epoch": 0.473918317832592, + "grad_norm": 0.5747443437576294, + "learning_rate": 0.00010822238813441853, + "loss": 0.0478, + "step": 20217 + }, + { + "epoch": 0.47394175940739697, + "grad_norm": 0.8520621061325073, + "learning_rate": 0.00010821504689695268, + "loss": 0.1923, + "step": 20218 + }, + { + "epoch": 0.47396520098220196, + "grad_norm": 0.3552887737751007, + "learning_rate": 0.00010820770561491176, + "loss": 0.0766, + "step": 20219 + }, + { + "epoch": 0.47398864255700696, + "grad_norm": 0.1537596583366394, + "learning_rate": 0.00010820036428833555, + "loss": 0.0093, + "step": 20220 + }, + { + "epoch": 0.47401208413181195, + "grad_norm": 0.3330828547477722, + "learning_rate": 0.00010819302291726397, + "loss": 0.059, + "step": 20221 + }, + { + "epoch": 0.47403552570661694, + "grad_norm": 0.38412341475486755, + "learning_rate": 0.00010818568150173682, + "loss": 0.0753, + "step": 20222 + }, + { + "epoch": 0.47405896728142194, + "grad_norm": 0.6271442174911499, + "learning_rate": 0.00010817834004179395, + "loss": 0.0949, + "step": 20223 + }, + { + "epoch": 0.474082408856227, + "grad_norm": 0.6476218700408936, + "learning_rate": 0.00010817099853747515, + "loss": 0.0757, + "step": 20224 + }, + { + "epoch": 0.474105850431032, + "grad_norm": 0.32753631472587585, + "learning_rate": 0.0001081636569888203, + "loss": 0.0544, + "step": 20225 + }, + { + "epoch": 0.474129292005837, + "grad_norm": 0.6926935315132141, + "learning_rate": 0.00010815631539586917, + "loss": 0.1021, + "step": 20226 + }, + { + "epoch": 0.47415273358064197, + "grad_norm": 0.7115145921707153, + "learning_rate": 0.00010814897375866166, + "loss": 0.1345, + "step": 20227 + }, + { + "epoch": 0.47417617515544697, + "grad_norm": 0.4504236578941345, + "learning_rate": 0.00010814163207723762, + "loss": 0.1041, + "step": 20228 + }, + { + "epoch": 0.47419961673025196, + "grad_norm": 0.31782275438308716, + "learning_rate": 0.00010813429035163686, + "loss": 0.0334, + "step": 20229 + }, + { + "epoch": 0.47422305830505695, + "grad_norm": 0.5903654098510742, + "learning_rate": 0.00010812694858189917, + "loss": 0.6214, + "step": 20230 + }, + { + "epoch": 0.47424649987986195, + "grad_norm": 0.3043060302734375, + "learning_rate": 0.00010811960676806448, + "loss": 0.0401, + "step": 20231 + }, + { + "epoch": 0.47426994145466694, + "grad_norm": 0.36345911026000977, + "learning_rate": 0.00010811226491017254, + "loss": 0.0212, + "step": 20232 + }, + { + "epoch": 0.47429338302947194, + "grad_norm": 0.44445085525512695, + "learning_rate": 0.00010810492300826323, + "loss": 0.0794, + "step": 20233 + }, + { + "epoch": 0.47431682460427693, + "grad_norm": 0.5342833995819092, + "learning_rate": 0.00010809758106237637, + "loss": 0.1076, + "step": 20234 + }, + { + "epoch": 0.4743402661790819, + "grad_norm": 0.38129499554634094, + "learning_rate": 0.00010809023907255183, + "loss": 0.0711, + "step": 20235 + }, + { + "epoch": 0.4743637077538869, + "grad_norm": 0.4242454171180725, + "learning_rate": 0.00010808289703882943, + "loss": 0.1432, + "step": 20236 + }, + { + "epoch": 0.4743871493286919, + "grad_norm": 0.43048179149627686, + "learning_rate": 0.00010807555496124896, + "loss": 0.0499, + "step": 20237 + }, + { + "epoch": 0.4744105909034969, + "grad_norm": 0.30638429522514343, + "learning_rate": 0.00010806821283985034, + "loss": 0.0546, + "step": 20238 + }, + { + "epoch": 0.4744340324783019, + "grad_norm": 0.467538446187973, + "learning_rate": 0.00010806087067467337, + "loss": 0.0772, + "step": 20239 + }, + { + "epoch": 0.4744574740531069, + "grad_norm": 0.18845708668231964, + "learning_rate": 0.00010805352846575785, + "loss": 0.0287, + "step": 20240 + }, + { + "epoch": 0.4744809156279119, + "grad_norm": 0.3592550754547119, + "learning_rate": 0.0001080461862131437, + "loss": 0.0469, + "step": 20241 + }, + { + "epoch": 0.4745043572027169, + "grad_norm": 0.4856680929660797, + "learning_rate": 0.00010803884391687071, + "loss": 0.1015, + "step": 20242 + }, + { + "epoch": 0.4745277987775219, + "grad_norm": 0.6514611840248108, + "learning_rate": 0.0001080315015769787, + "loss": 0.063, + "step": 20243 + }, + { + "epoch": 0.4745512403523269, + "grad_norm": 0.3949550688266754, + "learning_rate": 0.00010802415919350755, + "loss": 0.0713, + "step": 20244 + }, + { + "epoch": 0.47457468192713187, + "grad_norm": 0.40146276354789734, + "learning_rate": 0.0001080168167664971, + "loss": 0.0764, + "step": 20245 + }, + { + "epoch": 0.47459812350193686, + "grad_norm": 0.5637713670730591, + "learning_rate": 0.00010800947429598719, + "loss": 0.0544, + "step": 20246 + }, + { + "epoch": 0.47462156507674186, + "grad_norm": 0.48772722482681274, + "learning_rate": 0.00010800213178201762, + "loss": 0.1387, + "step": 20247 + }, + { + "epoch": 0.47464500665154685, + "grad_norm": 0.32132044434547424, + "learning_rate": 0.00010799478922462827, + "loss": 0.3317, + "step": 20248 + }, + { + "epoch": 0.47466844822635185, + "grad_norm": 0.6801361441612244, + "learning_rate": 0.00010798744662385896, + "loss": 0.2178, + "step": 20249 + }, + { + "epoch": 0.47469188980115684, + "grad_norm": 0.7274887561798096, + "learning_rate": 0.00010798010397974954, + "loss": 0.0871, + "step": 20250 + }, + { + "epoch": 0.47471533137596184, + "grad_norm": 0.6506035327911377, + "learning_rate": 0.00010797276129233984, + "loss": 0.6177, + "step": 20251 + }, + { + "epoch": 0.47473877295076683, + "grad_norm": 0.36660608649253845, + "learning_rate": 0.00010796541856166975, + "loss": 0.0554, + "step": 20252 + }, + { + "epoch": 0.4747622145255718, + "grad_norm": 0.32393980026245117, + "learning_rate": 0.00010795807578777903, + "loss": 0.0636, + "step": 20253 + }, + { + "epoch": 0.4747856561003768, + "grad_norm": 0.39454519748687744, + "learning_rate": 0.00010795073297070762, + "loss": 0.0402, + "step": 20254 + }, + { + "epoch": 0.4748090976751818, + "grad_norm": 0.7958886027336121, + "learning_rate": 0.00010794339011049528, + "loss": 0.1114, + "step": 20255 + }, + { + "epoch": 0.4748325392499868, + "grad_norm": 0.5753253102302551, + "learning_rate": 0.00010793604720718184, + "loss": 0.0806, + "step": 20256 + }, + { + "epoch": 0.4748559808247918, + "grad_norm": 0.16911627352237701, + "learning_rate": 0.00010792870426080725, + "loss": 0.0201, + "step": 20257 + }, + { + "epoch": 0.4748794223995968, + "grad_norm": 0.5976841449737549, + "learning_rate": 0.00010792136127141128, + "loss": 0.0777, + "step": 20258 + }, + { + "epoch": 0.4749028639744018, + "grad_norm": 0.13116493821144104, + "learning_rate": 0.00010791401823903375, + "loss": 0.0192, + "step": 20259 + }, + { + "epoch": 0.4749263055492068, + "grad_norm": 0.3948703110218048, + "learning_rate": 0.00010790667516371455, + "loss": 0.1263, + "step": 20260 + }, + { + "epoch": 0.4749497471240118, + "grad_norm": 0.35825785994529724, + "learning_rate": 0.00010789933204549352, + "loss": 0.0805, + "step": 20261 + }, + { + "epoch": 0.4749731886988168, + "grad_norm": 0.16989727318286896, + "learning_rate": 0.00010789198888441048, + "loss": 0.0437, + "step": 20262 + }, + { + "epoch": 0.47499663027362177, + "grad_norm": 0.9162877798080444, + "learning_rate": 0.00010788464568050526, + "loss": 0.1041, + "step": 20263 + }, + { + "epoch": 0.47502007184842676, + "grad_norm": 0.3104822039604187, + "learning_rate": 0.00010787730243381777, + "loss": 0.0407, + "step": 20264 + }, + { + "epoch": 0.47504351342323176, + "grad_norm": 0.5366043448448181, + "learning_rate": 0.0001078699591443878, + "loss": 0.1085, + "step": 20265 + }, + { + "epoch": 0.47506695499803675, + "grad_norm": 0.6133686304092407, + "learning_rate": 0.00010786261581225523, + "loss": 0.4606, + "step": 20266 + }, + { + "epoch": 0.47509039657284174, + "grad_norm": 0.3697766363620758, + "learning_rate": 0.00010785527243745987, + "loss": 0.0828, + "step": 20267 + }, + { + "epoch": 0.47511383814764674, + "grad_norm": 0.45049938559532166, + "learning_rate": 0.00010784792902004157, + "loss": 0.0548, + "step": 20268 + }, + { + "epoch": 0.47513727972245173, + "grad_norm": 0.5432690382003784, + "learning_rate": 0.00010784058556004018, + "loss": 0.1037, + "step": 20269 + }, + { + "epoch": 0.4751607212972567, + "grad_norm": 0.5350978374481201, + "learning_rate": 0.00010783324205749558, + "loss": 0.1986, + "step": 20270 + }, + { + "epoch": 0.4751841628720617, + "grad_norm": 0.11822153627872467, + "learning_rate": 0.0001078258985124476, + "loss": 0.03, + "step": 20271 + }, + { + "epoch": 0.4752076044468667, + "grad_norm": 0.5978589057922363, + "learning_rate": 0.00010781855492493602, + "loss": 0.5625, + "step": 20272 + }, + { + "epoch": 0.4752310460216717, + "grad_norm": 0.4121100604534149, + "learning_rate": 0.0001078112112950008, + "loss": 0.1182, + "step": 20273 + }, + { + "epoch": 0.4752544875964767, + "grad_norm": 0.39312243461608887, + "learning_rate": 0.00010780386762268169, + "loss": 0.0506, + "step": 20274 + }, + { + "epoch": 0.47527792917128175, + "grad_norm": 0.4296286106109619, + "learning_rate": 0.00010779652390801858, + "loss": 0.102, + "step": 20275 + }, + { + "epoch": 0.47530137074608675, + "grad_norm": 0.21150697767734528, + "learning_rate": 0.00010778918015105131, + "loss": 0.0345, + "step": 20276 + }, + { + "epoch": 0.47532481232089174, + "grad_norm": 0.2908577024936676, + "learning_rate": 0.00010778183635181975, + "loss": 0.053, + "step": 20277 + }, + { + "epoch": 0.47534825389569674, + "grad_norm": 0.14350484311580658, + "learning_rate": 0.00010777449251036371, + "loss": 0.0319, + "step": 20278 + }, + { + "epoch": 0.47537169547050173, + "grad_norm": 0.4003809094429016, + "learning_rate": 0.00010776714862672306, + "loss": 0.0442, + "step": 20279 + }, + { + "epoch": 0.4753951370453067, + "grad_norm": 0.7524863481521606, + "learning_rate": 0.00010775980470093762, + "loss": 0.0686, + "step": 20280 + }, + { + "epoch": 0.4754185786201117, + "grad_norm": 0.4497239589691162, + "learning_rate": 0.0001077524607330473, + "loss": 0.1407, + "step": 20281 + }, + { + "epoch": 0.4754420201949167, + "grad_norm": 0.3384794592857361, + "learning_rate": 0.00010774511672309189, + "loss": 0.0987, + "step": 20282 + }, + { + "epoch": 0.4754654617697217, + "grad_norm": 0.5707058906555176, + "learning_rate": 0.00010773777267111127, + "loss": 0.1213, + "step": 20283 + }, + { + "epoch": 0.4754889033445267, + "grad_norm": 0.4246211349964142, + "learning_rate": 0.00010773042857714526, + "loss": 0.054, + "step": 20284 + }, + { + "epoch": 0.4755123449193317, + "grad_norm": 0.1635132133960724, + "learning_rate": 0.00010772308444123376, + "loss": 0.044, + "step": 20285 + }, + { + "epoch": 0.4755357864941367, + "grad_norm": 0.5691508650779724, + "learning_rate": 0.00010771574026341656, + "loss": 0.1711, + "step": 20286 + }, + { + "epoch": 0.4755592280689417, + "grad_norm": 0.22442413866519928, + "learning_rate": 0.00010770839604373353, + "loss": 0.0228, + "step": 20287 + }, + { + "epoch": 0.4755826696437467, + "grad_norm": 0.33304259181022644, + "learning_rate": 0.00010770105178222454, + "loss": 0.0307, + "step": 20288 + }, + { + "epoch": 0.4756061112185517, + "grad_norm": 0.1545875072479248, + "learning_rate": 0.00010769370747892942, + "loss": 0.0222, + "step": 20289 + }, + { + "epoch": 0.47562955279335667, + "grad_norm": 0.18071630597114563, + "learning_rate": 0.00010768636313388804, + "loss": 0.0603, + "step": 20290 + }, + { + "epoch": 0.47565299436816166, + "grad_norm": 0.41595786809921265, + "learning_rate": 0.00010767901874714022, + "loss": 0.0411, + "step": 20291 + }, + { + "epoch": 0.47567643594296666, + "grad_norm": 0.1674448549747467, + "learning_rate": 0.00010767167431872582, + "loss": 0.0479, + "step": 20292 + }, + { + "epoch": 0.47569987751777165, + "grad_norm": 0.45236924290657043, + "learning_rate": 0.00010766432984868473, + "loss": 0.0673, + "step": 20293 + }, + { + "epoch": 0.47572331909257665, + "grad_norm": 0.6442564129829407, + "learning_rate": 0.00010765698533705675, + "loss": 0.1348, + "step": 20294 + }, + { + "epoch": 0.47574676066738164, + "grad_norm": 0.6599524617195129, + "learning_rate": 0.00010764964078388178, + "loss": 0.0847, + "step": 20295 + }, + { + "epoch": 0.47577020224218664, + "grad_norm": 0.4416791498661041, + "learning_rate": 0.00010764229618919963, + "loss": 0.1282, + "step": 20296 + }, + { + "epoch": 0.47579364381699163, + "grad_norm": 0.4191441237926483, + "learning_rate": 0.00010763495155305015, + "loss": 0.0654, + "step": 20297 + }, + { + "epoch": 0.4758170853917966, + "grad_norm": 0.5356311202049255, + "learning_rate": 0.00010762760687547322, + "loss": 0.1241, + "step": 20298 + }, + { + "epoch": 0.4758405269666016, + "grad_norm": 0.19440512359142303, + "learning_rate": 0.00010762026215650866, + "loss": 0.0267, + "step": 20299 + }, + { + "epoch": 0.4758639685414066, + "grad_norm": 0.601672351360321, + "learning_rate": 0.00010761291739619636, + "loss": 0.0994, + "step": 20300 + }, + { + "epoch": 0.4758874101162116, + "grad_norm": 0.11938919872045517, + "learning_rate": 0.00010760557259457615, + "loss": 0.0228, + "step": 20301 + }, + { + "epoch": 0.4759108516910166, + "grad_norm": 0.2799374759197235, + "learning_rate": 0.0001075982277516879, + "loss": 0.0483, + "step": 20302 + }, + { + "epoch": 0.4759342932658216, + "grad_norm": 0.7609199285507202, + "learning_rate": 0.00010759088286757147, + "loss": 0.0684, + "step": 20303 + }, + { + "epoch": 0.4759577348406266, + "grad_norm": 0.45727360248565674, + "learning_rate": 0.00010758353794226668, + "loss": 0.0816, + "step": 20304 + }, + { + "epoch": 0.4759811764154316, + "grad_norm": 0.5181230306625366, + "learning_rate": 0.00010757619297581339, + "loss": 0.0777, + "step": 20305 + }, + { + "epoch": 0.4760046179902366, + "grad_norm": 0.6343857049942017, + "learning_rate": 0.00010756884796825147, + "loss": 0.133, + "step": 20306 + }, + { + "epoch": 0.4760280595650416, + "grad_norm": 0.3833533227443695, + "learning_rate": 0.00010756150291962075, + "loss": 0.1013, + "step": 20307 + }, + { + "epoch": 0.47605150113984657, + "grad_norm": 0.4899483621120453, + "learning_rate": 0.00010755415782996112, + "loss": 0.0774, + "step": 20308 + }, + { + "epoch": 0.47607494271465156, + "grad_norm": 0.5175746083259583, + "learning_rate": 0.00010754681269931243, + "loss": 0.1458, + "step": 20309 + }, + { + "epoch": 0.47609838428945656, + "grad_norm": 0.6211109757423401, + "learning_rate": 0.00010753946752771452, + "loss": 0.0683, + "step": 20310 + }, + { + "epoch": 0.47612182586426155, + "grad_norm": 0.45602947473526, + "learning_rate": 0.00010753212231520721, + "loss": 0.0798, + "step": 20311 + }, + { + "epoch": 0.47614526743906654, + "grad_norm": 0.34056854248046875, + "learning_rate": 0.00010752477706183045, + "loss": 0.0798, + "step": 20312 + }, + { + "epoch": 0.47616870901387154, + "grad_norm": 0.26293978095054626, + "learning_rate": 0.00010751743176762397, + "loss": 0.0596, + "step": 20313 + }, + { + "epoch": 0.47619215058867653, + "grad_norm": 0.33864012360572815, + "learning_rate": 0.00010751008643262774, + "loss": 0.0746, + "step": 20314 + }, + { + "epoch": 0.4762155921634815, + "grad_norm": 0.5842623114585876, + "learning_rate": 0.00010750274105688158, + "loss": 0.1388, + "step": 20315 + }, + { + "epoch": 0.4762390337382865, + "grad_norm": 0.3096619248390198, + "learning_rate": 0.0001074953956404253, + "loss": 0.0623, + "step": 20316 + }, + { + "epoch": 0.4762624753130915, + "grad_norm": 0.7100460529327393, + "learning_rate": 0.0001074880501832988, + "loss": 0.1266, + "step": 20317 + }, + { + "epoch": 0.4762859168878965, + "grad_norm": 0.7619110941886902, + "learning_rate": 0.00010748070468554194, + "loss": 0.1474, + "step": 20318 + }, + { + "epoch": 0.4763093584627015, + "grad_norm": 0.16279828548431396, + "learning_rate": 0.00010747335914719459, + "loss": 0.0307, + "step": 20319 + }, + { + "epoch": 0.4763328000375065, + "grad_norm": 0.34300971031188965, + "learning_rate": 0.00010746601356829653, + "loss": 0.0583, + "step": 20320 + }, + { + "epoch": 0.4763562416123115, + "grad_norm": 0.5203723311424255, + "learning_rate": 0.00010745866794888773, + "loss": 0.0807, + "step": 20321 + }, + { + "epoch": 0.4763796831871165, + "grad_norm": 0.47619396448135376, + "learning_rate": 0.00010745132228900794, + "loss": 0.0982, + "step": 20322 + }, + { + "epoch": 0.4764031247619215, + "grad_norm": 0.5715903043746948, + "learning_rate": 0.00010744397658869707, + "loss": 0.0361, + "step": 20323 + }, + { + "epoch": 0.4764265663367265, + "grad_norm": 0.5331969857215881, + "learning_rate": 0.00010743663084799498, + "loss": 0.07, + "step": 20324 + }, + { + "epoch": 0.47645000791153147, + "grad_norm": 0.8678425550460815, + "learning_rate": 0.00010742928506694155, + "loss": 0.1523, + "step": 20325 + }, + { + "epoch": 0.47647344948633646, + "grad_norm": 0.22538131475448608, + "learning_rate": 0.00010742193924557658, + "loss": 0.0434, + "step": 20326 + }, + { + "epoch": 0.4764968910611415, + "grad_norm": 0.15986360609531403, + "learning_rate": 0.00010741459338393998, + "loss": 0.0083, + "step": 20327 + }, + { + "epoch": 0.4765203326359465, + "grad_norm": 0.2572174370288849, + "learning_rate": 0.00010740724748207156, + "loss": 0.0588, + "step": 20328 + }, + { + "epoch": 0.4765437742107515, + "grad_norm": 0.5137073397636414, + "learning_rate": 0.0001073999015400112, + "loss": 0.0796, + "step": 20329 + }, + { + "epoch": 0.4765672157855565, + "grad_norm": 0.3484264016151428, + "learning_rate": 0.00010739255555779878, + "loss": 0.0477, + "step": 20330 + }, + { + "epoch": 0.4765906573603615, + "grad_norm": 0.47683843970298767, + "learning_rate": 0.00010738520953547418, + "loss": 0.1224, + "step": 20331 + }, + { + "epoch": 0.4766140989351665, + "grad_norm": 0.3694380521774292, + "learning_rate": 0.00010737786347307719, + "loss": 0.0493, + "step": 20332 + }, + { + "epoch": 0.4766375405099715, + "grad_norm": 0.3770787715911865, + "learning_rate": 0.00010737051737064772, + "loss": 0.0731, + "step": 20333 + }, + { + "epoch": 0.4766609820847765, + "grad_norm": 0.23681840300559998, + "learning_rate": 0.00010736317122822562, + "loss": 0.0378, + "step": 20334 + }, + { + "epoch": 0.47668442365958147, + "grad_norm": 0.5782850384712219, + "learning_rate": 0.00010735582504585074, + "loss": 0.5597, + "step": 20335 + }, + { + "epoch": 0.47670786523438646, + "grad_norm": 0.9659883975982666, + "learning_rate": 0.00010734847882356291, + "loss": 0.1386, + "step": 20336 + }, + { + "epoch": 0.47673130680919146, + "grad_norm": 0.48935604095458984, + "learning_rate": 0.00010734113256140208, + "loss": 0.0687, + "step": 20337 + }, + { + "epoch": 0.47675474838399645, + "grad_norm": 0.3610857427120209, + "learning_rate": 0.00010733378625940801, + "loss": 0.0635, + "step": 20338 + }, + { + "epoch": 0.47677818995880145, + "grad_norm": 0.4017939865589142, + "learning_rate": 0.00010732643991762066, + "loss": 0.1015, + "step": 20339 + }, + { + "epoch": 0.47680163153360644, + "grad_norm": 0.35484209656715393, + "learning_rate": 0.00010731909353607984, + "loss": 0.0937, + "step": 20340 + }, + { + "epoch": 0.47682507310841143, + "grad_norm": 0.629790723323822, + "learning_rate": 0.00010731174711482537, + "loss": 0.7789, + "step": 20341 + }, + { + "epoch": 0.47684851468321643, + "grad_norm": 0.362526535987854, + "learning_rate": 0.00010730440065389716, + "loss": 0.0551, + "step": 20342 + }, + { + "epoch": 0.4768719562580214, + "grad_norm": 0.5552268624305725, + "learning_rate": 0.00010729705415333508, + "loss": 0.5251, + "step": 20343 + }, + { + "epoch": 0.4768953978328264, + "grad_norm": 0.34316402673721313, + "learning_rate": 0.000107289707613179, + "loss": 0.0771, + "step": 20344 + }, + { + "epoch": 0.4769188394076314, + "grad_norm": 0.29884541034698486, + "learning_rate": 0.00010728236103346874, + "loss": 0.0668, + "step": 20345 + }, + { + "epoch": 0.4769422809824364, + "grad_norm": 0.6410134434700012, + "learning_rate": 0.00010727501441424419, + "loss": 0.159, + "step": 20346 + }, + { + "epoch": 0.4769657225572414, + "grad_norm": 0.702966570854187, + "learning_rate": 0.0001072676677555452, + "loss": 0.137, + "step": 20347 + }, + { + "epoch": 0.4769891641320464, + "grad_norm": 0.2415502369403839, + "learning_rate": 0.00010726032105741161, + "loss": 0.0415, + "step": 20348 + }, + { + "epoch": 0.4770126057068514, + "grad_norm": 0.2255169302225113, + "learning_rate": 0.00010725297431988337, + "loss": 0.0185, + "step": 20349 + }, + { + "epoch": 0.4770360472816564, + "grad_norm": 0.7730315923690796, + "learning_rate": 0.00010724562754300027, + "loss": 0.1145, + "step": 20350 + }, + { + "epoch": 0.4770594888564614, + "grad_norm": 0.41105493903160095, + "learning_rate": 0.00010723828072680218, + "loss": 0.0997, + "step": 20351 + }, + { + "epoch": 0.47708293043126637, + "grad_norm": 0.3436294496059418, + "learning_rate": 0.00010723093387132898, + "loss": 0.0641, + "step": 20352 + }, + { + "epoch": 0.47710637200607137, + "grad_norm": 0.20667658746242523, + "learning_rate": 0.00010722358697662053, + "loss": 0.0411, + "step": 20353 + }, + { + "epoch": 0.47712981358087636, + "grad_norm": 0.5750250220298767, + "learning_rate": 0.00010721624004271671, + "loss": 0.6902, + "step": 20354 + }, + { + "epoch": 0.47715325515568136, + "grad_norm": 0.5129160284996033, + "learning_rate": 0.00010720889306965734, + "loss": 0.1552, + "step": 20355 + }, + { + "epoch": 0.47717669673048635, + "grad_norm": 0.5505222082138062, + "learning_rate": 0.00010720154605748233, + "loss": 0.821, + "step": 20356 + }, + { + "epoch": 0.47720013830529134, + "grad_norm": 0.6682679057121277, + "learning_rate": 0.00010719419900623153, + "loss": 0.1733, + "step": 20357 + }, + { + "epoch": 0.47722357988009634, + "grad_norm": 0.15910297632217407, + "learning_rate": 0.0001071868519159448, + "loss": 0.0217, + "step": 20358 + }, + { + "epoch": 0.47724702145490133, + "grad_norm": 0.3200729489326477, + "learning_rate": 0.00010717950478666201, + "loss": 0.0663, + "step": 20359 + }, + { + "epoch": 0.4772704630297063, + "grad_norm": 0.44228121638298035, + "learning_rate": 0.00010717215761842303, + "loss": 0.1348, + "step": 20360 + }, + { + "epoch": 0.4772939046045113, + "grad_norm": 0.4846785366535187, + "learning_rate": 0.0001071648104112677, + "loss": 0.1287, + "step": 20361 + }, + { + "epoch": 0.4773173461793163, + "grad_norm": 0.17143042385578156, + "learning_rate": 0.00010715746316523593, + "loss": 0.0341, + "step": 20362 + }, + { + "epoch": 0.4773407877541213, + "grad_norm": 0.2691715955734253, + "learning_rate": 0.00010715011588036756, + "loss": 0.0501, + "step": 20363 + }, + { + "epoch": 0.4773642293289263, + "grad_norm": 0.1885516345500946, + "learning_rate": 0.00010714276855670245, + "loss": 0.0325, + "step": 20364 + }, + { + "epoch": 0.4773876709037313, + "grad_norm": 0.4185888171195984, + "learning_rate": 0.00010713542119428049, + "loss": 0.0805, + "step": 20365 + }, + { + "epoch": 0.4774111124785363, + "grad_norm": 0.5368296504020691, + "learning_rate": 0.00010712807379314154, + "loss": 0.1346, + "step": 20366 + }, + { + "epoch": 0.4774345540533413, + "grad_norm": 0.2725452482700348, + "learning_rate": 0.00010712072635332542, + "loss": 0.0471, + "step": 20367 + }, + { + "epoch": 0.4774579956281463, + "grad_norm": 0.33311304450035095, + "learning_rate": 0.00010711337887487207, + "loss": 0.0359, + "step": 20368 + }, + { + "epoch": 0.4774814372029513, + "grad_norm": 0.5787428617477417, + "learning_rate": 0.00010710603135782133, + "loss": 0.072, + "step": 20369 + }, + { + "epoch": 0.47750487877775627, + "grad_norm": 0.13524514436721802, + "learning_rate": 0.00010709868380221306, + "loss": 0.0249, + "step": 20370 + }, + { + "epoch": 0.47752832035256126, + "grad_norm": 0.7368330359458923, + "learning_rate": 0.00010709133620808714, + "loss": 1.0186, + "step": 20371 + }, + { + "epoch": 0.47755176192736626, + "grad_norm": 0.27530115842819214, + "learning_rate": 0.0001070839885754834, + "loss": 0.0296, + "step": 20372 + }, + { + "epoch": 0.47757520350217125, + "grad_norm": 0.5238315463066101, + "learning_rate": 0.00010707664090444178, + "loss": 0.6666, + "step": 20373 + }, + { + "epoch": 0.47759864507697625, + "grad_norm": 1.0424643754959106, + "learning_rate": 0.00010706929319500208, + "loss": 0.134, + "step": 20374 + }, + { + "epoch": 0.47762208665178124, + "grad_norm": 0.2269383817911148, + "learning_rate": 0.0001070619454472042, + "loss": 0.0362, + "step": 20375 + }, + { + "epoch": 0.47764552822658624, + "grad_norm": 0.5722994208335876, + "learning_rate": 0.00010705459766108801, + "loss": 0.1788, + "step": 20376 + }, + { + "epoch": 0.47766896980139123, + "grad_norm": 0.2810693681240082, + "learning_rate": 0.00010704724983669339, + "loss": 0.0673, + "step": 20377 + }, + { + "epoch": 0.4776924113761963, + "grad_norm": 0.24394184350967407, + "learning_rate": 0.00010703990197406016, + "loss": 0.0306, + "step": 20378 + }, + { + "epoch": 0.4777158529510013, + "grad_norm": 0.5101906061172485, + "learning_rate": 0.00010703255407322826, + "loss": 0.1173, + "step": 20379 + }, + { + "epoch": 0.47773929452580627, + "grad_norm": 0.14122040569782257, + "learning_rate": 0.00010702520613423747, + "loss": 0.0104, + "step": 20380 + }, + { + "epoch": 0.47776273610061126, + "grad_norm": 0.3599218726158142, + "learning_rate": 0.00010701785815712777, + "loss": 0.052, + "step": 20381 + }, + { + "epoch": 0.47778617767541626, + "grad_norm": 0.1959630846977234, + "learning_rate": 0.00010701051014193897, + "loss": 0.0312, + "step": 20382 + }, + { + "epoch": 0.47780961925022125, + "grad_norm": 0.3065344989299774, + "learning_rate": 0.00010700316208871091, + "loss": 0.0864, + "step": 20383 + }, + { + "epoch": 0.47783306082502625, + "grad_norm": 0.45738232135772705, + "learning_rate": 0.00010699581399748351, + "loss": 0.1156, + "step": 20384 + }, + { + "epoch": 0.47785650239983124, + "grad_norm": 0.3478207588195801, + "learning_rate": 0.00010698846586829663, + "loss": 0.059, + "step": 20385 + }, + { + "epoch": 0.47787994397463623, + "grad_norm": 0.29810500144958496, + "learning_rate": 0.00010698111770119013, + "loss": 0.0403, + "step": 20386 + }, + { + "epoch": 0.47790338554944123, + "grad_norm": 0.2782338261604309, + "learning_rate": 0.0001069737694962039, + "loss": 0.0492, + "step": 20387 + }, + { + "epoch": 0.4779268271242462, + "grad_norm": 0.23769697546958923, + "learning_rate": 0.00010696642125337781, + "loss": 0.0385, + "step": 20388 + }, + { + "epoch": 0.4779502686990512, + "grad_norm": 0.34168583154678345, + "learning_rate": 0.00010695907297275169, + "loss": 0.061, + "step": 20389 + }, + { + "epoch": 0.4779737102738562, + "grad_norm": 0.6721623539924622, + "learning_rate": 0.00010695172465436547, + "loss": 0.0921, + "step": 20390 + }, + { + "epoch": 0.4779971518486612, + "grad_norm": 0.1703735888004303, + "learning_rate": 0.000106944376298259, + "loss": 0.0186, + "step": 20391 + }, + { + "epoch": 0.4780205934234662, + "grad_norm": 1.0209308862686157, + "learning_rate": 0.00010693702790447216, + "loss": 0.2482, + "step": 20392 + }, + { + "epoch": 0.4780440349982712, + "grad_norm": 1.1619174480438232, + "learning_rate": 0.00010692967947304477, + "loss": 0.102, + "step": 20393 + }, + { + "epoch": 0.4780674765730762, + "grad_norm": 0.45429253578186035, + "learning_rate": 0.00010692233100401676, + "loss": 0.0782, + "step": 20394 + }, + { + "epoch": 0.4780909181478812, + "grad_norm": 0.4777520000934601, + "learning_rate": 0.00010691498249742799, + "loss": 0.0661, + "step": 20395 + }, + { + "epoch": 0.4781143597226862, + "grad_norm": 0.3962370753288269, + "learning_rate": 0.00010690763395331833, + "loss": 0.0693, + "step": 20396 + }, + { + "epoch": 0.47813780129749117, + "grad_norm": 0.40982162952423096, + "learning_rate": 0.00010690028537172767, + "loss": 0.0565, + "step": 20397 + }, + { + "epoch": 0.47816124287229617, + "grad_norm": 0.21790650486946106, + "learning_rate": 0.00010689293675269584, + "loss": 0.0579, + "step": 20398 + }, + { + "epoch": 0.47818468444710116, + "grad_norm": 0.34873172640800476, + "learning_rate": 0.00010688558809626277, + "loss": 0.0608, + "step": 20399 + }, + { + "epoch": 0.47820812602190615, + "grad_norm": 0.5431953072547913, + "learning_rate": 0.00010687823940246828, + "loss": 0.6627, + "step": 20400 + }, + { + "epoch": 0.47823156759671115, + "grad_norm": 0.6878191828727722, + "learning_rate": 0.0001068708906713523, + "loss": 0.0918, + "step": 20401 + }, + { + "epoch": 0.47825500917151614, + "grad_norm": 0.38414669036865234, + "learning_rate": 0.00010686354190295462, + "loss": 0.0798, + "step": 20402 + }, + { + "epoch": 0.47827845074632114, + "grad_norm": 0.49195536971092224, + "learning_rate": 0.00010685619309731523, + "loss": 0.0645, + "step": 20403 + }, + { + "epoch": 0.47830189232112613, + "grad_norm": 0.12494263797998428, + "learning_rate": 0.00010684884425447393, + "loss": 0.022, + "step": 20404 + }, + { + "epoch": 0.4783253338959311, + "grad_norm": 0.5810005068778992, + "learning_rate": 0.00010684149537447057, + "loss": 0.6747, + "step": 20405 + }, + { + "epoch": 0.4783487754707361, + "grad_norm": 0.13774146139621735, + "learning_rate": 0.00010683414645734511, + "loss": 0.0259, + "step": 20406 + }, + { + "epoch": 0.4783722170455411, + "grad_norm": 0.39221736788749695, + "learning_rate": 0.00010682679750313735, + "loss": 0.0628, + "step": 20407 + }, + { + "epoch": 0.4783956586203461, + "grad_norm": 0.31327781081199646, + "learning_rate": 0.00010681944851188721, + "loss": 0.0531, + "step": 20408 + }, + { + "epoch": 0.4784191001951511, + "grad_norm": 0.5631215572357178, + "learning_rate": 0.00010681209948363453, + "loss": 0.1617, + "step": 20409 + }, + { + "epoch": 0.4784425417699561, + "grad_norm": 0.6553922891616821, + "learning_rate": 0.00010680475041841922, + "loss": 0.1201, + "step": 20410 + }, + { + "epoch": 0.4784659833447611, + "grad_norm": 0.5094660520553589, + "learning_rate": 0.00010679740131628118, + "loss": 0.6618, + "step": 20411 + }, + { + "epoch": 0.4784894249195661, + "grad_norm": 0.4314234256744385, + "learning_rate": 0.00010679005217726019, + "loss": 0.0874, + "step": 20412 + }, + { + "epoch": 0.4785128664943711, + "grad_norm": 0.2990248203277588, + "learning_rate": 0.00010678270300139621, + "loss": 0.0713, + "step": 20413 + }, + { + "epoch": 0.4785363080691761, + "grad_norm": 0.37326428294181824, + "learning_rate": 0.00010677535378872909, + "loss": 0.0942, + "step": 20414 + }, + { + "epoch": 0.47855974964398107, + "grad_norm": 0.1653972715139389, + "learning_rate": 0.0001067680045392987, + "loss": 0.0417, + "step": 20415 + }, + { + "epoch": 0.47858319121878606, + "grad_norm": 0.22771362960338593, + "learning_rate": 0.00010676065525314496, + "loss": 0.0367, + "step": 20416 + }, + { + "epoch": 0.47860663279359106, + "grad_norm": 0.5894268155097961, + "learning_rate": 0.0001067533059303077, + "loss": 0.1263, + "step": 20417 + }, + { + "epoch": 0.47863007436839605, + "grad_norm": 0.32599371671676636, + "learning_rate": 0.00010674595657082678, + "loss": 0.0977, + "step": 20418 + }, + { + "epoch": 0.47865351594320105, + "grad_norm": 0.6941648125648499, + "learning_rate": 0.00010673860717474216, + "loss": 0.2282, + "step": 20419 + }, + { + "epoch": 0.47867695751800604, + "grad_norm": 0.32021427154541016, + "learning_rate": 0.00010673125774209362, + "loss": 0.0638, + "step": 20420 + }, + { + "epoch": 0.47870039909281104, + "grad_norm": 0.6423079967498779, + "learning_rate": 0.0001067239082729211, + "loss": 0.0931, + "step": 20421 + }, + { + "epoch": 0.47872384066761603, + "grad_norm": 0.2528528571128845, + "learning_rate": 0.0001067165587672645, + "loss": 0.0231, + "step": 20422 + }, + { + "epoch": 0.478747282242421, + "grad_norm": 0.188017800450325, + "learning_rate": 0.00010670920922516363, + "loss": 0.0207, + "step": 20423 + }, + { + "epoch": 0.478770723817226, + "grad_norm": 0.11635629087686539, + "learning_rate": 0.0001067018596466584, + "loss": 0.0287, + "step": 20424 + }, + { + "epoch": 0.478794165392031, + "grad_norm": 0.3679741621017456, + "learning_rate": 0.00010669451003178871, + "loss": 0.0562, + "step": 20425 + }, + { + "epoch": 0.478817606966836, + "grad_norm": 0.09649485349655151, + "learning_rate": 0.00010668716038059442, + "loss": 0.0091, + "step": 20426 + }, + { + "epoch": 0.478841048541641, + "grad_norm": 0.21908879280090332, + "learning_rate": 0.0001066798106931154, + "loss": 0.0286, + "step": 20427 + }, + { + "epoch": 0.478864490116446, + "grad_norm": 0.5378134250640869, + "learning_rate": 0.00010667246096939152, + "loss": 0.0755, + "step": 20428 + }, + { + "epoch": 0.47888793169125105, + "grad_norm": 0.5600655674934387, + "learning_rate": 0.00010666511120946272, + "loss": 0.5165, + "step": 20429 + }, + { + "epoch": 0.47891137326605604, + "grad_norm": 0.6112837791442871, + "learning_rate": 0.00010665776141336883, + "loss": 0.1404, + "step": 20430 + }, + { + "epoch": 0.47893481484086103, + "grad_norm": 0.5764932632446289, + "learning_rate": 0.0001066504115811497, + "loss": 0.0774, + "step": 20431 + }, + { + "epoch": 0.47895825641566603, + "grad_norm": 0.442355751991272, + "learning_rate": 0.00010664306171284529, + "loss": 0.0601, + "step": 20432 + }, + { + "epoch": 0.478981697990471, + "grad_norm": 0.17492204904556274, + "learning_rate": 0.00010663571180849543, + "loss": 0.0087, + "step": 20433 + }, + { + "epoch": 0.479005139565276, + "grad_norm": 0.8522022366523743, + "learning_rate": 0.00010662836186814001, + "loss": 0.2152, + "step": 20434 + }, + { + "epoch": 0.479028581140081, + "grad_norm": 0.4215419590473175, + "learning_rate": 0.0001066210118918189, + "loss": 0.0578, + "step": 20435 + }, + { + "epoch": 0.479052022714886, + "grad_norm": 0.3760961592197418, + "learning_rate": 0.00010661366187957203, + "loss": 0.0665, + "step": 20436 + }, + { + "epoch": 0.479075464289691, + "grad_norm": 0.5400648713111877, + "learning_rate": 0.00010660631183143919, + "loss": 0.1058, + "step": 20437 + }, + { + "epoch": 0.479098905864496, + "grad_norm": 0.49719998240470886, + "learning_rate": 0.00010659896174746034, + "loss": 0.1552, + "step": 20438 + }, + { + "epoch": 0.479122347439301, + "grad_norm": 0.31611281633377075, + "learning_rate": 0.00010659161162767534, + "loss": 0.0387, + "step": 20439 + }, + { + "epoch": 0.479145789014106, + "grad_norm": 0.40755853056907654, + "learning_rate": 0.00010658426147212408, + "loss": 0.0982, + "step": 20440 + }, + { + "epoch": 0.479169230588911, + "grad_norm": 0.10761087387800217, + "learning_rate": 0.00010657691128084644, + "loss": 0.0104, + "step": 20441 + }, + { + "epoch": 0.47919267216371597, + "grad_norm": 0.40089496970176697, + "learning_rate": 0.0001065695610538823, + "loss": 0.0827, + "step": 20442 + }, + { + "epoch": 0.47921611373852097, + "grad_norm": 0.15004146099090576, + "learning_rate": 0.00010656221079127148, + "loss": 0.0277, + "step": 20443 + }, + { + "epoch": 0.47923955531332596, + "grad_norm": 0.5466628074645996, + "learning_rate": 0.00010655486049305396, + "loss": 0.1769, + "step": 20444 + }, + { + "epoch": 0.47926299688813095, + "grad_norm": 0.3373716175556183, + "learning_rate": 0.00010654751015926959, + "loss": 0.0833, + "step": 20445 + }, + { + "epoch": 0.47928643846293595, + "grad_norm": 0.5641312003135681, + "learning_rate": 0.00010654015978995824, + "loss": 0.142, + "step": 20446 + }, + { + "epoch": 0.47930988003774094, + "grad_norm": 0.31160664558410645, + "learning_rate": 0.0001065328093851598, + "loss": 0.0531, + "step": 20447 + }, + { + "epoch": 0.47933332161254594, + "grad_norm": 0.4518442451953888, + "learning_rate": 0.00010652545894491415, + "loss": 0.0556, + "step": 20448 + }, + { + "epoch": 0.47935676318735093, + "grad_norm": 0.4594556391239166, + "learning_rate": 0.00010651810846926117, + "loss": 0.1249, + "step": 20449 + }, + { + "epoch": 0.4793802047621559, + "grad_norm": 0.3505662679672241, + "learning_rate": 0.00010651075795824075, + "loss": 0.067, + "step": 20450 + }, + { + "epoch": 0.4794036463369609, + "grad_norm": 0.6479877233505249, + "learning_rate": 0.00010650340741189278, + "loss": 0.1309, + "step": 20451 + }, + { + "epoch": 0.4794270879117659, + "grad_norm": 0.43194589018821716, + "learning_rate": 0.00010649605683025714, + "loss": 0.1109, + "step": 20452 + }, + { + "epoch": 0.4794505294865709, + "grad_norm": 0.4747781455516815, + "learning_rate": 0.0001064887062133737, + "loss": 0.1678, + "step": 20453 + }, + { + "epoch": 0.4794739710613759, + "grad_norm": 0.4237464368343353, + "learning_rate": 0.00010648135556128236, + "loss": 0.0961, + "step": 20454 + }, + { + "epoch": 0.4794974126361809, + "grad_norm": 0.514293909072876, + "learning_rate": 0.00010647400487402303, + "loss": 0.1167, + "step": 20455 + }, + { + "epoch": 0.4795208542109859, + "grad_norm": 0.11709953099489212, + "learning_rate": 0.00010646665415163552, + "loss": 0.0094, + "step": 20456 + }, + { + "epoch": 0.4795442957857909, + "grad_norm": 0.46436455845832825, + "learning_rate": 0.00010645930339415979, + "loss": 0.0901, + "step": 20457 + }, + { + "epoch": 0.4795677373605959, + "grad_norm": 0.22947312891483307, + "learning_rate": 0.00010645195260163572, + "loss": 0.029, + "step": 20458 + }, + { + "epoch": 0.4795911789354009, + "grad_norm": 0.3137860596179962, + "learning_rate": 0.00010644460177410314, + "loss": 0.3203, + "step": 20459 + }, + { + "epoch": 0.47961462051020587, + "grad_norm": 0.7318207025527954, + "learning_rate": 0.00010643725091160197, + "loss": 0.1486, + "step": 20460 + }, + { + "epoch": 0.47963806208501086, + "grad_norm": 0.5015605688095093, + "learning_rate": 0.00010642990001417212, + "loss": 0.156, + "step": 20461 + }, + { + "epoch": 0.47966150365981586, + "grad_norm": 0.1605854332447052, + "learning_rate": 0.00010642254908185342, + "loss": 0.0461, + "step": 20462 + }, + { + "epoch": 0.47968494523462085, + "grad_norm": 0.6575950980186462, + "learning_rate": 0.0001064151981146858, + "loss": 0.1639, + "step": 20463 + }, + { + "epoch": 0.47970838680942585, + "grad_norm": 0.10541920363903046, + "learning_rate": 0.00010640784711270918, + "loss": 0.0211, + "step": 20464 + }, + { + "epoch": 0.47973182838423084, + "grad_norm": 0.4805333912372589, + "learning_rate": 0.00010640049607596336, + "loss": 0.0786, + "step": 20465 + }, + { + "epoch": 0.47975526995903583, + "grad_norm": 0.3957715630531311, + "learning_rate": 0.00010639314500448824, + "loss": 0.0813, + "step": 20466 + }, + { + "epoch": 0.47977871153384083, + "grad_norm": 0.174647256731987, + "learning_rate": 0.0001063857938983238, + "loss": 0.0372, + "step": 20467 + }, + { + "epoch": 0.4798021531086458, + "grad_norm": 0.48587608337402344, + "learning_rate": 0.00010637844275750983, + "loss": 0.0499, + "step": 20468 + }, + { + "epoch": 0.4798255946834508, + "grad_norm": 0.3416605591773987, + "learning_rate": 0.00010637109158208622, + "loss": 0.4405, + "step": 20469 + }, + { + "epoch": 0.4798490362582558, + "grad_norm": 0.41392767429351807, + "learning_rate": 0.00010636374037209294, + "loss": 0.0932, + "step": 20470 + }, + { + "epoch": 0.4798724778330608, + "grad_norm": 0.5703502297401428, + "learning_rate": 0.0001063563891275698, + "loss": 0.6337, + "step": 20471 + }, + { + "epoch": 0.4798959194078658, + "grad_norm": 0.5692912340164185, + "learning_rate": 0.0001063490378485567, + "loss": 0.1318, + "step": 20472 + }, + { + "epoch": 0.4799193609826708, + "grad_norm": 0.12601807713508606, + "learning_rate": 0.00010634168653509358, + "loss": 0.014, + "step": 20473 + }, + { + "epoch": 0.4799428025574758, + "grad_norm": 0.2598402202129364, + "learning_rate": 0.00010633433518722027, + "loss": 0.0658, + "step": 20474 + }, + { + "epoch": 0.4799662441322808, + "grad_norm": 0.15165981650352478, + "learning_rate": 0.00010632698380497665, + "loss": 0.0257, + "step": 20475 + }, + { + "epoch": 0.4799896857070858, + "grad_norm": 0.4008025825023651, + "learning_rate": 0.00010631963238840268, + "loss": 0.1217, + "step": 20476 + }, + { + "epoch": 0.4800131272818908, + "grad_norm": 0.7351489663124084, + "learning_rate": 0.0001063122809375382, + "loss": 0.0944, + "step": 20477 + }, + { + "epoch": 0.48003656885669577, + "grad_norm": 0.20544834434986115, + "learning_rate": 0.0001063049294524231, + "loss": 0.0292, + "step": 20478 + }, + { + "epoch": 0.48006001043150076, + "grad_norm": 0.22695910930633545, + "learning_rate": 0.00010629757793309727, + "loss": 0.0517, + "step": 20479 + }, + { + "epoch": 0.4800834520063058, + "grad_norm": 0.38197213411331177, + "learning_rate": 0.00010629022637960063, + "loss": 0.0909, + "step": 20480 + }, + { + "epoch": 0.4801068935811108, + "grad_norm": 0.4440311789512634, + "learning_rate": 0.00010628287479197301, + "loss": 0.1233, + "step": 20481 + }, + { + "epoch": 0.4801303351559158, + "grad_norm": 0.34710928797721863, + "learning_rate": 0.00010627552317025433, + "loss": 0.082, + "step": 20482 + }, + { + "epoch": 0.4801537767307208, + "grad_norm": 0.46941545605659485, + "learning_rate": 0.00010626817151448452, + "loss": 0.1015, + "step": 20483 + }, + { + "epoch": 0.4801772183055258, + "grad_norm": 0.09453120082616806, + "learning_rate": 0.00010626081982470343, + "loss": 0.0198, + "step": 20484 + }, + { + "epoch": 0.4802006598803308, + "grad_norm": 0.39818429946899414, + "learning_rate": 0.00010625346810095093, + "loss": 0.0598, + "step": 20485 + }, + { + "epoch": 0.4802241014551358, + "grad_norm": 0.7438320517539978, + "learning_rate": 0.00010624611634326698, + "loss": 0.6274, + "step": 20486 + }, + { + "epoch": 0.48024754302994077, + "grad_norm": 0.192435160279274, + "learning_rate": 0.00010623876455169137, + "loss": 0.0388, + "step": 20487 + }, + { + "epoch": 0.48027098460474577, + "grad_norm": 0.5046127438545227, + "learning_rate": 0.00010623141272626407, + "loss": 0.1478, + "step": 20488 + }, + { + "epoch": 0.48029442617955076, + "grad_norm": 0.5692020654678345, + "learning_rate": 0.00010622406086702497, + "loss": 0.115, + "step": 20489 + }, + { + "epoch": 0.48031786775435575, + "grad_norm": 0.5714141130447388, + "learning_rate": 0.00010621670897401395, + "loss": 0.1593, + "step": 20490 + }, + { + "epoch": 0.48034130932916075, + "grad_norm": 0.2236902266740799, + "learning_rate": 0.00010620935704727084, + "loss": 0.0409, + "step": 20491 + }, + { + "epoch": 0.48036475090396574, + "grad_norm": 0.39729440212249756, + "learning_rate": 0.00010620200508683564, + "loss": 0.0813, + "step": 20492 + }, + { + "epoch": 0.48038819247877074, + "grad_norm": 0.4869646728038788, + "learning_rate": 0.00010619465309274814, + "loss": 0.0746, + "step": 20493 + }, + { + "epoch": 0.48041163405357573, + "grad_norm": 0.3593880534172058, + "learning_rate": 0.00010618730106504829, + "loss": 0.0692, + "step": 20494 + }, + { + "epoch": 0.4804350756283807, + "grad_norm": 0.7849469780921936, + "learning_rate": 0.000106179949003776, + "loss": 0.191, + "step": 20495 + }, + { + "epoch": 0.4804585172031857, + "grad_norm": 0.3999945819377899, + "learning_rate": 0.0001061725969089711, + "loss": 0.043, + "step": 20496 + }, + { + "epoch": 0.4804819587779907, + "grad_norm": 0.5882455706596375, + "learning_rate": 0.00010616524478067351, + "loss": 0.1259, + "step": 20497 + }, + { + "epoch": 0.4805054003527957, + "grad_norm": 0.1929321438074112, + "learning_rate": 0.00010615789261892317, + "loss": 0.0412, + "step": 20498 + }, + { + "epoch": 0.4805288419276007, + "grad_norm": 0.5124331116676331, + "learning_rate": 0.00010615054042375989, + "loss": 0.1462, + "step": 20499 + }, + { + "epoch": 0.4805522835024057, + "grad_norm": 0.2950015366077423, + "learning_rate": 0.00010614318819522364, + "loss": 0.0479, + "step": 20500 + }, + { + "epoch": 0.4805757250772107, + "grad_norm": 0.708165168762207, + "learning_rate": 0.00010613583593335423, + "loss": 0.1472, + "step": 20501 + }, + { + "epoch": 0.4805991666520157, + "grad_norm": 0.5027030110359192, + "learning_rate": 0.00010612848363819165, + "loss": 0.1591, + "step": 20502 + }, + { + "epoch": 0.4806226082268207, + "grad_norm": 0.4745770990848541, + "learning_rate": 0.00010612113130977575, + "loss": 0.0721, + "step": 20503 + }, + { + "epoch": 0.4806460498016257, + "grad_norm": 0.2877620458602905, + "learning_rate": 0.00010611377894814639, + "loss": 0.0675, + "step": 20504 + }, + { + "epoch": 0.48066949137643067, + "grad_norm": 0.5524049401283264, + "learning_rate": 0.00010610642655334346, + "loss": 0.1087, + "step": 20505 + }, + { + "epoch": 0.48069293295123566, + "grad_norm": 0.48984795808792114, + "learning_rate": 0.00010609907412540696, + "loss": 0.1261, + "step": 20506 + }, + { + "epoch": 0.48071637452604066, + "grad_norm": 0.5086389780044556, + "learning_rate": 0.00010609172166437667, + "loss": 0.1055, + "step": 20507 + }, + { + "epoch": 0.48073981610084565, + "grad_norm": 0.4340720474720001, + "learning_rate": 0.00010608436917029254, + "loss": 0.1222, + "step": 20508 + }, + { + "epoch": 0.48076325767565065, + "grad_norm": 0.5339847207069397, + "learning_rate": 0.00010607701664319445, + "loss": 0.1424, + "step": 20509 + }, + { + "epoch": 0.48078669925045564, + "grad_norm": 1.0164892673492432, + "learning_rate": 0.00010606966408312229, + "loss": 0.1464, + "step": 20510 + }, + { + "epoch": 0.48081014082526063, + "grad_norm": 0.44439372420310974, + "learning_rate": 0.00010606231149011596, + "loss": 0.0702, + "step": 20511 + }, + { + "epoch": 0.48083358240006563, + "grad_norm": 0.46115559339523315, + "learning_rate": 0.00010605495886421538, + "loss": 0.075, + "step": 20512 + }, + { + "epoch": 0.4808570239748706, + "grad_norm": 0.4805966019630432, + "learning_rate": 0.00010604760620546039, + "loss": 0.0502, + "step": 20513 + }, + { + "epoch": 0.4808804655496756, + "grad_norm": 0.6110399961471558, + "learning_rate": 0.00010604025351389096, + "loss": 0.1042, + "step": 20514 + }, + { + "epoch": 0.4809039071244806, + "grad_norm": 0.4559907615184784, + "learning_rate": 0.00010603290078954691, + "loss": 0.1148, + "step": 20515 + }, + { + "epoch": 0.4809273486992856, + "grad_norm": 0.6758543252944946, + "learning_rate": 0.00010602554803246818, + "loss": 0.1656, + "step": 20516 + }, + { + "epoch": 0.4809507902740906, + "grad_norm": 0.43389081954956055, + "learning_rate": 0.00010601819524269465, + "loss": 0.347, + "step": 20517 + }, + { + "epoch": 0.4809742318488956, + "grad_norm": 0.21367420256137848, + "learning_rate": 0.00010601084242026625, + "loss": 0.0437, + "step": 20518 + }, + { + "epoch": 0.4809976734237006, + "grad_norm": 0.54384845495224, + "learning_rate": 0.00010600348956522284, + "loss": 0.4768, + "step": 20519 + }, + { + "epoch": 0.4810211149985056, + "grad_norm": 0.3192734122276306, + "learning_rate": 0.00010599613667760429, + "loss": 0.0946, + "step": 20520 + }, + { + "epoch": 0.4810445565733106, + "grad_norm": 0.17921306192874908, + "learning_rate": 0.00010598878375745058, + "loss": 0.0228, + "step": 20521 + }, + { + "epoch": 0.48106799814811557, + "grad_norm": 0.402714341878891, + "learning_rate": 0.00010598143080480157, + "loss": 0.0613, + "step": 20522 + }, + { + "epoch": 0.48109143972292057, + "grad_norm": 0.5701057314872742, + "learning_rate": 0.00010597407781969706, + "loss": 0.0653, + "step": 20523 + }, + { + "epoch": 0.48111488129772556, + "grad_norm": 0.31713154911994934, + "learning_rate": 0.0001059667248021771, + "loss": 0.0391, + "step": 20524 + }, + { + "epoch": 0.48113832287253055, + "grad_norm": 0.2590137720108032, + "learning_rate": 0.00010595937175228154, + "loss": 0.0731, + "step": 20525 + }, + { + "epoch": 0.48116176444733555, + "grad_norm": 0.21326683461666107, + "learning_rate": 0.00010595201867005025, + "loss": 0.0347, + "step": 20526 + }, + { + "epoch": 0.48118520602214054, + "grad_norm": 0.32572227716445923, + "learning_rate": 0.00010594466555552314, + "loss": 0.0583, + "step": 20527 + }, + { + "epoch": 0.48120864759694554, + "grad_norm": 0.4776190221309662, + "learning_rate": 0.0001059373124087401, + "loss": 0.0508, + "step": 20528 + }, + { + "epoch": 0.48123208917175053, + "grad_norm": 0.5684109926223755, + "learning_rate": 0.00010592995922974103, + "loss": 0.0843, + "step": 20529 + }, + { + "epoch": 0.4812555307465555, + "grad_norm": 0.3179238736629486, + "learning_rate": 0.00010592260601856583, + "loss": 0.0569, + "step": 20530 + }, + { + "epoch": 0.4812789723213606, + "grad_norm": 0.2564438581466675, + "learning_rate": 0.00010591525277525443, + "loss": 0.041, + "step": 20531 + }, + { + "epoch": 0.48130241389616557, + "grad_norm": 0.35246962308883667, + "learning_rate": 0.00010590789949984665, + "loss": 0.1131, + "step": 20532 + }, + { + "epoch": 0.48132585547097056, + "grad_norm": 0.3648236393928528, + "learning_rate": 0.0001059005461923825, + "loss": 0.0645, + "step": 20533 + }, + { + "epoch": 0.48134929704577556, + "grad_norm": 0.8870677947998047, + "learning_rate": 0.0001058931928529018, + "loss": 0.0967, + "step": 20534 + }, + { + "epoch": 0.48137273862058055, + "grad_norm": 0.5509538054466248, + "learning_rate": 0.00010588583948144446, + "loss": 0.0668, + "step": 20535 + }, + { + "epoch": 0.48139618019538555, + "grad_norm": 0.20535705983638763, + "learning_rate": 0.00010587848607805037, + "loss": 0.0522, + "step": 20536 + }, + { + "epoch": 0.48141962177019054, + "grad_norm": 0.24452249705791473, + "learning_rate": 0.00010587113264275949, + "loss": 0.1633, + "step": 20537 + }, + { + "epoch": 0.48144306334499554, + "grad_norm": 0.16706573963165283, + "learning_rate": 0.00010586377917561167, + "loss": 0.0291, + "step": 20538 + }, + { + "epoch": 0.48146650491980053, + "grad_norm": 0.47638338804244995, + "learning_rate": 0.00010585642567664682, + "loss": 0.074, + "step": 20539 + }, + { + "epoch": 0.4814899464946055, + "grad_norm": 0.8387273550033569, + "learning_rate": 0.00010584907214590485, + "loss": 0.2723, + "step": 20540 + }, + { + "epoch": 0.4815133880694105, + "grad_norm": 0.8714443445205688, + "learning_rate": 0.00010584171858342564, + "loss": 0.1821, + "step": 20541 + }, + { + "epoch": 0.4815368296442155, + "grad_norm": 0.10304686427116394, + "learning_rate": 0.00010583436498924906, + "loss": 0.0143, + "step": 20542 + }, + { + "epoch": 0.4815602712190205, + "grad_norm": 0.665654182434082, + "learning_rate": 0.00010582701136341508, + "loss": 0.1462, + "step": 20543 + }, + { + "epoch": 0.4815837127938255, + "grad_norm": 0.3232108950614929, + "learning_rate": 0.00010581965770596361, + "loss": 0.0787, + "step": 20544 + }, + { + "epoch": 0.4816071543686305, + "grad_norm": 0.2869909703731537, + "learning_rate": 0.00010581230401693447, + "loss": 0.0279, + "step": 20545 + }, + { + "epoch": 0.4816305959434355, + "grad_norm": 0.6071177124977112, + "learning_rate": 0.00010580495029636762, + "loss": 0.1151, + "step": 20546 + }, + { + "epoch": 0.4816540375182405, + "grad_norm": 0.5687691569328308, + "learning_rate": 0.00010579759654430295, + "loss": 0.1423, + "step": 20547 + }, + { + "epoch": 0.4816774790930455, + "grad_norm": 0.6235194206237793, + "learning_rate": 0.00010579024276078034, + "loss": 0.2026, + "step": 20548 + }, + { + "epoch": 0.4817009206678505, + "grad_norm": 0.21346066892147064, + "learning_rate": 0.00010578288894583972, + "loss": 0.0377, + "step": 20549 + }, + { + "epoch": 0.48172436224265547, + "grad_norm": 0.7373991012573242, + "learning_rate": 0.00010577553509952098, + "loss": 0.1423, + "step": 20550 + }, + { + "epoch": 0.48174780381746046, + "grad_norm": 0.21115374565124512, + "learning_rate": 0.00010576818122186402, + "loss": 0.0256, + "step": 20551 + }, + { + "epoch": 0.48177124539226546, + "grad_norm": 0.6302988529205322, + "learning_rate": 0.00010576082731290878, + "loss": 0.16, + "step": 20552 + }, + { + "epoch": 0.48179468696707045, + "grad_norm": 0.7380772233009338, + "learning_rate": 0.0001057534733726951, + "loss": 0.1227, + "step": 20553 + }, + { + "epoch": 0.48181812854187545, + "grad_norm": 0.5435645580291748, + "learning_rate": 0.00010574611940126291, + "loss": 0.0928, + "step": 20554 + }, + { + "epoch": 0.48184157011668044, + "grad_norm": 0.4637534022331238, + "learning_rate": 0.0001057387653986521, + "loss": 0.0743, + "step": 20555 + }, + { + "epoch": 0.48186501169148543, + "grad_norm": 0.3388792872428894, + "learning_rate": 0.00010573141136490261, + "loss": 0.0681, + "step": 20556 + }, + { + "epoch": 0.48188845326629043, + "grad_norm": 0.4707542359828949, + "learning_rate": 0.00010572405730005434, + "loss": 0.1172, + "step": 20557 + }, + { + "epoch": 0.4819118948410954, + "grad_norm": 0.24348153173923492, + "learning_rate": 0.00010571670320414712, + "loss": 0.0457, + "step": 20558 + }, + { + "epoch": 0.4819353364159004, + "grad_norm": 0.22194203734397888, + "learning_rate": 0.00010570934907722096, + "loss": 0.0543, + "step": 20559 + }, + { + "epoch": 0.4819587779907054, + "grad_norm": 0.6235191822052002, + "learning_rate": 0.0001057019949193157, + "loss": 0.4887, + "step": 20560 + }, + { + "epoch": 0.4819822195655104, + "grad_norm": 0.5911219120025635, + "learning_rate": 0.00010569464073047121, + "loss": 0.1024, + "step": 20561 + }, + { + "epoch": 0.4820056611403154, + "grad_norm": 0.6903539299964905, + "learning_rate": 0.0001056872865107275, + "loss": 0.6087, + "step": 20562 + }, + { + "epoch": 0.4820291027151204, + "grad_norm": 0.2013607770204544, + "learning_rate": 0.0001056799322601244, + "loss": 0.0315, + "step": 20563 + }, + { + "epoch": 0.4820525442899254, + "grad_norm": 0.4296470582485199, + "learning_rate": 0.0001056725779787018, + "loss": 0.0593, + "step": 20564 + }, + { + "epoch": 0.4820759858647304, + "grad_norm": 0.451619029045105, + "learning_rate": 0.00010566522366649967, + "loss": 0.0962, + "step": 20565 + }, + { + "epoch": 0.4820994274395354, + "grad_norm": 0.39428189396858215, + "learning_rate": 0.00010565786932355784, + "loss": 0.0559, + "step": 20566 + }, + { + "epoch": 0.48212286901434037, + "grad_norm": 0.4920404553413391, + "learning_rate": 0.00010565051494991627, + "loss": 0.0729, + "step": 20567 + }, + { + "epoch": 0.48214631058914537, + "grad_norm": 0.20644177496433258, + "learning_rate": 0.00010564316054561483, + "loss": 0.0481, + "step": 20568 + }, + { + "epoch": 0.48216975216395036, + "grad_norm": 0.17860457301139832, + "learning_rate": 0.00010563580611069348, + "loss": 0.0357, + "step": 20569 + }, + { + "epoch": 0.48219319373875535, + "grad_norm": 0.5711818337440491, + "learning_rate": 0.00010562845164519206, + "loss": 0.1379, + "step": 20570 + }, + { + "epoch": 0.48221663531356035, + "grad_norm": 0.6015787720680237, + "learning_rate": 0.00010562109714915051, + "loss": 0.6887, + "step": 20571 + }, + { + "epoch": 0.48224007688836534, + "grad_norm": 0.31670770049095154, + "learning_rate": 0.00010561374262260872, + "loss": 0.0492, + "step": 20572 + }, + { + "epoch": 0.48226351846317034, + "grad_norm": 0.6371698379516602, + "learning_rate": 0.0001056063880656066, + "loss": 0.5533, + "step": 20573 + }, + { + "epoch": 0.48228696003797533, + "grad_norm": 0.11016331613063812, + "learning_rate": 0.00010559903347818407, + "loss": 0.0219, + "step": 20574 + }, + { + "epoch": 0.4823104016127803, + "grad_norm": 0.04713801294565201, + "learning_rate": 0.00010559167886038104, + "loss": 0.0072, + "step": 20575 + }, + { + "epoch": 0.4823338431875853, + "grad_norm": 0.416365385055542, + "learning_rate": 0.0001055843242122374, + "loss": 0.0632, + "step": 20576 + }, + { + "epoch": 0.4823572847623903, + "grad_norm": 0.4937526285648346, + "learning_rate": 0.00010557696953379306, + "loss": 0.1102, + "step": 20577 + }, + { + "epoch": 0.4823807263371953, + "grad_norm": 0.3246285319328308, + "learning_rate": 0.00010556961482508789, + "loss": 0.0946, + "step": 20578 + }, + { + "epoch": 0.4824041679120003, + "grad_norm": 0.1942477524280548, + "learning_rate": 0.00010556226008616188, + "loss": 0.0203, + "step": 20579 + }, + { + "epoch": 0.4824276094868053, + "grad_norm": 0.7345314025878906, + "learning_rate": 0.00010555490531705486, + "loss": 0.147, + "step": 20580 + }, + { + "epoch": 0.4824510510616103, + "grad_norm": 0.46005910634994507, + "learning_rate": 0.00010554755051780678, + "loss": 0.1013, + "step": 20581 + }, + { + "epoch": 0.48247449263641534, + "grad_norm": 0.456654816865921, + "learning_rate": 0.00010554019568845757, + "loss": 0.0812, + "step": 20582 + }, + { + "epoch": 0.48249793421122034, + "grad_norm": 0.42060789465904236, + "learning_rate": 0.00010553284082904705, + "loss": 0.0883, + "step": 20583 + }, + { + "epoch": 0.48252137578602533, + "grad_norm": 0.2049238234758377, + "learning_rate": 0.00010552548593961517, + "loss": 0.0217, + "step": 20584 + }, + { + "epoch": 0.4825448173608303, + "grad_norm": 0.4922522008419037, + "learning_rate": 0.00010551813102020189, + "loss": 0.0854, + "step": 20585 + }, + { + "epoch": 0.4825682589356353, + "grad_norm": 0.34519684314727783, + "learning_rate": 0.00010551077607084707, + "loss": 0.0884, + "step": 20586 + }, + { + "epoch": 0.4825917005104403, + "grad_norm": 0.38651642203330994, + "learning_rate": 0.00010550342109159062, + "loss": 0.0385, + "step": 20587 + }, + { + "epoch": 0.4826151420852453, + "grad_norm": 0.702690064907074, + "learning_rate": 0.00010549606608247244, + "loss": 0.1238, + "step": 20588 + }, + { + "epoch": 0.4826385836600503, + "grad_norm": 0.4694875180721283, + "learning_rate": 0.00010548871104353246, + "loss": 0.0944, + "step": 20589 + }, + { + "epoch": 0.4826620252348553, + "grad_norm": 0.5567561388015747, + "learning_rate": 0.00010548135597481057, + "loss": 0.1041, + "step": 20590 + }, + { + "epoch": 0.4826854668096603, + "grad_norm": 0.32799533009529114, + "learning_rate": 0.00010547400087634668, + "loss": 0.0798, + "step": 20591 + }, + { + "epoch": 0.4827089083844653, + "grad_norm": 0.6837202906608582, + "learning_rate": 0.00010546664574818075, + "loss": 0.2011, + "step": 20592 + }, + { + "epoch": 0.4827323499592703, + "grad_norm": 0.5192931890487671, + "learning_rate": 0.00010545929059035262, + "loss": 0.0655, + "step": 20593 + }, + { + "epoch": 0.4827557915340753, + "grad_norm": 0.7939948439598083, + "learning_rate": 0.00010545193540290223, + "loss": 0.1197, + "step": 20594 + }, + { + "epoch": 0.48277923310888027, + "grad_norm": 0.5111783742904663, + "learning_rate": 0.0001054445801858695, + "loss": 0.6162, + "step": 20595 + }, + { + "epoch": 0.48280267468368526, + "grad_norm": 0.828451931476593, + "learning_rate": 0.00010543722493929427, + "loss": 0.2496, + "step": 20596 + }, + { + "epoch": 0.48282611625849026, + "grad_norm": 0.30512452125549316, + "learning_rate": 0.00010542986966321655, + "loss": 0.3325, + "step": 20597 + }, + { + "epoch": 0.48284955783329525, + "grad_norm": 0.42116791009902954, + "learning_rate": 0.00010542251435767621, + "loss": 0.0572, + "step": 20598 + }, + { + "epoch": 0.48287299940810025, + "grad_norm": 0.640053927898407, + "learning_rate": 0.00010541515902271312, + "loss": 0.1551, + "step": 20599 + }, + { + "epoch": 0.48289644098290524, + "grad_norm": 0.5766233801841736, + "learning_rate": 0.00010540780365836726, + "loss": 0.1572, + "step": 20600 + }, + { + "epoch": 0.48291988255771023, + "grad_norm": 0.5452268123626709, + "learning_rate": 0.0001054004482646785, + "loss": 0.0994, + "step": 20601 + }, + { + "epoch": 0.48294332413251523, + "grad_norm": 0.48483678698539734, + "learning_rate": 0.00010539309284168673, + "loss": 0.1126, + "step": 20602 + }, + { + "epoch": 0.4829667657073202, + "grad_norm": 0.16022254526615143, + "learning_rate": 0.0001053857373894319, + "loss": 0.0205, + "step": 20603 + }, + { + "epoch": 0.4829902072821252, + "grad_norm": 0.38491782546043396, + "learning_rate": 0.0001053783819079539, + "loss": 0.128, + "step": 20604 + }, + { + "epoch": 0.4830136488569302, + "grad_norm": 0.44637972116470337, + "learning_rate": 0.00010537102639729268, + "loss": 0.7912, + "step": 20605 + }, + { + "epoch": 0.4830370904317352, + "grad_norm": 0.35593658685684204, + "learning_rate": 0.00010536367085748808, + "loss": 0.0709, + "step": 20606 + }, + { + "epoch": 0.4830605320065402, + "grad_norm": 0.17121605575084686, + "learning_rate": 0.0001053563152885801, + "loss": 0.0574, + "step": 20607 + }, + { + "epoch": 0.4830839735813452, + "grad_norm": 0.582280695438385, + "learning_rate": 0.00010534895969060854, + "loss": 0.1052, + "step": 20608 + }, + { + "epoch": 0.4831074151561502, + "grad_norm": 0.3117411732673645, + "learning_rate": 0.0001053416040636134, + "loss": 0.0734, + "step": 20609 + }, + { + "epoch": 0.4831308567309552, + "grad_norm": 0.31358492374420166, + "learning_rate": 0.0001053342484076346, + "loss": 0.1284, + "step": 20610 + }, + { + "epoch": 0.4831542983057602, + "grad_norm": 0.3570827543735504, + "learning_rate": 0.00010532689272271198, + "loss": 0.0496, + "step": 20611 + }, + { + "epoch": 0.48317773988056517, + "grad_norm": 0.4875655472278595, + "learning_rate": 0.0001053195370088855, + "loss": 0.0922, + "step": 20612 + }, + { + "epoch": 0.48320118145537017, + "grad_norm": 0.3833281993865967, + "learning_rate": 0.00010531218126619506, + "loss": 0.0653, + "step": 20613 + }, + { + "epoch": 0.48322462303017516, + "grad_norm": 0.504862368106842, + "learning_rate": 0.00010530482549468062, + "loss": 0.0956, + "step": 20614 + }, + { + "epoch": 0.48324806460498015, + "grad_norm": 0.5791900753974915, + "learning_rate": 0.00010529746969438197, + "loss": 0.136, + "step": 20615 + }, + { + "epoch": 0.48327150617978515, + "grad_norm": 0.1549966186285019, + "learning_rate": 0.00010529011386533913, + "loss": 0.0251, + "step": 20616 + }, + { + "epoch": 0.48329494775459014, + "grad_norm": 0.5885168313980103, + "learning_rate": 0.000105282758007592, + "loss": 0.6703, + "step": 20617 + }, + { + "epoch": 0.48331838932939514, + "grad_norm": 0.4338751435279846, + "learning_rate": 0.00010527540212118045, + "loss": 0.0854, + "step": 20618 + }, + { + "epoch": 0.48334183090420013, + "grad_norm": 0.23039941489696503, + "learning_rate": 0.00010526804620614446, + "loss": 0.0611, + "step": 20619 + }, + { + "epoch": 0.4833652724790051, + "grad_norm": 0.6389977335929871, + "learning_rate": 0.0001052606902625239, + "loss": 0.0985, + "step": 20620 + }, + { + "epoch": 0.4833887140538101, + "grad_norm": 0.4808570444583893, + "learning_rate": 0.00010525333429035864, + "loss": 0.0749, + "step": 20621 + }, + { + "epoch": 0.4834121556286151, + "grad_norm": 0.1258140653371811, + "learning_rate": 0.00010524597828968865, + "loss": 0.0338, + "step": 20622 + }, + { + "epoch": 0.4834355972034201, + "grad_norm": 0.5102112293243408, + "learning_rate": 0.00010523862226055387, + "loss": 0.1344, + "step": 20623 + }, + { + "epoch": 0.4834590387782251, + "grad_norm": 0.6912498474121094, + "learning_rate": 0.00010523126620299417, + "loss": 0.164, + "step": 20624 + }, + { + "epoch": 0.4834824803530301, + "grad_norm": 0.10481465607881546, + "learning_rate": 0.00010522391011704946, + "loss": 0.0245, + "step": 20625 + }, + { + "epoch": 0.4835059219278351, + "grad_norm": 0.5391634702682495, + "learning_rate": 0.00010521655400275967, + "loss": 0.1439, + "step": 20626 + }, + { + "epoch": 0.4835293635026401, + "grad_norm": 0.07136697322130203, + "learning_rate": 0.0001052091978601647, + "loss": 0.0127, + "step": 20627 + }, + { + "epoch": 0.4835528050774451, + "grad_norm": 0.21188656985759735, + "learning_rate": 0.00010520184168930447, + "loss": 0.0466, + "step": 20628 + }, + { + "epoch": 0.4835762466522501, + "grad_norm": 0.4228361248970032, + "learning_rate": 0.00010519448549021893, + "loss": 0.0929, + "step": 20629 + }, + { + "epoch": 0.48359968822705507, + "grad_norm": 0.7429179549217224, + "learning_rate": 0.00010518712926294796, + "loss": 0.1476, + "step": 20630 + }, + { + "epoch": 0.48362312980186006, + "grad_norm": 0.5578019022941589, + "learning_rate": 0.00010517977300753146, + "loss": 0.5779, + "step": 20631 + }, + { + "epoch": 0.48364657137666506, + "grad_norm": 0.4155348241329193, + "learning_rate": 0.00010517241672400939, + "loss": 0.0844, + "step": 20632 + }, + { + "epoch": 0.4836700129514701, + "grad_norm": 0.34692972898483276, + "learning_rate": 0.00010516506041242162, + "loss": 0.0811, + "step": 20633 + }, + { + "epoch": 0.4836934545262751, + "grad_norm": 0.363433301448822, + "learning_rate": 0.0001051577040728081, + "loss": 0.4135, + "step": 20634 + }, + { + "epoch": 0.4837168961010801, + "grad_norm": 0.1933552771806717, + "learning_rate": 0.00010515034770520874, + "loss": 0.0356, + "step": 20635 + }, + { + "epoch": 0.4837403376758851, + "grad_norm": 0.3912930190563202, + "learning_rate": 0.00010514299130966346, + "loss": 0.0717, + "step": 20636 + }, + { + "epoch": 0.4837637792506901, + "grad_norm": 0.5415555834770203, + "learning_rate": 0.00010513563488621211, + "loss": 0.1368, + "step": 20637 + }, + { + "epoch": 0.4837872208254951, + "grad_norm": 0.5573801398277283, + "learning_rate": 0.0001051282784348947, + "loss": 0.1305, + "step": 20638 + }, + { + "epoch": 0.4838106624003001, + "grad_norm": 0.5105118751525879, + "learning_rate": 0.00010512092195575109, + "loss": 0.1027, + "step": 20639 + }, + { + "epoch": 0.48383410397510507, + "grad_norm": 0.3621477782726288, + "learning_rate": 0.00010511356544882125, + "loss": 0.0723, + "step": 20640 + }, + { + "epoch": 0.48385754554991006, + "grad_norm": 0.5093762278556824, + "learning_rate": 0.00010510620891414502, + "loss": 0.0605, + "step": 20641 + }, + { + "epoch": 0.48388098712471506, + "grad_norm": 0.43147775530815125, + "learning_rate": 0.00010509885235176237, + "loss": 0.135, + "step": 20642 + }, + { + "epoch": 0.48390442869952005, + "grad_norm": 0.39787009358406067, + "learning_rate": 0.00010509149576171322, + "loss": 0.0796, + "step": 20643 + }, + { + "epoch": 0.48392787027432504, + "grad_norm": 0.2697898745536804, + "learning_rate": 0.00010508413914403745, + "loss": 0.0594, + "step": 20644 + }, + { + "epoch": 0.48395131184913004, + "grad_norm": 0.1402331292629242, + "learning_rate": 0.000105076782498775, + "loss": 0.0181, + "step": 20645 + }, + { + "epoch": 0.48397475342393503, + "grad_norm": 0.18082375824451447, + "learning_rate": 0.00010506942582596579, + "loss": 0.0243, + "step": 20646 + }, + { + "epoch": 0.48399819499874003, + "grad_norm": 0.3721274137496948, + "learning_rate": 0.00010506206912564973, + "loss": 0.0701, + "step": 20647 + }, + { + "epoch": 0.484021636573545, + "grad_norm": 0.3875863254070282, + "learning_rate": 0.00010505471239786675, + "loss": 0.0929, + "step": 20648 + }, + { + "epoch": 0.48404507814835, + "grad_norm": 0.44526004791259766, + "learning_rate": 0.00010504735564265675, + "loss": 0.097, + "step": 20649 + }, + { + "epoch": 0.484068519723155, + "grad_norm": 0.4441003203392029, + "learning_rate": 0.00010503999886005967, + "loss": 0.1195, + "step": 20650 + }, + { + "epoch": 0.48409196129796, + "grad_norm": 0.5362022519111633, + "learning_rate": 0.00010503264205011538, + "loss": 0.0928, + "step": 20651 + }, + { + "epoch": 0.484115402872765, + "grad_norm": 0.41721123456954956, + "learning_rate": 0.00010502528521286386, + "loss": 0.1141, + "step": 20652 + }, + { + "epoch": 0.48413884444757, + "grad_norm": 0.1475883275270462, + "learning_rate": 0.00010501792834834499, + "loss": 0.015, + "step": 20653 + }, + { + "epoch": 0.484162286022375, + "grad_norm": 0.48336780071258545, + "learning_rate": 0.00010501057145659872, + "loss": 0.0801, + "step": 20654 + }, + { + "epoch": 0.48418572759718, + "grad_norm": 0.4005151689052582, + "learning_rate": 0.00010500321453766495, + "loss": 0.0601, + "step": 20655 + }, + { + "epoch": 0.484209169171985, + "grad_norm": 0.35375741124153137, + "learning_rate": 0.00010499585759158357, + "loss": 0.0649, + "step": 20656 + }, + { + "epoch": 0.48423261074678997, + "grad_norm": 0.39191338419914246, + "learning_rate": 0.00010498850061839451, + "loss": 0.043, + "step": 20657 + }, + { + "epoch": 0.48425605232159497, + "grad_norm": 0.5775128602981567, + "learning_rate": 0.00010498114361813776, + "loss": 0.1033, + "step": 20658 + }, + { + "epoch": 0.48427949389639996, + "grad_norm": 0.10395722091197968, + "learning_rate": 0.00010497378659085315, + "loss": 0.0144, + "step": 20659 + }, + { + "epoch": 0.48430293547120495, + "grad_norm": 0.41231754422187805, + "learning_rate": 0.00010496642953658065, + "loss": 0.0521, + "step": 20660 + }, + { + "epoch": 0.48432637704600995, + "grad_norm": 0.2189989537000656, + "learning_rate": 0.00010495907245536016, + "loss": 0.0387, + "step": 20661 + }, + { + "epoch": 0.48434981862081494, + "grad_norm": 0.4409846067428589, + "learning_rate": 0.00010495171534723158, + "loss": 0.0958, + "step": 20662 + }, + { + "epoch": 0.48437326019561994, + "grad_norm": 0.23584097623825073, + "learning_rate": 0.00010494435821223488, + "loss": 0.0444, + "step": 20663 + }, + { + "epoch": 0.48439670177042493, + "grad_norm": 0.2971862852573395, + "learning_rate": 0.00010493700105040994, + "loss": 0.0598, + "step": 20664 + }, + { + "epoch": 0.4844201433452299, + "grad_norm": 0.29215380549430847, + "learning_rate": 0.00010492964386179671, + "loss": 0.0731, + "step": 20665 + }, + { + "epoch": 0.4844435849200349, + "grad_norm": 0.4274473786354065, + "learning_rate": 0.00010492228664643506, + "loss": 0.0615, + "step": 20666 + }, + { + "epoch": 0.4844670264948399, + "grad_norm": 0.35790231823921204, + "learning_rate": 0.00010491492940436498, + "loss": 0.0531, + "step": 20667 + }, + { + "epoch": 0.4844904680696449, + "grad_norm": 0.7951161861419678, + "learning_rate": 0.00010490757213562637, + "loss": 0.0934, + "step": 20668 + }, + { + "epoch": 0.4845139096444499, + "grad_norm": 0.2194969356060028, + "learning_rate": 0.00010490021484025907, + "loss": 0.0261, + "step": 20669 + }, + { + "epoch": 0.4845373512192549, + "grad_norm": 0.21141259372234344, + "learning_rate": 0.00010489285751830311, + "loss": 0.021, + "step": 20670 + }, + { + "epoch": 0.4845607927940599, + "grad_norm": 0.5737904906272888, + "learning_rate": 0.00010488550016979837, + "loss": 0.0927, + "step": 20671 + }, + { + "epoch": 0.4845842343688649, + "grad_norm": 0.29359012842178345, + "learning_rate": 0.00010487814279478477, + "loss": 0.0694, + "step": 20672 + }, + { + "epoch": 0.4846076759436699, + "grad_norm": 0.3855222165584564, + "learning_rate": 0.00010487078539330222, + "loss": 0.1007, + "step": 20673 + }, + { + "epoch": 0.4846311175184749, + "grad_norm": 0.35709869861602783, + "learning_rate": 0.00010486342796539065, + "loss": 0.0752, + "step": 20674 + }, + { + "epoch": 0.48465455909327987, + "grad_norm": 0.3911980986595154, + "learning_rate": 0.00010485607051108998, + "loss": 0.1025, + "step": 20675 + }, + { + "epoch": 0.48467800066808486, + "grad_norm": 0.44292521476745605, + "learning_rate": 0.00010484871303044012, + "loss": 0.7943, + "step": 20676 + }, + { + "epoch": 0.48470144224288986, + "grad_norm": 0.5228833556175232, + "learning_rate": 0.00010484135552348104, + "loss": 0.163, + "step": 20677 + }, + { + "epoch": 0.48472488381769485, + "grad_norm": 0.11809944361448288, + "learning_rate": 0.00010483399799025264, + "loss": 0.0229, + "step": 20678 + }, + { + "epoch": 0.48474832539249985, + "grad_norm": 0.3674418330192566, + "learning_rate": 0.0001048266404307948, + "loss": 0.0905, + "step": 20679 + }, + { + "epoch": 0.48477176696730484, + "grad_norm": 0.5114224553108215, + "learning_rate": 0.00010481928284514748, + "loss": 0.1643, + "step": 20680 + }, + { + "epoch": 0.48479520854210983, + "grad_norm": 0.3423417806625366, + "learning_rate": 0.0001048119252333506, + "loss": 0.0337, + "step": 20681 + }, + { + "epoch": 0.48481865011691483, + "grad_norm": 0.5154797434806824, + "learning_rate": 0.00010480456759544406, + "loss": 0.1089, + "step": 20682 + }, + { + "epoch": 0.4848420916917198, + "grad_norm": 0.3988187313079834, + "learning_rate": 0.00010479720993146782, + "loss": 0.0569, + "step": 20683 + }, + { + "epoch": 0.4848655332665248, + "grad_norm": 0.46732887625694275, + "learning_rate": 0.00010478985224146178, + "loss": 0.0582, + "step": 20684 + }, + { + "epoch": 0.48488897484132987, + "grad_norm": 0.21971012651920319, + "learning_rate": 0.00010478249452546587, + "loss": 0.0484, + "step": 20685 + }, + { + "epoch": 0.48491241641613486, + "grad_norm": 0.3772428035736084, + "learning_rate": 0.00010477513678352, + "loss": 0.358, + "step": 20686 + }, + { + "epoch": 0.48493585799093986, + "grad_norm": 0.27960461378097534, + "learning_rate": 0.00010476777901566412, + "loss": 0.0295, + "step": 20687 + }, + { + "epoch": 0.48495929956574485, + "grad_norm": 0.7351235151290894, + "learning_rate": 0.0001047604212219381, + "loss": 0.1534, + "step": 20688 + }, + { + "epoch": 0.48498274114054984, + "grad_norm": 0.70087069272995, + "learning_rate": 0.00010475306340238193, + "loss": 0.1242, + "step": 20689 + }, + { + "epoch": 0.48500618271535484, + "grad_norm": 0.5115194916725159, + "learning_rate": 0.00010474570555703551, + "loss": 0.1, + "step": 20690 + }, + { + "epoch": 0.48502962429015983, + "grad_norm": 0.3889065384864807, + "learning_rate": 0.00010473834768593875, + "loss": 0.0883, + "step": 20691 + }, + { + "epoch": 0.4850530658649648, + "grad_norm": 0.5044077038764954, + "learning_rate": 0.00010473098978913158, + "loss": 0.1133, + "step": 20692 + }, + { + "epoch": 0.4850765074397698, + "grad_norm": 0.4542126953601837, + "learning_rate": 0.00010472363186665395, + "loss": 0.1143, + "step": 20693 + }, + { + "epoch": 0.4850999490145748, + "grad_norm": 0.204049214720726, + "learning_rate": 0.00010471627391854573, + "loss": 0.0324, + "step": 20694 + }, + { + "epoch": 0.4851233905893798, + "grad_norm": 0.34377434849739075, + "learning_rate": 0.00010470891594484687, + "loss": 0.0625, + "step": 20695 + }, + { + "epoch": 0.4851468321641848, + "grad_norm": 0.6040670275688171, + "learning_rate": 0.00010470155794559733, + "loss": 0.0686, + "step": 20696 + }, + { + "epoch": 0.4851702737389898, + "grad_norm": 0.4388623833656311, + "learning_rate": 0.000104694199920837, + "loss": 0.047, + "step": 20697 + }, + { + "epoch": 0.4851937153137948, + "grad_norm": 0.38810694217681885, + "learning_rate": 0.00010468684187060577, + "loss": 0.0738, + "step": 20698 + }, + { + "epoch": 0.4852171568885998, + "grad_norm": 0.6913837790489197, + "learning_rate": 0.00010467948379494366, + "loss": 0.6836, + "step": 20699 + }, + { + "epoch": 0.4852405984634048, + "grad_norm": 0.09848837554454803, + "learning_rate": 0.0001046721256938905, + "loss": 0.0197, + "step": 20700 + }, + { + "epoch": 0.4852640400382098, + "grad_norm": 0.8089907765388489, + "learning_rate": 0.00010466476756748624, + "loss": 0.1703, + "step": 20701 + }, + { + "epoch": 0.48528748161301477, + "grad_norm": 0.9550922513008118, + "learning_rate": 0.00010465740941577087, + "loss": 0.7034, + "step": 20702 + }, + { + "epoch": 0.48531092318781976, + "grad_norm": 0.7107685208320618, + "learning_rate": 0.00010465005123878423, + "loss": 0.0762, + "step": 20703 + }, + { + "epoch": 0.48533436476262476, + "grad_norm": 0.2639281153678894, + "learning_rate": 0.0001046426930365663, + "loss": 0.0658, + "step": 20704 + }, + { + "epoch": 0.48535780633742975, + "grad_norm": 0.867262601852417, + "learning_rate": 0.00010463533480915699, + "loss": 0.1388, + "step": 20705 + }, + { + "epoch": 0.48538124791223475, + "grad_norm": 0.606471836566925, + "learning_rate": 0.0001046279765565962, + "loss": 0.0727, + "step": 20706 + }, + { + "epoch": 0.48540468948703974, + "grad_norm": 0.2760442793369293, + "learning_rate": 0.00010462061827892389, + "loss": 0.0389, + "step": 20707 + }, + { + "epoch": 0.48542813106184474, + "grad_norm": 0.39960938692092896, + "learning_rate": 0.00010461325997617998, + "loss": 0.0496, + "step": 20708 + }, + { + "epoch": 0.48545157263664973, + "grad_norm": 0.45881929993629456, + "learning_rate": 0.00010460590164840441, + "loss": 0.0653, + "step": 20709 + }, + { + "epoch": 0.4854750142114547, + "grad_norm": 0.5269656777381897, + "learning_rate": 0.00010459854329563705, + "loss": 0.1127, + "step": 20710 + }, + { + "epoch": 0.4854984557862597, + "grad_norm": 0.3740833103656769, + "learning_rate": 0.00010459118491791788, + "loss": 0.1085, + "step": 20711 + }, + { + "epoch": 0.4855218973610647, + "grad_norm": 0.3690965473651886, + "learning_rate": 0.00010458382651528682, + "loss": 0.0419, + "step": 20712 + }, + { + "epoch": 0.4855453389358697, + "grad_norm": 0.4275684356689453, + "learning_rate": 0.00010457646808778378, + "loss": 0.1008, + "step": 20713 + }, + { + "epoch": 0.4855687805106747, + "grad_norm": 0.20335470139980316, + "learning_rate": 0.00010456910963544868, + "loss": 0.0586, + "step": 20714 + }, + { + "epoch": 0.4855922220854797, + "grad_norm": 0.6547207832336426, + "learning_rate": 0.0001045617511583215, + "loss": 0.1353, + "step": 20715 + }, + { + "epoch": 0.4856156636602847, + "grad_norm": 0.2347927838563919, + "learning_rate": 0.00010455439265644211, + "loss": 0.0277, + "step": 20716 + }, + { + "epoch": 0.4856391052350897, + "grad_norm": 0.299451619386673, + "learning_rate": 0.00010454703412985045, + "loss": 0.0381, + "step": 20717 + }, + { + "epoch": 0.4856625468098947, + "grad_norm": 0.3783546984195709, + "learning_rate": 0.00010453967557858646, + "loss": 0.0637, + "step": 20718 + }, + { + "epoch": 0.4856859883846997, + "grad_norm": 0.42932552099227905, + "learning_rate": 0.00010453231700269008, + "loss": 0.106, + "step": 20719 + }, + { + "epoch": 0.48570942995950467, + "grad_norm": 0.26787081360816956, + "learning_rate": 0.00010452495840220118, + "loss": 0.0369, + "step": 20720 + }, + { + "epoch": 0.48573287153430966, + "grad_norm": 0.16648782789707184, + "learning_rate": 0.00010451759977715978, + "loss": 0.0196, + "step": 20721 + }, + { + "epoch": 0.48575631310911466, + "grad_norm": 0.13065581023693085, + "learning_rate": 0.00010451024112760573, + "loss": 0.0274, + "step": 20722 + }, + { + "epoch": 0.48577975468391965, + "grad_norm": 0.3207824230194092, + "learning_rate": 0.00010450288245357897, + "loss": 0.0457, + "step": 20723 + }, + { + "epoch": 0.48580319625872465, + "grad_norm": 0.3772318661212921, + "learning_rate": 0.00010449552375511946, + "loss": 0.0673, + "step": 20724 + }, + { + "epoch": 0.48582663783352964, + "grad_norm": 0.4340544044971466, + "learning_rate": 0.0001044881650322671, + "loss": 0.0753, + "step": 20725 + }, + { + "epoch": 0.48585007940833463, + "grad_norm": 0.7986019253730774, + "learning_rate": 0.00010448080628506184, + "loss": 0.548, + "step": 20726 + }, + { + "epoch": 0.48587352098313963, + "grad_norm": 0.11471486836671829, + "learning_rate": 0.00010447344751354362, + "loss": 0.022, + "step": 20727 + }, + { + "epoch": 0.4858969625579446, + "grad_norm": 0.5864428877830505, + "learning_rate": 0.00010446608871775234, + "loss": 0.7737, + "step": 20728 + }, + { + "epoch": 0.4859204041327496, + "grad_norm": 0.35640478134155273, + "learning_rate": 0.00010445872989772792, + "loss": 0.032, + "step": 20729 + }, + { + "epoch": 0.4859438457075546, + "grad_norm": 0.5139234066009521, + "learning_rate": 0.00010445137105351028, + "loss": 0.5111, + "step": 20730 + }, + { + "epoch": 0.4859672872823596, + "grad_norm": 0.5377563238143921, + "learning_rate": 0.00010444401218513942, + "loss": 0.0887, + "step": 20731 + }, + { + "epoch": 0.4859907288571646, + "grad_norm": 0.2832418978214264, + "learning_rate": 0.00010443665329265522, + "loss": 0.0431, + "step": 20732 + }, + { + "epoch": 0.4860141704319696, + "grad_norm": 0.6638471484184265, + "learning_rate": 0.00010442929437609761, + "loss": 0.0455, + "step": 20733 + }, + { + "epoch": 0.4860376120067746, + "grad_norm": 0.6878257989883423, + "learning_rate": 0.00010442193543550652, + "loss": 0.528, + "step": 20734 + }, + { + "epoch": 0.4860610535815796, + "grad_norm": 0.3201475441455841, + "learning_rate": 0.0001044145764709219, + "loss": 0.0788, + "step": 20735 + }, + { + "epoch": 0.48608449515638463, + "grad_norm": 0.37912294268608093, + "learning_rate": 0.0001044072174823836, + "loss": 0.0764, + "step": 20736 + }, + { + "epoch": 0.4861079367311896, + "grad_norm": 0.4647120535373688, + "learning_rate": 0.00010439985846993167, + "loss": 0.0664, + "step": 20737 + }, + { + "epoch": 0.4861313783059946, + "grad_norm": 0.7194682359695435, + "learning_rate": 0.000104392499433606, + "loss": 0.1302, + "step": 20738 + }, + { + "epoch": 0.4861548198807996, + "grad_norm": 0.14268887042999268, + "learning_rate": 0.00010438514037344647, + "loss": 0.0173, + "step": 20739 + }, + { + "epoch": 0.4861782614556046, + "grad_norm": 0.590287983417511, + "learning_rate": 0.00010437778128949307, + "loss": 0.1167, + "step": 20740 + }, + { + "epoch": 0.4862017030304096, + "grad_norm": 0.6298198103904724, + "learning_rate": 0.0001043704221817857, + "loss": 0.0578, + "step": 20741 + }, + { + "epoch": 0.4862251446052146, + "grad_norm": 0.539138674736023, + "learning_rate": 0.00010436306305036428, + "loss": 0.0746, + "step": 20742 + }, + { + "epoch": 0.4862485861800196, + "grad_norm": 0.5723258852958679, + "learning_rate": 0.00010435570389526875, + "loss": 0.1261, + "step": 20743 + }, + { + "epoch": 0.4862720277548246, + "grad_norm": 0.09212710708379745, + "learning_rate": 0.00010434834471653907, + "loss": 0.0226, + "step": 20744 + }, + { + "epoch": 0.4862954693296296, + "grad_norm": 0.4360906779766083, + "learning_rate": 0.00010434098551421514, + "loss": 0.1038, + "step": 20745 + }, + { + "epoch": 0.4863189109044346, + "grad_norm": 0.5575463771820068, + "learning_rate": 0.0001043336262883369, + "loss": 0.0948, + "step": 20746 + }, + { + "epoch": 0.48634235247923957, + "grad_norm": 0.4752744436264038, + "learning_rate": 0.0001043262670389443, + "loss": 0.0755, + "step": 20747 + }, + { + "epoch": 0.48636579405404456, + "grad_norm": 0.5388439893722534, + "learning_rate": 0.00010431890776607722, + "loss": 0.0853, + "step": 20748 + }, + { + "epoch": 0.48638923562884956, + "grad_norm": 0.5392056107521057, + "learning_rate": 0.00010431154846977564, + "loss": 0.1469, + "step": 20749 + }, + { + "epoch": 0.48641267720365455, + "grad_norm": 0.6909751296043396, + "learning_rate": 0.00010430418915007949, + "loss": 0.6933, + "step": 20750 + }, + { + "epoch": 0.48643611877845955, + "grad_norm": 0.5084943175315857, + "learning_rate": 0.00010429682980702868, + "loss": 0.101, + "step": 20751 + }, + { + "epoch": 0.48645956035326454, + "grad_norm": 0.7219595909118652, + "learning_rate": 0.00010428947044066315, + "loss": 0.6448, + "step": 20752 + }, + { + "epoch": 0.48648300192806954, + "grad_norm": 0.2022770345211029, + "learning_rate": 0.00010428211105102283, + "loss": 0.0388, + "step": 20753 + }, + { + "epoch": 0.48650644350287453, + "grad_norm": 0.3611328899860382, + "learning_rate": 0.00010427475163814766, + "loss": 0.1078, + "step": 20754 + }, + { + "epoch": 0.4865298850776795, + "grad_norm": 0.1680176854133606, + "learning_rate": 0.00010426739220207755, + "loss": 0.0237, + "step": 20755 + }, + { + "epoch": 0.4865533266524845, + "grad_norm": 0.595996618270874, + "learning_rate": 0.00010426003274285248, + "loss": 0.11, + "step": 20756 + }, + { + "epoch": 0.4865767682272895, + "grad_norm": 0.30489107966423035, + "learning_rate": 0.00010425267326051234, + "loss": 0.0525, + "step": 20757 + }, + { + "epoch": 0.4866002098020945, + "grad_norm": 0.6216658353805542, + "learning_rate": 0.00010424531375509707, + "loss": 0.1222, + "step": 20758 + }, + { + "epoch": 0.4866236513768995, + "grad_norm": 0.7027052640914917, + "learning_rate": 0.00010423795422664662, + "loss": 0.5716, + "step": 20759 + }, + { + "epoch": 0.4866470929517045, + "grad_norm": 0.5497867465019226, + "learning_rate": 0.00010423059467520091, + "loss": 0.1445, + "step": 20760 + }, + { + "epoch": 0.4866705345265095, + "grad_norm": 0.12367154657840729, + "learning_rate": 0.00010422323510079985, + "loss": 0.0245, + "step": 20761 + }, + { + "epoch": 0.4866939761013145, + "grad_norm": 0.409902960062027, + "learning_rate": 0.0001042158755034834, + "loss": 0.0643, + "step": 20762 + }, + { + "epoch": 0.4867174176761195, + "grad_norm": 0.48549163341522217, + "learning_rate": 0.00010420851588329152, + "loss": 0.0852, + "step": 20763 + }, + { + "epoch": 0.4867408592509245, + "grad_norm": 0.5862799882888794, + "learning_rate": 0.0001042011562402641, + "loss": 0.1733, + "step": 20764 + }, + { + "epoch": 0.48676430082572947, + "grad_norm": 0.6822332739830017, + "learning_rate": 0.00010419379657444108, + "loss": 0.1521, + "step": 20765 + }, + { + "epoch": 0.48678774240053446, + "grad_norm": 0.5390506982803345, + "learning_rate": 0.00010418643688586243, + "loss": 0.0847, + "step": 20766 + }, + { + "epoch": 0.48681118397533946, + "grad_norm": 0.5959839820861816, + "learning_rate": 0.00010417907717456803, + "loss": 0.1919, + "step": 20767 + }, + { + "epoch": 0.48683462555014445, + "grad_norm": 0.450313001871109, + "learning_rate": 0.00010417171744059782, + "loss": 0.033, + "step": 20768 + }, + { + "epoch": 0.48685806712494945, + "grad_norm": 0.34355226159095764, + "learning_rate": 0.0001041643576839918, + "loss": 0.1345, + "step": 20769 + }, + { + "epoch": 0.48688150869975444, + "grad_norm": 0.2998731732368469, + "learning_rate": 0.00010415699790478983, + "loss": 0.0572, + "step": 20770 + }, + { + "epoch": 0.48690495027455943, + "grad_norm": 0.1181228831410408, + "learning_rate": 0.00010414963810303189, + "loss": 0.0308, + "step": 20771 + }, + { + "epoch": 0.48692839184936443, + "grad_norm": 0.4748249351978302, + "learning_rate": 0.00010414227827875789, + "loss": 0.0671, + "step": 20772 + }, + { + "epoch": 0.4869518334241694, + "grad_norm": 0.7509123682975769, + "learning_rate": 0.00010413491843200775, + "loss": 0.262, + "step": 20773 + }, + { + "epoch": 0.4869752749989744, + "grad_norm": 0.4464747905731201, + "learning_rate": 0.00010412755856282143, + "loss": 0.5251, + "step": 20774 + }, + { + "epoch": 0.4869987165737794, + "grad_norm": 0.3334181606769562, + "learning_rate": 0.00010412019867123888, + "loss": 0.0581, + "step": 20775 + }, + { + "epoch": 0.4870221581485844, + "grad_norm": 0.8651728630065918, + "learning_rate": 0.00010411283875730002, + "loss": 0.7926, + "step": 20776 + }, + { + "epoch": 0.4870455997233894, + "grad_norm": 0.48362669348716736, + "learning_rate": 0.00010410547882104478, + "loss": 0.1293, + "step": 20777 + }, + { + "epoch": 0.4870690412981944, + "grad_norm": 0.5047246813774109, + "learning_rate": 0.00010409811886251309, + "loss": 0.4448, + "step": 20778 + }, + { + "epoch": 0.4870924828729994, + "grad_norm": 0.632426381111145, + "learning_rate": 0.0001040907588817449, + "loss": 0.1344, + "step": 20779 + }, + { + "epoch": 0.4871159244478044, + "grad_norm": 0.4568912982940674, + "learning_rate": 0.00010408339887878012, + "loss": 0.111, + "step": 20780 + }, + { + "epoch": 0.4871393660226094, + "grad_norm": 0.6256784796714783, + "learning_rate": 0.0001040760388536587, + "loss": 0.1702, + "step": 20781 + }, + { + "epoch": 0.48716280759741437, + "grad_norm": 0.5029202699661255, + "learning_rate": 0.0001040686788064206, + "loss": 0.5644, + "step": 20782 + }, + { + "epoch": 0.48718624917221937, + "grad_norm": 0.23931993544101715, + "learning_rate": 0.00010406131873710571, + "loss": 0.0628, + "step": 20783 + }, + { + "epoch": 0.48720969074702436, + "grad_norm": 0.6864917874336243, + "learning_rate": 0.00010405395864575402, + "loss": 0.1867, + "step": 20784 + }, + { + "epoch": 0.48723313232182935, + "grad_norm": 0.691143274307251, + "learning_rate": 0.00010404659853240541, + "loss": 0.0714, + "step": 20785 + }, + { + "epoch": 0.48725657389663435, + "grad_norm": 0.0829833596944809, + "learning_rate": 0.00010403923839709986, + "loss": 0.0114, + "step": 20786 + }, + { + "epoch": 0.4872800154714394, + "grad_norm": 0.4907161295413971, + "learning_rate": 0.00010403187823987727, + "loss": 0.1979, + "step": 20787 + }, + { + "epoch": 0.4873034570462444, + "grad_norm": 0.7444694638252258, + "learning_rate": 0.00010402451806077762, + "loss": 0.2137, + "step": 20788 + }, + { + "epoch": 0.4873268986210494, + "grad_norm": 0.9737781882286072, + "learning_rate": 0.00010401715785984083, + "loss": 0.0625, + "step": 20789 + }, + { + "epoch": 0.4873503401958544, + "grad_norm": 0.20597200095653534, + "learning_rate": 0.00010400979763710678, + "loss": 0.0657, + "step": 20790 + }, + { + "epoch": 0.4873737817706594, + "grad_norm": 0.5870506763458252, + "learning_rate": 0.00010400243739261548, + "loss": 0.193, + "step": 20791 + }, + { + "epoch": 0.48739722334546437, + "grad_norm": 0.7376627922058105, + "learning_rate": 0.00010399507712640686, + "loss": 0.1792, + "step": 20792 + }, + { + "epoch": 0.48742066492026936, + "grad_norm": 0.741640567779541, + "learning_rate": 0.0001039877168385208, + "loss": 0.1801, + "step": 20793 + }, + { + "epoch": 0.48744410649507436, + "grad_norm": 0.1765630841255188, + "learning_rate": 0.00010398035652899731, + "loss": 0.0378, + "step": 20794 + }, + { + "epoch": 0.48746754806987935, + "grad_norm": 0.40018951892852783, + "learning_rate": 0.0001039729961978763, + "loss": 0.0431, + "step": 20795 + }, + { + "epoch": 0.48749098964468435, + "grad_norm": 0.7456658482551575, + "learning_rate": 0.00010396563584519767, + "loss": 0.1983, + "step": 20796 + }, + { + "epoch": 0.48751443121948934, + "grad_norm": 0.3294600248336792, + "learning_rate": 0.00010395827547100141, + "loss": 0.0678, + "step": 20797 + }, + { + "epoch": 0.48753787279429434, + "grad_norm": 0.2999517321586609, + "learning_rate": 0.00010395091507532744, + "loss": 0.0381, + "step": 20798 + }, + { + "epoch": 0.48756131436909933, + "grad_norm": 0.3139117658138275, + "learning_rate": 0.00010394355465821566, + "loss": 0.3058, + "step": 20799 + }, + { + "epoch": 0.4875847559439043, + "grad_norm": 0.6063418388366699, + "learning_rate": 0.00010393619421970609, + "loss": 0.0302, + "step": 20800 + }, + { + "epoch": 0.4876081975187093, + "grad_norm": 0.6761831045150757, + "learning_rate": 0.0001039288337598386, + "loss": 0.1069, + "step": 20801 + }, + { + "epoch": 0.4876316390935143, + "grad_norm": 0.4563533365726471, + "learning_rate": 0.00010392147327865314, + "loss": 0.087, + "step": 20802 + }, + { + "epoch": 0.4876550806683193, + "grad_norm": 0.5262436866760254, + "learning_rate": 0.00010391411277618963, + "loss": 0.1104, + "step": 20803 + }, + { + "epoch": 0.4876785222431243, + "grad_norm": 0.4913216829299927, + "learning_rate": 0.00010390675225248807, + "loss": 0.1546, + "step": 20804 + }, + { + "epoch": 0.4877019638179293, + "grad_norm": 0.5366776585578918, + "learning_rate": 0.00010389939170758838, + "loss": 0.4931, + "step": 20805 + }, + { + "epoch": 0.4877254053927343, + "grad_norm": 0.42104771733283997, + "learning_rate": 0.00010389203114153044, + "loss": 0.1041, + "step": 20806 + }, + { + "epoch": 0.4877488469675393, + "grad_norm": 0.10675670951604843, + "learning_rate": 0.00010388467055435424, + "loss": 0.0226, + "step": 20807 + }, + { + "epoch": 0.4877722885423443, + "grad_norm": 0.8216441869735718, + "learning_rate": 0.00010387730994609972, + "loss": 0.1222, + "step": 20808 + }, + { + "epoch": 0.4877957301171493, + "grad_norm": 0.3715285062789917, + "learning_rate": 0.00010386994931680678, + "loss": 0.0811, + "step": 20809 + }, + { + "epoch": 0.48781917169195427, + "grad_norm": 0.7102490663528442, + "learning_rate": 0.00010386258866651541, + "loss": 0.0648, + "step": 20810 + }, + { + "epoch": 0.48784261326675926, + "grad_norm": 0.3164599537849426, + "learning_rate": 0.00010385522799526554, + "loss": 0.0629, + "step": 20811 + }, + { + "epoch": 0.48786605484156426, + "grad_norm": 0.7406403422355652, + "learning_rate": 0.00010384786730309705, + "loss": 0.086, + "step": 20812 + }, + { + "epoch": 0.48788949641636925, + "grad_norm": 0.45076602697372437, + "learning_rate": 0.00010384050659004997, + "loss": 0.1758, + "step": 20813 + }, + { + "epoch": 0.48791293799117424, + "grad_norm": 0.29064321517944336, + "learning_rate": 0.00010383314585616419, + "loss": 0.0573, + "step": 20814 + }, + { + "epoch": 0.48793637956597924, + "grad_norm": 0.4894053041934967, + "learning_rate": 0.00010382578510147963, + "loss": 0.0999, + "step": 20815 + }, + { + "epoch": 0.48795982114078423, + "grad_norm": 0.20557758212089539, + "learning_rate": 0.00010381842432603624, + "loss": 0.0298, + "step": 20816 + }, + { + "epoch": 0.4879832627155892, + "grad_norm": 0.9469200372695923, + "learning_rate": 0.000103811063529874, + "loss": 0.1574, + "step": 20817 + }, + { + "epoch": 0.4880067042903942, + "grad_norm": 0.3890448212623596, + "learning_rate": 0.00010380370271303282, + "loss": 0.0483, + "step": 20818 + }, + { + "epoch": 0.4880301458651992, + "grad_norm": 0.24887815117835999, + "learning_rate": 0.00010379634187555263, + "loss": 0.0277, + "step": 20819 + }, + { + "epoch": 0.4880535874400042, + "grad_norm": 0.48455432057380676, + "learning_rate": 0.00010378898101747341, + "loss": 0.0902, + "step": 20820 + }, + { + "epoch": 0.4880770290148092, + "grad_norm": 0.7757640480995178, + "learning_rate": 0.00010378162013883503, + "loss": 0.4775, + "step": 20821 + }, + { + "epoch": 0.4881004705896142, + "grad_norm": 0.3093080222606659, + "learning_rate": 0.00010377425923967747, + "loss": 0.0653, + "step": 20822 + }, + { + "epoch": 0.4881239121644192, + "grad_norm": 0.12905627489089966, + "learning_rate": 0.00010376689832004072, + "loss": 0.0179, + "step": 20823 + }, + { + "epoch": 0.4881473537392242, + "grad_norm": 0.4817023277282715, + "learning_rate": 0.00010375953737996464, + "loss": 0.0531, + "step": 20824 + }, + { + "epoch": 0.4881707953140292, + "grad_norm": 0.7134148478507996, + "learning_rate": 0.00010375217641948923, + "loss": 0.129, + "step": 20825 + }, + { + "epoch": 0.4881942368888342, + "grad_norm": 0.6032009124755859, + "learning_rate": 0.0001037448154386544, + "loss": 0.1416, + "step": 20826 + }, + { + "epoch": 0.48821767846363917, + "grad_norm": 0.35089588165283203, + "learning_rate": 0.00010373745443750007, + "loss": 0.0674, + "step": 20827 + }, + { + "epoch": 0.48824112003844417, + "grad_norm": 0.519616961479187, + "learning_rate": 0.00010373009341606621, + "loss": 0.1392, + "step": 20828 + }, + { + "epoch": 0.48826456161324916, + "grad_norm": 0.4511457085609436, + "learning_rate": 0.00010372273237439279, + "loss": 0.5882, + "step": 20829 + }, + { + "epoch": 0.48828800318805415, + "grad_norm": 0.293250173330307, + "learning_rate": 0.00010371537131251969, + "loss": 0.0325, + "step": 20830 + }, + { + "epoch": 0.48831144476285915, + "grad_norm": 0.3903014659881592, + "learning_rate": 0.00010370801023048689, + "loss": 0.4646, + "step": 20831 + }, + { + "epoch": 0.48833488633766414, + "grad_norm": 0.458454966545105, + "learning_rate": 0.00010370064912833432, + "loss": 0.0919, + "step": 20832 + }, + { + "epoch": 0.48835832791246914, + "grad_norm": 0.7489913702011108, + "learning_rate": 0.00010369328800610192, + "loss": 0.1719, + "step": 20833 + }, + { + "epoch": 0.48838176948727413, + "grad_norm": 0.48573294281959534, + "learning_rate": 0.00010368592686382966, + "loss": 0.4531, + "step": 20834 + }, + { + "epoch": 0.4884052110620791, + "grad_norm": 0.4248761832714081, + "learning_rate": 0.00010367856570155744, + "loss": 0.0464, + "step": 20835 + }, + { + "epoch": 0.4884286526368841, + "grad_norm": 0.35661664605140686, + "learning_rate": 0.00010367120451932521, + "loss": 0.0999, + "step": 20836 + }, + { + "epoch": 0.4884520942116891, + "grad_norm": 0.17137721180915833, + "learning_rate": 0.00010366384331717292, + "loss": 0.0319, + "step": 20837 + }, + { + "epoch": 0.48847553578649416, + "grad_norm": 0.893258273601532, + "learning_rate": 0.00010365648209514054, + "loss": 0.2102, + "step": 20838 + }, + { + "epoch": 0.48849897736129916, + "grad_norm": 0.12135003507137299, + "learning_rate": 0.000103649120853268, + "loss": 0.0151, + "step": 20839 + }, + { + "epoch": 0.48852241893610415, + "grad_norm": 0.4161592125892639, + "learning_rate": 0.00010364175959159517, + "loss": 0.0853, + "step": 20840 + }, + { + "epoch": 0.48854586051090915, + "grad_norm": 0.19413907825946808, + "learning_rate": 0.00010363439831016207, + "loss": 0.0275, + "step": 20841 + }, + { + "epoch": 0.48856930208571414, + "grad_norm": 0.3082592487335205, + "learning_rate": 0.00010362703700900863, + "loss": 0.0353, + "step": 20842 + }, + { + "epoch": 0.48859274366051914, + "grad_norm": 0.11584697663784027, + "learning_rate": 0.00010361967568817479, + "loss": 0.0349, + "step": 20843 + }, + { + "epoch": 0.48861618523532413, + "grad_norm": 0.4132208526134491, + "learning_rate": 0.00010361231434770047, + "loss": 0.1016, + "step": 20844 + }, + { + "epoch": 0.4886396268101291, + "grad_norm": 0.34015166759490967, + "learning_rate": 0.00010360495298762564, + "loss": 0.0384, + "step": 20845 + }, + { + "epoch": 0.4886630683849341, + "grad_norm": 0.460327684879303, + "learning_rate": 0.00010359759160799024, + "loss": 0.1, + "step": 20846 + }, + { + "epoch": 0.4886865099597391, + "grad_norm": 0.16200648248195648, + "learning_rate": 0.00010359023020883418, + "loss": 0.0275, + "step": 20847 + }, + { + "epoch": 0.4887099515345441, + "grad_norm": 1.43291175365448, + "learning_rate": 0.00010358286879019747, + "loss": 0.0865, + "step": 20848 + }, + { + "epoch": 0.4887333931093491, + "grad_norm": 0.24366439878940582, + "learning_rate": 0.00010357550735212, + "loss": 0.0439, + "step": 20849 + }, + { + "epoch": 0.4887568346841541, + "grad_norm": 0.40809720754623413, + "learning_rate": 0.00010356814589464173, + "loss": 0.0695, + "step": 20850 + }, + { + "epoch": 0.4887802762589591, + "grad_norm": 0.5665215849876404, + "learning_rate": 0.0001035607844178026, + "loss": 0.1401, + "step": 20851 + }, + { + "epoch": 0.4888037178337641, + "grad_norm": 0.7286104559898376, + "learning_rate": 0.00010355342292164253, + "loss": 0.1637, + "step": 20852 + }, + { + "epoch": 0.4888271594085691, + "grad_norm": 0.20345249772071838, + "learning_rate": 0.00010354606140620151, + "loss": 0.0446, + "step": 20853 + }, + { + "epoch": 0.4888506009833741, + "grad_norm": 0.2960551977157593, + "learning_rate": 0.00010353869987151945, + "loss": 0.079, + "step": 20854 + }, + { + "epoch": 0.48887404255817907, + "grad_norm": 0.3543388843536377, + "learning_rate": 0.00010353133831763633, + "loss": 0.0465, + "step": 20855 + }, + { + "epoch": 0.48889748413298406, + "grad_norm": 0.3582390546798706, + "learning_rate": 0.00010352397674459202, + "loss": 0.0806, + "step": 20856 + }, + { + "epoch": 0.48892092570778906, + "grad_norm": 0.36865657567977905, + "learning_rate": 0.00010351661515242657, + "loss": 0.3804, + "step": 20857 + }, + { + "epoch": 0.48894436728259405, + "grad_norm": 0.33930256962776184, + "learning_rate": 0.00010350925354117982, + "loss": 0.0987, + "step": 20858 + }, + { + "epoch": 0.48896780885739904, + "grad_norm": 0.11095818877220154, + "learning_rate": 0.00010350189191089179, + "loss": 0.0122, + "step": 20859 + }, + { + "epoch": 0.48899125043220404, + "grad_norm": 0.10155574232339859, + "learning_rate": 0.00010349453026160239, + "loss": 0.018, + "step": 20860 + }, + { + "epoch": 0.48901469200700903, + "grad_norm": 0.044022466987371445, + "learning_rate": 0.00010348716859335157, + "loss": 0.0042, + "step": 20861 + }, + { + "epoch": 0.489038133581814, + "grad_norm": 0.3310277462005615, + "learning_rate": 0.00010347980690617929, + "loss": 0.0089, + "step": 20862 + }, + { + "epoch": 0.489061575156619, + "grad_norm": 0.45102399587631226, + "learning_rate": 0.00010347244520012545, + "loss": 0.0472, + "step": 20863 + }, + { + "epoch": 0.489085016731424, + "grad_norm": 0.605817437171936, + "learning_rate": 0.00010346508347523001, + "loss": 0.1761, + "step": 20864 + }, + { + "epoch": 0.489108458306229, + "grad_norm": 0.24265669286251068, + "learning_rate": 0.00010345772173153296, + "loss": 0.044, + "step": 20865 + }, + { + "epoch": 0.489131899881034, + "grad_norm": 0.8839640617370605, + "learning_rate": 0.00010345035996907422, + "loss": 0.0758, + "step": 20866 + }, + { + "epoch": 0.489155341455839, + "grad_norm": 0.27099958062171936, + "learning_rate": 0.00010344299818789371, + "loss": 0.0398, + "step": 20867 + }, + { + "epoch": 0.489178783030644, + "grad_norm": 0.9911618232727051, + "learning_rate": 0.00010343563638803144, + "loss": 0.1449, + "step": 20868 + }, + { + "epoch": 0.489202224605449, + "grad_norm": 0.9245631694793701, + "learning_rate": 0.00010342827456952726, + "loss": 0.1811, + "step": 20869 + }, + { + "epoch": 0.489225666180254, + "grad_norm": 0.3964498043060303, + "learning_rate": 0.00010342091273242116, + "loss": 0.1069, + "step": 20870 + }, + { + "epoch": 0.489249107755059, + "grad_norm": 0.13851112127304077, + "learning_rate": 0.00010341355087675313, + "loss": 0.0279, + "step": 20871 + }, + { + "epoch": 0.48927254932986397, + "grad_norm": 0.42210036516189575, + "learning_rate": 0.00010340618900256306, + "loss": 0.1185, + "step": 20872 + }, + { + "epoch": 0.48929599090466896, + "grad_norm": 1.0290817022323608, + "learning_rate": 0.0001033988271098909, + "loss": 0.1701, + "step": 20873 + }, + { + "epoch": 0.48931943247947396, + "grad_norm": 0.3984617590904236, + "learning_rate": 0.00010339146519877665, + "loss": 0.0847, + "step": 20874 + }, + { + "epoch": 0.48934287405427895, + "grad_norm": 0.42406684160232544, + "learning_rate": 0.00010338410326926018, + "loss": 0.0788, + "step": 20875 + }, + { + "epoch": 0.48936631562908395, + "grad_norm": 0.06120245158672333, + "learning_rate": 0.00010337674132138146, + "loss": 0.0053, + "step": 20876 + }, + { + "epoch": 0.48938975720388894, + "grad_norm": 0.8480834364891052, + "learning_rate": 0.00010336937935518047, + "loss": 0.2119, + "step": 20877 + }, + { + "epoch": 0.48941319877869394, + "grad_norm": 0.5636926293373108, + "learning_rate": 0.00010336201737069716, + "loss": 0.1598, + "step": 20878 + }, + { + "epoch": 0.48943664035349893, + "grad_norm": 0.32923030853271484, + "learning_rate": 0.00010335465536797137, + "loss": 0.0498, + "step": 20879 + }, + { + "epoch": 0.4894600819283039, + "grad_norm": 0.69510817527771, + "learning_rate": 0.0001033472933470432, + "loss": 0.1345, + "step": 20880 + }, + { + "epoch": 0.4894835235031089, + "grad_norm": 0.6722105145454407, + "learning_rate": 0.0001033399313079525, + "loss": 0.1687, + "step": 20881 + }, + { + "epoch": 0.4895069650779139, + "grad_norm": 0.4641258418560028, + "learning_rate": 0.0001033325692507392, + "loss": 0.0856, + "step": 20882 + }, + { + "epoch": 0.4895304066527189, + "grad_norm": 0.5468127727508545, + "learning_rate": 0.00010332520717544334, + "loss": 0.1276, + "step": 20883 + }, + { + "epoch": 0.4895538482275239, + "grad_norm": 0.5437325835227966, + "learning_rate": 0.0001033178450821048, + "loss": 0.6272, + "step": 20884 + }, + { + "epoch": 0.4895772898023289, + "grad_norm": 0.4712264835834503, + "learning_rate": 0.0001033104829707635, + "loss": 0.1113, + "step": 20885 + }, + { + "epoch": 0.4896007313771339, + "grad_norm": 0.5399673581123352, + "learning_rate": 0.00010330312084145948, + "loss": 0.0308, + "step": 20886 + }, + { + "epoch": 0.4896241729519389, + "grad_norm": 0.179212749004364, + "learning_rate": 0.00010329575869423262, + "loss": 0.0182, + "step": 20887 + }, + { + "epoch": 0.4896476145267439, + "grad_norm": 0.27251115441322327, + "learning_rate": 0.00010328839652912286, + "loss": 0.0551, + "step": 20888 + }, + { + "epoch": 0.48967105610154893, + "grad_norm": 0.3597577214241028, + "learning_rate": 0.00010328103434617016, + "loss": 0.12, + "step": 20889 + }, + { + "epoch": 0.4896944976763539, + "grad_norm": 0.17734113335609436, + "learning_rate": 0.00010327367214541451, + "loss": 0.0504, + "step": 20890 + }, + { + "epoch": 0.4897179392511589, + "grad_norm": 0.5263267755508423, + "learning_rate": 0.0001032663099268958, + "loss": 0.6509, + "step": 20891 + }, + { + "epoch": 0.4897413808259639, + "grad_norm": 0.6325234174728394, + "learning_rate": 0.00010325894769065401, + "loss": 0.6141, + "step": 20892 + }, + { + "epoch": 0.4897648224007689, + "grad_norm": 0.5961052775382996, + "learning_rate": 0.00010325158543672909, + "loss": 0.158, + "step": 20893 + }, + { + "epoch": 0.4897882639755739, + "grad_norm": 0.868608832359314, + "learning_rate": 0.00010324422316516094, + "loss": 0.6141, + "step": 20894 + }, + { + "epoch": 0.4898117055503789, + "grad_norm": 0.3802376091480255, + "learning_rate": 0.00010323686087598957, + "loss": 0.0504, + "step": 20895 + }, + { + "epoch": 0.4898351471251839, + "grad_norm": 0.20266450941562653, + "learning_rate": 0.0001032294985692549, + "loss": 0.0503, + "step": 20896 + }, + { + "epoch": 0.4898585886999889, + "grad_norm": 0.3665265142917633, + "learning_rate": 0.00010322213624499687, + "loss": 0.3226, + "step": 20897 + }, + { + "epoch": 0.4898820302747939, + "grad_norm": 0.16741739213466644, + "learning_rate": 0.00010321477390325545, + "loss": 0.0355, + "step": 20898 + }, + { + "epoch": 0.4899054718495989, + "grad_norm": 0.17624492943286896, + "learning_rate": 0.00010320741154407058, + "loss": 0.0257, + "step": 20899 + }, + { + "epoch": 0.48992891342440387, + "grad_norm": 0.29377689957618713, + "learning_rate": 0.00010320004916748218, + "loss": 0.0521, + "step": 20900 + }, + { + "epoch": 0.48995235499920886, + "grad_norm": 0.11405996233224869, + "learning_rate": 0.00010319268677353023, + "loss": 0.0212, + "step": 20901 + }, + { + "epoch": 0.48997579657401386, + "grad_norm": 0.6127824187278748, + "learning_rate": 0.0001031853243622547, + "loss": 0.1356, + "step": 20902 + }, + { + "epoch": 0.48999923814881885, + "grad_norm": 0.7576543092727661, + "learning_rate": 0.00010317796193369549, + "loss": 0.2549, + "step": 20903 + }, + { + "epoch": 0.49002267972362384, + "grad_norm": 0.2695333659648895, + "learning_rate": 0.00010317059948789256, + "loss": 0.0509, + "step": 20904 + }, + { + "epoch": 0.49004612129842884, + "grad_norm": 0.21481755375862122, + "learning_rate": 0.00010316323702488589, + "loss": 0.0443, + "step": 20905 + }, + { + "epoch": 0.49006956287323383, + "grad_norm": 0.6403017044067383, + "learning_rate": 0.00010315587454471538, + "loss": 0.5507, + "step": 20906 + }, + { + "epoch": 0.4900930044480388, + "grad_norm": 0.12268751859664917, + "learning_rate": 0.00010314851204742105, + "loss": 0.0311, + "step": 20907 + }, + { + "epoch": 0.4901164460228438, + "grad_norm": 0.5579503774642944, + "learning_rate": 0.00010314114953304275, + "loss": 0.1379, + "step": 20908 + }, + { + "epoch": 0.4901398875976488, + "grad_norm": 0.4797722101211548, + "learning_rate": 0.00010313378700162053, + "loss": 0.0978, + "step": 20909 + }, + { + "epoch": 0.4901633291724538, + "grad_norm": 0.9390314817428589, + "learning_rate": 0.00010312642445319429, + "loss": 0.1883, + "step": 20910 + }, + { + "epoch": 0.4901867707472588, + "grad_norm": 0.20147600769996643, + "learning_rate": 0.00010311906188780396, + "loss": 0.0825, + "step": 20911 + }, + { + "epoch": 0.4902102123220638, + "grad_norm": 0.4575027525424957, + "learning_rate": 0.00010311169930548951, + "loss": 0.1212, + "step": 20912 + }, + { + "epoch": 0.4902336538968688, + "grad_norm": 0.17805634438991547, + "learning_rate": 0.0001031043367062909, + "loss": 0.0326, + "step": 20913 + }, + { + "epoch": 0.4902570954716738, + "grad_norm": 0.9070152640342712, + "learning_rate": 0.00010309697409024805, + "loss": 0.3032, + "step": 20914 + }, + { + "epoch": 0.4902805370464788, + "grad_norm": 0.41463083028793335, + "learning_rate": 0.00010308961145740098, + "loss": 0.0834, + "step": 20915 + }, + { + "epoch": 0.4903039786212838, + "grad_norm": 0.35968446731567383, + "learning_rate": 0.00010308224880778958, + "loss": 0.0416, + "step": 20916 + }, + { + "epoch": 0.49032742019608877, + "grad_norm": 1.278324007987976, + "learning_rate": 0.00010307488614145378, + "loss": 0.1011, + "step": 20917 + }, + { + "epoch": 0.49035086177089376, + "grad_norm": 0.15910054743289948, + "learning_rate": 0.0001030675234584336, + "loss": 0.0364, + "step": 20918 + }, + { + "epoch": 0.49037430334569876, + "grad_norm": 0.33561813831329346, + "learning_rate": 0.00010306016075876892, + "loss": 0.1342, + "step": 20919 + }, + { + "epoch": 0.49039774492050375, + "grad_norm": 0.37500762939453125, + "learning_rate": 0.00010305279804249972, + "loss": 0.0859, + "step": 20920 + }, + { + "epoch": 0.49042118649530875, + "grad_norm": 0.5636974573135376, + "learning_rate": 0.00010304543530966597, + "loss": 0.1129, + "step": 20921 + }, + { + "epoch": 0.49044462807011374, + "grad_norm": 0.44851526618003845, + "learning_rate": 0.0001030380725603076, + "loss": 0.0788, + "step": 20922 + }, + { + "epoch": 0.49046806964491874, + "grad_norm": 0.6871599555015564, + "learning_rate": 0.00010303070979446456, + "loss": 0.09, + "step": 20923 + }, + { + "epoch": 0.49049151121972373, + "grad_norm": 0.41819635033607483, + "learning_rate": 0.00010302334701217681, + "loss": 0.0914, + "step": 20924 + }, + { + "epoch": 0.4905149527945287, + "grad_norm": 0.6131742596626282, + "learning_rate": 0.00010301598421348429, + "loss": 0.109, + "step": 20925 + }, + { + "epoch": 0.4905383943693337, + "grad_norm": 0.43096959590911865, + "learning_rate": 0.00010300862139842695, + "loss": 0.407, + "step": 20926 + }, + { + "epoch": 0.4905618359441387, + "grad_norm": 0.7270746231079102, + "learning_rate": 0.00010300125856704474, + "loss": 0.1468, + "step": 20927 + }, + { + "epoch": 0.4905852775189437, + "grad_norm": 0.4053921401500702, + "learning_rate": 0.00010299389571937762, + "loss": 0.0663, + "step": 20928 + }, + { + "epoch": 0.4906087190937487, + "grad_norm": 0.4340194761753082, + "learning_rate": 0.00010298653285546555, + "loss": 0.0381, + "step": 20929 + }, + { + "epoch": 0.4906321606685537, + "grad_norm": 0.5915300250053406, + "learning_rate": 0.00010297916997534847, + "loss": 0.1199, + "step": 20930 + }, + { + "epoch": 0.4906556022433587, + "grad_norm": 0.7356961369514465, + "learning_rate": 0.00010297180707906631, + "loss": 0.126, + "step": 20931 + }, + { + "epoch": 0.4906790438181637, + "grad_norm": 0.5388287901878357, + "learning_rate": 0.00010296444416665904, + "loss": 0.1525, + "step": 20932 + }, + { + "epoch": 0.4907024853929687, + "grad_norm": 1.0005862712860107, + "learning_rate": 0.00010295708123816664, + "loss": 0.1453, + "step": 20933 + }, + { + "epoch": 0.4907259269677737, + "grad_norm": 0.6867221593856812, + "learning_rate": 0.00010294971829362902, + "loss": 0.1081, + "step": 20934 + }, + { + "epoch": 0.49074936854257867, + "grad_norm": 0.36111024022102356, + "learning_rate": 0.00010294235533308616, + "loss": 0.0694, + "step": 20935 + }, + { + "epoch": 0.49077281011738366, + "grad_norm": 0.3407791554927826, + "learning_rate": 0.00010293499235657795, + "loss": 0.0607, + "step": 20936 + }, + { + "epoch": 0.49079625169218866, + "grad_norm": 0.3094044327735901, + "learning_rate": 0.00010292762936414442, + "loss": 0.0471, + "step": 20937 + }, + { + "epoch": 0.49081969326699365, + "grad_norm": 0.17189067602157593, + "learning_rate": 0.0001029202663558255, + "loss": 0.0137, + "step": 20938 + }, + { + "epoch": 0.49084313484179865, + "grad_norm": 0.28113675117492676, + "learning_rate": 0.00010291290333166112, + "loss": 0.0542, + "step": 20939 + }, + { + "epoch": 0.4908665764166037, + "grad_norm": 0.40736061334609985, + "learning_rate": 0.00010290554029169126, + "loss": 0.1264, + "step": 20940 + }, + { + "epoch": 0.4908900179914087, + "grad_norm": 0.5755582451820374, + "learning_rate": 0.00010289817723595585, + "loss": 0.0973, + "step": 20941 + }, + { + "epoch": 0.4909134595662137, + "grad_norm": 0.11225768178701401, + "learning_rate": 0.00010289081416449484, + "loss": 0.0136, + "step": 20942 + }, + { + "epoch": 0.4909369011410187, + "grad_norm": 0.3306494355201721, + "learning_rate": 0.00010288345107734819, + "loss": 0.0504, + "step": 20943 + }, + { + "epoch": 0.49096034271582367, + "grad_norm": 0.5430710315704346, + "learning_rate": 0.00010287608797455586, + "loss": 0.1117, + "step": 20944 + }, + { + "epoch": 0.49098378429062867, + "grad_norm": 0.570068895816803, + "learning_rate": 0.0001028687248561578, + "loss": 0.0986, + "step": 20945 + }, + { + "epoch": 0.49100722586543366, + "grad_norm": 0.14492978155612946, + "learning_rate": 0.00010286136172219394, + "loss": 0.0307, + "step": 20946 + }, + { + "epoch": 0.49103066744023866, + "grad_norm": 0.6975023150444031, + "learning_rate": 0.00010285399857270427, + "loss": 0.1522, + "step": 20947 + }, + { + "epoch": 0.49105410901504365, + "grad_norm": 0.5503929853439331, + "learning_rate": 0.00010284663540772874, + "loss": 0.1626, + "step": 20948 + }, + { + "epoch": 0.49107755058984864, + "grad_norm": 0.3177453875541687, + "learning_rate": 0.00010283927222730722, + "loss": 0.07, + "step": 20949 + }, + { + "epoch": 0.49110099216465364, + "grad_norm": 0.24033597111701965, + "learning_rate": 0.00010283190903147979, + "loss": 0.0475, + "step": 20950 + }, + { + "epoch": 0.49112443373945863, + "grad_norm": 0.20243316888809204, + "learning_rate": 0.00010282454582028634, + "loss": 0.0659, + "step": 20951 + }, + { + "epoch": 0.4911478753142636, + "grad_norm": 0.308294415473938, + "learning_rate": 0.00010281718259376681, + "loss": 0.059, + "step": 20952 + }, + { + "epoch": 0.4911713168890686, + "grad_norm": 0.3951771855354309, + "learning_rate": 0.00010280981935196117, + "loss": 0.0921, + "step": 20953 + }, + { + "epoch": 0.4911947584638736, + "grad_norm": 0.6374572515487671, + "learning_rate": 0.00010280245609490937, + "loss": 0.726, + "step": 20954 + }, + { + "epoch": 0.4912182000386786, + "grad_norm": 0.9596251845359802, + "learning_rate": 0.00010279509282265132, + "loss": 0.1217, + "step": 20955 + }, + { + "epoch": 0.4912416416134836, + "grad_norm": 0.5256126523017883, + "learning_rate": 0.00010278772953522709, + "loss": 0.1632, + "step": 20956 + }, + { + "epoch": 0.4912650831882886, + "grad_norm": 0.8061120510101318, + "learning_rate": 0.00010278036623267652, + "loss": 0.235, + "step": 20957 + }, + { + "epoch": 0.4912885247630936, + "grad_norm": 0.2570796012878418, + "learning_rate": 0.00010277300291503962, + "loss": 0.038, + "step": 20958 + }, + { + "epoch": 0.4913119663378986, + "grad_norm": 0.47107285261154175, + "learning_rate": 0.00010276563958235634, + "loss": 0.1453, + "step": 20959 + }, + { + "epoch": 0.4913354079127036, + "grad_norm": 0.3581514060497284, + "learning_rate": 0.0001027582762346666, + "loss": 0.0495, + "step": 20960 + }, + { + "epoch": 0.4913588494875086, + "grad_norm": 0.45801040530204773, + "learning_rate": 0.00010275091287201037, + "loss": 0.125, + "step": 20961 + }, + { + "epoch": 0.49138229106231357, + "grad_norm": 0.5795212984085083, + "learning_rate": 0.00010274354949442761, + "loss": 0.6681, + "step": 20962 + }, + { + "epoch": 0.49140573263711856, + "grad_norm": 0.4070345163345337, + "learning_rate": 0.0001027361861019583, + "loss": 0.0502, + "step": 20963 + }, + { + "epoch": 0.49142917421192356, + "grad_norm": 0.3248746395111084, + "learning_rate": 0.00010272882269464237, + "loss": 0.0706, + "step": 20964 + }, + { + "epoch": 0.49145261578672855, + "grad_norm": 0.6273117065429688, + "learning_rate": 0.00010272145927251975, + "loss": 0.1019, + "step": 20965 + }, + { + "epoch": 0.49147605736153355, + "grad_norm": 0.528184175491333, + "learning_rate": 0.00010271409583563043, + "loss": 0.1066, + "step": 20966 + }, + { + "epoch": 0.49149949893633854, + "grad_norm": 0.4595838487148285, + "learning_rate": 0.00010270673238401432, + "loss": 0.1076, + "step": 20967 + }, + { + "epoch": 0.49152294051114354, + "grad_norm": 0.17384454607963562, + "learning_rate": 0.00010269936891771142, + "loss": 0.0315, + "step": 20968 + }, + { + "epoch": 0.49154638208594853, + "grad_norm": 0.5861511826515198, + "learning_rate": 0.00010269200543676168, + "loss": 0.1103, + "step": 20969 + }, + { + "epoch": 0.4915698236607535, + "grad_norm": 0.29543784260749817, + "learning_rate": 0.00010268464194120504, + "loss": 0.0546, + "step": 20970 + }, + { + "epoch": 0.4915932652355585, + "grad_norm": 0.3947504162788391, + "learning_rate": 0.00010267727843108143, + "loss": 0.0913, + "step": 20971 + }, + { + "epoch": 0.4916167068103635, + "grad_norm": 0.8919373154640198, + "learning_rate": 0.00010266991490643088, + "loss": 0.1003, + "step": 20972 + }, + { + "epoch": 0.4916401483851685, + "grad_norm": 0.18612372875213623, + "learning_rate": 0.00010266255136729324, + "loss": 0.0311, + "step": 20973 + }, + { + "epoch": 0.4916635899599735, + "grad_norm": 0.4558051526546478, + "learning_rate": 0.00010265518781370855, + "loss": 0.1226, + "step": 20974 + }, + { + "epoch": 0.4916870315347785, + "grad_norm": 0.4603787660598755, + "learning_rate": 0.00010264782424571675, + "loss": 0.0882, + "step": 20975 + }, + { + "epoch": 0.4917104731095835, + "grad_norm": 0.4310189187526703, + "learning_rate": 0.00010264046066335777, + "loss": 0.1213, + "step": 20976 + }, + { + "epoch": 0.4917339146843885, + "grad_norm": 0.18438296020030975, + "learning_rate": 0.00010263309706667157, + "loss": 0.0364, + "step": 20977 + }, + { + "epoch": 0.4917573562591935, + "grad_norm": 0.11551163345575333, + "learning_rate": 0.00010262573345569813, + "loss": 0.0273, + "step": 20978 + }, + { + "epoch": 0.4917807978339985, + "grad_norm": 0.3246322274208069, + "learning_rate": 0.00010261836983047734, + "loss": 0.0744, + "step": 20979 + }, + { + "epoch": 0.49180423940880347, + "grad_norm": 0.597655177116394, + "learning_rate": 0.00010261100619104926, + "loss": 0.7059, + "step": 20980 + }, + { + "epoch": 0.49182768098360846, + "grad_norm": 0.4109647870063782, + "learning_rate": 0.00010260364253745376, + "loss": 0.0663, + "step": 20981 + }, + { + "epoch": 0.49185112255841346, + "grad_norm": 0.588392436504364, + "learning_rate": 0.00010259627886973082, + "loss": 0.6448, + "step": 20982 + }, + { + "epoch": 0.49187456413321845, + "grad_norm": 0.5166953802108765, + "learning_rate": 0.00010258891518792042, + "loss": 0.1327, + "step": 20983 + }, + { + "epoch": 0.49189800570802344, + "grad_norm": 0.47901615500450134, + "learning_rate": 0.00010258155149206245, + "loss": 0.1477, + "step": 20984 + }, + { + "epoch": 0.49192144728282844, + "grad_norm": 0.5340576171875, + "learning_rate": 0.00010257418778219693, + "loss": 0.0942, + "step": 20985 + }, + { + "epoch": 0.49194488885763343, + "grad_norm": 0.6028409004211426, + "learning_rate": 0.00010256682405836382, + "loss": 0.6129, + "step": 20986 + }, + { + "epoch": 0.4919683304324384, + "grad_norm": 0.62510085105896, + "learning_rate": 0.00010255946032060302, + "loss": 0.758, + "step": 20987 + }, + { + "epoch": 0.4919917720072434, + "grad_norm": 0.17618335783481598, + "learning_rate": 0.00010255209656895452, + "loss": 0.0323, + "step": 20988 + }, + { + "epoch": 0.4920152135820484, + "grad_norm": 0.38393083214759827, + "learning_rate": 0.00010254473280345828, + "loss": 0.0956, + "step": 20989 + }, + { + "epoch": 0.4920386551568534, + "grad_norm": 0.08380149304866791, + "learning_rate": 0.00010253736902415425, + "loss": 0.0116, + "step": 20990 + }, + { + "epoch": 0.49206209673165846, + "grad_norm": 0.5354222655296326, + "learning_rate": 0.00010253000523108238, + "loss": 0.5774, + "step": 20991 + }, + { + "epoch": 0.49208553830646345, + "grad_norm": 0.09563137590885162, + "learning_rate": 0.00010252264142428263, + "loss": 0.0327, + "step": 20992 + }, + { + "epoch": 0.49210897988126845, + "grad_norm": 0.3542593717575073, + "learning_rate": 0.00010251527760379494, + "loss": 0.0544, + "step": 20993 + }, + { + "epoch": 0.49213242145607344, + "grad_norm": 0.3612827658653259, + "learning_rate": 0.00010250791376965932, + "loss": 0.0356, + "step": 20994 + }, + { + "epoch": 0.49215586303087844, + "grad_norm": 0.6699166893959045, + "learning_rate": 0.00010250054992191567, + "loss": 0.0906, + "step": 20995 + }, + { + "epoch": 0.49217930460568343, + "grad_norm": 0.27442869544029236, + "learning_rate": 0.00010249318606060396, + "loss": 0.0999, + "step": 20996 + }, + { + "epoch": 0.4922027461804884, + "grad_norm": 0.5359426736831665, + "learning_rate": 0.00010248582218576417, + "loss": 0.1081, + "step": 20997 + }, + { + "epoch": 0.4922261877552934, + "grad_norm": 0.09906399995088577, + "learning_rate": 0.0001024784582974362, + "loss": 0.0144, + "step": 20998 + }, + { + "epoch": 0.4922496293300984, + "grad_norm": 0.7780545353889465, + "learning_rate": 0.0001024710943956601, + "loss": 0.1765, + "step": 20999 + }, + { + "epoch": 0.4922730709049034, + "grad_norm": 0.4574901759624481, + "learning_rate": 0.00010246373048047573, + "loss": 0.1192, + "step": 21000 + }, + { + "epoch": 0.4922965124797084, + "grad_norm": 0.3414593040943146, + "learning_rate": 0.00010245636655192313, + "loss": 0.0624, + "step": 21001 + }, + { + "epoch": 0.4923199540545134, + "grad_norm": 0.2941637635231018, + "learning_rate": 0.00010244900261004218, + "loss": 0.0486, + "step": 21002 + }, + { + "epoch": 0.4923433956293184, + "grad_norm": 0.3814239501953125, + "learning_rate": 0.00010244163865487289, + "loss": 0.0739, + "step": 21003 + }, + { + "epoch": 0.4923668372041234, + "grad_norm": 0.6830896735191345, + "learning_rate": 0.00010243427468645517, + "loss": 0.11, + "step": 21004 + }, + { + "epoch": 0.4923902787789284, + "grad_norm": 0.21504008769989014, + "learning_rate": 0.00010242691070482904, + "loss": 0.0528, + "step": 21005 + }, + { + "epoch": 0.4924137203537334, + "grad_norm": 0.4710172712802887, + "learning_rate": 0.0001024195467100344, + "loss": 0.0606, + "step": 21006 + }, + { + "epoch": 0.49243716192853837, + "grad_norm": 0.14891815185546875, + "learning_rate": 0.00010241218270211126, + "loss": 0.0364, + "step": 21007 + }, + { + "epoch": 0.49246060350334336, + "grad_norm": 0.1266077309846878, + "learning_rate": 0.00010240481868109955, + "loss": 0.0295, + "step": 21008 + }, + { + "epoch": 0.49248404507814836, + "grad_norm": 0.5326628684997559, + "learning_rate": 0.00010239745464703918, + "loss": 0.5598, + "step": 21009 + }, + { + "epoch": 0.49250748665295335, + "grad_norm": 0.41867896914482117, + "learning_rate": 0.00010239009059997018, + "loss": 0.0546, + "step": 21010 + }, + { + "epoch": 0.49253092822775835, + "grad_norm": 0.6740723252296448, + "learning_rate": 0.0001023827265399325, + "loss": 0.1148, + "step": 21011 + }, + { + "epoch": 0.49255436980256334, + "grad_norm": 0.4321833550930023, + "learning_rate": 0.00010237536246696606, + "loss": 0.0861, + "step": 21012 + }, + { + "epoch": 0.49257781137736834, + "grad_norm": 0.1655508279800415, + "learning_rate": 0.00010236799838111085, + "loss": 0.0514, + "step": 21013 + }, + { + "epoch": 0.49260125295217333, + "grad_norm": 0.3749769926071167, + "learning_rate": 0.0001023606342824068, + "loss": 0.065, + "step": 21014 + }, + { + "epoch": 0.4926246945269783, + "grad_norm": 0.5640936493873596, + "learning_rate": 0.00010235327017089389, + "loss": 0.1612, + "step": 21015 + }, + { + "epoch": 0.4926481361017833, + "grad_norm": 0.5601734519004822, + "learning_rate": 0.00010234590604661206, + "loss": 0.0825, + "step": 21016 + }, + { + "epoch": 0.4926715776765883, + "grad_norm": 0.7160149216651917, + "learning_rate": 0.00010233854190960128, + "loss": 0.6965, + "step": 21017 + }, + { + "epoch": 0.4926950192513933, + "grad_norm": 0.5298972129821777, + "learning_rate": 0.00010233117775990153, + "loss": 0.1414, + "step": 21018 + }, + { + "epoch": 0.4927184608261983, + "grad_norm": 0.4602333903312683, + "learning_rate": 0.00010232381359755272, + "loss": 0.0734, + "step": 21019 + }, + { + "epoch": 0.4927419024010033, + "grad_norm": 0.363985151052475, + "learning_rate": 0.00010231644942259484, + "loss": 0.0513, + "step": 21020 + }, + { + "epoch": 0.4927653439758083, + "grad_norm": 0.3431227505207062, + "learning_rate": 0.00010230908523506784, + "loss": 0.0455, + "step": 21021 + }, + { + "epoch": 0.4927887855506133, + "grad_norm": 0.36270761489868164, + "learning_rate": 0.00010230172103501165, + "loss": 0.0479, + "step": 21022 + }, + { + "epoch": 0.4928122271254183, + "grad_norm": 0.4339991509914398, + "learning_rate": 0.00010229435682246629, + "loss": 0.0591, + "step": 21023 + }, + { + "epoch": 0.4928356687002233, + "grad_norm": 0.11283321678638458, + "learning_rate": 0.0001022869925974717, + "loss": 0.0312, + "step": 21024 + }, + { + "epoch": 0.49285911027502827, + "grad_norm": 0.6156319975852966, + "learning_rate": 0.00010227962836006777, + "loss": 0.125, + "step": 21025 + }, + { + "epoch": 0.49288255184983326, + "grad_norm": 0.5898153781890869, + "learning_rate": 0.00010227226411029456, + "loss": 0.0979, + "step": 21026 + }, + { + "epoch": 0.49290599342463826, + "grad_norm": 0.07953882962465286, + "learning_rate": 0.00010226489984819198, + "loss": 0.0117, + "step": 21027 + }, + { + "epoch": 0.49292943499944325, + "grad_norm": 1.0764174461364746, + "learning_rate": 0.00010225753557379996, + "loss": 0.0858, + "step": 21028 + }, + { + "epoch": 0.49295287657424824, + "grad_norm": 0.5013346672058105, + "learning_rate": 0.00010225017128715851, + "loss": 0.073, + "step": 21029 + }, + { + "epoch": 0.49297631814905324, + "grad_norm": 0.316080778837204, + "learning_rate": 0.00010224280698830755, + "loss": 0.0702, + "step": 21030 + }, + { + "epoch": 0.49299975972385823, + "grad_norm": 0.5572065711021423, + "learning_rate": 0.00010223544267728707, + "loss": 0.1296, + "step": 21031 + }, + { + "epoch": 0.4930232012986632, + "grad_norm": 0.28859537839889526, + "learning_rate": 0.00010222807835413702, + "loss": 0.0474, + "step": 21032 + }, + { + "epoch": 0.4930466428734682, + "grad_norm": 0.4489482343196869, + "learning_rate": 0.00010222071401889735, + "loss": 0.0705, + "step": 21033 + }, + { + "epoch": 0.4930700844482732, + "grad_norm": 0.47895348072052, + "learning_rate": 0.000102213349671608, + "loss": 0.0716, + "step": 21034 + }, + { + "epoch": 0.4930935260230782, + "grad_norm": 0.8023959994316101, + "learning_rate": 0.00010220598531230899, + "loss": 0.6363, + "step": 21035 + }, + { + "epoch": 0.4931169675978832, + "grad_norm": 0.33531755208969116, + "learning_rate": 0.00010219862094104023, + "loss": 0.0869, + "step": 21036 + }, + { + "epoch": 0.4931404091726882, + "grad_norm": 0.47267812490463257, + "learning_rate": 0.00010219125655784167, + "loss": 0.138, + "step": 21037 + }, + { + "epoch": 0.4931638507474932, + "grad_norm": 0.5046987533569336, + "learning_rate": 0.00010218389216275331, + "loss": 0.0806, + "step": 21038 + }, + { + "epoch": 0.4931872923222982, + "grad_norm": 0.5609793663024902, + "learning_rate": 0.0001021765277558151, + "loss": 0.0835, + "step": 21039 + }, + { + "epoch": 0.4932107338971032, + "grad_norm": 0.6662328243255615, + "learning_rate": 0.00010216916333706697, + "loss": 0.0955, + "step": 21040 + }, + { + "epoch": 0.4932341754719082, + "grad_norm": 0.3726671040058136, + "learning_rate": 0.00010216179890654887, + "loss": 0.0601, + "step": 21041 + }, + { + "epoch": 0.4932576170467132, + "grad_norm": 0.4214897155761719, + "learning_rate": 0.00010215443446430084, + "loss": 0.0967, + "step": 21042 + }, + { + "epoch": 0.4932810586215182, + "grad_norm": 0.10209403187036514, + "learning_rate": 0.00010214707001036279, + "loss": 0.0209, + "step": 21043 + }, + { + "epoch": 0.4933045001963232, + "grad_norm": 0.5404754281044006, + "learning_rate": 0.00010213970554477463, + "loss": 0.054, + "step": 21044 + }, + { + "epoch": 0.4933279417711282, + "grad_norm": 0.3744899034500122, + "learning_rate": 0.00010213234106757641, + "loss": 0.501, + "step": 21045 + }, + { + "epoch": 0.4933513833459332, + "grad_norm": 0.2544693648815155, + "learning_rate": 0.00010212497657880801, + "loss": 0.0341, + "step": 21046 + }, + { + "epoch": 0.4933748249207382, + "grad_norm": 0.68529212474823, + "learning_rate": 0.00010211761207850947, + "loss": 0.8369, + "step": 21047 + }, + { + "epoch": 0.4933982664955432, + "grad_norm": 0.4102308750152588, + "learning_rate": 0.00010211024756672067, + "loss": 0.1005, + "step": 21048 + }, + { + "epoch": 0.4934217080703482, + "grad_norm": 0.518735945224762, + "learning_rate": 0.00010210288304348163, + "loss": 0.1253, + "step": 21049 + }, + { + "epoch": 0.4934451496451532, + "grad_norm": 0.279151052236557, + "learning_rate": 0.00010209551850883228, + "loss": 0.0442, + "step": 21050 + }, + { + "epoch": 0.4934685912199582, + "grad_norm": 0.8384619355201721, + "learning_rate": 0.0001020881539628126, + "loss": 0.1146, + "step": 21051 + }, + { + "epoch": 0.49349203279476317, + "grad_norm": 0.3830450475215912, + "learning_rate": 0.0001020807894054625, + "loss": 0.0442, + "step": 21052 + }, + { + "epoch": 0.49351547436956816, + "grad_norm": 0.7945303320884705, + "learning_rate": 0.00010207342483682202, + "loss": 0.1306, + "step": 21053 + }, + { + "epoch": 0.49353891594437316, + "grad_norm": 0.7542617321014404, + "learning_rate": 0.00010206606025693106, + "loss": 0.8004, + "step": 21054 + }, + { + "epoch": 0.49356235751917815, + "grad_norm": 0.7081255316734314, + "learning_rate": 0.00010205869566582962, + "loss": 0.1131, + "step": 21055 + }, + { + "epoch": 0.49358579909398315, + "grad_norm": 0.5713784098625183, + "learning_rate": 0.00010205133106355762, + "loss": 0.4709, + "step": 21056 + }, + { + "epoch": 0.49360924066878814, + "grad_norm": 0.1355937421321869, + "learning_rate": 0.00010204396645015504, + "loss": 0.0168, + "step": 21057 + }, + { + "epoch": 0.49363268224359313, + "grad_norm": 0.3384620249271393, + "learning_rate": 0.00010203660182566184, + "loss": 0.0473, + "step": 21058 + }, + { + "epoch": 0.49365612381839813, + "grad_norm": 0.3781973719596863, + "learning_rate": 0.00010202923719011799, + "loss": 0.0649, + "step": 21059 + }, + { + "epoch": 0.4936795653932031, + "grad_norm": 0.4806217551231384, + "learning_rate": 0.00010202187254356341, + "loss": 0.0889, + "step": 21060 + }, + { + "epoch": 0.4937030069680081, + "grad_norm": 0.6121742129325867, + "learning_rate": 0.00010201450788603815, + "loss": 0.1299, + "step": 21061 + }, + { + "epoch": 0.4937264485428131, + "grad_norm": 0.7971411347389221, + "learning_rate": 0.00010200714321758207, + "loss": 0.2015, + "step": 21062 + }, + { + "epoch": 0.4937498901176181, + "grad_norm": 0.5100234150886536, + "learning_rate": 0.00010199977853823519, + "loss": 0.1302, + "step": 21063 + }, + { + "epoch": 0.4937733316924231, + "grad_norm": 0.6072593927383423, + "learning_rate": 0.00010199241384803743, + "loss": 0.0873, + "step": 21064 + }, + { + "epoch": 0.4937967732672281, + "grad_norm": 0.5899966359138489, + "learning_rate": 0.00010198504914702882, + "loss": 0.1732, + "step": 21065 + }, + { + "epoch": 0.4938202148420331, + "grad_norm": 0.5464089512825012, + "learning_rate": 0.00010197768443524927, + "loss": 0.0754, + "step": 21066 + }, + { + "epoch": 0.4938436564168381, + "grad_norm": 0.35355886816978455, + "learning_rate": 0.00010197031971273872, + "loss": 0.0531, + "step": 21067 + }, + { + "epoch": 0.4938670979916431, + "grad_norm": 0.35768434405326843, + "learning_rate": 0.00010196295497953718, + "loss": 0.0429, + "step": 21068 + }, + { + "epoch": 0.4938905395664481, + "grad_norm": 0.20209825038909912, + "learning_rate": 0.00010195559023568457, + "loss": 0.025, + "step": 21069 + }, + { + "epoch": 0.49391398114125307, + "grad_norm": 0.3997269570827484, + "learning_rate": 0.00010194822548122089, + "loss": 0.1277, + "step": 21070 + }, + { + "epoch": 0.49393742271605806, + "grad_norm": 0.31846368312835693, + "learning_rate": 0.00010194086071618607, + "loss": 0.0719, + "step": 21071 + }, + { + "epoch": 0.49396086429086306, + "grad_norm": 0.37347087264060974, + "learning_rate": 0.00010193349594062012, + "loss": 0.0889, + "step": 21072 + }, + { + "epoch": 0.49398430586566805, + "grad_norm": 0.3530932366847992, + "learning_rate": 0.00010192613115456293, + "loss": 0.2718, + "step": 21073 + }, + { + "epoch": 0.49400774744047304, + "grad_norm": 0.15432102978229523, + "learning_rate": 0.00010191876635805453, + "loss": 0.0352, + "step": 21074 + }, + { + "epoch": 0.49403118901527804, + "grad_norm": 0.34924039244651794, + "learning_rate": 0.00010191140155113484, + "loss": 0.0416, + "step": 21075 + }, + { + "epoch": 0.49405463059008303, + "grad_norm": 0.14937102794647217, + "learning_rate": 0.00010190403673384381, + "loss": 0.0229, + "step": 21076 + }, + { + "epoch": 0.494078072164888, + "grad_norm": 0.5599930882453918, + "learning_rate": 0.0001018966719062214, + "loss": 0.0827, + "step": 21077 + }, + { + "epoch": 0.494101513739693, + "grad_norm": 0.15821239352226257, + "learning_rate": 0.00010188930706830766, + "loss": 0.0228, + "step": 21078 + }, + { + "epoch": 0.494124955314498, + "grad_norm": 0.43810930848121643, + "learning_rate": 0.00010188194222014244, + "loss": 0.1169, + "step": 21079 + }, + { + "epoch": 0.494148396889303, + "grad_norm": 0.346708744764328, + "learning_rate": 0.00010187457736176577, + "loss": 0.025, + "step": 21080 + }, + { + "epoch": 0.494171838464108, + "grad_norm": 0.5084566473960876, + "learning_rate": 0.00010186721249321761, + "loss": 0.0853, + "step": 21081 + }, + { + "epoch": 0.494195280038913, + "grad_norm": 0.6025157570838928, + "learning_rate": 0.00010185984761453786, + "loss": 0.058, + "step": 21082 + }, + { + "epoch": 0.494218721613718, + "grad_norm": 0.1296512633562088, + "learning_rate": 0.00010185248272576655, + "loss": 0.026, + "step": 21083 + }, + { + "epoch": 0.494242163188523, + "grad_norm": 0.6050021648406982, + "learning_rate": 0.00010184511782694362, + "loss": 0.1027, + "step": 21084 + }, + { + "epoch": 0.494265604763328, + "grad_norm": 0.46779516339302063, + "learning_rate": 0.00010183775291810902, + "loss": 0.0549, + "step": 21085 + }, + { + "epoch": 0.494289046338133, + "grad_norm": 0.362400621175766, + "learning_rate": 0.00010183038799930271, + "loss": 0.0604, + "step": 21086 + }, + { + "epoch": 0.49431248791293797, + "grad_norm": 0.4845142960548401, + "learning_rate": 0.00010182302307056468, + "loss": 0.4763, + "step": 21087 + }, + { + "epoch": 0.49433592948774296, + "grad_norm": 0.48683616518974304, + "learning_rate": 0.00010181565813193485, + "loss": 0.0734, + "step": 21088 + }, + { + "epoch": 0.49435937106254796, + "grad_norm": 0.5910977721214294, + "learning_rate": 0.00010180829318345322, + "loss": 0.6459, + "step": 21089 + }, + { + "epoch": 0.49438281263735295, + "grad_norm": 0.37651437520980835, + "learning_rate": 0.00010180092822515976, + "loss": 0.1137, + "step": 21090 + }, + { + "epoch": 0.49440625421215795, + "grad_norm": 0.6224974989891052, + "learning_rate": 0.00010179356325709441, + "loss": 0.125, + "step": 21091 + }, + { + "epoch": 0.49442969578696294, + "grad_norm": 0.26857495307922363, + "learning_rate": 0.00010178619827929712, + "loss": 0.0529, + "step": 21092 + }, + { + "epoch": 0.49445313736176794, + "grad_norm": 0.5442458391189575, + "learning_rate": 0.00010177883329180788, + "loss": 0.1414, + "step": 21093 + }, + { + "epoch": 0.494476578936573, + "grad_norm": 0.5129140019416809, + "learning_rate": 0.00010177146829466666, + "loss": 0.1007, + "step": 21094 + }, + { + "epoch": 0.494500020511378, + "grad_norm": 0.8525775671005249, + "learning_rate": 0.00010176410328791332, + "loss": 0.085, + "step": 21095 + }, + { + "epoch": 0.494523462086183, + "grad_norm": 0.3534892797470093, + "learning_rate": 0.00010175673827158798, + "loss": 0.0956, + "step": 21096 + }, + { + "epoch": 0.49454690366098797, + "grad_norm": 0.13759221136569977, + "learning_rate": 0.00010174937324573052, + "loss": 0.0185, + "step": 21097 + }, + { + "epoch": 0.49457034523579296, + "grad_norm": 0.29970496892929077, + "learning_rate": 0.00010174200821038092, + "loss": 0.0758, + "step": 21098 + }, + { + "epoch": 0.49459378681059796, + "grad_norm": 0.4123709201812744, + "learning_rate": 0.0001017346431655791, + "loss": 0.1081, + "step": 21099 + }, + { + "epoch": 0.49461722838540295, + "grad_norm": 0.5714116096496582, + "learning_rate": 0.00010172727811136508, + "loss": 0.0908, + "step": 21100 + }, + { + "epoch": 0.49464066996020795, + "grad_norm": 0.36846089363098145, + "learning_rate": 0.00010171991304777879, + "loss": 0.0498, + "step": 21101 + }, + { + "epoch": 0.49466411153501294, + "grad_norm": 0.8200199007987976, + "learning_rate": 0.00010171254797486021, + "loss": 0.2169, + "step": 21102 + }, + { + "epoch": 0.49468755310981793, + "grad_norm": 0.169973686337471, + "learning_rate": 0.0001017051828926493, + "loss": 0.0354, + "step": 21103 + }, + { + "epoch": 0.49471099468462293, + "grad_norm": 0.30172064900398254, + "learning_rate": 0.00010169781780118604, + "loss": 0.0438, + "step": 21104 + }, + { + "epoch": 0.4947344362594279, + "grad_norm": 0.554142951965332, + "learning_rate": 0.00010169045270051032, + "loss": 0.6278, + "step": 21105 + }, + { + "epoch": 0.4947578778342329, + "grad_norm": 0.24081318080425262, + "learning_rate": 0.0001016830875906622, + "loss": 0.048, + "step": 21106 + }, + { + "epoch": 0.4947813194090379, + "grad_norm": 0.49112144112586975, + "learning_rate": 0.00010167572247168158, + "loss": 0.1157, + "step": 21107 + }, + { + "epoch": 0.4948047609838429, + "grad_norm": 0.7560669183731079, + "learning_rate": 0.00010166835734360843, + "loss": 0.0465, + "step": 21108 + }, + { + "epoch": 0.4948282025586479, + "grad_norm": 0.5480161905288696, + "learning_rate": 0.00010166099220648276, + "loss": 0.1466, + "step": 21109 + }, + { + "epoch": 0.4948516441334529, + "grad_norm": 0.5874426960945129, + "learning_rate": 0.0001016536270603445, + "loss": 0.0699, + "step": 21110 + }, + { + "epoch": 0.4948750857082579, + "grad_norm": 0.22411587834358215, + "learning_rate": 0.00010164626190523357, + "loss": 0.0304, + "step": 21111 + }, + { + "epoch": 0.4948985272830629, + "grad_norm": 0.27429330348968506, + "learning_rate": 0.00010163889674119, + "loss": 0.0736, + "step": 21112 + }, + { + "epoch": 0.4949219688578679, + "grad_norm": 0.44103166460990906, + "learning_rate": 0.00010163153156825373, + "loss": 0.0999, + "step": 21113 + }, + { + "epoch": 0.49494541043267287, + "grad_norm": 0.7998479604721069, + "learning_rate": 0.00010162416638646472, + "loss": 0.7665, + "step": 21114 + }, + { + "epoch": 0.49496885200747787, + "grad_norm": 0.4226112365722656, + "learning_rate": 0.00010161680119586295, + "loss": 0.0653, + "step": 21115 + }, + { + "epoch": 0.49499229358228286, + "grad_norm": 0.49969369173049927, + "learning_rate": 0.00010160943599648838, + "loss": 0.1377, + "step": 21116 + }, + { + "epoch": 0.49501573515708786, + "grad_norm": 0.6187413334846497, + "learning_rate": 0.00010160207078838093, + "loss": 0.1266, + "step": 21117 + }, + { + "epoch": 0.49503917673189285, + "grad_norm": 0.10374809801578522, + "learning_rate": 0.00010159470557158064, + "loss": 0.0126, + "step": 21118 + }, + { + "epoch": 0.49506261830669784, + "grad_norm": 0.5225189924240112, + "learning_rate": 0.00010158734034612741, + "loss": 0.1277, + "step": 21119 + }, + { + "epoch": 0.49508605988150284, + "grad_norm": 0.2006402462720871, + "learning_rate": 0.00010157997511206123, + "loss": 0.028, + "step": 21120 + }, + { + "epoch": 0.49510950145630783, + "grad_norm": 0.24818089604377747, + "learning_rate": 0.00010157260986942204, + "loss": 0.0417, + "step": 21121 + }, + { + "epoch": 0.4951329430311128, + "grad_norm": 0.17741377651691437, + "learning_rate": 0.00010156524461824986, + "loss": 0.0266, + "step": 21122 + }, + { + "epoch": 0.4951563846059178, + "grad_norm": 0.4809693396091461, + "learning_rate": 0.00010155787935858458, + "loss": 0.0545, + "step": 21123 + }, + { + "epoch": 0.4951798261807228, + "grad_norm": 0.5208175778388977, + "learning_rate": 0.00010155051409046623, + "loss": 0.0722, + "step": 21124 + }, + { + "epoch": 0.4952032677555278, + "grad_norm": 0.7805020213127136, + "learning_rate": 0.00010154314881393474, + "loss": 0.645, + "step": 21125 + }, + { + "epoch": 0.4952267093303328, + "grad_norm": 0.6614019870758057, + "learning_rate": 0.00010153578352903008, + "loss": 0.1278, + "step": 21126 + }, + { + "epoch": 0.4952501509051378, + "grad_norm": 0.47707539796829224, + "learning_rate": 0.00010152841823579223, + "loss": 0.0747, + "step": 21127 + }, + { + "epoch": 0.4952735924799428, + "grad_norm": 0.4849538803100586, + "learning_rate": 0.00010152105293426113, + "loss": 0.0681, + "step": 21128 + }, + { + "epoch": 0.4952970340547478, + "grad_norm": 0.2841537296772003, + "learning_rate": 0.00010151368762447676, + "loss": 0.0225, + "step": 21129 + }, + { + "epoch": 0.4953204756295528, + "grad_norm": 0.42735755443573, + "learning_rate": 0.00010150632230647905, + "loss": 0.0695, + "step": 21130 + }, + { + "epoch": 0.4953439172043578, + "grad_norm": 0.4460643529891968, + "learning_rate": 0.000101498956980308, + "loss": 0.0826, + "step": 21131 + }, + { + "epoch": 0.49536735877916277, + "grad_norm": 0.39191102981567383, + "learning_rate": 0.00010149159164600359, + "loss": 0.0876, + "step": 21132 + }, + { + "epoch": 0.49539080035396776, + "grad_norm": 0.5298686027526855, + "learning_rate": 0.00010148422630360574, + "loss": 0.0921, + "step": 21133 + }, + { + "epoch": 0.49541424192877276, + "grad_norm": 0.49464789032936096, + "learning_rate": 0.00010147686095315447, + "loss": 0.0688, + "step": 21134 + }, + { + "epoch": 0.49543768350357775, + "grad_norm": 0.5158405900001526, + "learning_rate": 0.0001014694955946897, + "loss": 0.0661, + "step": 21135 + }, + { + "epoch": 0.49546112507838275, + "grad_norm": 0.3790174424648285, + "learning_rate": 0.0001014621302282514, + "loss": 0.061, + "step": 21136 + }, + { + "epoch": 0.49548456665318774, + "grad_norm": 0.45856377482414246, + "learning_rate": 0.0001014547648538795, + "loss": 0.0827, + "step": 21137 + }, + { + "epoch": 0.49550800822799274, + "grad_norm": 0.7495695948600769, + "learning_rate": 0.00010144739947161405, + "loss": 0.1589, + "step": 21138 + }, + { + "epoch": 0.49553144980279773, + "grad_norm": 0.34242430329322815, + "learning_rate": 0.00010144003408149498, + "loss": 0.0711, + "step": 21139 + }, + { + "epoch": 0.4955548913776027, + "grad_norm": 0.6933420300483704, + "learning_rate": 0.0001014326686835622, + "loss": 0.1658, + "step": 21140 + }, + { + "epoch": 0.4955783329524077, + "grad_norm": 0.4753311574459076, + "learning_rate": 0.00010142530327785578, + "loss": 0.0746, + "step": 21141 + }, + { + "epoch": 0.4956017745272127, + "grad_norm": 1.166125774383545, + "learning_rate": 0.00010141793786441556, + "loss": 0.198, + "step": 21142 + }, + { + "epoch": 0.4956252161020177, + "grad_norm": 0.4697283208370209, + "learning_rate": 0.0001014105724432816, + "loss": 0.1011, + "step": 21143 + }, + { + "epoch": 0.4956486576768227, + "grad_norm": 0.7500843405723572, + "learning_rate": 0.00010140320701449384, + "loss": 0.1588, + "step": 21144 + }, + { + "epoch": 0.49567209925162775, + "grad_norm": 0.32664066553115845, + "learning_rate": 0.00010139584157809224, + "loss": 0.0763, + "step": 21145 + }, + { + "epoch": 0.49569554082643275, + "grad_norm": 0.19148457050323486, + "learning_rate": 0.00010138847613411675, + "loss": 0.0216, + "step": 21146 + }, + { + "epoch": 0.49571898240123774, + "grad_norm": 0.6308172345161438, + "learning_rate": 0.00010138111068260736, + "loss": 0.111, + "step": 21147 + }, + { + "epoch": 0.49574242397604273, + "grad_norm": 0.4357950687408447, + "learning_rate": 0.00010137374522360404, + "loss": 0.0628, + "step": 21148 + }, + { + "epoch": 0.49576586555084773, + "grad_norm": 0.8281998038291931, + "learning_rate": 0.00010136637975714669, + "loss": 0.1886, + "step": 21149 + }, + { + "epoch": 0.4957893071256527, + "grad_norm": 0.6543229222297668, + "learning_rate": 0.00010135901428327534, + "loss": 0.4971, + "step": 21150 + }, + { + "epoch": 0.4958127487004577, + "grad_norm": 0.29155874252319336, + "learning_rate": 0.00010135164880202996, + "loss": 0.0293, + "step": 21151 + }, + { + "epoch": 0.4958361902752627, + "grad_norm": 0.5223374366760254, + "learning_rate": 0.0001013442833134505, + "loss": 0.1306, + "step": 21152 + }, + { + "epoch": 0.4958596318500677, + "grad_norm": 0.47367650270462036, + "learning_rate": 0.0001013369178175769, + "loss": 0.4333, + "step": 21153 + }, + { + "epoch": 0.4958830734248727, + "grad_norm": 0.2985786497592926, + "learning_rate": 0.00010132955231444919, + "loss": 0.0754, + "step": 21154 + }, + { + "epoch": 0.4959065149996777, + "grad_norm": 0.18437843024730682, + "learning_rate": 0.00010132218680410723, + "loss": 0.0389, + "step": 21155 + }, + { + "epoch": 0.4959299565744827, + "grad_norm": 0.06822656840085983, + "learning_rate": 0.00010131482128659106, + "loss": 0.0122, + "step": 21156 + }, + { + "epoch": 0.4959533981492877, + "grad_norm": 0.3390793800354004, + "learning_rate": 0.00010130745576194068, + "loss": 0.448, + "step": 21157 + }, + { + "epoch": 0.4959768397240927, + "grad_norm": 0.49030745029449463, + "learning_rate": 0.00010130009023019598, + "loss": 0.7409, + "step": 21158 + }, + { + "epoch": 0.49600028129889767, + "grad_norm": 0.5734884738922119, + "learning_rate": 0.00010129272469139694, + "loss": 0.1285, + "step": 21159 + }, + { + "epoch": 0.49602372287370267, + "grad_norm": 0.21161368489265442, + "learning_rate": 0.00010128535914558357, + "loss": 0.0316, + "step": 21160 + }, + { + "epoch": 0.49604716444850766, + "grad_norm": 0.4165385961532593, + "learning_rate": 0.00010127799359279577, + "loss": 0.1238, + "step": 21161 + }, + { + "epoch": 0.49607060602331265, + "grad_norm": 0.42696303129196167, + "learning_rate": 0.00010127062803307354, + "loss": 0.0825, + "step": 21162 + }, + { + "epoch": 0.49609404759811765, + "grad_norm": 0.271982878446579, + "learning_rate": 0.00010126326246645688, + "loss": 0.0235, + "step": 21163 + }, + { + "epoch": 0.49611748917292264, + "grad_norm": 0.8170852661132812, + "learning_rate": 0.00010125589689298571, + "loss": 0.1123, + "step": 21164 + }, + { + "epoch": 0.49614093074772764, + "grad_norm": 0.760523796081543, + "learning_rate": 0.0001012485313127, + "loss": 0.1737, + "step": 21165 + }, + { + "epoch": 0.49616437232253263, + "grad_norm": 0.2986840307712555, + "learning_rate": 0.00010124116572563973, + "loss": 0.2873, + "step": 21166 + }, + { + "epoch": 0.4961878138973376, + "grad_norm": 0.1639566421508789, + "learning_rate": 0.00010123380013184487, + "loss": 0.0263, + "step": 21167 + }, + { + "epoch": 0.4962112554721426, + "grad_norm": 0.09357868880033493, + "learning_rate": 0.00010122643453135534, + "loss": 0.0128, + "step": 21168 + }, + { + "epoch": 0.4962346970469476, + "grad_norm": 0.3372979462146759, + "learning_rate": 0.00010121906892421118, + "loss": 0.0461, + "step": 21169 + }, + { + "epoch": 0.4962581386217526, + "grad_norm": 0.43745821714401245, + "learning_rate": 0.00010121170331045231, + "loss": 0.1117, + "step": 21170 + }, + { + "epoch": 0.4962815801965576, + "grad_norm": 0.47027766704559326, + "learning_rate": 0.00010120433769011868, + "loss": 0.0718, + "step": 21171 + }, + { + "epoch": 0.4963050217713626, + "grad_norm": 0.1486883908510208, + "learning_rate": 0.00010119697206325031, + "loss": 0.034, + "step": 21172 + }, + { + "epoch": 0.4963284633461676, + "grad_norm": 0.7159925103187561, + "learning_rate": 0.00010118960642988712, + "loss": 0.1037, + "step": 21173 + }, + { + "epoch": 0.4963519049209726, + "grad_norm": 0.7056784629821777, + "learning_rate": 0.0001011822407900691, + "loss": 0.1081, + "step": 21174 + }, + { + "epoch": 0.4963753464957776, + "grad_norm": 0.257468044757843, + "learning_rate": 0.00010117487514383619, + "loss": 0.0365, + "step": 21175 + }, + { + "epoch": 0.4963987880705826, + "grad_norm": 1.1080697774887085, + "learning_rate": 0.0001011675094912284, + "loss": 0.0723, + "step": 21176 + }, + { + "epoch": 0.49642222964538757, + "grad_norm": 0.6759201884269714, + "learning_rate": 0.00010116014383228567, + "loss": 0.0999, + "step": 21177 + }, + { + "epoch": 0.49644567122019256, + "grad_norm": 0.4243153929710388, + "learning_rate": 0.00010115277816704794, + "loss": 0.086, + "step": 21178 + }, + { + "epoch": 0.49646911279499756, + "grad_norm": 0.4493730664253235, + "learning_rate": 0.00010114541249555523, + "loss": 0.0399, + "step": 21179 + }, + { + "epoch": 0.49649255436980255, + "grad_norm": 0.20850449800491333, + "learning_rate": 0.00010113804681784744, + "loss": 0.0281, + "step": 21180 + }, + { + "epoch": 0.49651599594460755, + "grad_norm": 0.5225971937179565, + "learning_rate": 0.0001011306811339646, + "loss": 0.1017, + "step": 21181 + }, + { + "epoch": 0.49653943751941254, + "grad_norm": 0.4343242049217224, + "learning_rate": 0.00010112331544394667, + "loss": 0.0747, + "step": 21182 + }, + { + "epoch": 0.49656287909421754, + "grad_norm": 1.0823702812194824, + "learning_rate": 0.00010111594974783361, + "loss": 0.2099, + "step": 21183 + }, + { + "epoch": 0.49658632066902253, + "grad_norm": 0.5657541751861572, + "learning_rate": 0.00010110858404566533, + "loss": 0.6219, + "step": 21184 + }, + { + "epoch": 0.4966097622438275, + "grad_norm": 0.21393832564353943, + "learning_rate": 0.00010110121833748188, + "loss": 0.0215, + "step": 21185 + }, + { + "epoch": 0.4966332038186325, + "grad_norm": 0.3505036532878876, + "learning_rate": 0.00010109385262332314, + "loss": 0.0808, + "step": 21186 + }, + { + "epoch": 0.4966566453934375, + "grad_norm": 0.1155078113079071, + "learning_rate": 0.00010108648690322915, + "loss": 0.0114, + "step": 21187 + }, + { + "epoch": 0.4966800869682425, + "grad_norm": 1.457289218902588, + "learning_rate": 0.00010107912117723987, + "loss": 0.1209, + "step": 21188 + }, + { + "epoch": 0.4967035285430475, + "grad_norm": 0.6979296207427979, + "learning_rate": 0.00010107175544539524, + "loss": 0.1312, + "step": 21189 + }, + { + "epoch": 0.4967269701178525, + "grad_norm": 0.41256770491600037, + "learning_rate": 0.00010106438970773522, + "loss": 0.0749, + "step": 21190 + }, + { + "epoch": 0.4967504116926575, + "grad_norm": 0.5348238348960876, + "learning_rate": 0.00010105702396429981, + "loss": 0.1192, + "step": 21191 + }, + { + "epoch": 0.4967738532674625, + "grad_norm": 0.6577622294425964, + "learning_rate": 0.00010104965821512895, + "loss": 0.1455, + "step": 21192 + }, + { + "epoch": 0.4967972948422675, + "grad_norm": 0.1603386104106903, + "learning_rate": 0.00010104229246026262, + "loss": 0.0337, + "step": 21193 + }, + { + "epoch": 0.4968207364170725, + "grad_norm": 0.70279860496521, + "learning_rate": 0.00010103492669974077, + "loss": 0.4898, + "step": 21194 + }, + { + "epoch": 0.49684417799187747, + "grad_norm": 0.3423418700695038, + "learning_rate": 0.00010102756093360339, + "loss": 0.0499, + "step": 21195 + }, + { + "epoch": 0.4968676195666825, + "grad_norm": 0.6755719780921936, + "learning_rate": 0.00010102019516189042, + "loss": 0.1308, + "step": 21196 + }, + { + "epoch": 0.4968910611414875, + "grad_norm": 0.2975069284439087, + "learning_rate": 0.00010101282938464187, + "loss": 0.0538, + "step": 21197 + }, + { + "epoch": 0.4969145027162925, + "grad_norm": 0.3935098946094513, + "learning_rate": 0.00010100546360189765, + "loss": 0.0838, + "step": 21198 + }, + { + "epoch": 0.4969379442910975, + "grad_norm": 0.3899317979812622, + "learning_rate": 0.00010099809781369777, + "loss": 0.0778, + "step": 21199 + }, + { + "epoch": 0.4969613858659025, + "grad_norm": 0.484941303730011, + "learning_rate": 0.00010099073202008218, + "loss": 0.6376, + "step": 21200 + }, + { + "epoch": 0.4969848274407075, + "grad_norm": 0.2593361735343933, + "learning_rate": 0.00010098336622109085, + "loss": 0.0406, + "step": 21201 + }, + { + "epoch": 0.4970082690155125, + "grad_norm": 0.3341239094734192, + "learning_rate": 0.00010097600041676375, + "loss": 0.0928, + "step": 21202 + }, + { + "epoch": 0.4970317105903175, + "grad_norm": 0.37289440631866455, + "learning_rate": 0.00010096863460714086, + "loss": 0.0411, + "step": 21203 + }, + { + "epoch": 0.49705515216512247, + "grad_norm": 0.32042360305786133, + "learning_rate": 0.00010096126879226209, + "loss": 0.0899, + "step": 21204 + }, + { + "epoch": 0.49707859373992747, + "grad_norm": 0.9098353385925293, + "learning_rate": 0.00010095390297216749, + "loss": 0.1049, + "step": 21205 + }, + { + "epoch": 0.49710203531473246, + "grad_norm": 0.19439779222011566, + "learning_rate": 0.00010094653714689694, + "loss": 0.0375, + "step": 21206 + }, + { + "epoch": 0.49712547688953745, + "grad_norm": 1.0029807090759277, + "learning_rate": 0.00010093917131649051, + "loss": 0.096, + "step": 21207 + }, + { + "epoch": 0.49714891846434245, + "grad_norm": 0.8207777142524719, + "learning_rate": 0.00010093180548098809, + "loss": 0.116, + "step": 21208 + }, + { + "epoch": 0.49717236003914744, + "grad_norm": 0.44861018657684326, + "learning_rate": 0.00010092443964042964, + "loss": 0.0446, + "step": 21209 + }, + { + "epoch": 0.49719580161395244, + "grad_norm": 0.11502167582511902, + "learning_rate": 0.00010091707379485515, + "loss": 0.0283, + "step": 21210 + }, + { + "epoch": 0.49721924318875743, + "grad_norm": 0.5856948494911194, + "learning_rate": 0.00010090970794430464, + "loss": 0.1357, + "step": 21211 + }, + { + "epoch": 0.4972426847635624, + "grad_norm": 0.15422475337982178, + "learning_rate": 0.00010090234208881801, + "loss": 0.0214, + "step": 21212 + }, + { + "epoch": 0.4972661263383674, + "grad_norm": 0.2609299421310425, + "learning_rate": 0.00010089497622843522, + "loss": 0.0567, + "step": 21213 + }, + { + "epoch": 0.4972895679131724, + "grad_norm": 0.5555539131164551, + "learning_rate": 0.00010088761036319632, + "loss": 0.1625, + "step": 21214 + }, + { + "epoch": 0.4973130094879774, + "grad_norm": 0.24970203638076782, + "learning_rate": 0.0001008802444931412, + "loss": 0.0483, + "step": 21215 + }, + { + "epoch": 0.4973364510627824, + "grad_norm": 0.3324066400527954, + "learning_rate": 0.00010087287861830979, + "loss": 0.07, + "step": 21216 + }, + { + "epoch": 0.4973598926375874, + "grad_norm": 0.7574015259742737, + "learning_rate": 0.00010086551273874216, + "loss": 0.6659, + "step": 21217 + }, + { + "epoch": 0.4973833342123924, + "grad_norm": 0.5266520977020264, + "learning_rate": 0.00010085814685447825, + "loss": 0.0705, + "step": 21218 + }, + { + "epoch": 0.4974067757871974, + "grad_norm": 0.5639177560806274, + "learning_rate": 0.00010085078096555798, + "loss": 0.0567, + "step": 21219 + }, + { + "epoch": 0.4974302173620024, + "grad_norm": 0.5840886235237122, + "learning_rate": 0.00010084341507202137, + "loss": 0.1324, + "step": 21220 + }, + { + "epoch": 0.4974536589368074, + "grad_norm": 1.0535719394683838, + "learning_rate": 0.00010083604917390839, + "loss": 0.1546, + "step": 21221 + }, + { + "epoch": 0.49747710051161237, + "grad_norm": 1.0160596370697021, + "learning_rate": 0.00010082868327125894, + "loss": 0.126, + "step": 21222 + }, + { + "epoch": 0.49750054208641736, + "grad_norm": 0.3006971776485443, + "learning_rate": 0.00010082131736411302, + "loss": 0.0452, + "step": 21223 + }, + { + "epoch": 0.49752398366122236, + "grad_norm": 0.13933061063289642, + "learning_rate": 0.00010081395145251065, + "loss": 0.0335, + "step": 21224 + }, + { + "epoch": 0.49754742523602735, + "grad_norm": 0.4901881217956543, + "learning_rate": 0.00010080658553649174, + "loss": 0.12, + "step": 21225 + }, + { + "epoch": 0.49757086681083235, + "grad_norm": 0.4730667471885681, + "learning_rate": 0.00010079921961609629, + "loss": 0.0893, + "step": 21226 + }, + { + "epoch": 0.49759430838563734, + "grad_norm": 0.14803631603717804, + "learning_rate": 0.00010079185369136424, + "loss": 0.0256, + "step": 21227 + }, + { + "epoch": 0.49761774996044233, + "grad_norm": 0.2694111466407776, + "learning_rate": 0.00010078448776233556, + "loss": 0.0301, + "step": 21228 + }, + { + "epoch": 0.49764119153524733, + "grad_norm": 0.41871851682662964, + "learning_rate": 0.00010077712182905021, + "loss": 0.0729, + "step": 21229 + }, + { + "epoch": 0.4976646331100523, + "grad_norm": 0.4248657822608948, + "learning_rate": 0.0001007697558915482, + "loss": 0.0907, + "step": 21230 + }, + { + "epoch": 0.4976880746848573, + "grad_norm": 0.1929493248462677, + "learning_rate": 0.00010076238994986948, + "loss": 0.0477, + "step": 21231 + }, + { + "epoch": 0.4977115162596623, + "grad_norm": 0.388908326625824, + "learning_rate": 0.00010075502400405399, + "loss": 0.0782, + "step": 21232 + }, + { + "epoch": 0.4977349578344673, + "grad_norm": 0.5486009120941162, + "learning_rate": 0.00010074765805414175, + "loss": 0.0605, + "step": 21233 + }, + { + "epoch": 0.4977583994092723, + "grad_norm": 0.5715928673744202, + "learning_rate": 0.00010074029210017268, + "loss": 0.1009, + "step": 21234 + }, + { + "epoch": 0.4977818409840773, + "grad_norm": 0.32814356684684753, + "learning_rate": 0.00010073292614218674, + "loss": 0.0343, + "step": 21235 + }, + { + "epoch": 0.4978052825588823, + "grad_norm": 0.533097505569458, + "learning_rate": 0.00010072556018022394, + "loss": 0.0911, + "step": 21236 + }, + { + "epoch": 0.4978287241336873, + "grad_norm": 0.2191225290298462, + "learning_rate": 0.00010071819421432425, + "loss": 0.0399, + "step": 21237 + }, + { + "epoch": 0.4978521657084923, + "grad_norm": 0.36431044340133667, + "learning_rate": 0.0001007108282445276, + "loss": 0.0475, + "step": 21238 + }, + { + "epoch": 0.4978756072832973, + "grad_norm": 0.5550832748413086, + "learning_rate": 0.00010070346227087397, + "loss": 0.1339, + "step": 21239 + }, + { + "epoch": 0.49789904885810227, + "grad_norm": 0.5366477370262146, + "learning_rate": 0.00010069609629340336, + "loss": 0.1749, + "step": 21240 + }, + { + "epoch": 0.49792249043290726, + "grad_norm": 0.7346373200416565, + "learning_rate": 0.00010068873031215566, + "loss": 0.1927, + "step": 21241 + }, + { + "epoch": 0.49794593200771226, + "grad_norm": 0.46460267901420593, + "learning_rate": 0.00010068136432717095, + "loss": 0.059, + "step": 21242 + }, + { + "epoch": 0.49796937358251725, + "grad_norm": 0.3939318358898163, + "learning_rate": 0.00010067399833848911, + "loss": 0.0894, + "step": 21243 + }, + { + "epoch": 0.49799281515732224, + "grad_norm": 0.5479933619499207, + "learning_rate": 0.00010066663234615012, + "loss": 0.1317, + "step": 21244 + }, + { + "epoch": 0.49801625673212724, + "grad_norm": 0.17002242803573608, + "learning_rate": 0.000100659266350194, + "loss": 0.0332, + "step": 21245 + }, + { + "epoch": 0.49803969830693223, + "grad_norm": 0.19952169060707092, + "learning_rate": 0.00010065190035066068, + "loss": 0.0305, + "step": 21246 + }, + { + "epoch": 0.4980631398817373, + "grad_norm": 0.47908711433410645, + "learning_rate": 0.0001006445343475901, + "loss": 0.1092, + "step": 21247 + }, + { + "epoch": 0.4980865814565423, + "grad_norm": 0.5972097516059875, + "learning_rate": 0.00010063716834102225, + "loss": 0.1324, + "step": 21248 + }, + { + "epoch": 0.49811002303134727, + "grad_norm": 0.5834901928901672, + "learning_rate": 0.00010062980233099715, + "loss": 0.1036, + "step": 21249 + }, + { + "epoch": 0.49813346460615227, + "grad_norm": 0.3228580057621002, + "learning_rate": 0.0001006224363175547, + "loss": 0.0572, + "step": 21250 + }, + { + "epoch": 0.49815690618095726, + "grad_norm": 0.8087433576583862, + "learning_rate": 0.00010061507030073489, + "loss": 0.153, + "step": 21251 + }, + { + "epoch": 0.49818034775576225, + "grad_norm": 0.4940379559993744, + "learning_rate": 0.00010060770428057771, + "loss": 0.4785, + "step": 21252 + }, + { + "epoch": 0.49820378933056725, + "grad_norm": 0.31915247440338135, + "learning_rate": 0.0001006003382571231, + "loss": 0.0945, + "step": 21253 + }, + { + "epoch": 0.49822723090537224, + "grad_norm": 0.3433857858181, + "learning_rate": 0.00010059297223041102, + "loss": 0.0957, + "step": 21254 + }, + { + "epoch": 0.49825067248017724, + "grad_norm": 0.4340677559375763, + "learning_rate": 0.00010058560620048147, + "loss": 0.0727, + "step": 21255 + }, + { + "epoch": 0.49827411405498223, + "grad_norm": 0.1565234512090683, + "learning_rate": 0.00010057824016737443, + "loss": 0.0266, + "step": 21256 + }, + { + "epoch": 0.4982975556297872, + "grad_norm": 0.25759372115135193, + "learning_rate": 0.0001005708741311298, + "loss": 0.0467, + "step": 21257 + }, + { + "epoch": 0.4983209972045922, + "grad_norm": 0.18197283148765564, + "learning_rate": 0.0001005635080917876, + "loss": 0.0263, + "step": 21258 + }, + { + "epoch": 0.4983444387793972, + "grad_norm": 0.40240129828453064, + "learning_rate": 0.00010055614204938781, + "loss": 0.0614, + "step": 21259 + }, + { + "epoch": 0.4983678803542022, + "grad_norm": 0.47130286693573, + "learning_rate": 0.00010054877600397035, + "loss": 0.0573, + "step": 21260 + }, + { + "epoch": 0.4983913219290072, + "grad_norm": 0.5261305570602417, + "learning_rate": 0.00010054140995557524, + "loss": 0.2099, + "step": 21261 + }, + { + "epoch": 0.4984147635038122, + "grad_norm": 0.4320019483566284, + "learning_rate": 0.00010053404390424243, + "loss": 0.0948, + "step": 21262 + }, + { + "epoch": 0.4984382050786172, + "grad_norm": 0.34063899517059326, + "learning_rate": 0.00010052667785001186, + "loss": 0.0487, + "step": 21263 + }, + { + "epoch": 0.4984616466534222, + "grad_norm": 0.20138780772686005, + "learning_rate": 0.00010051931179292353, + "loss": 0.0491, + "step": 21264 + }, + { + "epoch": 0.4984850882282272, + "grad_norm": 0.6852312088012695, + "learning_rate": 0.00010051194573301738, + "loss": 0.0992, + "step": 21265 + }, + { + "epoch": 0.4985085298030322, + "grad_norm": 0.8778024315834045, + "learning_rate": 0.00010050457967033345, + "loss": 0.0924, + "step": 21266 + }, + { + "epoch": 0.49853197137783717, + "grad_norm": 0.36275652050971985, + "learning_rate": 0.0001004972136049116, + "loss": 0.0756, + "step": 21267 + }, + { + "epoch": 0.49855541295264216, + "grad_norm": 0.5167210698127747, + "learning_rate": 0.0001004898475367919, + "loss": 0.0752, + "step": 21268 + }, + { + "epoch": 0.49857885452744716, + "grad_norm": 0.558165967464447, + "learning_rate": 0.00010048248146601426, + "loss": 0.087, + "step": 21269 + }, + { + "epoch": 0.49860229610225215, + "grad_norm": 0.5214546322822571, + "learning_rate": 0.00010047511539261864, + "loss": 0.1229, + "step": 21270 + }, + { + "epoch": 0.49862573767705715, + "grad_norm": 0.28047889471054077, + "learning_rate": 0.00010046774931664504, + "loss": 0.0536, + "step": 21271 + }, + { + "epoch": 0.49864917925186214, + "grad_norm": 0.21844986081123352, + "learning_rate": 0.00010046038323813343, + "loss": 0.0421, + "step": 21272 + }, + { + "epoch": 0.49867262082666713, + "grad_norm": 0.5246370434761047, + "learning_rate": 0.00010045301715712377, + "loss": 0.1185, + "step": 21273 + }, + { + "epoch": 0.49869606240147213, + "grad_norm": 0.1725511997938156, + "learning_rate": 0.00010044565107365603, + "loss": 0.024, + "step": 21274 + }, + { + "epoch": 0.4987195039762771, + "grad_norm": 0.11594903469085693, + "learning_rate": 0.00010043828498777015, + "loss": 0.026, + "step": 21275 + }, + { + "epoch": 0.4987429455510821, + "grad_norm": 0.45909932255744934, + "learning_rate": 0.00010043091889950613, + "loss": 0.043, + "step": 21276 + }, + { + "epoch": 0.4987663871258871, + "grad_norm": 1.0970115661621094, + "learning_rate": 0.00010042355280890392, + "loss": 0.0627, + "step": 21277 + }, + { + "epoch": 0.4987898287006921, + "grad_norm": 0.6818090081214905, + "learning_rate": 0.00010041618671600355, + "loss": 0.1157, + "step": 21278 + }, + { + "epoch": 0.4988132702754971, + "grad_norm": 0.07403623312711716, + "learning_rate": 0.0001004088206208449, + "loss": 0.015, + "step": 21279 + }, + { + "epoch": 0.4988367118503021, + "grad_norm": 0.5470428466796875, + "learning_rate": 0.00010040145452346798, + "loss": 0.0914, + "step": 21280 + }, + { + "epoch": 0.4988601534251071, + "grad_norm": 0.38641422986984253, + "learning_rate": 0.00010039408842391278, + "loss": 0.0745, + "step": 21281 + }, + { + "epoch": 0.4988835949999121, + "grad_norm": 0.40442416071891785, + "learning_rate": 0.00010038672232221922, + "loss": 0.0623, + "step": 21282 + }, + { + "epoch": 0.4989070365747171, + "grad_norm": 0.2604157626628876, + "learning_rate": 0.0001003793562184273, + "loss": 0.0566, + "step": 21283 + }, + { + "epoch": 0.49893047814952207, + "grad_norm": 0.46299147605895996, + "learning_rate": 0.00010037199011257699, + "loss": 0.065, + "step": 21284 + }, + { + "epoch": 0.49895391972432707, + "grad_norm": 0.20587657392024994, + "learning_rate": 0.00010036462400470825, + "loss": 0.0407, + "step": 21285 + }, + { + "epoch": 0.49897736129913206, + "grad_norm": 0.6811830401420593, + "learning_rate": 0.00010035725789486104, + "loss": 0.1065, + "step": 21286 + }, + { + "epoch": 0.49900080287393705, + "grad_norm": 0.4401889741420746, + "learning_rate": 0.00010034989178307535, + "loss": 0.0636, + "step": 21287 + }, + { + "epoch": 0.49902424444874205, + "grad_norm": 0.8494977951049805, + "learning_rate": 0.00010034252566939114, + "loss": 0.0872, + "step": 21288 + }, + { + "epoch": 0.49904768602354704, + "grad_norm": 0.517641007900238, + "learning_rate": 0.00010033515955384833, + "loss": 0.1163, + "step": 21289 + }, + { + "epoch": 0.49907112759835204, + "grad_norm": 0.355362206697464, + "learning_rate": 0.00010032779343648697, + "loss": 0.3763, + "step": 21290 + }, + { + "epoch": 0.49909456917315703, + "grad_norm": 0.2922329604625702, + "learning_rate": 0.00010032042731734701, + "loss": 0.0279, + "step": 21291 + }, + { + "epoch": 0.499118010747962, + "grad_norm": 0.3784184157848358, + "learning_rate": 0.00010031306119646837, + "loss": 0.2477, + "step": 21292 + }, + { + "epoch": 0.499141452322767, + "grad_norm": 0.588246762752533, + "learning_rate": 0.0001003056950738911, + "loss": 0.5568, + "step": 21293 + }, + { + "epoch": 0.499164893897572, + "grad_norm": 0.4045722782611847, + "learning_rate": 0.0001002983289496551, + "loss": 0.0454, + "step": 21294 + }, + { + "epoch": 0.499188335472377, + "grad_norm": 0.5148916840553284, + "learning_rate": 0.00010029096282380033, + "loss": 0.1523, + "step": 21295 + }, + { + "epoch": 0.499211777047182, + "grad_norm": 0.43388864398002625, + "learning_rate": 0.00010028359669636677, + "loss": 0.6254, + "step": 21296 + }, + { + "epoch": 0.499235218621987, + "grad_norm": 0.5735114812850952, + "learning_rate": 0.00010027623056739448, + "loss": 0.613, + "step": 21297 + }, + { + "epoch": 0.49925866019679205, + "grad_norm": 0.7810203433036804, + "learning_rate": 0.0001002688644369233, + "loss": 0.0936, + "step": 21298 + }, + { + "epoch": 0.49928210177159704, + "grad_norm": 0.42182087898254395, + "learning_rate": 0.00010026149830499328, + "loss": 0.0816, + "step": 21299 + }, + { + "epoch": 0.49930554334640204, + "grad_norm": 0.25911277532577515, + "learning_rate": 0.00010025413217164436, + "loss": 0.0386, + "step": 21300 + }, + { + "epoch": 0.49932898492120703, + "grad_norm": 0.19275610148906708, + "learning_rate": 0.00010024676603691649, + "loss": 0.0442, + "step": 21301 + }, + { + "epoch": 0.499352426496012, + "grad_norm": 0.44593149423599243, + "learning_rate": 0.00010023939990084967, + "loss": 0.0835, + "step": 21302 + }, + { + "epoch": 0.499375868070817, + "grad_norm": 0.33913034200668335, + "learning_rate": 0.00010023203376348387, + "loss": 0.0671, + "step": 21303 + }, + { + "epoch": 0.499399309645622, + "grad_norm": 0.3970039188861847, + "learning_rate": 0.00010022466762485905, + "loss": 0.0988, + "step": 21304 + }, + { + "epoch": 0.499422751220427, + "grad_norm": 0.4804578125476837, + "learning_rate": 0.00010021730148501515, + "loss": 0.1218, + "step": 21305 + }, + { + "epoch": 0.499446192795232, + "grad_norm": 0.9809905290603638, + "learning_rate": 0.00010020993534399222, + "loss": 0.1391, + "step": 21306 + }, + { + "epoch": 0.499469634370037, + "grad_norm": 0.23123770952224731, + "learning_rate": 0.00010020256920183016, + "loss": 0.0352, + "step": 21307 + }, + { + "epoch": 0.499493075944842, + "grad_norm": 0.4422304034233093, + "learning_rate": 0.00010019520305856889, + "loss": 0.052, + "step": 21308 + }, + { + "epoch": 0.499516517519647, + "grad_norm": 0.29256123304367065, + "learning_rate": 0.0001001878369142485, + "loss": 0.0687, + "step": 21309 + }, + { + "epoch": 0.499539959094452, + "grad_norm": 0.7467201948165894, + "learning_rate": 0.0001001804707689089, + "loss": 0.1141, + "step": 21310 + }, + { + "epoch": 0.499563400669257, + "grad_norm": 0.17054998874664307, + "learning_rate": 0.00010017310462259004, + "loss": 0.0415, + "step": 21311 + }, + { + "epoch": 0.49958684224406197, + "grad_norm": 0.5814245939254761, + "learning_rate": 0.00010016573847533192, + "loss": 0.0989, + "step": 21312 + }, + { + "epoch": 0.49961028381886696, + "grad_norm": 0.9023683071136475, + "learning_rate": 0.00010015837232717452, + "loss": 0.1723, + "step": 21313 + }, + { + "epoch": 0.49963372539367196, + "grad_norm": 0.43799638748168945, + "learning_rate": 0.00010015100617815775, + "loss": 0.0405, + "step": 21314 + }, + { + "epoch": 0.49965716696847695, + "grad_norm": 0.626358687877655, + "learning_rate": 0.00010014364002832162, + "loss": 0.1178, + "step": 21315 + }, + { + "epoch": 0.49968060854328195, + "grad_norm": 0.5270189046859741, + "learning_rate": 0.00010013627387770612, + "loss": 0.0965, + "step": 21316 + }, + { + "epoch": 0.49970405011808694, + "grad_norm": 0.43551498651504517, + "learning_rate": 0.00010012890772635117, + "loss": 0.1025, + "step": 21317 + }, + { + "epoch": 0.49972749169289193, + "grad_norm": 0.3342578709125519, + "learning_rate": 0.00010012154157429679, + "loss": 0.0774, + "step": 21318 + }, + { + "epoch": 0.49975093326769693, + "grad_norm": 0.24200314283370972, + "learning_rate": 0.0001001141754215829, + "loss": 0.0546, + "step": 21319 + }, + { + "epoch": 0.4997743748425019, + "grad_norm": 0.4702569544315338, + "learning_rate": 0.00010010680926824949, + "loss": 0.1593, + "step": 21320 + }, + { + "epoch": 0.4997978164173069, + "grad_norm": 0.1947007179260254, + "learning_rate": 0.00010009944311433652, + "loss": 0.0514, + "step": 21321 + }, + { + "epoch": 0.4998212579921119, + "grad_norm": 0.7269644737243652, + "learning_rate": 0.00010009207695988401, + "loss": 0.1822, + "step": 21322 + }, + { + "epoch": 0.4998446995669169, + "grad_norm": 0.5849777460098267, + "learning_rate": 0.00010008471080493187, + "loss": 0.164, + "step": 21323 + }, + { + "epoch": 0.4998681411417219, + "grad_norm": 0.094455786049366, + "learning_rate": 0.00010007734464952008, + "loss": 0.0169, + "step": 21324 + }, + { + "epoch": 0.4998915827165269, + "grad_norm": 0.7821498513221741, + "learning_rate": 0.00010006997849368862, + "loss": 0.0901, + "step": 21325 + }, + { + "epoch": 0.4999150242913319, + "grad_norm": 0.40723705291748047, + "learning_rate": 0.00010006261233747743, + "loss": 0.0292, + "step": 21326 + }, + { + "epoch": 0.4999384658661369, + "grad_norm": 0.22586536407470703, + "learning_rate": 0.00010005524618092652, + "loss": 0.0745, + "step": 21327 + }, + { + "epoch": 0.4999619074409419, + "grad_norm": 0.20473413169384003, + "learning_rate": 0.00010004788002407585, + "loss": 0.0547, + "step": 21328 + }, + { + "epoch": 0.49998534901574687, + "grad_norm": 0.32725879549980164, + "learning_rate": 0.0001000405138669654, + "loss": 0.0323, + "step": 21329 + }, + { + "epoch": 0.5000087905905519, + "grad_norm": 0.4166288673877716, + "learning_rate": 0.00010003314770963507, + "loss": 0.1101, + "step": 21330 + }, + { + "epoch": 0.5000322321653569, + "grad_norm": 0.2605593502521515, + "learning_rate": 0.00010002578155212492, + "loss": 0.0555, + "step": 21331 + }, + { + "epoch": 0.5000556737401619, + "grad_norm": 0.32736533880233765, + "learning_rate": 0.00010001841539447487, + "loss": 0.057, + "step": 21332 + }, + { + "epoch": 0.5000791153149668, + "grad_norm": 0.09388776868581772, + "learning_rate": 0.0001000110492367249, + "loss": 0.0092, + "step": 21333 + }, + { + "epoch": 0.5001025568897719, + "grad_norm": 0.33221569657325745, + "learning_rate": 0.00010000368307891496, + "loss": 0.0625, + "step": 21334 + }, + { + "epoch": 0.5001259984645768, + "grad_norm": 0.38412487506866455, + "learning_rate": 9.999631692108506e-05, + "loss": 0.4261, + "step": 21335 + }, + { + "epoch": 0.5001494400393819, + "grad_norm": 0.10928503423929214, + "learning_rate": 9.998895076327512e-05, + "loss": 0.0254, + "step": 21336 + }, + { + "epoch": 0.5001728816141868, + "grad_norm": 0.28890764713287354, + "learning_rate": 9.998158460552515e-05, + "loss": 0.0841, + "step": 21337 + }, + { + "epoch": 0.5001963231889919, + "grad_norm": 0.36560365557670593, + "learning_rate": 9.99742184478751e-05, + "loss": 0.0609, + "step": 21338 + }, + { + "epoch": 0.5002197647637968, + "grad_norm": 0.14729857444763184, + "learning_rate": 9.996685229036495e-05, + "loss": 0.0254, + "step": 21339 + }, + { + "epoch": 0.5002432063386019, + "grad_norm": 0.6105349063873291, + "learning_rate": 9.995948613303465e-05, + "loss": 0.1449, + "step": 21340 + }, + { + "epoch": 0.5002666479134068, + "grad_norm": 0.42766180634498596, + "learning_rate": 9.995211997592414e-05, + "loss": 0.1023, + "step": 21341 + }, + { + "epoch": 0.5002900894882119, + "grad_norm": 0.6426076889038086, + "learning_rate": 9.994475381907349e-05, + "loss": 0.0738, + "step": 21342 + }, + { + "epoch": 0.5003135310630168, + "grad_norm": 0.430464506149292, + "learning_rate": 9.993738766252257e-05, + "loss": 0.0857, + "step": 21343 + }, + { + "epoch": 0.5003369726378218, + "grad_norm": 0.3430795967578888, + "learning_rate": 9.993002150631141e-05, + "loss": 0.0833, + "step": 21344 + }, + { + "epoch": 0.5003604142126268, + "grad_norm": 0.34144189953804016, + "learning_rate": 9.992265535047995e-05, + "loss": 0.0866, + "step": 21345 + }, + { + "epoch": 0.5003838557874318, + "grad_norm": 0.674159824848175, + "learning_rate": 9.991528919506816e-05, + "loss": 0.1503, + "step": 21346 + }, + { + "epoch": 0.5004072973622368, + "grad_norm": 0.5031772255897522, + "learning_rate": 9.990792304011601e-05, + "loss": 0.6611, + "step": 21347 + }, + { + "epoch": 0.5004307389370418, + "grad_norm": 0.516179621219635, + "learning_rate": 9.99005568856635e-05, + "loss": 0.1493, + "step": 21348 + }, + { + "epoch": 0.5004541805118468, + "grad_norm": 0.8184710741043091, + "learning_rate": 9.989319073175053e-05, + "loss": 0.1765, + "step": 21349 + }, + { + "epoch": 0.5004776220866518, + "grad_norm": 0.31455880403518677, + "learning_rate": 9.988582457841714e-05, + "loss": 0.0666, + "step": 21350 + }, + { + "epoch": 0.5005010636614567, + "grad_norm": 0.8079864382743835, + "learning_rate": 9.987845842570326e-05, + "loss": 0.1472, + "step": 21351 + }, + { + "epoch": 0.5005245052362618, + "grad_norm": 0.2702834904193878, + "learning_rate": 9.987109227364885e-05, + "loss": 0.0747, + "step": 21352 + }, + { + "epoch": 0.5005479468110667, + "grad_norm": 0.6538974642753601, + "learning_rate": 9.986372612229389e-05, + "loss": 0.1089, + "step": 21353 + }, + { + "epoch": 0.5005713883858718, + "grad_norm": 0.4336119592189789, + "learning_rate": 9.985635997167837e-05, + "loss": 0.081, + "step": 21354 + }, + { + "epoch": 0.5005948299606767, + "grad_norm": 0.7820320129394531, + "learning_rate": 9.984899382184227e-05, + "loss": 0.1468, + "step": 21355 + }, + { + "epoch": 0.5006182715354818, + "grad_norm": 0.47971564531326294, + "learning_rate": 9.984162767282551e-05, + "loss": 0.3503, + "step": 21356 + }, + { + "epoch": 0.5006417131102867, + "grad_norm": 0.08871588855981827, + "learning_rate": 9.98342615246681e-05, + "loss": 0.0051, + "step": 21357 + }, + { + "epoch": 0.5006651546850918, + "grad_norm": 0.46086427569389343, + "learning_rate": 9.982689537741e-05, + "loss": 0.1677, + "step": 21358 + }, + { + "epoch": 0.5006885962598967, + "grad_norm": 0.14857618510723114, + "learning_rate": 9.981952923109112e-05, + "loss": 0.0152, + "step": 21359 + }, + { + "epoch": 0.5007120378347018, + "grad_norm": 0.3514200747013092, + "learning_rate": 9.981216308575151e-05, + "loss": 0.0535, + "step": 21360 + }, + { + "epoch": 0.5007354794095067, + "grad_norm": 0.3710920512676239, + "learning_rate": 9.980479694143114e-05, + "loss": 0.0629, + "step": 21361 + }, + { + "epoch": 0.5007589209843117, + "grad_norm": 0.3326025903224945, + "learning_rate": 9.97974307981699e-05, + "loss": 0.1137, + "step": 21362 + }, + { + "epoch": 0.5007823625591167, + "grad_norm": 0.3916105031967163, + "learning_rate": 9.979006465600783e-05, + "loss": 0.057, + "step": 21363 + }, + { + "epoch": 0.5008058041339217, + "grad_norm": 0.5334809422492981, + "learning_rate": 9.978269851498487e-05, + "loss": 0.4939, + "step": 21364 + }, + { + "epoch": 0.5008292457087267, + "grad_norm": 0.48604294657707214, + "learning_rate": 9.977533237514096e-05, + "loss": 0.077, + "step": 21365 + }, + { + "epoch": 0.5008526872835317, + "grad_norm": 0.3755936622619629, + "learning_rate": 9.976796623651613e-05, + "loss": 0.0877, + "step": 21366 + }, + { + "epoch": 0.5008761288583367, + "grad_norm": 0.5409164428710938, + "learning_rate": 9.976060009915034e-05, + "loss": 0.0982, + "step": 21367 + }, + { + "epoch": 0.5008995704331417, + "grad_norm": 0.6073734760284424, + "learning_rate": 9.975323396308353e-05, + "loss": 0.0333, + "step": 21368 + }, + { + "epoch": 0.5009230120079466, + "grad_norm": 0.3547095060348511, + "learning_rate": 9.974586782835566e-05, + "loss": 0.0608, + "step": 21369 + }, + { + "epoch": 0.5009464535827517, + "grad_norm": 0.4375295639038086, + "learning_rate": 9.973850169500675e-05, + "loss": 0.097, + "step": 21370 + }, + { + "epoch": 0.5009698951575566, + "grad_norm": 0.15955716371536255, + "learning_rate": 9.973113556307673e-05, + "loss": 0.0218, + "step": 21371 + }, + { + "epoch": 0.5009933367323617, + "grad_norm": 0.4900253713130951, + "learning_rate": 9.972376943260555e-05, + "loss": 0.138, + "step": 21372 + }, + { + "epoch": 0.5010167783071666, + "grad_norm": 0.3294816017150879, + "learning_rate": 9.971640330363324e-05, + "loss": 0.0522, + "step": 21373 + }, + { + "epoch": 0.5010402198819717, + "grad_norm": 0.35139796137809753, + "learning_rate": 9.970903717619969e-05, + "loss": 0.056, + "step": 21374 + }, + { + "epoch": 0.5010636614567767, + "grad_norm": 0.501694917678833, + "learning_rate": 9.970167105034496e-05, + "loss": 0.6645, + "step": 21375 + }, + { + "epoch": 0.5010871030315817, + "grad_norm": 0.5387786626815796, + "learning_rate": 9.969430492610896e-05, + "loss": 0.5716, + "step": 21376 + }, + { + "epoch": 0.5011105446063867, + "grad_norm": 0.7073926329612732, + "learning_rate": 9.968693880353165e-05, + "loss": 0.171, + "step": 21377 + }, + { + "epoch": 0.5011339861811916, + "grad_norm": 0.5214034914970398, + "learning_rate": 9.967957268265299e-05, + "loss": 0.0748, + "step": 21378 + }, + { + "epoch": 0.5011574277559967, + "grad_norm": 0.22934004664421082, + "learning_rate": 9.967220656351303e-05, + "loss": 0.0449, + "step": 21379 + }, + { + "epoch": 0.5011808693308016, + "grad_norm": 0.22559556365013123, + "learning_rate": 9.966484044615168e-05, + "loss": 0.041, + "step": 21380 + }, + { + "epoch": 0.5012043109056067, + "grad_norm": 0.6116759777069092, + "learning_rate": 9.965747433060889e-05, + "loss": 0.1537, + "step": 21381 + }, + { + "epoch": 0.5012277524804116, + "grad_norm": 0.3492768108844757, + "learning_rate": 9.965010821692469e-05, + "loss": 0.0828, + "step": 21382 + }, + { + "epoch": 0.5012511940552167, + "grad_norm": 0.6136812567710876, + "learning_rate": 9.9642742105139e-05, + "loss": 0.0735, + "step": 21383 + }, + { + "epoch": 0.5012746356300216, + "grad_norm": 0.2807212173938751, + "learning_rate": 9.963537599529178e-05, + "loss": 0.0498, + "step": 21384 + }, + { + "epoch": 0.5012980772048267, + "grad_norm": 0.09217959642410278, + "learning_rate": 9.962800988742305e-05, + "loss": 0.0145, + "step": 21385 + }, + { + "epoch": 0.5013215187796316, + "grad_norm": 0.46483632922172546, + "learning_rate": 9.962064378157274e-05, + "loss": 0.0585, + "step": 21386 + }, + { + "epoch": 0.5013449603544367, + "grad_norm": 0.7705459594726562, + "learning_rate": 9.961327767778082e-05, + "loss": 0.0396, + "step": 21387 + }, + { + "epoch": 0.5013684019292416, + "grad_norm": 0.4413532316684723, + "learning_rate": 9.960591157608727e-05, + "loss": 0.0931, + "step": 21388 + }, + { + "epoch": 0.5013918435040466, + "grad_norm": 0.30009227991104126, + "learning_rate": 9.959854547653206e-05, + "loss": 0.0387, + "step": 21389 + }, + { + "epoch": 0.5014152850788516, + "grad_norm": 0.5337884426116943, + "learning_rate": 9.959117937915511e-05, + "loss": 0.0901, + "step": 21390 + }, + { + "epoch": 0.5014387266536566, + "grad_norm": 0.18400201201438904, + "learning_rate": 9.958381328399646e-05, + "loss": 0.0135, + "step": 21391 + }, + { + "epoch": 0.5014621682284616, + "grad_norm": 0.8882917165756226, + "learning_rate": 9.957644719109607e-05, + "loss": 0.1745, + "step": 21392 + }, + { + "epoch": 0.5014856098032666, + "grad_norm": 0.5357938408851624, + "learning_rate": 9.956908110049388e-05, + "loss": 0.1078, + "step": 21393 + }, + { + "epoch": 0.5015090513780716, + "grad_norm": 0.34041720628738403, + "learning_rate": 9.956171501222986e-05, + "loss": 0.0399, + "step": 21394 + }, + { + "epoch": 0.5015324929528766, + "grad_norm": 0.44059357047080994, + "learning_rate": 9.955434892634401e-05, + "loss": 0.1217, + "step": 21395 + }, + { + "epoch": 0.5015559345276815, + "grad_norm": 0.6719505190849304, + "learning_rate": 9.954698284287627e-05, + "loss": 0.1607, + "step": 21396 + }, + { + "epoch": 0.5015793761024866, + "grad_norm": 0.4197363257408142, + "learning_rate": 9.953961676186658e-05, + "loss": 0.121, + "step": 21397 + }, + { + "epoch": 0.5016028176772915, + "grad_norm": 0.49523770809173584, + "learning_rate": 9.953225068335498e-05, + "loss": 0.1027, + "step": 21398 + }, + { + "epoch": 0.5016262592520966, + "grad_norm": 0.3316394090652466, + "learning_rate": 9.95248846073814e-05, + "loss": 0.0208, + "step": 21399 + }, + { + "epoch": 0.5016497008269015, + "grad_norm": 0.42284849286079407, + "learning_rate": 9.951751853398579e-05, + "loss": 0.1025, + "step": 21400 + }, + { + "epoch": 0.5016731424017066, + "grad_norm": 0.5264827609062195, + "learning_rate": 9.951015246320816e-05, + "loss": 0.1452, + "step": 21401 + }, + { + "epoch": 0.5016965839765115, + "grad_norm": 0.5522547960281372, + "learning_rate": 9.950278639508838e-05, + "loss": 0.1154, + "step": 21402 + }, + { + "epoch": 0.5017200255513166, + "grad_norm": 0.21101799607276917, + "learning_rate": 9.949542032966658e-05, + "loss": 0.0516, + "step": 21403 + }, + { + "epoch": 0.5017434671261215, + "grad_norm": 0.49828433990478516, + "learning_rate": 9.94880542669826e-05, + "loss": 0.0834, + "step": 21404 + }, + { + "epoch": 0.5017669087009266, + "grad_norm": 0.7282685041427612, + "learning_rate": 9.94806882070765e-05, + "loss": 0.6934, + "step": 21405 + }, + { + "epoch": 0.5017903502757315, + "grad_norm": 0.4826507866382599, + "learning_rate": 9.947332214998816e-05, + "loss": 0.093, + "step": 21406 + }, + { + "epoch": 0.5018137918505365, + "grad_norm": 0.2976408004760742, + "learning_rate": 9.946595609575759e-05, + "loss": 0.0778, + "step": 21407 + }, + { + "epoch": 0.5018372334253415, + "grad_norm": 0.6424932479858398, + "learning_rate": 9.945859004442479e-05, + "loss": 0.0757, + "step": 21408 + }, + { + "epoch": 0.5018606750001465, + "grad_norm": 0.6790807247161865, + "learning_rate": 9.945122399602967e-05, + "loss": 0.0852, + "step": 21409 + }, + { + "epoch": 0.5018841165749515, + "grad_norm": 0.2835141122341156, + "learning_rate": 9.944385795061222e-05, + "loss": 0.0524, + "step": 21410 + }, + { + "epoch": 0.5019075581497565, + "grad_norm": 0.7070468068122864, + "learning_rate": 9.943649190821242e-05, + "loss": 0.1577, + "step": 21411 + }, + { + "epoch": 0.5019309997245615, + "grad_norm": 0.37816163897514343, + "learning_rate": 9.942912586887022e-05, + "loss": 0.0482, + "step": 21412 + }, + { + "epoch": 0.5019544412993665, + "grad_norm": 0.32597479224205017, + "learning_rate": 9.942175983262563e-05, + "loss": 0.3953, + "step": 21413 + }, + { + "epoch": 0.5019778828741714, + "grad_norm": 0.38923394680023193, + "learning_rate": 9.941439379951853e-05, + "loss": 0.1131, + "step": 21414 + }, + { + "epoch": 0.5020013244489765, + "grad_norm": 0.08924531191587448, + "learning_rate": 9.940702776958899e-05, + "loss": 0.0161, + "step": 21415 + }, + { + "epoch": 0.5020247660237814, + "grad_norm": 0.6476414203643799, + "learning_rate": 9.939966174287691e-05, + "loss": 0.1488, + "step": 21416 + }, + { + "epoch": 0.5020482075985865, + "grad_norm": 0.36476272344589233, + "learning_rate": 9.939229571942231e-05, + "loss": 0.0698, + "step": 21417 + }, + { + "epoch": 0.5020716491733914, + "grad_norm": 0.4855140149593353, + "learning_rate": 9.938492969926513e-05, + "loss": 0.129, + "step": 21418 + }, + { + "epoch": 0.5020950907481965, + "grad_norm": 0.5874426364898682, + "learning_rate": 9.937756368244532e-05, + "loss": 0.1244, + "step": 21419 + }, + { + "epoch": 0.5021185323230014, + "grad_norm": 0.3540107309818268, + "learning_rate": 9.937019766900289e-05, + "loss": 0.3337, + "step": 21420 + }, + { + "epoch": 0.5021419738978065, + "grad_norm": 0.5156068801879883, + "learning_rate": 9.936283165897777e-05, + "loss": 0.0902, + "step": 21421 + }, + { + "epoch": 0.5021654154726114, + "grad_norm": 0.4040417969226837, + "learning_rate": 9.935546565240993e-05, + "loss": 0.0637, + "step": 21422 + }, + { + "epoch": 0.5021888570474164, + "grad_norm": 0.3219768702983856, + "learning_rate": 9.934809964933938e-05, + "loss": 0.0436, + "step": 21423 + }, + { + "epoch": 0.5022122986222214, + "grad_norm": 0.3040231764316559, + "learning_rate": 9.934073364980605e-05, + "loss": 0.0413, + "step": 21424 + }, + { + "epoch": 0.5022357401970264, + "grad_norm": 0.37552380561828613, + "learning_rate": 9.93333676538499e-05, + "loss": 0.0549, + "step": 21425 + }, + { + "epoch": 0.5022591817718315, + "grad_norm": 0.3145922124385834, + "learning_rate": 9.93260016615109e-05, + "loss": 0.04, + "step": 21426 + }, + { + "epoch": 0.5022826233466364, + "grad_norm": 0.443544864654541, + "learning_rate": 9.931863567282906e-05, + "loss": 0.0542, + "step": 21427 + }, + { + "epoch": 0.5023060649214415, + "grad_norm": 0.2455935925245285, + "learning_rate": 9.931126968784433e-05, + "loss": 0.0193, + "step": 21428 + }, + { + "epoch": 0.5023295064962464, + "grad_norm": 0.5859149694442749, + "learning_rate": 9.930390370659665e-05, + "loss": 0.5845, + "step": 21429 + }, + { + "epoch": 0.5023529480710515, + "grad_norm": 0.07628173381090164, + "learning_rate": 9.929653772912604e-05, + "loss": 0.0092, + "step": 21430 + }, + { + "epoch": 0.5023763896458564, + "grad_norm": 0.2772391140460968, + "learning_rate": 9.928917175547244e-05, + "loss": 0.0561, + "step": 21431 + }, + { + "epoch": 0.5023998312206615, + "grad_norm": 0.5508426427841187, + "learning_rate": 9.928180578567577e-05, + "loss": 0.5653, + "step": 21432 + }, + { + "epoch": 0.5024232727954664, + "grad_norm": 0.37678390741348267, + "learning_rate": 9.927443981977608e-05, + "loss": 0.0491, + "step": 21433 + }, + { + "epoch": 0.5024467143702714, + "grad_norm": 0.7560335397720337, + "learning_rate": 9.926707385781328e-05, + "loss": 0.7435, + "step": 21434 + }, + { + "epoch": 0.5024701559450764, + "grad_norm": 0.4658089876174927, + "learning_rate": 9.925970789982736e-05, + "loss": 0.0365, + "step": 21435 + }, + { + "epoch": 0.5024935975198814, + "grad_norm": 0.45173946022987366, + "learning_rate": 9.92523419458583e-05, + "loss": 0.0945, + "step": 21436 + }, + { + "epoch": 0.5025170390946864, + "grad_norm": 0.2974279820919037, + "learning_rate": 9.924497599594605e-05, + "loss": 0.058, + "step": 21437 + }, + { + "epoch": 0.5025404806694914, + "grad_norm": 0.16705471277236938, + "learning_rate": 9.923761005013054e-05, + "loss": 0.034, + "step": 21438 + }, + { + "epoch": 0.5025639222442964, + "grad_norm": 0.3921283185482025, + "learning_rate": 9.923024410845179e-05, + "loss": 0.1043, + "step": 21439 + }, + { + "epoch": 0.5025873638191014, + "grad_norm": 0.580627977848053, + "learning_rate": 9.92228781709498e-05, + "loss": 0.0772, + "step": 21440 + }, + { + "epoch": 0.5026108053939063, + "grad_norm": 0.4235326647758484, + "learning_rate": 9.921551223766448e-05, + "loss": 0.0544, + "step": 21441 + }, + { + "epoch": 0.5026342469687114, + "grad_norm": 0.8598755598068237, + "learning_rate": 9.920814630863578e-05, + "loss": 0.1393, + "step": 21442 + }, + { + "epoch": 0.5026576885435163, + "grad_norm": 0.1690118908882141, + "learning_rate": 9.920078038390375e-05, + "loss": 0.0273, + "step": 21443 + }, + { + "epoch": 0.5026811301183214, + "grad_norm": 0.48665857315063477, + "learning_rate": 9.91934144635083e-05, + "loss": 0.0555, + "step": 21444 + }, + { + "epoch": 0.5027045716931263, + "grad_norm": 0.30701330304145813, + "learning_rate": 9.918604854748937e-05, + "loss": 0.4629, + "step": 21445 + }, + { + "epoch": 0.5027280132679314, + "grad_norm": 0.3524387776851654, + "learning_rate": 9.9178682635887e-05, + "loss": 0.0822, + "step": 21446 + }, + { + "epoch": 0.5027514548427363, + "grad_norm": 0.24238666892051697, + "learning_rate": 9.91713167287411e-05, + "loss": 0.0444, + "step": 21447 + }, + { + "epoch": 0.5027748964175414, + "grad_norm": 0.47517088055610657, + "learning_rate": 9.916395082609166e-05, + "loss": 0.047, + "step": 21448 + }, + { + "epoch": 0.5027983379923463, + "grad_norm": 0.5397379398345947, + "learning_rate": 9.915658492797866e-05, + "loss": 0.1212, + "step": 21449 + }, + { + "epoch": 0.5028217795671514, + "grad_norm": 0.4331654906272888, + "learning_rate": 9.914921903444201e-05, + "loss": 0.0607, + "step": 21450 + }, + { + "epoch": 0.5028452211419563, + "grad_norm": 0.4824064373970032, + "learning_rate": 9.914185314552176e-05, + "loss": 0.0971, + "step": 21451 + }, + { + "epoch": 0.5028686627167613, + "grad_norm": 0.758773148059845, + "learning_rate": 9.913448726125783e-05, + "loss": 0.664, + "step": 21452 + }, + { + "epoch": 0.5028921042915663, + "grad_norm": 0.5532277822494507, + "learning_rate": 9.912712138169021e-05, + "loss": 0.6171, + "step": 21453 + }, + { + "epoch": 0.5029155458663713, + "grad_norm": 0.5485040545463562, + "learning_rate": 9.911975550685884e-05, + "loss": 0.2588, + "step": 21454 + }, + { + "epoch": 0.5029389874411763, + "grad_norm": 0.5474147796630859, + "learning_rate": 9.911238963680372e-05, + "loss": 0.1556, + "step": 21455 + }, + { + "epoch": 0.5029624290159813, + "grad_norm": 0.531798243522644, + "learning_rate": 9.910502377156479e-05, + "loss": 0.0664, + "step": 21456 + }, + { + "epoch": 0.5029858705907863, + "grad_norm": 0.35698720812797546, + "learning_rate": 9.909765791118201e-05, + "loss": 0.0705, + "step": 21457 + }, + { + "epoch": 0.5030093121655913, + "grad_norm": 0.12158969044685364, + "learning_rate": 9.909029205569537e-05, + "loss": 0.025, + "step": 21458 + }, + { + "epoch": 0.5030327537403962, + "grad_norm": 0.1333918422460556, + "learning_rate": 9.908292620514486e-05, + "loss": 0.0264, + "step": 21459 + }, + { + "epoch": 0.5030561953152013, + "grad_norm": 0.8452542424201965, + "learning_rate": 9.907556035957039e-05, + "loss": 0.1293, + "step": 21460 + }, + { + "epoch": 0.5030796368900062, + "grad_norm": 0.4767395257949829, + "learning_rate": 9.906819451901196e-05, + "loss": 0.0622, + "step": 21461 + }, + { + "epoch": 0.5031030784648113, + "grad_norm": 0.3780533969402313, + "learning_rate": 9.906082868350954e-05, + "loss": 0.0661, + "step": 21462 + }, + { + "epoch": 0.5031265200396162, + "grad_norm": 0.37208348512649536, + "learning_rate": 9.905346285310306e-05, + "loss": 0.0952, + "step": 21463 + }, + { + "epoch": 0.5031499616144213, + "grad_norm": 0.6204017400741577, + "learning_rate": 9.904609702783252e-05, + "loss": 0.161, + "step": 21464 + }, + { + "epoch": 0.5031734031892262, + "grad_norm": 0.6288276314735413, + "learning_rate": 9.903873120773791e-05, + "loss": 0.1097, + "step": 21465 + }, + { + "epoch": 0.5031968447640313, + "grad_norm": 0.4524208903312683, + "learning_rate": 9.903136539285918e-05, + "loss": 0.0878, + "step": 21466 + }, + { + "epoch": 0.5032202863388362, + "grad_norm": 0.41662460565567017, + "learning_rate": 9.902399958323626e-05, + "loss": 0.0784, + "step": 21467 + }, + { + "epoch": 0.5032437279136412, + "grad_norm": 0.41756242513656616, + "learning_rate": 9.901663377890917e-05, + "loss": 0.0849, + "step": 21468 + }, + { + "epoch": 0.5032671694884462, + "grad_norm": 0.42498648166656494, + "learning_rate": 9.900926797991785e-05, + "loss": 0.0691, + "step": 21469 + }, + { + "epoch": 0.5032906110632512, + "grad_norm": 0.2351992279291153, + "learning_rate": 9.900190218630225e-05, + "loss": 0.108, + "step": 21470 + }, + { + "epoch": 0.5033140526380562, + "grad_norm": 0.27613475918769836, + "learning_rate": 9.899453639810238e-05, + "loss": 0.026, + "step": 21471 + }, + { + "epoch": 0.5033374942128612, + "grad_norm": 0.39381107687950134, + "learning_rate": 9.898717061535818e-05, + "loss": 0.0842, + "step": 21472 + }, + { + "epoch": 0.5033609357876662, + "grad_norm": 0.5637950897216797, + "learning_rate": 9.89798048381096e-05, + "loss": 0.1025, + "step": 21473 + }, + { + "epoch": 0.5033843773624712, + "grad_norm": 0.373688668012619, + "learning_rate": 9.897243906639666e-05, + "loss": 0.0369, + "step": 21474 + }, + { + "epoch": 0.5034078189372762, + "grad_norm": 0.6783633828163147, + "learning_rate": 9.896507330025924e-05, + "loss": 0.1472, + "step": 21475 + }, + { + "epoch": 0.5034312605120812, + "grad_norm": 0.5109262466430664, + "learning_rate": 9.895770753973739e-05, + "loss": 0.0451, + "step": 21476 + }, + { + "epoch": 0.5034547020868863, + "grad_norm": 0.5773867964744568, + "learning_rate": 9.895034178487106e-05, + "loss": 0.0887, + "step": 21477 + }, + { + "epoch": 0.5034781436616912, + "grad_norm": 0.26712602376937866, + "learning_rate": 9.894297603570021e-05, + "loss": 0.045, + "step": 21478 + }, + { + "epoch": 0.5035015852364962, + "grad_norm": 0.4138089716434479, + "learning_rate": 9.89356102922648e-05, + "loss": 0.0749, + "step": 21479 + }, + { + "epoch": 0.5035250268113012, + "grad_norm": 0.4609624445438385, + "learning_rate": 9.892824455460478e-05, + "loss": 0.077, + "step": 21480 + }, + { + "epoch": 0.5035484683861062, + "grad_norm": 0.14021722972393036, + "learning_rate": 9.892087882276015e-05, + "loss": 0.0142, + "step": 21481 + }, + { + "epoch": 0.5035719099609112, + "grad_norm": 0.47585010528564453, + "learning_rate": 9.891351309677087e-05, + "loss": 0.5412, + "step": 21482 + }, + { + "epoch": 0.5035953515357162, + "grad_norm": 0.4365462064743042, + "learning_rate": 9.890614737667687e-05, + "loss": 0.0655, + "step": 21483 + }, + { + "epoch": 0.5036187931105212, + "grad_norm": 0.34465721249580383, + "learning_rate": 9.889878166251819e-05, + "loss": 0.049, + "step": 21484 + }, + { + "epoch": 0.5036422346853262, + "grad_norm": 0.4816058576107025, + "learning_rate": 9.88914159543347e-05, + "loss": 0.0874, + "step": 21485 + }, + { + "epoch": 0.5036656762601311, + "grad_norm": 0.4331834316253662, + "learning_rate": 9.888405025216645e-05, + "loss": 0.0926, + "step": 21486 + }, + { + "epoch": 0.5036891178349362, + "grad_norm": 0.5614374279975891, + "learning_rate": 9.887668455605333e-05, + "loss": 0.113, + "step": 21487 + }, + { + "epoch": 0.5037125594097411, + "grad_norm": 0.5133069753646851, + "learning_rate": 9.88693188660354e-05, + "loss": 0.089, + "step": 21488 + }, + { + "epoch": 0.5037360009845462, + "grad_norm": 0.25303396582603455, + "learning_rate": 9.886195318215257e-05, + "loss": 0.0405, + "step": 21489 + }, + { + "epoch": 0.5037594425593511, + "grad_norm": 0.3598216772079468, + "learning_rate": 9.885458750444479e-05, + "loss": 0.0423, + "step": 21490 + }, + { + "epoch": 0.5037828841341562, + "grad_norm": 0.4546228051185608, + "learning_rate": 9.884722183295208e-05, + "loss": 0.098, + "step": 21491 + }, + { + "epoch": 0.5038063257089611, + "grad_norm": 0.6890984773635864, + "learning_rate": 9.883985616771437e-05, + "loss": 0.1148, + "step": 21492 + }, + { + "epoch": 0.5038297672837662, + "grad_norm": 0.7333186268806458, + "learning_rate": 9.883249050877163e-05, + "loss": 0.6964, + "step": 21493 + }, + { + "epoch": 0.5038532088585711, + "grad_norm": 0.6649259328842163, + "learning_rate": 9.882512485616384e-05, + "loss": 0.1003, + "step": 21494 + }, + { + "epoch": 0.5038766504333761, + "grad_norm": 0.42428886890411377, + "learning_rate": 9.881775920993092e-05, + "loss": 0.0879, + "step": 21495 + }, + { + "epoch": 0.5039000920081811, + "grad_norm": 0.3036535680294037, + "learning_rate": 9.881039357011291e-05, + "loss": 0.0293, + "step": 21496 + }, + { + "epoch": 0.5039235335829861, + "grad_norm": 0.07581663876771927, + "learning_rate": 9.880302793674974e-05, + "loss": 0.0122, + "step": 21497 + }, + { + "epoch": 0.5039469751577911, + "grad_norm": 0.5098483562469482, + "learning_rate": 9.879566230988134e-05, + "loss": 0.7698, + "step": 21498 + }, + { + "epoch": 0.5039704167325961, + "grad_norm": 0.42877039313316345, + "learning_rate": 9.87882966895477e-05, + "loss": 0.0734, + "step": 21499 + }, + { + "epoch": 0.5039938583074011, + "grad_norm": 0.5114029049873352, + "learning_rate": 9.878093107578883e-05, + "loss": 0.0957, + "step": 21500 + }, + { + "epoch": 0.5040172998822061, + "grad_norm": 0.22744733095169067, + "learning_rate": 9.877356546864467e-05, + "loss": 0.0272, + "step": 21501 + }, + { + "epoch": 0.504040741457011, + "grad_norm": 0.7818481922149658, + "learning_rate": 9.876619986815516e-05, + "loss": 0.5271, + "step": 21502 + }, + { + "epoch": 0.5040641830318161, + "grad_norm": 0.3568758964538574, + "learning_rate": 9.875883427436028e-05, + "loss": 0.0413, + "step": 21503 + }, + { + "epoch": 0.504087624606621, + "grad_norm": 0.6616005897521973, + "learning_rate": 9.875146868730002e-05, + "loss": 0.0874, + "step": 21504 + }, + { + "epoch": 0.5041110661814261, + "grad_norm": 0.49635249376296997, + "learning_rate": 9.87441031070143e-05, + "loss": 0.1176, + "step": 21505 + }, + { + "epoch": 0.504134507756231, + "grad_norm": 0.5775406956672668, + "learning_rate": 9.873673753354315e-05, + "loss": 0.0967, + "step": 21506 + }, + { + "epoch": 0.5041579493310361, + "grad_norm": 0.41247910261154175, + "learning_rate": 9.872937196692648e-05, + "loss": 0.0643, + "step": 21507 + }, + { + "epoch": 0.504181390905841, + "grad_norm": 0.32763388752937317, + "learning_rate": 9.872200640720424e-05, + "loss": 0.0372, + "step": 21508 + }, + { + "epoch": 0.5042048324806461, + "grad_norm": 0.365288645029068, + "learning_rate": 9.871464085441647e-05, + "loss": 0.0871, + "step": 21509 + }, + { + "epoch": 0.504228274055451, + "grad_norm": 0.39902588725090027, + "learning_rate": 9.870727530860309e-05, + "loss": 0.0866, + "step": 21510 + }, + { + "epoch": 0.5042517156302561, + "grad_norm": 0.18412485718727112, + "learning_rate": 9.869990976980403e-05, + "loss": 0.0261, + "step": 21511 + }, + { + "epoch": 0.504275157205061, + "grad_norm": 0.5287142992019653, + "learning_rate": 9.869254423805932e-05, + "loss": 0.1229, + "step": 21512 + }, + { + "epoch": 0.504298598779866, + "grad_norm": 0.44081801176071167, + "learning_rate": 9.868517871340892e-05, + "loss": 0.0333, + "step": 21513 + }, + { + "epoch": 0.504322040354671, + "grad_norm": 0.3841378688812256, + "learning_rate": 9.867781319589278e-05, + "loss": 0.0802, + "step": 21514 + }, + { + "epoch": 0.504345481929476, + "grad_norm": 0.6284140944480896, + "learning_rate": 9.867044768555084e-05, + "loss": 0.0512, + "step": 21515 + }, + { + "epoch": 0.504368923504281, + "grad_norm": 0.08054893463850021, + "learning_rate": 9.866308218242311e-05, + "loss": 0.0217, + "step": 21516 + }, + { + "epoch": 0.504392365079086, + "grad_norm": 0.40042680501937866, + "learning_rate": 9.865571668654953e-05, + "loss": 0.0446, + "step": 21517 + }, + { + "epoch": 0.504415806653891, + "grad_norm": 0.48172423243522644, + "learning_rate": 9.864835119797005e-05, + "loss": 0.1087, + "step": 21518 + }, + { + "epoch": 0.504439248228696, + "grad_norm": 0.7387813329696655, + "learning_rate": 9.864098571672468e-05, + "loss": 0.1208, + "step": 21519 + }, + { + "epoch": 0.504462689803501, + "grad_norm": 0.1144927367568016, + "learning_rate": 9.863362024285335e-05, + "loss": 0.0112, + "step": 21520 + }, + { + "epoch": 0.504486131378306, + "grad_norm": 0.370938777923584, + "learning_rate": 9.862625477639603e-05, + "loss": 0.0525, + "step": 21521 + }, + { + "epoch": 0.5045095729531109, + "grad_norm": 0.5554680228233337, + "learning_rate": 9.861888931739269e-05, + "loss": 0.6183, + "step": 21522 + }, + { + "epoch": 0.504533014527916, + "grad_norm": 0.4156605899333954, + "learning_rate": 9.861152386588326e-05, + "loss": 0.0879, + "step": 21523 + }, + { + "epoch": 0.5045564561027209, + "grad_norm": 0.5996803641319275, + "learning_rate": 9.860415842190779e-05, + "loss": 0.1482, + "step": 21524 + }, + { + "epoch": 0.504579897677526, + "grad_norm": 0.38856709003448486, + "learning_rate": 9.859679298550617e-05, + "loss": 0.0438, + "step": 21525 + }, + { + "epoch": 0.5046033392523309, + "grad_norm": 0.8005272150039673, + "learning_rate": 9.858942755671841e-05, + "loss": 0.1175, + "step": 21526 + }, + { + "epoch": 0.504626780827136, + "grad_norm": 0.14307932555675507, + "learning_rate": 9.858206213558445e-05, + "loss": 0.0396, + "step": 21527 + }, + { + "epoch": 0.504650222401941, + "grad_norm": 0.1894749253988266, + "learning_rate": 9.857469672214426e-05, + "loss": 0.0392, + "step": 21528 + }, + { + "epoch": 0.504673663976746, + "grad_norm": 0.3462470471858978, + "learning_rate": 9.85673313164378e-05, + "loss": 0.053, + "step": 21529 + }, + { + "epoch": 0.504697105551551, + "grad_norm": 0.6080081462860107, + "learning_rate": 9.855996591850504e-05, + "loss": 0.1309, + "step": 21530 + }, + { + "epoch": 0.504720547126356, + "grad_norm": 0.2625477612018585, + "learning_rate": 9.855260052838596e-05, + "loss": 0.0425, + "step": 21531 + }, + { + "epoch": 0.504743988701161, + "grad_norm": 0.3757103383541107, + "learning_rate": 9.854523514612051e-05, + "loss": 0.044, + "step": 21532 + }, + { + "epoch": 0.5047674302759659, + "grad_norm": 0.34189626574516296, + "learning_rate": 9.853786977174864e-05, + "loss": 0.0584, + "step": 21533 + }, + { + "epoch": 0.504790871850771, + "grad_norm": 0.1558627337217331, + "learning_rate": 9.853050440531035e-05, + "loss": 0.0495, + "step": 21534 + }, + { + "epoch": 0.5048143134255759, + "grad_norm": 0.6088181734085083, + "learning_rate": 9.852313904684558e-05, + "loss": 0.0939, + "step": 21535 + }, + { + "epoch": 0.504837755000381, + "grad_norm": 0.6198761463165283, + "learning_rate": 9.851577369639424e-05, + "loss": 0.1439, + "step": 21536 + }, + { + "epoch": 0.5048611965751859, + "grad_norm": 0.5495381355285645, + "learning_rate": 9.85084083539964e-05, + "loss": 0.4991, + "step": 21537 + }, + { + "epoch": 0.504884638149991, + "grad_norm": 0.15099884569644928, + "learning_rate": 9.850104301969199e-05, + "loss": 0.014, + "step": 21538 + }, + { + "epoch": 0.5049080797247959, + "grad_norm": 0.3318725526332855, + "learning_rate": 9.849367769352096e-05, + "loss": 0.0604, + "step": 21539 + }, + { + "epoch": 0.504931521299601, + "grad_norm": 1.5126423835754395, + "learning_rate": 9.848631237552325e-05, + "loss": 0.29, + "step": 21540 + }, + { + "epoch": 0.5049549628744059, + "grad_norm": 0.4825141429901123, + "learning_rate": 9.847894706573889e-05, + "loss": 0.099, + "step": 21541 + }, + { + "epoch": 0.5049784044492109, + "grad_norm": 0.2833386957645416, + "learning_rate": 9.84715817642078e-05, + "loss": 0.0376, + "step": 21542 + }, + { + "epoch": 0.5050018460240159, + "grad_norm": 0.5610377788543701, + "learning_rate": 9.846421647096993e-05, + "loss": 0.1726, + "step": 21543 + }, + { + "epoch": 0.5050252875988209, + "grad_norm": 0.41358160972595215, + "learning_rate": 9.845685118606528e-05, + "loss": 0.0802, + "step": 21544 + }, + { + "epoch": 0.5050487291736259, + "grad_norm": 0.246430903673172, + "learning_rate": 9.84494859095338e-05, + "loss": 0.0802, + "step": 21545 + }, + { + "epoch": 0.5050721707484309, + "grad_norm": 0.2488251030445099, + "learning_rate": 9.844212064141544e-05, + "loss": 0.0288, + "step": 21546 + }, + { + "epoch": 0.5050956123232359, + "grad_norm": 0.6332094073295593, + "learning_rate": 9.843475538175019e-05, + "loss": 0.8301, + "step": 21547 + }, + { + "epoch": 0.5051190538980409, + "grad_norm": 0.48100629448890686, + "learning_rate": 9.842739013057796e-05, + "loss": 0.1206, + "step": 21548 + }, + { + "epoch": 0.5051424954728458, + "grad_norm": 0.22656957805156708, + "learning_rate": 9.84200248879388e-05, + "loss": 0.0529, + "step": 21549 + }, + { + "epoch": 0.5051659370476509, + "grad_norm": 0.5526788234710693, + "learning_rate": 9.84126596538726e-05, + "loss": 0.0866, + "step": 21550 + }, + { + "epoch": 0.5051893786224558, + "grad_norm": 0.40069347620010376, + "learning_rate": 9.840529442841939e-05, + "loss": 0.099, + "step": 21551 + }, + { + "epoch": 0.5052128201972609, + "grad_norm": 0.30716365575790405, + "learning_rate": 9.83979292116191e-05, + "loss": 0.0829, + "step": 21552 + }, + { + "epoch": 0.5052362617720658, + "grad_norm": 1.0883152484893799, + "learning_rate": 9.839056400351164e-05, + "loss": 0.2313, + "step": 21553 + }, + { + "epoch": 0.5052597033468709, + "grad_norm": 0.2625840902328491, + "learning_rate": 9.838319880413706e-05, + "loss": 0.0875, + "step": 21554 + }, + { + "epoch": 0.5052831449216758, + "grad_norm": 0.27493229508399963, + "learning_rate": 9.83758336135353e-05, + "loss": 0.0553, + "step": 21555 + }, + { + "epoch": 0.5053065864964809, + "grad_norm": 0.6536187529563904, + "learning_rate": 9.836846843174629e-05, + "loss": 0.1591, + "step": 21556 + }, + { + "epoch": 0.5053300280712858, + "grad_norm": 0.3536751866340637, + "learning_rate": 9.836110325881003e-05, + "loss": 0.0555, + "step": 21557 + }, + { + "epoch": 0.5053534696460908, + "grad_norm": 0.28895553946495056, + "learning_rate": 9.835373809476648e-05, + "loss": 0.0606, + "step": 21558 + }, + { + "epoch": 0.5053769112208958, + "grad_norm": 0.3382940888404846, + "learning_rate": 9.834637293965557e-05, + "loss": 0.0519, + "step": 21559 + }, + { + "epoch": 0.5054003527957008, + "grad_norm": 0.08692570775747299, + "learning_rate": 9.833900779351725e-05, + "loss": 0.0189, + "step": 21560 + }, + { + "epoch": 0.5054237943705058, + "grad_norm": 0.32167667150497437, + "learning_rate": 9.833164265639157e-05, + "loss": 0.0838, + "step": 21561 + }, + { + "epoch": 0.5054472359453108, + "grad_norm": 0.07831834256649017, + "learning_rate": 9.832427752831844e-05, + "loss": 0.014, + "step": 21562 + }, + { + "epoch": 0.5054706775201158, + "grad_norm": 0.5451489686965942, + "learning_rate": 9.831691240933782e-05, + "loss": 0.1643, + "step": 21563 + }, + { + "epoch": 0.5054941190949208, + "grad_norm": 0.5881475806236267, + "learning_rate": 9.830954729948969e-05, + "loss": 0.8373, + "step": 21564 + }, + { + "epoch": 0.5055175606697258, + "grad_norm": 0.40354663133621216, + "learning_rate": 9.8302182198814e-05, + "loss": 0.0879, + "step": 21565 + }, + { + "epoch": 0.5055410022445308, + "grad_norm": 0.46346524357795715, + "learning_rate": 9.829481710735073e-05, + "loss": 0.1126, + "step": 21566 + }, + { + "epoch": 0.5055644438193357, + "grad_norm": 0.24824966490268707, + "learning_rate": 9.828745202513982e-05, + "loss": 0.0499, + "step": 21567 + }, + { + "epoch": 0.5055878853941408, + "grad_norm": 0.49389469623565674, + "learning_rate": 9.828008695222122e-05, + "loss": 0.1357, + "step": 21568 + }, + { + "epoch": 0.5056113269689457, + "grad_norm": 0.4846523702144623, + "learning_rate": 9.827272188863494e-05, + "loss": 0.1035, + "step": 21569 + }, + { + "epoch": 0.5056347685437508, + "grad_norm": 0.3894192576408386, + "learning_rate": 9.826535683442093e-05, + "loss": 0.05, + "step": 21570 + }, + { + "epoch": 0.5056582101185557, + "grad_norm": 0.2903822958469391, + "learning_rate": 9.825799178961912e-05, + "loss": 0.0473, + "step": 21571 + }, + { + "epoch": 0.5056816516933608, + "grad_norm": 0.3694990575313568, + "learning_rate": 9.825062675426947e-05, + "loss": 0.044, + "step": 21572 + }, + { + "epoch": 0.5057050932681657, + "grad_norm": 0.1387915015220642, + "learning_rate": 9.824326172841202e-05, + "loss": 0.0205, + "step": 21573 + }, + { + "epoch": 0.5057285348429708, + "grad_norm": 0.4310961067676544, + "learning_rate": 9.823589671208666e-05, + "loss": 0.113, + "step": 21574 + }, + { + "epoch": 0.5057519764177757, + "grad_norm": 0.3355894386768341, + "learning_rate": 9.822853170533338e-05, + "loss": 0.0506, + "step": 21575 + }, + { + "epoch": 0.5057754179925807, + "grad_norm": 0.44364216923713684, + "learning_rate": 9.822116670819214e-05, + "loss": 0.0974, + "step": 21576 + }, + { + "epoch": 0.5057988595673857, + "grad_norm": 0.5964653491973877, + "learning_rate": 9.82138017207029e-05, + "loss": 0.1286, + "step": 21577 + }, + { + "epoch": 0.5058223011421907, + "grad_norm": 0.32525596022605896, + "learning_rate": 9.820643674290561e-05, + "loss": 0.0438, + "step": 21578 + }, + { + "epoch": 0.5058457427169958, + "grad_norm": 0.20447295904159546, + "learning_rate": 9.819907177484026e-05, + "loss": 0.0667, + "step": 21579 + }, + { + "epoch": 0.5058691842918007, + "grad_norm": 0.3599911630153656, + "learning_rate": 9.819170681654679e-05, + "loss": 0.0834, + "step": 21580 + }, + { + "epoch": 0.5058926258666058, + "grad_norm": 0.2692370116710663, + "learning_rate": 9.818434186806515e-05, + "loss": 0.0312, + "step": 21581 + }, + { + "epoch": 0.5059160674414107, + "grad_norm": 0.0625365823507309, + "learning_rate": 9.817697692943535e-05, + "loss": 0.012, + "step": 21582 + }, + { + "epoch": 0.5059395090162158, + "grad_norm": 0.42987263202667236, + "learning_rate": 9.816961200069733e-05, + "loss": 0.0649, + "step": 21583 + }, + { + "epoch": 0.5059629505910207, + "grad_norm": 0.20988789200782776, + "learning_rate": 9.8162247081891e-05, + "loss": 0.0494, + "step": 21584 + }, + { + "epoch": 0.5059863921658257, + "grad_norm": 0.18767400085926056, + "learning_rate": 9.815488217305639e-05, + "loss": 0.0422, + "step": 21585 + }, + { + "epoch": 0.5060098337406307, + "grad_norm": 0.46529513597488403, + "learning_rate": 9.814751727423346e-05, + "loss": 0.105, + "step": 21586 + }, + { + "epoch": 0.5060332753154357, + "grad_norm": 0.3642081916332245, + "learning_rate": 9.814015238546215e-05, + "loss": 0.0326, + "step": 21587 + }, + { + "epoch": 0.5060567168902407, + "grad_norm": 0.4457550048828125, + "learning_rate": 9.81327875067824e-05, + "loss": 0.1129, + "step": 21588 + }, + { + "epoch": 0.5060801584650457, + "grad_norm": 0.2868448793888092, + "learning_rate": 9.812542263823424e-05, + "loss": 0.0653, + "step": 21589 + }, + { + "epoch": 0.5061036000398507, + "grad_norm": 0.4267842769622803, + "learning_rate": 9.811805777985757e-05, + "loss": 0.0337, + "step": 21590 + }, + { + "epoch": 0.5061270416146557, + "grad_norm": 0.15233612060546875, + "learning_rate": 9.811069293169237e-05, + "loss": 0.0321, + "step": 21591 + }, + { + "epoch": 0.5061504831894607, + "grad_norm": 0.49587514996528625, + "learning_rate": 9.81033280937786e-05, + "loss": 0.1161, + "step": 21592 + }, + { + "epoch": 0.5061739247642657, + "grad_norm": 0.7490708231925964, + "learning_rate": 9.809596326615624e-05, + "loss": 0.1475, + "step": 21593 + }, + { + "epoch": 0.5061973663390706, + "grad_norm": 0.5871614217758179, + "learning_rate": 9.808859844886521e-05, + "loss": 0.1187, + "step": 21594 + }, + { + "epoch": 0.5062208079138757, + "grad_norm": 0.19024547934532166, + "learning_rate": 9.808123364194552e-05, + "loss": 0.0229, + "step": 21595 + }, + { + "epoch": 0.5062442494886806, + "grad_norm": 0.506367027759552, + "learning_rate": 9.807386884543706e-05, + "loss": 0.0918, + "step": 21596 + }, + { + "epoch": 0.5062676910634857, + "grad_norm": 0.16781187057495117, + "learning_rate": 9.806650405937991e-05, + "loss": 0.0238, + "step": 21597 + }, + { + "epoch": 0.5062911326382906, + "grad_norm": 0.36994463205337524, + "learning_rate": 9.805913928381391e-05, + "loss": 0.3024, + "step": 21598 + }, + { + "epoch": 0.5063145742130957, + "grad_norm": 0.5146498680114746, + "learning_rate": 9.805177451877912e-05, + "loss": 0.0517, + "step": 21599 + }, + { + "epoch": 0.5063380157879006, + "grad_norm": 0.31743112206459045, + "learning_rate": 9.804440976431545e-05, + "loss": 0.0613, + "step": 21600 + }, + { + "epoch": 0.5063614573627057, + "grad_norm": 0.612046480178833, + "learning_rate": 9.803704502046285e-05, + "loss": 0.1448, + "step": 21601 + }, + { + "epoch": 0.5063848989375106, + "grad_norm": 0.8108476400375366, + "learning_rate": 9.80296802872613e-05, + "loss": 0.6414, + "step": 21602 + }, + { + "epoch": 0.5064083405123156, + "grad_norm": 0.5594377517700195, + "learning_rate": 9.802231556475077e-05, + "loss": 0.1069, + "step": 21603 + }, + { + "epoch": 0.5064317820871206, + "grad_norm": 0.46562260389328003, + "learning_rate": 9.801495085297122e-05, + "loss": 0.0787, + "step": 21604 + }, + { + "epoch": 0.5064552236619256, + "grad_norm": 0.42202213406562805, + "learning_rate": 9.800758615196259e-05, + "loss": 0.0924, + "step": 21605 + }, + { + "epoch": 0.5064786652367306, + "grad_norm": 0.5284475684165955, + "learning_rate": 9.800022146176484e-05, + "loss": 0.1323, + "step": 21606 + }, + { + "epoch": 0.5065021068115356, + "grad_norm": 0.43060314655303955, + "learning_rate": 9.799285678241797e-05, + "loss": 0.0984, + "step": 21607 + }, + { + "epoch": 0.5065255483863406, + "grad_norm": 0.08203243464231491, + "learning_rate": 9.79854921139619e-05, + "loss": 0.0113, + "step": 21608 + }, + { + "epoch": 0.5065489899611456, + "grad_norm": 0.43333882093429565, + "learning_rate": 9.797812745643659e-05, + "loss": 0.0456, + "step": 21609 + }, + { + "epoch": 0.5065724315359506, + "grad_norm": 0.1901283860206604, + "learning_rate": 9.797076280988202e-05, + "loss": 0.0237, + "step": 21610 + }, + { + "epoch": 0.5065958731107556, + "grad_norm": 0.14064349234104156, + "learning_rate": 9.796339817433817e-05, + "loss": 0.0193, + "step": 21611 + }, + { + "epoch": 0.5066193146855605, + "grad_norm": 0.49965447187423706, + "learning_rate": 9.795603354984499e-05, + "loss": 0.107, + "step": 21612 + }, + { + "epoch": 0.5066427562603656, + "grad_norm": 0.5197572112083435, + "learning_rate": 9.79486689364424e-05, + "loss": 0.073, + "step": 21613 + }, + { + "epoch": 0.5066661978351705, + "grad_norm": 0.5062830448150635, + "learning_rate": 9.79413043341704e-05, + "loss": 0.1269, + "step": 21614 + }, + { + "epoch": 0.5066896394099756, + "grad_norm": 0.3570843040943146, + "learning_rate": 9.793393974306896e-05, + "loss": 0.0644, + "step": 21615 + }, + { + "epoch": 0.5067130809847805, + "grad_norm": 0.4838702380657196, + "learning_rate": 9.792657516317799e-05, + "loss": 0.5737, + "step": 21616 + }, + { + "epoch": 0.5067365225595856, + "grad_norm": 0.5039934515953064, + "learning_rate": 9.791921059453752e-05, + "loss": 0.6516, + "step": 21617 + }, + { + "epoch": 0.5067599641343905, + "grad_norm": 0.5050910115242004, + "learning_rate": 9.791184603718744e-05, + "loss": 0.1161, + "step": 21618 + }, + { + "epoch": 0.5067834057091956, + "grad_norm": 0.35762113332748413, + "learning_rate": 9.790448149116774e-05, + "loss": 0.0527, + "step": 21619 + }, + { + "epoch": 0.5068068472840005, + "grad_norm": 0.1354375034570694, + "learning_rate": 9.78971169565184e-05, + "loss": 0.0062, + "step": 21620 + }, + { + "epoch": 0.5068302888588055, + "grad_norm": 0.43792060017585754, + "learning_rate": 9.788975243327932e-05, + "loss": 0.0618, + "step": 21621 + }, + { + "epoch": 0.5068537304336105, + "grad_norm": 0.1803559809923172, + "learning_rate": 9.788238792149055e-05, + "loss": 0.0369, + "step": 21622 + }, + { + "epoch": 0.5068771720084155, + "grad_norm": 0.5926923155784607, + "learning_rate": 9.787502342119198e-05, + "loss": 0.0953, + "step": 21623 + }, + { + "epoch": 0.5069006135832205, + "grad_norm": 0.3946874737739563, + "learning_rate": 9.786765893242362e-05, + "loss": 0.0394, + "step": 21624 + }, + { + "epoch": 0.5069240551580255, + "grad_norm": 0.313736230134964, + "learning_rate": 9.786029445522538e-05, + "loss": 0.0735, + "step": 21625 + }, + { + "epoch": 0.5069474967328305, + "grad_norm": 0.3845953643321991, + "learning_rate": 9.785292998963724e-05, + "loss": 0.1062, + "step": 21626 + }, + { + "epoch": 0.5069709383076355, + "grad_norm": 0.5211442112922668, + "learning_rate": 9.784556553569919e-05, + "loss": 0.1098, + "step": 21627 + }, + { + "epoch": 0.5069943798824404, + "grad_norm": 0.5020673871040344, + "learning_rate": 9.783820109345114e-05, + "loss": 0.0906, + "step": 21628 + }, + { + "epoch": 0.5070178214572455, + "grad_norm": 0.3815505802631378, + "learning_rate": 9.783083666293307e-05, + "loss": 0.2545, + "step": 21629 + }, + { + "epoch": 0.5070412630320504, + "grad_norm": 0.5355131030082703, + "learning_rate": 9.782347224418496e-05, + "loss": 0.1001, + "step": 21630 + }, + { + "epoch": 0.5070647046068555, + "grad_norm": 0.5390394926071167, + "learning_rate": 9.781610783724674e-05, + "loss": 0.065, + "step": 21631 + }, + { + "epoch": 0.5070881461816605, + "grad_norm": 0.3857334852218628, + "learning_rate": 9.780874344215836e-05, + "loss": 0.0759, + "step": 21632 + }, + { + "epoch": 0.5071115877564655, + "grad_norm": 0.30124011635780334, + "learning_rate": 9.780137905895978e-05, + "loss": 0.0614, + "step": 21633 + }, + { + "epoch": 0.5071350293312705, + "grad_norm": 0.3156091272830963, + "learning_rate": 9.779401468769102e-05, + "loss": 0.0272, + "step": 21634 + }, + { + "epoch": 0.5071584709060755, + "grad_norm": 0.4067196547985077, + "learning_rate": 9.778665032839201e-05, + "loss": 0.7157, + "step": 21635 + }, + { + "epoch": 0.5071819124808805, + "grad_norm": 0.4419681429862976, + "learning_rate": 9.777928598110266e-05, + "loss": 0.0682, + "step": 21636 + }, + { + "epoch": 0.5072053540556855, + "grad_norm": 0.28966882824897766, + "learning_rate": 9.7771921645863e-05, + "loss": 0.0555, + "step": 21637 + }, + { + "epoch": 0.5072287956304905, + "grad_norm": 0.44031086564064026, + "learning_rate": 9.776455732271296e-05, + "loss": 0.0827, + "step": 21638 + }, + { + "epoch": 0.5072522372052954, + "grad_norm": 0.5040519833564758, + "learning_rate": 9.775719301169246e-05, + "loss": 0.0584, + "step": 21639 + }, + { + "epoch": 0.5072756787801005, + "grad_norm": 0.3388882577419281, + "learning_rate": 9.774982871284153e-05, + "loss": 0.0812, + "step": 21640 + }, + { + "epoch": 0.5072991203549054, + "grad_norm": 0.29945605993270874, + "learning_rate": 9.774246442620005e-05, + "loss": 0.0518, + "step": 21641 + }, + { + "epoch": 0.5073225619297105, + "grad_norm": 0.329913854598999, + "learning_rate": 9.773510015180807e-05, + "loss": 0.0333, + "step": 21642 + }, + { + "epoch": 0.5073460035045154, + "grad_norm": 0.38992199301719666, + "learning_rate": 9.772773588970548e-05, + "loss": 0.0466, + "step": 21643 + }, + { + "epoch": 0.5073694450793205, + "grad_norm": 0.4112875461578369, + "learning_rate": 9.772037163993224e-05, + "loss": 0.0778, + "step": 21644 + }, + { + "epoch": 0.5073928866541254, + "grad_norm": 0.2401932328939438, + "learning_rate": 9.771300740252833e-05, + "loss": 0.0329, + "step": 21645 + }, + { + "epoch": 0.5074163282289305, + "grad_norm": 0.4434085786342621, + "learning_rate": 9.770564317753371e-05, + "loss": 0.1109, + "step": 21646 + }, + { + "epoch": 0.5074397698037354, + "grad_norm": 0.40951958298683167, + "learning_rate": 9.769827896498836e-05, + "loss": 0.0684, + "step": 21647 + }, + { + "epoch": 0.5074632113785404, + "grad_norm": 0.48106586933135986, + "learning_rate": 9.769091476493217e-05, + "loss": 0.0327, + "step": 21648 + }, + { + "epoch": 0.5074866529533454, + "grad_norm": 0.7349977493286133, + "learning_rate": 9.768355057740518e-05, + "loss": 0.1485, + "step": 21649 + }, + { + "epoch": 0.5075100945281504, + "grad_norm": 0.47467467188835144, + "learning_rate": 9.767618640244732e-05, + "loss": 0.0914, + "step": 21650 + }, + { + "epoch": 0.5075335361029554, + "grad_norm": 0.5770235657691956, + "learning_rate": 9.766882224009849e-05, + "loss": 0.1373, + "step": 21651 + }, + { + "epoch": 0.5075569776777604, + "grad_norm": 0.40302762389183044, + "learning_rate": 9.766145809039874e-05, + "loss": 0.0662, + "step": 21652 + }, + { + "epoch": 0.5075804192525654, + "grad_norm": 1.0645984411239624, + "learning_rate": 9.765409395338797e-05, + "loss": 0.1737, + "step": 21653 + }, + { + "epoch": 0.5076038608273704, + "grad_norm": 0.620578408241272, + "learning_rate": 9.764672982910613e-05, + "loss": 0.1157, + "step": 21654 + }, + { + "epoch": 0.5076273024021754, + "grad_norm": 0.33580344915390015, + "learning_rate": 9.763936571759323e-05, + "loss": 0.0604, + "step": 21655 + }, + { + "epoch": 0.5076507439769804, + "grad_norm": 0.5140570998191833, + "learning_rate": 9.76320016188892e-05, + "loss": 0.1146, + "step": 21656 + }, + { + "epoch": 0.5076741855517853, + "grad_norm": 0.5185297131538391, + "learning_rate": 9.762463753303395e-05, + "loss": 0.0438, + "step": 21657 + }, + { + "epoch": 0.5076976271265904, + "grad_norm": 0.3331601917743683, + "learning_rate": 9.761727346006751e-05, + "loss": 0.0707, + "step": 21658 + }, + { + "epoch": 0.5077210687013953, + "grad_norm": 0.6387835741043091, + "learning_rate": 9.760990940002982e-05, + "loss": 0.0801, + "step": 21659 + }, + { + "epoch": 0.5077445102762004, + "grad_norm": 0.4912797212600708, + "learning_rate": 9.760254535296083e-05, + "loss": 0.1042, + "step": 21660 + }, + { + "epoch": 0.5077679518510053, + "grad_norm": 0.5156233310699463, + "learning_rate": 9.759518131890049e-05, + "loss": 0.1125, + "step": 21661 + }, + { + "epoch": 0.5077913934258104, + "grad_norm": 0.4341222643852234, + "learning_rate": 9.758781729788876e-05, + "loss": 0.1437, + "step": 21662 + }, + { + "epoch": 0.5078148350006153, + "grad_norm": 0.06987357884645462, + "learning_rate": 9.758045328996562e-05, + "loss": 0.005, + "step": 21663 + }, + { + "epoch": 0.5078382765754204, + "grad_norm": 0.22872208058834076, + "learning_rate": 9.757308929517098e-05, + "loss": 0.065, + "step": 21664 + }, + { + "epoch": 0.5078617181502253, + "grad_norm": 0.4137830436229706, + "learning_rate": 9.756572531354485e-05, + "loss": 0.107, + "step": 21665 + }, + { + "epoch": 0.5078851597250303, + "grad_norm": 0.28324899077415466, + "learning_rate": 9.755836134512716e-05, + "loss": 0.017, + "step": 21666 + }, + { + "epoch": 0.5079086012998353, + "grad_norm": 0.43168365955352783, + "learning_rate": 9.755099738995787e-05, + "loss": 0.0765, + "step": 21667 + }, + { + "epoch": 0.5079320428746403, + "grad_norm": 0.1358930617570877, + "learning_rate": 9.754363344807694e-05, + "loss": 0.0274, + "step": 21668 + }, + { + "epoch": 0.5079554844494453, + "grad_norm": 0.4053719639778137, + "learning_rate": 9.753626951952428e-05, + "loss": 0.1185, + "step": 21669 + }, + { + "epoch": 0.5079789260242503, + "grad_norm": 0.3428092300891876, + "learning_rate": 9.752890560433992e-05, + "loss": 0.0471, + "step": 21670 + }, + { + "epoch": 0.5080023675990553, + "grad_norm": 0.38466766476631165, + "learning_rate": 9.752154170256379e-05, + "loss": 0.0966, + "step": 21671 + }, + { + "epoch": 0.5080258091738603, + "grad_norm": 0.545243501663208, + "learning_rate": 9.751417781423586e-05, + "loss": 0.1702, + "step": 21672 + }, + { + "epoch": 0.5080492507486652, + "grad_norm": 0.39916011691093445, + "learning_rate": 9.750681393939606e-05, + "loss": 0.067, + "step": 21673 + }, + { + "epoch": 0.5080726923234703, + "grad_norm": 0.264331579208374, + "learning_rate": 9.749945007808435e-05, + "loss": 0.0473, + "step": 21674 + }, + { + "epoch": 0.5080961338982752, + "grad_norm": 0.4891696274280548, + "learning_rate": 9.74920862303407e-05, + "loss": 0.1385, + "step": 21675 + }, + { + "epoch": 0.5081195754730803, + "grad_norm": 0.5541577339172363, + "learning_rate": 9.748472239620507e-05, + "loss": 0.1464, + "step": 21676 + }, + { + "epoch": 0.5081430170478852, + "grad_norm": 0.5718458890914917, + "learning_rate": 9.747735857571739e-05, + "loss": 0.1025, + "step": 21677 + }, + { + "epoch": 0.5081664586226903, + "grad_norm": 0.36577677726745605, + "learning_rate": 9.746999476891766e-05, + "loss": 0.098, + "step": 21678 + }, + { + "epoch": 0.5081899001974952, + "grad_norm": 0.2362353801727295, + "learning_rate": 9.746263097584579e-05, + "loss": 0.0398, + "step": 21679 + }, + { + "epoch": 0.5082133417723003, + "grad_norm": 0.9209792017936707, + "learning_rate": 9.745526719654176e-05, + "loss": 0.2416, + "step": 21680 + }, + { + "epoch": 0.5082367833471052, + "grad_norm": 0.6220048666000366, + "learning_rate": 9.744790343104547e-05, + "loss": 0.1628, + "step": 21681 + }, + { + "epoch": 0.5082602249219103, + "grad_norm": 0.48805898427963257, + "learning_rate": 9.744053967939699e-05, + "loss": 0.5304, + "step": 21682 + }, + { + "epoch": 0.5082836664967153, + "grad_norm": 0.2729153335094452, + "learning_rate": 9.74331759416362e-05, + "loss": 0.0491, + "step": 21683 + }, + { + "epoch": 0.5083071080715202, + "grad_norm": 0.5648029446601868, + "learning_rate": 9.742581221780307e-05, + "loss": 0.1512, + "step": 21684 + }, + { + "epoch": 0.5083305496463253, + "grad_norm": 0.4340246319770813, + "learning_rate": 9.741844850793757e-05, + "loss": 0.2659, + "step": 21685 + }, + { + "epoch": 0.5083539912211302, + "grad_norm": 0.3467492163181305, + "learning_rate": 9.741108481207961e-05, + "loss": 0.0726, + "step": 21686 + }, + { + "epoch": 0.5083774327959353, + "grad_norm": 0.5550252795219421, + "learning_rate": 9.74037211302692e-05, + "loss": 0.0952, + "step": 21687 + }, + { + "epoch": 0.5084008743707402, + "grad_norm": 0.6401856541633606, + "learning_rate": 9.739635746254627e-05, + "loss": 0.121, + "step": 21688 + }, + { + "epoch": 0.5084243159455453, + "grad_norm": 0.4325920045375824, + "learning_rate": 9.738899380895077e-05, + "loss": 0.0872, + "step": 21689 + }, + { + "epoch": 0.5084477575203502, + "grad_norm": 0.2415488064289093, + "learning_rate": 9.738163016952267e-05, + "loss": 0.0491, + "step": 21690 + }, + { + "epoch": 0.5084711990951553, + "grad_norm": 0.5604200959205627, + "learning_rate": 9.737426654430192e-05, + "loss": 0.0638, + "step": 21691 + }, + { + "epoch": 0.5084946406699602, + "grad_norm": 0.3895094394683838, + "learning_rate": 9.736690293332845e-05, + "loss": 0.0935, + "step": 21692 + }, + { + "epoch": 0.5085180822447652, + "grad_norm": 0.7485367059707642, + "learning_rate": 9.735953933664227e-05, + "loss": 0.1537, + "step": 21693 + }, + { + "epoch": 0.5085415238195702, + "grad_norm": 0.15121275186538696, + "learning_rate": 9.735217575428326e-05, + "loss": 0.0266, + "step": 21694 + }, + { + "epoch": 0.5085649653943752, + "grad_norm": 0.6121008396148682, + "learning_rate": 9.734481218629146e-05, + "loss": 0.7776, + "step": 21695 + }, + { + "epoch": 0.5085884069691802, + "grad_norm": 0.14842990040779114, + "learning_rate": 9.733744863270676e-05, + "loss": 0.0247, + "step": 21696 + }, + { + "epoch": 0.5086118485439852, + "grad_norm": 0.5546321868896484, + "learning_rate": 9.733008509356916e-05, + "loss": 0.0824, + "step": 21697 + }, + { + "epoch": 0.5086352901187902, + "grad_norm": 0.1807037889957428, + "learning_rate": 9.732272156891859e-05, + "loss": 0.0274, + "step": 21698 + }, + { + "epoch": 0.5086587316935952, + "grad_norm": 0.4934004843235016, + "learning_rate": 9.731535805879499e-05, + "loss": 0.1336, + "step": 21699 + }, + { + "epoch": 0.5086821732684002, + "grad_norm": 0.8147866725921631, + "learning_rate": 9.730799456323835e-05, + "loss": 0.1641, + "step": 21700 + }, + { + "epoch": 0.5087056148432052, + "grad_norm": 0.42121145129203796, + "learning_rate": 9.730063108228862e-05, + "loss": 0.1339, + "step": 21701 + }, + { + "epoch": 0.5087290564180101, + "grad_norm": 0.6147937178611755, + "learning_rate": 9.72932676159857e-05, + "loss": 0.0423, + "step": 21702 + }, + { + "epoch": 0.5087524979928152, + "grad_norm": 0.37981927394866943, + "learning_rate": 9.728590416436962e-05, + "loss": 0.0812, + "step": 21703 + }, + { + "epoch": 0.5087759395676201, + "grad_norm": 0.14534296095371246, + "learning_rate": 9.72785407274803e-05, + "loss": 0.0184, + "step": 21704 + }, + { + "epoch": 0.5087993811424252, + "grad_norm": 0.19615846872329712, + "learning_rate": 9.727117730535768e-05, + "loss": 0.0243, + "step": 21705 + }, + { + "epoch": 0.5088228227172301, + "grad_norm": 0.4399751126766205, + "learning_rate": 9.726381389804169e-05, + "loss": 0.0913, + "step": 21706 + }, + { + "epoch": 0.5088462642920352, + "grad_norm": 0.07396851480007172, + "learning_rate": 9.725645050557239e-05, + "loss": 0.0123, + "step": 21707 + }, + { + "epoch": 0.5088697058668401, + "grad_norm": 0.5258347988128662, + "learning_rate": 9.724908712798964e-05, + "loss": 0.0821, + "step": 21708 + }, + { + "epoch": 0.5088931474416452, + "grad_norm": 0.35123202204704285, + "learning_rate": 9.724172376533342e-05, + "loss": 0.0667, + "step": 21709 + }, + { + "epoch": 0.5089165890164501, + "grad_norm": 0.3535561263561249, + "learning_rate": 9.723436041764369e-05, + "loss": 0.0783, + "step": 21710 + }, + { + "epoch": 0.5089400305912551, + "grad_norm": 0.4874369204044342, + "learning_rate": 9.722699708496041e-05, + "loss": 0.0758, + "step": 21711 + }, + { + "epoch": 0.5089634721660601, + "grad_norm": 0.12132260948419571, + "learning_rate": 9.72196337673235e-05, + "loss": 0.0313, + "step": 21712 + }, + { + "epoch": 0.5089869137408651, + "grad_norm": 0.9226230382919312, + "learning_rate": 9.721227046477294e-05, + "loss": 0.0775, + "step": 21713 + }, + { + "epoch": 0.5090103553156701, + "grad_norm": 0.7810960412025452, + "learning_rate": 9.720490717734869e-05, + "loss": 0.096, + "step": 21714 + }, + { + "epoch": 0.5090337968904751, + "grad_norm": 0.13376013934612274, + "learning_rate": 9.719754390509067e-05, + "loss": 0.0292, + "step": 21715 + }, + { + "epoch": 0.5090572384652801, + "grad_norm": 0.39316731691360474, + "learning_rate": 9.719018064803888e-05, + "loss": 0.0777, + "step": 21716 + }, + { + "epoch": 0.5090806800400851, + "grad_norm": 0.0755663737654686, + "learning_rate": 9.718281740623323e-05, + "loss": 0.0123, + "step": 21717 + }, + { + "epoch": 0.50910412161489, + "grad_norm": 0.24759964644908905, + "learning_rate": 9.717545417971369e-05, + "loss": 0.0392, + "step": 21718 + }, + { + "epoch": 0.5091275631896951, + "grad_norm": 0.4977351129055023, + "learning_rate": 9.716809096852021e-05, + "loss": 0.1082, + "step": 21719 + }, + { + "epoch": 0.5091510047645, + "grad_norm": 0.473836213350296, + "learning_rate": 9.716072777269278e-05, + "loss": 0.1257, + "step": 21720 + }, + { + "epoch": 0.5091744463393051, + "grad_norm": 0.263376921415329, + "learning_rate": 9.715336459227129e-05, + "loss": 0.0495, + "step": 21721 + }, + { + "epoch": 0.50919788791411, + "grad_norm": 0.33717960119247437, + "learning_rate": 9.714600142729574e-05, + "loss": 0.0687, + "step": 21722 + }, + { + "epoch": 0.5092213294889151, + "grad_norm": 0.6062596440315247, + "learning_rate": 9.713863827780609e-05, + "loss": 0.1181, + "step": 21723 + }, + { + "epoch": 0.50924477106372, + "grad_norm": 0.6660478711128235, + "learning_rate": 9.713127514384221e-05, + "loss": 0.1495, + "step": 21724 + }, + { + "epoch": 0.5092682126385251, + "grad_norm": 0.126527801156044, + "learning_rate": 9.712391202544416e-05, + "loss": 0.0164, + "step": 21725 + }, + { + "epoch": 0.50929165421333, + "grad_norm": 0.37094199657440186, + "learning_rate": 9.711654892265184e-05, + "loss": 0.0607, + "step": 21726 + }, + { + "epoch": 0.509315095788135, + "grad_norm": 0.5163360238075256, + "learning_rate": 9.71091858355052e-05, + "loss": 0.1011, + "step": 21727 + }, + { + "epoch": 0.50933853736294, + "grad_norm": 0.42665427923202515, + "learning_rate": 9.71018227640442e-05, + "loss": 0.1063, + "step": 21728 + }, + { + "epoch": 0.509361978937745, + "grad_norm": 0.3526286780834198, + "learning_rate": 9.709445970830878e-05, + "loss": 0.0796, + "step": 21729 + }, + { + "epoch": 0.50938542051255, + "grad_norm": 0.3246719539165497, + "learning_rate": 9.708709666833889e-05, + "loss": 0.0971, + "step": 21730 + }, + { + "epoch": 0.509408862087355, + "grad_norm": 0.2934071719646454, + "learning_rate": 9.70797336441745e-05, + "loss": 0.0586, + "step": 21731 + }, + { + "epoch": 0.50943230366216, + "grad_norm": 0.5043566823005676, + "learning_rate": 9.707237063585558e-05, + "loss": 0.1192, + "step": 21732 + }, + { + "epoch": 0.509455745236965, + "grad_norm": 0.33514928817749023, + "learning_rate": 9.706500764342206e-05, + "loss": 0.0401, + "step": 21733 + }, + { + "epoch": 0.5094791868117701, + "grad_norm": 0.29955190420150757, + "learning_rate": 9.705764466691388e-05, + "loss": 0.0946, + "step": 21734 + }, + { + "epoch": 0.509502628386575, + "grad_norm": 0.4374115467071533, + "learning_rate": 9.7050281706371e-05, + "loss": 0.0859, + "step": 21735 + }, + { + "epoch": 0.5095260699613801, + "grad_norm": 0.19424767792224884, + "learning_rate": 9.704291876183339e-05, + "loss": 0.0378, + "step": 21736 + }, + { + "epoch": 0.509549511536185, + "grad_norm": 0.5048691630363464, + "learning_rate": 9.703555583334096e-05, + "loss": 0.0824, + "step": 21737 + }, + { + "epoch": 0.50957295311099, + "grad_norm": 0.5250634551048279, + "learning_rate": 9.702819292093372e-05, + "loss": 0.1374, + "step": 21738 + }, + { + "epoch": 0.509596394685795, + "grad_norm": 0.4737042486667633, + "learning_rate": 9.702083002465158e-05, + "loss": 0.0977, + "step": 21739 + }, + { + "epoch": 0.5096198362606, + "grad_norm": 0.40462905168533325, + "learning_rate": 9.701346714453448e-05, + "loss": 0.0609, + "step": 21740 + }, + { + "epoch": 0.509643277835405, + "grad_norm": 0.5578381419181824, + "learning_rate": 9.700610428062241e-05, + "loss": 0.1486, + "step": 21741 + }, + { + "epoch": 0.50966671941021, + "grad_norm": 0.39361709356307983, + "learning_rate": 9.699874143295527e-05, + "loss": 0.1115, + "step": 21742 + }, + { + "epoch": 0.509690160985015, + "grad_norm": 0.3813949525356293, + "learning_rate": 9.699137860157307e-05, + "loss": 0.4129, + "step": 21743 + }, + { + "epoch": 0.50971360255982, + "grad_norm": 0.39959368109703064, + "learning_rate": 9.698401578651572e-05, + "loss": 0.1025, + "step": 21744 + }, + { + "epoch": 0.509737044134625, + "grad_norm": 0.7043309807777405, + "learning_rate": 9.697665298782321e-05, + "loss": 0.1781, + "step": 21745 + }, + { + "epoch": 0.50976048570943, + "grad_norm": 0.1584295779466629, + "learning_rate": 9.696929020553546e-05, + "loss": 0.0196, + "step": 21746 + }, + { + "epoch": 0.5097839272842349, + "grad_norm": 0.5870487093925476, + "learning_rate": 9.696192743969241e-05, + "loss": 0.0768, + "step": 21747 + }, + { + "epoch": 0.50980736885904, + "grad_norm": 0.35168570280075073, + "learning_rate": 9.695456469033405e-05, + "loss": 0.0796, + "step": 21748 + }, + { + "epoch": 0.5098308104338449, + "grad_norm": 0.4437965750694275, + "learning_rate": 9.694720195750032e-05, + "loss": 0.0851, + "step": 21749 + }, + { + "epoch": 0.50985425200865, + "grad_norm": 0.5234010815620422, + "learning_rate": 9.693983924123112e-05, + "loss": 0.1009, + "step": 21750 + }, + { + "epoch": 0.5098776935834549, + "grad_norm": 0.36025387048721313, + "learning_rate": 9.693247654156645e-05, + "loss": 0.0515, + "step": 21751 + }, + { + "epoch": 0.50990113515826, + "grad_norm": 0.2949362099170685, + "learning_rate": 9.692511385854626e-05, + "loss": 0.0518, + "step": 21752 + }, + { + "epoch": 0.5099245767330649, + "grad_norm": 0.43669936060905457, + "learning_rate": 9.691775119221047e-05, + "loss": 0.0915, + "step": 21753 + }, + { + "epoch": 0.50994801830787, + "grad_norm": 0.4998357892036438, + "learning_rate": 9.691038854259902e-05, + "loss": 0.0923, + "step": 21754 + }, + { + "epoch": 0.5099714598826749, + "grad_norm": 0.32806456089019775, + "learning_rate": 9.690302590975193e-05, + "loss": 0.0766, + "step": 21755 + }, + { + "epoch": 0.5099949014574799, + "grad_norm": 0.6574185490608215, + "learning_rate": 9.689566329370913e-05, + "loss": 0.1332, + "step": 21756 + }, + { + "epoch": 0.5100183430322849, + "grad_norm": 0.5213078856468201, + "learning_rate": 9.688830069451051e-05, + "loss": 0.651, + "step": 21757 + }, + { + "epoch": 0.5100417846070899, + "grad_norm": 0.4425674378871918, + "learning_rate": 9.688093811219607e-05, + "loss": 0.0276, + "step": 21758 + }, + { + "epoch": 0.5100652261818949, + "grad_norm": 0.41660600900650024, + "learning_rate": 9.687357554680574e-05, + "loss": 0.0498, + "step": 21759 + }, + { + "epoch": 0.5100886677566999, + "grad_norm": 0.3150046467781067, + "learning_rate": 9.686621299837949e-05, + "loss": 0.0377, + "step": 21760 + }, + { + "epoch": 0.5101121093315049, + "grad_norm": 0.4202827215194702, + "learning_rate": 9.685885046695726e-05, + "loss": 0.0902, + "step": 21761 + }, + { + "epoch": 0.5101355509063099, + "grad_norm": 0.24932602047920227, + "learning_rate": 9.685148795257898e-05, + "loss": 0.0364, + "step": 21762 + }, + { + "epoch": 0.5101589924811148, + "grad_norm": 0.1997404396533966, + "learning_rate": 9.684412545528464e-05, + "loss": 0.0539, + "step": 21763 + }, + { + "epoch": 0.5101824340559199, + "grad_norm": 0.31964927911758423, + "learning_rate": 9.683676297511415e-05, + "loss": 0.0756, + "step": 21764 + }, + { + "epoch": 0.5102058756307248, + "grad_norm": 0.3523429036140442, + "learning_rate": 9.682940051210746e-05, + "loss": 0.0361, + "step": 21765 + }, + { + "epoch": 0.5102293172055299, + "grad_norm": 0.2440207600593567, + "learning_rate": 9.682203806630456e-05, + "loss": 0.0501, + "step": 21766 + }, + { + "epoch": 0.5102527587803348, + "grad_norm": 0.2021532952785492, + "learning_rate": 9.681467563774531e-05, + "loss": 0.0468, + "step": 21767 + }, + { + "epoch": 0.5102762003551399, + "grad_norm": 0.8074203133583069, + "learning_rate": 9.680731322646977e-05, + "loss": 0.1453, + "step": 21768 + }, + { + "epoch": 0.5102996419299448, + "grad_norm": 0.43001607060432434, + "learning_rate": 9.679995083251782e-05, + "loss": 0.0934, + "step": 21769 + }, + { + "epoch": 0.5103230835047499, + "grad_norm": 0.391966849565506, + "learning_rate": 9.679258845592944e-05, + "loss": 0.0351, + "step": 21770 + }, + { + "epoch": 0.5103465250795548, + "grad_norm": 0.6347590684890747, + "learning_rate": 9.678522609674457e-05, + "loss": 0.1489, + "step": 21771 + }, + { + "epoch": 0.5103699666543599, + "grad_norm": 0.2933749854564667, + "learning_rate": 9.677786375500314e-05, + "loss": 0.0723, + "step": 21772 + }, + { + "epoch": 0.5103934082291648, + "grad_norm": 0.598939836025238, + "learning_rate": 9.677050143074513e-05, + "loss": 0.1082, + "step": 21773 + }, + { + "epoch": 0.5104168498039698, + "grad_norm": 0.33147814869880676, + "learning_rate": 9.676313912401046e-05, + "loss": 0.0493, + "step": 21774 + }, + { + "epoch": 0.5104402913787748, + "grad_norm": 0.724039614200592, + "learning_rate": 9.675577683483907e-05, + "loss": 0.1154, + "step": 21775 + }, + { + "epoch": 0.5104637329535798, + "grad_norm": 0.8747588992118835, + "learning_rate": 9.674841456327096e-05, + "loss": 0.0497, + "step": 21776 + }, + { + "epoch": 0.5104871745283848, + "grad_norm": 0.27316513657569885, + "learning_rate": 9.674105230934604e-05, + "loss": 0.0394, + "step": 21777 + }, + { + "epoch": 0.5105106161031898, + "grad_norm": 0.5341894626617432, + "learning_rate": 9.673369007310423e-05, + "loss": 0.1499, + "step": 21778 + }, + { + "epoch": 0.5105340576779948, + "grad_norm": 0.46586981415748596, + "learning_rate": 9.67263278545855e-05, + "loss": 0.1041, + "step": 21779 + }, + { + "epoch": 0.5105574992527998, + "grad_norm": 0.26502877473831177, + "learning_rate": 9.671896565382984e-05, + "loss": 0.0557, + "step": 21780 + }, + { + "epoch": 0.5105809408276047, + "grad_norm": 0.28668129444122314, + "learning_rate": 9.671160347087717e-05, + "loss": 0.0415, + "step": 21781 + }, + { + "epoch": 0.5106043824024098, + "grad_norm": 0.35103368759155273, + "learning_rate": 9.67042413057674e-05, + "loss": 0.0691, + "step": 21782 + }, + { + "epoch": 0.5106278239772147, + "grad_norm": 0.2648119032382965, + "learning_rate": 9.669687915854055e-05, + "loss": 0.0277, + "step": 21783 + }, + { + "epoch": 0.5106512655520198, + "grad_norm": 0.5962219834327698, + "learning_rate": 9.66895170292365e-05, + "loss": 0.1579, + "step": 21784 + }, + { + "epoch": 0.5106747071268248, + "grad_norm": 0.2433043271303177, + "learning_rate": 9.668215491789522e-05, + "loss": 0.0609, + "step": 21785 + }, + { + "epoch": 0.5106981487016298, + "grad_norm": 0.40980255603790283, + "learning_rate": 9.667479282455669e-05, + "loss": 0.0616, + "step": 21786 + }, + { + "epoch": 0.5107215902764348, + "grad_norm": 0.5960814952850342, + "learning_rate": 9.666743074926082e-05, + "loss": 0.1766, + "step": 21787 + }, + { + "epoch": 0.5107450318512398, + "grad_norm": 0.851340651512146, + "learning_rate": 9.666006869204755e-05, + "loss": 0.137, + "step": 21788 + }, + { + "epoch": 0.5107684734260448, + "grad_norm": 0.1499423235654831, + "learning_rate": 9.665270665295685e-05, + "loss": 0.0418, + "step": 21789 + }, + { + "epoch": 0.5107919150008497, + "grad_norm": 0.273145854473114, + "learning_rate": 9.664534463202864e-05, + "loss": 0.0339, + "step": 21790 + }, + { + "epoch": 0.5108153565756548, + "grad_norm": 0.3254678249359131, + "learning_rate": 9.663798262930288e-05, + "loss": 0.0649, + "step": 21791 + }, + { + "epoch": 0.5108387981504597, + "grad_norm": 0.4400757849216461, + "learning_rate": 9.663062064481953e-05, + "loss": 0.0783, + "step": 21792 + }, + { + "epoch": 0.5108622397252648, + "grad_norm": 0.5210138559341431, + "learning_rate": 9.662325867861855e-05, + "loss": 0.0703, + "step": 21793 + }, + { + "epoch": 0.5108856813000697, + "grad_norm": 0.2861093282699585, + "learning_rate": 9.661589673073986e-05, + "loss": 0.0746, + "step": 21794 + }, + { + "epoch": 0.5109091228748748, + "grad_norm": 0.09748425334692001, + "learning_rate": 9.660853480122337e-05, + "loss": 0.0267, + "step": 21795 + }, + { + "epoch": 0.5109325644496797, + "grad_norm": 0.41261643171310425, + "learning_rate": 9.660117289010911e-05, + "loss": 0.106, + "step": 21796 + }, + { + "epoch": 0.5109560060244848, + "grad_norm": 0.7173625826835632, + "learning_rate": 9.659381099743695e-05, + "loss": 0.1732, + "step": 21797 + }, + { + "epoch": 0.5109794475992897, + "grad_norm": 0.1292707622051239, + "learning_rate": 9.65864491232469e-05, + "loss": 0.0345, + "step": 21798 + }, + { + "epoch": 0.5110028891740948, + "grad_norm": 0.29107213020324707, + "learning_rate": 9.657908726757885e-05, + "loss": 0.0748, + "step": 21799 + }, + { + "epoch": 0.5110263307488997, + "grad_norm": 0.1206681877374649, + "learning_rate": 9.657172543047275e-05, + "loss": 0.0124, + "step": 21800 + }, + { + "epoch": 0.5110497723237047, + "grad_norm": 0.2644123136997223, + "learning_rate": 9.656436361196861e-05, + "loss": 0.0308, + "step": 21801 + }, + { + "epoch": 0.5110732138985097, + "grad_norm": 1.2769867181777954, + "learning_rate": 9.655700181210632e-05, + "loss": 0.3448, + "step": 21802 + }, + { + "epoch": 0.5110966554733147, + "grad_norm": 0.8695811033248901, + "learning_rate": 9.654964003092579e-05, + "loss": 0.1582, + "step": 21803 + }, + { + "epoch": 0.5111200970481197, + "grad_norm": 0.43115007877349854, + "learning_rate": 9.654227826846702e-05, + "loss": 0.123, + "step": 21804 + }, + { + "epoch": 0.5111435386229247, + "grad_norm": 0.3560348153114319, + "learning_rate": 9.653491652476999e-05, + "loss": 0.0949, + "step": 21805 + }, + { + "epoch": 0.5111669801977297, + "grad_norm": 0.45856356620788574, + "learning_rate": 9.652755479987459e-05, + "loss": 0.0872, + "step": 21806 + }, + { + "epoch": 0.5111904217725347, + "grad_norm": 0.40693122148513794, + "learning_rate": 9.652019309382075e-05, + "loss": 0.1013, + "step": 21807 + }, + { + "epoch": 0.5112138633473396, + "grad_norm": 0.4017747640609741, + "learning_rate": 9.651283140664846e-05, + "loss": 0.0938, + "step": 21808 + }, + { + "epoch": 0.5112373049221447, + "grad_norm": 0.2399645745754242, + "learning_rate": 9.650546973839765e-05, + "loss": 0.0364, + "step": 21809 + }, + { + "epoch": 0.5112607464969496, + "grad_norm": 0.49756142497062683, + "learning_rate": 9.649810808910823e-05, + "loss": 0.151, + "step": 21810 + }, + { + "epoch": 0.5112841880717547, + "grad_norm": 0.16671058535575867, + "learning_rate": 9.64907464588202e-05, + "loss": 0.0355, + "step": 21811 + }, + { + "epoch": 0.5113076296465596, + "grad_norm": 0.5257607102394104, + "learning_rate": 9.648338484757349e-05, + "loss": 0.1242, + "step": 21812 + }, + { + "epoch": 0.5113310712213647, + "grad_norm": 0.2072737216949463, + "learning_rate": 9.647602325540799e-05, + "loss": 0.0389, + "step": 21813 + }, + { + "epoch": 0.5113545127961696, + "grad_norm": 0.47317415475845337, + "learning_rate": 9.646866168236373e-05, + "loss": 0.1069, + "step": 21814 + }, + { + "epoch": 0.5113779543709747, + "grad_norm": 0.6057534217834473, + "learning_rate": 9.646130012848055e-05, + "loss": 0.7162, + "step": 21815 + }, + { + "epoch": 0.5114013959457796, + "grad_norm": 0.3331948220729828, + "learning_rate": 9.645393859379851e-05, + "loss": 0.052, + "step": 21816 + }, + { + "epoch": 0.5114248375205847, + "grad_norm": 0.2684746980667114, + "learning_rate": 9.644657707835748e-05, + "loss": 0.0492, + "step": 21817 + }, + { + "epoch": 0.5114482790953896, + "grad_norm": 0.3750413656234741, + "learning_rate": 9.643921558219743e-05, + "loss": 0.0711, + "step": 21818 + }, + { + "epoch": 0.5114717206701946, + "grad_norm": 0.24618539214134216, + "learning_rate": 9.64318541053583e-05, + "loss": 0.0257, + "step": 21819 + }, + { + "epoch": 0.5114951622449996, + "grad_norm": 0.4266427457332611, + "learning_rate": 9.642449264788001e-05, + "loss": 0.1157, + "step": 21820 + }, + { + "epoch": 0.5115186038198046, + "grad_norm": 0.6240895390510559, + "learning_rate": 9.641713120980255e-05, + "loss": 0.1075, + "step": 21821 + }, + { + "epoch": 0.5115420453946096, + "grad_norm": 0.640593945980072, + "learning_rate": 9.640976979116583e-05, + "loss": 0.1639, + "step": 21822 + }, + { + "epoch": 0.5115654869694146, + "grad_norm": 0.7230424284934998, + "learning_rate": 9.640240839200977e-05, + "loss": 0.155, + "step": 21823 + }, + { + "epoch": 0.5115889285442196, + "grad_norm": 0.5673004388809204, + "learning_rate": 9.639504701237438e-05, + "loss": 0.1246, + "step": 21824 + }, + { + "epoch": 0.5116123701190246, + "grad_norm": 0.4247950613498688, + "learning_rate": 9.638768565229957e-05, + "loss": 0.0766, + "step": 21825 + }, + { + "epoch": 0.5116358116938295, + "grad_norm": 0.19916580617427826, + "learning_rate": 9.638032431182526e-05, + "loss": 0.0398, + "step": 21826 + }, + { + "epoch": 0.5116592532686346, + "grad_norm": 0.25327202677726746, + "learning_rate": 9.637296299099138e-05, + "loss": 0.0761, + "step": 21827 + }, + { + "epoch": 0.5116826948434395, + "grad_norm": 0.6727803349494934, + "learning_rate": 9.636560168983794e-05, + "loss": 0.636, + "step": 21828 + }, + { + "epoch": 0.5117061364182446, + "grad_norm": 0.2184504121541977, + "learning_rate": 9.635824040840487e-05, + "loss": 0.0256, + "step": 21829 + }, + { + "epoch": 0.5117295779930495, + "grad_norm": 0.47585558891296387, + "learning_rate": 9.635087914673204e-05, + "loss": 0.0802, + "step": 21830 + }, + { + "epoch": 0.5117530195678546, + "grad_norm": 0.39917972683906555, + "learning_rate": 9.634351790485948e-05, + "loss": 0.0641, + "step": 21831 + }, + { + "epoch": 0.5117764611426595, + "grad_norm": 0.17239880561828613, + "learning_rate": 9.63361566828271e-05, + "loss": 0.0313, + "step": 21832 + }, + { + "epoch": 0.5117999027174646, + "grad_norm": 0.41547688841819763, + "learning_rate": 9.632879548067481e-05, + "loss": 0.1323, + "step": 21833 + }, + { + "epoch": 0.5118233442922695, + "grad_norm": 0.2044031322002411, + "learning_rate": 9.632143429844259e-05, + "loss": 0.0625, + "step": 21834 + }, + { + "epoch": 0.5118467858670745, + "grad_norm": 0.1473008692264557, + "learning_rate": 9.631407313617036e-05, + "loss": 0.0261, + "step": 21835 + }, + { + "epoch": 0.5118702274418796, + "grad_norm": 0.5454517602920532, + "learning_rate": 9.63067119938981e-05, + "loss": 0.1468, + "step": 21836 + }, + { + "epoch": 0.5118936690166845, + "grad_norm": 0.10431485623121262, + "learning_rate": 9.629935087166571e-05, + "loss": 0.0158, + "step": 21837 + }, + { + "epoch": 0.5119171105914896, + "grad_norm": 0.4189643859863281, + "learning_rate": 9.629198976951314e-05, + "loss": 0.1086, + "step": 21838 + }, + { + "epoch": 0.5119405521662945, + "grad_norm": 0.29713237285614014, + "learning_rate": 9.628462868748035e-05, + "loss": 0.0439, + "step": 21839 + }, + { + "epoch": 0.5119639937410996, + "grad_norm": 0.39629456400871277, + "learning_rate": 9.627726762560722e-05, + "loss": 0.6709, + "step": 21840 + }, + { + "epoch": 0.5119874353159045, + "grad_norm": 0.15811733901500702, + "learning_rate": 9.62699065839338e-05, + "loss": 0.0172, + "step": 21841 + }, + { + "epoch": 0.5120108768907096, + "grad_norm": 0.21676874160766602, + "learning_rate": 9.626254556249994e-05, + "loss": 0.0531, + "step": 21842 + }, + { + "epoch": 0.5120343184655145, + "grad_norm": 0.3801862299442291, + "learning_rate": 9.625518456134564e-05, + "loss": 0.469, + "step": 21843 + }, + { + "epoch": 0.5120577600403196, + "grad_norm": 0.3345971405506134, + "learning_rate": 9.624782358051081e-05, + "loss": 0.0629, + "step": 21844 + }, + { + "epoch": 0.5120812016151245, + "grad_norm": 0.5351186990737915, + "learning_rate": 9.624046262003537e-05, + "loss": 0.7693, + "step": 21845 + }, + { + "epoch": 0.5121046431899295, + "grad_norm": 0.3573726713657379, + "learning_rate": 9.62331016799593e-05, + "loss": 0.0752, + "step": 21846 + }, + { + "epoch": 0.5121280847647345, + "grad_norm": 0.40744370222091675, + "learning_rate": 9.622574076032254e-05, + "loss": 0.0823, + "step": 21847 + }, + { + "epoch": 0.5121515263395395, + "grad_norm": 0.1716766357421875, + "learning_rate": 9.6218379861165e-05, + "loss": 0.0448, + "step": 21848 + }, + { + "epoch": 0.5121749679143445, + "grad_norm": 0.6053537726402283, + "learning_rate": 9.621101898252664e-05, + "loss": 0.0737, + "step": 21849 + }, + { + "epoch": 0.5121984094891495, + "grad_norm": 0.48415979743003845, + "learning_rate": 9.620365812444741e-05, + "loss": 0.5606, + "step": 21850 + }, + { + "epoch": 0.5122218510639545, + "grad_norm": 0.43910446763038635, + "learning_rate": 9.619629728696722e-05, + "loss": 0.0792, + "step": 21851 + }, + { + "epoch": 0.5122452926387595, + "grad_norm": 0.4184935688972473, + "learning_rate": 9.6188936470126e-05, + "loss": 0.0518, + "step": 21852 + }, + { + "epoch": 0.5122687342135644, + "grad_norm": 0.44248807430267334, + "learning_rate": 9.618157567396376e-05, + "loss": 0.1054, + "step": 21853 + }, + { + "epoch": 0.5122921757883695, + "grad_norm": 0.4689567983150482, + "learning_rate": 9.61742148985204e-05, + "loss": 0.0903, + "step": 21854 + }, + { + "epoch": 0.5123156173631744, + "grad_norm": 0.4448280334472656, + "learning_rate": 9.616685414383583e-05, + "loss": 0.0995, + "step": 21855 + }, + { + "epoch": 0.5123390589379795, + "grad_norm": 0.5588496327400208, + "learning_rate": 9.615949340995005e-05, + "loss": 0.1191, + "step": 21856 + }, + { + "epoch": 0.5123625005127844, + "grad_norm": 0.48669853806495667, + "learning_rate": 9.615213269690296e-05, + "loss": 0.0714, + "step": 21857 + }, + { + "epoch": 0.5123859420875895, + "grad_norm": 0.1496635526418686, + "learning_rate": 9.614477200473448e-05, + "loss": 0.0173, + "step": 21858 + }, + { + "epoch": 0.5124093836623944, + "grad_norm": 0.24065208435058594, + "learning_rate": 9.613741133348461e-05, + "loss": 0.0344, + "step": 21859 + }, + { + "epoch": 0.5124328252371995, + "grad_norm": 0.14079709351062775, + "learning_rate": 9.613005068319324e-05, + "loss": 0.0367, + "step": 21860 + }, + { + "epoch": 0.5124562668120044, + "grad_norm": 0.15147826075553894, + "learning_rate": 9.61226900539003e-05, + "loss": 0.0288, + "step": 21861 + }, + { + "epoch": 0.5124797083868095, + "grad_norm": 0.5088819861412048, + "learning_rate": 9.61153294456458e-05, + "loss": 0.049, + "step": 21862 + }, + { + "epoch": 0.5125031499616144, + "grad_norm": 0.5828189849853516, + "learning_rate": 9.610796885846961e-05, + "loss": 0.6472, + "step": 21863 + }, + { + "epoch": 0.5125265915364194, + "grad_norm": 0.5677000284194946, + "learning_rate": 9.610060829241166e-05, + "loss": 0.0831, + "step": 21864 + }, + { + "epoch": 0.5125500331112244, + "grad_norm": 0.28964439034461975, + "learning_rate": 9.609324774751192e-05, + "loss": 0.0518, + "step": 21865 + }, + { + "epoch": 0.5125734746860294, + "grad_norm": 0.33425477147102356, + "learning_rate": 9.608588722381037e-05, + "loss": 0.0571, + "step": 21866 + }, + { + "epoch": 0.5125969162608344, + "grad_norm": 0.1903677135705948, + "learning_rate": 9.60785267213469e-05, + "loss": 0.0401, + "step": 21867 + }, + { + "epoch": 0.5126203578356394, + "grad_norm": 0.5568642616271973, + "learning_rate": 9.607116624016143e-05, + "loss": 0.0804, + "step": 21868 + }, + { + "epoch": 0.5126437994104444, + "grad_norm": 0.3191477954387665, + "learning_rate": 9.606380578029395e-05, + "loss": 0.0516, + "step": 21869 + }, + { + "epoch": 0.5126672409852494, + "grad_norm": 0.44284534454345703, + "learning_rate": 9.605644534178437e-05, + "loss": 0.1266, + "step": 21870 + }, + { + "epoch": 0.5126906825600543, + "grad_norm": 0.303489625453949, + "learning_rate": 9.604908492467259e-05, + "loss": 0.0499, + "step": 21871 + }, + { + "epoch": 0.5127141241348594, + "grad_norm": 0.12732048332691193, + "learning_rate": 9.604172452899862e-05, + "loss": 0.0155, + "step": 21872 + }, + { + "epoch": 0.5127375657096643, + "grad_norm": 0.2899143099784851, + "learning_rate": 9.603436415480235e-05, + "loss": 0.0519, + "step": 21873 + }, + { + "epoch": 0.5127610072844694, + "grad_norm": 0.11124575883150101, + "learning_rate": 9.602700380212374e-05, + "loss": 0.0266, + "step": 21874 + }, + { + "epoch": 0.5127844488592743, + "grad_norm": 0.3652481734752655, + "learning_rate": 9.601964347100272e-05, + "loss": 0.0711, + "step": 21875 + }, + { + "epoch": 0.5128078904340794, + "grad_norm": 0.2999521791934967, + "learning_rate": 9.601228316147918e-05, + "loss": 0.0732, + "step": 21876 + }, + { + "epoch": 0.5128313320088843, + "grad_norm": 0.17804226279258728, + "learning_rate": 9.600492287359314e-05, + "loss": 0.0336, + "step": 21877 + }, + { + "epoch": 0.5128547735836894, + "grad_norm": 0.49842894077301025, + "learning_rate": 9.599756260738452e-05, + "loss": 0.1475, + "step": 21878 + }, + { + "epoch": 0.5128782151584943, + "grad_norm": 0.5756306052207947, + "learning_rate": 9.599020236289323e-05, + "loss": 0.1102, + "step": 21879 + }, + { + "epoch": 0.5129016567332993, + "grad_norm": 0.5822443962097168, + "learning_rate": 9.59828421401592e-05, + "loss": 0.1214, + "step": 21880 + }, + { + "epoch": 0.5129250983081043, + "grad_norm": 0.13042955100536346, + "learning_rate": 9.59754819392224e-05, + "loss": 0.025, + "step": 21881 + }, + { + "epoch": 0.5129485398829093, + "grad_norm": 0.5900571942329407, + "learning_rate": 9.596812176012275e-05, + "loss": 0.126, + "step": 21882 + }, + { + "epoch": 0.5129719814577143, + "grad_norm": 0.9619256854057312, + "learning_rate": 9.596076160290017e-05, + "loss": 0.0624, + "step": 21883 + }, + { + "epoch": 0.5129954230325193, + "grad_norm": 0.44478994607925415, + "learning_rate": 9.595340146759462e-05, + "loss": 0.0793, + "step": 21884 + }, + { + "epoch": 0.5130188646073243, + "grad_norm": 0.5415692925453186, + "learning_rate": 9.594604135424603e-05, + "loss": 0.0537, + "step": 21885 + }, + { + "epoch": 0.5130423061821293, + "grad_norm": 0.5932862162590027, + "learning_rate": 9.593868126289431e-05, + "loss": 0.0517, + "step": 21886 + }, + { + "epoch": 0.5130657477569344, + "grad_norm": 0.551578938961029, + "learning_rate": 9.593132119357945e-05, + "loss": 0.1877, + "step": 21887 + }, + { + "epoch": 0.5130891893317393, + "grad_norm": 0.2841675877571106, + "learning_rate": 9.59239611463413e-05, + "loss": 0.0633, + "step": 21888 + }, + { + "epoch": 0.5131126309065444, + "grad_norm": 0.5609574913978577, + "learning_rate": 9.591660112121991e-05, + "loss": 0.1114, + "step": 21889 + }, + { + "epoch": 0.5131360724813493, + "grad_norm": 0.2106209546327591, + "learning_rate": 9.590924111825512e-05, + "loss": 0.0411, + "step": 21890 + }, + { + "epoch": 0.5131595140561543, + "grad_norm": 0.4497002363204956, + "learning_rate": 9.590188113748692e-05, + "loss": 0.0557, + "step": 21891 + }, + { + "epoch": 0.5131829556309593, + "grad_norm": 0.5571048855781555, + "learning_rate": 9.589452117895526e-05, + "loss": 0.6791, + "step": 21892 + }, + { + "epoch": 0.5132063972057643, + "grad_norm": 0.4876849055290222, + "learning_rate": 9.588716124269999e-05, + "loss": 0.1291, + "step": 21893 + }, + { + "epoch": 0.5132298387805693, + "grad_norm": 0.4380270838737488, + "learning_rate": 9.587980132876113e-05, + "loss": 0.0836, + "step": 21894 + }, + { + "epoch": 0.5132532803553743, + "grad_norm": 0.2774110436439514, + "learning_rate": 9.587244143717859e-05, + "loss": 0.0384, + "step": 21895 + }, + { + "epoch": 0.5132767219301793, + "grad_norm": 0.10469508171081543, + "learning_rate": 9.586508156799227e-05, + "loss": 0.0179, + "step": 21896 + }, + { + "epoch": 0.5133001635049843, + "grad_norm": 0.903109610080719, + "learning_rate": 9.585772172124216e-05, + "loss": 0.1007, + "step": 21897 + }, + { + "epoch": 0.5133236050797892, + "grad_norm": 0.3931778073310852, + "learning_rate": 9.585036189696816e-05, + "loss": 0.0878, + "step": 21898 + }, + { + "epoch": 0.5133470466545943, + "grad_norm": 0.3470735251903534, + "learning_rate": 9.58430020952102e-05, + "loss": 0.0647, + "step": 21899 + }, + { + "epoch": 0.5133704882293992, + "grad_norm": 0.13266132771968842, + "learning_rate": 9.583564231600821e-05, + "loss": 0.0313, + "step": 21900 + }, + { + "epoch": 0.5133939298042043, + "grad_norm": 0.19125109910964966, + "learning_rate": 9.582828255940218e-05, + "loss": 0.0296, + "step": 21901 + }, + { + "epoch": 0.5134173713790092, + "grad_norm": 0.4485977590084076, + "learning_rate": 9.5820922825432e-05, + "loss": 0.0796, + "step": 21902 + }, + { + "epoch": 0.5134408129538143, + "grad_norm": 0.5615864992141724, + "learning_rate": 9.58135631141376e-05, + "loss": 0.7105, + "step": 21903 + }, + { + "epoch": 0.5134642545286192, + "grad_norm": 0.5056420564651489, + "learning_rate": 9.580620342555893e-05, + "loss": 0.12, + "step": 21904 + }, + { + "epoch": 0.5134876961034243, + "grad_norm": 0.4349902272224426, + "learning_rate": 9.579884375973593e-05, + "loss": 0.0687, + "step": 21905 + }, + { + "epoch": 0.5135111376782292, + "grad_norm": 0.5285981297492981, + "learning_rate": 9.57914841167085e-05, + "loss": 0.1415, + "step": 21906 + }, + { + "epoch": 0.5135345792530343, + "grad_norm": 0.6097385287284851, + "learning_rate": 9.578412449651661e-05, + "loss": 0.0821, + "step": 21907 + }, + { + "epoch": 0.5135580208278392, + "grad_norm": 0.5580295324325562, + "learning_rate": 9.577676489920017e-05, + "loss": 0.1038, + "step": 21908 + }, + { + "epoch": 0.5135814624026442, + "grad_norm": 0.28552910685539246, + "learning_rate": 9.576940532479914e-05, + "loss": 0.0639, + "step": 21909 + }, + { + "epoch": 0.5136049039774492, + "grad_norm": 0.7145234942436218, + "learning_rate": 9.576204577335343e-05, + "loss": 0.1545, + "step": 21910 + }, + { + "epoch": 0.5136283455522542, + "grad_norm": 0.5442980527877808, + "learning_rate": 9.575468624490296e-05, + "loss": 0.1716, + "step": 21911 + }, + { + "epoch": 0.5136517871270592, + "grad_norm": 0.633125364780426, + "learning_rate": 9.574732673948767e-05, + "loss": 0.1529, + "step": 21912 + }, + { + "epoch": 0.5136752287018642, + "grad_norm": 0.1390790343284607, + "learning_rate": 9.573996725714752e-05, + "loss": 0.0273, + "step": 21913 + }, + { + "epoch": 0.5136986702766692, + "grad_norm": 0.4255453050136566, + "learning_rate": 9.573260779792245e-05, + "loss": 0.0904, + "step": 21914 + }, + { + "epoch": 0.5137221118514742, + "grad_norm": 0.152167409658432, + "learning_rate": 9.572524836185235e-05, + "loss": 0.0226, + "step": 21915 + }, + { + "epoch": 0.5137455534262791, + "grad_norm": 0.3187344968318939, + "learning_rate": 9.571788894897719e-05, + "loss": 0.0722, + "step": 21916 + }, + { + "epoch": 0.5137689950010842, + "grad_norm": 0.4427192807197571, + "learning_rate": 9.571052955933687e-05, + "loss": 0.0373, + "step": 21917 + }, + { + "epoch": 0.5137924365758891, + "grad_norm": 0.20565930008888245, + "learning_rate": 9.570317019297133e-05, + "loss": 0.0343, + "step": 21918 + }, + { + "epoch": 0.5138158781506942, + "grad_norm": 0.3614037036895752, + "learning_rate": 9.569581084992053e-05, + "loss": 0.0763, + "step": 21919 + }, + { + "epoch": 0.5138393197254991, + "grad_norm": 0.23953412473201752, + "learning_rate": 9.568845153022439e-05, + "loss": 0.018, + "step": 21920 + }, + { + "epoch": 0.5138627613003042, + "grad_norm": 0.21602538228034973, + "learning_rate": 9.568109223392279e-05, + "loss": 0.0551, + "step": 21921 + }, + { + "epoch": 0.5138862028751091, + "grad_norm": 0.540074348449707, + "learning_rate": 9.567373296105574e-05, + "loss": 0.1367, + "step": 21922 + }, + { + "epoch": 0.5139096444499142, + "grad_norm": 0.3568074107170105, + "learning_rate": 9.566637371166314e-05, + "loss": 0.0526, + "step": 21923 + }, + { + "epoch": 0.5139330860247191, + "grad_norm": 0.3675135672092438, + "learning_rate": 9.565901448578489e-05, + "loss": 0.0756, + "step": 21924 + }, + { + "epoch": 0.5139565275995241, + "grad_norm": 0.45237624645233154, + "learning_rate": 9.565165528346093e-05, + "loss": 0.0902, + "step": 21925 + }, + { + "epoch": 0.5139799691743291, + "grad_norm": 0.5271822214126587, + "learning_rate": 9.564429610473125e-05, + "loss": 0.0571, + "step": 21926 + }, + { + "epoch": 0.5140034107491341, + "grad_norm": 1.1698460578918457, + "learning_rate": 9.563693694963575e-05, + "loss": 0.5816, + "step": 21927 + }, + { + "epoch": 0.5140268523239391, + "grad_norm": 0.5377801656723022, + "learning_rate": 9.562957781821433e-05, + "loss": 0.1205, + "step": 21928 + }, + { + "epoch": 0.5140502938987441, + "grad_norm": 0.4273894429206848, + "learning_rate": 9.562221871050697e-05, + "loss": 0.1295, + "step": 21929 + }, + { + "epoch": 0.5140737354735491, + "grad_norm": 0.5168319940567017, + "learning_rate": 9.561485962655356e-05, + "loss": 0.108, + "step": 21930 + }, + { + "epoch": 0.5140971770483541, + "grad_norm": 0.7136479020118713, + "learning_rate": 9.560750056639402e-05, + "loss": 0.1814, + "step": 21931 + }, + { + "epoch": 0.514120618623159, + "grad_norm": 0.12631440162658691, + "learning_rate": 9.560014153006834e-05, + "loss": 0.02, + "step": 21932 + }, + { + "epoch": 0.5141440601979641, + "grad_norm": 0.2859966456890106, + "learning_rate": 9.559278251761641e-05, + "loss": 0.2005, + "step": 21933 + }, + { + "epoch": 0.514167501772769, + "grad_norm": 0.5987848043441772, + "learning_rate": 9.558542352907816e-05, + "loss": 0.5925, + "step": 21934 + }, + { + "epoch": 0.5141909433475741, + "grad_norm": 0.5798603296279907, + "learning_rate": 9.557806456449352e-05, + "loss": 0.1136, + "step": 21935 + }, + { + "epoch": 0.514214384922379, + "grad_norm": 0.510867714881897, + "learning_rate": 9.557070562390244e-05, + "loss": 0.0925, + "step": 21936 + }, + { + "epoch": 0.5142378264971841, + "grad_norm": 0.5033459663391113, + "learning_rate": 9.556334670734479e-05, + "loss": 0.6656, + "step": 21937 + }, + { + "epoch": 0.5142612680719891, + "grad_norm": 0.46104246377944946, + "learning_rate": 9.555598781486058e-05, + "loss": 0.1236, + "step": 21938 + }, + { + "epoch": 0.5142847096467941, + "grad_norm": 0.21990618109703064, + "learning_rate": 9.554862894648971e-05, + "loss": 0.0452, + "step": 21939 + }, + { + "epoch": 0.5143081512215991, + "grad_norm": 0.49152177572250366, + "learning_rate": 9.554127010227212e-05, + "loss": 0.0393, + "step": 21940 + }, + { + "epoch": 0.5143315927964041, + "grad_norm": 0.602749228477478, + "learning_rate": 9.55339112822477e-05, + "loss": 0.1341, + "step": 21941 + }, + { + "epoch": 0.5143550343712091, + "grad_norm": 0.4456186592578888, + "learning_rate": 9.552655248645641e-05, + "loss": 0.1066, + "step": 21942 + }, + { + "epoch": 0.514378475946014, + "grad_norm": 0.13313353061676025, + "learning_rate": 9.551919371493818e-05, + "loss": 0.0224, + "step": 21943 + }, + { + "epoch": 0.5144019175208191, + "grad_norm": 0.36984801292419434, + "learning_rate": 9.551183496773291e-05, + "loss": 0.0534, + "step": 21944 + }, + { + "epoch": 0.514425359095624, + "grad_norm": 0.36849266290664673, + "learning_rate": 9.550447624488057e-05, + "loss": 0.0382, + "step": 21945 + }, + { + "epoch": 0.5144488006704291, + "grad_norm": 0.3122164309024811, + "learning_rate": 9.549711754642104e-05, + "loss": 0.073, + "step": 21946 + }, + { + "epoch": 0.514472242245234, + "grad_norm": 0.28140631318092346, + "learning_rate": 9.54897588723943e-05, + "loss": 0.0349, + "step": 21947 + }, + { + "epoch": 0.5144956838200391, + "grad_norm": 0.11264284700155258, + "learning_rate": 9.548240022284028e-05, + "loss": 0.0178, + "step": 21948 + }, + { + "epoch": 0.514519125394844, + "grad_norm": 0.15815968811511993, + "learning_rate": 9.54750415977988e-05, + "loss": 0.0251, + "step": 21949 + }, + { + "epoch": 0.5145425669696491, + "grad_norm": 0.32260164618492126, + "learning_rate": 9.546768299730995e-05, + "loss": 0.0809, + "step": 21950 + }, + { + "epoch": 0.514566008544454, + "grad_norm": 0.7337551712989807, + "learning_rate": 9.546032442141353e-05, + "loss": 0.0727, + "step": 21951 + }, + { + "epoch": 0.514589450119259, + "grad_norm": 0.3582892417907715, + "learning_rate": 9.545296587014957e-05, + "loss": 0.3861, + "step": 21952 + }, + { + "epoch": 0.514612891694064, + "grad_norm": 0.6451979875564575, + "learning_rate": 9.54456073435579e-05, + "loss": 0.0965, + "step": 21953 + }, + { + "epoch": 0.514636333268869, + "grad_norm": 0.09992896020412445, + "learning_rate": 9.543824884167852e-05, + "loss": 0.0235, + "step": 21954 + }, + { + "epoch": 0.514659774843674, + "grad_norm": 0.11875538527965546, + "learning_rate": 9.543089036455133e-05, + "loss": 0.0126, + "step": 21955 + }, + { + "epoch": 0.514683216418479, + "grad_norm": 0.1800742894411087, + "learning_rate": 9.542353191221623e-05, + "loss": 0.0157, + "step": 21956 + }, + { + "epoch": 0.514706657993284, + "grad_norm": 0.5195176601409912, + "learning_rate": 9.541617348471322e-05, + "loss": 0.0729, + "step": 21957 + }, + { + "epoch": 0.514730099568089, + "grad_norm": 0.30189791321754456, + "learning_rate": 9.540881508208216e-05, + "loss": 0.0725, + "step": 21958 + }, + { + "epoch": 0.514753541142894, + "grad_norm": 0.27221184968948364, + "learning_rate": 9.540145670436298e-05, + "loss": 0.0244, + "step": 21959 + }, + { + "epoch": 0.514776982717699, + "grad_norm": 0.4954189360141754, + "learning_rate": 9.539409835159564e-05, + "loss": 0.4543, + "step": 21960 + }, + { + "epoch": 0.5148004242925039, + "grad_norm": 0.5433716773986816, + "learning_rate": 9.538674002382003e-05, + "loss": 0.0878, + "step": 21961 + }, + { + "epoch": 0.514823865867309, + "grad_norm": 0.5154224634170532, + "learning_rate": 9.537938172107612e-05, + "loss": 0.1027, + "step": 21962 + }, + { + "epoch": 0.5148473074421139, + "grad_norm": 0.14414438605308533, + "learning_rate": 9.537202344340379e-05, + "loss": 0.0313, + "step": 21963 + }, + { + "epoch": 0.514870749016919, + "grad_norm": 0.4489872455596924, + "learning_rate": 9.536466519084303e-05, + "loss": 0.0722, + "step": 21964 + }, + { + "epoch": 0.5148941905917239, + "grad_norm": 0.12259652465581894, + "learning_rate": 9.535730696343372e-05, + "loss": 0.0237, + "step": 21965 + }, + { + "epoch": 0.514917632166529, + "grad_norm": 0.1273145228624344, + "learning_rate": 9.534994876121578e-05, + "loss": 0.024, + "step": 21966 + }, + { + "epoch": 0.5149410737413339, + "grad_norm": 0.17660737037658691, + "learning_rate": 9.534259058422915e-05, + "loss": 0.0367, + "step": 21967 + }, + { + "epoch": 0.514964515316139, + "grad_norm": 0.34391796588897705, + "learning_rate": 9.533523243251377e-05, + "loss": 0.0651, + "step": 21968 + }, + { + "epoch": 0.5149879568909439, + "grad_norm": 0.6168725490570068, + "learning_rate": 9.532787430610952e-05, + "loss": 0.0921, + "step": 21969 + }, + { + "epoch": 0.515011398465749, + "grad_norm": 0.509727418422699, + "learning_rate": 9.532051620505639e-05, + "loss": 0.0594, + "step": 21970 + }, + { + "epoch": 0.5150348400405539, + "grad_norm": 0.5259776711463928, + "learning_rate": 9.531315812939426e-05, + "loss": 0.0768, + "step": 21971 + }, + { + "epoch": 0.5150582816153589, + "grad_norm": 0.499076783657074, + "learning_rate": 9.530580007916305e-05, + "loss": 0.0735, + "step": 21972 + }, + { + "epoch": 0.5150817231901639, + "grad_norm": 0.6215426921844482, + "learning_rate": 9.529844205440267e-05, + "loss": 0.1144, + "step": 21973 + }, + { + "epoch": 0.5151051647649689, + "grad_norm": 0.6763386726379395, + "learning_rate": 9.529108405515313e-05, + "loss": 0.1143, + "step": 21974 + }, + { + "epoch": 0.5151286063397739, + "grad_norm": 0.2998329699039459, + "learning_rate": 9.52837260814543e-05, + "loss": 0.0726, + "step": 21975 + }, + { + "epoch": 0.5151520479145789, + "grad_norm": 0.5786031484603882, + "learning_rate": 9.527636813334607e-05, + "loss": 0.0891, + "step": 21976 + }, + { + "epoch": 0.5151754894893839, + "grad_norm": 0.58368319272995, + "learning_rate": 9.526901021086843e-05, + "loss": 0.0741, + "step": 21977 + }, + { + "epoch": 0.5151989310641889, + "grad_norm": 0.6288090944290161, + "learning_rate": 9.526165231406128e-05, + "loss": 0.6222, + "step": 21978 + }, + { + "epoch": 0.5152223726389938, + "grad_norm": 0.5144168734550476, + "learning_rate": 9.52542944429645e-05, + "loss": 0.1077, + "step": 21979 + }, + { + "epoch": 0.5152458142137989, + "grad_norm": 0.20030324161052704, + "learning_rate": 9.524693659761808e-05, + "loss": 0.0235, + "step": 21980 + }, + { + "epoch": 0.5152692557886038, + "grad_norm": 0.4910202622413635, + "learning_rate": 9.523957877806192e-05, + "loss": 0.0863, + "step": 21981 + }, + { + "epoch": 0.5152926973634089, + "grad_norm": 0.6642935872077942, + "learning_rate": 9.523222098433592e-05, + "loss": 0.0902, + "step": 21982 + }, + { + "epoch": 0.5153161389382138, + "grad_norm": 0.3028661608695984, + "learning_rate": 9.522486321648005e-05, + "loss": 0.0589, + "step": 21983 + }, + { + "epoch": 0.5153395805130189, + "grad_norm": 0.15512195229530334, + "learning_rate": 9.521750547453416e-05, + "loss": 0.017, + "step": 21984 + }, + { + "epoch": 0.5153630220878238, + "grad_norm": 0.48564720153808594, + "learning_rate": 9.521014775853824e-05, + "loss": 0.0803, + "step": 21985 + }, + { + "epoch": 0.5153864636626289, + "grad_norm": 0.6121883988380432, + "learning_rate": 9.520279006853219e-05, + "loss": 0.088, + "step": 21986 + }, + { + "epoch": 0.5154099052374338, + "grad_norm": 0.12153158336877823, + "learning_rate": 9.519543240455595e-05, + "loss": 0.0161, + "step": 21987 + }, + { + "epoch": 0.5154333468122388, + "grad_norm": 0.7352392673492432, + "learning_rate": 9.518807476664944e-05, + "loss": 0.16, + "step": 21988 + }, + { + "epoch": 0.5154567883870439, + "grad_norm": 0.522594690322876, + "learning_rate": 9.518071715485255e-05, + "loss": 0.0586, + "step": 21989 + }, + { + "epoch": 0.5154802299618488, + "grad_norm": 0.5913299322128296, + "learning_rate": 9.517335956920524e-05, + "loss": 0.1364, + "step": 21990 + }, + { + "epoch": 0.5155036715366539, + "grad_norm": 0.35325419902801514, + "learning_rate": 9.516600200974739e-05, + "loss": 0.0554, + "step": 21991 + }, + { + "epoch": 0.5155271131114588, + "grad_norm": 0.586086630821228, + "learning_rate": 9.515864447651897e-05, + "loss": 0.0957, + "step": 21992 + }, + { + "epoch": 0.5155505546862639, + "grad_norm": 0.15498360991477966, + "learning_rate": 9.515128696955989e-05, + "loss": 0.0187, + "step": 21993 + }, + { + "epoch": 0.5155739962610688, + "grad_norm": 0.3477592170238495, + "learning_rate": 9.514392948891004e-05, + "loss": 0.0713, + "step": 21994 + }, + { + "epoch": 0.5155974378358739, + "grad_norm": 0.5930087566375732, + "learning_rate": 9.513657203460938e-05, + "loss": 0.1126, + "step": 21995 + }, + { + "epoch": 0.5156208794106788, + "grad_norm": 0.49622488021850586, + "learning_rate": 9.512921460669783e-05, + "loss": 0.0958, + "step": 21996 + }, + { + "epoch": 0.5156443209854839, + "grad_norm": 0.19211280345916748, + "learning_rate": 9.512185720521527e-05, + "loss": 0.0284, + "step": 21997 + }, + { + "epoch": 0.5156677625602888, + "grad_norm": 0.2874576449394226, + "learning_rate": 9.511449983020163e-05, + "loss": 0.0604, + "step": 21998 + }, + { + "epoch": 0.5156912041350938, + "grad_norm": 0.7582759261131287, + "learning_rate": 9.510714248169689e-05, + "loss": 0.1727, + "step": 21999 + }, + { + "epoch": 0.5157146457098988, + "grad_norm": 0.21079221367835999, + "learning_rate": 9.509978515974093e-05, + "loss": 0.0352, + "step": 22000 + }, + { + "epoch": 0.5157380872847038, + "grad_norm": 0.21460652351379395, + "learning_rate": 9.509242786437365e-05, + "loss": 0.0458, + "step": 22001 + }, + { + "epoch": 0.5157615288595088, + "grad_norm": 0.38459715247154236, + "learning_rate": 9.508507059563503e-05, + "loss": 0.1373, + "step": 22002 + }, + { + "epoch": 0.5157849704343138, + "grad_norm": 0.45962536334991455, + "learning_rate": 9.507771335356495e-05, + "loss": 0.107, + "step": 22003 + }, + { + "epoch": 0.5158084120091188, + "grad_norm": 0.5920459628105164, + "learning_rate": 9.507035613820332e-05, + "loss": 0.1369, + "step": 22004 + }, + { + "epoch": 0.5158318535839238, + "grad_norm": 0.37568384408950806, + "learning_rate": 9.506299894959007e-05, + "loss": 0.0691, + "step": 22005 + }, + { + "epoch": 0.5158552951587287, + "grad_norm": 0.4231555461883545, + "learning_rate": 9.505564178776516e-05, + "loss": 0.0571, + "step": 22006 + }, + { + "epoch": 0.5158787367335338, + "grad_norm": 0.37878403067588806, + "learning_rate": 9.504828465276844e-05, + "loss": 0.0911, + "step": 22007 + }, + { + "epoch": 0.5159021783083387, + "grad_norm": 0.5708250999450684, + "learning_rate": 9.504092754463989e-05, + "loss": 0.1177, + "step": 22008 + }, + { + "epoch": 0.5159256198831438, + "grad_norm": 0.1146901473402977, + "learning_rate": 9.50335704634194e-05, + "loss": 0.0211, + "step": 22009 + }, + { + "epoch": 0.5159490614579487, + "grad_norm": 0.5730403065681458, + "learning_rate": 9.502621340914686e-05, + "loss": 0.0683, + "step": 22010 + }, + { + "epoch": 0.5159725030327538, + "grad_norm": 0.5773929357528687, + "learning_rate": 9.501885638186224e-05, + "loss": 0.0896, + "step": 22011 + }, + { + "epoch": 0.5159959446075587, + "grad_norm": 0.35915982723236084, + "learning_rate": 9.501149938160549e-05, + "loss": 0.064, + "step": 22012 + }, + { + "epoch": 0.5160193861823638, + "grad_norm": 1.1916804313659668, + "learning_rate": 9.500414240841647e-05, + "loss": 0.0731, + "step": 22013 + }, + { + "epoch": 0.5160428277571687, + "grad_norm": 0.2847045958042145, + "learning_rate": 9.499678546233508e-05, + "loss": 0.0472, + "step": 22014 + }, + { + "epoch": 0.5160662693319737, + "grad_norm": 0.40159475803375244, + "learning_rate": 9.498942854340131e-05, + "loss": 0.0862, + "step": 22015 + }, + { + "epoch": 0.5160897109067787, + "grad_norm": 0.18068158626556396, + "learning_rate": 9.498207165165503e-05, + "loss": 0.0286, + "step": 22016 + }, + { + "epoch": 0.5161131524815837, + "grad_norm": 0.509515106678009, + "learning_rate": 9.497471478713615e-05, + "loss": 0.534, + "step": 22017 + }, + { + "epoch": 0.5161365940563887, + "grad_norm": 0.6309571266174316, + "learning_rate": 9.496735794988465e-05, + "loss": 0.0899, + "step": 22018 + }, + { + "epoch": 0.5161600356311937, + "grad_norm": 0.4736175239086151, + "learning_rate": 9.49600011399404e-05, + "loss": 0.0762, + "step": 22019 + }, + { + "epoch": 0.5161834772059987, + "grad_norm": 0.4207325875759125, + "learning_rate": 9.495264435734327e-05, + "loss": 0.0626, + "step": 22020 + }, + { + "epoch": 0.5162069187808037, + "grad_norm": 0.6502305269241333, + "learning_rate": 9.494528760213329e-05, + "loss": 0.0967, + "step": 22021 + }, + { + "epoch": 0.5162303603556087, + "grad_norm": 0.35876408219337463, + "learning_rate": 9.493793087435027e-05, + "loss": 0.1039, + "step": 22022 + }, + { + "epoch": 0.5162538019304137, + "grad_norm": 0.1676425188779831, + "learning_rate": 9.493057417403422e-05, + "loss": 0.034, + "step": 22023 + }, + { + "epoch": 0.5162772435052186, + "grad_norm": 0.4094688594341278, + "learning_rate": 9.492321750122501e-05, + "loss": 0.0831, + "step": 22024 + }, + { + "epoch": 0.5163006850800237, + "grad_norm": 0.23073388636112213, + "learning_rate": 9.491586085596257e-05, + "loss": 0.0275, + "step": 22025 + }, + { + "epoch": 0.5163241266548286, + "grad_norm": 0.6339430809020996, + "learning_rate": 9.49085042382868e-05, + "loss": 0.1744, + "step": 22026 + }, + { + "epoch": 0.5163475682296337, + "grad_norm": 0.8245176672935486, + "learning_rate": 9.490114764823765e-05, + "loss": 0.2487, + "step": 22027 + }, + { + "epoch": 0.5163710098044386, + "grad_norm": 0.6817692518234253, + "learning_rate": 9.4893791085855e-05, + "loss": 0.6138, + "step": 22028 + }, + { + "epoch": 0.5163944513792437, + "grad_norm": 0.40486544370651245, + "learning_rate": 9.488643455117879e-05, + "loss": 0.0903, + "step": 22029 + }, + { + "epoch": 0.5164178929540486, + "grad_norm": 0.7144465446472168, + "learning_rate": 9.487907804424893e-05, + "loss": 0.0668, + "step": 22030 + }, + { + "epoch": 0.5164413345288537, + "grad_norm": 0.2039932757616043, + "learning_rate": 9.487172156510533e-05, + "loss": 0.0453, + "step": 22031 + }, + { + "epoch": 0.5164647761036586, + "grad_norm": 0.47092071175575256, + "learning_rate": 9.486436511378791e-05, + "loss": 0.0589, + "step": 22032 + }, + { + "epoch": 0.5164882176784636, + "grad_norm": 0.7678199410438538, + "learning_rate": 9.485700869033661e-05, + "loss": 0.1548, + "step": 22033 + }, + { + "epoch": 0.5165116592532686, + "grad_norm": 0.19440244138240814, + "learning_rate": 9.484965229479127e-05, + "loss": 0.0279, + "step": 22034 + }, + { + "epoch": 0.5165351008280736, + "grad_norm": 0.2821498215198517, + "learning_rate": 9.48422959271919e-05, + "loss": 0.065, + "step": 22035 + }, + { + "epoch": 0.5165585424028786, + "grad_norm": 0.40620186924934387, + "learning_rate": 9.483493958757838e-05, + "loss": 0.068, + "step": 22036 + }, + { + "epoch": 0.5165819839776836, + "grad_norm": 0.10009998083114624, + "learning_rate": 9.482758327599062e-05, + "loss": 0.0156, + "step": 22037 + }, + { + "epoch": 0.5166054255524886, + "grad_norm": 0.6721030473709106, + "learning_rate": 9.482022699246855e-05, + "loss": 0.1454, + "step": 22038 + }, + { + "epoch": 0.5166288671272936, + "grad_norm": 0.46037372946739197, + "learning_rate": 9.481287073705205e-05, + "loss": 0.0989, + "step": 22039 + }, + { + "epoch": 0.5166523087020987, + "grad_norm": 0.5580621957778931, + "learning_rate": 9.480551450978109e-05, + "loss": 0.5918, + "step": 22040 + }, + { + "epoch": 0.5166757502769036, + "grad_norm": 0.10227230191230774, + "learning_rate": 9.479815831069555e-05, + "loss": 0.0139, + "step": 22041 + }, + { + "epoch": 0.5166991918517087, + "grad_norm": 0.3619122803211212, + "learning_rate": 9.479080213983533e-05, + "loss": 0.0638, + "step": 22042 + }, + { + "epoch": 0.5167226334265136, + "grad_norm": 0.6702963709831238, + "learning_rate": 9.478344599724038e-05, + "loss": 0.7012, + "step": 22043 + }, + { + "epoch": 0.5167460750013186, + "grad_norm": 0.7750061750411987, + "learning_rate": 9.477608988295059e-05, + "loss": 0.2206, + "step": 22044 + }, + { + "epoch": 0.5167695165761236, + "grad_norm": 0.6329361796379089, + "learning_rate": 9.476873379700588e-05, + "loss": 0.1279, + "step": 22045 + }, + { + "epoch": 0.5167929581509286, + "grad_norm": 0.4525070786476135, + "learning_rate": 9.476137773944614e-05, + "loss": 0.0713, + "step": 22046 + }, + { + "epoch": 0.5168163997257336, + "grad_norm": 0.5493303537368774, + "learning_rate": 9.475402171031133e-05, + "loss": 0.1312, + "step": 22047 + }, + { + "epoch": 0.5168398413005386, + "grad_norm": 0.6422492861747742, + "learning_rate": 9.474666570964137e-05, + "loss": 0.1391, + "step": 22048 + }, + { + "epoch": 0.5168632828753436, + "grad_norm": 0.07143550366163254, + "learning_rate": 9.473930973747613e-05, + "loss": 0.0144, + "step": 22049 + }, + { + "epoch": 0.5168867244501486, + "grad_norm": 0.0801234170794487, + "learning_rate": 9.473195379385557e-05, + "loss": 0.0088, + "step": 22050 + }, + { + "epoch": 0.5169101660249535, + "grad_norm": 0.28852030634880066, + "learning_rate": 9.472459787881957e-05, + "loss": 0.2666, + "step": 22051 + }, + { + "epoch": 0.5169336075997586, + "grad_norm": 0.4029179811477661, + "learning_rate": 9.471724199240801e-05, + "loss": 0.0573, + "step": 22052 + }, + { + "epoch": 0.5169570491745635, + "grad_norm": 0.47900664806365967, + "learning_rate": 9.470988613466089e-05, + "loss": 0.0678, + "step": 22053 + }, + { + "epoch": 0.5169804907493686, + "grad_norm": 0.13060373067855835, + "learning_rate": 9.470253030561806e-05, + "loss": 0.0171, + "step": 22054 + }, + { + "epoch": 0.5170039323241735, + "grad_norm": 0.6032364964485168, + "learning_rate": 9.469517450531944e-05, + "loss": 0.178, + "step": 22055 + }, + { + "epoch": 0.5170273738989786, + "grad_norm": 0.1933315098285675, + "learning_rate": 9.468781873380497e-05, + "loss": 0.0453, + "step": 22056 + }, + { + "epoch": 0.5170508154737835, + "grad_norm": 0.649793267250061, + "learning_rate": 9.468046299111454e-05, + "loss": 0.1454, + "step": 22057 + }, + { + "epoch": 0.5170742570485886, + "grad_norm": 0.7712331414222717, + "learning_rate": 9.467310727728803e-05, + "loss": 0.6509, + "step": 22058 + }, + { + "epoch": 0.5170976986233935, + "grad_norm": 0.20588338375091553, + "learning_rate": 9.466575159236541e-05, + "loss": 0.0603, + "step": 22059 + }, + { + "epoch": 0.5171211401981985, + "grad_norm": 0.3916579782962799, + "learning_rate": 9.46583959363866e-05, + "loss": 0.1154, + "step": 22060 + }, + { + "epoch": 0.5171445817730035, + "grad_norm": 0.5554677844047546, + "learning_rate": 9.465104030939147e-05, + "loss": 0.0895, + "step": 22061 + }, + { + "epoch": 0.5171680233478085, + "grad_norm": 0.7268480062484741, + "learning_rate": 9.464368471141994e-05, + "loss": 0.1332, + "step": 22062 + }, + { + "epoch": 0.5171914649226135, + "grad_norm": 0.7744359374046326, + "learning_rate": 9.463632914251193e-05, + "loss": 0.0453, + "step": 22063 + }, + { + "epoch": 0.5172149064974185, + "grad_norm": 0.6614329218864441, + "learning_rate": 9.462897360270734e-05, + "loss": 0.1002, + "step": 22064 + }, + { + "epoch": 0.5172383480722235, + "grad_norm": 0.5045260787010193, + "learning_rate": 9.46216180920461e-05, + "loss": 0.6344, + "step": 22065 + }, + { + "epoch": 0.5172617896470285, + "grad_norm": 0.5076686143875122, + "learning_rate": 9.461426261056812e-05, + "loss": 0.196, + "step": 22066 + }, + { + "epoch": 0.5172852312218335, + "grad_norm": 0.49338874220848083, + "learning_rate": 9.460690715831328e-05, + "loss": 0.0771, + "step": 22067 + }, + { + "epoch": 0.5173086727966385, + "grad_norm": 0.7102185487747192, + "learning_rate": 9.459955173532154e-05, + "loss": 0.1205, + "step": 22068 + }, + { + "epoch": 0.5173321143714434, + "grad_norm": 0.33237627148628235, + "learning_rate": 9.459219634163279e-05, + "loss": 0.0775, + "step": 22069 + }, + { + "epoch": 0.5173555559462485, + "grad_norm": 0.3672868609428406, + "learning_rate": 9.45848409772869e-05, + "loss": 0.0383, + "step": 22070 + }, + { + "epoch": 0.5173789975210534, + "grad_norm": 0.3788504898548126, + "learning_rate": 9.45774856423238e-05, + "loss": 0.0889, + "step": 22071 + }, + { + "epoch": 0.5174024390958585, + "grad_norm": 0.058058228343725204, + "learning_rate": 9.457013033678346e-05, + "loss": 0.0035, + "step": 22072 + }, + { + "epoch": 0.5174258806706634, + "grad_norm": 0.5766521096229553, + "learning_rate": 9.456277506070573e-05, + "loss": 0.1471, + "step": 22073 + }, + { + "epoch": 0.5174493222454685, + "grad_norm": 0.424223393201828, + "learning_rate": 9.455541981413053e-05, + "loss": 0.0652, + "step": 22074 + }, + { + "epoch": 0.5174727638202734, + "grad_norm": 0.29154032468795776, + "learning_rate": 9.45480645970978e-05, + "loss": 0.0699, + "step": 22075 + }, + { + "epoch": 0.5174962053950785, + "grad_norm": 0.5255448818206787, + "learning_rate": 9.454070940964741e-05, + "loss": 0.1111, + "step": 22076 + }, + { + "epoch": 0.5175196469698834, + "grad_norm": 0.6319847106933594, + "learning_rate": 9.453335425181926e-05, + "loss": 0.119, + "step": 22077 + }, + { + "epoch": 0.5175430885446884, + "grad_norm": 0.34467583894729614, + "learning_rate": 9.452599912365333e-05, + "loss": 0.1072, + "step": 22078 + }, + { + "epoch": 0.5175665301194934, + "grad_norm": 0.41258755326271057, + "learning_rate": 9.451864402518947e-05, + "loss": 0.0847, + "step": 22079 + }, + { + "epoch": 0.5175899716942984, + "grad_norm": 0.4540838897228241, + "learning_rate": 9.451128895646758e-05, + "loss": 0.0977, + "step": 22080 + }, + { + "epoch": 0.5176134132691034, + "grad_norm": 0.380410760641098, + "learning_rate": 9.45039339175276e-05, + "loss": 0.0734, + "step": 22081 + }, + { + "epoch": 0.5176368548439084, + "grad_norm": 0.37277913093566895, + "learning_rate": 9.449657890840942e-05, + "loss": 0.1031, + "step": 22082 + }, + { + "epoch": 0.5176602964187134, + "grad_norm": 0.5873217582702637, + "learning_rate": 9.448922392915295e-05, + "loss": 0.1457, + "step": 22083 + }, + { + "epoch": 0.5176837379935184, + "grad_norm": 0.14532937109470367, + "learning_rate": 9.44818689797981e-05, + "loss": 0.0167, + "step": 22084 + }, + { + "epoch": 0.5177071795683234, + "grad_norm": 0.23847369849681854, + "learning_rate": 9.447451406038483e-05, + "loss": 0.0412, + "step": 22085 + }, + { + "epoch": 0.5177306211431284, + "grad_norm": 0.4162508249282837, + "learning_rate": 9.446715917095298e-05, + "loss": 0.0677, + "step": 22086 + }, + { + "epoch": 0.5177540627179333, + "grad_norm": 0.33519604802131653, + "learning_rate": 9.445980431154247e-05, + "loss": 0.0443, + "step": 22087 + }, + { + "epoch": 0.5177775042927384, + "grad_norm": 0.6100331544876099, + "learning_rate": 9.445244948219323e-05, + "loss": 0.2469, + "step": 22088 + }, + { + "epoch": 0.5178009458675433, + "grad_norm": 0.4356100559234619, + "learning_rate": 9.444509468294515e-05, + "loss": 0.0944, + "step": 22089 + }, + { + "epoch": 0.5178243874423484, + "grad_norm": 0.2695022225379944, + "learning_rate": 9.443773991383814e-05, + "loss": 0.0603, + "step": 22090 + }, + { + "epoch": 0.5178478290171534, + "grad_norm": 0.35051429271698, + "learning_rate": 9.443038517491212e-05, + "loss": 0.0368, + "step": 22091 + }, + { + "epoch": 0.5178712705919584, + "grad_norm": 0.28057390451431274, + "learning_rate": 9.442303046620699e-05, + "loss": 0.0669, + "step": 22092 + }, + { + "epoch": 0.5178947121667634, + "grad_norm": 0.2090083509683609, + "learning_rate": 9.441567578776263e-05, + "loss": 0.0518, + "step": 22093 + }, + { + "epoch": 0.5179181537415684, + "grad_norm": 0.09014929085969925, + "learning_rate": 9.440832113961899e-05, + "loss": 0.0119, + "step": 22094 + }, + { + "epoch": 0.5179415953163734, + "grad_norm": 0.3291054368019104, + "learning_rate": 9.440096652181593e-05, + "loss": 0.0685, + "step": 22095 + }, + { + "epoch": 0.5179650368911783, + "grad_norm": 0.5617273449897766, + "learning_rate": 9.43936119343934e-05, + "loss": 0.0796, + "step": 22096 + }, + { + "epoch": 0.5179884784659834, + "grad_norm": 0.7149775624275208, + "learning_rate": 9.438625737739129e-05, + "loss": 0.1152, + "step": 22097 + }, + { + "epoch": 0.5180119200407883, + "grad_norm": 0.10768170654773712, + "learning_rate": 9.437890285084951e-05, + "loss": 0.0124, + "step": 22098 + }, + { + "epoch": 0.5180353616155934, + "grad_norm": 0.42172348499298096, + "learning_rate": 9.437154835480797e-05, + "loss": 0.096, + "step": 22099 + }, + { + "epoch": 0.5180588031903983, + "grad_norm": 0.37859994173049927, + "learning_rate": 9.436419388930654e-05, + "loss": 0.0919, + "step": 22100 + }, + { + "epoch": 0.5180822447652034, + "grad_norm": 0.4621513783931732, + "learning_rate": 9.435683945438518e-05, + "loss": 0.1096, + "step": 22101 + }, + { + "epoch": 0.5181056863400083, + "grad_norm": 0.5817313194274902, + "learning_rate": 9.434948505008374e-05, + "loss": 0.1208, + "step": 22102 + }, + { + "epoch": 0.5181291279148134, + "grad_norm": 0.971594512462616, + "learning_rate": 9.434213067644219e-05, + "loss": 0.6498, + "step": 22103 + }, + { + "epoch": 0.5181525694896183, + "grad_norm": 0.5400933027267456, + "learning_rate": 9.433477633350038e-05, + "loss": 0.0701, + "step": 22104 + }, + { + "epoch": 0.5181760110644233, + "grad_norm": 0.46823838353157043, + "learning_rate": 9.432742202129822e-05, + "loss": 0.0784, + "step": 22105 + }, + { + "epoch": 0.5181994526392283, + "grad_norm": 0.4429363012313843, + "learning_rate": 9.432006773987564e-05, + "loss": 0.118, + "step": 22106 + }, + { + "epoch": 0.5182228942140333, + "grad_norm": 0.5783902406692505, + "learning_rate": 9.431271348927249e-05, + "loss": 0.1747, + "step": 22107 + }, + { + "epoch": 0.5182463357888383, + "grad_norm": 0.5150298476219177, + "learning_rate": 9.430535926952879e-05, + "loss": 0.11, + "step": 22108 + }, + { + "epoch": 0.5182697773636433, + "grad_norm": 0.4015548825263977, + "learning_rate": 9.429800508068431e-05, + "loss": 0.0573, + "step": 22109 + }, + { + "epoch": 0.5182932189384483, + "grad_norm": 0.8567149639129639, + "learning_rate": 9.429065092277905e-05, + "loss": 0.1557, + "step": 22110 + }, + { + "epoch": 0.5183166605132533, + "grad_norm": 0.15867100656032562, + "learning_rate": 9.428329679585289e-05, + "loss": 0.0285, + "step": 22111 + }, + { + "epoch": 0.5183401020880583, + "grad_norm": 0.2853047847747803, + "learning_rate": 9.427594269994568e-05, + "loss": 0.0455, + "step": 22112 + }, + { + "epoch": 0.5183635436628633, + "grad_norm": 0.20486557483673096, + "learning_rate": 9.42685886350974e-05, + "loss": 0.0304, + "step": 22113 + }, + { + "epoch": 0.5183869852376682, + "grad_norm": 0.7787456512451172, + "learning_rate": 9.426123460134792e-05, + "loss": 0.2359, + "step": 22114 + }, + { + "epoch": 0.5184104268124733, + "grad_norm": 0.5075376629829407, + "learning_rate": 9.425388059873713e-05, + "loss": 0.5625, + "step": 22115 + }, + { + "epoch": 0.5184338683872782, + "grad_norm": 0.5432745814323425, + "learning_rate": 9.424652662730495e-05, + "loss": 0.1467, + "step": 22116 + }, + { + "epoch": 0.5184573099620833, + "grad_norm": 0.6888120770454407, + "learning_rate": 9.423917268709129e-05, + "loss": 0.1077, + "step": 22117 + }, + { + "epoch": 0.5184807515368882, + "grad_norm": 0.3702698349952698, + "learning_rate": 9.4231818778136e-05, + "loss": 0.0734, + "step": 22118 + }, + { + "epoch": 0.5185041931116933, + "grad_norm": 0.29945164918899536, + "learning_rate": 9.422446490047902e-05, + "loss": 0.0465, + "step": 22119 + }, + { + "epoch": 0.5185276346864982, + "grad_norm": 0.5448703765869141, + "learning_rate": 9.421711105416029e-05, + "loss": 0.5671, + "step": 22120 + }, + { + "epoch": 0.5185510762613033, + "grad_norm": 0.29825669527053833, + "learning_rate": 9.420975723921967e-05, + "loss": 0.073, + "step": 22121 + }, + { + "epoch": 0.5185745178361082, + "grad_norm": 0.3408212661743164, + "learning_rate": 9.420240345569706e-05, + "loss": 0.4569, + "step": 22122 + }, + { + "epoch": 0.5185979594109132, + "grad_norm": 0.744284451007843, + "learning_rate": 9.419504970363239e-05, + "loss": 0.0812, + "step": 22123 + }, + { + "epoch": 0.5186214009857182, + "grad_norm": 0.41388148069381714, + "learning_rate": 9.418769598306556e-05, + "loss": 0.0661, + "step": 22124 + }, + { + "epoch": 0.5186448425605232, + "grad_norm": 0.7570905685424805, + "learning_rate": 9.418034229403641e-05, + "loss": 0.1357, + "step": 22125 + }, + { + "epoch": 0.5186682841353282, + "grad_norm": 0.495116651058197, + "learning_rate": 9.417298863658493e-05, + "loss": 0.1748, + "step": 22126 + }, + { + "epoch": 0.5186917257101332, + "grad_norm": 0.15945784747600555, + "learning_rate": 9.416563501075096e-05, + "loss": 0.013, + "step": 22127 + }, + { + "epoch": 0.5187151672849382, + "grad_norm": 0.5511173009872437, + "learning_rate": 9.415828141657441e-05, + "loss": 0.1098, + "step": 22128 + }, + { + "epoch": 0.5187386088597432, + "grad_norm": 0.618820309638977, + "learning_rate": 9.41509278540952e-05, + "loss": 0.1477, + "step": 22129 + }, + { + "epoch": 0.5187620504345481, + "grad_norm": 0.3162108063697815, + "learning_rate": 9.414357432335322e-05, + "loss": 0.0759, + "step": 22130 + }, + { + "epoch": 0.5187854920093532, + "grad_norm": 0.16074684262275696, + "learning_rate": 9.413622082438834e-05, + "loss": 0.04, + "step": 22131 + }, + { + "epoch": 0.5188089335841581, + "grad_norm": 0.3651781976222992, + "learning_rate": 9.41288673572405e-05, + "loss": 0.0836, + "step": 22132 + }, + { + "epoch": 0.5188323751589632, + "grad_norm": 0.5830720663070679, + "learning_rate": 9.412151392194961e-05, + "loss": 0.1056, + "step": 22133 + }, + { + "epoch": 0.5188558167337681, + "grad_norm": 0.23316729068756104, + "learning_rate": 9.411416051855556e-05, + "loss": 0.0593, + "step": 22134 + }, + { + "epoch": 0.5188792583085732, + "grad_norm": 0.2782762050628662, + "learning_rate": 9.410680714709822e-05, + "loss": 0.0503, + "step": 22135 + }, + { + "epoch": 0.5189026998833781, + "grad_norm": 0.2639179527759552, + "learning_rate": 9.409945380761752e-05, + "loss": 0.0151, + "step": 22136 + }, + { + "epoch": 0.5189261414581832, + "grad_norm": 0.4328261911869049, + "learning_rate": 9.409210050015336e-05, + "loss": 0.0755, + "step": 22137 + }, + { + "epoch": 0.5189495830329881, + "grad_norm": 0.28970006108283997, + "learning_rate": 9.408474722474561e-05, + "loss": 0.0441, + "step": 22138 + }, + { + "epoch": 0.5189730246077932, + "grad_norm": 0.16365118324756622, + "learning_rate": 9.407739398143419e-05, + "loss": 0.0167, + "step": 22139 + }, + { + "epoch": 0.5189964661825981, + "grad_norm": 0.6007880568504333, + "learning_rate": 9.4070040770259e-05, + "loss": 0.0883, + "step": 22140 + }, + { + "epoch": 0.5190199077574031, + "grad_norm": 0.09154165536165237, + "learning_rate": 9.406268759125995e-05, + "loss": 0.0113, + "step": 22141 + }, + { + "epoch": 0.5190433493322082, + "grad_norm": 0.4522203505039215, + "learning_rate": 9.405533444447691e-05, + "loss": 0.0713, + "step": 22142 + }, + { + "epoch": 0.5190667909070131, + "grad_norm": 0.2447546422481537, + "learning_rate": 9.404798132994976e-05, + "loss": 0.041, + "step": 22143 + }, + { + "epoch": 0.5190902324818182, + "grad_norm": 0.5233331322669983, + "learning_rate": 9.404062824771846e-05, + "loss": 0.0563, + "step": 22144 + }, + { + "epoch": 0.5191136740566231, + "grad_norm": 0.5109034180641174, + "learning_rate": 9.403327519782288e-05, + "loss": 0.4681, + "step": 22145 + }, + { + "epoch": 0.5191371156314282, + "grad_norm": 0.45984309911727905, + "learning_rate": 9.402592218030293e-05, + "loss": 0.0562, + "step": 22146 + }, + { + "epoch": 0.5191605572062331, + "grad_norm": 0.6638591885566711, + "learning_rate": 9.401856919519847e-05, + "loss": 0.1759, + "step": 22147 + }, + { + "epoch": 0.5191839987810382, + "grad_norm": 0.6703406572341919, + "learning_rate": 9.401121624254945e-05, + "loss": 0.1784, + "step": 22148 + }, + { + "epoch": 0.5192074403558431, + "grad_norm": 0.2655636966228485, + "learning_rate": 9.400386332239572e-05, + "loss": 0.0669, + "step": 22149 + }, + { + "epoch": 0.5192308819306481, + "grad_norm": 0.13216780126094818, + "learning_rate": 9.39965104347772e-05, + "loss": 0.0192, + "step": 22150 + }, + { + "epoch": 0.5192543235054531, + "grad_norm": 0.4737999737262726, + "learning_rate": 9.398915757973379e-05, + "loss": 0.0743, + "step": 22151 + }, + { + "epoch": 0.5192777650802581, + "grad_norm": 0.1415325105190277, + "learning_rate": 9.398180475730538e-05, + "loss": 0.0151, + "step": 22152 + }, + { + "epoch": 0.5193012066550631, + "grad_norm": 0.2514415979385376, + "learning_rate": 9.397445196753185e-05, + "loss": 0.0509, + "step": 22153 + }, + { + "epoch": 0.5193246482298681, + "grad_norm": 0.5001582503318787, + "learning_rate": 9.396709921045313e-05, + "loss": 0.1956, + "step": 22154 + }, + { + "epoch": 0.5193480898046731, + "grad_norm": 0.5377777814865112, + "learning_rate": 9.39597464861091e-05, + "loss": 0.5772, + "step": 22155 + }, + { + "epoch": 0.5193715313794781, + "grad_norm": 0.39266693592071533, + "learning_rate": 9.395239379453962e-05, + "loss": 0.0703, + "step": 22156 + }, + { + "epoch": 0.519394972954283, + "grad_norm": 0.4142817258834839, + "learning_rate": 9.394504113578463e-05, + "loss": 0.1151, + "step": 22157 + }, + { + "epoch": 0.5194184145290881, + "grad_norm": 0.1962127685546875, + "learning_rate": 9.393768850988405e-05, + "loss": 0.0397, + "step": 22158 + }, + { + "epoch": 0.519441856103893, + "grad_norm": 0.41981396079063416, + "learning_rate": 9.393033591687773e-05, + "loss": 0.0915, + "step": 22159 + }, + { + "epoch": 0.5194652976786981, + "grad_norm": 0.4079248905181885, + "learning_rate": 9.392298335680556e-05, + "loss": 0.6024, + "step": 22160 + }, + { + "epoch": 0.519488739253503, + "grad_norm": 0.8178441524505615, + "learning_rate": 9.391563082970749e-05, + "loss": 0.1429, + "step": 22161 + }, + { + "epoch": 0.5195121808283081, + "grad_norm": 0.630264401435852, + "learning_rate": 9.390827833562337e-05, + "loss": 0.1272, + "step": 22162 + }, + { + "epoch": 0.519535622403113, + "grad_norm": 1.348374366760254, + "learning_rate": 9.390092587459308e-05, + "loss": 0.3096, + "step": 22163 + }, + { + "epoch": 0.5195590639779181, + "grad_norm": 0.5066619515419006, + "learning_rate": 9.389357344665655e-05, + "loss": 0.0613, + "step": 22164 + }, + { + "epoch": 0.519582505552723, + "grad_norm": 0.15224386751651764, + "learning_rate": 9.388622105185368e-05, + "loss": 0.0306, + "step": 22165 + }, + { + "epoch": 0.5196059471275281, + "grad_norm": 0.2934759855270386, + "learning_rate": 9.38788686902243e-05, + "loss": 0.0523, + "step": 22166 + }, + { + "epoch": 0.519629388702333, + "grad_norm": 0.3967254161834717, + "learning_rate": 9.387151636180838e-05, + "loss": 0.1306, + "step": 22167 + }, + { + "epoch": 0.519652830277138, + "grad_norm": 0.44758427143096924, + "learning_rate": 9.386416406664575e-05, + "loss": 0.0944, + "step": 22168 + }, + { + "epoch": 0.519676271851943, + "grad_norm": 0.21989375352859497, + "learning_rate": 9.385681180477639e-05, + "loss": 0.048, + "step": 22169 + }, + { + "epoch": 0.519699713426748, + "grad_norm": 0.38217201828956604, + "learning_rate": 9.38494595762401e-05, + "loss": 0.0957, + "step": 22170 + }, + { + "epoch": 0.519723155001553, + "grad_norm": 0.05095373094081879, + "learning_rate": 9.384210738107685e-05, + "loss": 0.0046, + "step": 22171 + }, + { + "epoch": 0.519746596576358, + "grad_norm": 0.4934273660182953, + "learning_rate": 9.38347552193265e-05, + "loss": 0.0634, + "step": 22172 + }, + { + "epoch": 0.519770038151163, + "grad_norm": 0.3847058117389679, + "learning_rate": 9.38274030910289e-05, + "loss": 0.0924, + "step": 22173 + }, + { + "epoch": 0.519793479725968, + "grad_norm": 0.26920291781425476, + "learning_rate": 9.382005099622404e-05, + "loss": 0.0656, + "step": 22174 + }, + { + "epoch": 0.519816921300773, + "grad_norm": 0.2377912700176239, + "learning_rate": 9.381269893495173e-05, + "loss": 0.0551, + "step": 22175 + }, + { + "epoch": 0.519840362875578, + "grad_norm": 0.4140819013118744, + "learning_rate": 9.380534690725188e-05, + "loss": 0.0718, + "step": 22176 + }, + { + "epoch": 0.5198638044503829, + "grad_norm": 0.18827594816684723, + "learning_rate": 9.379799491316441e-05, + "loss": 0.0349, + "step": 22177 + }, + { + "epoch": 0.519887246025188, + "grad_norm": 0.44185981154441833, + "learning_rate": 9.379064295272917e-05, + "loss": 0.0764, + "step": 22178 + }, + { + "epoch": 0.5199106875999929, + "grad_norm": 0.3879150450229645, + "learning_rate": 9.378329102598611e-05, + "loss": 0.1433, + "step": 22179 + }, + { + "epoch": 0.519934129174798, + "grad_norm": 0.24994418025016785, + "learning_rate": 9.377593913297502e-05, + "loss": 0.0653, + "step": 22180 + }, + { + "epoch": 0.5199575707496029, + "grad_norm": 1.1369034051895142, + "learning_rate": 9.376858727373592e-05, + "loss": 0.0366, + "step": 22181 + }, + { + "epoch": 0.519981012324408, + "grad_norm": 0.4857557713985443, + "learning_rate": 9.376123544830862e-05, + "loss": 0.1023, + "step": 22182 + }, + { + "epoch": 0.5200044538992129, + "grad_norm": 0.42875179648399353, + "learning_rate": 9.375388365673305e-05, + "loss": 0.1102, + "step": 22183 + }, + { + "epoch": 0.520027895474018, + "grad_norm": 0.11460462957620621, + "learning_rate": 9.374653189904908e-05, + "loss": 0.0116, + "step": 22184 + }, + { + "epoch": 0.5200513370488229, + "grad_norm": 0.6037158966064453, + "learning_rate": 9.373918017529658e-05, + "loss": 0.1316, + "step": 22185 + }, + { + "epoch": 0.5200747786236279, + "grad_norm": 0.43183010816574097, + "learning_rate": 9.373182848551549e-05, + "loss": 0.0855, + "step": 22186 + }, + { + "epoch": 0.5200982201984329, + "grad_norm": 0.5219195485115051, + "learning_rate": 9.372447682974568e-05, + "loss": 0.7083, + "step": 22187 + }, + { + "epoch": 0.5201216617732379, + "grad_norm": 0.15491217374801636, + "learning_rate": 9.371712520802701e-05, + "loss": 0.0434, + "step": 22188 + }, + { + "epoch": 0.5201451033480429, + "grad_norm": 0.41840171813964844, + "learning_rate": 9.370977362039941e-05, + "loss": 0.0756, + "step": 22189 + }, + { + "epoch": 0.5201685449228479, + "grad_norm": 0.3520192503929138, + "learning_rate": 9.370242206690277e-05, + "loss": 0.0806, + "step": 22190 + }, + { + "epoch": 0.5201919864976529, + "grad_norm": 0.5258169174194336, + "learning_rate": 9.369507054757694e-05, + "loss": 0.0534, + "step": 22191 + }, + { + "epoch": 0.5202154280724579, + "grad_norm": 0.4597947895526886, + "learning_rate": 9.36877190624618e-05, + "loss": 0.0863, + "step": 22192 + }, + { + "epoch": 0.520238869647263, + "grad_norm": 0.6077679395675659, + "learning_rate": 9.368036761159732e-05, + "loss": 0.159, + "step": 22193 + }, + { + "epoch": 0.5202623112220679, + "grad_norm": 0.4302297830581665, + "learning_rate": 9.367301619502335e-05, + "loss": 0.0603, + "step": 22194 + }, + { + "epoch": 0.520285752796873, + "grad_norm": 0.6855745315551758, + "learning_rate": 9.366566481277974e-05, + "loss": 0.1572, + "step": 22195 + }, + { + "epoch": 0.5203091943716779, + "grad_norm": 0.5061175227165222, + "learning_rate": 9.365831346490645e-05, + "loss": 0.0838, + "step": 22196 + }, + { + "epoch": 0.5203326359464829, + "grad_norm": 0.4918203353881836, + "learning_rate": 9.365096215144332e-05, + "loss": 0.103, + "step": 22197 + }, + { + "epoch": 0.5203560775212879, + "grad_norm": 0.2383960485458374, + "learning_rate": 9.364361087243022e-05, + "loss": 0.0617, + "step": 22198 + }, + { + "epoch": 0.5203795190960929, + "grad_norm": 0.17460288107395172, + "learning_rate": 9.36362596279071e-05, + "loss": 0.032, + "step": 22199 + }, + { + "epoch": 0.5204029606708979, + "grad_norm": 0.37745705246925354, + "learning_rate": 9.362890841791381e-05, + "loss": 0.0785, + "step": 22200 + }, + { + "epoch": 0.5204264022457029, + "grad_norm": 0.35568058490753174, + "learning_rate": 9.36215572424902e-05, + "loss": 0.0916, + "step": 22201 + }, + { + "epoch": 0.5204498438205079, + "grad_norm": 0.3425253927707672, + "learning_rate": 9.361420610167625e-05, + "loss": 0.0501, + "step": 22202 + }, + { + "epoch": 0.5204732853953129, + "grad_norm": 0.8406018018722534, + "learning_rate": 9.360685499551178e-05, + "loss": 0.1114, + "step": 22203 + }, + { + "epoch": 0.5204967269701178, + "grad_norm": 0.35941582918167114, + "learning_rate": 9.359950392403666e-05, + "loss": 0.0873, + "step": 22204 + }, + { + "epoch": 0.5205201685449229, + "grad_norm": 0.3963052034378052, + "learning_rate": 9.359215288729083e-05, + "loss": 0.0818, + "step": 22205 + }, + { + "epoch": 0.5205436101197278, + "grad_norm": 0.42633283138275146, + "learning_rate": 9.358480188531419e-05, + "loss": 0.0622, + "step": 22206 + }, + { + "epoch": 0.5205670516945329, + "grad_norm": 0.421847939491272, + "learning_rate": 9.357745091814659e-05, + "loss": 0.0824, + "step": 22207 + }, + { + "epoch": 0.5205904932693378, + "grad_norm": 0.713718056678772, + "learning_rate": 9.357009998582789e-05, + "loss": 0.1794, + "step": 22208 + }, + { + "epoch": 0.5206139348441429, + "grad_norm": 0.19380216300487518, + "learning_rate": 9.356274908839804e-05, + "loss": 0.0301, + "step": 22209 + }, + { + "epoch": 0.5206373764189478, + "grad_norm": 0.3624328076839447, + "learning_rate": 9.35553982258969e-05, + "loss": 0.0941, + "step": 22210 + }, + { + "epoch": 0.5206608179937529, + "grad_norm": 0.32750335335731506, + "learning_rate": 9.354804739836432e-05, + "loss": 0.0936, + "step": 22211 + }, + { + "epoch": 0.5206842595685578, + "grad_norm": 0.6152448654174805, + "learning_rate": 9.354069660584023e-05, + "loss": 0.0637, + "step": 22212 + }, + { + "epoch": 0.5207077011433628, + "grad_norm": 0.40536320209503174, + "learning_rate": 9.353334584836449e-05, + "loss": 0.5653, + "step": 22213 + }, + { + "epoch": 0.5207311427181678, + "grad_norm": 0.3281627893447876, + "learning_rate": 9.352599512597702e-05, + "loss": 0.3883, + "step": 22214 + }, + { + "epoch": 0.5207545842929728, + "grad_norm": 0.5342617034912109, + "learning_rate": 9.351864443871768e-05, + "loss": 0.5315, + "step": 22215 + }, + { + "epoch": 0.5207780258677778, + "grad_norm": 0.3674444258213043, + "learning_rate": 9.351129378662631e-05, + "loss": 0.081, + "step": 22216 + }, + { + "epoch": 0.5208014674425828, + "grad_norm": 0.2640242278575897, + "learning_rate": 9.35039431697429e-05, + "loss": 0.0529, + "step": 22217 + }, + { + "epoch": 0.5208249090173878, + "grad_norm": 0.4744005799293518, + "learning_rate": 9.349659258810723e-05, + "loss": 0.0299, + "step": 22218 + }, + { + "epoch": 0.5208483505921928, + "grad_norm": 0.3105432987213135, + "learning_rate": 9.348924204175927e-05, + "loss": 0.0886, + "step": 22219 + }, + { + "epoch": 0.5208717921669977, + "grad_norm": 0.3728094696998596, + "learning_rate": 9.348189153073884e-05, + "loss": 0.0979, + "step": 22220 + }, + { + "epoch": 0.5208952337418028, + "grad_norm": 0.566947340965271, + "learning_rate": 9.347454105508587e-05, + "loss": 0.0429, + "step": 22221 + }, + { + "epoch": 0.5209186753166077, + "grad_norm": 0.437421053647995, + "learning_rate": 9.346719061484024e-05, + "loss": 0.083, + "step": 22222 + }, + { + "epoch": 0.5209421168914128, + "grad_norm": 0.45250293612480164, + "learning_rate": 9.345984021004179e-05, + "loss": 0.0258, + "step": 22223 + }, + { + "epoch": 0.5209655584662177, + "grad_norm": 0.13352732360363007, + "learning_rate": 9.345248984073042e-05, + "loss": 0.0289, + "step": 22224 + }, + { + "epoch": 0.5209890000410228, + "grad_norm": 0.17060771584510803, + "learning_rate": 9.344513950694607e-05, + "loss": 0.0241, + "step": 22225 + }, + { + "epoch": 0.5210124416158277, + "grad_norm": 0.2239663153886795, + "learning_rate": 9.343778920872853e-05, + "loss": 0.0641, + "step": 22226 + }, + { + "epoch": 0.5210358831906328, + "grad_norm": 0.5039916038513184, + "learning_rate": 9.343043894611775e-05, + "loss": 0.1084, + "step": 22227 + }, + { + "epoch": 0.5210593247654377, + "grad_norm": 0.5641341209411621, + "learning_rate": 9.342308871915361e-05, + "loss": 0.1364, + "step": 22228 + }, + { + "epoch": 0.5210827663402428, + "grad_norm": 0.1233968660235405, + "learning_rate": 9.341573852787591e-05, + "loss": 0.023, + "step": 22229 + }, + { + "epoch": 0.5211062079150477, + "grad_norm": 0.5596114993095398, + "learning_rate": 9.340838837232464e-05, + "loss": 0.6621, + "step": 22230 + }, + { + "epoch": 0.5211296494898527, + "grad_norm": 0.49882689118385315, + "learning_rate": 9.340103825253965e-05, + "loss": 0.682, + "step": 22231 + }, + { + "epoch": 0.5211530910646577, + "grad_norm": 1.0453616380691528, + "learning_rate": 9.339368816856082e-05, + "loss": 0.1391, + "step": 22232 + }, + { + "epoch": 0.5211765326394627, + "grad_norm": 0.3659493327140808, + "learning_rate": 9.338633812042801e-05, + "loss": 0.0615, + "step": 22233 + }, + { + "epoch": 0.5211999742142677, + "grad_norm": 0.8019707202911377, + "learning_rate": 9.337898810818112e-05, + "loss": 0.1303, + "step": 22234 + }, + { + "epoch": 0.5212234157890727, + "grad_norm": 0.4585851728916168, + "learning_rate": 9.337163813186003e-05, + "loss": 0.116, + "step": 22235 + }, + { + "epoch": 0.5212468573638777, + "grad_norm": 0.4007839858531952, + "learning_rate": 9.33642881915046e-05, + "loss": 0.0505, + "step": 22236 + }, + { + "epoch": 0.5212702989386827, + "grad_norm": 0.6799644827842712, + "learning_rate": 9.335693828715475e-05, + "loss": 0.0719, + "step": 22237 + }, + { + "epoch": 0.5212937405134876, + "grad_norm": 0.305998295545578, + "learning_rate": 9.334958841885034e-05, + "loss": 0.0463, + "step": 22238 + }, + { + "epoch": 0.5213171820882927, + "grad_norm": 0.4136706590652466, + "learning_rate": 9.334223858663122e-05, + "loss": 0.0776, + "step": 22239 + }, + { + "epoch": 0.5213406236630976, + "grad_norm": 0.40577107667922974, + "learning_rate": 9.333488879053733e-05, + "loss": 0.0539, + "step": 22240 + }, + { + "epoch": 0.5213640652379027, + "grad_norm": 0.957258939743042, + "learning_rate": 9.332753903060847e-05, + "loss": 0.297, + "step": 22241 + }, + { + "epoch": 0.5213875068127076, + "grad_norm": 0.8070530891418457, + "learning_rate": 9.332018930688462e-05, + "loss": 0.1463, + "step": 22242 + }, + { + "epoch": 0.5214109483875127, + "grad_norm": 0.37342405319213867, + "learning_rate": 9.33128396194056e-05, + "loss": 0.0859, + "step": 22243 + }, + { + "epoch": 0.5214343899623177, + "grad_norm": 0.230342298746109, + "learning_rate": 9.330548996821131e-05, + "loss": 0.0671, + "step": 22244 + }, + { + "epoch": 0.5214578315371227, + "grad_norm": 0.3447813391685486, + "learning_rate": 9.329814035334162e-05, + "loss": 0.0302, + "step": 22245 + }, + { + "epoch": 0.5214812731119277, + "grad_norm": 0.18209470808506012, + "learning_rate": 9.32907907748364e-05, + "loss": 0.0289, + "step": 22246 + }, + { + "epoch": 0.5215047146867327, + "grad_norm": 0.7345834970474243, + "learning_rate": 9.328344123273554e-05, + "loss": 0.1356, + "step": 22247 + }, + { + "epoch": 0.5215281562615377, + "grad_norm": 0.14830978214740753, + "learning_rate": 9.327609172707891e-05, + "loss": 0.0142, + "step": 22248 + }, + { + "epoch": 0.5215515978363426, + "grad_norm": 0.5110137462615967, + "learning_rate": 9.32687422579064e-05, + "loss": 0.5132, + "step": 22249 + }, + { + "epoch": 0.5215750394111477, + "grad_norm": 0.26005685329437256, + "learning_rate": 9.32613928252579e-05, + "loss": 0.0567, + "step": 22250 + }, + { + "epoch": 0.5215984809859526, + "grad_norm": 0.705130398273468, + "learning_rate": 9.325404342917323e-05, + "loss": 0.0673, + "step": 22251 + }, + { + "epoch": 0.5216219225607577, + "grad_norm": 0.3448377251625061, + "learning_rate": 9.324669406969235e-05, + "loss": 0.0507, + "step": 22252 + }, + { + "epoch": 0.5216453641355626, + "grad_norm": 0.4226605296134949, + "learning_rate": 9.323934474685505e-05, + "loss": 0.0989, + "step": 22253 + }, + { + "epoch": 0.5216688057103677, + "grad_norm": 0.5222977995872498, + "learning_rate": 9.32319954607013e-05, + "loss": 0.6468, + "step": 22254 + }, + { + "epoch": 0.5216922472851726, + "grad_norm": 0.3289535343647003, + "learning_rate": 9.322464621127094e-05, + "loss": 0.0353, + "step": 22255 + }, + { + "epoch": 0.5217156888599777, + "grad_norm": 0.24957740306854248, + "learning_rate": 9.321729699860381e-05, + "loss": 0.0319, + "step": 22256 + }, + { + "epoch": 0.5217391304347826, + "grad_norm": 0.4741501212120056, + "learning_rate": 9.320994782273984e-05, + "loss": 0.0698, + "step": 22257 + }, + { + "epoch": 0.5217625720095876, + "grad_norm": 0.18313255906105042, + "learning_rate": 9.320259868371886e-05, + "loss": 0.036, + "step": 22258 + }, + { + "epoch": 0.5217860135843926, + "grad_norm": 0.68290776014328, + "learning_rate": 9.319524958158079e-05, + "loss": 0.1421, + "step": 22259 + }, + { + "epoch": 0.5218094551591976, + "grad_norm": 0.5671905875205994, + "learning_rate": 9.318790051636548e-05, + "loss": 0.0892, + "step": 22260 + }, + { + "epoch": 0.5218328967340026, + "grad_norm": 0.23273919522762299, + "learning_rate": 9.318055148811281e-05, + "loss": 0.0401, + "step": 22261 + }, + { + "epoch": 0.5218563383088076, + "grad_norm": 0.48399847745895386, + "learning_rate": 9.317320249686267e-05, + "loss": 0.488, + "step": 22262 + }, + { + "epoch": 0.5218797798836126, + "grad_norm": 0.5193194150924683, + "learning_rate": 9.316585354265494e-05, + "loss": 0.0716, + "step": 22263 + }, + { + "epoch": 0.5219032214584176, + "grad_norm": 0.7416393160820007, + "learning_rate": 9.315850462552944e-05, + "loss": 0.5713, + "step": 22264 + }, + { + "epoch": 0.5219266630332225, + "grad_norm": 0.2543409466743469, + "learning_rate": 9.315115574552607e-05, + "loss": 0.04, + "step": 22265 + }, + { + "epoch": 0.5219501046080276, + "grad_norm": 0.31709733605384827, + "learning_rate": 9.314380690268478e-05, + "loss": 0.0591, + "step": 22266 + }, + { + "epoch": 0.5219735461828325, + "grad_norm": 0.4480564296245575, + "learning_rate": 9.313645809704537e-05, + "loss": 0.0799, + "step": 22267 + }, + { + "epoch": 0.5219969877576376, + "grad_norm": 0.33864572644233704, + "learning_rate": 9.31291093286477e-05, + "loss": 0.0476, + "step": 22268 + }, + { + "epoch": 0.5220204293324425, + "grad_norm": 0.5525569319725037, + "learning_rate": 9.312176059753173e-05, + "loss": 0.1407, + "step": 22269 + }, + { + "epoch": 0.5220438709072476, + "grad_norm": 0.3457653522491455, + "learning_rate": 9.311441190373726e-05, + "loss": 0.3546, + "step": 22270 + }, + { + "epoch": 0.5220673124820525, + "grad_norm": 0.4881519377231598, + "learning_rate": 9.310706324730416e-05, + "loss": 0.0961, + "step": 22271 + }, + { + "epoch": 0.5220907540568576, + "grad_norm": 0.040516629815101624, + "learning_rate": 9.309971462827235e-05, + "loss": 0.0116, + "step": 22272 + }, + { + "epoch": 0.5221141956316625, + "grad_norm": 0.5853396654129028, + "learning_rate": 9.30923660466817e-05, + "loss": 0.4019, + "step": 22273 + }, + { + "epoch": 0.5221376372064676, + "grad_norm": 0.4210304617881775, + "learning_rate": 9.308501750257203e-05, + "loss": 0.0877, + "step": 22274 + }, + { + "epoch": 0.5221610787812725, + "grad_norm": 0.657179057598114, + "learning_rate": 9.307766899598328e-05, + "loss": 0.0423, + "step": 22275 + }, + { + "epoch": 0.5221845203560775, + "grad_norm": 0.5613231658935547, + "learning_rate": 9.307032052695528e-05, + "loss": 0.1761, + "step": 22276 + }, + { + "epoch": 0.5222079619308825, + "grad_norm": 0.5430102944374084, + "learning_rate": 9.306297209552788e-05, + "loss": 0.0867, + "step": 22277 + }, + { + "epoch": 0.5222314035056875, + "grad_norm": 0.27257612347602844, + "learning_rate": 9.3055623701741e-05, + "loss": 0.0362, + "step": 22278 + }, + { + "epoch": 0.5222548450804925, + "grad_norm": 0.1878119707107544, + "learning_rate": 9.304827534563454e-05, + "loss": 0.0383, + "step": 22279 + }, + { + "epoch": 0.5222782866552975, + "grad_norm": 0.8372688889503479, + "learning_rate": 9.304092702724833e-05, + "loss": 0.1235, + "step": 22280 + }, + { + "epoch": 0.5223017282301025, + "grad_norm": 0.27967187762260437, + "learning_rate": 9.30335787466222e-05, + "loss": 0.0484, + "step": 22281 + }, + { + "epoch": 0.5223251698049075, + "grad_norm": 0.5053728818893433, + "learning_rate": 9.302623050379612e-05, + "loss": 0.1134, + "step": 22282 + }, + { + "epoch": 0.5223486113797124, + "grad_norm": 0.1874343305826187, + "learning_rate": 9.30188822988099e-05, + "loss": 0.0327, + "step": 22283 + }, + { + "epoch": 0.5223720529545175, + "grad_norm": 0.4036652147769928, + "learning_rate": 9.301153413170339e-05, + "loss": 0.1051, + "step": 22284 + }, + { + "epoch": 0.5223954945293224, + "grad_norm": 0.47994640469551086, + "learning_rate": 9.300418600251652e-05, + "loss": 0.1031, + "step": 22285 + }, + { + "epoch": 0.5224189361041275, + "grad_norm": 0.5055827498435974, + "learning_rate": 9.299683791128914e-05, + "loss": 0.1176, + "step": 22286 + }, + { + "epoch": 0.5224423776789324, + "grad_norm": 0.22985902428627014, + "learning_rate": 9.298948985806108e-05, + "loss": 0.029, + "step": 22287 + }, + { + "epoch": 0.5224658192537375, + "grad_norm": 0.48336756229400635, + "learning_rate": 9.298214184287229e-05, + "loss": 0.0848, + "step": 22288 + }, + { + "epoch": 0.5224892608285424, + "grad_norm": 0.31079158186912537, + "learning_rate": 9.297479386576252e-05, + "loss": 0.0418, + "step": 22289 + }, + { + "epoch": 0.5225127024033475, + "grad_norm": 0.6771708130836487, + "learning_rate": 9.296744592677178e-05, + "loss": 0.8198, + "step": 22290 + }, + { + "epoch": 0.5225361439781524, + "grad_norm": 0.1752234697341919, + "learning_rate": 9.296009802593984e-05, + "loss": 0.0122, + "step": 22291 + }, + { + "epoch": 0.5225595855529575, + "grad_norm": 0.5435231328010559, + "learning_rate": 9.295275016330665e-05, + "loss": 0.0898, + "step": 22292 + }, + { + "epoch": 0.5225830271277624, + "grad_norm": 0.34903448820114136, + "learning_rate": 9.294540233891202e-05, + "loss": 0.0468, + "step": 22293 + }, + { + "epoch": 0.5226064687025674, + "grad_norm": 0.5172184109687805, + "learning_rate": 9.293805455279582e-05, + "loss": 0.1256, + "step": 22294 + }, + { + "epoch": 0.5226299102773725, + "grad_norm": 0.5797253847122192, + "learning_rate": 9.293070680499795e-05, + "loss": 0.1266, + "step": 22295 + }, + { + "epoch": 0.5226533518521774, + "grad_norm": 0.5727055072784424, + "learning_rate": 9.292335909555825e-05, + "loss": 0.5606, + "step": 22296 + }, + { + "epoch": 0.5226767934269825, + "grad_norm": 0.6665028929710388, + "learning_rate": 9.291601142451661e-05, + "loss": 0.1988, + "step": 22297 + }, + { + "epoch": 0.5227002350017874, + "grad_norm": 0.6125810146331787, + "learning_rate": 9.29086637919129e-05, + "loss": 0.1683, + "step": 22298 + }, + { + "epoch": 0.5227236765765925, + "grad_norm": 0.4625032842159271, + "learning_rate": 9.290131619778696e-05, + "loss": 0.0932, + "step": 22299 + }, + { + "epoch": 0.5227471181513974, + "grad_norm": 0.5413926243782043, + "learning_rate": 9.28939686421787e-05, + "loss": 0.1135, + "step": 22300 + }, + { + "epoch": 0.5227705597262025, + "grad_norm": 0.6639466285705566, + "learning_rate": 9.288662112512797e-05, + "loss": 0.0979, + "step": 22301 + }, + { + "epoch": 0.5227940013010074, + "grad_norm": 0.5780974626541138, + "learning_rate": 9.287927364667458e-05, + "loss": 0.1049, + "step": 22302 + }, + { + "epoch": 0.5228174428758124, + "grad_norm": 0.39043086767196655, + "learning_rate": 9.287192620685847e-05, + "loss": 0.0715, + "step": 22303 + }, + { + "epoch": 0.5228408844506174, + "grad_norm": 0.15477776527404785, + "learning_rate": 9.286457880571952e-05, + "loss": 0.0252, + "step": 22304 + }, + { + "epoch": 0.5228643260254224, + "grad_norm": 0.4258773624897003, + "learning_rate": 9.285723144329756e-05, + "loss": 0.1095, + "step": 22305 + }, + { + "epoch": 0.5228877676002274, + "grad_norm": 0.3896631896495819, + "learning_rate": 9.284988411963245e-05, + "loss": 0.1052, + "step": 22306 + }, + { + "epoch": 0.5229112091750324, + "grad_norm": 0.633760392665863, + "learning_rate": 9.284253683476409e-05, + "loss": 0.0649, + "step": 22307 + }, + { + "epoch": 0.5229346507498374, + "grad_norm": 0.19287648797035217, + "learning_rate": 9.283518958873232e-05, + "loss": 0.0221, + "step": 22308 + }, + { + "epoch": 0.5229580923246424, + "grad_norm": 0.21908052265644073, + "learning_rate": 9.282784238157701e-05, + "loss": 0.0396, + "step": 22309 + }, + { + "epoch": 0.5229815338994473, + "grad_norm": 0.36512625217437744, + "learning_rate": 9.282049521333802e-05, + "loss": 0.0606, + "step": 22310 + }, + { + "epoch": 0.5230049754742524, + "grad_norm": 1.0327602624893188, + "learning_rate": 9.281314808405524e-05, + "loss": 0.1216, + "step": 22311 + }, + { + "epoch": 0.5230284170490573, + "grad_norm": 0.20102964341640472, + "learning_rate": 9.28058009937685e-05, + "loss": 0.0339, + "step": 22312 + }, + { + "epoch": 0.5230518586238624, + "grad_norm": 0.5569104552268982, + "learning_rate": 9.27984539425177e-05, + "loss": 0.0866, + "step": 22313 + }, + { + "epoch": 0.5230753001986673, + "grad_norm": 0.24195148050785065, + "learning_rate": 9.279110693034266e-05, + "loss": 0.0744, + "step": 22314 + }, + { + "epoch": 0.5230987417734724, + "grad_norm": 0.11072178930044174, + "learning_rate": 9.278375995728331e-05, + "loss": 0.0196, + "step": 22315 + }, + { + "epoch": 0.5231221833482773, + "grad_norm": 0.1695629060268402, + "learning_rate": 9.277641302337948e-05, + "loss": 0.0191, + "step": 22316 + }, + { + "epoch": 0.5231456249230824, + "grad_norm": 0.5852925777435303, + "learning_rate": 9.276906612867103e-05, + "loss": 0.1447, + "step": 22317 + }, + { + "epoch": 0.5231690664978873, + "grad_norm": 0.3827577233314514, + "learning_rate": 9.276171927319785e-05, + "loss": 0.1117, + "step": 22318 + }, + { + "epoch": 0.5231925080726924, + "grad_norm": 0.656864583492279, + "learning_rate": 9.275437245699974e-05, + "loss": 0.1642, + "step": 22319 + }, + { + "epoch": 0.5232159496474973, + "grad_norm": 0.359697550535202, + "learning_rate": 9.274702568011665e-05, + "loss": 0.0945, + "step": 22320 + }, + { + "epoch": 0.5232393912223023, + "grad_norm": 1.0323357582092285, + "learning_rate": 9.273967894258841e-05, + "loss": 0.042, + "step": 22321 + }, + { + "epoch": 0.5232628327971073, + "grad_norm": 0.37731263041496277, + "learning_rate": 9.273233224445483e-05, + "loss": 0.0693, + "step": 22322 + }, + { + "epoch": 0.5232862743719123, + "grad_norm": 0.7782991528511047, + "learning_rate": 9.272498558575585e-05, + "loss": 0.2045, + "step": 22323 + }, + { + "epoch": 0.5233097159467173, + "grad_norm": 0.4759513735771179, + "learning_rate": 9.271763896653131e-05, + "loss": 0.0539, + "step": 22324 + }, + { + "epoch": 0.5233331575215223, + "grad_norm": 0.2817886769771576, + "learning_rate": 9.271029238682104e-05, + "loss": 0.0178, + "step": 22325 + }, + { + "epoch": 0.5233565990963273, + "grad_norm": 0.49414777755737305, + "learning_rate": 9.270294584666491e-05, + "loss": 0.0758, + "step": 22326 + }, + { + "epoch": 0.5233800406711323, + "grad_norm": 0.22346198558807373, + "learning_rate": 9.269559934610285e-05, + "loss": 0.0323, + "step": 22327 + }, + { + "epoch": 0.5234034822459372, + "grad_norm": 0.8877311944961548, + "learning_rate": 9.268825288517466e-05, + "loss": 0.3454, + "step": 22328 + }, + { + "epoch": 0.5234269238207423, + "grad_norm": 0.18178120255470276, + "learning_rate": 9.26809064639202e-05, + "loss": 0.0438, + "step": 22329 + }, + { + "epoch": 0.5234503653955472, + "grad_norm": 0.32843026518821716, + "learning_rate": 9.267356008237936e-05, + "loss": 0.0769, + "step": 22330 + }, + { + "epoch": 0.5234738069703523, + "grad_norm": 0.09667973220348358, + "learning_rate": 9.2666213740592e-05, + "loss": 0.0223, + "step": 22331 + }, + { + "epoch": 0.5234972485451572, + "grad_norm": 0.360133558511734, + "learning_rate": 9.265886743859795e-05, + "loss": 0.0538, + "step": 22332 + }, + { + "epoch": 0.5235206901199623, + "grad_norm": 0.45465287566185, + "learning_rate": 9.265152117643711e-05, + "loss": 0.0631, + "step": 22333 + }, + { + "epoch": 0.5235441316947672, + "grad_norm": 0.3899829089641571, + "learning_rate": 9.26441749541493e-05, + "loss": 0.0929, + "step": 22334 + }, + { + "epoch": 0.5235675732695723, + "grad_norm": 0.39968252182006836, + "learning_rate": 9.263682877177442e-05, + "loss": 0.044, + "step": 22335 + }, + { + "epoch": 0.5235910148443772, + "grad_norm": 0.7326130867004395, + "learning_rate": 9.262948262935233e-05, + "loss": 0.0495, + "step": 22336 + }, + { + "epoch": 0.5236144564191823, + "grad_norm": 0.4913206100463867, + "learning_rate": 9.262213652692285e-05, + "loss": 0.3725, + "step": 22337 + }, + { + "epoch": 0.5236378979939872, + "grad_norm": 0.5748478770256042, + "learning_rate": 9.261479046452583e-05, + "loss": 0.0767, + "step": 22338 + }, + { + "epoch": 0.5236613395687922, + "grad_norm": 0.4346105754375458, + "learning_rate": 9.260744444220121e-05, + "loss": 0.0836, + "step": 22339 + }, + { + "epoch": 0.5236847811435972, + "grad_norm": 0.6112855672836304, + "learning_rate": 9.260009845998881e-05, + "loss": 0.0995, + "step": 22340 + }, + { + "epoch": 0.5237082227184022, + "grad_norm": 0.666168212890625, + "learning_rate": 9.259275251792845e-05, + "loss": 0.5942, + "step": 22341 + }, + { + "epoch": 0.5237316642932072, + "grad_norm": 0.3088066875934601, + "learning_rate": 9.258540661606006e-05, + "loss": 0.0565, + "step": 22342 + }, + { + "epoch": 0.5237551058680122, + "grad_norm": 0.3795009255409241, + "learning_rate": 9.257806075442346e-05, + "loss": 0.0621, + "step": 22343 + }, + { + "epoch": 0.5237785474428172, + "grad_norm": 0.5566991567611694, + "learning_rate": 9.257071493305848e-05, + "loss": 0.1044, + "step": 22344 + }, + { + "epoch": 0.5238019890176222, + "grad_norm": 0.14730817079544067, + "learning_rate": 9.256336915200504e-05, + "loss": 0.0194, + "step": 22345 + }, + { + "epoch": 0.5238254305924273, + "grad_norm": 0.37430667877197266, + "learning_rate": 9.255602341130297e-05, + "loss": 0.0941, + "step": 22346 + }, + { + "epoch": 0.5238488721672322, + "grad_norm": 0.7667904496192932, + "learning_rate": 9.25486777109921e-05, + "loss": 0.1312, + "step": 22347 + }, + { + "epoch": 0.5238723137420372, + "grad_norm": 0.08856586366891861, + "learning_rate": 9.254133205111233e-05, + "loss": 0.0123, + "step": 22348 + }, + { + "epoch": 0.5238957553168422, + "grad_norm": 0.23766392469406128, + "learning_rate": 9.253398643170351e-05, + "loss": 0.033, + "step": 22349 + }, + { + "epoch": 0.5239191968916472, + "grad_norm": 0.1549144983291626, + "learning_rate": 9.252664085280545e-05, + "loss": 0.0422, + "step": 22350 + }, + { + "epoch": 0.5239426384664522, + "grad_norm": 0.6520909667015076, + "learning_rate": 9.251929531445805e-05, + "loss": 0.1382, + "step": 22351 + }, + { + "epoch": 0.5239660800412572, + "grad_norm": 0.5668389201164246, + "learning_rate": 9.251194981670121e-05, + "loss": 0.097, + "step": 22352 + }, + { + "epoch": 0.5239895216160622, + "grad_norm": 0.16320358216762543, + "learning_rate": 9.250460435957472e-05, + "loss": 0.0433, + "step": 22353 + }, + { + "epoch": 0.5240129631908672, + "grad_norm": 0.41145607829093933, + "learning_rate": 9.249725894311845e-05, + "loss": 0.1086, + "step": 22354 + }, + { + "epoch": 0.5240364047656721, + "grad_norm": 0.21138502657413483, + "learning_rate": 9.248991356737228e-05, + "loss": 0.0398, + "step": 22355 + }, + { + "epoch": 0.5240598463404772, + "grad_norm": 0.26964840292930603, + "learning_rate": 9.248256823237604e-05, + "loss": 0.039, + "step": 22356 + }, + { + "epoch": 0.5240832879152821, + "grad_norm": 0.2971213459968567, + "learning_rate": 9.247522293816959e-05, + "loss": 0.0246, + "step": 22357 + }, + { + "epoch": 0.5241067294900872, + "grad_norm": 0.5934112668037415, + "learning_rate": 9.246787768479281e-05, + "loss": 0.1566, + "step": 22358 + }, + { + "epoch": 0.5241301710648921, + "grad_norm": 0.6543574333190918, + "learning_rate": 9.246053247228555e-05, + "loss": 0.0501, + "step": 22359 + }, + { + "epoch": 0.5241536126396972, + "grad_norm": 0.38529062271118164, + "learning_rate": 9.24531873006876e-05, + "loss": 0.501, + "step": 22360 + }, + { + "epoch": 0.5241770542145021, + "grad_norm": 0.4240790605545044, + "learning_rate": 9.24458421700389e-05, + "loss": 0.0976, + "step": 22361 + }, + { + "epoch": 0.5242004957893072, + "grad_norm": 0.256480872631073, + "learning_rate": 9.243849708037924e-05, + "loss": 0.024, + "step": 22362 + }, + { + "epoch": 0.5242239373641121, + "grad_norm": 0.4542481005191803, + "learning_rate": 9.243115203174856e-05, + "loss": 0.0771, + "step": 22363 + }, + { + "epoch": 0.5242473789389172, + "grad_norm": 0.595144510269165, + "learning_rate": 9.242380702418662e-05, + "loss": 0.0744, + "step": 22364 + }, + { + "epoch": 0.5242708205137221, + "grad_norm": 0.4960198700428009, + "learning_rate": 9.241646205773336e-05, + "loss": 0.0933, + "step": 22365 + }, + { + "epoch": 0.5242942620885271, + "grad_norm": 0.6122522354125977, + "learning_rate": 9.240911713242857e-05, + "loss": 0.1526, + "step": 22366 + }, + { + "epoch": 0.5243177036633321, + "grad_norm": 0.4766523838043213, + "learning_rate": 9.24017722483121e-05, + "loss": 0.1354, + "step": 22367 + }, + { + "epoch": 0.5243411452381371, + "grad_norm": 0.41918960213661194, + "learning_rate": 9.239442740542386e-05, + "loss": 0.0673, + "step": 22368 + }, + { + "epoch": 0.5243645868129421, + "grad_norm": 0.4321313798427582, + "learning_rate": 9.238708260380365e-05, + "loss": 0.0969, + "step": 22369 + }, + { + "epoch": 0.5243880283877471, + "grad_norm": 0.34838008880615234, + "learning_rate": 9.237973784349136e-05, + "loss": 0.0497, + "step": 22370 + }, + { + "epoch": 0.5244114699625521, + "grad_norm": 0.4372556507587433, + "learning_rate": 9.237239312452682e-05, + "loss": 0.1094, + "step": 22371 + }, + { + "epoch": 0.5244349115373571, + "grad_norm": 0.5452728867530823, + "learning_rate": 9.236504844694989e-05, + "loss": 0.1295, + "step": 22372 + }, + { + "epoch": 0.524458353112162, + "grad_norm": 0.6873915195465088, + "learning_rate": 9.235770381080042e-05, + "loss": 0.7975, + "step": 22373 + }, + { + "epoch": 0.5244817946869671, + "grad_norm": 0.527328610420227, + "learning_rate": 9.235035921611823e-05, + "loss": 0.0985, + "step": 22374 + }, + { + "epoch": 0.524505236261772, + "grad_norm": 0.14090079069137573, + "learning_rate": 9.234301466294326e-05, + "loss": 0.0249, + "step": 22375 + }, + { + "epoch": 0.5245286778365771, + "grad_norm": 0.2155076265335083, + "learning_rate": 9.233567015131526e-05, + "loss": 0.0382, + "step": 22376 + }, + { + "epoch": 0.524552119411382, + "grad_norm": 0.39725393056869507, + "learning_rate": 9.232832568127417e-05, + "loss": 0.0832, + "step": 22377 + }, + { + "epoch": 0.5245755609861871, + "grad_norm": 0.31864893436431885, + "learning_rate": 9.232098125285979e-05, + "loss": 0.0595, + "step": 22378 + }, + { + "epoch": 0.524599002560992, + "grad_norm": 0.1914251744747162, + "learning_rate": 9.231363686611198e-05, + "loss": 0.0381, + "step": 22379 + }, + { + "epoch": 0.5246224441357971, + "grad_norm": 0.23757831752300262, + "learning_rate": 9.230629252107059e-05, + "loss": 0.0478, + "step": 22380 + }, + { + "epoch": 0.524645885710602, + "grad_norm": 0.5925993323326111, + "learning_rate": 9.229894821777549e-05, + "loss": 0.1714, + "step": 22381 + }, + { + "epoch": 0.524669327285407, + "grad_norm": 0.5902407765388489, + "learning_rate": 9.229160395626648e-05, + "loss": 0.1133, + "step": 22382 + }, + { + "epoch": 0.524692768860212, + "grad_norm": 0.3963911831378937, + "learning_rate": 9.228425973658348e-05, + "loss": 0.0924, + "step": 22383 + }, + { + "epoch": 0.524716210435017, + "grad_norm": 0.3660443127155304, + "learning_rate": 9.22769155587663e-05, + "loss": 0.0349, + "step": 22384 + }, + { + "epoch": 0.524739652009822, + "grad_norm": 0.3112928867340088, + "learning_rate": 9.226957142285476e-05, + "loss": 0.0534, + "step": 22385 + }, + { + "epoch": 0.524763093584627, + "grad_norm": 0.43519988656044006, + "learning_rate": 9.226222732888878e-05, + "loss": 0.0976, + "step": 22386 + }, + { + "epoch": 0.524786535159432, + "grad_norm": 0.31142064929008484, + "learning_rate": 9.22548832769081e-05, + "loss": 0.0592, + "step": 22387 + }, + { + "epoch": 0.524809976734237, + "grad_norm": 0.4655996859073639, + "learning_rate": 9.224753926695271e-05, + "loss": 0.1002, + "step": 22388 + }, + { + "epoch": 0.524833418309042, + "grad_norm": 0.21198943257331848, + "learning_rate": 9.224019529906236e-05, + "loss": 0.0489, + "step": 22389 + }, + { + "epoch": 0.524856859883847, + "grad_norm": 0.3645693063735962, + "learning_rate": 9.223285137327695e-05, + "loss": 0.0809, + "step": 22390 + }, + { + "epoch": 0.5248803014586519, + "grad_norm": 0.47384393215179443, + "learning_rate": 9.222550748963633e-05, + "loss": 0.1175, + "step": 22391 + }, + { + "epoch": 0.524903743033457, + "grad_norm": 0.089212566614151, + "learning_rate": 9.221816364818027e-05, + "loss": 0.0191, + "step": 22392 + }, + { + "epoch": 0.5249271846082619, + "grad_norm": 0.49804815649986267, + "learning_rate": 9.22108198489487e-05, + "loss": 0.021, + "step": 22393 + }, + { + "epoch": 0.524950626183067, + "grad_norm": 0.6800670623779297, + "learning_rate": 9.220347609198146e-05, + "loss": 0.0663, + "step": 22394 + }, + { + "epoch": 0.5249740677578719, + "grad_norm": 0.5923932194709778, + "learning_rate": 9.219613237731833e-05, + "loss": 0.1675, + "step": 22395 + }, + { + "epoch": 0.524997509332677, + "grad_norm": 0.0815378949046135, + "learning_rate": 9.218878870499925e-05, + "loss": 0.0084, + "step": 22396 + }, + { + "epoch": 0.5250209509074819, + "grad_norm": 0.8385384678840637, + "learning_rate": 9.2181445075064e-05, + "loss": 0.1924, + "step": 22397 + }, + { + "epoch": 0.525044392482287, + "grad_norm": 0.35326817631721497, + "learning_rate": 9.217410148755245e-05, + "loss": 0.0857, + "step": 22398 + }, + { + "epoch": 0.525067834057092, + "grad_norm": 0.40918272733688354, + "learning_rate": 9.216675794250441e-05, + "loss": 0.0647, + "step": 22399 + }, + { + "epoch": 0.525091275631897, + "grad_norm": 0.6827799081802368, + "learning_rate": 9.215941443995982e-05, + "loss": 0.1422, + "step": 22400 + }, + { + "epoch": 0.525114717206702, + "grad_norm": 0.3148789405822754, + "learning_rate": 9.215207097995845e-05, + "loss": 0.1052, + "step": 22401 + }, + { + "epoch": 0.5251381587815069, + "grad_norm": 0.13238298892974854, + "learning_rate": 9.214472756254014e-05, + "loss": 0.0251, + "step": 22402 + }, + { + "epoch": 0.525161600356312, + "grad_norm": 0.6889440417289734, + "learning_rate": 9.21373841877448e-05, + "loss": 0.0822, + "step": 22403 + }, + { + "epoch": 0.5251850419311169, + "grad_norm": 0.6443667411804199, + "learning_rate": 9.213004085561222e-05, + "loss": 0.0971, + "step": 22404 + }, + { + "epoch": 0.525208483505922, + "grad_norm": 0.4939020574092865, + "learning_rate": 9.212269756618225e-05, + "loss": 0.1735, + "step": 22405 + }, + { + "epoch": 0.5252319250807269, + "grad_norm": 0.4094700217247009, + "learning_rate": 9.211535431949475e-05, + "loss": 0.0891, + "step": 22406 + }, + { + "epoch": 0.525255366655532, + "grad_norm": 0.21161051094532013, + "learning_rate": 9.210801111558955e-05, + "loss": 0.0336, + "step": 22407 + }, + { + "epoch": 0.5252788082303369, + "grad_norm": 0.4070558249950409, + "learning_rate": 9.210066795450651e-05, + "loss": 0.1091, + "step": 22408 + }, + { + "epoch": 0.525302249805142, + "grad_norm": 0.15883533656597137, + "learning_rate": 9.209332483628549e-05, + "loss": 0.0287, + "step": 22409 + }, + { + "epoch": 0.5253256913799469, + "grad_norm": 0.45167404413223267, + "learning_rate": 9.208598176096628e-05, + "loss": 0.0741, + "step": 22410 + }, + { + "epoch": 0.5253491329547519, + "grad_norm": 0.6137272119522095, + "learning_rate": 9.207863872858875e-05, + "loss": 0.1038, + "step": 22411 + }, + { + "epoch": 0.5253725745295569, + "grad_norm": 0.2854105830192566, + "learning_rate": 9.207129573919276e-05, + "loss": 0.0365, + "step": 22412 + }, + { + "epoch": 0.5253960161043619, + "grad_norm": 0.4182838201522827, + "learning_rate": 9.206395279281814e-05, + "loss": 0.1048, + "step": 22413 + }, + { + "epoch": 0.5254194576791669, + "grad_norm": 0.5430601835250854, + "learning_rate": 9.205660988950474e-05, + "loss": 0.1067, + "step": 22414 + }, + { + "epoch": 0.5254428992539719, + "grad_norm": 0.5589776039123535, + "learning_rate": 9.204926702929241e-05, + "loss": 0.0802, + "step": 22415 + }, + { + "epoch": 0.5254663408287769, + "grad_norm": 0.44960474967956543, + "learning_rate": 9.204192421222098e-05, + "loss": 0.0947, + "step": 22416 + }, + { + "epoch": 0.5254897824035819, + "grad_norm": 0.2388063222169876, + "learning_rate": 9.203458143833027e-05, + "loss": 0.0445, + "step": 22417 + }, + { + "epoch": 0.5255132239783868, + "grad_norm": 0.1314336508512497, + "learning_rate": 9.202723870766017e-05, + "loss": 0.0281, + "step": 22418 + }, + { + "epoch": 0.5255366655531919, + "grad_norm": 0.32334402203559875, + "learning_rate": 9.20198960202505e-05, + "loss": 0.0627, + "step": 22419 + }, + { + "epoch": 0.5255601071279968, + "grad_norm": 0.4348858594894409, + "learning_rate": 9.201255337614107e-05, + "loss": 0.5357, + "step": 22420 + }, + { + "epoch": 0.5255835487028019, + "grad_norm": 0.2743828296661377, + "learning_rate": 9.200521077537178e-05, + "loss": 0.0271, + "step": 22421 + }, + { + "epoch": 0.5256069902776068, + "grad_norm": 0.40371790528297424, + "learning_rate": 9.199786821798243e-05, + "loss": 0.1224, + "step": 22422 + }, + { + "epoch": 0.5256304318524119, + "grad_norm": 0.34824487566947937, + "learning_rate": 9.199052570401282e-05, + "loss": 0.0871, + "step": 22423 + }, + { + "epoch": 0.5256538734272168, + "grad_norm": 0.7098196148872375, + "learning_rate": 9.198318323350289e-05, + "loss": 0.2426, + "step": 22424 + }, + { + "epoch": 0.5256773150020219, + "grad_norm": 0.21809981763362885, + "learning_rate": 9.197584080649244e-05, + "loss": 0.0311, + "step": 22425 + }, + { + "epoch": 0.5257007565768268, + "grad_norm": 0.37716951966285706, + "learning_rate": 9.19684984230213e-05, + "loss": 0.0551, + "step": 22426 + }, + { + "epoch": 0.5257241981516318, + "grad_norm": 0.39583006501197815, + "learning_rate": 9.196115608312931e-05, + "loss": 0.2988, + "step": 22427 + }, + { + "epoch": 0.5257476397264368, + "grad_norm": 0.13154816627502441, + "learning_rate": 9.195381378685632e-05, + "loss": 0.0282, + "step": 22428 + }, + { + "epoch": 0.5257710813012418, + "grad_norm": 0.46440431475639343, + "learning_rate": 9.194647153424217e-05, + "loss": 0.1353, + "step": 22429 + }, + { + "epoch": 0.5257945228760468, + "grad_norm": 0.44250884652137756, + "learning_rate": 9.193912932532667e-05, + "loss": 0.1136, + "step": 22430 + }, + { + "epoch": 0.5258179644508518, + "grad_norm": 0.08188498765230179, + "learning_rate": 9.19317871601497e-05, + "loss": 0.0152, + "step": 22431 + }, + { + "epoch": 0.5258414060256568, + "grad_norm": 0.7289205193519592, + "learning_rate": 9.192444503875107e-05, + "loss": 0.5991, + "step": 22432 + }, + { + "epoch": 0.5258648476004618, + "grad_norm": 0.24736163020133972, + "learning_rate": 9.191710296117062e-05, + "loss": 0.0281, + "step": 22433 + }, + { + "epoch": 0.5258882891752668, + "grad_norm": 0.24862872064113617, + "learning_rate": 9.190976092744821e-05, + "loss": 0.0436, + "step": 22434 + }, + { + "epoch": 0.5259117307500718, + "grad_norm": 0.6363619565963745, + "learning_rate": 9.190241893762362e-05, + "loss": 0.1069, + "step": 22435 + }, + { + "epoch": 0.5259351723248767, + "grad_norm": 0.42784106731414795, + "learning_rate": 9.189507699173678e-05, + "loss": 0.0443, + "step": 22436 + }, + { + "epoch": 0.5259586138996818, + "grad_norm": 0.6186965107917786, + "learning_rate": 9.188773508982746e-05, + "loss": 0.1045, + "step": 22437 + }, + { + "epoch": 0.5259820554744867, + "grad_norm": 0.14959879219532013, + "learning_rate": 9.188039323193555e-05, + "loss": 0.0216, + "step": 22438 + }, + { + "epoch": 0.5260054970492918, + "grad_norm": 0.6126819252967834, + "learning_rate": 9.187305141810084e-05, + "loss": 0.0892, + "step": 22439 + }, + { + "epoch": 0.5260289386240967, + "grad_norm": 0.07098228484392166, + "learning_rate": 9.186570964836316e-05, + "loss": 0.0114, + "step": 22440 + }, + { + "epoch": 0.5260523801989018, + "grad_norm": 0.5356602072715759, + "learning_rate": 9.185836792276239e-05, + "loss": 0.0868, + "step": 22441 + }, + { + "epoch": 0.5260758217737067, + "grad_norm": 0.6041731238365173, + "learning_rate": 9.185102624133835e-05, + "loss": 0.1242, + "step": 22442 + }, + { + "epoch": 0.5260992633485118, + "grad_norm": 0.2333989441394806, + "learning_rate": 9.184368460413084e-05, + "loss": 0.0423, + "step": 22443 + }, + { + "epoch": 0.5261227049233167, + "grad_norm": 0.6695714592933655, + "learning_rate": 9.183634301117976e-05, + "loss": 0.6135, + "step": 22444 + }, + { + "epoch": 0.5261461464981217, + "grad_norm": 0.47644495964050293, + "learning_rate": 9.182900146252489e-05, + "loss": 0.102, + "step": 22445 + }, + { + "epoch": 0.5261695880729267, + "grad_norm": 0.19203682243824005, + "learning_rate": 9.18216599582061e-05, + "loss": 0.0296, + "step": 22446 + }, + { + "epoch": 0.5261930296477317, + "grad_norm": 0.5150144696235657, + "learning_rate": 9.181431849826318e-05, + "loss": 0.0559, + "step": 22447 + }, + { + "epoch": 0.5262164712225367, + "grad_norm": 0.284453421831131, + "learning_rate": 9.180697708273603e-05, + "loss": 0.0748, + "step": 22448 + }, + { + "epoch": 0.5262399127973417, + "grad_norm": 0.5031340718269348, + "learning_rate": 9.179963571166445e-05, + "loss": 0.1342, + "step": 22449 + }, + { + "epoch": 0.5262633543721468, + "grad_norm": 0.5732572674751282, + "learning_rate": 9.179229438508828e-05, + "loss": 0.0758, + "step": 22450 + }, + { + "epoch": 0.5262867959469517, + "grad_norm": 0.6255761384963989, + "learning_rate": 9.178495310304736e-05, + "loss": 0.1116, + "step": 22451 + }, + { + "epoch": 0.5263102375217568, + "grad_norm": 0.34055617451667786, + "learning_rate": 9.177761186558148e-05, + "loss": 0.0539, + "step": 22452 + }, + { + "epoch": 0.5263336790965617, + "grad_norm": 0.11796021461486816, + "learning_rate": 9.177027067273054e-05, + "loss": 0.0185, + "step": 22453 + }, + { + "epoch": 0.5263571206713668, + "grad_norm": 0.46510785818099976, + "learning_rate": 9.176292952453435e-05, + "loss": 0.0823, + "step": 22454 + }, + { + "epoch": 0.5263805622461717, + "grad_norm": 0.34394165873527527, + "learning_rate": 9.175558842103272e-05, + "loss": 0.0219, + "step": 22455 + }, + { + "epoch": 0.5264040038209767, + "grad_norm": 0.12143642455339432, + "learning_rate": 9.17482473622655e-05, + "loss": 0.018, + "step": 22456 + }, + { + "epoch": 0.5264274453957817, + "grad_norm": 0.14398479461669922, + "learning_rate": 9.174090634827253e-05, + "loss": 0.0265, + "step": 22457 + }, + { + "epoch": 0.5264508869705867, + "grad_norm": 0.6537600755691528, + "learning_rate": 9.173356537909361e-05, + "loss": 0.0766, + "step": 22458 + }, + { + "epoch": 0.5264743285453917, + "grad_norm": 0.37584903836250305, + "learning_rate": 9.172622445476862e-05, + "loss": 0.0864, + "step": 22459 + }, + { + "epoch": 0.5264977701201967, + "grad_norm": 0.1381734311580658, + "learning_rate": 9.171888357533733e-05, + "loss": 0.0268, + "step": 22460 + }, + { + "epoch": 0.5265212116950017, + "grad_norm": 0.6398982405662537, + "learning_rate": 9.171154274083965e-05, + "loss": 0.1092, + "step": 22461 + }, + { + "epoch": 0.5265446532698067, + "grad_norm": 0.3864938020706177, + "learning_rate": 9.170420195131535e-05, + "loss": 0.0587, + "step": 22462 + }, + { + "epoch": 0.5265680948446116, + "grad_norm": 0.1998835951089859, + "learning_rate": 9.169686120680431e-05, + "loss": 0.0377, + "step": 22463 + }, + { + "epoch": 0.5265915364194167, + "grad_norm": 0.4333779215812683, + "learning_rate": 9.168952050734634e-05, + "loss": 0.0962, + "step": 22464 + }, + { + "epoch": 0.5266149779942216, + "grad_norm": 0.20307284593582153, + "learning_rate": 9.168217985298123e-05, + "loss": 0.0177, + "step": 22465 + }, + { + "epoch": 0.5266384195690267, + "grad_norm": 0.4468538463115692, + "learning_rate": 9.167483924374888e-05, + "loss": 0.1116, + "step": 22466 + }, + { + "epoch": 0.5266618611438316, + "grad_norm": 0.8930294513702393, + "learning_rate": 9.166749867968908e-05, + "loss": 0.1842, + "step": 22467 + }, + { + "epoch": 0.5266853027186367, + "grad_norm": 0.16117389500141144, + "learning_rate": 9.166015816084164e-05, + "loss": 0.0185, + "step": 22468 + }, + { + "epoch": 0.5267087442934416, + "grad_norm": 0.6049567461013794, + "learning_rate": 9.165281768724644e-05, + "loss": 0.5438, + "step": 22469 + }, + { + "epoch": 0.5267321858682467, + "grad_norm": 0.5180573463439941, + "learning_rate": 9.16454772589433e-05, + "loss": 0.0782, + "step": 22470 + }, + { + "epoch": 0.5267556274430516, + "grad_norm": 0.4543913006782532, + "learning_rate": 9.163813687597199e-05, + "loss": 0.085, + "step": 22471 + }, + { + "epoch": 0.5267790690178566, + "grad_norm": 0.6209509968757629, + "learning_rate": 9.163079653837237e-05, + "loss": 0.1616, + "step": 22472 + }, + { + "epoch": 0.5268025105926616, + "grad_norm": 0.7432235479354858, + "learning_rate": 9.162345624618433e-05, + "loss": 0.0986, + "step": 22473 + }, + { + "epoch": 0.5268259521674666, + "grad_norm": 0.5386156439781189, + "learning_rate": 9.161611599944766e-05, + "loss": 0.1263, + "step": 22474 + }, + { + "epoch": 0.5268493937422716, + "grad_norm": 0.7108360528945923, + "learning_rate": 9.160877579820215e-05, + "loss": 0.1377, + "step": 22475 + }, + { + "epoch": 0.5268728353170766, + "grad_norm": 0.708281934261322, + "learning_rate": 9.16014356424877e-05, + "loss": 0.1627, + "step": 22476 + }, + { + "epoch": 0.5268962768918816, + "grad_norm": 0.4021065831184387, + "learning_rate": 9.159409553234406e-05, + "loss": 0.1117, + "step": 22477 + }, + { + "epoch": 0.5269197184666866, + "grad_norm": 0.24664290249347687, + "learning_rate": 9.15867554678111e-05, + "loss": 0.0307, + "step": 22478 + }, + { + "epoch": 0.5269431600414916, + "grad_norm": 0.17950937151908875, + "learning_rate": 9.157941544892865e-05, + "loss": 0.0134, + "step": 22479 + }, + { + "epoch": 0.5269666016162966, + "grad_norm": 0.8688785433769226, + "learning_rate": 9.157207547573652e-05, + "loss": 0.6255, + "step": 22480 + }, + { + "epoch": 0.5269900431911015, + "grad_norm": 0.3590347170829773, + "learning_rate": 9.156473554827454e-05, + "loss": 0.0646, + "step": 22481 + }, + { + "epoch": 0.5270134847659066, + "grad_norm": 0.49584031105041504, + "learning_rate": 9.155739566658256e-05, + "loss": 0.11, + "step": 22482 + }, + { + "epoch": 0.5270369263407115, + "grad_norm": 0.519548773765564, + "learning_rate": 9.155005583070038e-05, + "loss": 0.555, + "step": 22483 + }, + { + "epoch": 0.5270603679155166, + "grad_norm": 0.11336284130811691, + "learning_rate": 9.154271604066782e-05, + "loss": 0.0221, + "step": 22484 + }, + { + "epoch": 0.5270838094903215, + "grad_norm": 0.5854629874229431, + "learning_rate": 9.153537629652471e-05, + "loss": 0.0989, + "step": 22485 + }, + { + "epoch": 0.5271072510651266, + "grad_norm": 0.41814687848091125, + "learning_rate": 9.152803659831092e-05, + "loss": 0.1254, + "step": 22486 + }, + { + "epoch": 0.5271306926399315, + "grad_norm": 0.175177663564682, + "learning_rate": 9.152069694606623e-05, + "loss": 0.0349, + "step": 22487 + }, + { + "epoch": 0.5271541342147366, + "grad_norm": 0.34552741050720215, + "learning_rate": 9.151335733983049e-05, + "loss": 0.0731, + "step": 22488 + }, + { + "epoch": 0.5271775757895415, + "grad_norm": 0.38623541593551636, + "learning_rate": 9.150601777964351e-05, + "loss": 0.0746, + "step": 22489 + }, + { + "epoch": 0.5272010173643465, + "grad_norm": 1.0142009258270264, + "learning_rate": 9.149867826554508e-05, + "loss": 0.2026, + "step": 22490 + }, + { + "epoch": 0.5272244589391515, + "grad_norm": 0.5053521990776062, + "learning_rate": 9.149133879757511e-05, + "loss": 0.0622, + "step": 22491 + }, + { + "epoch": 0.5272479005139565, + "grad_norm": 0.4290541112422943, + "learning_rate": 9.148399937577336e-05, + "loss": 0.0453, + "step": 22492 + }, + { + "epoch": 0.5272713420887615, + "grad_norm": 0.2995300590991974, + "learning_rate": 9.147666000017965e-05, + "loss": 0.0666, + "step": 22493 + }, + { + "epoch": 0.5272947836635665, + "grad_norm": 0.40747886896133423, + "learning_rate": 9.146932067083384e-05, + "loss": 0.0615, + "step": 22494 + }, + { + "epoch": 0.5273182252383715, + "grad_norm": 0.5310425162315369, + "learning_rate": 9.146198138777574e-05, + "loss": 0.0831, + "step": 22495 + }, + { + "epoch": 0.5273416668131765, + "grad_norm": 0.14945511519908905, + "learning_rate": 9.145464215104512e-05, + "loss": 0.0201, + "step": 22496 + }, + { + "epoch": 0.5273651083879815, + "grad_norm": 0.6577538251876831, + "learning_rate": 9.144730296068189e-05, + "loss": 0.0531, + "step": 22497 + }, + { + "epoch": 0.5273885499627865, + "grad_norm": 0.09897755831480026, + "learning_rate": 9.143996381672584e-05, + "loss": 0.0136, + "step": 22498 + }, + { + "epoch": 0.5274119915375914, + "grad_norm": 0.4818022549152374, + "learning_rate": 9.143262471921679e-05, + "loss": 0.0731, + "step": 22499 + }, + { + "epoch": 0.5274354331123965, + "grad_norm": 0.5253801941871643, + "learning_rate": 9.142528566819454e-05, + "loss": 0.1301, + "step": 22500 + }, + { + "epoch": 0.5274588746872015, + "grad_norm": 0.22679641842842102, + "learning_rate": 9.141794666369895e-05, + "loss": 0.0284, + "step": 22501 + }, + { + "epoch": 0.5274823162620065, + "grad_norm": 0.5782351493835449, + "learning_rate": 9.141060770576983e-05, + "loss": 0.6299, + "step": 22502 + }, + { + "epoch": 0.5275057578368115, + "grad_norm": 0.3942677080631256, + "learning_rate": 9.140326879444697e-05, + "loss": 0.0757, + "step": 22503 + }, + { + "epoch": 0.5275291994116165, + "grad_norm": 0.5251184105873108, + "learning_rate": 9.139592992977023e-05, + "loss": 0.0665, + "step": 22504 + }, + { + "epoch": 0.5275526409864215, + "grad_norm": 0.1852346658706665, + "learning_rate": 9.138859111177942e-05, + "loss": 0.0392, + "step": 22505 + }, + { + "epoch": 0.5275760825612265, + "grad_norm": 0.6816840171813965, + "learning_rate": 9.138125234051434e-05, + "loss": 0.0944, + "step": 22506 + }, + { + "epoch": 0.5275995241360315, + "grad_norm": 0.9191866517066956, + "learning_rate": 9.137391361601484e-05, + "loss": 0.1402, + "step": 22507 + }, + { + "epoch": 0.5276229657108364, + "grad_norm": 0.4500378668308258, + "learning_rate": 9.136657493832069e-05, + "loss": 0.1006, + "step": 22508 + }, + { + "epoch": 0.5276464072856415, + "grad_norm": 0.5011281371116638, + "learning_rate": 9.135923630747179e-05, + "loss": 0.0672, + "step": 22509 + }, + { + "epoch": 0.5276698488604464, + "grad_norm": 0.47180792689323425, + "learning_rate": 9.135189772350789e-05, + "loss": 0.0746, + "step": 22510 + }, + { + "epoch": 0.5276932904352515, + "grad_norm": 0.4108973741531372, + "learning_rate": 9.134455918646886e-05, + "loss": 0.0815, + "step": 22511 + }, + { + "epoch": 0.5277167320100564, + "grad_norm": 0.546562671661377, + "learning_rate": 9.133722069639449e-05, + "loss": 0.0605, + "step": 22512 + }, + { + "epoch": 0.5277401735848615, + "grad_norm": 0.28013285994529724, + "learning_rate": 9.132988225332459e-05, + "loss": 0.0407, + "step": 22513 + }, + { + "epoch": 0.5277636151596664, + "grad_norm": 0.5824460983276367, + "learning_rate": 9.132254385729902e-05, + "loss": 0.1277, + "step": 22514 + }, + { + "epoch": 0.5277870567344715, + "grad_norm": 0.7265088558197021, + "learning_rate": 9.131520550835755e-05, + "loss": 0.1075, + "step": 22515 + }, + { + "epoch": 0.5278104983092764, + "grad_norm": 0.4954991042613983, + "learning_rate": 9.130786720654003e-05, + "loss": 0.0504, + "step": 22516 + }, + { + "epoch": 0.5278339398840814, + "grad_norm": 0.3523952066898346, + "learning_rate": 9.130052895188626e-05, + "loss": 0.5086, + "step": 22517 + }, + { + "epoch": 0.5278573814588864, + "grad_norm": 0.37012091279029846, + "learning_rate": 9.129319074443608e-05, + "loss": 0.0416, + "step": 22518 + }, + { + "epoch": 0.5278808230336914, + "grad_norm": 0.561802327632904, + "learning_rate": 9.128585258422927e-05, + "loss": 0.0825, + "step": 22519 + }, + { + "epoch": 0.5279042646084964, + "grad_norm": 0.27441656589508057, + "learning_rate": 9.127851447130565e-05, + "loss": 0.0593, + "step": 22520 + }, + { + "epoch": 0.5279277061833014, + "grad_norm": 0.3561156988143921, + "learning_rate": 9.127117640570508e-05, + "loss": 0.3678, + "step": 22521 + }, + { + "epoch": 0.5279511477581064, + "grad_norm": 0.3751300871372223, + "learning_rate": 9.126383838746736e-05, + "loss": 0.0706, + "step": 22522 + }, + { + "epoch": 0.5279745893329114, + "grad_norm": 0.26242542266845703, + "learning_rate": 9.125650041663228e-05, + "loss": 0.0303, + "step": 22523 + }, + { + "epoch": 0.5279980309077164, + "grad_norm": 0.28249526023864746, + "learning_rate": 9.124916249323971e-05, + "loss": 0.0252, + "step": 22524 + }, + { + "epoch": 0.5280214724825214, + "grad_norm": 0.6520360708236694, + "learning_rate": 9.124182461732938e-05, + "loss": 0.1298, + "step": 22525 + }, + { + "epoch": 0.5280449140573263, + "grad_norm": 0.3233683109283447, + "learning_rate": 9.12344867889412e-05, + "loss": 0.0729, + "step": 22526 + }, + { + "epoch": 0.5280683556321314, + "grad_norm": 0.38157397508621216, + "learning_rate": 9.122714900811493e-05, + "loss": 0.0655, + "step": 22527 + }, + { + "epoch": 0.5280917972069363, + "grad_norm": 0.47632142901420593, + "learning_rate": 9.121981127489039e-05, + "loss": 0.08, + "step": 22528 + }, + { + "epoch": 0.5281152387817414, + "grad_norm": 0.4276581406593323, + "learning_rate": 9.12124735893074e-05, + "loss": 0.0646, + "step": 22529 + }, + { + "epoch": 0.5281386803565463, + "grad_norm": 0.09407822787761688, + "learning_rate": 9.120513595140579e-05, + "loss": 0.0135, + "step": 22530 + }, + { + "epoch": 0.5281621219313514, + "grad_norm": 0.419828861951828, + "learning_rate": 9.119779836122534e-05, + "loss": 0.045, + "step": 22531 + }, + { + "epoch": 0.5281855635061563, + "grad_norm": 0.6641074419021606, + "learning_rate": 9.11904608188059e-05, + "loss": 0.1296, + "step": 22532 + }, + { + "epoch": 0.5282090050809614, + "grad_norm": 0.2681846022605896, + "learning_rate": 9.118312332418722e-05, + "loss": 0.0671, + "step": 22533 + }, + { + "epoch": 0.5282324466557663, + "grad_norm": 0.33228403329849243, + "learning_rate": 9.11757858774092e-05, + "loss": 0.0702, + "step": 22534 + }, + { + "epoch": 0.5282558882305713, + "grad_norm": 0.6137348413467407, + "learning_rate": 9.11684484785116e-05, + "loss": 0.0991, + "step": 22535 + }, + { + "epoch": 0.5282793298053763, + "grad_norm": 0.6149294376373291, + "learning_rate": 9.116111112753428e-05, + "loss": 0.1073, + "step": 22536 + }, + { + "epoch": 0.5283027713801813, + "grad_norm": 0.280708372592926, + "learning_rate": 9.115377382451701e-05, + "loss": 0.052, + "step": 22537 + }, + { + "epoch": 0.5283262129549863, + "grad_norm": 0.34310102462768555, + "learning_rate": 9.114643656949958e-05, + "loss": 0.0524, + "step": 22538 + }, + { + "epoch": 0.5283496545297913, + "grad_norm": 0.25273117423057556, + "learning_rate": 9.113909936252187e-05, + "loss": 0.0307, + "step": 22539 + }, + { + "epoch": 0.5283730961045963, + "grad_norm": 0.3627215325832367, + "learning_rate": 9.113176220362365e-05, + "loss": 0.0831, + "step": 22540 + }, + { + "epoch": 0.5283965376794013, + "grad_norm": 0.5654078722000122, + "learning_rate": 9.112442509284472e-05, + "loss": 0.0993, + "step": 22541 + }, + { + "epoch": 0.5284199792542063, + "grad_norm": 0.5904461741447449, + "learning_rate": 9.111708803022492e-05, + "loss": 0.1002, + "step": 22542 + }, + { + "epoch": 0.5284434208290113, + "grad_norm": 0.39517608284950256, + "learning_rate": 9.110975101580405e-05, + "loss": 0.0619, + "step": 22543 + }, + { + "epoch": 0.5284668624038162, + "grad_norm": 0.4950483739376068, + "learning_rate": 9.110241404962192e-05, + "loss": 0.0723, + "step": 22544 + }, + { + "epoch": 0.5284903039786213, + "grad_norm": 1.0809205770492554, + "learning_rate": 9.10950771317183e-05, + "loss": 0.1225, + "step": 22545 + }, + { + "epoch": 0.5285137455534262, + "grad_norm": 0.5824036598205566, + "learning_rate": 9.108774026213309e-05, + "loss": 0.0528, + "step": 22546 + }, + { + "epoch": 0.5285371871282313, + "grad_norm": 0.5394022464752197, + "learning_rate": 9.108040344090606e-05, + "loss": 0.0423, + "step": 22547 + }, + { + "epoch": 0.5285606287030362, + "grad_norm": 0.09418050199747086, + "learning_rate": 9.107306666807698e-05, + "loss": 0.0176, + "step": 22548 + }, + { + "epoch": 0.5285840702778413, + "grad_norm": 0.7706369757652283, + "learning_rate": 9.106572994368572e-05, + "loss": 0.0705, + "step": 22549 + }, + { + "epoch": 0.5286075118526462, + "grad_norm": 0.19232605397701263, + "learning_rate": 9.105839326777206e-05, + "loss": 0.0372, + "step": 22550 + }, + { + "epoch": 0.5286309534274513, + "grad_norm": 0.5219674110412598, + "learning_rate": 9.105105664037577e-05, + "loss": 0.0681, + "step": 22551 + }, + { + "epoch": 0.5286543950022563, + "grad_norm": 0.7393280267715454, + "learning_rate": 9.104372006153674e-05, + "loss": 0.0674, + "step": 22552 + }, + { + "epoch": 0.5286778365770612, + "grad_norm": 0.6304383277893066, + "learning_rate": 9.103638353129474e-05, + "loss": 0.1363, + "step": 22553 + }, + { + "epoch": 0.5287012781518663, + "grad_norm": 0.16635465621948242, + "learning_rate": 9.102904704968954e-05, + "loss": 0.0479, + "step": 22554 + }, + { + "epoch": 0.5287247197266712, + "grad_norm": 0.27264079451560974, + "learning_rate": 9.102171061676101e-05, + "loss": 0.0455, + "step": 22555 + }, + { + "epoch": 0.5287481613014763, + "grad_norm": 0.48055997490882874, + "learning_rate": 9.10143742325489e-05, + "loss": 0.0643, + "step": 22556 + }, + { + "epoch": 0.5287716028762812, + "grad_norm": 0.5167385339736938, + "learning_rate": 9.100703789709306e-05, + "loss": 0.0678, + "step": 22557 + }, + { + "epoch": 0.5287950444510863, + "grad_norm": 0.4188370108604431, + "learning_rate": 9.099970161043329e-05, + "loss": 0.0579, + "step": 22558 + }, + { + "epoch": 0.5288184860258912, + "grad_norm": 0.3517085313796997, + "learning_rate": 9.09923653726094e-05, + "loss": 0.0708, + "step": 22559 + }, + { + "epoch": 0.5288419276006963, + "grad_norm": 0.4934426248073578, + "learning_rate": 9.09850291836612e-05, + "loss": 0.0842, + "step": 22560 + }, + { + "epoch": 0.5288653691755012, + "grad_norm": 0.402022123336792, + "learning_rate": 9.097769304362846e-05, + "loss": 0.0646, + "step": 22561 + }, + { + "epoch": 0.5288888107503062, + "grad_norm": 0.42714565992355347, + "learning_rate": 9.097035695255103e-05, + "loss": 0.0928, + "step": 22562 + }, + { + "epoch": 0.5289122523251112, + "grad_norm": 0.6473987102508545, + "learning_rate": 9.096302091046867e-05, + "loss": 0.1647, + "step": 22563 + }, + { + "epoch": 0.5289356938999162, + "grad_norm": 0.29791906476020813, + "learning_rate": 9.095568491742125e-05, + "loss": 0.2336, + "step": 22564 + }, + { + "epoch": 0.5289591354747212, + "grad_norm": 0.4376683235168457, + "learning_rate": 9.094834897344851e-05, + "loss": 0.0824, + "step": 22565 + }, + { + "epoch": 0.5289825770495262, + "grad_norm": 0.8463603854179382, + "learning_rate": 9.094101307859029e-05, + "loss": 0.1807, + "step": 22566 + }, + { + "epoch": 0.5290060186243312, + "grad_norm": 0.1622058004140854, + "learning_rate": 9.093367723288639e-05, + "loss": 0.0253, + "step": 22567 + }, + { + "epoch": 0.5290294601991362, + "grad_norm": 0.667337954044342, + "learning_rate": 9.092634143637663e-05, + "loss": 0.1252, + "step": 22568 + }, + { + "epoch": 0.5290529017739412, + "grad_norm": 0.35901081562042236, + "learning_rate": 9.091900568910075e-05, + "loss": 0.0515, + "step": 22569 + }, + { + "epoch": 0.5290763433487462, + "grad_norm": 0.7892700433731079, + "learning_rate": 9.091166999109861e-05, + "loss": 0.229, + "step": 22570 + }, + { + "epoch": 0.5290997849235511, + "grad_norm": 0.5094596147537231, + "learning_rate": 9.090433434241004e-05, + "loss": 0.0628, + "step": 22571 + }, + { + "epoch": 0.5291232264983562, + "grad_norm": 0.4945354163646698, + "learning_rate": 9.089699874307479e-05, + "loss": 0.045, + "step": 22572 + }, + { + "epoch": 0.5291466680731611, + "grad_norm": 0.47798046469688416, + "learning_rate": 9.088966319313268e-05, + "loss": 0.124, + "step": 22573 + }, + { + "epoch": 0.5291701096479662, + "grad_norm": 0.2463950514793396, + "learning_rate": 9.088232769262351e-05, + "loss": 0.0381, + "step": 22574 + }, + { + "epoch": 0.5291935512227711, + "grad_norm": 0.39282888174057007, + "learning_rate": 9.087499224158711e-05, + "loss": 0.0768, + "step": 22575 + }, + { + "epoch": 0.5292169927975762, + "grad_norm": 0.15421557426452637, + "learning_rate": 9.086765684006323e-05, + "loss": 0.036, + "step": 22576 + }, + { + "epoch": 0.5292404343723811, + "grad_norm": 0.4193587601184845, + "learning_rate": 9.086032148809172e-05, + "loss": 0.0571, + "step": 22577 + }, + { + "epoch": 0.5292638759471862, + "grad_norm": 0.34648269414901733, + "learning_rate": 9.085298618571236e-05, + "loss": 0.252, + "step": 22578 + }, + { + "epoch": 0.5292873175219911, + "grad_norm": 0.15209712088108063, + "learning_rate": 9.084565093296493e-05, + "loss": 0.0348, + "step": 22579 + }, + { + "epoch": 0.5293107590967961, + "grad_norm": 0.6237833499908447, + "learning_rate": 9.08383157298893e-05, + "loss": 0.4748, + "step": 22580 + }, + { + "epoch": 0.5293342006716011, + "grad_norm": 0.27047815918922424, + "learning_rate": 9.083098057652515e-05, + "loss": 0.0379, + "step": 22581 + }, + { + "epoch": 0.5293576422464061, + "grad_norm": 0.1279347687959671, + "learning_rate": 9.082364547291242e-05, + "loss": 0.0163, + "step": 22582 + }, + { + "epoch": 0.5293810838212111, + "grad_norm": 0.6185315251350403, + "learning_rate": 9.081631041909083e-05, + "loss": 0.1816, + "step": 22583 + }, + { + "epoch": 0.5294045253960161, + "grad_norm": 0.16749508678913116, + "learning_rate": 9.08089754151002e-05, + "loss": 0.024, + "step": 22584 + }, + { + "epoch": 0.5294279669708211, + "grad_norm": 0.6184630990028381, + "learning_rate": 9.080164046098034e-05, + "loss": 0.1596, + "step": 22585 + }, + { + "epoch": 0.5294514085456261, + "grad_norm": 0.4008861780166626, + "learning_rate": 9.079430555677102e-05, + "loss": 0.0721, + "step": 22586 + }, + { + "epoch": 0.529474850120431, + "grad_norm": 0.19518956542015076, + "learning_rate": 9.078697070251207e-05, + "loss": 0.0259, + "step": 22587 + }, + { + "epoch": 0.5294982916952361, + "grad_norm": 0.10712389647960663, + "learning_rate": 9.077963589824329e-05, + "loss": 0.0194, + "step": 22588 + }, + { + "epoch": 0.529521733270041, + "grad_norm": 0.40304023027420044, + "learning_rate": 9.077230114400442e-05, + "loss": 0.5517, + "step": 22589 + }, + { + "epoch": 0.5295451748448461, + "grad_norm": 1.0438579320907593, + "learning_rate": 9.076496643983535e-05, + "loss": 0.2251, + "step": 22590 + }, + { + "epoch": 0.529568616419651, + "grad_norm": 0.25478410720825195, + "learning_rate": 9.075763178577581e-05, + "loss": 0.0493, + "step": 22591 + }, + { + "epoch": 0.5295920579944561, + "grad_norm": 0.47873467206954956, + "learning_rate": 9.07502971818656e-05, + "loss": 0.1059, + "step": 22592 + }, + { + "epoch": 0.529615499569261, + "grad_norm": 0.2881753146648407, + "learning_rate": 9.074296262814453e-05, + "loss": 0.0309, + "step": 22593 + }, + { + "epoch": 0.5296389411440661, + "grad_norm": 0.5197072625160217, + "learning_rate": 9.073562812465245e-05, + "loss": 0.0829, + "step": 22594 + }, + { + "epoch": 0.529662382718871, + "grad_norm": 0.9258034825325012, + "learning_rate": 9.07282936714291e-05, + "loss": 0.1399, + "step": 22595 + }, + { + "epoch": 0.529685824293676, + "grad_norm": 0.4132530093193054, + "learning_rate": 9.072095926851427e-05, + "loss": 0.0684, + "step": 22596 + }, + { + "epoch": 0.529709265868481, + "grad_norm": 0.526017427444458, + "learning_rate": 9.07136249159478e-05, + "loss": 0.2132, + "step": 22597 + }, + { + "epoch": 0.529732707443286, + "grad_norm": 0.5953244566917419, + "learning_rate": 9.070629061376947e-05, + "loss": 0.4698, + "step": 22598 + }, + { + "epoch": 0.529756149018091, + "grad_norm": 0.389804869890213, + "learning_rate": 9.069895636201904e-05, + "loss": 0.071, + "step": 22599 + }, + { + "epoch": 0.529779590592896, + "grad_norm": 0.44698792695999146, + "learning_rate": 9.069162216073636e-05, + "loss": 0.0827, + "step": 22600 + }, + { + "epoch": 0.529803032167701, + "grad_norm": 0.429398775100708, + "learning_rate": 9.068428800996116e-05, + "loss": 0.0577, + "step": 22601 + }, + { + "epoch": 0.529826473742506, + "grad_norm": 0.6934991478919983, + "learning_rate": 9.067695390973332e-05, + "loss": 0.0902, + "step": 22602 + }, + { + "epoch": 0.5298499153173111, + "grad_norm": 0.4148961305618286, + "learning_rate": 9.066961986009257e-05, + "loss": 0.098, + "step": 22603 + }, + { + "epoch": 0.529873356892116, + "grad_norm": 0.3404313027858734, + "learning_rate": 9.066228586107873e-05, + "loss": 0.0719, + "step": 22604 + }, + { + "epoch": 0.5298967984669211, + "grad_norm": 0.11357269436120987, + "learning_rate": 9.065495191273154e-05, + "loss": 0.0176, + "step": 22605 + }, + { + "epoch": 0.529920240041726, + "grad_norm": 0.6092438697814941, + "learning_rate": 9.06476180150909e-05, + "loss": 0.0965, + "step": 22606 + }, + { + "epoch": 0.529943681616531, + "grad_norm": 0.5789566040039062, + "learning_rate": 9.064028416819654e-05, + "loss": 0.1003, + "step": 22607 + }, + { + "epoch": 0.529967123191336, + "grad_norm": 0.6179075837135315, + "learning_rate": 9.063295037208826e-05, + "loss": 0.083, + "step": 22608 + }, + { + "epoch": 0.529990564766141, + "grad_norm": 0.23049794137477875, + "learning_rate": 9.062561662680586e-05, + "loss": 0.028, + "step": 22609 + }, + { + "epoch": 0.530014006340946, + "grad_norm": 0.38432013988494873, + "learning_rate": 9.061828293238912e-05, + "loss": 0.0782, + "step": 22610 + }, + { + "epoch": 0.530037447915751, + "grad_norm": 0.9706987738609314, + "learning_rate": 9.061094928887784e-05, + "loss": 0.1552, + "step": 22611 + }, + { + "epoch": 0.530060889490556, + "grad_norm": 0.1214715987443924, + "learning_rate": 9.060361569631182e-05, + "loss": 0.0175, + "step": 22612 + }, + { + "epoch": 0.530084331065361, + "grad_norm": 0.8026210069656372, + "learning_rate": 9.059628215473084e-05, + "loss": 0.7666, + "step": 22613 + }, + { + "epoch": 0.530107772640166, + "grad_norm": 0.23461847007274628, + "learning_rate": 9.058894866417469e-05, + "loss": 0.0594, + "step": 22614 + }, + { + "epoch": 0.530131214214971, + "grad_norm": 0.209276020526886, + "learning_rate": 9.058161522468317e-05, + "loss": 0.0355, + "step": 22615 + }, + { + "epoch": 0.5301546557897759, + "grad_norm": 0.42448416352272034, + "learning_rate": 9.05742818362961e-05, + "loss": 0.0605, + "step": 22616 + }, + { + "epoch": 0.530178097364581, + "grad_norm": 0.32093197107315063, + "learning_rate": 9.056694849905321e-05, + "loss": 0.0527, + "step": 22617 + }, + { + "epoch": 0.5302015389393859, + "grad_norm": 0.7359897494316101, + "learning_rate": 9.05596152129943e-05, + "loss": 0.0857, + "step": 22618 + }, + { + "epoch": 0.530224980514191, + "grad_norm": 0.09442243725061417, + "learning_rate": 9.055228197815922e-05, + "loss": 0.02, + "step": 22619 + }, + { + "epoch": 0.5302484220889959, + "grad_norm": 0.2344633936882019, + "learning_rate": 9.054494879458773e-05, + "loss": 0.0321, + "step": 22620 + }, + { + "epoch": 0.530271863663801, + "grad_norm": 0.2705893814563751, + "learning_rate": 9.053761566231959e-05, + "loss": 0.0394, + "step": 22621 + }, + { + "epoch": 0.5302953052386059, + "grad_norm": 0.4820016324520111, + "learning_rate": 9.053028258139463e-05, + "loss": 0.1246, + "step": 22622 + }, + { + "epoch": 0.530318746813411, + "grad_norm": 0.10681835561990738, + "learning_rate": 9.052294955185262e-05, + "loss": 0.0057, + "step": 22623 + }, + { + "epoch": 0.5303421883882159, + "grad_norm": 0.7964795827865601, + "learning_rate": 9.051561657373334e-05, + "loss": 0.1998, + "step": 22624 + }, + { + "epoch": 0.530365629963021, + "grad_norm": 0.6578893661499023, + "learning_rate": 9.050828364707662e-05, + "loss": 0.1096, + "step": 22625 + }, + { + "epoch": 0.5303890715378259, + "grad_norm": 0.5852024555206299, + "learning_rate": 9.05009507719222e-05, + "loss": 0.6173, + "step": 22626 + }, + { + "epoch": 0.5304125131126309, + "grad_norm": 0.4566068947315216, + "learning_rate": 9.049361794830989e-05, + "loss": 0.0761, + "step": 22627 + }, + { + "epoch": 0.5304359546874359, + "grad_norm": 0.7151980996131897, + "learning_rate": 9.048628517627948e-05, + "loss": 0.1145, + "step": 22628 + }, + { + "epoch": 0.5304593962622409, + "grad_norm": 0.4751600921154022, + "learning_rate": 9.047895245587076e-05, + "loss": 0.0587, + "step": 22629 + }, + { + "epoch": 0.5304828378370459, + "grad_norm": 0.3359461724758148, + "learning_rate": 9.047161978712346e-05, + "loss": 0.2067, + "step": 22630 + }, + { + "epoch": 0.5305062794118509, + "grad_norm": 0.30754125118255615, + "learning_rate": 9.046428717007745e-05, + "loss": 0.0546, + "step": 22631 + }, + { + "epoch": 0.5305297209866559, + "grad_norm": 0.2674303948879242, + "learning_rate": 9.04569546047725e-05, + "loss": 0.043, + "step": 22632 + }, + { + "epoch": 0.5305531625614609, + "grad_norm": 0.5565329194068909, + "learning_rate": 9.04496220912484e-05, + "loss": 0.1162, + "step": 22633 + }, + { + "epoch": 0.5305766041362658, + "grad_norm": 0.22991938889026642, + "learning_rate": 9.044228962954489e-05, + "loss": 0.0713, + "step": 22634 + }, + { + "epoch": 0.5306000457110709, + "grad_norm": 0.5388286709785461, + "learning_rate": 9.04349572197018e-05, + "loss": 0.1162, + "step": 22635 + }, + { + "epoch": 0.5306234872858758, + "grad_norm": 0.6026670336723328, + "learning_rate": 9.04276248617589e-05, + "loss": 0.1558, + "step": 22636 + }, + { + "epoch": 0.5306469288606809, + "grad_norm": 0.33756229281425476, + "learning_rate": 9.042029255575595e-05, + "loss": 0.0869, + "step": 22637 + }, + { + "epoch": 0.5306703704354858, + "grad_norm": 0.5446579456329346, + "learning_rate": 9.04129603017328e-05, + "loss": 0.0999, + "step": 22638 + }, + { + "epoch": 0.5306938120102909, + "grad_norm": 0.27201205492019653, + "learning_rate": 9.040562809972917e-05, + "loss": 0.0513, + "step": 22639 + }, + { + "epoch": 0.5307172535850958, + "grad_norm": 0.4390787184238434, + "learning_rate": 9.03982959497849e-05, + "loss": 0.0959, + "step": 22640 + }, + { + "epoch": 0.5307406951599009, + "grad_norm": 0.39653706550598145, + "learning_rate": 9.039096385193973e-05, + "loss": 0.0714, + "step": 22641 + }, + { + "epoch": 0.5307641367347058, + "grad_norm": 0.41354867815971375, + "learning_rate": 9.038363180623342e-05, + "loss": 0.0935, + "step": 22642 + }, + { + "epoch": 0.5307875783095108, + "grad_norm": 0.22630241513252258, + "learning_rate": 9.037629981270582e-05, + "loss": 0.0338, + "step": 22643 + }, + { + "epoch": 0.5308110198843158, + "grad_norm": 0.273226797580719, + "learning_rate": 9.036896787139672e-05, + "loss": 0.0636, + "step": 22644 + }, + { + "epoch": 0.5308344614591208, + "grad_norm": 0.4644102454185486, + "learning_rate": 9.036163598234586e-05, + "loss": 0.0998, + "step": 22645 + }, + { + "epoch": 0.5308579030339258, + "grad_norm": 0.28797465562820435, + "learning_rate": 9.035430414559301e-05, + "loss": 0.0693, + "step": 22646 + }, + { + "epoch": 0.5308813446087308, + "grad_norm": 0.45106586813926697, + "learning_rate": 9.034697236117799e-05, + "loss": 0.1099, + "step": 22647 + }, + { + "epoch": 0.5309047861835358, + "grad_norm": 0.671511709690094, + "learning_rate": 9.033964062914059e-05, + "loss": 0.1787, + "step": 22648 + }, + { + "epoch": 0.5309282277583408, + "grad_norm": 0.380607008934021, + "learning_rate": 9.033230894952054e-05, + "loss": 0.1028, + "step": 22649 + }, + { + "epoch": 0.5309516693331457, + "grad_norm": 0.37397095561027527, + "learning_rate": 9.032497732235767e-05, + "loss": 0.1028, + "step": 22650 + }, + { + "epoch": 0.5309751109079508, + "grad_norm": 0.31539106369018555, + "learning_rate": 9.031764574769176e-05, + "loss": 0.0693, + "step": 22651 + }, + { + "epoch": 0.5309985524827557, + "grad_norm": 0.5566034913063049, + "learning_rate": 9.031031422556254e-05, + "loss": 0.0893, + "step": 22652 + }, + { + "epoch": 0.5310219940575608, + "grad_norm": 0.41317474842071533, + "learning_rate": 9.030298275600986e-05, + "loss": 0.0374, + "step": 22653 + }, + { + "epoch": 0.5310454356323658, + "grad_norm": 0.45495179295539856, + "learning_rate": 9.029565133907343e-05, + "loss": 0.0626, + "step": 22654 + }, + { + "epoch": 0.5310688772071708, + "grad_norm": 0.7430252432823181, + "learning_rate": 9.02883199747931e-05, + "loss": 0.1956, + "step": 22655 + }, + { + "epoch": 0.5310923187819758, + "grad_norm": 0.39432841539382935, + "learning_rate": 9.02809886632086e-05, + "loss": 0.0891, + "step": 22656 + }, + { + "epoch": 0.5311157603567808, + "grad_norm": 0.38297370076179504, + "learning_rate": 9.027365740435974e-05, + "loss": 0.0836, + "step": 22657 + }, + { + "epoch": 0.5311392019315858, + "grad_norm": 0.6667554378509521, + "learning_rate": 9.02663261982863e-05, + "loss": 0.5861, + "step": 22658 + }, + { + "epoch": 0.5311626435063908, + "grad_norm": 0.3656706213951111, + "learning_rate": 9.025899504502802e-05, + "loss": 0.0853, + "step": 22659 + }, + { + "epoch": 0.5311860850811958, + "grad_norm": 0.33120661973953247, + "learning_rate": 9.025166394462474e-05, + "loss": 0.0508, + "step": 22660 + }, + { + "epoch": 0.5312095266560007, + "grad_norm": 0.4299107491970062, + "learning_rate": 9.02443328971162e-05, + "loss": 0.074, + "step": 22661 + }, + { + "epoch": 0.5312329682308058, + "grad_norm": 0.36046212911605835, + "learning_rate": 9.023700190254217e-05, + "loss": 0.041, + "step": 22662 + }, + { + "epoch": 0.5312564098056107, + "grad_norm": 0.22204796969890594, + "learning_rate": 9.022967096094245e-05, + "loss": 0.0327, + "step": 22663 + }, + { + "epoch": 0.5312798513804158, + "grad_norm": 0.5150210857391357, + "learning_rate": 9.022234007235681e-05, + "loss": 0.0883, + "step": 22664 + }, + { + "epoch": 0.5313032929552207, + "grad_norm": 0.39004677534103394, + "learning_rate": 9.021500923682501e-05, + "loss": 0.0705, + "step": 22665 + }, + { + "epoch": 0.5313267345300258, + "grad_norm": 0.5024076104164124, + "learning_rate": 9.020767845438683e-05, + "loss": 0.1015, + "step": 22666 + }, + { + "epoch": 0.5313501761048307, + "grad_norm": 0.6013475656509399, + "learning_rate": 9.02003477250821e-05, + "loss": 0.2019, + "step": 22667 + }, + { + "epoch": 0.5313736176796358, + "grad_norm": 0.3615729808807373, + "learning_rate": 9.019301704895056e-05, + "loss": 0.0433, + "step": 22668 + }, + { + "epoch": 0.5313970592544407, + "grad_norm": 0.37874072790145874, + "learning_rate": 9.018568642603196e-05, + "loss": 0.0775, + "step": 22669 + }, + { + "epoch": 0.5314205008292457, + "grad_norm": 0.5005304217338562, + "learning_rate": 9.017835585636612e-05, + "loss": 0.0656, + "step": 22670 + }, + { + "epoch": 0.5314439424040507, + "grad_norm": 0.5854547619819641, + "learning_rate": 9.01710253399928e-05, + "loss": 0.0665, + "step": 22671 + }, + { + "epoch": 0.5314673839788557, + "grad_norm": 0.4791566729545593, + "learning_rate": 9.016369487695175e-05, + "loss": 0.0626, + "step": 22672 + }, + { + "epoch": 0.5314908255536607, + "grad_norm": 0.2799932360649109, + "learning_rate": 9.015636446728279e-05, + "loss": 0.0628, + "step": 22673 + }, + { + "epoch": 0.5315142671284657, + "grad_norm": 0.31100690364837646, + "learning_rate": 9.014903411102567e-05, + "loss": 0.0557, + "step": 22674 + }, + { + "epoch": 0.5315377087032707, + "grad_norm": 0.3780404329299927, + "learning_rate": 9.014170380822016e-05, + "loss": 0.0676, + "step": 22675 + }, + { + "epoch": 0.5315611502780757, + "grad_norm": 0.42013102769851685, + "learning_rate": 9.013437355890606e-05, + "loss": 0.5825, + "step": 22676 + }, + { + "epoch": 0.5315845918528807, + "grad_norm": 0.30691635608673096, + "learning_rate": 9.01270433631231e-05, + "loss": 0.047, + "step": 22677 + }, + { + "epoch": 0.5316080334276857, + "grad_norm": 0.36792314052581787, + "learning_rate": 9.011971322091108e-05, + "loss": 0.0679, + "step": 22678 + }, + { + "epoch": 0.5316314750024906, + "grad_norm": 0.5455208420753479, + "learning_rate": 9.011238313230978e-05, + "loss": 0.1535, + "step": 22679 + }, + { + "epoch": 0.5316549165772957, + "grad_norm": 0.4121691584587097, + "learning_rate": 9.010505309735898e-05, + "loss": 0.0797, + "step": 22680 + }, + { + "epoch": 0.5316783581521006, + "grad_norm": 0.7767332196235657, + "learning_rate": 9.009772311609843e-05, + "loss": 0.0627, + "step": 22681 + }, + { + "epoch": 0.5317017997269057, + "grad_norm": 0.1276315152645111, + "learning_rate": 9.009039318856793e-05, + "loss": 0.0359, + "step": 22682 + }, + { + "epoch": 0.5317252413017106, + "grad_norm": 0.30391454696655273, + "learning_rate": 9.008306331480723e-05, + "loss": 0.0252, + "step": 22683 + }, + { + "epoch": 0.5317486828765157, + "grad_norm": 0.5070129632949829, + "learning_rate": 9.007573349485609e-05, + "loss": 0.0976, + "step": 22684 + }, + { + "epoch": 0.5317721244513206, + "grad_norm": 0.36551955342292786, + "learning_rate": 9.006840372875432e-05, + "loss": 0.0865, + "step": 22685 + }, + { + "epoch": 0.5317955660261257, + "grad_norm": 0.2343035638332367, + "learning_rate": 9.006107401654167e-05, + "loss": 0.0621, + "step": 22686 + }, + { + "epoch": 0.5318190076009306, + "grad_norm": 0.6280043721199036, + "learning_rate": 9.005374435825788e-05, + "loss": 0.084, + "step": 22687 + }, + { + "epoch": 0.5318424491757356, + "grad_norm": 0.23138703405857086, + "learning_rate": 9.004641475394279e-05, + "loss": 0.0449, + "step": 22688 + }, + { + "epoch": 0.5318658907505406, + "grad_norm": 0.4838981032371521, + "learning_rate": 9.003908520363614e-05, + "loss": 0.0428, + "step": 22689 + }, + { + "epoch": 0.5318893323253456, + "grad_norm": 0.4137870967388153, + "learning_rate": 9.003175570737765e-05, + "loss": 0.0695, + "step": 22690 + }, + { + "epoch": 0.5319127739001506, + "grad_norm": 0.22047017514705658, + "learning_rate": 9.002442626520712e-05, + "loss": 0.0393, + "step": 22691 + }, + { + "epoch": 0.5319362154749556, + "grad_norm": 0.07274557650089264, + "learning_rate": 9.001709687716437e-05, + "loss": 0.0108, + "step": 22692 + }, + { + "epoch": 0.5319596570497606, + "grad_norm": 0.5133265256881714, + "learning_rate": 9.000976754328915e-05, + "loss": 0.0567, + "step": 22693 + }, + { + "epoch": 0.5319830986245656, + "grad_norm": 0.5086338520050049, + "learning_rate": 9.000243826362118e-05, + "loss": 0.782, + "step": 22694 + }, + { + "epoch": 0.5320065401993705, + "grad_norm": 0.3346916735172272, + "learning_rate": 8.999510903820028e-05, + "loss": 0.0605, + "step": 22695 + }, + { + "epoch": 0.5320299817741756, + "grad_norm": 0.2603572905063629, + "learning_rate": 8.99877798670662e-05, + "loss": 0.0501, + "step": 22696 + }, + { + "epoch": 0.5320534233489805, + "grad_norm": 0.1582212746143341, + "learning_rate": 8.998045075025871e-05, + "loss": 0.0272, + "step": 22697 + }, + { + "epoch": 0.5320768649237856, + "grad_norm": 0.11966338753700256, + "learning_rate": 8.997312168781758e-05, + "loss": 0.0212, + "step": 22698 + }, + { + "epoch": 0.5321003064985905, + "grad_norm": 0.13867752254009247, + "learning_rate": 8.996579267978255e-05, + "loss": 0.0236, + "step": 22699 + }, + { + "epoch": 0.5321237480733956, + "grad_norm": 0.6717884540557861, + "learning_rate": 8.995846372619343e-05, + "loss": 0.0856, + "step": 22700 + }, + { + "epoch": 0.5321471896482005, + "grad_norm": 0.17489661276340485, + "learning_rate": 8.995113482708997e-05, + "loss": 0.0306, + "step": 22701 + }, + { + "epoch": 0.5321706312230056, + "grad_norm": 0.24768255650997162, + "learning_rate": 8.994380598251193e-05, + "loss": 0.1012, + "step": 22702 + }, + { + "epoch": 0.5321940727978105, + "grad_norm": 0.2004292607307434, + "learning_rate": 8.993647719249904e-05, + "loss": 0.033, + "step": 22703 + }, + { + "epoch": 0.5322175143726156, + "grad_norm": 0.11151626706123352, + "learning_rate": 8.992914845709114e-05, + "loss": 0.0212, + "step": 22704 + }, + { + "epoch": 0.5322409559474206, + "grad_norm": 0.5367186069488525, + "learning_rate": 8.992181977632798e-05, + "loss": 0.0968, + "step": 22705 + }, + { + "epoch": 0.5322643975222255, + "grad_norm": 0.5140998959541321, + "learning_rate": 8.99144911502493e-05, + "loss": 0.0634, + "step": 22706 + }, + { + "epoch": 0.5322878390970306, + "grad_norm": 0.4613790512084961, + "learning_rate": 8.990716257889486e-05, + "loss": 0.0868, + "step": 22707 + }, + { + "epoch": 0.5323112806718355, + "grad_norm": 0.2799144387245178, + "learning_rate": 8.989983406230447e-05, + "loss": 0.0302, + "step": 22708 + }, + { + "epoch": 0.5323347222466406, + "grad_norm": 0.3230119049549103, + "learning_rate": 8.989250560051786e-05, + "loss": 0.0685, + "step": 22709 + }, + { + "epoch": 0.5323581638214455, + "grad_norm": 0.16510914266109467, + "learning_rate": 8.988517719357478e-05, + "loss": 0.02, + "step": 22710 + }, + { + "epoch": 0.5323816053962506, + "grad_norm": 0.5295974016189575, + "learning_rate": 8.987784884151502e-05, + "loss": 0.0715, + "step": 22711 + }, + { + "epoch": 0.5324050469710555, + "grad_norm": 0.3773594796657562, + "learning_rate": 8.987052054437832e-05, + "loss": 0.0467, + "step": 22712 + }, + { + "epoch": 0.5324284885458606, + "grad_norm": 0.11355438083410263, + "learning_rate": 8.986319230220449e-05, + "loss": 0.0175, + "step": 22713 + }, + { + "epoch": 0.5324519301206655, + "grad_norm": 0.1864515095949173, + "learning_rate": 8.985586411503326e-05, + "loss": 0.0438, + "step": 22714 + }, + { + "epoch": 0.5324753716954705, + "grad_norm": 0.21060426533222198, + "learning_rate": 8.984853598290435e-05, + "loss": 0.0508, + "step": 22715 + }, + { + "epoch": 0.5324988132702755, + "grad_norm": 0.39235490560531616, + "learning_rate": 8.984120790585762e-05, + "loss": 0.0523, + "step": 22716 + }, + { + "epoch": 0.5325222548450805, + "grad_norm": 0.20714956521987915, + "learning_rate": 8.983387988393275e-05, + "loss": 0.0375, + "step": 22717 + }, + { + "epoch": 0.5325456964198855, + "grad_norm": 0.27816519141197205, + "learning_rate": 8.982655191716956e-05, + "loss": 0.0483, + "step": 22718 + }, + { + "epoch": 0.5325691379946905, + "grad_norm": 0.6334831118583679, + "learning_rate": 8.981922400560777e-05, + "loss": 0.1153, + "step": 22719 + }, + { + "epoch": 0.5325925795694955, + "grad_norm": 0.4160178601741791, + "learning_rate": 8.981189614928718e-05, + "loss": 0.0756, + "step": 22720 + }, + { + "epoch": 0.5326160211443005, + "grad_norm": 0.1656975895166397, + "learning_rate": 8.98045683482475e-05, + "loss": 0.0247, + "step": 22721 + }, + { + "epoch": 0.5326394627191054, + "grad_norm": 0.12251132726669312, + "learning_rate": 8.979724060252854e-05, + "loss": 0.0171, + "step": 22722 + }, + { + "epoch": 0.5326629042939105, + "grad_norm": 0.10784266144037247, + "learning_rate": 8.978991291217003e-05, + "loss": 0.0158, + "step": 22723 + }, + { + "epoch": 0.5326863458687154, + "grad_norm": 0.5293658971786499, + "learning_rate": 8.978258527721175e-05, + "loss": 0.0905, + "step": 22724 + }, + { + "epoch": 0.5327097874435205, + "grad_norm": 0.14682449400424957, + "learning_rate": 8.977525769769343e-05, + "loss": 0.0127, + "step": 22725 + }, + { + "epoch": 0.5327332290183254, + "grad_norm": 1.2790815830230713, + "learning_rate": 8.976793017365487e-05, + "loss": 0.0878, + "step": 22726 + }, + { + "epoch": 0.5327566705931305, + "grad_norm": 0.49944984912872314, + "learning_rate": 8.976060270513575e-05, + "loss": 0.0996, + "step": 22727 + }, + { + "epoch": 0.5327801121679354, + "grad_norm": 0.4983568787574768, + "learning_rate": 8.975327529217596e-05, + "loss": 0.1149, + "step": 22728 + }, + { + "epoch": 0.5328035537427405, + "grad_norm": 0.6858383417129517, + "learning_rate": 8.974594793481513e-05, + "loss": 0.1589, + "step": 22729 + }, + { + "epoch": 0.5328269953175454, + "grad_norm": 0.48894330859184265, + "learning_rate": 8.973862063309312e-05, + "loss": 0.0862, + "step": 22730 + }, + { + "epoch": 0.5328504368923505, + "grad_norm": 0.2856331765651703, + "learning_rate": 8.973129338704964e-05, + "loss": 0.0491, + "step": 22731 + }, + { + "epoch": 0.5328738784671554, + "grad_norm": 0.3601313531398773, + "learning_rate": 8.972396619672443e-05, + "loss": 0.0605, + "step": 22732 + }, + { + "epoch": 0.5328973200419604, + "grad_norm": 0.12947234511375427, + "learning_rate": 8.971663906215728e-05, + "loss": 0.0185, + "step": 22733 + }, + { + "epoch": 0.5329207616167654, + "grad_norm": 0.30342796444892883, + "learning_rate": 8.970931198338794e-05, + "loss": 0.0658, + "step": 22734 + }, + { + "epoch": 0.5329442031915704, + "grad_norm": 0.10806110501289368, + "learning_rate": 8.970198496045616e-05, + "loss": 0.017, + "step": 22735 + }, + { + "epoch": 0.5329676447663754, + "grad_norm": 0.3626348376274109, + "learning_rate": 8.96946579934017e-05, + "loss": 0.0643, + "step": 22736 + }, + { + "epoch": 0.5329910863411804, + "grad_norm": 0.2025643140077591, + "learning_rate": 8.968733108226432e-05, + "loss": 0.0257, + "step": 22737 + }, + { + "epoch": 0.5330145279159854, + "grad_norm": 0.4232948422431946, + "learning_rate": 8.968000422708375e-05, + "loss": 0.0698, + "step": 22738 + }, + { + "epoch": 0.5330379694907904, + "grad_norm": 0.2020391821861267, + "learning_rate": 8.967267742789974e-05, + "loss": 0.0246, + "step": 22739 + }, + { + "epoch": 0.5330614110655953, + "grad_norm": 0.9943932890892029, + "learning_rate": 8.966535068475211e-05, + "loss": 0.1002, + "step": 22740 + }, + { + "epoch": 0.5330848526404004, + "grad_norm": 0.654761552810669, + "learning_rate": 8.96580239976806e-05, + "loss": 0.11, + "step": 22741 + }, + { + "epoch": 0.5331082942152053, + "grad_norm": 0.2347099483013153, + "learning_rate": 8.965069736672491e-05, + "loss": 0.0395, + "step": 22742 + }, + { + "epoch": 0.5331317357900104, + "grad_norm": 0.385044664144516, + "learning_rate": 8.964337079192484e-05, + "loss": 0.0595, + "step": 22743 + }, + { + "epoch": 0.5331551773648153, + "grad_norm": 0.3588669002056122, + "learning_rate": 8.963604427332014e-05, + "loss": 0.0379, + "step": 22744 + }, + { + "epoch": 0.5331786189396204, + "grad_norm": 0.4332166910171509, + "learning_rate": 8.962871781095053e-05, + "loss": 0.0573, + "step": 22745 + }, + { + "epoch": 0.5332020605144253, + "grad_norm": 0.4173870384693146, + "learning_rate": 8.962139140485582e-05, + "loss": 0.0807, + "step": 22746 + }, + { + "epoch": 0.5332255020892304, + "grad_norm": 0.2909199595451355, + "learning_rate": 8.961406505507572e-05, + "loss": 0.0512, + "step": 22747 + }, + { + "epoch": 0.5332489436640353, + "grad_norm": 0.33676978945732117, + "learning_rate": 8.960673876165e-05, + "loss": 0.0693, + "step": 22748 + }, + { + "epoch": 0.5332723852388404, + "grad_norm": 0.215544193983078, + "learning_rate": 8.95994125246184e-05, + "loss": 0.0359, + "step": 22749 + }, + { + "epoch": 0.5332958268136453, + "grad_norm": 0.1491340547800064, + "learning_rate": 8.959208634402067e-05, + "loss": 0.0218, + "step": 22750 + }, + { + "epoch": 0.5333192683884503, + "grad_norm": 0.2863418459892273, + "learning_rate": 8.958476021989656e-05, + "loss": 0.0249, + "step": 22751 + }, + { + "epoch": 0.5333427099632553, + "grad_norm": 0.5016613006591797, + "learning_rate": 8.957743415228584e-05, + "loss": 0.0544, + "step": 22752 + }, + { + "epoch": 0.5333661515380603, + "grad_norm": 0.20454178750514984, + "learning_rate": 8.957010814122828e-05, + "loss": 0.0334, + "step": 22753 + }, + { + "epoch": 0.5333895931128653, + "grad_norm": 0.8792877197265625, + "learning_rate": 8.956278218676361e-05, + "loss": 0.7067, + "step": 22754 + }, + { + "epoch": 0.5334130346876703, + "grad_norm": 0.5990207195281982, + "learning_rate": 8.955545628893154e-05, + "loss": 0.094, + "step": 22755 + }, + { + "epoch": 0.5334364762624754, + "grad_norm": 0.603545606136322, + "learning_rate": 8.95481304477719e-05, + "loss": 0.1113, + "step": 22756 + }, + { + "epoch": 0.5334599178372803, + "grad_norm": 0.9707116484642029, + "learning_rate": 8.954080466332438e-05, + "loss": 0.8121, + "step": 22757 + }, + { + "epoch": 0.5334833594120854, + "grad_norm": 0.1739872395992279, + "learning_rate": 8.953347893562874e-05, + "loss": 0.0361, + "step": 22758 + }, + { + "epoch": 0.5335068009868903, + "grad_norm": 0.10931946337223053, + "learning_rate": 8.952615326472476e-05, + "loss": 0.0233, + "step": 22759 + }, + { + "epoch": 0.5335302425616953, + "grad_norm": 0.48805901408195496, + "learning_rate": 8.951882765065215e-05, + "loss": 0.0781, + "step": 22760 + }, + { + "epoch": 0.5335536841365003, + "grad_norm": 0.787784218788147, + "learning_rate": 8.951150209345068e-05, + "loss": 0.1451, + "step": 22761 + }, + { + "epoch": 0.5335771257113053, + "grad_norm": 0.7936659455299377, + "learning_rate": 8.95041765931601e-05, + "loss": 0.7339, + "step": 22762 + }, + { + "epoch": 0.5336005672861103, + "grad_norm": 0.6961209177970886, + "learning_rate": 8.949685114982014e-05, + "loss": 0.2007, + "step": 22763 + }, + { + "epoch": 0.5336240088609153, + "grad_norm": 0.383909672498703, + "learning_rate": 8.948952576347053e-05, + "loss": 0.0761, + "step": 22764 + }, + { + "epoch": 0.5336474504357203, + "grad_norm": 0.5656422972679138, + "learning_rate": 8.94822004341511e-05, + "loss": 0.0878, + "step": 22765 + }, + { + "epoch": 0.5336708920105253, + "grad_norm": 0.3270052373409271, + "learning_rate": 8.947487516190154e-05, + "loss": 0.0747, + "step": 22766 + }, + { + "epoch": 0.5336943335853302, + "grad_norm": 0.17368751764297485, + "learning_rate": 8.946754994676158e-05, + "loss": 0.0321, + "step": 22767 + }, + { + "epoch": 0.5337177751601353, + "grad_norm": 0.43249526619911194, + "learning_rate": 8.946022478877101e-05, + "loss": 0.0745, + "step": 22768 + }, + { + "epoch": 0.5337412167349402, + "grad_norm": 0.5111855864524841, + "learning_rate": 8.945289968796957e-05, + "loss": 0.1102, + "step": 22769 + }, + { + "epoch": 0.5337646583097453, + "grad_norm": 0.594323456287384, + "learning_rate": 8.944557464439696e-05, + "loss": 0.1151, + "step": 22770 + }, + { + "epoch": 0.5337880998845502, + "grad_norm": 0.4029034972190857, + "learning_rate": 8.943824965809299e-05, + "loss": 0.0819, + "step": 22771 + }, + { + "epoch": 0.5338115414593553, + "grad_norm": 0.6479580998420715, + "learning_rate": 8.943092472909736e-05, + "loss": 0.1331, + "step": 22772 + }, + { + "epoch": 0.5338349830341602, + "grad_norm": 0.3099082410335541, + "learning_rate": 8.942359985744982e-05, + "loss": 0.0378, + "step": 22773 + }, + { + "epoch": 0.5338584246089653, + "grad_norm": 0.9135605096817017, + "learning_rate": 8.941627504319014e-05, + "loss": 0.1224, + "step": 22774 + }, + { + "epoch": 0.5338818661837702, + "grad_norm": 0.7518431544303894, + "learning_rate": 8.940895028635805e-05, + "loss": 0.1778, + "step": 22775 + }, + { + "epoch": 0.5339053077585753, + "grad_norm": 0.27088385820388794, + "learning_rate": 8.940162558699325e-05, + "loss": 0.0632, + "step": 22776 + }, + { + "epoch": 0.5339287493333802, + "grad_norm": 0.09247621148824692, + "learning_rate": 8.939430094513554e-05, + "loss": 0.0142, + "step": 22777 + }, + { + "epoch": 0.5339521909081852, + "grad_norm": 0.44259408116340637, + "learning_rate": 8.93869763608247e-05, + "loss": 0.0695, + "step": 22778 + }, + { + "epoch": 0.5339756324829902, + "grad_norm": 0.6216095685958862, + "learning_rate": 8.93796518341004e-05, + "loss": 0.1536, + "step": 22779 + }, + { + "epoch": 0.5339990740577952, + "grad_norm": 0.2862398624420166, + "learning_rate": 8.93723273650024e-05, + "loss": 0.0512, + "step": 22780 + }, + { + "epoch": 0.5340225156326002, + "grad_norm": 0.7018606662750244, + "learning_rate": 8.936500295357046e-05, + "loss": 0.0869, + "step": 22781 + }, + { + "epoch": 0.5340459572074052, + "grad_norm": 0.4572046399116516, + "learning_rate": 8.935767859984432e-05, + "loss": 0.0644, + "step": 22782 + }, + { + "epoch": 0.5340693987822102, + "grad_norm": 0.3099846839904785, + "learning_rate": 8.935035430386369e-05, + "loss": 0.0851, + "step": 22783 + }, + { + "epoch": 0.5340928403570152, + "grad_norm": 0.10894393175840378, + "learning_rate": 8.934303006566837e-05, + "loss": 0.0171, + "step": 22784 + }, + { + "epoch": 0.5341162819318201, + "grad_norm": 0.21244311332702637, + "learning_rate": 8.933570588529805e-05, + "loss": 0.0358, + "step": 22785 + }, + { + "epoch": 0.5341397235066252, + "grad_norm": 0.2663443088531494, + "learning_rate": 8.932838176279248e-05, + "loss": 0.0764, + "step": 22786 + }, + { + "epoch": 0.5341631650814301, + "grad_norm": 0.08334814757108688, + "learning_rate": 8.932105769819143e-05, + "loss": 0.0182, + "step": 22787 + }, + { + "epoch": 0.5341866066562352, + "grad_norm": 0.49775266647338867, + "learning_rate": 8.931373369153457e-05, + "loss": 0.0822, + "step": 22788 + }, + { + "epoch": 0.5342100482310401, + "grad_norm": 0.5038346648216248, + "learning_rate": 8.930640974286175e-05, + "loss": 0.1518, + "step": 22789 + }, + { + "epoch": 0.5342334898058452, + "grad_norm": 0.21374452114105225, + "learning_rate": 8.929908585221263e-05, + "loss": 0.0279, + "step": 22790 + }, + { + "epoch": 0.5342569313806501, + "grad_norm": 0.2787570357322693, + "learning_rate": 8.929176201962697e-05, + "loss": 0.266, + "step": 22791 + }, + { + "epoch": 0.5342803729554552, + "grad_norm": 0.07279495149850845, + "learning_rate": 8.928443824514451e-05, + "loss": 0.0074, + "step": 22792 + }, + { + "epoch": 0.5343038145302601, + "grad_norm": 0.4591330289840698, + "learning_rate": 8.927711452880501e-05, + "loss": 0.199, + "step": 22793 + }, + { + "epoch": 0.5343272561050652, + "grad_norm": 0.30936199426651, + "learning_rate": 8.926979087064819e-05, + "loss": 0.0413, + "step": 22794 + }, + { + "epoch": 0.5343506976798701, + "grad_norm": 0.4305765926837921, + "learning_rate": 8.926246727071374e-05, + "loss": 0.1042, + "step": 22795 + }, + { + "epoch": 0.5343741392546751, + "grad_norm": 0.7034476399421692, + "learning_rate": 8.925514372904148e-05, + "loss": 0.1724, + "step": 22796 + }, + { + "epoch": 0.5343975808294801, + "grad_norm": 0.4853130280971527, + "learning_rate": 8.924782024567113e-05, + "loss": 0.0638, + "step": 22797 + }, + { + "epoch": 0.5344210224042851, + "grad_norm": 0.578088104724884, + "learning_rate": 8.924049682064237e-05, + "loss": 0.0488, + "step": 22798 + }, + { + "epoch": 0.5344444639790901, + "grad_norm": 0.3378787040710449, + "learning_rate": 8.923317345399501e-05, + "loss": 0.0507, + "step": 22799 + }, + { + "epoch": 0.5344679055538951, + "grad_norm": 0.3669504225254059, + "learning_rate": 8.92258501457687e-05, + "loss": 0.086, + "step": 22800 + }, + { + "epoch": 0.5344913471287001, + "grad_norm": 0.1913919597864151, + "learning_rate": 8.921852689600328e-05, + "loss": 0.0191, + "step": 22801 + }, + { + "epoch": 0.5345147887035051, + "grad_norm": 0.43485257029533386, + "learning_rate": 8.921120370473842e-05, + "loss": 0.0579, + "step": 22802 + }, + { + "epoch": 0.53453823027831, + "grad_norm": 0.11996745318174362, + "learning_rate": 8.920388057201387e-05, + "loss": 0.0351, + "step": 22803 + }, + { + "epoch": 0.5345616718531151, + "grad_norm": 0.3709407150745392, + "learning_rate": 8.919655749786938e-05, + "loss": 0.0563, + "step": 22804 + }, + { + "epoch": 0.53458511342792, + "grad_norm": 0.147906094789505, + "learning_rate": 8.918923448234465e-05, + "loss": 0.0218, + "step": 22805 + }, + { + "epoch": 0.5346085550027251, + "grad_norm": 0.5358210206031799, + "learning_rate": 8.918191152547947e-05, + "loss": 0.0903, + "step": 22806 + }, + { + "epoch": 0.5346319965775301, + "grad_norm": 0.35159236192703247, + "learning_rate": 8.917458862731354e-05, + "loss": 0.0719, + "step": 22807 + }, + { + "epoch": 0.5346554381523351, + "grad_norm": 0.5457218289375305, + "learning_rate": 8.916726578788657e-05, + "loss": 0.6498, + "step": 22808 + }, + { + "epoch": 0.5346788797271401, + "grad_norm": 0.26266762614250183, + "learning_rate": 8.915994300723833e-05, + "loss": 0.0391, + "step": 22809 + }, + { + "epoch": 0.5347023213019451, + "grad_norm": 0.30120038986206055, + "learning_rate": 8.915262028540856e-05, + "loss": 0.0523, + "step": 22810 + }, + { + "epoch": 0.5347257628767501, + "grad_norm": 0.4429152309894562, + "learning_rate": 8.914529762243695e-05, + "loss": 0.0858, + "step": 22811 + }, + { + "epoch": 0.534749204451555, + "grad_norm": 0.17867222428321838, + "learning_rate": 8.913797501836324e-05, + "loss": 0.0168, + "step": 22812 + }, + { + "epoch": 0.5347726460263601, + "grad_norm": 0.5856115818023682, + "learning_rate": 8.913065247322723e-05, + "loss": 0.0796, + "step": 22813 + }, + { + "epoch": 0.534796087601165, + "grad_norm": 0.43195998668670654, + "learning_rate": 8.91233299870686e-05, + "loss": 0.0775, + "step": 22814 + }, + { + "epoch": 0.5348195291759701, + "grad_norm": 0.45692065358161926, + "learning_rate": 8.911600755992705e-05, + "loss": 0.045, + "step": 22815 + }, + { + "epoch": 0.534842970750775, + "grad_norm": 0.3514106273651123, + "learning_rate": 8.910868519184241e-05, + "loss": 0.0552, + "step": 22816 + }, + { + "epoch": 0.5348664123255801, + "grad_norm": 0.6383962035179138, + "learning_rate": 8.910136288285433e-05, + "loss": 0.0333, + "step": 22817 + }, + { + "epoch": 0.534889853900385, + "grad_norm": 0.7893348336219788, + "learning_rate": 8.909404063300253e-05, + "loss": 0.106, + "step": 22818 + }, + { + "epoch": 0.5349132954751901, + "grad_norm": 0.4497811198234558, + "learning_rate": 8.908671844232681e-05, + "loss": 0.0723, + "step": 22819 + }, + { + "epoch": 0.534936737049995, + "grad_norm": 0.541216254234314, + "learning_rate": 8.907939631086685e-05, + "loss": 0.0903, + "step": 22820 + }, + { + "epoch": 0.5349601786248, + "grad_norm": 0.607061505317688, + "learning_rate": 8.907207423866237e-05, + "loss": 0.7852, + "step": 22821 + }, + { + "epoch": 0.534983620199605, + "grad_norm": 0.2721564471721649, + "learning_rate": 8.906475222575316e-05, + "loss": 0.065, + "step": 22822 + }, + { + "epoch": 0.53500706177441, + "grad_norm": 0.46260061860084534, + "learning_rate": 8.90574302721789e-05, + "loss": 0.0753, + "step": 22823 + }, + { + "epoch": 0.535030503349215, + "grad_norm": 0.3769504427909851, + "learning_rate": 8.905010837797929e-05, + "loss": 0.0735, + "step": 22824 + }, + { + "epoch": 0.53505394492402, + "grad_norm": 0.2866513133049011, + "learning_rate": 8.90427865431941e-05, + "loss": 0.0538, + "step": 22825 + }, + { + "epoch": 0.535077386498825, + "grad_norm": 0.15428754687309265, + "learning_rate": 8.903546476786312e-05, + "loss": 0.0144, + "step": 22826 + }, + { + "epoch": 0.53510082807363, + "grad_norm": 0.3612827956676483, + "learning_rate": 8.9028143052026e-05, + "loss": 0.0279, + "step": 22827 + }, + { + "epoch": 0.535124269648435, + "grad_norm": 0.5692062377929688, + "learning_rate": 8.902082139572245e-05, + "loss": 0.0588, + "step": 22828 + }, + { + "epoch": 0.53514771122324, + "grad_norm": 0.4457428455352783, + "learning_rate": 8.901349979899227e-05, + "loss": 0.1106, + "step": 22829 + }, + { + "epoch": 0.535171152798045, + "grad_norm": 0.3985097408294678, + "learning_rate": 8.900617826187512e-05, + "loss": 0.0616, + "step": 22830 + }, + { + "epoch": 0.53519459437285, + "grad_norm": 0.8381977081298828, + "learning_rate": 8.899885678441078e-05, + "loss": 0.1943, + "step": 22831 + }, + { + "epoch": 0.5352180359476549, + "grad_norm": 0.4541665017604828, + "learning_rate": 8.899153536663895e-05, + "loss": 0.084, + "step": 22832 + }, + { + "epoch": 0.53524147752246, + "grad_norm": 0.2922523319721222, + "learning_rate": 8.898421400859934e-05, + "loss": 0.0369, + "step": 22833 + }, + { + "epoch": 0.5352649190972649, + "grad_norm": 0.5376083254814148, + "learning_rate": 8.89768927103317e-05, + "loss": 0.1115, + "step": 22834 + }, + { + "epoch": 0.53528836067207, + "grad_norm": 0.11319663375616074, + "learning_rate": 8.896957147187577e-05, + "loss": 0.0236, + "step": 22835 + }, + { + "epoch": 0.5353118022468749, + "grad_norm": 0.5209466218948364, + "learning_rate": 8.896225029327121e-05, + "loss": 0.5868, + "step": 22836 + }, + { + "epoch": 0.53533524382168, + "grad_norm": 0.6609506607055664, + "learning_rate": 8.895492917455781e-05, + "loss": 0.1088, + "step": 22837 + }, + { + "epoch": 0.5353586853964849, + "grad_norm": 0.22807587683200836, + "learning_rate": 8.894760811577529e-05, + "loss": 0.0396, + "step": 22838 + }, + { + "epoch": 0.53538212697129, + "grad_norm": 0.5360878705978394, + "learning_rate": 8.894028711696335e-05, + "loss": 0.0626, + "step": 22839 + }, + { + "epoch": 0.5354055685460949, + "grad_norm": 0.385193407535553, + "learning_rate": 8.893296617816172e-05, + "loss": 0.0717, + "step": 22840 + }, + { + "epoch": 0.5354290101208999, + "grad_norm": 0.6427736878395081, + "learning_rate": 8.892564529941016e-05, + "loss": 0.0726, + "step": 22841 + }, + { + "epoch": 0.5354524516957049, + "grad_norm": 0.6299512982368469, + "learning_rate": 8.891832448074833e-05, + "loss": 0.6809, + "step": 22842 + }, + { + "epoch": 0.5354758932705099, + "grad_norm": 0.7912598252296448, + "learning_rate": 8.891100372221598e-05, + "loss": 0.1525, + "step": 22843 + }, + { + "epoch": 0.5354993348453149, + "grad_norm": 0.5766164064407349, + "learning_rate": 8.890368302385286e-05, + "loss": 0.0545, + "step": 22844 + }, + { + "epoch": 0.5355227764201199, + "grad_norm": 0.5774990320205688, + "learning_rate": 8.889636238569865e-05, + "loss": 0.1146, + "step": 22845 + }, + { + "epoch": 0.5355462179949249, + "grad_norm": 0.1899343878030777, + "learning_rate": 8.888904180779308e-05, + "loss": 0.0409, + "step": 22846 + }, + { + "epoch": 0.5355696595697299, + "grad_norm": 0.6731668710708618, + "learning_rate": 8.888172129017591e-05, + "loss": 0.0901, + "step": 22847 + }, + { + "epoch": 0.5355931011445348, + "grad_norm": 0.4934043288230896, + "learning_rate": 8.887440083288682e-05, + "loss": 0.0871, + "step": 22848 + }, + { + "epoch": 0.5356165427193399, + "grad_norm": 0.2119048833847046, + "learning_rate": 8.88670804359655e-05, + "loss": 0.0441, + "step": 22849 + }, + { + "epoch": 0.5356399842941448, + "grad_norm": 0.34879270195961, + "learning_rate": 8.885976009945173e-05, + "loss": 0.0488, + "step": 22850 + }, + { + "epoch": 0.5356634258689499, + "grad_norm": 0.715220034122467, + "learning_rate": 8.885243982338525e-05, + "loss": 0.183, + "step": 22851 + }, + { + "epoch": 0.5356868674437548, + "grad_norm": 0.1444714516401291, + "learning_rate": 8.884511960780575e-05, + "loss": 0.039, + "step": 22852 + }, + { + "epoch": 0.5357103090185599, + "grad_norm": 0.51035076379776, + "learning_rate": 8.883779945275292e-05, + "loss": 0.0878, + "step": 22853 + }, + { + "epoch": 0.5357337505933648, + "grad_norm": 0.20126454532146454, + "learning_rate": 8.883047935826652e-05, + "loss": 0.0462, + "step": 22854 + }, + { + "epoch": 0.5357571921681699, + "grad_norm": 0.5695136189460754, + "learning_rate": 8.882315932438626e-05, + "loss": 0.1034, + "step": 22855 + }, + { + "epoch": 0.5357806337429748, + "grad_norm": 0.45197784900665283, + "learning_rate": 8.881583935115182e-05, + "loss": 0.0976, + "step": 22856 + }, + { + "epoch": 0.5358040753177798, + "grad_norm": 0.40230655670166016, + "learning_rate": 8.880851943860298e-05, + "loss": 0.0832, + "step": 22857 + }, + { + "epoch": 0.5358275168925849, + "grad_norm": 0.4774268567562103, + "learning_rate": 8.880119958677941e-05, + "loss": 0.0856, + "step": 22858 + }, + { + "epoch": 0.5358509584673898, + "grad_norm": 0.5990945100784302, + "learning_rate": 8.879387979572084e-05, + "loss": 0.0992, + "step": 22859 + }, + { + "epoch": 0.5358744000421949, + "grad_norm": 0.2811915874481201, + "learning_rate": 8.878656006546701e-05, + "loss": 0.0334, + "step": 22860 + }, + { + "epoch": 0.5358978416169998, + "grad_norm": 0.43834179639816284, + "learning_rate": 8.877924039605758e-05, + "loss": 0.0437, + "step": 22861 + }, + { + "epoch": 0.5359212831918049, + "grad_norm": 0.05123518407344818, + "learning_rate": 8.877192078753236e-05, + "loss": 0.0054, + "step": 22862 + }, + { + "epoch": 0.5359447247666098, + "grad_norm": 0.21792612969875336, + "learning_rate": 8.876460123993098e-05, + "loss": 0.0587, + "step": 22863 + }, + { + "epoch": 0.5359681663414149, + "grad_norm": 0.46648937463760376, + "learning_rate": 8.875728175329323e-05, + "loss": 0.1193, + "step": 22864 + }, + { + "epoch": 0.5359916079162198, + "grad_norm": 0.32792553305625916, + "learning_rate": 8.874996232765875e-05, + "loss": 0.0393, + "step": 22865 + }, + { + "epoch": 0.5360150494910249, + "grad_norm": 0.7161449193954468, + "learning_rate": 8.87426429630673e-05, + "loss": 0.1442, + "step": 22866 + }, + { + "epoch": 0.5360384910658298, + "grad_norm": 0.8118685483932495, + "learning_rate": 8.873532365955858e-05, + "loss": 0.1828, + "step": 22867 + }, + { + "epoch": 0.5360619326406348, + "grad_norm": 0.25788745284080505, + "learning_rate": 8.872800441717231e-05, + "loss": 0.0302, + "step": 22868 + }, + { + "epoch": 0.5360853742154398, + "grad_norm": 0.43352094292640686, + "learning_rate": 8.872068523594823e-05, + "loss": 0.0463, + "step": 22869 + }, + { + "epoch": 0.5361088157902448, + "grad_norm": 0.131772980093956, + "learning_rate": 8.871336611592602e-05, + "loss": 0.0348, + "step": 22870 + }, + { + "epoch": 0.5361322573650498, + "grad_norm": 0.45933568477630615, + "learning_rate": 8.870604705714539e-05, + "loss": 0.0619, + "step": 22871 + }, + { + "epoch": 0.5361556989398548, + "grad_norm": 0.22288942337036133, + "learning_rate": 8.869872805964607e-05, + "loss": 0.0704, + "step": 22872 + }, + { + "epoch": 0.5361791405146598, + "grad_norm": 0.6011106371879578, + "learning_rate": 8.869140912346774e-05, + "loss": 0.1307, + "step": 22873 + }, + { + "epoch": 0.5362025820894648, + "grad_norm": 0.1538594663143158, + "learning_rate": 8.868409024865018e-05, + "loss": 0.0372, + "step": 22874 + }, + { + "epoch": 0.5362260236642697, + "grad_norm": 0.20001032948493958, + "learning_rate": 8.867677143523304e-05, + "loss": 0.0175, + "step": 22875 + }, + { + "epoch": 0.5362494652390748, + "grad_norm": 0.4013829529285431, + "learning_rate": 8.866945268325609e-05, + "loss": 0.0862, + "step": 22876 + }, + { + "epoch": 0.5362729068138797, + "grad_norm": 0.4839843809604645, + "learning_rate": 8.866213399275899e-05, + "loss": 0.068, + "step": 22877 + }, + { + "epoch": 0.5362963483886848, + "grad_norm": 0.3115387558937073, + "learning_rate": 8.865481536378147e-05, + "loss": 0.0648, + "step": 22878 + }, + { + "epoch": 0.5363197899634897, + "grad_norm": 0.2883963882923126, + "learning_rate": 8.864749679636324e-05, + "loss": 0.0338, + "step": 22879 + }, + { + "epoch": 0.5363432315382948, + "grad_norm": 0.2682667374610901, + "learning_rate": 8.864017829054402e-05, + "loss": 0.0272, + "step": 22880 + }, + { + "epoch": 0.5363666731130997, + "grad_norm": 0.1452396959066391, + "learning_rate": 8.863285984636349e-05, + "loss": 0.0231, + "step": 22881 + }, + { + "epoch": 0.5363901146879048, + "grad_norm": 0.4594372808933258, + "learning_rate": 8.862554146386142e-05, + "loss": 0.1356, + "step": 22882 + }, + { + "epoch": 0.5364135562627097, + "grad_norm": 0.12060166150331497, + "learning_rate": 8.861822314307746e-05, + "loss": 0.014, + "step": 22883 + }, + { + "epoch": 0.5364369978375148, + "grad_norm": 0.5747095346450806, + "learning_rate": 8.861090488405132e-05, + "loss": 0.1172, + "step": 22884 + }, + { + "epoch": 0.5364604394123197, + "grad_norm": 0.22085367143154144, + "learning_rate": 8.860358668682273e-05, + "loss": 0.0211, + "step": 22885 + }, + { + "epoch": 0.5364838809871247, + "grad_norm": 0.40912261605262756, + "learning_rate": 8.859626855143142e-05, + "loss": 0.0454, + "step": 22886 + }, + { + "epoch": 0.5365073225619297, + "grad_norm": 0.6416525840759277, + "learning_rate": 8.858895047791709e-05, + "loss": 0.0996, + "step": 22887 + }, + { + "epoch": 0.5365307641367347, + "grad_norm": 0.19615769386291504, + "learning_rate": 8.858163246631941e-05, + "loss": 0.032, + "step": 22888 + }, + { + "epoch": 0.5365542057115397, + "grad_norm": 0.7579649686813354, + "learning_rate": 8.857431451667813e-05, + "loss": 0.1322, + "step": 22889 + }, + { + "epoch": 0.5365776472863447, + "grad_norm": 0.6556891798973083, + "learning_rate": 8.856699662903296e-05, + "loss": 0.7593, + "step": 22890 + }, + { + "epoch": 0.5366010888611497, + "grad_norm": 1.2967966794967651, + "learning_rate": 8.855967880342355e-05, + "loss": 0.0697, + "step": 22891 + }, + { + "epoch": 0.5366245304359547, + "grad_norm": 0.15124118328094482, + "learning_rate": 8.855236103988966e-05, + "loss": 0.0102, + "step": 22892 + }, + { + "epoch": 0.5366479720107596, + "grad_norm": 0.22082574665546417, + "learning_rate": 8.8545043338471e-05, + "loss": 0.0321, + "step": 22893 + }, + { + "epoch": 0.5366714135855647, + "grad_norm": 0.5907754898071289, + "learning_rate": 8.853772569920723e-05, + "loss": 0.1698, + "step": 22894 + }, + { + "epoch": 0.5366948551603696, + "grad_norm": 0.9290260672569275, + "learning_rate": 8.853040812213809e-05, + "loss": 0.0943, + "step": 22895 + }, + { + "epoch": 0.5367182967351747, + "grad_norm": 0.5094513297080994, + "learning_rate": 8.852309060730328e-05, + "loss": 0.1011, + "step": 22896 + }, + { + "epoch": 0.5367417383099796, + "grad_norm": 0.4668499827384949, + "learning_rate": 8.851577315474247e-05, + "loss": 0.0283, + "step": 22897 + }, + { + "epoch": 0.5367651798847847, + "grad_norm": 0.14223088324069977, + "learning_rate": 8.850845576449543e-05, + "loss": 0.0351, + "step": 22898 + }, + { + "epoch": 0.5367886214595896, + "grad_norm": 0.10176017880439758, + "learning_rate": 8.850113843660182e-05, + "loss": 0.0099, + "step": 22899 + }, + { + "epoch": 0.5368120630343947, + "grad_norm": 0.4177655577659607, + "learning_rate": 8.849382117110138e-05, + "loss": 0.0317, + "step": 22900 + }, + { + "epoch": 0.5368355046091996, + "grad_norm": 0.6333528757095337, + "learning_rate": 8.848650396803377e-05, + "loss": 0.0745, + "step": 22901 + }, + { + "epoch": 0.5368589461840046, + "grad_norm": 0.27883294224739075, + "learning_rate": 8.847918682743874e-05, + "loss": 0.0588, + "step": 22902 + }, + { + "epoch": 0.5368823877588096, + "grad_norm": 0.595903217792511, + "learning_rate": 8.847186974935594e-05, + "loss": 0.1717, + "step": 22903 + }, + { + "epoch": 0.5369058293336146, + "grad_norm": 0.3791347146034241, + "learning_rate": 8.84645527338251e-05, + "loss": 0.0536, + "step": 22904 + }, + { + "epoch": 0.5369292709084196, + "grad_norm": 0.10487882792949677, + "learning_rate": 8.845723578088593e-05, + "loss": 0.0157, + "step": 22905 + }, + { + "epoch": 0.5369527124832246, + "grad_norm": 0.5663549900054932, + "learning_rate": 8.844991889057811e-05, + "loss": 0.0565, + "step": 22906 + }, + { + "epoch": 0.5369761540580296, + "grad_norm": 0.4312683343887329, + "learning_rate": 8.844260206294138e-05, + "loss": 0.0992, + "step": 22907 + }, + { + "epoch": 0.5369995956328346, + "grad_norm": 0.5191663503646851, + "learning_rate": 8.84352852980154e-05, + "loss": 0.0975, + "step": 22908 + }, + { + "epoch": 0.5370230372076397, + "grad_norm": 0.8307141661643982, + "learning_rate": 8.842796859583984e-05, + "loss": 0.0721, + "step": 22909 + }, + { + "epoch": 0.5370464787824446, + "grad_norm": 0.5201578140258789, + "learning_rate": 8.842065195645452e-05, + "loss": 0.1005, + "step": 22910 + }, + { + "epoch": 0.5370699203572497, + "grad_norm": 0.29306960105895996, + "learning_rate": 8.841333537989903e-05, + "loss": 0.043, + "step": 22911 + }, + { + "epoch": 0.5370933619320546, + "grad_norm": 0.5780682563781738, + "learning_rate": 8.840601886621314e-05, + "loss": 0.1181, + "step": 22912 + }, + { + "epoch": 0.5371168035068596, + "grad_norm": 0.21890583634376526, + "learning_rate": 8.839870241543647e-05, + "loss": 0.0513, + "step": 22913 + }, + { + "epoch": 0.5371402450816646, + "grad_norm": 0.33878159523010254, + "learning_rate": 8.839138602760882e-05, + "loss": 0.0863, + "step": 22914 + }, + { + "epoch": 0.5371636866564696, + "grad_norm": 0.45728734135627747, + "learning_rate": 8.838406970276981e-05, + "loss": 0.0598, + "step": 22915 + }, + { + "epoch": 0.5371871282312746, + "grad_norm": 0.5903640389442444, + "learning_rate": 8.837675344095915e-05, + "loss": 0.1056, + "step": 22916 + }, + { + "epoch": 0.5372105698060796, + "grad_norm": 0.3659321367740631, + "learning_rate": 8.836943724221658e-05, + "loss": 0.0496, + "step": 22917 + }, + { + "epoch": 0.5372340113808846, + "grad_norm": 0.33694663643836975, + "learning_rate": 8.836212110658177e-05, + "loss": 0.4008, + "step": 22918 + }, + { + "epoch": 0.5372574529556896, + "grad_norm": 0.3155875504016876, + "learning_rate": 8.83548050340944e-05, + "loss": 0.047, + "step": 22919 + }, + { + "epoch": 0.5372808945304945, + "grad_norm": 1.225754976272583, + "learning_rate": 8.834748902479421e-05, + "loss": 0.1568, + "step": 22920 + }, + { + "epoch": 0.5373043361052996, + "grad_norm": 0.35809969902038574, + "learning_rate": 8.834017307872085e-05, + "loss": 0.0716, + "step": 22921 + }, + { + "epoch": 0.5373277776801045, + "grad_norm": 0.28644683957099915, + "learning_rate": 8.833285719591402e-05, + "loss": 0.0381, + "step": 22922 + }, + { + "epoch": 0.5373512192549096, + "grad_norm": 0.39377352595329285, + "learning_rate": 8.832554137641344e-05, + "loss": 0.0513, + "step": 22923 + }, + { + "epoch": 0.5373746608297145, + "grad_norm": 0.2583913505077362, + "learning_rate": 8.831822562025883e-05, + "loss": 0.0416, + "step": 22924 + }, + { + "epoch": 0.5373981024045196, + "grad_norm": 0.25922492146492004, + "learning_rate": 8.831090992748986e-05, + "loss": 0.0253, + "step": 22925 + }, + { + "epoch": 0.5374215439793245, + "grad_norm": 0.3308858871459961, + "learning_rate": 8.830359429814619e-05, + "loss": 0.2768, + "step": 22926 + }, + { + "epoch": 0.5374449855541296, + "grad_norm": 0.5297684669494629, + "learning_rate": 8.829627873226757e-05, + "loss": 0.0778, + "step": 22927 + }, + { + "epoch": 0.5374684271289345, + "grad_norm": 0.6234508156776428, + "learning_rate": 8.828896322989368e-05, + "loss": 0.1555, + "step": 22928 + }, + { + "epoch": 0.5374918687037396, + "grad_norm": 0.22395439445972443, + "learning_rate": 8.828164779106418e-05, + "loss": 0.0391, + "step": 22929 + }, + { + "epoch": 0.5375153102785445, + "grad_norm": 0.6691876649856567, + "learning_rate": 8.827433241581879e-05, + "loss": 0.1217, + "step": 22930 + }, + { + "epoch": 0.5375387518533495, + "grad_norm": 0.4559021592140198, + "learning_rate": 8.826701710419722e-05, + "loss": 0.0292, + "step": 22931 + }, + { + "epoch": 0.5375621934281545, + "grad_norm": 0.32836002111434937, + "learning_rate": 8.825970185623913e-05, + "loss": 0.0454, + "step": 22932 + }, + { + "epoch": 0.5375856350029595, + "grad_norm": 0.5428371429443359, + "learning_rate": 8.825238667198425e-05, + "loss": 0.1397, + "step": 22933 + }, + { + "epoch": 0.5376090765777645, + "grad_norm": 0.30922725796699524, + "learning_rate": 8.824507155147219e-05, + "loss": 0.0329, + "step": 22934 + }, + { + "epoch": 0.5376325181525695, + "grad_norm": 0.47204238176345825, + "learning_rate": 8.823775649474275e-05, + "loss": 0.082, + "step": 22935 + }, + { + "epoch": 0.5376559597273745, + "grad_norm": 0.18626362085342407, + "learning_rate": 8.823044150183555e-05, + "loss": 0.0382, + "step": 22936 + }, + { + "epoch": 0.5376794013021795, + "grad_norm": 0.4525390565395355, + "learning_rate": 8.822312657279033e-05, + "loss": 0.1068, + "step": 22937 + }, + { + "epoch": 0.5377028428769844, + "grad_norm": 0.6324640512466431, + "learning_rate": 8.821581170764677e-05, + "loss": 0.6985, + "step": 22938 + }, + { + "epoch": 0.5377262844517895, + "grad_norm": 0.6340934634208679, + "learning_rate": 8.820849690644452e-05, + "loss": 0.3193, + "step": 22939 + }, + { + "epoch": 0.5377497260265944, + "grad_norm": 0.26443126797676086, + "learning_rate": 8.82011821692233e-05, + "loss": 0.0553, + "step": 22940 + }, + { + "epoch": 0.5377731676013995, + "grad_norm": 0.217394158244133, + "learning_rate": 8.819386749602283e-05, + "loss": 0.034, + "step": 22941 + }, + { + "epoch": 0.5377966091762044, + "grad_norm": 0.2632301151752472, + "learning_rate": 8.81865528868827e-05, + "loss": 0.0359, + "step": 22942 + }, + { + "epoch": 0.5378200507510095, + "grad_norm": 0.2209286242723465, + "learning_rate": 8.817923834184273e-05, + "loss": 0.0474, + "step": 22943 + }, + { + "epoch": 0.5378434923258144, + "grad_norm": 0.5872166156768799, + "learning_rate": 8.817192386094251e-05, + "loss": 0.096, + "step": 22944 + }, + { + "epoch": 0.5378669339006195, + "grad_norm": 1.2491310834884644, + "learning_rate": 8.816460944422178e-05, + "loss": 0.1272, + "step": 22945 + }, + { + "epoch": 0.5378903754754244, + "grad_norm": 0.1650291234254837, + "learning_rate": 8.815729509172017e-05, + "loss": 0.0203, + "step": 22946 + }, + { + "epoch": 0.5379138170502294, + "grad_norm": 0.35150256752967834, + "learning_rate": 8.814998080347746e-05, + "loss": 0.0719, + "step": 22947 + }, + { + "epoch": 0.5379372586250344, + "grad_norm": 0.4840770959854126, + "learning_rate": 8.814266657953327e-05, + "loss": 0.1147, + "step": 22948 + }, + { + "epoch": 0.5379607001998394, + "grad_norm": 0.258266806602478, + "learning_rate": 8.81353524199273e-05, + "loss": 0.0253, + "step": 22949 + }, + { + "epoch": 0.5379841417746444, + "grad_norm": 0.11344209313392639, + "learning_rate": 8.812803832469927e-05, + "loss": 0.0173, + "step": 22950 + }, + { + "epoch": 0.5380075833494494, + "grad_norm": 0.11413814127445221, + "learning_rate": 8.81207242938888e-05, + "loss": 0.0196, + "step": 22951 + }, + { + "epoch": 0.5380310249242544, + "grad_norm": 0.15176376700401306, + "learning_rate": 8.811341032753564e-05, + "loss": 0.0216, + "step": 22952 + }, + { + "epoch": 0.5380544664990594, + "grad_norm": 0.5619122385978699, + "learning_rate": 8.810609642567945e-05, + "loss": 0.117, + "step": 22953 + }, + { + "epoch": 0.5380779080738644, + "grad_norm": 0.04552000015974045, + "learning_rate": 8.80987825883599e-05, + "loss": 0.0027, + "step": 22954 + }, + { + "epoch": 0.5381013496486694, + "grad_norm": 1.0336947441101074, + "learning_rate": 8.809146881561669e-05, + "loss": 0.2113, + "step": 22955 + }, + { + "epoch": 0.5381247912234743, + "grad_norm": 0.4692055583000183, + "learning_rate": 8.808415510748953e-05, + "loss": 0.1045, + "step": 22956 + }, + { + "epoch": 0.5381482327982794, + "grad_norm": 0.6021854281425476, + "learning_rate": 8.807684146401804e-05, + "loss": 0.1448, + "step": 22957 + }, + { + "epoch": 0.5381716743730843, + "grad_norm": 0.6276039481163025, + "learning_rate": 8.806952788524195e-05, + "loss": 0.1336, + "step": 22958 + }, + { + "epoch": 0.5381951159478894, + "grad_norm": 0.5169024467468262, + "learning_rate": 8.806221437120095e-05, + "loss": 0.1199, + "step": 22959 + }, + { + "epoch": 0.5382185575226944, + "grad_norm": 0.13842187821865082, + "learning_rate": 8.805490092193472e-05, + "loss": 0.021, + "step": 22960 + }, + { + "epoch": 0.5382419990974994, + "grad_norm": 0.31051596999168396, + "learning_rate": 8.804758753748292e-05, + "loss": 0.0185, + "step": 22961 + }, + { + "epoch": 0.5382654406723044, + "grad_norm": 0.39718204736709595, + "learning_rate": 8.804027421788526e-05, + "loss": 0.0505, + "step": 22962 + }, + { + "epoch": 0.5382888822471094, + "grad_norm": 0.2762250602245331, + "learning_rate": 8.80329609631814e-05, + "loss": 0.0542, + "step": 22963 + }, + { + "epoch": 0.5383123238219144, + "grad_norm": 0.29851239919662476, + "learning_rate": 8.802564777341103e-05, + "loss": 0.0392, + "step": 22964 + }, + { + "epoch": 0.5383357653967193, + "grad_norm": 0.18027673661708832, + "learning_rate": 8.801833464861383e-05, + "loss": 0.0293, + "step": 22965 + }, + { + "epoch": 0.5383592069715244, + "grad_norm": 0.1135687381029129, + "learning_rate": 8.80110215888295e-05, + "loss": 0.0134, + "step": 22966 + }, + { + "epoch": 0.5383826485463293, + "grad_norm": 0.26358115673065186, + "learning_rate": 8.800370859409767e-05, + "loss": 0.068, + "step": 22967 + }, + { + "epoch": 0.5384060901211344, + "grad_norm": 0.4447176456451416, + "learning_rate": 8.799639566445808e-05, + "loss": 0.0829, + "step": 22968 + }, + { + "epoch": 0.5384295316959393, + "grad_norm": 0.14768053591251373, + "learning_rate": 8.79890827999504e-05, + "loss": 0.0191, + "step": 22969 + }, + { + "epoch": 0.5384529732707444, + "grad_norm": 0.0904545933008194, + "learning_rate": 8.798177000061422e-05, + "loss": 0.016, + "step": 22970 + }, + { + "epoch": 0.5384764148455493, + "grad_norm": 0.2611059546470642, + "learning_rate": 8.797445726648934e-05, + "loss": 0.0805, + "step": 22971 + }, + { + "epoch": 0.5384998564203544, + "grad_norm": 0.3943690359592438, + "learning_rate": 8.796714459761541e-05, + "loss": 0.0646, + "step": 22972 + }, + { + "epoch": 0.5385232979951593, + "grad_norm": 0.32886844873428345, + "learning_rate": 8.795983199403209e-05, + "loss": 0.0699, + "step": 22973 + }, + { + "epoch": 0.5385467395699644, + "grad_norm": 0.41260096430778503, + "learning_rate": 8.795251945577904e-05, + "loss": 0.0616, + "step": 22974 + }, + { + "epoch": 0.5385701811447693, + "grad_norm": 0.5101522207260132, + "learning_rate": 8.794520698289598e-05, + "loss": 0.1042, + "step": 22975 + }, + { + "epoch": 0.5385936227195743, + "grad_norm": 0.5458811521530151, + "learning_rate": 8.793789457542256e-05, + "loss": 0.5884, + "step": 22976 + }, + { + "epoch": 0.5386170642943793, + "grad_norm": 0.12359121441841125, + "learning_rate": 8.793058223339845e-05, + "loss": 0.029, + "step": 22977 + }, + { + "epoch": 0.5386405058691843, + "grad_norm": 0.5957169532775879, + "learning_rate": 8.792326995686335e-05, + "loss": 0.0869, + "step": 22978 + }, + { + "epoch": 0.5386639474439893, + "grad_norm": 0.9329640865325928, + "learning_rate": 8.791595774585695e-05, + "loss": 0.1135, + "step": 22979 + }, + { + "epoch": 0.5386873890187943, + "grad_norm": 0.7358602285385132, + "learning_rate": 8.790864560041885e-05, + "loss": 0.0726, + "step": 22980 + }, + { + "epoch": 0.5387108305935993, + "grad_norm": 0.5539270639419556, + "learning_rate": 8.790133352058883e-05, + "loss": 0.1337, + "step": 22981 + }, + { + "epoch": 0.5387342721684043, + "grad_norm": 0.2312021553516388, + "learning_rate": 8.789402150640646e-05, + "loss": 0.0208, + "step": 22982 + }, + { + "epoch": 0.5387577137432092, + "grad_norm": 0.5468871593475342, + "learning_rate": 8.788670955791151e-05, + "loss": 0.1188, + "step": 22983 + }, + { + "epoch": 0.5387811553180143, + "grad_norm": 0.1792008876800537, + "learning_rate": 8.787939767514359e-05, + "loss": 0.0377, + "step": 22984 + }, + { + "epoch": 0.5388045968928192, + "grad_norm": 0.5736010074615479, + "learning_rate": 8.787208585814242e-05, + "loss": 0.1106, + "step": 22985 + }, + { + "epoch": 0.5388280384676243, + "grad_norm": 0.20657850801944733, + "learning_rate": 8.786477410694764e-05, + "loss": 0.0634, + "step": 22986 + }, + { + "epoch": 0.5388514800424292, + "grad_norm": 0.23812638223171234, + "learning_rate": 8.785746242159896e-05, + "loss": 0.0296, + "step": 22987 + }, + { + "epoch": 0.5388749216172343, + "grad_norm": 0.2670639455318451, + "learning_rate": 8.785015080213602e-05, + "loss": 0.07, + "step": 22988 + }, + { + "epoch": 0.5388983631920392, + "grad_norm": 0.6039549112319946, + "learning_rate": 8.784283924859849e-05, + "loss": 0.0971, + "step": 22989 + }, + { + "epoch": 0.5389218047668443, + "grad_norm": 0.6011580228805542, + "learning_rate": 8.783552776102607e-05, + "loss": 0.6348, + "step": 22990 + }, + { + "epoch": 0.5389452463416492, + "grad_norm": 0.24819821119308472, + "learning_rate": 8.782821633945842e-05, + "loss": 0.023, + "step": 22991 + }, + { + "epoch": 0.5389686879164542, + "grad_norm": 0.35627198219299316, + "learning_rate": 8.78209049839352e-05, + "loss": 0.0275, + "step": 22992 + }, + { + "epoch": 0.5389921294912592, + "grad_norm": 0.38994458317756653, + "learning_rate": 8.781359369449609e-05, + "loss": 0.0501, + "step": 22993 + }, + { + "epoch": 0.5390155710660642, + "grad_norm": 0.5726454854011536, + "learning_rate": 8.780628247118078e-05, + "loss": 0.1152, + "step": 22994 + }, + { + "epoch": 0.5390390126408692, + "grad_norm": 0.4887075126171112, + "learning_rate": 8.779897131402888e-05, + "loss": 0.081, + "step": 22995 + }, + { + "epoch": 0.5390624542156742, + "grad_norm": 0.41975587606430054, + "learning_rate": 8.779166022308013e-05, + "loss": 0.0809, + "step": 22996 + }, + { + "epoch": 0.5390858957904792, + "grad_norm": 0.5852124094963074, + "learning_rate": 8.77843491983742e-05, + "loss": 0.115, + "step": 22997 + }, + { + "epoch": 0.5391093373652842, + "grad_norm": 0.4090828597545624, + "learning_rate": 8.777703823995071e-05, + "loss": 0.0758, + "step": 22998 + }, + { + "epoch": 0.5391327789400892, + "grad_norm": 1.3493127822875977, + "learning_rate": 8.776972734784937e-05, + "loss": 0.1543, + "step": 22999 + }, + { + "epoch": 0.5391562205148942, + "grad_norm": 0.854579746723175, + "learning_rate": 8.776241652210983e-05, + "loss": 0.5213, + "step": 23000 + }, + { + "epoch": 0.5391796620896991, + "grad_norm": 0.4852105677127838, + "learning_rate": 8.775510576277177e-05, + "loss": 0.0877, + "step": 23001 + }, + { + "epoch": 0.5392031036645042, + "grad_norm": 0.14943194389343262, + "learning_rate": 8.774779506987482e-05, + "loss": 0.0287, + "step": 23002 + }, + { + "epoch": 0.5392265452393091, + "grad_norm": 0.5995780229568481, + "learning_rate": 8.77404844434587e-05, + "loss": 0.0581, + "step": 23003 + }, + { + "epoch": 0.5392499868141142, + "grad_norm": 0.58470618724823, + "learning_rate": 8.773317388356306e-05, + "loss": 0.1457, + "step": 23004 + }, + { + "epoch": 0.5392734283889191, + "grad_norm": 0.5097311735153198, + "learning_rate": 8.772586339022755e-05, + "loss": 0.0892, + "step": 23005 + }, + { + "epoch": 0.5392968699637242, + "grad_norm": 0.1582256406545639, + "learning_rate": 8.771855296349187e-05, + "loss": 0.0252, + "step": 23006 + }, + { + "epoch": 0.5393203115385291, + "grad_norm": 0.6150368452072144, + "learning_rate": 8.77112426033956e-05, + "loss": 0.1115, + "step": 23007 + }, + { + "epoch": 0.5393437531133342, + "grad_norm": 0.4634038805961609, + "learning_rate": 8.770393230997854e-05, + "loss": 0.0721, + "step": 23008 + }, + { + "epoch": 0.5393671946881391, + "grad_norm": 0.4753890633583069, + "learning_rate": 8.769662208328027e-05, + "loss": 0.0299, + "step": 23009 + }, + { + "epoch": 0.5393906362629441, + "grad_norm": 0.7823320031166077, + "learning_rate": 8.768931192334049e-05, + "loss": 0.6721, + "step": 23010 + }, + { + "epoch": 0.5394140778377492, + "grad_norm": 0.3101484775543213, + "learning_rate": 8.768200183019884e-05, + "loss": 0.0507, + "step": 23011 + }, + { + "epoch": 0.5394375194125541, + "grad_norm": 0.49003127217292786, + "learning_rate": 8.767469180389499e-05, + "loss": 0.0877, + "step": 23012 + }, + { + "epoch": 0.5394609609873592, + "grad_norm": 0.5082215070724487, + "learning_rate": 8.766738184446861e-05, + "loss": 0.0794, + "step": 23013 + }, + { + "epoch": 0.5394844025621641, + "grad_norm": 0.43087366223335266, + "learning_rate": 8.766007195195938e-05, + "loss": 0.1155, + "step": 23014 + }, + { + "epoch": 0.5395078441369692, + "grad_norm": 0.07151354104280472, + "learning_rate": 8.765276212640691e-05, + "loss": 0.0178, + "step": 23015 + }, + { + "epoch": 0.5395312857117741, + "grad_norm": 0.5905258655548096, + "learning_rate": 8.764545236785095e-05, + "loss": 0.1388, + "step": 23016 + }, + { + "epoch": 0.5395547272865792, + "grad_norm": 0.8818987607955933, + "learning_rate": 8.763814267633105e-05, + "loss": 0.1639, + "step": 23017 + }, + { + "epoch": 0.5395781688613841, + "grad_norm": 0.37316277623176575, + "learning_rate": 8.7630833051887e-05, + "loss": 0.0748, + "step": 23018 + }, + { + "epoch": 0.5396016104361891, + "grad_norm": 0.2819267213344574, + "learning_rate": 8.762352349455833e-05, + "loss": 0.0826, + "step": 23019 + }, + { + "epoch": 0.5396250520109941, + "grad_norm": 0.29953014850616455, + "learning_rate": 8.76162140043848e-05, + "loss": 0.0409, + "step": 23020 + }, + { + "epoch": 0.5396484935857991, + "grad_norm": 0.2757132649421692, + "learning_rate": 8.760890458140607e-05, + "loss": 0.0219, + "step": 23021 + }, + { + "epoch": 0.5396719351606041, + "grad_norm": 0.3254759907722473, + "learning_rate": 8.760159522566173e-05, + "loss": 0.0456, + "step": 23022 + }, + { + "epoch": 0.5396953767354091, + "grad_norm": 0.5426414012908936, + "learning_rate": 8.759428593719153e-05, + "loss": 0.0946, + "step": 23023 + }, + { + "epoch": 0.5397188183102141, + "grad_norm": 0.4678879380226135, + "learning_rate": 8.758697671603503e-05, + "loss": 0.0691, + "step": 23024 + }, + { + "epoch": 0.5397422598850191, + "grad_norm": 0.4522473216056824, + "learning_rate": 8.757966756223198e-05, + "loss": 0.0852, + "step": 23025 + }, + { + "epoch": 0.539765701459824, + "grad_norm": 0.5093165636062622, + "learning_rate": 8.7572358475822e-05, + "loss": 0.391, + "step": 23026 + }, + { + "epoch": 0.5397891430346291, + "grad_norm": 0.2518584132194519, + "learning_rate": 8.756504945684474e-05, + "loss": 0.0499, + "step": 23027 + }, + { + "epoch": 0.539812584609434, + "grad_norm": 0.3234892189502716, + "learning_rate": 8.755774050533988e-05, + "loss": 0.0627, + "step": 23028 + }, + { + "epoch": 0.5398360261842391, + "grad_norm": 0.6201798319816589, + "learning_rate": 8.755043162134709e-05, + "loss": 0.0989, + "step": 23029 + }, + { + "epoch": 0.539859467759044, + "grad_norm": 0.8249682188034058, + "learning_rate": 8.754312280490596e-05, + "loss": 0.1948, + "step": 23030 + }, + { + "epoch": 0.5398829093338491, + "grad_norm": 0.6536405682563782, + "learning_rate": 8.753581405605619e-05, + "loss": 0.1473, + "step": 23031 + }, + { + "epoch": 0.539906350908654, + "grad_norm": 0.44332173466682434, + "learning_rate": 8.752850537483747e-05, + "loss": 0.0859, + "step": 23032 + }, + { + "epoch": 0.5399297924834591, + "grad_norm": 0.39512714743614197, + "learning_rate": 8.752119676128945e-05, + "loss": 0.0928, + "step": 23033 + }, + { + "epoch": 0.539953234058264, + "grad_norm": 0.7501262426376343, + "learning_rate": 8.751388821545175e-05, + "loss": 0.5429, + "step": 23034 + }, + { + "epoch": 0.5399766756330691, + "grad_norm": 0.31717485189437866, + "learning_rate": 8.750657973736404e-05, + "loss": 0.0436, + "step": 23035 + }, + { + "epoch": 0.540000117207874, + "grad_norm": 0.5515079498291016, + "learning_rate": 8.7499271327066e-05, + "loss": 0.1018, + "step": 23036 + }, + { + "epoch": 0.540023558782679, + "grad_norm": 0.271457701921463, + "learning_rate": 8.749196298459725e-05, + "loss": 0.0564, + "step": 23037 + }, + { + "epoch": 0.540047000357484, + "grad_norm": 0.23750245571136475, + "learning_rate": 8.748465470999746e-05, + "loss": 0.0526, + "step": 23038 + }, + { + "epoch": 0.540070441932289, + "grad_norm": 0.12315760552883148, + "learning_rate": 8.747734650330629e-05, + "loss": 0.0167, + "step": 23039 + }, + { + "epoch": 0.540093883507094, + "grad_norm": 0.42242878675460815, + "learning_rate": 8.747003836456337e-05, + "loss": 0.1013, + "step": 23040 + }, + { + "epoch": 0.540117325081899, + "grad_norm": 0.16828393936157227, + "learning_rate": 8.74627302938084e-05, + "loss": 0.0392, + "step": 23041 + }, + { + "epoch": 0.540140766656704, + "grad_norm": 0.07599515467882156, + "learning_rate": 8.745542229108102e-05, + "loss": 0.0101, + "step": 23042 + }, + { + "epoch": 0.540164208231509, + "grad_norm": 0.3590262532234192, + "learning_rate": 8.744811435642081e-05, + "loss": 0.0718, + "step": 23043 + }, + { + "epoch": 0.540187649806314, + "grad_norm": 0.21170274913311005, + "learning_rate": 8.744080648986751e-05, + "loss": 0.0315, + "step": 23044 + }, + { + "epoch": 0.540211091381119, + "grad_norm": 0.566978931427002, + "learning_rate": 8.743349869146077e-05, + "loss": 0.0868, + "step": 23045 + }, + { + "epoch": 0.5402345329559239, + "grad_norm": 0.40981706976890564, + "learning_rate": 8.742619096124021e-05, + "loss": 0.0547, + "step": 23046 + }, + { + "epoch": 0.540257974530729, + "grad_norm": 0.8137564659118652, + "learning_rate": 8.74188832992455e-05, + "loss": 0.1391, + "step": 23047 + }, + { + "epoch": 0.5402814161055339, + "grad_norm": 0.507659375667572, + "learning_rate": 8.741157570551628e-05, + "loss": 0.0681, + "step": 23048 + }, + { + "epoch": 0.540304857680339, + "grad_norm": 0.30724167823791504, + "learning_rate": 8.740426818009219e-05, + "loss": 0.0293, + "step": 23049 + }, + { + "epoch": 0.5403282992551439, + "grad_norm": 0.38449952006340027, + "learning_rate": 8.73969607230129e-05, + "loss": 0.0593, + "step": 23050 + }, + { + "epoch": 0.540351740829949, + "grad_norm": 0.14394009113311768, + "learning_rate": 8.738965333431808e-05, + "loss": 0.0133, + "step": 23051 + }, + { + "epoch": 0.5403751824047539, + "grad_norm": 0.5935835242271423, + "learning_rate": 8.738234601404733e-05, + "loss": 0.612, + "step": 23052 + }, + { + "epoch": 0.540398623979559, + "grad_norm": 0.6286822557449341, + "learning_rate": 8.737503876224032e-05, + "loss": 0.1719, + "step": 23053 + }, + { + "epoch": 0.5404220655543639, + "grad_norm": 0.7262802124023438, + "learning_rate": 8.736773157893673e-05, + "loss": 0.1235, + "step": 23054 + }, + { + "epoch": 0.540445507129169, + "grad_norm": 0.36974599957466125, + "learning_rate": 8.736042446417613e-05, + "loss": 0.052, + "step": 23055 + }, + { + "epoch": 0.5404689487039739, + "grad_norm": 0.32300522923469543, + "learning_rate": 8.735311741799827e-05, + "loss": 0.0787, + "step": 23056 + }, + { + "epoch": 0.5404923902787789, + "grad_norm": 0.4345012903213501, + "learning_rate": 8.734581044044273e-05, + "loss": 0.0588, + "step": 23057 + }, + { + "epoch": 0.5405158318535839, + "grad_norm": 0.155609592795372, + "learning_rate": 8.73385035315492e-05, + "loss": 0.0162, + "step": 23058 + }, + { + "epoch": 0.5405392734283889, + "grad_norm": 0.5734956860542297, + "learning_rate": 8.733119669135731e-05, + "loss": 0.597, + "step": 23059 + }, + { + "epoch": 0.5405627150031939, + "grad_norm": 0.31588321924209595, + "learning_rate": 8.732388991990668e-05, + "loss": 0.048, + "step": 23060 + }, + { + "epoch": 0.5405861565779989, + "grad_norm": 0.46063339710235596, + "learning_rate": 8.731658321723699e-05, + "loss": 0.0727, + "step": 23061 + }, + { + "epoch": 0.540609598152804, + "grad_norm": 0.12247415632009506, + "learning_rate": 8.730927658338787e-05, + "loss": 0.0224, + "step": 23062 + }, + { + "epoch": 0.5406330397276089, + "grad_norm": 0.6119765639305115, + "learning_rate": 8.730197001839898e-05, + "loss": 0.1528, + "step": 23063 + }, + { + "epoch": 0.540656481302414, + "grad_norm": 0.5357038378715515, + "learning_rate": 8.729466352230997e-05, + "loss": 0.0841, + "step": 23064 + }, + { + "epoch": 0.5406799228772189, + "grad_norm": 0.5282727479934692, + "learning_rate": 8.728735709516044e-05, + "loss": 0.0794, + "step": 23065 + }, + { + "epoch": 0.5407033644520239, + "grad_norm": 0.626081109046936, + "learning_rate": 8.72800507369901e-05, + "loss": 0.6126, + "step": 23066 + }, + { + "epoch": 0.5407268060268289, + "grad_norm": 0.3952521085739136, + "learning_rate": 8.727274444783851e-05, + "loss": 0.0779, + "step": 23067 + }, + { + "epoch": 0.5407502476016339, + "grad_norm": 0.30923372507095337, + "learning_rate": 8.72654382277454e-05, + "loss": 0.0385, + "step": 23068 + }, + { + "epoch": 0.5407736891764389, + "grad_norm": 0.5169662237167358, + "learning_rate": 8.725813207675037e-05, + "loss": 0.0803, + "step": 23069 + }, + { + "epoch": 0.5407971307512439, + "grad_norm": 0.5999913215637207, + "learning_rate": 8.725082599489311e-05, + "loss": 0.6873, + "step": 23070 + }, + { + "epoch": 0.5408205723260489, + "grad_norm": 0.197014719247818, + "learning_rate": 8.72435199822132e-05, + "loss": 0.0465, + "step": 23071 + }, + { + "epoch": 0.5408440139008539, + "grad_norm": 0.19690530002117157, + "learning_rate": 8.723621403875031e-05, + "loss": 0.0242, + "step": 23072 + }, + { + "epoch": 0.5408674554756588, + "grad_norm": 0.33163830637931824, + "learning_rate": 8.72289081645441e-05, + "loss": 0.0625, + "step": 23073 + }, + { + "epoch": 0.5408908970504639, + "grad_norm": 0.458927184343338, + "learning_rate": 8.722160235963417e-05, + "loss": 0.0631, + "step": 23074 + }, + { + "epoch": 0.5409143386252688, + "grad_norm": 0.19495472311973572, + "learning_rate": 8.72142966240602e-05, + "loss": 0.0249, + "step": 23075 + }, + { + "epoch": 0.5409377802000739, + "grad_norm": 0.34797951579093933, + "learning_rate": 8.720699095786181e-05, + "loss": 0.0648, + "step": 23076 + }, + { + "epoch": 0.5409612217748788, + "grad_norm": 0.18326056003570557, + "learning_rate": 8.719968536107866e-05, + "loss": 0.0134, + "step": 23077 + }, + { + "epoch": 0.5409846633496839, + "grad_norm": 0.5482969880104065, + "learning_rate": 8.719237983375035e-05, + "loss": 0.1262, + "step": 23078 + }, + { + "epoch": 0.5410081049244888, + "grad_norm": 0.1412101536989212, + "learning_rate": 8.718507437591656e-05, + "loss": 0.0157, + "step": 23079 + }, + { + "epoch": 0.5410315464992939, + "grad_norm": 0.7988640069961548, + "learning_rate": 8.717776898761689e-05, + "loss": 0.1532, + "step": 23080 + }, + { + "epoch": 0.5410549880740988, + "grad_norm": 0.36684924364089966, + "learning_rate": 8.717046366889105e-05, + "loss": 0.1163, + "step": 23081 + }, + { + "epoch": 0.5410784296489038, + "grad_norm": 0.293709933757782, + "learning_rate": 8.71631584197786e-05, + "loss": 0.0554, + "step": 23082 + }, + { + "epoch": 0.5411018712237088, + "grad_norm": 0.30503755807876587, + "learning_rate": 8.715585324031923e-05, + "loss": 0.038, + "step": 23083 + }, + { + "epoch": 0.5411253127985138, + "grad_norm": 0.43039366602897644, + "learning_rate": 8.714854813055258e-05, + "loss": 0.0986, + "step": 23084 + }, + { + "epoch": 0.5411487543733188, + "grad_norm": 0.3449662923812866, + "learning_rate": 8.714124309051824e-05, + "loss": 0.0524, + "step": 23085 + }, + { + "epoch": 0.5411721959481238, + "grad_norm": 0.23495769500732422, + "learning_rate": 8.71339381202559e-05, + "loss": 0.0446, + "step": 23086 + }, + { + "epoch": 0.5411956375229288, + "grad_norm": 0.11818215996026993, + "learning_rate": 8.712663321980518e-05, + "loss": 0.0391, + "step": 23087 + }, + { + "epoch": 0.5412190790977338, + "grad_norm": 0.25864943861961365, + "learning_rate": 8.711932838920567e-05, + "loss": 0.0424, + "step": 23088 + }, + { + "epoch": 0.5412425206725388, + "grad_norm": 0.40289220213890076, + "learning_rate": 8.711202362849708e-05, + "loss": 0.0911, + "step": 23089 + }, + { + "epoch": 0.5412659622473438, + "grad_norm": 0.12442183494567871, + "learning_rate": 8.710471893771901e-05, + "loss": 0.0165, + "step": 23090 + }, + { + "epoch": 0.5412894038221487, + "grad_norm": 0.3858296573162079, + "learning_rate": 8.709741431691107e-05, + "loss": 0.219, + "step": 23091 + }, + { + "epoch": 0.5413128453969538, + "grad_norm": 0.0876302495598793, + "learning_rate": 8.709010976611291e-05, + "loss": 0.0166, + "step": 23092 + }, + { + "epoch": 0.5413362869717587, + "grad_norm": 0.0724998489022255, + "learning_rate": 8.708280528536422e-05, + "loss": 0.0109, + "step": 23093 + }, + { + "epoch": 0.5413597285465638, + "grad_norm": 0.35234734416007996, + "learning_rate": 8.70755008747046e-05, + "loss": 0.0677, + "step": 23094 + }, + { + "epoch": 0.5413831701213687, + "grad_norm": 0.406849205493927, + "learning_rate": 8.706819653417363e-05, + "loss": 0.0951, + "step": 23095 + }, + { + "epoch": 0.5414066116961738, + "grad_norm": 0.1598537713289261, + "learning_rate": 8.706089226381102e-05, + "loss": 0.0255, + "step": 23096 + }, + { + "epoch": 0.5414300532709787, + "grad_norm": 0.351351797580719, + "learning_rate": 8.705358806365639e-05, + "loss": 0.04, + "step": 23097 + }, + { + "epoch": 0.5414534948457838, + "grad_norm": 0.37567633390426636, + "learning_rate": 8.704628393374932e-05, + "loss": 0.074, + "step": 23098 + }, + { + "epoch": 0.5414769364205887, + "grad_norm": 0.193365678191185, + "learning_rate": 8.70389798741295e-05, + "loss": 0.0363, + "step": 23099 + }, + { + "epoch": 0.5415003779953937, + "grad_norm": 0.6518443822860718, + "learning_rate": 8.703167588483652e-05, + "loss": 0.8145, + "step": 23100 + }, + { + "epoch": 0.5415238195701987, + "grad_norm": 0.1438988298177719, + "learning_rate": 8.702437196591004e-05, + "loss": 0.0295, + "step": 23101 + }, + { + "epoch": 0.5415472611450037, + "grad_norm": 0.15463078022003174, + "learning_rate": 8.70170681173897e-05, + "loss": 0.0183, + "step": 23102 + }, + { + "epoch": 0.5415707027198087, + "grad_norm": 0.2071203738451004, + "learning_rate": 8.700976433931509e-05, + "loss": 0.0351, + "step": 23103 + }, + { + "epoch": 0.5415941442946137, + "grad_norm": 0.3368207812309265, + "learning_rate": 8.700246063172583e-05, + "loss": 0.0242, + "step": 23104 + }, + { + "epoch": 0.5416175858694187, + "grad_norm": 0.48883485794067383, + "learning_rate": 8.699515699466163e-05, + "loss": 0.0888, + "step": 23105 + }, + { + "epoch": 0.5416410274442237, + "grad_norm": 0.39930784702301025, + "learning_rate": 8.698785342816208e-05, + "loss": 0.1098, + "step": 23106 + }, + { + "epoch": 0.5416644690190286, + "grad_norm": 0.31404465436935425, + "learning_rate": 8.698054993226678e-05, + "loss": 0.0569, + "step": 23107 + }, + { + "epoch": 0.5416879105938337, + "grad_norm": 0.5245173573493958, + "learning_rate": 8.69732465070154e-05, + "loss": 0.0992, + "step": 23108 + }, + { + "epoch": 0.5417113521686386, + "grad_norm": 0.29148223996162415, + "learning_rate": 8.696594315244756e-05, + "loss": 0.0284, + "step": 23109 + }, + { + "epoch": 0.5417347937434437, + "grad_norm": 0.46410953998565674, + "learning_rate": 8.695863986860285e-05, + "loss": 0.0555, + "step": 23110 + }, + { + "epoch": 0.5417582353182486, + "grad_norm": 0.8893687725067139, + "learning_rate": 8.695133665552095e-05, + "loss": 0.1774, + "step": 23111 + }, + { + "epoch": 0.5417816768930537, + "grad_norm": 0.5238153338432312, + "learning_rate": 8.694403351324146e-05, + "loss": 0.5931, + "step": 23112 + }, + { + "epoch": 0.5418051184678587, + "grad_norm": 0.4540095031261444, + "learning_rate": 8.693673044180399e-05, + "loss": 0.0506, + "step": 23113 + }, + { + "epoch": 0.5418285600426637, + "grad_norm": 0.1534147709608078, + "learning_rate": 8.692942744124823e-05, + "loss": 0.0251, + "step": 23114 + }, + { + "epoch": 0.5418520016174687, + "grad_norm": 0.5494937300682068, + "learning_rate": 8.692212451161374e-05, + "loss": 0.5994, + "step": 23115 + }, + { + "epoch": 0.5418754431922737, + "grad_norm": 0.5596886873245239, + "learning_rate": 8.691482165294013e-05, + "loss": 0.3854, + "step": 23116 + }, + { + "epoch": 0.5418988847670787, + "grad_norm": 0.6219000816345215, + "learning_rate": 8.690751886526709e-05, + "loss": 0.6329, + "step": 23117 + }, + { + "epoch": 0.5419223263418836, + "grad_norm": 0.6382867693901062, + "learning_rate": 8.690021614863425e-05, + "loss": 0.0787, + "step": 23118 + }, + { + "epoch": 0.5419457679166887, + "grad_norm": 0.5691304802894592, + "learning_rate": 8.68929135030812e-05, + "loss": 0.1057, + "step": 23119 + }, + { + "epoch": 0.5419692094914936, + "grad_norm": 0.9489820003509521, + "learning_rate": 8.688561092864755e-05, + "loss": 0.126, + "step": 23120 + }, + { + "epoch": 0.5419926510662987, + "grad_norm": 0.38091808557510376, + "learning_rate": 8.687830842537297e-05, + "loss": 0.0498, + "step": 23121 + }, + { + "epoch": 0.5420160926411036, + "grad_norm": 0.5112861394882202, + "learning_rate": 8.687100599329705e-05, + "loss": 0.0766, + "step": 23122 + }, + { + "epoch": 0.5420395342159087, + "grad_norm": 0.3855912387371063, + "learning_rate": 8.68637036324594e-05, + "loss": 0.0622, + "step": 23123 + }, + { + "epoch": 0.5420629757907136, + "grad_norm": 0.5395631790161133, + "learning_rate": 8.685640134289968e-05, + "loss": 0.0786, + "step": 23124 + }, + { + "epoch": 0.5420864173655187, + "grad_norm": 0.6639795303344727, + "learning_rate": 8.684909912465749e-05, + "loss": 0.0539, + "step": 23125 + }, + { + "epoch": 0.5421098589403236, + "grad_norm": 0.12901785969734192, + "learning_rate": 8.684179697777246e-05, + "loss": 0.0134, + "step": 23126 + }, + { + "epoch": 0.5421333005151286, + "grad_norm": 0.4683445692062378, + "learning_rate": 8.683449490228421e-05, + "loss": 0.0966, + "step": 23127 + }, + { + "epoch": 0.5421567420899336, + "grad_norm": 0.575253963470459, + "learning_rate": 8.682719289823234e-05, + "loss": 0.1004, + "step": 23128 + }, + { + "epoch": 0.5421801836647386, + "grad_norm": 0.20045697689056396, + "learning_rate": 8.681989096565651e-05, + "loss": 0.0427, + "step": 23129 + }, + { + "epoch": 0.5422036252395436, + "grad_norm": 0.7718073725700378, + "learning_rate": 8.68125891045963e-05, + "loss": 0.2843, + "step": 23130 + }, + { + "epoch": 0.5422270668143486, + "grad_norm": 0.17342469096183777, + "learning_rate": 8.68052873150914e-05, + "loss": 0.0206, + "step": 23131 + }, + { + "epoch": 0.5422505083891536, + "grad_norm": 0.20466096699237823, + "learning_rate": 8.679798559718136e-05, + "loss": 0.041, + "step": 23132 + }, + { + "epoch": 0.5422739499639586, + "grad_norm": 0.6737682819366455, + "learning_rate": 8.679068395090581e-05, + "loss": 0.1108, + "step": 23133 + }, + { + "epoch": 0.5422973915387636, + "grad_norm": 0.4379879832267761, + "learning_rate": 8.67833823763044e-05, + "loss": 0.1215, + "step": 23134 + }, + { + "epoch": 0.5423208331135686, + "grad_norm": 0.5221585035324097, + "learning_rate": 8.677608087341671e-05, + "loss": 0.0686, + "step": 23135 + }, + { + "epoch": 0.5423442746883735, + "grad_norm": 0.5515698790550232, + "learning_rate": 8.676877944228238e-05, + "loss": 0.1023, + "step": 23136 + }, + { + "epoch": 0.5423677162631786, + "grad_norm": 0.4515795111656189, + "learning_rate": 8.676147808294105e-05, + "loss": 0.1112, + "step": 23137 + }, + { + "epoch": 0.5423911578379835, + "grad_norm": 0.45844361186027527, + "learning_rate": 8.675417679543229e-05, + "loss": 0.1453, + "step": 23138 + }, + { + "epoch": 0.5424145994127886, + "grad_norm": 0.25072652101516724, + "learning_rate": 8.674687557979575e-05, + "loss": 0.0496, + "step": 23139 + }, + { + "epoch": 0.5424380409875935, + "grad_norm": 0.18167273700237274, + "learning_rate": 8.673957443607099e-05, + "loss": 0.0374, + "step": 23140 + }, + { + "epoch": 0.5424614825623986, + "grad_norm": 0.2876220643520355, + "learning_rate": 8.67322733642977e-05, + "loss": 0.0579, + "step": 23141 + }, + { + "epoch": 0.5424849241372035, + "grad_norm": 0.09269440919160843, + "learning_rate": 8.672497236451548e-05, + "loss": 0.0175, + "step": 23142 + }, + { + "epoch": 0.5425083657120086, + "grad_norm": 0.7223796248435974, + "learning_rate": 8.671767143676394e-05, + "loss": 0.1024, + "step": 23143 + }, + { + "epoch": 0.5425318072868135, + "grad_norm": 0.3999447226524353, + "learning_rate": 8.67103705810827e-05, + "loss": 0.0607, + "step": 23144 + }, + { + "epoch": 0.5425552488616185, + "grad_norm": 0.3147115409374237, + "learning_rate": 8.670306979751132e-05, + "loss": 0.3405, + "step": 23145 + }, + { + "epoch": 0.5425786904364235, + "grad_norm": 0.5822736620903015, + "learning_rate": 8.669576908608949e-05, + "loss": 0.1404, + "step": 23146 + }, + { + "epoch": 0.5426021320112285, + "grad_norm": 0.12243454903364182, + "learning_rate": 8.668846844685678e-05, + "loss": 0.0273, + "step": 23147 + }, + { + "epoch": 0.5426255735860335, + "grad_norm": 0.4845184087753296, + "learning_rate": 8.66811678798528e-05, + "loss": 0.1902, + "step": 23148 + }, + { + "epoch": 0.5426490151608385, + "grad_norm": 0.3385649025440216, + "learning_rate": 8.66738673851172e-05, + "loss": 0.0738, + "step": 23149 + }, + { + "epoch": 0.5426724567356435, + "grad_norm": 0.3198621869087219, + "learning_rate": 8.666656696268958e-05, + "loss": 0.0459, + "step": 23150 + }, + { + "epoch": 0.5426958983104485, + "grad_norm": 0.25421077013015747, + "learning_rate": 8.665926661260951e-05, + "loss": 0.0442, + "step": 23151 + }, + { + "epoch": 0.5427193398852534, + "grad_norm": 0.1514129638671875, + "learning_rate": 8.665196633491666e-05, + "loss": 0.0126, + "step": 23152 + }, + { + "epoch": 0.5427427814600585, + "grad_norm": 0.7207462191581726, + "learning_rate": 8.664466612965056e-05, + "loss": 0.2121, + "step": 23153 + }, + { + "epoch": 0.5427662230348634, + "grad_norm": 0.5537415146827698, + "learning_rate": 8.663736599685094e-05, + "loss": 0.1022, + "step": 23154 + }, + { + "epoch": 0.5427896646096685, + "grad_norm": 0.519960880279541, + "learning_rate": 8.663006593655731e-05, + "loss": 0.1334, + "step": 23155 + }, + { + "epoch": 0.5428131061844734, + "grad_norm": 0.3869013786315918, + "learning_rate": 8.662276594880935e-05, + "loss": 0.0575, + "step": 23156 + }, + { + "epoch": 0.5428365477592785, + "grad_norm": 0.15854763984680176, + "learning_rate": 8.661546603364665e-05, + "loss": 0.0252, + "step": 23157 + }, + { + "epoch": 0.5428599893340834, + "grad_norm": 0.37546834349632263, + "learning_rate": 8.660816619110876e-05, + "loss": 0.0693, + "step": 23158 + }, + { + "epoch": 0.5428834309088885, + "grad_norm": 0.48428475856781006, + "learning_rate": 8.660086642123537e-05, + "loss": 0.0757, + "step": 23159 + }, + { + "epoch": 0.5429068724836934, + "grad_norm": 0.5575950145721436, + "learning_rate": 8.659356672406606e-05, + "loss": 0.0991, + "step": 23160 + }, + { + "epoch": 0.5429303140584985, + "grad_norm": 0.16056407988071442, + "learning_rate": 8.658626709964041e-05, + "loss": 0.0156, + "step": 23161 + }, + { + "epoch": 0.5429537556333034, + "grad_norm": 0.48779726028442383, + "learning_rate": 8.657896754799807e-05, + "loss": 0.0707, + "step": 23162 + }, + { + "epoch": 0.5429771972081084, + "grad_norm": 0.13618184626102448, + "learning_rate": 8.657166806917863e-05, + "loss": 0.0318, + "step": 23163 + }, + { + "epoch": 0.5430006387829134, + "grad_norm": 0.5219429135322571, + "learning_rate": 8.656436866322167e-05, + "loss": 0.0757, + "step": 23164 + }, + { + "epoch": 0.5430240803577184, + "grad_norm": 0.38212376832962036, + "learning_rate": 8.65570693301668e-05, + "loss": 0.0821, + "step": 23165 + }, + { + "epoch": 0.5430475219325235, + "grad_norm": 0.6016777753829956, + "learning_rate": 8.654977007005372e-05, + "loss": 0.0613, + "step": 23166 + }, + { + "epoch": 0.5430709635073284, + "grad_norm": 0.4493749141693115, + "learning_rate": 8.654247088292193e-05, + "loss": 0.0979, + "step": 23167 + }, + { + "epoch": 0.5430944050821335, + "grad_norm": 0.7840275764465332, + "learning_rate": 8.653517176881107e-05, + "loss": 0.1378, + "step": 23168 + }, + { + "epoch": 0.5431178466569384, + "grad_norm": 0.14274388551712036, + "learning_rate": 8.652787272776077e-05, + "loss": 0.0191, + "step": 23169 + }, + { + "epoch": 0.5431412882317435, + "grad_norm": 0.5317254662513733, + "learning_rate": 8.652057375981061e-05, + "loss": 0.0915, + "step": 23170 + }, + { + "epoch": 0.5431647298065484, + "grad_norm": 0.48244380950927734, + "learning_rate": 8.651327486500016e-05, + "loss": 0.1074, + "step": 23171 + }, + { + "epoch": 0.5431881713813534, + "grad_norm": 0.1738773137331009, + "learning_rate": 8.650597604336909e-05, + "loss": 0.0275, + "step": 23172 + }, + { + "epoch": 0.5432116129561584, + "grad_norm": 0.4672448933124542, + "learning_rate": 8.649867729495696e-05, + "loss": 0.081, + "step": 23173 + }, + { + "epoch": 0.5432350545309634, + "grad_norm": 0.3262220025062561, + "learning_rate": 8.649137861980339e-05, + "loss": 0.0398, + "step": 23174 + }, + { + "epoch": 0.5432584961057684, + "grad_norm": 0.6250187754631042, + "learning_rate": 8.6484080017948e-05, + "loss": 0.0987, + "step": 23175 + }, + { + "epoch": 0.5432819376805734, + "grad_norm": 0.5390422344207764, + "learning_rate": 8.647678148943034e-05, + "loss": 0.0807, + "step": 23176 + }, + { + "epoch": 0.5433053792553784, + "grad_norm": 0.3029971718788147, + "learning_rate": 8.646948303429006e-05, + "loss": 0.0561, + "step": 23177 + }, + { + "epoch": 0.5433288208301834, + "grad_norm": 0.2399018406867981, + "learning_rate": 8.646218465256673e-05, + "loss": 0.0316, + "step": 23178 + }, + { + "epoch": 0.5433522624049884, + "grad_norm": 0.7509223222732544, + "learning_rate": 8.645488634429998e-05, + "loss": 0.2069, + "step": 23179 + }, + { + "epoch": 0.5433757039797934, + "grad_norm": 0.09139078855514526, + "learning_rate": 8.644758810952937e-05, + "loss": 0.0113, + "step": 23180 + }, + { + "epoch": 0.5433991455545983, + "grad_norm": 0.4112582206726074, + "learning_rate": 8.644028994829457e-05, + "loss": 0.0601, + "step": 23181 + }, + { + "epoch": 0.5434225871294034, + "grad_norm": 0.5653544068336487, + "learning_rate": 8.643299186063514e-05, + "loss": 0.6511, + "step": 23182 + }, + { + "epoch": 0.5434460287042083, + "grad_norm": 0.18627682328224182, + "learning_rate": 8.642569384659066e-05, + "loss": 0.0327, + "step": 23183 + }, + { + "epoch": 0.5434694702790134, + "grad_norm": 0.5110827684402466, + "learning_rate": 8.641839590620074e-05, + "loss": 0.0806, + "step": 23184 + }, + { + "epoch": 0.5434929118538183, + "grad_norm": 0.6907532215118408, + "learning_rate": 8.6411098039505e-05, + "loss": 0.1339, + "step": 23185 + }, + { + "epoch": 0.5435163534286234, + "grad_norm": 0.0540250726044178, + "learning_rate": 8.640380024654301e-05, + "loss": 0.0053, + "step": 23186 + }, + { + "epoch": 0.5435397950034283, + "grad_norm": 0.43776828050613403, + "learning_rate": 8.639650252735441e-05, + "loss": 0.1036, + "step": 23187 + }, + { + "epoch": 0.5435632365782334, + "grad_norm": 0.6844182014465332, + "learning_rate": 8.638920488197875e-05, + "loss": 0.1001, + "step": 23188 + }, + { + "epoch": 0.5435866781530383, + "grad_norm": 0.32351037859916687, + "learning_rate": 8.63819073104556e-05, + "loss": 0.0521, + "step": 23189 + }, + { + "epoch": 0.5436101197278433, + "grad_norm": 0.14579354226589203, + "learning_rate": 8.637460981282465e-05, + "loss": 0.0162, + "step": 23190 + }, + { + "epoch": 0.5436335613026483, + "grad_norm": 0.5004226565361023, + "learning_rate": 8.636731238912545e-05, + "loss": 0.5931, + "step": 23191 + }, + { + "epoch": 0.5436570028774533, + "grad_norm": 0.11993992328643799, + "learning_rate": 8.636001503939762e-05, + "loss": 0.0155, + "step": 23192 + }, + { + "epoch": 0.5436804444522583, + "grad_norm": 0.311722993850708, + "learning_rate": 8.635271776368069e-05, + "loss": 0.054, + "step": 23193 + }, + { + "epoch": 0.5437038860270633, + "grad_norm": 0.5829364657402039, + "learning_rate": 8.634542056201432e-05, + "loss": 0.1336, + "step": 23194 + }, + { + "epoch": 0.5437273276018683, + "grad_norm": 0.5233567953109741, + "learning_rate": 8.633812343443808e-05, + "loss": 0.089, + "step": 23195 + }, + { + "epoch": 0.5437507691766733, + "grad_norm": 0.5783617496490479, + "learning_rate": 8.633082638099157e-05, + "loss": 0.7036, + "step": 23196 + }, + { + "epoch": 0.5437742107514782, + "grad_norm": 0.5322703123092651, + "learning_rate": 8.632352940171438e-05, + "loss": 0.0747, + "step": 23197 + }, + { + "epoch": 0.5437976523262833, + "grad_norm": 0.274894654750824, + "learning_rate": 8.631623249664611e-05, + "loss": 0.0352, + "step": 23198 + }, + { + "epoch": 0.5438210939010882, + "grad_norm": 0.5118321776390076, + "learning_rate": 8.630893566582631e-05, + "loss": 0.0947, + "step": 23199 + }, + { + "epoch": 0.5438445354758933, + "grad_norm": 0.6743553280830383, + "learning_rate": 8.630163890929466e-05, + "loss": 0.1036, + "step": 23200 + }, + { + "epoch": 0.5438679770506982, + "grad_norm": 0.45388129353523254, + "learning_rate": 8.629434222709064e-05, + "loss": 0.1076, + "step": 23201 + }, + { + "epoch": 0.5438914186255033, + "grad_norm": 0.15873849391937256, + "learning_rate": 8.628704561925395e-05, + "loss": 0.0209, + "step": 23202 + }, + { + "epoch": 0.5439148602003082, + "grad_norm": 0.3863013684749603, + "learning_rate": 8.627974908582411e-05, + "loss": 0.0764, + "step": 23203 + }, + { + "epoch": 0.5439383017751133, + "grad_norm": 0.43793490529060364, + "learning_rate": 8.627245262684077e-05, + "loss": 0.0609, + "step": 23204 + }, + { + "epoch": 0.5439617433499182, + "grad_norm": 0.4981388747692108, + "learning_rate": 8.626515624234349e-05, + "loss": 0.0459, + "step": 23205 + }, + { + "epoch": 0.5439851849247233, + "grad_norm": 0.3414331078529358, + "learning_rate": 8.625785993237184e-05, + "loss": 0.0869, + "step": 23206 + }, + { + "epoch": 0.5440086264995282, + "grad_norm": 1.055829644203186, + "learning_rate": 8.625056369696543e-05, + "loss": 0.1851, + "step": 23207 + }, + { + "epoch": 0.5440320680743332, + "grad_norm": 0.4322068393230438, + "learning_rate": 8.624326753616386e-05, + "loss": 0.0701, + "step": 23208 + }, + { + "epoch": 0.5440555096491382, + "grad_norm": 0.20275090634822845, + "learning_rate": 8.623597145000668e-05, + "loss": 0.0349, + "step": 23209 + }, + { + "epoch": 0.5440789512239432, + "grad_norm": 0.08948756009340286, + "learning_rate": 8.622867543853353e-05, + "loss": 0.0201, + "step": 23210 + }, + { + "epoch": 0.5441023927987482, + "grad_norm": 0.9405078887939453, + "learning_rate": 8.622137950178395e-05, + "loss": 0.3127, + "step": 23211 + }, + { + "epoch": 0.5441258343735532, + "grad_norm": 0.3897518813610077, + "learning_rate": 8.621408363979759e-05, + "loss": 0.0496, + "step": 23212 + }, + { + "epoch": 0.5441492759483582, + "grad_norm": 0.13599447906017303, + "learning_rate": 8.620678785261394e-05, + "loss": 0.0201, + "step": 23213 + }, + { + "epoch": 0.5441727175231632, + "grad_norm": 0.4633468687534332, + "learning_rate": 8.619949214027269e-05, + "loss": 0.6654, + "step": 23214 + }, + { + "epoch": 0.5441961590979681, + "grad_norm": 0.4903029203414917, + "learning_rate": 8.619219650281339e-05, + "loss": 0.1504, + "step": 23215 + }, + { + "epoch": 0.5442196006727732, + "grad_norm": 0.5347883701324463, + "learning_rate": 8.618490094027559e-05, + "loss": 0.1224, + "step": 23216 + }, + { + "epoch": 0.5442430422475782, + "grad_norm": 0.17686446011066437, + "learning_rate": 8.617760545269893e-05, + "loss": 0.0244, + "step": 23217 + }, + { + "epoch": 0.5442664838223832, + "grad_norm": 0.3575534522533417, + "learning_rate": 8.617031004012295e-05, + "loss": 0.098, + "step": 23218 + }, + { + "epoch": 0.5442899253971882, + "grad_norm": 0.14302638173103333, + "learning_rate": 8.61630147025873e-05, + "loss": 0.0392, + "step": 23219 + }, + { + "epoch": 0.5443133669719932, + "grad_norm": 0.48528972268104553, + "learning_rate": 8.61557194401315e-05, + "loss": 0.7213, + "step": 23220 + }, + { + "epoch": 0.5443368085467982, + "grad_norm": 0.7071468234062195, + "learning_rate": 8.614842425279513e-05, + "loss": 0.1429, + "step": 23221 + }, + { + "epoch": 0.5443602501216032, + "grad_norm": 0.3308035433292389, + "learning_rate": 8.614112914061782e-05, + "loss": 0.0695, + "step": 23222 + }, + { + "epoch": 0.5443836916964082, + "grad_norm": 0.47625046968460083, + "learning_rate": 8.613383410363915e-05, + "loss": 0.5355, + "step": 23223 + }, + { + "epoch": 0.5444071332712132, + "grad_norm": 0.5338044762611389, + "learning_rate": 8.612653914189866e-05, + "loss": 0.0548, + "step": 23224 + }, + { + "epoch": 0.5444305748460182, + "grad_norm": 0.4212610721588135, + "learning_rate": 8.611924425543597e-05, + "loss": 0.1046, + "step": 23225 + }, + { + "epoch": 0.5444540164208231, + "grad_norm": 0.2979375123977661, + "learning_rate": 8.611194944429062e-05, + "loss": 0.0464, + "step": 23226 + }, + { + "epoch": 0.5444774579956282, + "grad_norm": 0.17492972314357758, + "learning_rate": 8.610465470850226e-05, + "loss": 0.0378, + "step": 23227 + }, + { + "epoch": 0.5445008995704331, + "grad_norm": 0.4569138288497925, + "learning_rate": 8.609736004811041e-05, + "loss": 0.486, + "step": 23228 + }, + { + "epoch": 0.5445243411452382, + "grad_norm": 0.35132691264152527, + "learning_rate": 8.60900654631547e-05, + "loss": 0.0434, + "step": 23229 + }, + { + "epoch": 0.5445477827200431, + "grad_norm": 0.348317414522171, + "learning_rate": 8.608277095367469e-05, + "loss": 0.0517, + "step": 23230 + }, + { + "epoch": 0.5445712242948482, + "grad_norm": 0.3990510404109955, + "learning_rate": 8.607547651970993e-05, + "loss": 0.4329, + "step": 23231 + }, + { + "epoch": 0.5445946658696531, + "grad_norm": 0.6785221695899963, + "learning_rate": 8.606818216130003e-05, + "loss": 0.115, + "step": 23232 + }, + { + "epoch": 0.5446181074444582, + "grad_norm": 0.34779417514801025, + "learning_rate": 8.60608878784846e-05, + "loss": 0.3941, + "step": 23233 + }, + { + "epoch": 0.5446415490192631, + "grad_norm": 0.5788614749908447, + "learning_rate": 8.605359367130314e-05, + "loss": 0.1228, + "step": 23234 + }, + { + "epoch": 0.5446649905940681, + "grad_norm": 0.13407914340496063, + "learning_rate": 8.604629953979531e-05, + "loss": 0.0179, + "step": 23235 + }, + { + "epoch": 0.5446884321688731, + "grad_norm": 0.22795143723487854, + "learning_rate": 8.603900548400063e-05, + "loss": 0.0601, + "step": 23236 + }, + { + "epoch": 0.5447118737436781, + "grad_norm": 0.13840621709823608, + "learning_rate": 8.603171150395869e-05, + "loss": 0.0325, + "step": 23237 + }, + { + "epoch": 0.5447353153184831, + "grad_norm": 0.2628374695777893, + "learning_rate": 8.602441759970905e-05, + "loss": 0.054, + "step": 23238 + }, + { + "epoch": 0.5447587568932881, + "grad_norm": 0.44649699330329895, + "learning_rate": 8.601712377129137e-05, + "loss": 0.1144, + "step": 23239 + }, + { + "epoch": 0.5447821984680931, + "grad_norm": 0.4289623200893402, + "learning_rate": 8.600983001874515e-05, + "loss": 0.1253, + "step": 23240 + }, + { + "epoch": 0.5448056400428981, + "grad_norm": 0.6758870482444763, + "learning_rate": 8.600253634210997e-05, + "loss": 0.4575, + "step": 23241 + }, + { + "epoch": 0.544829081617703, + "grad_norm": 0.5999852418899536, + "learning_rate": 8.599524274142545e-05, + "loss": 0.1277, + "step": 23242 + }, + { + "epoch": 0.5448525231925081, + "grad_norm": 0.7284667491912842, + "learning_rate": 8.598794921673114e-05, + "loss": 0.1248, + "step": 23243 + }, + { + "epoch": 0.544875964767313, + "grad_norm": 0.311442494392395, + "learning_rate": 8.598065576806657e-05, + "loss": 0.0661, + "step": 23244 + }, + { + "epoch": 0.5448994063421181, + "grad_norm": 0.46802154183387756, + "learning_rate": 8.597336239547139e-05, + "loss": 0.0878, + "step": 23245 + }, + { + "epoch": 0.544922847916923, + "grad_norm": 0.2233964055776596, + "learning_rate": 8.596606909898514e-05, + "loss": 0.034, + "step": 23246 + }, + { + "epoch": 0.5449462894917281, + "grad_norm": 0.6279354095458984, + "learning_rate": 8.595877587864737e-05, + "loss": 0.115, + "step": 23247 + }, + { + "epoch": 0.544969731066533, + "grad_norm": 0.20921021699905396, + "learning_rate": 8.59514827344977e-05, + "loss": 0.0638, + "step": 23248 + }, + { + "epoch": 0.5449931726413381, + "grad_norm": 0.4882664680480957, + "learning_rate": 8.594418966657567e-05, + "loss": 0.0719, + "step": 23249 + }, + { + "epoch": 0.545016614216143, + "grad_norm": 0.520892858505249, + "learning_rate": 8.593689667492084e-05, + "loss": 0.0782, + "step": 23250 + }, + { + "epoch": 0.545040055790948, + "grad_norm": 0.24521023035049438, + "learning_rate": 8.592960375957282e-05, + "loss": 0.0223, + "step": 23251 + }, + { + "epoch": 0.545063497365753, + "grad_norm": 0.1556108444929123, + "learning_rate": 8.592231092057119e-05, + "loss": 0.0289, + "step": 23252 + }, + { + "epoch": 0.545086938940558, + "grad_norm": 0.49508509039878845, + "learning_rate": 8.591501815795547e-05, + "loss": 0.1239, + "step": 23253 + }, + { + "epoch": 0.545110380515363, + "grad_norm": 0.46040982007980347, + "learning_rate": 8.590772547176527e-05, + "loss": 0.0462, + "step": 23254 + }, + { + "epoch": 0.545133822090168, + "grad_norm": 0.1768243908882141, + "learning_rate": 8.590043286204014e-05, + "loss": 0.0121, + "step": 23255 + }, + { + "epoch": 0.545157263664973, + "grad_norm": 0.39917615056037903, + "learning_rate": 8.589314032881965e-05, + "loss": 0.0608, + "step": 23256 + }, + { + "epoch": 0.545180705239778, + "grad_norm": 0.36599525809288025, + "learning_rate": 8.588584787214341e-05, + "loss": 0.0742, + "step": 23257 + }, + { + "epoch": 0.545204146814583, + "grad_norm": 0.4517713189125061, + "learning_rate": 8.587855549205094e-05, + "loss": 0.1303, + "step": 23258 + }, + { + "epoch": 0.545227588389388, + "grad_norm": 0.14990708231925964, + "learning_rate": 8.587126318858181e-05, + "loss": 0.034, + "step": 23259 + }, + { + "epoch": 0.545251029964193, + "grad_norm": 0.410275936126709, + "learning_rate": 8.586397096177562e-05, + "loss": 0.1126, + "step": 23260 + }, + { + "epoch": 0.545274471538998, + "grad_norm": 0.43739795684814453, + "learning_rate": 8.585667881167192e-05, + "loss": 0.0596, + "step": 23261 + }, + { + "epoch": 0.5452979131138029, + "grad_norm": 0.5222660899162292, + "learning_rate": 8.584938673831025e-05, + "loss": 0.0659, + "step": 23262 + }, + { + "epoch": 0.545321354688608, + "grad_norm": 0.5724724531173706, + "learning_rate": 8.584209474173022e-05, + "loss": 0.0949, + "step": 23263 + }, + { + "epoch": 0.5453447962634129, + "grad_norm": 0.8403461575508118, + "learning_rate": 8.58348028219714e-05, + "loss": 0.0897, + "step": 23264 + }, + { + "epoch": 0.545368237838218, + "grad_norm": 0.292894184589386, + "learning_rate": 8.582751097907333e-05, + "loss": 0.039, + "step": 23265 + }, + { + "epoch": 0.5453916794130229, + "grad_norm": 0.3148462772369385, + "learning_rate": 8.582021921307556e-05, + "loss": 0.0598, + "step": 23266 + }, + { + "epoch": 0.545415120987828, + "grad_norm": 0.16098955273628235, + "learning_rate": 8.581292752401773e-05, + "loss": 0.0163, + "step": 23267 + }, + { + "epoch": 0.545438562562633, + "grad_norm": 0.6272377371788025, + "learning_rate": 8.580563591193933e-05, + "loss": 0.1171, + "step": 23268 + }, + { + "epoch": 0.545462004137438, + "grad_norm": 0.5299947261810303, + "learning_rate": 8.579834437687994e-05, + "loss": 0.1143, + "step": 23269 + }, + { + "epoch": 0.545485445712243, + "grad_norm": 0.44802364706993103, + "learning_rate": 8.579105291887915e-05, + "loss": 0.0846, + "step": 23270 + }, + { + "epoch": 0.5455088872870479, + "grad_norm": 0.33945173025131226, + "learning_rate": 8.57837615379765e-05, + "loss": 0.0535, + "step": 23271 + }, + { + "epoch": 0.545532328861853, + "grad_norm": 0.1221139207482338, + "learning_rate": 8.577647023421154e-05, + "loss": 0.0255, + "step": 23272 + }, + { + "epoch": 0.5455557704366579, + "grad_norm": 0.4249729812145233, + "learning_rate": 8.576917900762387e-05, + "loss": 0.0786, + "step": 23273 + }, + { + "epoch": 0.545579212011463, + "grad_norm": 0.4413066804409027, + "learning_rate": 8.5761887858253e-05, + "loss": 0.5024, + "step": 23274 + }, + { + "epoch": 0.5456026535862679, + "grad_norm": 0.4455411434173584, + "learning_rate": 8.575459678613857e-05, + "loss": 0.1034, + "step": 23275 + }, + { + "epoch": 0.545626095161073, + "grad_norm": 0.42896801233291626, + "learning_rate": 8.574730579132006e-05, + "loss": 0.0636, + "step": 23276 + }, + { + "epoch": 0.5456495367358779, + "grad_norm": 0.35073143243789673, + "learning_rate": 8.574001487383711e-05, + "loss": 0.0415, + "step": 23277 + }, + { + "epoch": 0.545672978310683, + "grad_norm": 0.39999493956565857, + "learning_rate": 8.573272403372924e-05, + "loss": 0.2862, + "step": 23278 + }, + { + "epoch": 0.5456964198854879, + "grad_norm": 0.49914857745170593, + "learning_rate": 8.572543327103598e-05, + "loss": 0.0895, + "step": 23279 + }, + { + "epoch": 0.5457198614602929, + "grad_norm": 0.14274927973747253, + "learning_rate": 8.571814258579695e-05, + "loss": 0.0208, + "step": 23280 + }, + { + "epoch": 0.5457433030350979, + "grad_norm": 0.4391818046569824, + "learning_rate": 8.571085197805167e-05, + "loss": 0.0602, + "step": 23281 + }, + { + "epoch": 0.5457667446099029, + "grad_norm": 0.09450215101242065, + "learning_rate": 8.570356144783969e-05, + "loss": 0.0179, + "step": 23282 + }, + { + "epoch": 0.5457901861847079, + "grad_norm": 0.7132269144058228, + "learning_rate": 8.569627099520062e-05, + "loss": 0.0775, + "step": 23283 + }, + { + "epoch": 0.5458136277595129, + "grad_norm": 0.5247676372528076, + "learning_rate": 8.568898062017398e-05, + "loss": 0.119, + "step": 23284 + }, + { + "epoch": 0.5458370693343179, + "grad_norm": 0.5900933742523193, + "learning_rate": 8.568169032279931e-05, + "loss": 0.1473, + "step": 23285 + }, + { + "epoch": 0.5458605109091229, + "grad_norm": 0.5953505635261536, + "learning_rate": 8.567440010311617e-05, + "loss": 0.1506, + "step": 23286 + }, + { + "epoch": 0.5458839524839278, + "grad_norm": 0.8685161471366882, + "learning_rate": 8.566710996116419e-05, + "loss": 0.1161, + "step": 23287 + }, + { + "epoch": 0.5459073940587329, + "grad_norm": 0.4966977834701538, + "learning_rate": 8.565981989698287e-05, + "loss": 0.0601, + "step": 23288 + }, + { + "epoch": 0.5459308356335378, + "grad_norm": 0.35379984974861145, + "learning_rate": 8.565252991061173e-05, + "loss": 0.0581, + "step": 23289 + }, + { + "epoch": 0.5459542772083429, + "grad_norm": 0.7209131717681885, + "learning_rate": 8.56452400020904e-05, + "loss": 0.0934, + "step": 23290 + }, + { + "epoch": 0.5459777187831478, + "grad_norm": 0.16338913142681122, + "learning_rate": 8.56379501714584e-05, + "loss": 0.0176, + "step": 23291 + }, + { + "epoch": 0.5460011603579529, + "grad_norm": 0.5025994777679443, + "learning_rate": 8.563066041875526e-05, + "loss": 0.1171, + "step": 23292 + }, + { + "epoch": 0.5460246019327578, + "grad_norm": 1.4902725219726562, + "learning_rate": 8.56233707440206e-05, + "loss": 0.2722, + "step": 23293 + }, + { + "epoch": 0.5460480435075629, + "grad_norm": 0.5553012490272522, + "learning_rate": 8.561608114729391e-05, + "loss": 0.4246, + "step": 23294 + }, + { + "epoch": 0.5460714850823678, + "grad_norm": 0.42568373680114746, + "learning_rate": 8.560879162861478e-05, + "loss": 0.105, + "step": 23295 + }, + { + "epoch": 0.5460949266571729, + "grad_norm": 0.3906504511833191, + "learning_rate": 8.560150218802276e-05, + "loss": 0.0477, + "step": 23296 + }, + { + "epoch": 0.5461183682319778, + "grad_norm": 0.7627397775650024, + "learning_rate": 8.559421282555737e-05, + "loss": 0.1069, + "step": 23297 + }, + { + "epoch": 0.5461418098067828, + "grad_norm": 0.9152541756629944, + "learning_rate": 8.558692354125815e-05, + "loss": 0.5792, + "step": 23298 + }, + { + "epoch": 0.5461652513815878, + "grad_norm": 0.5554296374320984, + "learning_rate": 8.557963433516475e-05, + "loss": 0.5635, + "step": 23299 + }, + { + "epoch": 0.5461886929563928, + "grad_norm": 0.13248766958713531, + "learning_rate": 8.557234520731666e-05, + "loss": 0.0195, + "step": 23300 + }, + { + "epoch": 0.5462121345311978, + "grad_norm": 0.5200983881950378, + "learning_rate": 8.556505615775339e-05, + "loss": 0.0524, + "step": 23301 + }, + { + "epoch": 0.5462355761060028, + "grad_norm": 0.4527866244316101, + "learning_rate": 8.555776718651456e-05, + "loss": 0.1214, + "step": 23302 + }, + { + "epoch": 0.5462590176808078, + "grad_norm": 0.29905614256858826, + "learning_rate": 8.555047829363969e-05, + "loss": 0.0504, + "step": 23303 + }, + { + "epoch": 0.5462824592556128, + "grad_norm": 0.13381947576999664, + "learning_rate": 8.554318947916832e-05, + "loss": 0.0372, + "step": 23304 + }, + { + "epoch": 0.5463059008304177, + "grad_norm": 0.4005793035030365, + "learning_rate": 8.553590074314003e-05, + "loss": 0.0732, + "step": 23305 + }, + { + "epoch": 0.5463293424052228, + "grad_norm": 0.630850613117218, + "learning_rate": 8.552861208559434e-05, + "loss": 0.1024, + "step": 23306 + }, + { + "epoch": 0.5463527839800277, + "grad_norm": 0.3407174348831177, + "learning_rate": 8.55213235065708e-05, + "loss": 0.0477, + "step": 23307 + }, + { + "epoch": 0.5463762255548328, + "grad_norm": 0.2739395201206207, + "learning_rate": 8.551403500610897e-05, + "loss": 0.0739, + "step": 23308 + }, + { + "epoch": 0.5463996671296377, + "grad_norm": 0.6524683237075806, + "learning_rate": 8.550674658424841e-05, + "loss": 0.0734, + "step": 23309 + }, + { + "epoch": 0.5464231087044428, + "grad_norm": 0.3988727033138275, + "learning_rate": 8.549945824102862e-05, + "loss": 0.0365, + "step": 23310 + }, + { + "epoch": 0.5464465502792477, + "grad_norm": 0.5239477157592773, + "learning_rate": 8.549216997648916e-05, + "loss": 0.426, + "step": 23311 + }, + { + "epoch": 0.5464699918540528, + "grad_norm": 0.6967489719390869, + "learning_rate": 8.548488179066963e-05, + "loss": 0.1295, + "step": 23312 + }, + { + "epoch": 0.5464934334288577, + "grad_norm": 0.07643522322177887, + "learning_rate": 8.547759368360955e-05, + "loss": 0.0113, + "step": 23313 + }, + { + "epoch": 0.5465168750036627, + "grad_norm": 0.49004507064819336, + "learning_rate": 8.547030565534841e-05, + "loss": 0.0386, + "step": 23314 + }, + { + "epoch": 0.5465403165784677, + "grad_norm": 0.3733857274055481, + "learning_rate": 8.546301770592586e-05, + "loss": 0.0929, + "step": 23315 + }, + { + "epoch": 0.5465637581532727, + "grad_norm": 0.2647710144519806, + "learning_rate": 8.545572983538135e-05, + "loss": 0.0214, + "step": 23316 + }, + { + "epoch": 0.5465871997280777, + "grad_norm": 0.21312680840492249, + "learning_rate": 8.544844204375445e-05, + "loss": 0.0455, + "step": 23317 + }, + { + "epoch": 0.5466106413028827, + "grad_norm": 0.6791346073150635, + "learning_rate": 8.544115433108473e-05, + "loss": 0.0503, + "step": 23318 + }, + { + "epoch": 0.5466340828776878, + "grad_norm": 0.6465896368026733, + "learning_rate": 8.543386669741172e-05, + "loss": 0.6512, + "step": 23319 + }, + { + "epoch": 0.5466575244524927, + "grad_norm": 0.6052467823028564, + "learning_rate": 8.542657914277495e-05, + "loss": 0.6623, + "step": 23320 + }, + { + "epoch": 0.5466809660272978, + "grad_norm": 0.704903244972229, + "learning_rate": 8.541929166721397e-05, + "loss": 0.11, + "step": 23321 + }, + { + "epoch": 0.5467044076021027, + "grad_norm": 0.25206562876701355, + "learning_rate": 8.541200427076832e-05, + "loss": 0.0291, + "step": 23322 + }, + { + "epoch": 0.5467278491769078, + "grad_norm": 0.5544643998146057, + "learning_rate": 8.540471695347754e-05, + "loss": 0.1561, + "step": 23323 + }, + { + "epoch": 0.5467512907517127, + "grad_norm": 0.5823367238044739, + "learning_rate": 8.539742971538117e-05, + "loss": 0.1327, + "step": 23324 + }, + { + "epoch": 0.5467747323265177, + "grad_norm": 0.6390489935874939, + "learning_rate": 8.539014255651878e-05, + "loss": 0.6203, + "step": 23325 + }, + { + "epoch": 0.5467981739013227, + "grad_norm": 0.5713656544685364, + "learning_rate": 8.538285547692989e-05, + "loss": 0.1082, + "step": 23326 + }, + { + "epoch": 0.5468216154761277, + "grad_norm": 0.40737384557724, + "learning_rate": 8.537556847665402e-05, + "loss": 0.0592, + "step": 23327 + }, + { + "epoch": 0.5468450570509327, + "grad_norm": 0.38012897968292236, + "learning_rate": 8.536828155573074e-05, + "loss": 0.0514, + "step": 23328 + }, + { + "epoch": 0.5468684986257377, + "grad_norm": 0.22308826446533203, + "learning_rate": 8.536099471419955e-05, + "loss": 0.0487, + "step": 23329 + }, + { + "epoch": 0.5468919402005427, + "grad_norm": 0.474989116191864, + "learning_rate": 8.535370795210005e-05, + "loss": 0.1376, + "step": 23330 + }, + { + "epoch": 0.5469153817753477, + "grad_norm": 0.4775716960430145, + "learning_rate": 8.534642126947175e-05, + "loss": 0.0709, + "step": 23331 + }, + { + "epoch": 0.5469388233501526, + "grad_norm": 0.3206508159637451, + "learning_rate": 8.533913466635414e-05, + "loss": 0.0648, + "step": 23332 + }, + { + "epoch": 0.5469622649249577, + "grad_norm": 0.44429558515548706, + "learning_rate": 8.533184814278682e-05, + "loss": 0.0911, + "step": 23333 + }, + { + "epoch": 0.5469857064997626, + "grad_norm": 0.20835350453853607, + "learning_rate": 8.53245616988093e-05, + "loss": 0.0418, + "step": 23334 + }, + { + "epoch": 0.5470091480745677, + "grad_norm": 0.7340649962425232, + "learning_rate": 8.531727533446109e-05, + "loss": 0.1092, + "step": 23335 + }, + { + "epoch": 0.5470325896493726, + "grad_norm": 0.6209077835083008, + "learning_rate": 8.530998904978178e-05, + "loss": 0.1352, + "step": 23336 + }, + { + "epoch": 0.5470560312241777, + "grad_norm": 0.34401097893714905, + "learning_rate": 8.53027028448109e-05, + "loss": 0.0286, + "step": 23337 + }, + { + "epoch": 0.5470794727989826, + "grad_norm": 0.20185206830501556, + "learning_rate": 8.529541671958796e-05, + "loss": 0.036, + "step": 23338 + }, + { + "epoch": 0.5471029143737877, + "grad_norm": 0.45001572370529175, + "learning_rate": 8.528813067415249e-05, + "loss": 0.0551, + "step": 23339 + }, + { + "epoch": 0.5471263559485926, + "grad_norm": 0.3176442086696625, + "learning_rate": 8.528084470854405e-05, + "loss": 0.0698, + "step": 23340 + }, + { + "epoch": 0.5471497975233977, + "grad_norm": 0.3751319348812103, + "learning_rate": 8.527355882280217e-05, + "loss": 0.0822, + "step": 23341 + }, + { + "epoch": 0.5471732390982026, + "grad_norm": 0.11470764130353928, + "learning_rate": 8.526627301696634e-05, + "loss": 0.013, + "step": 23342 + }, + { + "epoch": 0.5471966806730076, + "grad_norm": 0.3241519331932068, + "learning_rate": 8.525898729107617e-05, + "loss": 0.0353, + "step": 23343 + }, + { + "epoch": 0.5472201222478126, + "grad_norm": 0.4373832941055298, + "learning_rate": 8.525170164517112e-05, + "loss": 0.099, + "step": 23344 + }, + { + "epoch": 0.5472435638226176, + "grad_norm": 0.563691258430481, + "learning_rate": 8.524441607929074e-05, + "loss": 0.095, + "step": 23345 + }, + { + "epoch": 0.5472670053974226, + "grad_norm": 0.4773753583431244, + "learning_rate": 8.523713059347462e-05, + "loss": 0.0697, + "step": 23346 + }, + { + "epoch": 0.5472904469722276, + "grad_norm": 0.8390437960624695, + "learning_rate": 8.522984518776216e-05, + "loss": 0.0826, + "step": 23347 + }, + { + "epoch": 0.5473138885470326, + "grad_norm": 0.5878220200538635, + "learning_rate": 8.522255986219306e-05, + "loss": 0.1367, + "step": 23348 + }, + { + "epoch": 0.5473373301218376, + "grad_norm": 0.432480126619339, + "learning_rate": 8.521527461680671e-05, + "loss": 0.0836, + "step": 23349 + }, + { + "epoch": 0.5473607716966425, + "grad_norm": 0.5172516703605652, + "learning_rate": 8.520798945164274e-05, + "loss": 0.0613, + "step": 23350 + }, + { + "epoch": 0.5473842132714476, + "grad_norm": 0.6251527070999146, + "learning_rate": 8.520070436674063e-05, + "loss": 0.0638, + "step": 23351 + }, + { + "epoch": 0.5474076548462525, + "grad_norm": 0.13834962248802185, + "learning_rate": 8.51934193621399e-05, + "loss": 0.023, + "step": 23352 + }, + { + "epoch": 0.5474310964210576, + "grad_norm": 0.4803439676761627, + "learning_rate": 8.518613443788008e-05, + "loss": 0.0633, + "step": 23353 + }, + { + "epoch": 0.5474545379958625, + "grad_norm": 0.6991757750511169, + "learning_rate": 8.517884959400075e-05, + "loss": 0.1544, + "step": 23354 + }, + { + "epoch": 0.5474779795706676, + "grad_norm": 0.2663195729255676, + "learning_rate": 8.517156483054136e-05, + "loss": 0.0614, + "step": 23355 + }, + { + "epoch": 0.5475014211454725, + "grad_norm": 0.1531415432691574, + "learning_rate": 8.51642801475415e-05, + "loss": 0.0345, + "step": 23356 + }, + { + "epoch": 0.5475248627202776, + "grad_norm": 0.19841444492340088, + "learning_rate": 8.515699554504066e-05, + "loss": 0.0581, + "step": 23357 + }, + { + "epoch": 0.5475483042950825, + "grad_norm": 0.3865766227245331, + "learning_rate": 8.514971102307838e-05, + "loss": 0.1015, + "step": 23358 + }, + { + "epoch": 0.5475717458698875, + "grad_norm": 0.5839787721633911, + "learning_rate": 8.514242658169415e-05, + "loss": 0.1264, + "step": 23359 + }, + { + "epoch": 0.5475951874446925, + "grad_norm": 0.8369814157485962, + "learning_rate": 8.513514222092757e-05, + "loss": 0.2174, + "step": 23360 + }, + { + "epoch": 0.5476186290194975, + "grad_norm": 0.7915257811546326, + "learning_rate": 8.512785794081815e-05, + "loss": 0.155, + "step": 23361 + }, + { + "epoch": 0.5476420705943025, + "grad_norm": 0.4990135729312897, + "learning_rate": 8.512057374140535e-05, + "loss": 0.1038, + "step": 23362 + }, + { + "epoch": 0.5476655121691075, + "grad_norm": 0.6069093942642212, + "learning_rate": 8.511328962272876e-05, + "loss": 0.1604, + "step": 23363 + }, + { + "epoch": 0.5476889537439125, + "grad_norm": 0.5228314995765686, + "learning_rate": 8.510600558482788e-05, + "loss": 0.0406, + "step": 23364 + }, + { + "epoch": 0.5477123953187175, + "grad_norm": 1.0584722757339478, + "learning_rate": 8.509872162774221e-05, + "loss": 0.1597, + "step": 23365 + }, + { + "epoch": 0.5477358368935225, + "grad_norm": 0.5384491086006165, + "learning_rate": 8.509143775151133e-05, + "loss": 0.1129, + "step": 23366 + }, + { + "epoch": 0.5477592784683275, + "grad_norm": 0.39641711115837097, + "learning_rate": 8.50841539561747e-05, + "loss": 0.0549, + "step": 23367 + }, + { + "epoch": 0.5477827200431324, + "grad_norm": 0.5819203853607178, + "learning_rate": 8.507687024177188e-05, + "loss": 0.0923, + "step": 23368 + }, + { + "epoch": 0.5478061616179375, + "grad_norm": 0.4449545741081238, + "learning_rate": 8.50695866083424e-05, + "loss": 0.0942, + "step": 23369 + }, + { + "epoch": 0.5478296031927425, + "grad_norm": 0.09371542185544968, + "learning_rate": 8.506230305592573e-05, + "loss": 0.0145, + "step": 23370 + }, + { + "epoch": 0.5478530447675475, + "grad_norm": 0.6195138096809387, + "learning_rate": 8.50550195845614e-05, + "loss": 0.1457, + "step": 23371 + }, + { + "epoch": 0.5478764863423525, + "grad_norm": 0.26426684856414795, + "learning_rate": 8.504773619428899e-05, + "loss": 0.0253, + "step": 23372 + }, + { + "epoch": 0.5478999279171575, + "grad_norm": 0.17143972218036652, + "learning_rate": 8.504045288514799e-05, + "loss": 0.0319, + "step": 23373 + }, + { + "epoch": 0.5479233694919625, + "grad_norm": 0.32452133297920227, + "learning_rate": 8.50331696571779e-05, + "loss": 0.0389, + "step": 23374 + }, + { + "epoch": 0.5479468110667675, + "grad_norm": 0.354819655418396, + "learning_rate": 8.502588651041826e-05, + "loss": 0.0843, + "step": 23375 + }, + { + "epoch": 0.5479702526415725, + "grad_norm": 1.0071736574172974, + "learning_rate": 8.501860344490859e-05, + "loss": 0.0884, + "step": 23376 + }, + { + "epoch": 0.5479936942163774, + "grad_norm": 0.5681928396224976, + "learning_rate": 8.501132046068838e-05, + "loss": 0.4968, + "step": 23377 + }, + { + "epoch": 0.5480171357911825, + "grad_norm": 0.25601622462272644, + "learning_rate": 8.500403755779718e-05, + "loss": 0.0667, + "step": 23378 + }, + { + "epoch": 0.5480405773659874, + "grad_norm": 0.4174903929233551, + "learning_rate": 8.499675473627449e-05, + "loss": 0.0981, + "step": 23379 + }, + { + "epoch": 0.5480640189407925, + "grad_norm": 0.5513438582420349, + "learning_rate": 8.498947199615983e-05, + "loss": 0.0629, + "step": 23380 + }, + { + "epoch": 0.5480874605155974, + "grad_norm": 0.40174373984336853, + "learning_rate": 8.498218933749273e-05, + "loss": 0.0884, + "step": 23381 + }, + { + "epoch": 0.5481109020904025, + "grad_norm": 0.41212648153305054, + "learning_rate": 8.497490676031267e-05, + "loss": 0.0662, + "step": 23382 + }, + { + "epoch": 0.5481343436652074, + "grad_norm": 0.29151055216789246, + "learning_rate": 8.49676242646592e-05, + "loss": 0.0737, + "step": 23383 + }, + { + "epoch": 0.5481577852400125, + "grad_norm": 0.4920252859592438, + "learning_rate": 8.496034185057178e-05, + "loss": 0.0869, + "step": 23384 + }, + { + "epoch": 0.5481812268148174, + "grad_norm": 0.16882245242595673, + "learning_rate": 8.495305951809003e-05, + "loss": 0.0391, + "step": 23385 + }, + { + "epoch": 0.5482046683896225, + "grad_norm": 0.6675832271575928, + "learning_rate": 8.494577726725339e-05, + "loss": 0.2072, + "step": 23386 + }, + { + "epoch": 0.5482281099644274, + "grad_norm": 0.6105228662490845, + "learning_rate": 8.493849509810136e-05, + "loss": 0.1146, + "step": 23387 + }, + { + "epoch": 0.5482515515392324, + "grad_norm": 0.6100192666053772, + "learning_rate": 8.49312130106735e-05, + "loss": 0.1025, + "step": 23388 + }, + { + "epoch": 0.5482749931140374, + "grad_norm": 0.10935720801353455, + "learning_rate": 8.492393100500932e-05, + "loss": 0.0325, + "step": 23389 + }, + { + "epoch": 0.5482984346888424, + "grad_norm": 0.4977729320526123, + "learning_rate": 8.491664908114827e-05, + "loss": 0.8313, + "step": 23390 + }, + { + "epoch": 0.5483218762636474, + "grad_norm": 0.18982547521591187, + "learning_rate": 8.490936723912994e-05, + "loss": 0.0385, + "step": 23391 + }, + { + "epoch": 0.5483453178384524, + "grad_norm": 0.08808314055204391, + "learning_rate": 8.49020854789938e-05, + "loss": 0.0186, + "step": 23392 + }, + { + "epoch": 0.5483687594132574, + "grad_norm": 0.5383508205413818, + "learning_rate": 8.489480380077937e-05, + "loss": 0.1342, + "step": 23393 + }, + { + "epoch": 0.5483922009880624, + "grad_norm": 0.38177189230918884, + "learning_rate": 8.488752220452616e-05, + "loss": 0.0593, + "step": 23394 + }, + { + "epoch": 0.5484156425628673, + "grad_norm": 0.264028936624527, + "learning_rate": 8.488024069027367e-05, + "loss": 0.0233, + "step": 23395 + }, + { + "epoch": 0.5484390841376724, + "grad_norm": 0.45637059211730957, + "learning_rate": 8.487295925806141e-05, + "loss": 0.4734, + "step": 23396 + }, + { + "epoch": 0.5484625257124773, + "grad_norm": 0.13798080384731293, + "learning_rate": 8.486567790792888e-05, + "loss": 0.0128, + "step": 23397 + }, + { + "epoch": 0.5484859672872824, + "grad_norm": 0.381277471780777, + "learning_rate": 8.485839663991566e-05, + "loss": 0.0593, + "step": 23398 + }, + { + "epoch": 0.5485094088620873, + "grad_norm": 0.7779362797737122, + "learning_rate": 8.485111545406119e-05, + "loss": 0.1471, + "step": 23399 + }, + { + "epoch": 0.5485328504368924, + "grad_norm": 0.34470200538635254, + "learning_rate": 8.484383435040497e-05, + "loss": 0.0785, + "step": 23400 + }, + { + "epoch": 0.5485562920116973, + "grad_norm": 0.1445535570383072, + "learning_rate": 8.483655332898656e-05, + "loss": 0.0342, + "step": 23401 + }, + { + "epoch": 0.5485797335865024, + "grad_norm": 0.11254745721817017, + "learning_rate": 8.482927238984543e-05, + "loss": 0.0153, + "step": 23402 + }, + { + "epoch": 0.5486031751613073, + "grad_norm": 0.25293633341789246, + "learning_rate": 8.482199153302109e-05, + "loss": 0.0415, + "step": 23403 + }, + { + "epoch": 0.5486266167361123, + "grad_norm": 0.508409321308136, + "learning_rate": 8.481471075855306e-05, + "loss": 0.0348, + "step": 23404 + }, + { + "epoch": 0.5486500583109173, + "grad_norm": 0.5210585594177246, + "learning_rate": 8.480743006648082e-05, + "loss": 0.0541, + "step": 23405 + }, + { + "epoch": 0.5486734998857223, + "grad_norm": 0.6798084378242493, + "learning_rate": 8.48001494568439e-05, + "loss": 0.1222, + "step": 23406 + }, + { + "epoch": 0.5486969414605273, + "grad_norm": 0.312700480222702, + "learning_rate": 8.479286892968181e-05, + "loss": 0.086, + "step": 23407 + }, + { + "epoch": 0.5487203830353323, + "grad_norm": 0.5051367878913879, + "learning_rate": 8.4785588485034e-05, + "loss": 0.0663, + "step": 23408 + }, + { + "epoch": 0.5487438246101373, + "grad_norm": 0.8073470592498779, + "learning_rate": 8.477830812294003e-05, + "loss": 0.0896, + "step": 23409 + }, + { + "epoch": 0.5487672661849423, + "grad_norm": 0.5770357251167297, + "learning_rate": 8.477102784343939e-05, + "loss": 0.4849, + "step": 23410 + }, + { + "epoch": 0.5487907077597473, + "grad_norm": 0.6448224186897278, + "learning_rate": 8.476374764657161e-05, + "loss": 0.1185, + "step": 23411 + }, + { + "epoch": 0.5488141493345523, + "grad_norm": 0.2355819195508957, + "learning_rate": 8.475646753237614e-05, + "loss": 0.0244, + "step": 23412 + }, + { + "epoch": 0.5488375909093572, + "grad_norm": 0.3359048068523407, + "learning_rate": 8.474918750089251e-05, + "loss": 0.066, + "step": 23413 + }, + { + "epoch": 0.5488610324841623, + "grad_norm": 0.4497922658920288, + "learning_rate": 8.474190755216022e-05, + "loss": 0.0818, + "step": 23414 + }, + { + "epoch": 0.5488844740589672, + "grad_norm": 0.5852001905441284, + "learning_rate": 8.473462768621878e-05, + "loss": 0.6464, + "step": 23415 + }, + { + "epoch": 0.5489079156337723, + "grad_norm": 0.28293660283088684, + "learning_rate": 8.472734790310765e-05, + "loss": 0.0348, + "step": 23416 + }, + { + "epoch": 0.5489313572085772, + "grad_norm": 0.14378470182418823, + "learning_rate": 8.47200682028664e-05, + "loss": 0.0332, + "step": 23417 + }, + { + "epoch": 0.5489547987833823, + "grad_norm": 0.3472127616405487, + "learning_rate": 8.471278858553445e-05, + "loss": 0.0546, + "step": 23418 + }, + { + "epoch": 0.5489782403581872, + "grad_norm": 0.39759063720703125, + "learning_rate": 8.470550905115137e-05, + "loss": 0.0755, + "step": 23419 + }, + { + "epoch": 0.5490016819329923, + "grad_norm": 0.3452449440956116, + "learning_rate": 8.469822959975657e-05, + "loss": 0.0425, + "step": 23420 + }, + { + "epoch": 0.5490251235077973, + "grad_norm": 0.6263824105262756, + "learning_rate": 8.469095023138967e-05, + "loss": 0.0813, + "step": 23421 + }, + { + "epoch": 0.5490485650826022, + "grad_norm": 0.10944779217243195, + "learning_rate": 8.468367094609007e-05, + "loss": 0.01, + "step": 23422 + }, + { + "epoch": 0.5490720066574073, + "grad_norm": 0.25298184156417847, + "learning_rate": 8.467639174389732e-05, + "loss": 0.0425, + "step": 23423 + }, + { + "epoch": 0.5490954482322122, + "grad_norm": 0.7325530648231506, + "learning_rate": 8.46691126248509e-05, + "loss": 0.1801, + "step": 23424 + }, + { + "epoch": 0.5491188898070173, + "grad_norm": 0.10613353550434113, + "learning_rate": 8.46618335889903e-05, + "loss": 0.0168, + "step": 23425 + }, + { + "epoch": 0.5491423313818222, + "grad_norm": 0.46554481983184814, + "learning_rate": 8.465455463635502e-05, + "loss": 0.1084, + "step": 23426 + }, + { + "epoch": 0.5491657729566273, + "grad_norm": 0.5268246531486511, + "learning_rate": 8.464727576698457e-05, + "loss": 0.0796, + "step": 23427 + }, + { + "epoch": 0.5491892145314322, + "grad_norm": 0.6286261677742004, + "learning_rate": 8.463999698091841e-05, + "loss": 0.036, + "step": 23428 + }, + { + "epoch": 0.5492126561062373, + "grad_norm": 0.2119712382555008, + "learning_rate": 8.463271827819609e-05, + "loss": 0.0129, + "step": 23429 + }, + { + "epoch": 0.5492360976810422, + "grad_norm": 0.19078627228736877, + "learning_rate": 8.462543965885705e-05, + "loss": 0.0364, + "step": 23430 + }, + { + "epoch": 0.5492595392558473, + "grad_norm": 0.3496353328227997, + "learning_rate": 8.46181611229408e-05, + "loss": 0.0856, + "step": 23431 + }, + { + "epoch": 0.5492829808306522, + "grad_norm": 0.4330250024795532, + "learning_rate": 8.461088267048682e-05, + "loss": 0.0708, + "step": 23432 + }, + { + "epoch": 0.5493064224054572, + "grad_norm": 0.3747493326663971, + "learning_rate": 8.460360430153466e-05, + "loss": 0.0774, + "step": 23433 + }, + { + "epoch": 0.5493298639802622, + "grad_norm": 0.2101515233516693, + "learning_rate": 8.45963260161238e-05, + "loss": 0.0534, + "step": 23434 + }, + { + "epoch": 0.5493533055550672, + "grad_norm": 0.31514328718185425, + "learning_rate": 8.458904781429366e-05, + "loss": 0.0635, + "step": 23435 + }, + { + "epoch": 0.5493767471298722, + "grad_norm": 0.6822181940078735, + "learning_rate": 8.45817696960838e-05, + "loss": 0.1198, + "step": 23436 + }, + { + "epoch": 0.5494001887046772, + "grad_norm": 0.3810606598854065, + "learning_rate": 8.45744916615337e-05, + "loss": 0.0605, + "step": 23437 + }, + { + "epoch": 0.5494236302794822, + "grad_norm": 0.5838484168052673, + "learning_rate": 8.456721371068282e-05, + "loss": 0.7414, + "step": 23438 + }, + { + "epoch": 0.5494470718542872, + "grad_norm": 0.26176151633262634, + "learning_rate": 8.45599358435707e-05, + "loss": 0.0741, + "step": 23439 + }, + { + "epoch": 0.5494705134290921, + "grad_norm": 0.3315688669681549, + "learning_rate": 8.45526580602368e-05, + "loss": 0.0516, + "step": 23440 + }, + { + "epoch": 0.5494939550038972, + "grad_norm": 0.31667569279670715, + "learning_rate": 8.454538036072058e-05, + "loss": 0.0586, + "step": 23441 + }, + { + "epoch": 0.5495173965787021, + "grad_norm": 0.8886712193489075, + "learning_rate": 8.45381027450616e-05, + "loss": 0.1397, + "step": 23442 + }, + { + "epoch": 0.5495408381535072, + "grad_norm": 0.14248040318489075, + "learning_rate": 8.453082521329927e-05, + "loss": 0.0155, + "step": 23443 + }, + { + "epoch": 0.5495642797283121, + "grad_norm": 0.41502222418785095, + "learning_rate": 8.452354776547312e-05, + "loss": 0.0663, + "step": 23444 + }, + { + "epoch": 0.5495877213031172, + "grad_norm": 0.2995275557041168, + "learning_rate": 8.451627040162264e-05, + "loss": 0.0485, + "step": 23445 + }, + { + "epoch": 0.5496111628779221, + "grad_norm": 0.3535245358943939, + "learning_rate": 8.450899312178733e-05, + "loss": 0.0585, + "step": 23446 + }, + { + "epoch": 0.5496346044527272, + "grad_norm": 0.3613906502723694, + "learning_rate": 8.450171592600664e-05, + "loss": 0.0452, + "step": 23447 + }, + { + "epoch": 0.5496580460275321, + "grad_norm": 0.7050965428352356, + "learning_rate": 8.449443881432011e-05, + "loss": 0.0916, + "step": 23448 + }, + { + "epoch": 0.5496814876023371, + "grad_norm": 0.6148957014083862, + "learning_rate": 8.448716178676716e-05, + "loss": 0.1042, + "step": 23449 + }, + { + "epoch": 0.5497049291771421, + "grad_norm": 0.38818874955177307, + "learning_rate": 8.447988484338731e-05, + "loss": 0.0996, + "step": 23450 + }, + { + "epoch": 0.5497283707519471, + "grad_norm": 0.36589357256889343, + "learning_rate": 8.447260798422006e-05, + "loss": 0.3639, + "step": 23451 + }, + { + "epoch": 0.5497518123267521, + "grad_norm": 0.6280344128608704, + "learning_rate": 8.446533120930487e-05, + "loss": 0.087, + "step": 23452 + }, + { + "epoch": 0.5497752539015571, + "grad_norm": 0.41700074076652527, + "learning_rate": 8.44580545186812e-05, + "loss": 0.0642, + "step": 23453 + }, + { + "epoch": 0.5497986954763621, + "grad_norm": 0.7949963212013245, + "learning_rate": 8.44507779123886e-05, + "loss": 0.1252, + "step": 23454 + }, + { + "epoch": 0.5498221370511671, + "grad_norm": 0.5503585934638977, + "learning_rate": 8.44435013904665e-05, + "loss": 0.1441, + "step": 23455 + }, + { + "epoch": 0.549845578625972, + "grad_norm": 0.4453699588775635, + "learning_rate": 8.44362249529544e-05, + "loss": 0.062, + "step": 23456 + }, + { + "epoch": 0.5498690202007771, + "grad_norm": 0.19686968624591827, + "learning_rate": 8.442894859989174e-05, + "loss": 0.0474, + "step": 23457 + }, + { + "epoch": 0.549892461775582, + "grad_norm": 0.33606991171836853, + "learning_rate": 8.44216723313181e-05, + "loss": 0.0527, + "step": 23458 + }, + { + "epoch": 0.5499159033503871, + "grad_norm": 0.42073917388916016, + "learning_rate": 8.441439614727289e-05, + "loss": 0.0452, + "step": 23459 + }, + { + "epoch": 0.549939344925192, + "grad_norm": 0.2514760196208954, + "learning_rate": 8.440712004779559e-05, + "loss": 0.0475, + "step": 23460 + }, + { + "epoch": 0.5499627864999971, + "grad_norm": 0.3039402961730957, + "learning_rate": 8.439984403292572e-05, + "loss": 0.0651, + "step": 23461 + }, + { + "epoch": 0.549986228074802, + "grad_norm": 0.1875697374343872, + "learning_rate": 8.439256810270273e-05, + "loss": 0.0418, + "step": 23462 + }, + { + "epoch": 0.5500096696496071, + "grad_norm": 0.3340395987033844, + "learning_rate": 8.438529225716609e-05, + "loss": 0.0437, + "step": 23463 + }, + { + "epoch": 0.550033111224412, + "grad_norm": 0.24004201591014862, + "learning_rate": 8.43780164963553e-05, + "loss": 0.0614, + "step": 23464 + }, + { + "epoch": 0.5500565527992171, + "grad_norm": 0.4305814504623413, + "learning_rate": 8.437074082030985e-05, + "loss": 0.0666, + "step": 23465 + }, + { + "epoch": 0.550079994374022, + "grad_norm": 0.5767718553543091, + "learning_rate": 8.436346522906915e-05, + "loss": 0.1006, + "step": 23466 + }, + { + "epoch": 0.550103435948827, + "grad_norm": 0.6676970720291138, + "learning_rate": 8.435618972267278e-05, + "loss": 0.1406, + "step": 23467 + }, + { + "epoch": 0.550126877523632, + "grad_norm": 0.6584795117378235, + "learning_rate": 8.434891430116016e-05, + "loss": 0.7255, + "step": 23468 + }, + { + "epoch": 0.550150319098437, + "grad_norm": 0.5882524251937866, + "learning_rate": 8.434163896457072e-05, + "loss": 0.1385, + "step": 23469 + }, + { + "epoch": 0.550173760673242, + "grad_norm": 0.11173036694526672, + "learning_rate": 8.4334363712944e-05, + "loss": 0.0256, + "step": 23470 + }, + { + "epoch": 0.550197202248047, + "grad_norm": 0.40063223242759705, + "learning_rate": 8.43270885463195e-05, + "loss": 0.0981, + "step": 23471 + }, + { + "epoch": 0.5502206438228521, + "grad_norm": 0.36428624391555786, + "learning_rate": 8.431981346473666e-05, + "loss": 0.0464, + "step": 23472 + }, + { + "epoch": 0.550244085397657, + "grad_norm": 0.5533102750778198, + "learning_rate": 8.43125384682349e-05, + "loss": 0.0623, + "step": 23473 + }, + { + "epoch": 0.5502675269724621, + "grad_norm": 0.5144340991973877, + "learning_rate": 8.430526355685379e-05, + "loss": 0.1162, + "step": 23474 + }, + { + "epoch": 0.550290968547267, + "grad_norm": 0.3675994575023651, + "learning_rate": 8.429798873063278e-05, + "loss": 0.084, + "step": 23475 + }, + { + "epoch": 0.550314410122072, + "grad_norm": 0.26927998661994934, + "learning_rate": 8.429071398961128e-05, + "loss": 0.0367, + "step": 23476 + }, + { + "epoch": 0.550337851696877, + "grad_norm": 0.7567132711410522, + "learning_rate": 8.428343933382884e-05, + "loss": 0.1695, + "step": 23477 + }, + { + "epoch": 0.550361293271682, + "grad_norm": 0.5593932867050171, + "learning_rate": 8.427616476332487e-05, + "loss": 0.055, + "step": 23478 + }, + { + "epoch": 0.550384734846487, + "grad_norm": 0.2341584861278534, + "learning_rate": 8.42688902781389e-05, + "loss": 0.0359, + "step": 23479 + }, + { + "epoch": 0.550408176421292, + "grad_norm": 0.08468177914619446, + "learning_rate": 8.426161587831037e-05, + "loss": 0.0217, + "step": 23480 + }, + { + "epoch": 0.550431617996097, + "grad_norm": 0.5323643088340759, + "learning_rate": 8.425434156387871e-05, + "loss": 0.1053, + "step": 23481 + }, + { + "epoch": 0.550455059570902, + "grad_norm": 0.41879764199256897, + "learning_rate": 8.424706733488347e-05, + "loss": 0.1047, + "step": 23482 + }, + { + "epoch": 0.550478501145707, + "grad_norm": 0.1398135870695114, + "learning_rate": 8.423979319136409e-05, + "loss": 0.0215, + "step": 23483 + }, + { + "epoch": 0.550501942720512, + "grad_norm": 0.07837189733982086, + "learning_rate": 8.423251913336004e-05, + "loss": 0.0223, + "step": 23484 + }, + { + "epoch": 0.5505253842953169, + "grad_norm": 1.1116114854812622, + "learning_rate": 8.422524516091077e-05, + "loss": 0.1297, + "step": 23485 + }, + { + "epoch": 0.550548825870122, + "grad_norm": 0.07795321196317673, + "learning_rate": 8.421797127405577e-05, + "loss": 0.0183, + "step": 23486 + }, + { + "epoch": 0.5505722674449269, + "grad_norm": 0.12114300578832626, + "learning_rate": 8.421069747283451e-05, + "loss": 0.027, + "step": 23487 + }, + { + "epoch": 0.550595709019732, + "grad_norm": 0.15286076068878174, + "learning_rate": 8.420342375728644e-05, + "loss": 0.0286, + "step": 23488 + }, + { + "epoch": 0.5506191505945369, + "grad_norm": 0.24757494032382965, + "learning_rate": 8.419615012745104e-05, + "loss": 0.0427, + "step": 23489 + }, + { + "epoch": 0.550642592169342, + "grad_norm": 0.3250032961368561, + "learning_rate": 8.41888765833678e-05, + "loss": 0.0703, + "step": 23490 + }, + { + "epoch": 0.5506660337441469, + "grad_norm": 0.4971979856491089, + "learning_rate": 8.41816031250761e-05, + "loss": 0.121, + "step": 23491 + }, + { + "epoch": 0.550689475318952, + "grad_norm": 0.5909038186073303, + "learning_rate": 8.417432975261552e-05, + "loss": 0.6301, + "step": 23492 + }, + { + "epoch": 0.5507129168937569, + "grad_norm": 0.39041435718536377, + "learning_rate": 8.416705646602541e-05, + "loss": 0.0645, + "step": 23493 + }, + { + "epoch": 0.550736358468562, + "grad_norm": 0.6450502872467041, + "learning_rate": 8.415978326534534e-05, + "loss": 0.1042, + "step": 23494 + }, + { + "epoch": 0.5507598000433669, + "grad_norm": 0.34000152349472046, + "learning_rate": 8.415251015061472e-05, + "loss": 0.0319, + "step": 23495 + }, + { + "epoch": 0.5507832416181719, + "grad_norm": 0.6102238297462463, + "learning_rate": 8.414523712187306e-05, + "loss": 0.5697, + "step": 23496 + }, + { + "epoch": 0.5508066831929769, + "grad_norm": 0.18297457695007324, + "learning_rate": 8.413796417915977e-05, + "loss": 0.0364, + "step": 23497 + }, + { + "epoch": 0.5508301247677819, + "grad_norm": 0.5487461686134338, + "learning_rate": 8.413069132251431e-05, + "loss": 0.061, + "step": 23498 + }, + { + "epoch": 0.5508535663425869, + "grad_norm": 1.0500478744506836, + "learning_rate": 8.412341855197618e-05, + "loss": 0.1164, + "step": 23499 + }, + { + "epoch": 0.5508770079173919, + "grad_norm": 0.11686645448207855, + "learning_rate": 8.411614586758485e-05, + "loss": 0.0183, + "step": 23500 + }, + { + "epoch": 0.5509004494921969, + "grad_norm": 0.5574228763580322, + "learning_rate": 8.410887326937972e-05, + "loss": 0.108, + "step": 23501 + }, + { + "epoch": 0.5509238910670019, + "grad_norm": 0.5789743065834045, + "learning_rate": 8.410160075740033e-05, + "loss": 0.1172, + "step": 23502 + }, + { + "epoch": 0.5509473326418068, + "grad_norm": 0.5069329142570496, + "learning_rate": 8.409432833168608e-05, + "loss": 0.1678, + "step": 23503 + }, + { + "epoch": 0.5509707742166119, + "grad_norm": 0.40872833132743835, + "learning_rate": 8.408705599227644e-05, + "loss": 0.0913, + "step": 23504 + }, + { + "epoch": 0.5509942157914168, + "grad_norm": 0.5158546566963196, + "learning_rate": 8.407978373921087e-05, + "loss": 0.1378, + "step": 23505 + }, + { + "epoch": 0.5510176573662219, + "grad_norm": 0.3247106671333313, + "learning_rate": 8.407251157252885e-05, + "loss": 0.069, + "step": 23506 + }, + { + "epoch": 0.5510410989410268, + "grad_norm": 0.1489814966917038, + "learning_rate": 8.406523949226986e-05, + "loss": 0.0246, + "step": 23507 + }, + { + "epoch": 0.5510645405158319, + "grad_norm": 0.3625519871711731, + "learning_rate": 8.40579674984733e-05, + "loss": 0.0304, + "step": 23508 + }, + { + "epoch": 0.5510879820906368, + "grad_norm": 0.17224512994289398, + "learning_rate": 8.405069559117869e-05, + "loss": 0.0257, + "step": 23509 + }, + { + "epoch": 0.5511114236654419, + "grad_norm": 0.31899648904800415, + "learning_rate": 8.404342377042543e-05, + "loss": 0.0821, + "step": 23510 + }, + { + "epoch": 0.5511348652402468, + "grad_norm": 0.6897502541542053, + "learning_rate": 8.403615203625299e-05, + "loss": 0.1335, + "step": 23511 + }, + { + "epoch": 0.5511583068150518, + "grad_norm": 0.6504631638526917, + "learning_rate": 8.402888038870087e-05, + "loss": 0.1249, + "step": 23512 + }, + { + "epoch": 0.5511817483898568, + "grad_norm": 0.20733892917633057, + "learning_rate": 8.402160882780848e-05, + "loss": 0.0414, + "step": 23513 + }, + { + "epoch": 0.5512051899646618, + "grad_norm": 0.568519115447998, + "learning_rate": 8.401433735361528e-05, + "loss": 0.1142, + "step": 23514 + }, + { + "epoch": 0.5512286315394668, + "grad_norm": 0.09541910141706467, + "learning_rate": 8.400706596616075e-05, + "loss": 0.0135, + "step": 23515 + }, + { + "epoch": 0.5512520731142718, + "grad_norm": 0.33329567313194275, + "learning_rate": 8.399979466548432e-05, + "loss": 0.0782, + "step": 23516 + }, + { + "epoch": 0.5512755146890768, + "grad_norm": 0.1383265107870102, + "learning_rate": 8.399252345162542e-05, + "loss": 0.0204, + "step": 23517 + }, + { + "epoch": 0.5512989562638818, + "grad_norm": 0.1693001538515091, + "learning_rate": 8.398525232462356e-05, + "loss": 0.0229, + "step": 23518 + }, + { + "epoch": 0.5513223978386868, + "grad_norm": 0.42046603560447693, + "learning_rate": 8.39779812845182e-05, + "loss": 0.0844, + "step": 23519 + }, + { + "epoch": 0.5513458394134918, + "grad_norm": 0.35810717940330505, + "learning_rate": 8.397071033134875e-05, + "loss": 0.0627, + "step": 23520 + }, + { + "epoch": 0.5513692809882967, + "grad_norm": 0.22913290560245514, + "learning_rate": 8.396343946515466e-05, + "loss": 0.0426, + "step": 23521 + }, + { + "epoch": 0.5513927225631018, + "grad_norm": 0.14325189590454102, + "learning_rate": 8.395616868597543e-05, + "loss": 0.0258, + "step": 23522 + }, + { + "epoch": 0.5514161641379068, + "grad_norm": 0.3975781798362732, + "learning_rate": 8.394889799385045e-05, + "loss": 0.0346, + "step": 23523 + }, + { + "epoch": 0.5514396057127118, + "grad_norm": 0.47021615505218506, + "learning_rate": 8.394162738881923e-05, + "loss": 0.0842, + "step": 23524 + }, + { + "epoch": 0.5514630472875168, + "grad_norm": 0.34675195813179016, + "learning_rate": 8.393435687092119e-05, + "loss": 0.0691, + "step": 23525 + }, + { + "epoch": 0.5514864888623218, + "grad_norm": 0.3319578766822815, + "learning_rate": 8.392708644019575e-05, + "loss": 0.0668, + "step": 23526 + }, + { + "epoch": 0.5515099304371268, + "grad_norm": 0.5633193254470825, + "learning_rate": 8.391981609668241e-05, + "loss": 0.0778, + "step": 23527 + }, + { + "epoch": 0.5515333720119318, + "grad_norm": 0.5801190733909607, + "learning_rate": 8.391254584042061e-05, + "loss": 0.0717, + "step": 23528 + }, + { + "epoch": 0.5515568135867368, + "grad_norm": 0.31737467646598816, + "learning_rate": 8.390527567144973e-05, + "loss": 0.0415, + "step": 23529 + }, + { + "epoch": 0.5515802551615417, + "grad_norm": 0.39348191022872925, + "learning_rate": 8.389800558980931e-05, + "loss": 0.0654, + "step": 23530 + }, + { + "epoch": 0.5516036967363468, + "grad_norm": 0.5269396305084229, + "learning_rate": 8.38907355955388e-05, + "loss": 0.1166, + "step": 23531 + }, + { + "epoch": 0.5516271383111517, + "grad_norm": 0.21223711967468262, + "learning_rate": 8.38834656886776e-05, + "loss": 0.0303, + "step": 23532 + }, + { + "epoch": 0.5516505798859568, + "grad_norm": 0.9526686072349548, + "learning_rate": 8.387619586926514e-05, + "loss": 0.1826, + "step": 23533 + }, + { + "epoch": 0.5516740214607617, + "grad_norm": 0.1473158746957779, + "learning_rate": 8.386892613734092e-05, + "loss": 0.0199, + "step": 23534 + }, + { + "epoch": 0.5516974630355668, + "grad_norm": 0.18389759957790375, + "learning_rate": 8.386165649294436e-05, + "loss": 0.0274, + "step": 23535 + }, + { + "epoch": 0.5517209046103717, + "grad_norm": 0.6354973912239075, + "learning_rate": 8.385438693611488e-05, + "loss": 0.0961, + "step": 23536 + }, + { + "epoch": 0.5517443461851768, + "grad_norm": 0.6681496500968933, + "learning_rate": 8.384711746689198e-05, + "loss": 0.6023, + "step": 23537 + }, + { + "epoch": 0.5517677877599817, + "grad_norm": 0.12061692029237747, + "learning_rate": 8.383984808531508e-05, + "loss": 0.0166, + "step": 23538 + }, + { + "epoch": 0.5517912293347867, + "grad_norm": 0.713694155216217, + "learning_rate": 8.383257879142358e-05, + "loss": 0.2112, + "step": 23539 + }, + { + "epoch": 0.5518146709095917, + "grad_norm": 0.27209052443504333, + "learning_rate": 8.3825309585257e-05, + "loss": 0.037, + "step": 23540 + }, + { + "epoch": 0.5518381124843967, + "grad_norm": 0.2340465784072876, + "learning_rate": 8.381804046685474e-05, + "loss": 0.0452, + "step": 23541 + }, + { + "epoch": 0.5518615540592017, + "grad_norm": 0.39961129426956177, + "learning_rate": 8.38107714362562e-05, + "loss": 0.091, + "step": 23542 + }, + { + "epoch": 0.5518849956340067, + "grad_norm": 0.16212770342826843, + "learning_rate": 8.380350249350088e-05, + "loss": 0.0351, + "step": 23543 + }, + { + "epoch": 0.5519084372088117, + "grad_norm": 1.0608056783676147, + "learning_rate": 8.379623363862826e-05, + "loss": 0.1293, + "step": 23544 + }, + { + "epoch": 0.5519318787836167, + "grad_norm": 0.5743598341941833, + "learning_rate": 8.37889648716777e-05, + "loss": 0.1303, + "step": 23545 + }, + { + "epoch": 0.5519553203584217, + "grad_norm": 0.08116581290960312, + "learning_rate": 8.378169619268868e-05, + "loss": 0.011, + "step": 23546 + }, + { + "epoch": 0.5519787619332267, + "grad_norm": 0.7048518657684326, + "learning_rate": 8.377442760170064e-05, + "loss": 0.1141, + "step": 23547 + }, + { + "epoch": 0.5520022035080316, + "grad_norm": 0.663253903388977, + "learning_rate": 8.376715909875301e-05, + "loss": 0.2076, + "step": 23548 + }, + { + "epoch": 0.5520256450828367, + "grad_norm": 0.5802831053733826, + "learning_rate": 8.375989068388522e-05, + "loss": 0.1353, + "step": 23549 + }, + { + "epoch": 0.5520490866576416, + "grad_norm": 0.23118634521961212, + "learning_rate": 8.375262235713674e-05, + "loss": 0.0134, + "step": 23550 + }, + { + "epoch": 0.5520725282324467, + "grad_norm": 0.28919634222984314, + "learning_rate": 8.374535411854698e-05, + "loss": 0.059, + "step": 23551 + }, + { + "epoch": 0.5520959698072516, + "grad_norm": 0.4065954387187958, + "learning_rate": 8.373808596815535e-05, + "loss": 0.0867, + "step": 23552 + }, + { + "epoch": 0.5521194113820567, + "grad_norm": 0.35497328639030457, + "learning_rate": 8.373081790600136e-05, + "loss": 0.0414, + "step": 23553 + }, + { + "epoch": 0.5521428529568616, + "grad_norm": 0.7463645935058594, + "learning_rate": 8.372354993212437e-05, + "loss": 0.0532, + "step": 23554 + }, + { + "epoch": 0.5521662945316667, + "grad_norm": 0.46781566739082336, + "learning_rate": 8.371628204656392e-05, + "loss": 0.4516, + "step": 23555 + }, + { + "epoch": 0.5521897361064716, + "grad_norm": 0.5793401598930359, + "learning_rate": 8.370901424935934e-05, + "loss": 0.4995, + "step": 23556 + }, + { + "epoch": 0.5522131776812766, + "grad_norm": 0.5521657466888428, + "learning_rate": 8.370174654055011e-05, + "loss": 0.0818, + "step": 23557 + }, + { + "epoch": 0.5522366192560816, + "grad_norm": 0.39814993739128113, + "learning_rate": 8.369447892017568e-05, + "loss": 0.0213, + "step": 23558 + }, + { + "epoch": 0.5522600608308866, + "grad_norm": 0.1681576371192932, + "learning_rate": 8.368721138827543e-05, + "loss": 0.034, + "step": 23559 + }, + { + "epoch": 0.5522835024056916, + "grad_norm": 0.6463558077812195, + "learning_rate": 8.367994394488887e-05, + "loss": 0.133, + "step": 23560 + }, + { + "epoch": 0.5523069439804966, + "grad_norm": 0.5411694645881653, + "learning_rate": 8.367267659005535e-05, + "loss": 0.1393, + "step": 23561 + }, + { + "epoch": 0.5523303855553016, + "grad_norm": 0.3013487756252289, + "learning_rate": 8.366540932381439e-05, + "loss": 0.0386, + "step": 23562 + }, + { + "epoch": 0.5523538271301066, + "grad_norm": 0.6673893928527832, + "learning_rate": 8.365814214620537e-05, + "loss": 0.0744, + "step": 23563 + }, + { + "epoch": 0.5523772687049116, + "grad_norm": 0.4209643602371216, + "learning_rate": 8.365087505726771e-05, + "loss": 0.0911, + "step": 23564 + }, + { + "epoch": 0.5524007102797166, + "grad_norm": 0.15284545719623566, + "learning_rate": 8.364360805704087e-05, + "loss": 0.0356, + "step": 23565 + }, + { + "epoch": 0.5524241518545215, + "grad_norm": 0.5291098356246948, + "learning_rate": 8.363634114556424e-05, + "loss": 0.0902, + "step": 23566 + }, + { + "epoch": 0.5524475934293266, + "grad_norm": 0.43381667137145996, + "learning_rate": 8.362907432287733e-05, + "loss": 0.0662, + "step": 23567 + }, + { + "epoch": 0.5524710350041315, + "grad_norm": 0.1821274757385254, + "learning_rate": 8.362180758901951e-05, + "loss": 0.0315, + "step": 23568 + }, + { + "epoch": 0.5524944765789366, + "grad_norm": 0.8191523551940918, + "learning_rate": 8.361454094403023e-05, + "loss": 0.1202, + "step": 23569 + }, + { + "epoch": 0.5525179181537415, + "grad_norm": 0.3201591670513153, + "learning_rate": 8.360727438794891e-05, + "loss": 0.0414, + "step": 23570 + }, + { + "epoch": 0.5525413597285466, + "grad_norm": 0.3467867374420166, + "learning_rate": 8.360000792081496e-05, + "loss": 0.0366, + "step": 23571 + }, + { + "epoch": 0.5525648013033515, + "grad_norm": 0.38738617300987244, + "learning_rate": 8.359274154266786e-05, + "loss": 0.0794, + "step": 23572 + }, + { + "epoch": 0.5525882428781566, + "grad_norm": 0.3991090953350067, + "learning_rate": 8.358547525354699e-05, + "loss": 0.0833, + "step": 23573 + }, + { + "epoch": 0.5526116844529616, + "grad_norm": 0.2067466825246811, + "learning_rate": 8.357820905349178e-05, + "loss": 0.0182, + "step": 23574 + }, + { + "epoch": 0.5526351260277665, + "grad_norm": 0.22770707309246063, + "learning_rate": 8.357094294254169e-05, + "loss": 0.0229, + "step": 23575 + }, + { + "epoch": 0.5526585676025716, + "grad_norm": 0.3710660934448242, + "learning_rate": 8.356367692073613e-05, + "loss": 0.0608, + "step": 23576 + }, + { + "epoch": 0.5526820091773765, + "grad_norm": 0.3333088755607605, + "learning_rate": 8.355641098811448e-05, + "loss": 0.0457, + "step": 23577 + }, + { + "epoch": 0.5527054507521816, + "grad_norm": 0.45754751563072205, + "learning_rate": 8.354914514471622e-05, + "loss": 0.0509, + "step": 23578 + }, + { + "epoch": 0.5527288923269865, + "grad_norm": 0.6588448286056519, + "learning_rate": 8.354187939058077e-05, + "loss": 0.1114, + "step": 23579 + }, + { + "epoch": 0.5527523339017916, + "grad_norm": 0.9885949492454529, + "learning_rate": 8.353461372574755e-05, + "loss": 0.7987, + "step": 23580 + }, + { + "epoch": 0.5527757754765965, + "grad_norm": 0.46627676486968994, + "learning_rate": 8.352734815025598e-05, + "loss": 0.1155, + "step": 23581 + }, + { + "epoch": 0.5527992170514016, + "grad_norm": 0.3102782666683197, + "learning_rate": 8.35200826641455e-05, + "loss": 0.0279, + "step": 23582 + }, + { + "epoch": 0.5528226586262065, + "grad_norm": 0.4228268265724182, + "learning_rate": 8.35128172674555e-05, + "loss": 0.0664, + "step": 23583 + }, + { + "epoch": 0.5528461002010115, + "grad_norm": 0.1747932881116867, + "learning_rate": 8.350555196022539e-05, + "loss": 0.0268, + "step": 23584 + }, + { + "epoch": 0.5528695417758165, + "grad_norm": 0.2526592016220093, + "learning_rate": 8.349828674249466e-05, + "loss": 0.0318, + "step": 23585 + }, + { + "epoch": 0.5528929833506215, + "grad_norm": 0.33717718720436096, + "learning_rate": 8.349102161430268e-05, + "loss": 0.0701, + "step": 23586 + }, + { + "epoch": 0.5529164249254265, + "grad_norm": 0.7636315226554871, + "learning_rate": 8.348375657568886e-05, + "loss": 0.1444, + "step": 23587 + }, + { + "epoch": 0.5529398665002315, + "grad_norm": 0.5954223275184631, + "learning_rate": 8.347649162669267e-05, + "loss": 0.0711, + "step": 23588 + }, + { + "epoch": 0.5529633080750365, + "grad_norm": 0.3974081575870514, + "learning_rate": 8.34692267673535e-05, + "loss": 0.0939, + "step": 23589 + }, + { + "epoch": 0.5529867496498415, + "grad_norm": 0.1322023719549179, + "learning_rate": 8.34619619977107e-05, + "loss": 0.0193, + "step": 23590 + }, + { + "epoch": 0.5530101912246465, + "grad_norm": 0.4459347426891327, + "learning_rate": 8.34546973178038e-05, + "loss": 0.0523, + "step": 23591 + }, + { + "epoch": 0.5530336327994515, + "grad_norm": 0.6118428707122803, + "learning_rate": 8.344743272767222e-05, + "loss": 0.5695, + "step": 23592 + }, + { + "epoch": 0.5530570743742564, + "grad_norm": 0.3478293716907501, + "learning_rate": 8.34401682273553e-05, + "loss": 0.0328, + "step": 23593 + }, + { + "epoch": 0.5530805159490615, + "grad_norm": 0.6486023664474487, + "learning_rate": 8.34329038168925e-05, + "loss": 0.1333, + "step": 23594 + }, + { + "epoch": 0.5531039575238664, + "grad_norm": 0.45484647154808044, + "learning_rate": 8.342563949632324e-05, + "loss": 0.0829, + "step": 23595 + }, + { + "epoch": 0.5531273990986715, + "grad_norm": 0.10621140897274017, + "learning_rate": 8.34183752656869e-05, + "loss": 0.0203, + "step": 23596 + }, + { + "epoch": 0.5531508406734764, + "grad_norm": 0.3984803855419159, + "learning_rate": 8.341111112502294e-05, + "loss": 0.0545, + "step": 23597 + }, + { + "epoch": 0.5531742822482815, + "grad_norm": 0.49297499656677246, + "learning_rate": 8.340384707437077e-05, + "loss": 0.0642, + "step": 23598 + }, + { + "epoch": 0.5531977238230864, + "grad_norm": 0.20379355549812317, + "learning_rate": 8.339658311376977e-05, + "loss": 0.036, + "step": 23599 + }, + { + "epoch": 0.5532211653978915, + "grad_norm": 0.39427220821380615, + "learning_rate": 8.338931924325939e-05, + "loss": 0.0653, + "step": 23600 + }, + { + "epoch": 0.5532446069726964, + "grad_norm": 0.36630040407180786, + "learning_rate": 8.338205546287904e-05, + "loss": 0.0668, + "step": 23601 + }, + { + "epoch": 0.5532680485475014, + "grad_norm": 0.3559831380844116, + "learning_rate": 8.337479177266806e-05, + "loss": 0.0946, + "step": 23602 + }, + { + "epoch": 0.5532914901223064, + "grad_norm": 0.8623394966125488, + "learning_rate": 8.336752817266598e-05, + "loss": 0.2353, + "step": 23603 + }, + { + "epoch": 0.5533149316971114, + "grad_norm": 0.7302821278572083, + "learning_rate": 8.336026466291215e-05, + "loss": 0.0749, + "step": 23604 + }, + { + "epoch": 0.5533383732719164, + "grad_norm": 0.39584776759147644, + "learning_rate": 8.335300124344603e-05, + "loss": 0.0358, + "step": 23605 + }, + { + "epoch": 0.5533618148467214, + "grad_norm": 0.7774975299835205, + "learning_rate": 8.334573791430694e-05, + "loss": 0.1957, + "step": 23606 + }, + { + "epoch": 0.5533852564215264, + "grad_norm": 0.5300292372703552, + "learning_rate": 8.333847467553437e-05, + "loss": 0.0965, + "step": 23607 + }, + { + "epoch": 0.5534086979963314, + "grad_norm": 0.5827982425689697, + "learning_rate": 8.333121152716772e-05, + "loss": 0.5452, + "step": 23608 + }, + { + "epoch": 0.5534321395711364, + "grad_norm": 0.3878721296787262, + "learning_rate": 8.332394846924635e-05, + "loss": 0.0559, + "step": 23609 + }, + { + "epoch": 0.5534555811459414, + "grad_norm": 0.1815311163663864, + "learning_rate": 8.331668550180974e-05, + "loss": 0.0346, + "step": 23610 + }, + { + "epoch": 0.5534790227207463, + "grad_norm": 0.19123691320419312, + "learning_rate": 8.330942262489725e-05, + "loss": 0.0383, + "step": 23611 + }, + { + "epoch": 0.5535024642955514, + "grad_norm": 0.3886093497276306, + "learning_rate": 8.330215983854828e-05, + "loss": 0.0529, + "step": 23612 + }, + { + "epoch": 0.5535259058703563, + "grad_norm": 0.15658719837665558, + "learning_rate": 8.329489714280229e-05, + "loss": 0.0116, + "step": 23613 + }, + { + "epoch": 0.5535493474451614, + "grad_norm": 0.5869392156600952, + "learning_rate": 8.328763453769866e-05, + "loss": 0.6239, + "step": 23614 + }, + { + "epoch": 0.5535727890199663, + "grad_norm": 0.11861753463745117, + "learning_rate": 8.328037202327674e-05, + "loss": 0.0256, + "step": 23615 + }, + { + "epoch": 0.5535962305947714, + "grad_norm": 0.407866895198822, + "learning_rate": 8.327310959957601e-05, + "loss": 0.022, + "step": 23616 + }, + { + "epoch": 0.5536196721695763, + "grad_norm": 0.7161434888839722, + "learning_rate": 8.326584726663589e-05, + "loss": 0.07, + "step": 23617 + }, + { + "epoch": 0.5536431137443814, + "grad_norm": 0.6911687254905701, + "learning_rate": 8.325858502449575e-05, + "loss": 0.1154, + "step": 23618 + }, + { + "epoch": 0.5536665553191863, + "grad_norm": 0.6658154726028442, + "learning_rate": 8.325132287319498e-05, + "loss": 0.0485, + "step": 23619 + }, + { + "epoch": 0.5536899968939913, + "grad_norm": 0.4113248884677887, + "learning_rate": 8.324406081277303e-05, + "loss": 0.369, + "step": 23620 + }, + { + "epoch": 0.5537134384687963, + "grad_norm": 0.34624290466308594, + "learning_rate": 8.323679884326927e-05, + "loss": 0.0405, + "step": 23621 + }, + { + "epoch": 0.5537368800436013, + "grad_norm": 0.760262668132782, + "learning_rate": 8.322953696472309e-05, + "loss": 0.5934, + "step": 23622 + }, + { + "epoch": 0.5537603216184063, + "grad_norm": 0.4880139231681824, + "learning_rate": 8.322227517717393e-05, + "loss": 0.0684, + "step": 23623 + }, + { + "epoch": 0.5537837631932113, + "grad_norm": 0.0703945979475975, + "learning_rate": 8.321501348066119e-05, + "loss": 0.0156, + "step": 23624 + }, + { + "epoch": 0.5538072047680164, + "grad_norm": 0.6009061336517334, + "learning_rate": 8.320775187522422e-05, + "loss": 0.1718, + "step": 23625 + }, + { + "epoch": 0.5538306463428213, + "grad_norm": 0.17913205921649933, + "learning_rate": 8.32004903609025e-05, + "loss": 0.033, + "step": 23626 + }, + { + "epoch": 0.5538540879176264, + "grad_norm": 0.28062430024147034, + "learning_rate": 8.319322893773533e-05, + "loss": 0.0369, + "step": 23627 + }, + { + "epoch": 0.5538775294924313, + "grad_norm": 0.6617304086685181, + "learning_rate": 8.318596760576222e-05, + "loss": 0.1434, + "step": 23628 + }, + { + "epoch": 0.5539009710672363, + "grad_norm": 0.19910676777362823, + "learning_rate": 8.317870636502251e-05, + "loss": 0.0356, + "step": 23629 + }, + { + "epoch": 0.5539244126420413, + "grad_norm": 0.4026508331298828, + "learning_rate": 8.317144521555563e-05, + "loss": 0.0472, + "step": 23630 + }, + { + "epoch": 0.5539478542168463, + "grad_norm": 0.4541773498058319, + "learning_rate": 8.316418415740096e-05, + "loss": 0.0768, + "step": 23631 + }, + { + "epoch": 0.5539712957916513, + "grad_norm": 0.3781459629535675, + "learning_rate": 8.315692319059788e-05, + "loss": 0.0354, + "step": 23632 + }, + { + "epoch": 0.5539947373664563, + "grad_norm": 0.32001587748527527, + "learning_rate": 8.314966231518582e-05, + "loss": 0.0389, + "step": 23633 + }, + { + "epoch": 0.5540181789412613, + "grad_norm": 0.31887954473495483, + "learning_rate": 8.314240153120416e-05, + "loss": 0.0509, + "step": 23634 + }, + { + "epoch": 0.5540416205160663, + "grad_norm": 0.5739172697067261, + "learning_rate": 8.313514083869232e-05, + "loss": 0.0921, + "step": 23635 + }, + { + "epoch": 0.5540650620908713, + "grad_norm": 0.5223550796508789, + "learning_rate": 8.312788023768968e-05, + "loss": 0.1551, + "step": 23636 + }, + { + "epoch": 0.5540885036656763, + "grad_norm": 0.4379429221153259, + "learning_rate": 8.312061972823561e-05, + "loss": 0.1281, + "step": 23637 + }, + { + "epoch": 0.5541119452404812, + "grad_norm": 0.35361990332603455, + "learning_rate": 8.311335931036956e-05, + "loss": 0.048, + "step": 23638 + }, + { + "epoch": 0.5541353868152863, + "grad_norm": 0.12176831066608429, + "learning_rate": 8.310609898413086e-05, + "loss": 0.0202, + "step": 23639 + }, + { + "epoch": 0.5541588283900912, + "grad_norm": 0.16502641141414642, + "learning_rate": 8.309883874955897e-05, + "loss": 0.0249, + "step": 23640 + }, + { + "epoch": 0.5541822699648963, + "grad_norm": 0.5600295662879944, + "learning_rate": 8.309157860669323e-05, + "loss": 0.1354, + "step": 23641 + }, + { + "epoch": 0.5542057115397012, + "grad_norm": 0.5176275372505188, + "learning_rate": 8.308431855557311e-05, + "loss": 0.0812, + "step": 23642 + }, + { + "epoch": 0.5542291531145063, + "grad_norm": 0.8056997656822205, + "learning_rate": 8.307705859623792e-05, + "loss": 0.0859, + "step": 23643 + }, + { + "epoch": 0.5542525946893112, + "grad_norm": 0.46304458379745483, + "learning_rate": 8.306979872872709e-05, + "loss": 0.0863, + "step": 23644 + }, + { + "epoch": 0.5542760362641163, + "grad_norm": 0.6022554636001587, + "learning_rate": 8.306253895308001e-05, + "loss": 0.0393, + "step": 23645 + }, + { + "epoch": 0.5542994778389212, + "grad_norm": 0.5034669041633606, + "learning_rate": 8.305527926933609e-05, + "loss": 0.0765, + "step": 23646 + }, + { + "epoch": 0.5543229194137262, + "grad_norm": 0.6385390758514404, + "learning_rate": 8.304801967753466e-05, + "loss": 0.1021, + "step": 23647 + }, + { + "epoch": 0.5543463609885312, + "grad_norm": 0.7535375952720642, + "learning_rate": 8.304076017771519e-05, + "loss": 0.1072, + "step": 23648 + }, + { + "epoch": 0.5543698025633362, + "grad_norm": 0.5141211152076721, + "learning_rate": 8.303350076991702e-05, + "loss": 0.1038, + "step": 23649 + }, + { + "epoch": 0.5543932441381412, + "grad_norm": 0.268254816532135, + "learning_rate": 8.302624145417953e-05, + "loss": 0.2293, + "step": 23650 + }, + { + "epoch": 0.5544166857129462, + "grad_norm": 0.41738295555114746, + "learning_rate": 8.301898223054212e-05, + "loss": 0.1063, + "step": 23651 + }, + { + "epoch": 0.5544401272877512, + "grad_norm": 0.6438717842102051, + "learning_rate": 8.301172309904422e-05, + "loss": 0.2202, + "step": 23652 + }, + { + "epoch": 0.5544635688625562, + "grad_norm": 0.49576446413993835, + "learning_rate": 8.30044640597252e-05, + "loss": 0.0761, + "step": 23653 + }, + { + "epoch": 0.5544870104373611, + "grad_norm": 0.10322816669940948, + "learning_rate": 8.299720511262441e-05, + "loss": 0.017, + "step": 23654 + }, + { + "epoch": 0.5545104520121662, + "grad_norm": 0.0873960480093956, + "learning_rate": 8.298994625778128e-05, + "loss": 0.009, + "step": 23655 + }, + { + "epoch": 0.5545338935869711, + "grad_norm": 0.6619318127632141, + "learning_rate": 8.298268749523519e-05, + "loss": 0.1485, + "step": 23656 + }, + { + "epoch": 0.5545573351617762, + "grad_norm": 0.5055177211761475, + "learning_rate": 8.297542882502548e-05, + "loss": 0.1017, + "step": 23657 + }, + { + "epoch": 0.5545807767365811, + "grad_norm": 0.22201284766197205, + "learning_rate": 8.296817024719161e-05, + "loss": 0.0453, + "step": 23658 + }, + { + "epoch": 0.5546042183113862, + "grad_norm": 0.38732174038887024, + "learning_rate": 8.296091176177293e-05, + "loss": 0.0806, + "step": 23659 + }, + { + "epoch": 0.5546276598861911, + "grad_norm": 0.13560374081134796, + "learning_rate": 8.295365336880878e-05, + "loss": 0.0352, + "step": 23660 + }, + { + "epoch": 0.5546511014609962, + "grad_norm": 0.44569075107574463, + "learning_rate": 8.294639506833861e-05, + "loss": 0.6073, + "step": 23661 + }, + { + "epoch": 0.5546745430358011, + "grad_norm": 0.5350340008735657, + "learning_rate": 8.29391368604018e-05, + "loss": 0.0545, + "step": 23662 + }, + { + "epoch": 0.5546979846106062, + "grad_norm": 0.3071407079696655, + "learning_rate": 8.293187874503766e-05, + "loss": 0.0552, + "step": 23663 + }, + { + "epoch": 0.5547214261854111, + "grad_norm": 0.21959978342056274, + "learning_rate": 8.292462072228563e-05, + "loss": 0.0216, + "step": 23664 + }, + { + "epoch": 0.5547448677602161, + "grad_norm": 0.17620068788528442, + "learning_rate": 8.291736279218512e-05, + "loss": 0.0373, + "step": 23665 + }, + { + "epoch": 0.5547683093350211, + "grad_norm": 0.3054983913898468, + "learning_rate": 8.291010495477549e-05, + "loss": 0.073, + "step": 23666 + }, + { + "epoch": 0.5547917509098261, + "grad_norm": 0.46806207299232483, + "learning_rate": 8.290284721009609e-05, + "loss": 0.0828, + "step": 23667 + }, + { + "epoch": 0.5548151924846311, + "grad_norm": 0.08261807262897491, + "learning_rate": 8.289558955818633e-05, + "loss": 0.0106, + "step": 23668 + }, + { + "epoch": 0.5548386340594361, + "grad_norm": 0.1529010534286499, + "learning_rate": 8.28883319990856e-05, + "loss": 0.0288, + "step": 23669 + }, + { + "epoch": 0.5548620756342411, + "grad_norm": 0.5332080721855164, + "learning_rate": 8.288107453283323e-05, + "loss": 0.0806, + "step": 23670 + }, + { + "epoch": 0.5548855172090461, + "grad_norm": 0.5302335619926453, + "learning_rate": 8.287381715946864e-05, + "loss": 0.1375, + "step": 23671 + }, + { + "epoch": 0.554908958783851, + "grad_norm": 0.5072125792503357, + "learning_rate": 8.286655987903119e-05, + "loss": 0.0613, + "step": 23672 + }, + { + "epoch": 0.5549324003586561, + "grad_norm": 0.5043341517448425, + "learning_rate": 8.28593026915603e-05, + "loss": 0.1146, + "step": 23673 + }, + { + "epoch": 0.554955841933461, + "grad_norm": 0.3664957284927368, + "learning_rate": 8.28520455970953e-05, + "loss": 0.0715, + "step": 23674 + }, + { + "epoch": 0.5549792835082661, + "grad_norm": 0.5189961194992065, + "learning_rate": 8.284478859567553e-05, + "loss": 0.1244, + "step": 23675 + }, + { + "epoch": 0.5550027250830711, + "grad_norm": 0.23539048433303833, + "learning_rate": 8.283753168734048e-05, + "loss": 0.0361, + "step": 23676 + }, + { + "epoch": 0.5550261666578761, + "grad_norm": 0.4983772337436676, + "learning_rate": 8.283027487212944e-05, + "loss": 0.0531, + "step": 23677 + }, + { + "epoch": 0.5550496082326811, + "grad_norm": 0.15919341146945953, + "learning_rate": 8.282301815008185e-05, + "loss": 0.0218, + "step": 23678 + }, + { + "epoch": 0.5550730498074861, + "grad_norm": 0.5864765048027039, + "learning_rate": 8.2815761521237e-05, + "loss": 0.1251, + "step": 23679 + }, + { + "epoch": 0.5550964913822911, + "grad_norm": 0.42226213216781616, + "learning_rate": 8.280850498563433e-05, + "loss": 0.1032, + "step": 23680 + }, + { + "epoch": 0.555119932957096, + "grad_norm": 0.2843101918697357, + "learning_rate": 8.280124854331322e-05, + "loss": 0.0355, + "step": 23681 + }, + { + "epoch": 0.5551433745319011, + "grad_norm": 0.7121860384941101, + "learning_rate": 8.279399219431298e-05, + "loss": 0.0813, + "step": 23682 + }, + { + "epoch": 0.555166816106706, + "grad_norm": 0.7423999905586243, + "learning_rate": 8.278673593867304e-05, + "loss": 0.1024, + "step": 23683 + }, + { + "epoch": 0.5551902576815111, + "grad_norm": 0.29477980732917786, + "learning_rate": 8.277947977643276e-05, + "loss": 0.0353, + "step": 23684 + }, + { + "epoch": 0.555213699256316, + "grad_norm": 0.47454747557640076, + "learning_rate": 8.27722237076315e-05, + "loss": 0.0956, + "step": 23685 + }, + { + "epoch": 0.5552371408311211, + "grad_norm": 0.5297451019287109, + "learning_rate": 8.276496773230864e-05, + "loss": 0.0726, + "step": 23686 + }, + { + "epoch": 0.555260582405926, + "grad_norm": 0.21967138350009918, + "learning_rate": 8.275771185050356e-05, + "loss": 0.0407, + "step": 23687 + }, + { + "epoch": 0.5552840239807311, + "grad_norm": 0.3028472065925598, + "learning_rate": 8.275045606225558e-05, + "loss": 0.0181, + "step": 23688 + }, + { + "epoch": 0.555307465555536, + "grad_norm": 0.4124543368816376, + "learning_rate": 8.274320036760412e-05, + "loss": 0.0548, + "step": 23689 + }, + { + "epoch": 0.5553309071303411, + "grad_norm": 0.4683479368686676, + "learning_rate": 8.273594476658857e-05, + "loss": 0.0816, + "step": 23690 + }, + { + "epoch": 0.555354348705146, + "grad_norm": 0.7272621989250183, + "learning_rate": 8.272868925924827e-05, + "loss": 0.1566, + "step": 23691 + }, + { + "epoch": 0.555377790279951, + "grad_norm": 0.4471890330314636, + "learning_rate": 8.272143384562256e-05, + "loss": 0.4889, + "step": 23692 + }, + { + "epoch": 0.555401231854756, + "grad_norm": 0.5741168856620789, + "learning_rate": 8.271417852575086e-05, + "loss": 0.0864, + "step": 23693 + }, + { + "epoch": 0.555424673429561, + "grad_norm": 0.8622278571128845, + "learning_rate": 8.270692329967252e-05, + "loss": 0.4259, + "step": 23694 + }, + { + "epoch": 0.555448115004366, + "grad_norm": 0.44204291701316833, + "learning_rate": 8.269966816742687e-05, + "loss": 0.0886, + "step": 23695 + }, + { + "epoch": 0.555471556579171, + "grad_norm": 0.21999983489513397, + "learning_rate": 8.269241312905333e-05, + "loss": 0.0443, + "step": 23696 + }, + { + "epoch": 0.555494998153976, + "grad_norm": 0.373271644115448, + "learning_rate": 8.268515818459125e-05, + "loss": 0.0814, + "step": 23697 + }, + { + "epoch": 0.555518439728781, + "grad_norm": 0.4301337003707886, + "learning_rate": 8.267790333407996e-05, + "loss": 0.1171, + "step": 23698 + }, + { + "epoch": 0.555541881303586, + "grad_norm": 0.42646342515945435, + "learning_rate": 8.267064857755888e-05, + "loss": 0.0535, + "step": 23699 + }, + { + "epoch": 0.555565322878391, + "grad_norm": 0.593396782875061, + "learning_rate": 8.26633939150673e-05, + "loss": 0.1642, + "step": 23700 + }, + { + "epoch": 0.5555887644531959, + "grad_norm": 0.21917849779129028, + "learning_rate": 8.265613934664467e-05, + "loss": 0.0234, + "step": 23701 + }, + { + "epoch": 0.555612206028001, + "grad_norm": 0.5695092678070068, + "learning_rate": 8.264888487233032e-05, + "loss": 0.6436, + "step": 23702 + }, + { + "epoch": 0.5556356476028059, + "grad_norm": 0.575880765914917, + "learning_rate": 8.26416304921636e-05, + "loss": 0.0794, + "step": 23703 + }, + { + "epoch": 0.555659089177611, + "grad_norm": 0.5990747213363647, + "learning_rate": 8.263437620618389e-05, + "loss": 0.1354, + "step": 23704 + }, + { + "epoch": 0.5556825307524159, + "grad_norm": 0.6959443092346191, + "learning_rate": 8.262712201443052e-05, + "loss": 0.0557, + "step": 23705 + }, + { + "epoch": 0.555705972327221, + "grad_norm": 0.650028645992279, + "learning_rate": 8.261986791694291e-05, + "loss": 0.5555, + "step": 23706 + }, + { + "epoch": 0.5557294139020259, + "grad_norm": 0.42762213945388794, + "learning_rate": 8.261261391376038e-05, + "loss": 0.0486, + "step": 23707 + }, + { + "epoch": 0.555752855476831, + "grad_norm": 0.39093151688575745, + "learning_rate": 8.260536000492226e-05, + "loss": 0.0212, + "step": 23708 + }, + { + "epoch": 0.5557762970516359, + "grad_norm": 0.37150442600250244, + "learning_rate": 8.259810619046797e-05, + "loss": 0.0833, + "step": 23709 + }, + { + "epoch": 0.5557997386264409, + "grad_norm": 0.44949787855148315, + "learning_rate": 8.259085247043683e-05, + "loss": 0.0581, + "step": 23710 + }, + { + "epoch": 0.5558231802012459, + "grad_norm": 0.08506599813699722, + "learning_rate": 8.258359884486824e-05, + "loss": 0.0115, + "step": 23711 + }, + { + "epoch": 0.5558466217760509, + "grad_norm": 0.40541866421699524, + "learning_rate": 8.257634531380148e-05, + "loss": 0.0718, + "step": 23712 + }, + { + "epoch": 0.5558700633508559, + "grad_norm": 0.5424754023551941, + "learning_rate": 8.256909187727601e-05, + "loss": 0.1046, + "step": 23713 + }, + { + "epoch": 0.5558935049256609, + "grad_norm": 0.3583996295928955, + "learning_rate": 8.256183853533112e-05, + "loss": 0.096, + "step": 23714 + }, + { + "epoch": 0.5559169465004659, + "grad_norm": 0.9252868294715881, + "learning_rate": 8.25545852880062e-05, + "loss": 0.1294, + "step": 23715 + }, + { + "epoch": 0.5559403880752709, + "grad_norm": 0.2426014542579651, + "learning_rate": 8.254733213534059e-05, + "loss": 0.044, + "step": 23716 + }, + { + "epoch": 0.5559638296500758, + "grad_norm": 0.3892747759819031, + "learning_rate": 8.254007907737362e-05, + "loss": 0.0547, + "step": 23717 + }, + { + "epoch": 0.5559872712248809, + "grad_norm": 0.21447697281837463, + "learning_rate": 8.25328261141447e-05, + "loss": 0.0322, + "step": 23718 + }, + { + "epoch": 0.5560107127996858, + "grad_norm": 0.5134793519973755, + "learning_rate": 8.252557324569317e-05, + "loss": 0.0795, + "step": 23719 + }, + { + "epoch": 0.5560341543744909, + "grad_norm": 0.333360880613327, + "learning_rate": 8.251832047205833e-05, + "loss": 0.0562, + "step": 23720 + }, + { + "epoch": 0.5560575959492958, + "grad_norm": 0.2796810269355774, + "learning_rate": 8.25110677932796e-05, + "loss": 0.0323, + "step": 23721 + }, + { + "epoch": 0.5560810375241009, + "grad_norm": 0.33875295519828796, + "learning_rate": 8.25038152093963e-05, + "loss": 0.0507, + "step": 23722 + }, + { + "epoch": 0.5561044790989058, + "grad_norm": 0.15186792612075806, + "learning_rate": 8.249656272044778e-05, + "loss": 0.0418, + "step": 23723 + }, + { + "epoch": 0.5561279206737109, + "grad_norm": 0.5765690207481384, + "learning_rate": 8.248931032647338e-05, + "loss": 0.7236, + "step": 23724 + }, + { + "epoch": 0.5561513622485158, + "grad_norm": 0.522561252117157, + "learning_rate": 8.248205802751252e-05, + "loss": 0.4928, + "step": 23725 + }, + { + "epoch": 0.5561748038233209, + "grad_norm": 0.15696051716804504, + "learning_rate": 8.247480582360451e-05, + "loss": 0.0356, + "step": 23726 + }, + { + "epoch": 0.5561982453981259, + "grad_norm": 0.18836036324501038, + "learning_rate": 8.246755371478867e-05, + "loss": 0.0384, + "step": 23727 + }, + { + "epoch": 0.5562216869729308, + "grad_norm": 0.102237768471241, + "learning_rate": 8.246030170110439e-05, + "loss": 0.0073, + "step": 23728 + }, + { + "epoch": 0.5562451285477359, + "grad_norm": 0.6517042517662048, + "learning_rate": 8.2453049782591e-05, + "loss": 0.109, + "step": 23729 + }, + { + "epoch": 0.5562685701225408, + "grad_norm": 0.13908538222312927, + "learning_rate": 8.244579795928787e-05, + "loss": 0.0137, + "step": 23730 + }, + { + "epoch": 0.5562920116973459, + "grad_norm": 0.6253851652145386, + "learning_rate": 8.243854623123433e-05, + "loss": 0.107, + "step": 23731 + }, + { + "epoch": 0.5563154532721508, + "grad_norm": 0.4175955355167389, + "learning_rate": 8.243129459846972e-05, + "loss": 0.0563, + "step": 23732 + }, + { + "epoch": 0.5563388948469559, + "grad_norm": 0.622667670249939, + "learning_rate": 8.24240430610334e-05, + "loss": 0.0638, + "step": 23733 + }, + { + "epoch": 0.5563623364217608, + "grad_norm": 0.3826988637447357, + "learning_rate": 8.241679161896472e-05, + "loss": 0.3718, + "step": 23734 + }, + { + "epoch": 0.5563857779965659, + "grad_norm": 0.6453925371170044, + "learning_rate": 8.240954027230303e-05, + "loss": 0.1239, + "step": 23735 + }, + { + "epoch": 0.5564092195713708, + "grad_norm": 0.36393290758132935, + "learning_rate": 8.240228902108764e-05, + "loss": 0.0575, + "step": 23736 + }, + { + "epoch": 0.5564326611461758, + "grad_norm": 0.5003228187561035, + "learning_rate": 8.239503786535793e-05, + "loss": 0.0765, + "step": 23737 + }, + { + "epoch": 0.5564561027209808, + "grad_norm": 0.35191610455513, + "learning_rate": 8.238778680515327e-05, + "loss": 0.0783, + "step": 23738 + }, + { + "epoch": 0.5564795442957858, + "grad_norm": 0.2956724464893341, + "learning_rate": 8.238053584051296e-05, + "loss": 0.0794, + "step": 23739 + }, + { + "epoch": 0.5565029858705908, + "grad_norm": 0.4171120226383209, + "learning_rate": 8.237328497147635e-05, + "loss": 0.0799, + "step": 23740 + }, + { + "epoch": 0.5565264274453958, + "grad_norm": 0.3901827931404114, + "learning_rate": 8.23660341980828e-05, + "loss": 0.0391, + "step": 23741 + }, + { + "epoch": 0.5565498690202008, + "grad_norm": 0.46217674016952515, + "learning_rate": 8.235878352037166e-05, + "loss": 0.1027, + "step": 23742 + }, + { + "epoch": 0.5565733105950058, + "grad_norm": 0.5507052540779114, + "learning_rate": 8.235153293838223e-05, + "loss": 0.1168, + "step": 23743 + }, + { + "epoch": 0.5565967521698107, + "grad_norm": 0.44377413392066956, + "learning_rate": 8.234428245215389e-05, + "loss": 0.0818, + "step": 23744 + }, + { + "epoch": 0.5566201937446158, + "grad_norm": 0.13378621637821198, + "learning_rate": 8.233703206172598e-05, + "loss": 0.0166, + "step": 23745 + }, + { + "epoch": 0.5566436353194207, + "grad_norm": 0.336891770362854, + "learning_rate": 8.232978176713782e-05, + "loss": 0.057, + "step": 23746 + }, + { + "epoch": 0.5566670768942258, + "grad_norm": 0.7750824093818665, + "learning_rate": 8.232253156842876e-05, + "loss": 0.1586, + "step": 23747 + }, + { + "epoch": 0.5566905184690307, + "grad_norm": 0.31230831146240234, + "learning_rate": 8.231528146563811e-05, + "loss": 0.0495, + "step": 23748 + }, + { + "epoch": 0.5567139600438358, + "grad_norm": 0.5486161708831787, + "learning_rate": 8.230803145880527e-05, + "loss": 0.0915, + "step": 23749 + }, + { + "epoch": 0.5567374016186407, + "grad_norm": 0.42376694083213806, + "learning_rate": 8.230078154796955e-05, + "loss": 0.061, + "step": 23750 + }, + { + "epoch": 0.5567608431934458, + "grad_norm": 0.3039965033531189, + "learning_rate": 8.229353173317029e-05, + "loss": 0.0582, + "step": 23751 + }, + { + "epoch": 0.5567842847682507, + "grad_norm": 0.3315909206867218, + "learning_rate": 8.228628201444682e-05, + "loss": 0.05, + "step": 23752 + }, + { + "epoch": 0.5568077263430558, + "grad_norm": 0.3002086877822876, + "learning_rate": 8.227903239183848e-05, + "loss": 0.0362, + "step": 23753 + }, + { + "epoch": 0.5568311679178607, + "grad_norm": 0.3269197642803192, + "learning_rate": 8.227178286538463e-05, + "loss": 0.044, + "step": 23754 + }, + { + "epoch": 0.5568546094926657, + "grad_norm": 0.3550558090209961, + "learning_rate": 8.226453343512456e-05, + "loss": 0.6762, + "step": 23755 + }, + { + "epoch": 0.5568780510674707, + "grad_norm": 0.16144783794879913, + "learning_rate": 8.225728410109764e-05, + "loss": 0.0186, + "step": 23756 + }, + { + "epoch": 0.5569014926422757, + "grad_norm": 0.675696849822998, + "learning_rate": 8.22500348633432e-05, + "loss": 0.1159, + "step": 23757 + }, + { + "epoch": 0.5569249342170807, + "grad_norm": 0.3019779920578003, + "learning_rate": 8.224278572190055e-05, + "loss": 0.0411, + "step": 23758 + }, + { + "epoch": 0.5569483757918857, + "grad_norm": 0.09089785814285278, + "learning_rate": 8.223553667680906e-05, + "loss": 0.0168, + "step": 23759 + }, + { + "epoch": 0.5569718173666907, + "grad_norm": 0.6444457769393921, + "learning_rate": 8.222828772810802e-05, + "loss": 0.0977, + "step": 23760 + }, + { + "epoch": 0.5569952589414957, + "grad_norm": 0.2972893714904785, + "learning_rate": 8.222103887583682e-05, + "loss": 0.0656, + "step": 23761 + }, + { + "epoch": 0.5570187005163006, + "grad_norm": 0.22289389371871948, + "learning_rate": 8.221379012003474e-05, + "loss": 0.0239, + "step": 23762 + }, + { + "epoch": 0.5570421420911057, + "grad_norm": 0.17688898742198944, + "learning_rate": 8.220654146074116e-05, + "loss": 0.0293, + "step": 23763 + }, + { + "epoch": 0.5570655836659106, + "grad_norm": 0.08085493743419647, + "learning_rate": 8.21992928979954e-05, + "loss": 0.0059, + "step": 23764 + }, + { + "epoch": 0.5570890252407157, + "grad_norm": 0.39986905455589294, + "learning_rate": 8.219204443183674e-05, + "loss": 0.0951, + "step": 23765 + }, + { + "epoch": 0.5571124668155206, + "grad_norm": 0.5716445446014404, + "learning_rate": 8.218479606230457e-05, + "loss": 0.14, + "step": 23766 + }, + { + "epoch": 0.5571359083903257, + "grad_norm": 0.21554073691368103, + "learning_rate": 8.217754778943821e-05, + "loss": 0.0274, + "step": 23767 + }, + { + "epoch": 0.5571593499651306, + "grad_norm": 0.332102507352829, + "learning_rate": 8.217029961327693e-05, + "loss": 0.04, + "step": 23768 + }, + { + "epoch": 0.5571827915399357, + "grad_norm": 0.463993638753891, + "learning_rate": 8.216305153386015e-05, + "loss": 0.0774, + "step": 23769 + }, + { + "epoch": 0.5572062331147406, + "grad_norm": 0.41691499948501587, + "learning_rate": 8.215580355122713e-05, + "loss": 0.1356, + "step": 23770 + }, + { + "epoch": 0.5572296746895457, + "grad_norm": 0.7197339534759521, + "learning_rate": 8.214855566541724e-05, + "loss": 0.1507, + "step": 23771 + }, + { + "epoch": 0.5572531162643506, + "grad_norm": 0.08937548100948334, + "learning_rate": 8.214130787646977e-05, + "loss": 0.0119, + "step": 23772 + }, + { + "epoch": 0.5572765578391556, + "grad_norm": 0.49656009674072266, + "learning_rate": 8.213406018442405e-05, + "loss": 0.0793, + "step": 23773 + }, + { + "epoch": 0.5572999994139606, + "grad_norm": 0.25640854239463806, + "learning_rate": 8.212681258931945e-05, + "loss": 0.054, + "step": 23774 + }, + { + "epoch": 0.5573234409887656, + "grad_norm": 0.5240603685379028, + "learning_rate": 8.211956509119525e-05, + "loss": 0.6728, + "step": 23775 + }, + { + "epoch": 0.5573468825635706, + "grad_norm": 0.07740968465805054, + "learning_rate": 8.211231769009082e-05, + "loss": 0.0138, + "step": 23776 + }, + { + "epoch": 0.5573703241383756, + "grad_norm": 0.5416850447654724, + "learning_rate": 8.210507038604545e-05, + "loss": 0.0977, + "step": 23777 + }, + { + "epoch": 0.5573937657131807, + "grad_norm": 0.6735635995864868, + "learning_rate": 8.209782317909844e-05, + "loss": 0.2021, + "step": 23778 + }, + { + "epoch": 0.5574172072879856, + "grad_norm": 0.6580974459648132, + "learning_rate": 8.209057606928918e-05, + "loss": 0.0967, + "step": 23779 + }, + { + "epoch": 0.5574406488627907, + "grad_norm": 0.3856992721557617, + "learning_rate": 8.208332905665697e-05, + "loss": 0.0673, + "step": 23780 + }, + { + "epoch": 0.5574640904375956, + "grad_norm": 0.5723627805709839, + "learning_rate": 8.207608214124108e-05, + "loss": 0.6326, + "step": 23781 + }, + { + "epoch": 0.5574875320124006, + "grad_norm": 0.6852423548698425, + "learning_rate": 8.206883532308089e-05, + "loss": 0.1113, + "step": 23782 + }, + { + "epoch": 0.5575109735872056, + "grad_norm": 0.551133930683136, + "learning_rate": 8.206158860221569e-05, + "loss": 0.0988, + "step": 23783 + }, + { + "epoch": 0.5575344151620106, + "grad_norm": 0.3820474445819855, + "learning_rate": 8.205434197868484e-05, + "loss": 0.0307, + "step": 23784 + }, + { + "epoch": 0.5575578567368156, + "grad_norm": 0.25434887409210205, + "learning_rate": 8.204709545252759e-05, + "loss": 0.0567, + "step": 23785 + }, + { + "epoch": 0.5575812983116206, + "grad_norm": 0.19530580937862396, + "learning_rate": 8.203984902378334e-05, + "loss": 0.0256, + "step": 23786 + }, + { + "epoch": 0.5576047398864256, + "grad_norm": 0.10287927091121674, + "learning_rate": 8.203260269249139e-05, + "loss": 0.0173, + "step": 23787 + }, + { + "epoch": 0.5576281814612306, + "grad_norm": 0.4762510061264038, + "learning_rate": 8.2025356458691e-05, + "loss": 0.1329, + "step": 23788 + }, + { + "epoch": 0.5576516230360355, + "grad_norm": 0.2383406013250351, + "learning_rate": 8.201811032242157e-05, + "loss": 0.0684, + "step": 23789 + }, + { + "epoch": 0.5576750646108406, + "grad_norm": 0.2856440842151642, + "learning_rate": 8.201086428372236e-05, + "loss": 0.0247, + "step": 23790 + }, + { + "epoch": 0.5576985061856455, + "grad_norm": 0.3287726640701294, + "learning_rate": 8.200361834263272e-05, + "loss": 0.069, + "step": 23791 + }, + { + "epoch": 0.5577219477604506, + "grad_norm": 0.40758973360061646, + "learning_rate": 8.199637249919194e-05, + "loss": 0.0563, + "step": 23792 + }, + { + "epoch": 0.5577453893352555, + "grad_norm": 0.4616345763206482, + "learning_rate": 8.198912675343935e-05, + "loss": 0.1105, + "step": 23793 + }, + { + "epoch": 0.5577688309100606, + "grad_norm": 0.33196619153022766, + "learning_rate": 8.198188110541427e-05, + "loss": 0.4052, + "step": 23794 + }, + { + "epoch": 0.5577922724848655, + "grad_norm": 0.47315919399261475, + "learning_rate": 8.197463555515603e-05, + "loss": 0.1126, + "step": 23795 + }, + { + "epoch": 0.5578157140596706, + "grad_norm": 0.792264461517334, + "learning_rate": 8.196739010270388e-05, + "loss": 0.8835, + "step": 23796 + }, + { + "epoch": 0.5578391556344755, + "grad_norm": 0.48602181673049927, + "learning_rate": 8.196014474809717e-05, + "loss": 0.7439, + "step": 23797 + }, + { + "epoch": 0.5578625972092806, + "grad_norm": 0.26258423924446106, + "learning_rate": 8.195289949137525e-05, + "loss": 0.028, + "step": 23798 + }, + { + "epoch": 0.5578860387840855, + "grad_norm": 0.20532871782779694, + "learning_rate": 8.194565433257741e-05, + "loss": 0.028, + "step": 23799 + }, + { + "epoch": 0.5579094803588905, + "grad_norm": 0.14513817429542542, + "learning_rate": 8.193840927174295e-05, + "loss": 0.0391, + "step": 23800 + }, + { + "epoch": 0.5579329219336955, + "grad_norm": 0.40627944469451904, + "learning_rate": 8.19311643089112e-05, + "loss": 0.0893, + "step": 23801 + }, + { + "epoch": 0.5579563635085005, + "grad_norm": 0.21692398190498352, + "learning_rate": 8.192391944412146e-05, + "loss": 0.0358, + "step": 23802 + }, + { + "epoch": 0.5579798050833055, + "grad_norm": 0.11185061186552048, + "learning_rate": 8.191667467741302e-05, + "loss": 0.0085, + "step": 23803 + }, + { + "epoch": 0.5580032466581105, + "grad_norm": 0.48082858324050903, + "learning_rate": 8.190943000882524e-05, + "loss": 0.0484, + "step": 23804 + }, + { + "epoch": 0.5580266882329155, + "grad_norm": 0.29004791378974915, + "learning_rate": 8.19021854383974e-05, + "loss": 0.0696, + "step": 23805 + }, + { + "epoch": 0.5580501298077205, + "grad_norm": 0.25927412509918213, + "learning_rate": 8.189494096616879e-05, + "loss": 0.0777, + "step": 23806 + }, + { + "epoch": 0.5580735713825254, + "grad_norm": 0.2175617665052414, + "learning_rate": 8.188769659217875e-05, + "loss": 0.0556, + "step": 23807 + }, + { + "epoch": 0.5580970129573305, + "grad_norm": 0.542327880859375, + "learning_rate": 8.188045231646658e-05, + "loss": 0.1025, + "step": 23808 + }, + { + "epoch": 0.5581204545321354, + "grad_norm": 0.4358844757080078, + "learning_rate": 8.187320813907155e-05, + "loss": 0.0654, + "step": 23809 + }, + { + "epoch": 0.5581438961069405, + "grad_norm": 0.14588727056980133, + "learning_rate": 8.186596406003302e-05, + "loss": 0.0445, + "step": 23810 + }, + { + "epoch": 0.5581673376817454, + "grad_norm": 0.5420334339141846, + "learning_rate": 8.185872007939031e-05, + "loss": 0.1606, + "step": 23811 + }, + { + "epoch": 0.5581907792565505, + "grad_norm": 0.13039568066596985, + "learning_rate": 8.185147619718268e-05, + "loss": 0.0201, + "step": 23812 + }, + { + "epoch": 0.5582142208313554, + "grad_norm": 0.4788607358932495, + "learning_rate": 8.184423241344943e-05, + "loss": 0.0955, + "step": 23813 + }, + { + "epoch": 0.5582376624061605, + "grad_norm": 0.1238732561469078, + "learning_rate": 8.183698872822991e-05, + "loss": 0.0261, + "step": 23814 + }, + { + "epoch": 0.5582611039809654, + "grad_norm": 0.29080140590667725, + "learning_rate": 8.182974514156341e-05, + "loss": 0.0333, + "step": 23815 + }, + { + "epoch": 0.5582845455557705, + "grad_norm": 0.25125381350517273, + "learning_rate": 8.182250165348919e-05, + "loss": 0.0347, + "step": 23816 + }, + { + "epoch": 0.5583079871305754, + "grad_norm": 0.5685376524925232, + "learning_rate": 8.181525826404661e-05, + "loss": 0.041, + "step": 23817 + }, + { + "epoch": 0.5583314287053804, + "grad_norm": 0.629757821559906, + "learning_rate": 8.180801497327495e-05, + "loss": 0.1096, + "step": 23818 + }, + { + "epoch": 0.5583548702801854, + "grad_norm": 0.334922730922699, + "learning_rate": 8.18007717812135e-05, + "loss": 0.0639, + "step": 23819 + }, + { + "epoch": 0.5583783118549904, + "grad_norm": 0.4595947861671448, + "learning_rate": 8.179352868790158e-05, + "loss": 0.1286, + "step": 23820 + }, + { + "epoch": 0.5584017534297954, + "grad_norm": 0.22530293464660645, + "learning_rate": 8.178628569337847e-05, + "loss": 0.0101, + "step": 23821 + }, + { + "epoch": 0.5584251950046004, + "grad_norm": 0.368961364030838, + "learning_rate": 8.177904279768349e-05, + "loss": 0.0777, + "step": 23822 + }, + { + "epoch": 0.5584486365794054, + "grad_norm": 0.2550239861011505, + "learning_rate": 8.177180000085594e-05, + "loss": 0.0202, + "step": 23823 + }, + { + "epoch": 0.5584720781542104, + "grad_norm": 0.39155516028404236, + "learning_rate": 8.176455730293514e-05, + "loss": 0.0952, + "step": 23824 + }, + { + "epoch": 0.5584955197290153, + "grad_norm": 0.35581570863723755, + "learning_rate": 8.175731470396036e-05, + "loss": 0.0486, + "step": 23825 + }, + { + "epoch": 0.5585189613038204, + "grad_norm": 0.7806900143623352, + "learning_rate": 8.175007220397087e-05, + "loss": 0.178, + "step": 23826 + }, + { + "epoch": 0.5585424028786253, + "grad_norm": 0.794240415096283, + "learning_rate": 8.174282980300603e-05, + "loss": 0.1333, + "step": 23827 + }, + { + "epoch": 0.5585658444534304, + "grad_norm": 0.10694719851016998, + "learning_rate": 8.173558750110509e-05, + "loss": 0.0163, + "step": 23828 + }, + { + "epoch": 0.5585892860282354, + "grad_norm": 0.215941920876503, + "learning_rate": 8.172834529830739e-05, + "loss": 0.0291, + "step": 23829 + }, + { + "epoch": 0.5586127276030404, + "grad_norm": 1.3071171045303345, + "learning_rate": 8.172110319465219e-05, + "loss": 0.0867, + "step": 23830 + }, + { + "epoch": 0.5586361691778454, + "grad_norm": 0.11495815962553024, + "learning_rate": 8.171386119017879e-05, + "loss": 0.0243, + "step": 23831 + }, + { + "epoch": 0.5586596107526504, + "grad_norm": 0.8598188757896423, + "learning_rate": 8.170661928492651e-05, + "loss": 0.1574, + "step": 23832 + }, + { + "epoch": 0.5586830523274554, + "grad_norm": 0.3699814975261688, + "learning_rate": 8.169937747893458e-05, + "loss": 0.0576, + "step": 23833 + }, + { + "epoch": 0.5587064939022603, + "grad_norm": 0.22192539274692535, + "learning_rate": 8.16921357722424e-05, + "loss": 0.0376, + "step": 23834 + }, + { + "epoch": 0.5587299354770654, + "grad_norm": 0.40812650322914124, + "learning_rate": 8.168489416488916e-05, + "loss": 0.0607, + "step": 23835 + }, + { + "epoch": 0.5587533770518703, + "grad_norm": 0.4533635973930359, + "learning_rate": 8.167765265691424e-05, + "loss": 0.0968, + "step": 23836 + }, + { + "epoch": 0.5587768186266754, + "grad_norm": 0.564350962638855, + "learning_rate": 8.167041124835688e-05, + "loss": 0.1209, + "step": 23837 + }, + { + "epoch": 0.5588002602014803, + "grad_norm": 0.3567432165145874, + "learning_rate": 8.166316993925637e-05, + "loss": 0.072, + "step": 23838 + }, + { + "epoch": 0.5588237017762854, + "grad_norm": 0.472951740026474, + "learning_rate": 8.165592872965204e-05, + "loss": 0.1075, + "step": 23839 + }, + { + "epoch": 0.5588471433510903, + "grad_norm": 0.5200668573379517, + "learning_rate": 8.164868761958315e-05, + "loss": 0.1174, + "step": 23840 + }, + { + "epoch": 0.5588705849258954, + "grad_norm": 0.1575564593076706, + "learning_rate": 8.164144660908898e-05, + "loss": 0.0351, + "step": 23841 + }, + { + "epoch": 0.5588940265007003, + "grad_norm": 0.32670077681541443, + "learning_rate": 8.163420569820886e-05, + "loss": 0.0433, + "step": 23842 + }, + { + "epoch": 0.5589174680755054, + "grad_norm": 0.0960952416062355, + "learning_rate": 8.162696488698205e-05, + "loss": 0.0141, + "step": 23843 + }, + { + "epoch": 0.5589409096503103, + "grad_norm": 0.6429228186607361, + "learning_rate": 8.161972417544783e-05, + "loss": 0.5262, + "step": 23844 + }, + { + "epoch": 0.5589643512251153, + "grad_norm": 0.5937825441360474, + "learning_rate": 8.161248356364552e-05, + "loss": 0.1093, + "step": 23845 + }, + { + "epoch": 0.5589877927999203, + "grad_norm": 0.3827628195285797, + "learning_rate": 8.160524305161436e-05, + "loss": 0.0546, + "step": 23846 + }, + { + "epoch": 0.5590112343747253, + "grad_norm": 0.5502282977104187, + "learning_rate": 8.15980026393937e-05, + "loss": 0.8144, + "step": 23847 + }, + { + "epoch": 0.5590346759495303, + "grad_norm": 0.5642492175102234, + "learning_rate": 8.159076232702277e-05, + "loss": 0.1792, + "step": 23848 + }, + { + "epoch": 0.5590581175243353, + "grad_norm": 0.40801721811294556, + "learning_rate": 8.15835221145409e-05, + "loss": 0.072, + "step": 23849 + }, + { + "epoch": 0.5590815590991403, + "grad_norm": 0.24196264147758484, + "learning_rate": 8.157628200198737e-05, + "loss": 0.0268, + "step": 23850 + }, + { + "epoch": 0.5591050006739453, + "grad_norm": 0.3209019601345062, + "learning_rate": 8.156904198940144e-05, + "loss": 0.0704, + "step": 23851 + }, + { + "epoch": 0.5591284422487502, + "grad_norm": 0.3439646363258362, + "learning_rate": 8.15618020768224e-05, + "loss": 0.0806, + "step": 23852 + }, + { + "epoch": 0.5591518838235553, + "grad_norm": 0.3510136902332306, + "learning_rate": 8.155456226428954e-05, + "loss": 0.0948, + "step": 23853 + }, + { + "epoch": 0.5591753253983602, + "grad_norm": 0.36734530329704285, + "learning_rate": 8.154732255184214e-05, + "loss": 0.0748, + "step": 23854 + }, + { + "epoch": 0.5591987669731653, + "grad_norm": 0.7096545100212097, + "learning_rate": 8.154008293951952e-05, + "loss": 0.1353, + "step": 23855 + }, + { + "epoch": 0.5592222085479702, + "grad_norm": 0.3424215614795685, + "learning_rate": 8.15328434273609e-05, + "loss": 0.039, + "step": 23856 + }, + { + "epoch": 0.5592456501227753, + "grad_norm": 0.6569095849990845, + "learning_rate": 8.152560401540559e-05, + "loss": 0.0663, + "step": 23857 + }, + { + "epoch": 0.5592690916975802, + "grad_norm": 0.2621670365333557, + "learning_rate": 8.151836470369282e-05, + "loss": 0.063, + "step": 23858 + }, + { + "epoch": 0.5592925332723853, + "grad_norm": 0.20316913723945618, + "learning_rate": 8.151112549226201e-05, + "loss": 0.0392, + "step": 23859 + }, + { + "epoch": 0.5593159748471902, + "grad_norm": 0.3987743854522705, + "learning_rate": 8.150388638115232e-05, + "loss": 0.0603, + "step": 23860 + }, + { + "epoch": 0.5593394164219953, + "grad_norm": 0.47712522745132446, + "learning_rate": 8.149664737040306e-05, + "loss": 0.0645, + "step": 23861 + }, + { + "epoch": 0.5593628579968002, + "grad_norm": 0.34487631916999817, + "learning_rate": 8.148940846005353e-05, + "loss": 0.0631, + "step": 23862 + }, + { + "epoch": 0.5593862995716052, + "grad_norm": 0.401715487241745, + "learning_rate": 8.148216965014298e-05, + "loss": 0.0683, + "step": 23863 + }, + { + "epoch": 0.5594097411464102, + "grad_norm": 0.13124623894691467, + "learning_rate": 8.147493094071068e-05, + "loss": 0.0313, + "step": 23864 + }, + { + "epoch": 0.5594331827212152, + "grad_norm": 0.316557914018631, + "learning_rate": 8.146769233179595e-05, + "loss": 0.0652, + "step": 23865 + }, + { + "epoch": 0.5594566242960202, + "grad_norm": 0.8024784922599792, + "learning_rate": 8.146045382343803e-05, + "loss": 0.1228, + "step": 23866 + }, + { + "epoch": 0.5594800658708252, + "grad_norm": 0.7190836071968079, + "learning_rate": 8.14532154156762e-05, + "loss": 0.0699, + "step": 23867 + }, + { + "epoch": 0.5595035074456302, + "grad_norm": 0.380562961101532, + "learning_rate": 8.144597710854978e-05, + "loss": 0.1054, + "step": 23868 + }, + { + "epoch": 0.5595269490204352, + "grad_norm": 0.5044623017311096, + "learning_rate": 8.143873890209797e-05, + "loss": 0.0961, + "step": 23869 + }, + { + "epoch": 0.5595503905952401, + "grad_norm": 0.6028769016265869, + "learning_rate": 8.143150079636006e-05, + "loss": 0.7485, + "step": 23870 + }, + { + "epoch": 0.5595738321700452, + "grad_norm": 0.16148287057876587, + "learning_rate": 8.142426279137541e-05, + "loss": 0.0383, + "step": 23871 + }, + { + "epoch": 0.5595972737448501, + "grad_norm": 0.3681776821613312, + "learning_rate": 8.141702488718321e-05, + "loss": 0.0635, + "step": 23872 + }, + { + "epoch": 0.5596207153196552, + "grad_norm": 0.05542927235364914, + "learning_rate": 8.140978708382275e-05, + "loss": 0.0102, + "step": 23873 + }, + { + "epoch": 0.5596441568944601, + "grad_norm": 0.10689659416675568, + "learning_rate": 8.140254938133333e-05, + "loss": 0.0199, + "step": 23874 + }, + { + "epoch": 0.5596675984692652, + "grad_norm": 0.2550494074821472, + "learning_rate": 8.13953117797542e-05, + "loss": 0.0561, + "step": 23875 + }, + { + "epoch": 0.5596910400440701, + "grad_norm": 0.2878575325012207, + "learning_rate": 8.13880742791246e-05, + "loss": 0.0256, + "step": 23876 + }, + { + "epoch": 0.5597144816188752, + "grad_norm": 0.3300733268260956, + "learning_rate": 8.138083687948385e-05, + "loss": 0.0394, + "step": 23877 + }, + { + "epoch": 0.5597379231936801, + "grad_norm": 0.7292587757110596, + "learning_rate": 8.137359958087121e-05, + "loss": 0.098, + "step": 23878 + }, + { + "epoch": 0.5597613647684851, + "grad_norm": 0.3236267566680908, + "learning_rate": 8.136636238332592e-05, + "loss": 0.4482, + "step": 23879 + }, + { + "epoch": 0.5597848063432902, + "grad_norm": 0.5844895243644714, + "learning_rate": 8.135912528688729e-05, + "loss": 0.7122, + "step": 23880 + }, + { + "epoch": 0.5598082479180951, + "grad_norm": 0.5229164958000183, + "learning_rate": 8.135188829159458e-05, + "loss": 0.0826, + "step": 23881 + }, + { + "epoch": 0.5598316894929002, + "grad_norm": 0.08630381524562836, + "learning_rate": 8.1344651397487e-05, + "loss": 0.0141, + "step": 23882 + }, + { + "epoch": 0.5598551310677051, + "grad_norm": 0.212244912981987, + "learning_rate": 8.133741460460389e-05, + "loss": 0.0647, + "step": 23883 + }, + { + "epoch": 0.5598785726425102, + "grad_norm": 0.15010201930999756, + "learning_rate": 8.133017791298451e-05, + "loss": 0.0164, + "step": 23884 + }, + { + "epoch": 0.5599020142173151, + "grad_norm": 0.2345314770936966, + "learning_rate": 8.13229413226681e-05, + "loss": 0.0086, + "step": 23885 + }, + { + "epoch": 0.5599254557921202, + "grad_norm": 0.3121756911277771, + "learning_rate": 8.131570483369394e-05, + "loss": 0.0402, + "step": 23886 + }, + { + "epoch": 0.5599488973669251, + "grad_norm": 0.3840656280517578, + "learning_rate": 8.13084684461013e-05, + "loss": 0.0598, + "step": 23887 + }, + { + "epoch": 0.5599723389417302, + "grad_norm": 0.05501572787761688, + "learning_rate": 8.130123215992944e-05, + "loss": 0.0064, + "step": 23888 + }, + { + "epoch": 0.5599957805165351, + "grad_norm": 1.3030812740325928, + "learning_rate": 8.129399597521759e-05, + "loss": 0.1248, + "step": 23889 + }, + { + "epoch": 0.5600192220913401, + "grad_norm": 0.40621817111968994, + "learning_rate": 8.128675989200507e-05, + "loss": 0.0344, + "step": 23890 + }, + { + "epoch": 0.5600426636661451, + "grad_norm": 0.3999440670013428, + "learning_rate": 8.127952391033111e-05, + "loss": 0.0718, + "step": 23891 + }, + { + "epoch": 0.5600661052409501, + "grad_norm": 0.24460656940937042, + "learning_rate": 8.127228803023496e-05, + "loss": 0.0347, + "step": 23892 + }, + { + "epoch": 0.5600895468157551, + "grad_norm": 0.3850669860839844, + "learning_rate": 8.126505225175593e-05, + "loss": 0.0566, + "step": 23893 + }, + { + "epoch": 0.5601129883905601, + "grad_norm": 0.560232400894165, + "learning_rate": 8.125781657493321e-05, + "loss": 0.0927, + "step": 23894 + }, + { + "epoch": 0.5601364299653651, + "grad_norm": 0.27393728494644165, + "learning_rate": 8.125058099980615e-05, + "loss": 0.0403, + "step": 23895 + }, + { + "epoch": 0.5601598715401701, + "grad_norm": 0.6614543795585632, + "learning_rate": 8.124334552641393e-05, + "loss": 0.1118, + "step": 23896 + }, + { + "epoch": 0.560183313114975, + "grad_norm": 0.12710925936698914, + "learning_rate": 8.123611015479588e-05, + "loss": 0.0243, + "step": 23897 + }, + { + "epoch": 0.5602067546897801, + "grad_norm": 0.3035387694835663, + "learning_rate": 8.122887488499121e-05, + "loss": 0.0606, + "step": 23898 + }, + { + "epoch": 0.560230196264585, + "grad_norm": 0.2694835662841797, + "learning_rate": 8.122163971703918e-05, + "loss": 0.0551, + "step": 23899 + }, + { + "epoch": 0.5602536378393901, + "grad_norm": 0.2499072253704071, + "learning_rate": 8.121440465097908e-05, + "loss": 0.0409, + "step": 23900 + }, + { + "epoch": 0.560277079414195, + "grad_norm": 0.17181611061096191, + "learning_rate": 8.120716968685013e-05, + "loss": 0.0375, + "step": 23901 + }, + { + "epoch": 0.5603005209890001, + "grad_norm": 0.23607438802719116, + "learning_rate": 8.119993482469162e-05, + "loss": 0.0409, + "step": 23902 + }, + { + "epoch": 0.560323962563805, + "grad_norm": 0.544080376625061, + "learning_rate": 8.11927000645428e-05, + "loss": 0.1095, + "step": 23903 + }, + { + "epoch": 0.5603474041386101, + "grad_norm": 0.3684292435646057, + "learning_rate": 8.118546540644289e-05, + "loss": 0.0722, + "step": 23904 + }, + { + "epoch": 0.560370845713415, + "grad_norm": 0.5187355279922485, + "learning_rate": 8.11782308504312e-05, + "loss": 0.551, + "step": 23905 + }, + { + "epoch": 0.56039428728822, + "grad_norm": 0.4341696500778198, + "learning_rate": 8.117099639654693e-05, + "loss": 0.0698, + "step": 23906 + }, + { + "epoch": 0.560417728863025, + "grad_norm": 0.4751228094100952, + "learning_rate": 8.116376204482939e-05, + "loss": 0.0554, + "step": 23907 + }, + { + "epoch": 0.56044117043783, + "grad_norm": 0.792789101600647, + "learning_rate": 8.115652779531778e-05, + "loss": 0.1671, + "step": 23908 + }, + { + "epoch": 0.560464612012635, + "grad_norm": 0.2722955048084259, + "learning_rate": 8.114929364805141e-05, + "loss": 0.0417, + "step": 23909 + }, + { + "epoch": 0.56048805358744, + "grad_norm": 0.7416477799415588, + "learning_rate": 8.11420596030695e-05, + "loss": 0.674, + "step": 23910 + }, + { + "epoch": 0.560511495162245, + "grad_norm": 0.2315642237663269, + "learning_rate": 8.113482566041128e-05, + "loss": 0.0478, + "step": 23911 + }, + { + "epoch": 0.56053493673705, + "grad_norm": 0.3858935534954071, + "learning_rate": 8.112759182011605e-05, + "loss": 0.1139, + "step": 23912 + }, + { + "epoch": 0.560558378311855, + "grad_norm": 0.08722130954265594, + "learning_rate": 8.112035808222305e-05, + "loss": 0.0065, + "step": 23913 + }, + { + "epoch": 0.56058181988666, + "grad_norm": 0.03728542849421501, + "learning_rate": 8.111312444677149e-05, + "loss": 0.0023, + "step": 23914 + }, + { + "epoch": 0.5606052614614649, + "grad_norm": 0.5125826001167297, + "learning_rate": 8.110589091380065e-05, + "loss": 0.1451, + "step": 23915 + }, + { + "epoch": 0.56062870303627, + "grad_norm": 0.4282246530056, + "learning_rate": 8.109865748334979e-05, + "loss": 0.137, + "step": 23916 + }, + { + "epoch": 0.5606521446110749, + "grad_norm": 0.2585207223892212, + "learning_rate": 8.109142415545813e-05, + "loss": 0.063, + "step": 23917 + }, + { + "epoch": 0.56067558618588, + "grad_norm": 0.4580410420894623, + "learning_rate": 8.10841909301649e-05, + "loss": 0.0896, + "step": 23918 + }, + { + "epoch": 0.5606990277606849, + "grad_norm": 0.4817797839641571, + "learning_rate": 8.107695780750945e-05, + "loss": 0.0738, + "step": 23919 + }, + { + "epoch": 0.56072246933549, + "grad_norm": 0.31615710258483887, + "learning_rate": 8.106972478753094e-05, + "loss": 0.0403, + "step": 23920 + }, + { + "epoch": 0.5607459109102949, + "grad_norm": 0.2621224522590637, + "learning_rate": 8.106249187026861e-05, + "loss": 0.0314, + "step": 23921 + }, + { + "epoch": 0.5607693524851, + "grad_norm": 0.490269273519516, + "learning_rate": 8.105525905576176e-05, + "loss": 0.1057, + "step": 23922 + }, + { + "epoch": 0.5607927940599049, + "grad_norm": 0.537460207939148, + "learning_rate": 8.104802634404961e-05, + "loss": 0.1274, + "step": 23923 + }, + { + "epoch": 0.56081623563471, + "grad_norm": 0.4402539134025574, + "learning_rate": 8.104079373517138e-05, + "loss": 0.0635, + "step": 23924 + }, + { + "epoch": 0.5608396772095149, + "grad_norm": 0.5473488569259644, + "learning_rate": 8.103356122916635e-05, + "loss": 0.1266, + "step": 23925 + }, + { + "epoch": 0.5608631187843199, + "grad_norm": 0.168406143784523, + "learning_rate": 8.102632882607376e-05, + "loss": 0.0297, + "step": 23926 + }, + { + "epoch": 0.5608865603591249, + "grad_norm": 0.22478805482387543, + "learning_rate": 8.101909652593283e-05, + "loss": 0.0383, + "step": 23927 + }, + { + "epoch": 0.5609100019339299, + "grad_norm": 0.415676474571228, + "learning_rate": 8.101186432878282e-05, + "loss": 0.0249, + "step": 23928 + }, + { + "epoch": 0.5609334435087349, + "grad_norm": 0.6904425621032715, + "learning_rate": 8.100463223466296e-05, + "loss": 0.1218, + "step": 23929 + }, + { + "epoch": 0.5609568850835399, + "grad_norm": 0.17020420730113983, + "learning_rate": 8.09974002436125e-05, + "loss": 0.0545, + "step": 23930 + }, + { + "epoch": 0.5609803266583449, + "grad_norm": 0.7728256583213806, + "learning_rate": 8.099016835567065e-05, + "loss": 0.1627, + "step": 23931 + }, + { + "epoch": 0.5610037682331499, + "grad_norm": 0.5658969283103943, + "learning_rate": 8.098293657087673e-05, + "loss": 0.0838, + "step": 23932 + }, + { + "epoch": 0.561027209807955, + "grad_norm": 0.288906067609787, + "learning_rate": 8.097570488926991e-05, + "loss": 0.0304, + "step": 23933 + }, + { + "epoch": 0.5610506513827599, + "grad_norm": 0.313198059797287, + "learning_rate": 8.096847331088945e-05, + "loss": 0.0456, + "step": 23934 + }, + { + "epoch": 0.5610740929575649, + "grad_norm": 0.4245946705341339, + "learning_rate": 8.09612418357746e-05, + "loss": 0.0602, + "step": 23935 + }, + { + "epoch": 0.5610975345323699, + "grad_norm": 0.20291173458099365, + "learning_rate": 8.095401046396459e-05, + "loss": 0.0459, + "step": 23936 + }, + { + "epoch": 0.5611209761071749, + "grad_norm": 0.17960911989212036, + "learning_rate": 8.094677919549864e-05, + "loss": 0.0388, + "step": 23937 + }, + { + "epoch": 0.5611444176819799, + "grad_norm": 0.4910365045070648, + "learning_rate": 8.093954803041602e-05, + "loss": 0.7652, + "step": 23938 + }, + { + "epoch": 0.5611678592567849, + "grad_norm": 0.3462724983692169, + "learning_rate": 8.093231696875592e-05, + "loss": 0.0494, + "step": 23939 + }, + { + "epoch": 0.5611913008315899, + "grad_norm": 0.47183558344841003, + "learning_rate": 8.092508601055763e-05, + "loss": 0.0733, + "step": 23940 + }, + { + "epoch": 0.5612147424063949, + "grad_norm": 0.15497377514839172, + "learning_rate": 8.091785515586036e-05, + "loss": 0.018, + "step": 23941 + }, + { + "epoch": 0.5612381839811998, + "grad_norm": 0.44573941826820374, + "learning_rate": 8.091062440470333e-05, + "loss": 0.0844, + "step": 23942 + }, + { + "epoch": 0.5612616255560049, + "grad_norm": 0.3535064458847046, + "learning_rate": 8.090339375712578e-05, + "loss": 0.0613, + "step": 23943 + }, + { + "epoch": 0.5612850671308098, + "grad_norm": 0.4350404739379883, + "learning_rate": 8.089616321316696e-05, + "loss": 0.0765, + "step": 23944 + }, + { + "epoch": 0.5613085087056149, + "grad_norm": 0.3507615625858307, + "learning_rate": 8.08889327728661e-05, + "loss": 0.0515, + "step": 23945 + }, + { + "epoch": 0.5613319502804198, + "grad_norm": 0.5694718360900879, + "learning_rate": 8.088170243626242e-05, + "loss": 0.11, + "step": 23946 + }, + { + "epoch": 0.5613553918552249, + "grad_norm": 0.57123863697052, + "learning_rate": 8.087447220339519e-05, + "loss": 0.0914, + "step": 23947 + }, + { + "epoch": 0.5613788334300298, + "grad_norm": 0.5692154765129089, + "learning_rate": 8.086724207430358e-05, + "loss": 0.1083, + "step": 23948 + }, + { + "epoch": 0.5614022750048349, + "grad_norm": 0.23839171230793, + "learning_rate": 8.086001204902686e-05, + "loss": 0.0512, + "step": 23949 + }, + { + "epoch": 0.5614257165796398, + "grad_norm": 0.5086683034896851, + "learning_rate": 8.085278212760426e-05, + "loss": 0.0928, + "step": 23950 + }, + { + "epoch": 0.5614491581544448, + "grad_norm": 0.40895208716392517, + "learning_rate": 8.084555231007502e-05, + "loss": 0.0467, + "step": 23951 + }, + { + "epoch": 0.5614725997292498, + "grad_norm": 0.4036644995212555, + "learning_rate": 8.083832259647833e-05, + "loss": 0.0149, + "step": 23952 + }, + { + "epoch": 0.5614960413040548, + "grad_norm": 0.5532739162445068, + "learning_rate": 8.083109298685345e-05, + "loss": 0.1269, + "step": 23953 + }, + { + "epoch": 0.5615194828788598, + "grad_norm": 0.48322054743766785, + "learning_rate": 8.082386348123958e-05, + "loss": 0.0687, + "step": 23954 + }, + { + "epoch": 0.5615429244536648, + "grad_norm": 0.6186321377754211, + "learning_rate": 8.081663407967595e-05, + "loss": 0.1343, + "step": 23955 + }, + { + "epoch": 0.5615663660284698, + "grad_norm": 0.4194965362548828, + "learning_rate": 8.080940478220181e-05, + "loss": 0.0524, + "step": 23956 + }, + { + "epoch": 0.5615898076032748, + "grad_norm": 0.9299081563949585, + "learning_rate": 8.08021755888564e-05, + "loss": 0.2157, + "step": 23957 + }, + { + "epoch": 0.5616132491780798, + "grad_norm": 0.10031404346227646, + "learning_rate": 8.079494649967895e-05, + "loss": 0.0165, + "step": 23958 + }, + { + "epoch": 0.5616366907528848, + "grad_norm": 0.6681910157203674, + "learning_rate": 8.078771751470862e-05, + "loss": 0.5924, + "step": 23959 + }, + { + "epoch": 0.5616601323276897, + "grad_norm": 0.3400794267654419, + "learning_rate": 8.07804886339847e-05, + "loss": 0.0912, + "step": 23960 + }, + { + "epoch": 0.5616835739024948, + "grad_norm": 0.2298535257577896, + "learning_rate": 8.077325985754638e-05, + "loss": 0.0569, + "step": 23961 + }, + { + "epoch": 0.5617070154772997, + "grad_norm": 0.5890882611274719, + "learning_rate": 8.076603118543288e-05, + "loss": 0.1408, + "step": 23962 + }, + { + "epoch": 0.5617304570521048, + "grad_norm": 0.33897408843040466, + "learning_rate": 8.075880261768344e-05, + "loss": 0.4195, + "step": 23963 + }, + { + "epoch": 0.5617538986269097, + "grad_norm": 0.23111867904663086, + "learning_rate": 8.07515741543373e-05, + "loss": 0.0368, + "step": 23964 + }, + { + "epoch": 0.5617773402017148, + "grad_norm": 0.5207710266113281, + "learning_rate": 8.074434579543364e-05, + "loss": 0.6329, + "step": 23965 + }, + { + "epoch": 0.5618007817765197, + "grad_norm": 0.22552400827407837, + "learning_rate": 8.073711754101171e-05, + "loss": 0.0308, + "step": 23966 + }, + { + "epoch": 0.5618242233513248, + "grad_norm": 0.7052634954452515, + "learning_rate": 8.072988939111068e-05, + "loss": 0.0406, + "step": 23967 + }, + { + "epoch": 0.5618476649261297, + "grad_norm": 0.3904842138290405, + "learning_rate": 8.072266134576987e-05, + "loss": 0.0426, + "step": 23968 + }, + { + "epoch": 0.5618711065009347, + "grad_norm": 0.8218950033187866, + "learning_rate": 8.07154334050284e-05, + "loss": 0.2239, + "step": 23969 + }, + { + "epoch": 0.5618945480757397, + "grad_norm": 0.19854457676410675, + "learning_rate": 8.070820556892556e-05, + "loss": 0.0154, + "step": 23970 + }, + { + "epoch": 0.5619179896505447, + "grad_norm": 0.16233877837657928, + "learning_rate": 8.070097783750053e-05, + "loss": 0.0166, + "step": 23971 + }, + { + "epoch": 0.5619414312253497, + "grad_norm": 0.4368203282356262, + "learning_rate": 8.069375021079253e-05, + "loss": 0.1041, + "step": 23972 + }, + { + "epoch": 0.5619648728001547, + "grad_norm": 0.3710160553455353, + "learning_rate": 8.068652268884081e-05, + "loss": 0.0765, + "step": 23973 + }, + { + "epoch": 0.5619883143749597, + "grad_norm": 0.6509361267089844, + "learning_rate": 8.067929527168454e-05, + "loss": 0.1551, + "step": 23974 + }, + { + "epoch": 0.5620117559497647, + "grad_norm": 0.11364419758319855, + "learning_rate": 8.067206795936295e-05, + "loss": 0.0095, + "step": 23975 + }, + { + "epoch": 0.5620351975245697, + "grad_norm": 0.7010259628295898, + "learning_rate": 8.066484075191528e-05, + "loss": 0.1472, + "step": 23976 + }, + { + "epoch": 0.5620586390993747, + "grad_norm": 0.2392575889825821, + "learning_rate": 8.065761364938071e-05, + "loss": 0.027, + "step": 23977 + }, + { + "epoch": 0.5620820806741796, + "grad_norm": 0.2802910804748535, + "learning_rate": 8.065038665179849e-05, + "loss": 0.0436, + "step": 23978 + }, + { + "epoch": 0.5621055222489847, + "grad_norm": 0.24471648037433624, + "learning_rate": 8.064315975920778e-05, + "loss": 0.0323, + "step": 23979 + }, + { + "epoch": 0.5621289638237896, + "grad_norm": 0.23871944844722748, + "learning_rate": 8.063593297164787e-05, + "loss": 0.0543, + "step": 23980 + }, + { + "epoch": 0.5621524053985947, + "grad_norm": 0.1932598352432251, + "learning_rate": 8.062870628915792e-05, + "loss": 0.0341, + "step": 23981 + }, + { + "epoch": 0.5621758469733996, + "grad_norm": 0.512214720249176, + "learning_rate": 8.062147971177714e-05, + "loss": 0.109, + "step": 23982 + }, + { + "epoch": 0.5621992885482047, + "grad_norm": 0.37573736906051636, + "learning_rate": 8.061425323954478e-05, + "loss": 0.0463, + "step": 23983 + }, + { + "epoch": 0.5622227301230097, + "grad_norm": 0.13272498548030853, + "learning_rate": 8.06070268725e-05, + "loss": 0.024, + "step": 23984 + }, + { + "epoch": 0.5622461716978147, + "grad_norm": 0.22397467494010925, + "learning_rate": 8.059980061068206e-05, + "loss": 0.0439, + "step": 23985 + }, + { + "epoch": 0.5622696132726197, + "grad_norm": 0.48058846592903137, + "learning_rate": 8.059257445413016e-05, + "loss": 0.0773, + "step": 23986 + }, + { + "epoch": 0.5622930548474246, + "grad_norm": 0.4956182837486267, + "learning_rate": 8.058534840288346e-05, + "loss": 0.0712, + "step": 23987 + }, + { + "epoch": 0.5623164964222297, + "grad_norm": 0.5161957144737244, + "learning_rate": 8.057812245698123e-05, + "loss": 0.092, + "step": 23988 + }, + { + "epoch": 0.5623399379970346, + "grad_norm": 0.646400511264801, + "learning_rate": 8.057089661646267e-05, + "loss": 0.067, + "step": 23989 + }, + { + "epoch": 0.5623633795718397, + "grad_norm": 0.4143316149711609, + "learning_rate": 8.056367088136692e-05, + "loss": 0.077, + "step": 23990 + }, + { + "epoch": 0.5623868211466446, + "grad_norm": 0.5434496998786926, + "learning_rate": 8.055644525173323e-05, + "loss": 0.1019, + "step": 23991 + }, + { + "epoch": 0.5624102627214497, + "grad_norm": 0.8015005588531494, + "learning_rate": 8.054921972760088e-05, + "loss": 0.1407, + "step": 23992 + }, + { + "epoch": 0.5624337042962546, + "grad_norm": 0.920569658279419, + "learning_rate": 8.054199430900899e-05, + "loss": 0.04, + "step": 23993 + }, + { + "epoch": 0.5624571458710597, + "grad_norm": 0.6720635890960693, + "learning_rate": 8.053476899599676e-05, + "loss": 0.1479, + "step": 23994 + }, + { + "epoch": 0.5624805874458646, + "grad_norm": 0.23335346579551697, + "learning_rate": 8.052754378860344e-05, + "loss": 0.0388, + "step": 23995 + }, + { + "epoch": 0.5625040290206696, + "grad_norm": 0.5692105293273926, + "learning_rate": 8.052031868686822e-05, + "loss": 0.0784, + "step": 23996 + }, + { + "epoch": 0.5625274705954746, + "grad_norm": 0.30019283294677734, + "learning_rate": 8.051309369083029e-05, + "loss": 0.3356, + "step": 23997 + }, + { + "epoch": 0.5625509121702796, + "grad_norm": 0.24417610466480255, + "learning_rate": 8.050586880052888e-05, + "loss": 0.0397, + "step": 23998 + }, + { + "epoch": 0.5625743537450846, + "grad_norm": 0.5724149346351624, + "learning_rate": 8.049864401600315e-05, + "loss": 0.6135, + "step": 23999 + }, + { + "epoch": 0.5625977953198896, + "grad_norm": 0.35200682282447815, + "learning_rate": 8.049141933729233e-05, + "loss": 0.0978, + "step": 24000 + }, + { + "epoch": 0.5626212368946946, + "grad_norm": 0.6742205023765564, + "learning_rate": 8.048419476443563e-05, + "loss": 0.1082, + "step": 24001 + }, + { + "epoch": 0.5626446784694996, + "grad_norm": 0.5924486517906189, + "learning_rate": 8.047697029747225e-05, + "loss": 0.151, + "step": 24002 + }, + { + "epoch": 0.5626681200443046, + "grad_norm": 0.44351816177368164, + "learning_rate": 8.046974593644134e-05, + "loss": 0.1058, + "step": 24003 + }, + { + "epoch": 0.5626915616191096, + "grad_norm": 0.3242492973804474, + "learning_rate": 8.046252168138212e-05, + "loss": 0.0681, + "step": 24004 + }, + { + "epoch": 0.5627150031939145, + "grad_norm": 0.5493506193161011, + "learning_rate": 8.045529753233385e-05, + "loss": 0.0963, + "step": 24005 + }, + { + "epoch": 0.5627384447687196, + "grad_norm": 0.14071443676948547, + "learning_rate": 8.044807348933568e-05, + "loss": 0.0156, + "step": 24006 + }, + { + "epoch": 0.5627618863435245, + "grad_norm": 0.46607401967048645, + "learning_rate": 8.04408495524268e-05, + "loss": 0.0943, + "step": 24007 + }, + { + "epoch": 0.5627853279183296, + "grad_norm": 0.14130628108978271, + "learning_rate": 8.043362572164643e-05, + "loss": 0.0219, + "step": 24008 + }, + { + "epoch": 0.5628087694931345, + "grad_norm": 0.10481686145067215, + "learning_rate": 8.042640199703377e-05, + "loss": 0.0182, + "step": 24009 + }, + { + "epoch": 0.5628322110679396, + "grad_norm": 0.2274448573589325, + "learning_rate": 8.041917837862798e-05, + "loss": 0.0373, + "step": 24010 + }, + { + "epoch": 0.5628556526427445, + "grad_norm": 0.49563759565353394, + "learning_rate": 8.04119548664683e-05, + "loss": 0.0632, + "step": 24011 + }, + { + "epoch": 0.5628790942175496, + "grad_norm": 0.18504630029201508, + "learning_rate": 8.04047314605939e-05, + "loss": 0.0474, + "step": 24012 + }, + { + "epoch": 0.5629025357923545, + "grad_norm": 0.45424333214759827, + "learning_rate": 8.039750816104396e-05, + "loss": 0.0828, + "step": 24013 + }, + { + "epoch": 0.5629259773671595, + "grad_norm": 0.31980669498443604, + "learning_rate": 8.039028496785771e-05, + "loss": 0.0397, + "step": 24014 + }, + { + "epoch": 0.5629494189419645, + "grad_norm": 0.44165974855422974, + "learning_rate": 8.03830618810743e-05, + "loss": 0.0692, + "step": 24015 + }, + { + "epoch": 0.5629728605167695, + "grad_norm": 0.5939223170280457, + "learning_rate": 8.037583890073295e-05, + "loss": 0.1483, + "step": 24016 + }, + { + "epoch": 0.5629963020915745, + "grad_norm": 0.5181925892829895, + "learning_rate": 8.036861602687286e-05, + "loss": 0.075, + "step": 24017 + }, + { + "epoch": 0.5630197436663795, + "grad_norm": 0.5532749891281128, + "learning_rate": 8.036139325953323e-05, + "loss": 0.1378, + "step": 24018 + }, + { + "epoch": 0.5630431852411845, + "grad_norm": 0.653038501739502, + "learning_rate": 8.03541705987532e-05, + "loss": 0.068, + "step": 24019 + }, + { + "epoch": 0.5630666268159895, + "grad_norm": 0.5362669825553894, + "learning_rate": 8.034694804457203e-05, + "loss": 0.0648, + "step": 24020 + }, + { + "epoch": 0.5630900683907945, + "grad_norm": 0.046084675937891006, + "learning_rate": 8.033972559702886e-05, + "loss": 0.0066, + "step": 24021 + }, + { + "epoch": 0.5631135099655995, + "grad_norm": 0.18545375764369965, + "learning_rate": 8.033250325616287e-05, + "loss": 0.0394, + "step": 24022 + }, + { + "epoch": 0.5631369515404044, + "grad_norm": 0.34905293583869934, + "learning_rate": 8.03252810220133e-05, + "loss": 0.0381, + "step": 24023 + }, + { + "epoch": 0.5631603931152095, + "grad_norm": 0.3119356334209442, + "learning_rate": 8.03180588946193e-05, + "loss": 0.0838, + "step": 24024 + }, + { + "epoch": 0.5631838346900144, + "grad_norm": 0.36174631118774414, + "learning_rate": 8.031083687402006e-05, + "loss": 0.0696, + "step": 24025 + }, + { + "epoch": 0.5632072762648195, + "grad_norm": 0.4851083755493164, + "learning_rate": 8.030361496025476e-05, + "loss": 0.0536, + "step": 24026 + }, + { + "epoch": 0.5632307178396244, + "grad_norm": 0.8241202235221863, + "learning_rate": 8.029639315336264e-05, + "loss": 0.6844, + "step": 24027 + }, + { + "epoch": 0.5632541594144295, + "grad_norm": 0.603424072265625, + "learning_rate": 8.028917145338277e-05, + "loss": 0.1118, + "step": 24028 + }, + { + "epoch": 0.5632776009892344, + "grad_norm": 0.16777142882347107, + "learning_rate": 8.028194986035446e-05, + "loss": 0.0175, + "step": 24029 + }, + { + "epoch": 0.5633010425640395, + "grad_norm": 0.4785551428794861, + "learning_rate": 8.027472837431685e-05, + "loss": 0.1263, + "step": 24030 + }, + { + "epoch": 0.5633244841388444, + "grad_norm": 0.4084182679653168, + "learning_rate": 8.026750699530912e-05, + "loss": 0.1073, + "step": 24031 + }, + { + "epoch": 0.5633479257136494, + "grad_norm": 0.3469308018684387, + "learning_rate": 8.026028572337043e-05, + "loss": 0.0398, + "step": 24032 + }, + { + "epoch": 0.5633713672884544, + "grad_norm": 0.14185792207717896, + "learning_rate": 8.025306455854e-05, + "loss": 0.0115, + "step": 24033 + }, + { + "epoch": 0.5633948088632594, + "grad_norm": 0.3851804733276367, + "learning_rate": 8.0245843500857e-05, + "loss": 0.0593, + "step": 24034 + }, + { + "epoch": 0.5634182504380645, + "grad_norm": 0.13739652931690216, + "learning_rate": 8.023862255036059e-05, + "loss": 0.0362, + "step": 24035 + }, + { + "epoch": 0.5634416920128694, + "grad_norm": 0.5941241979598999, + "learning_rate": 8.023140170709e-05, + "loss": 0.0834, + "step": 24036 + }, + { + "epoch": 0.5634651335876745, + "grad_norm": 0.7231615781784058, + "learning_rate": 8.022418097108436e-05, + "loss": 0.1093, + "step": 24037 + }, + { + "epoch": 0.5634885751624794, + "grad_norm": 0.4735304117202759, + "learning_rate": 8.021696034238287e-05, + "loss": 0.0841, + "step": 24038 + }, + { + "epoch": 0.5635120167372845, + "grad_norm": 0.41381004452705383, + "learning_rate": 8.02097398210247e-05, + "loss": 0.0832, + "step": 24039 + }, + { + "epoch": 0.5635354583120894, + "grad_norm": 0.16826845705509186, + "learning_rate": 8.020251940704903e-05, + "loss": 0.0333, + "step": 24040 + }, + { + "epoch": 0.5635588998868944, + "grad_norm": 0.03464260697364807, + "learning_rate": 8.019529910049506e-05, + "loss": 0.0022, + "step": 24041 + }, + { + "epoch": 0.5635823414616994, + "grad_norm": 0.18405310809612274, + "learning_rate": 8.018807890140196e-05, + "loss": 0.0412, + "step": 24042 + }, + { + "epoch": 0.5636057830365044, + "grad_norm": 0.14163081347942352, + "learning_rate": 8.018085880980891e-05, + "loss": 0.0286, + "step": 24043 + }, + { + "epoch": 0.5636292246113094, + "grad_norm": 0.12733912467956543, + "learning_rate": 8.017363882575508e-05, + "loss": 0.0152, + "step": 24044 + }, + { + "epoch": 0.5636526661861144, + "grad_norm": 0.6525368094444275, + "learning_rate": 8.016641894927962e-05, + "loss": 0.0976, + "step": 24045 + }, + { + "epoch": 0.5636761077609194, + "grad_norm": 0.5226966142654419, + "learning_rate": 8.015919918042174e-05, + "loss": 0.0745, + "step": 24046 + }, + { + "epoch": 0.5636995493357244, + "grad_norm": 0.15267260372638702, + "learning_rate": 8.015197951922062e-05, + "loss": 0.0318, + "step": 24047 + }, + { + "epoch": 0.5637229909105294, + "grad_norm": 0.5642207264900208, + "learning_rate": 8.014475996571538e-05, + "loss": 0.1141, + "step": 24048 + }, + { + "epoch": 0.5637464324853344, + "grad_norm": 0.4782247543334961, + "learning_rate": 8.013754051994526e-05, + "loss": 0.0873, + "step": 24049 + }, + { + "epoch": 0.5637698740601393, + "grad_norm": 0.4141899347305298, + "learning_rate": 8.013032118194941e-05, + "loss": 0.0456, + "step": 24050 + }, + { + "epoch": 0.5637933156349444, + "grad_norm": 0.6457732319831848, + "learning_rate": 8.012310195176697e-05, + "loss": 0.113, + "step": 24051 + }, + { + "epoch": 0.5638167572097493, + "grad_norm": 0.35310062766075134, + "learning_rate": 8.011588282943712e-05, + "loss": 0.0605, + "step": 24052 + }, + { + "epoch": 0.5638401987845544, + "grad_norm": 0.08337496221065521, + "learning_rate": 8.010866381499908e-05, + "loss": 0.0133, + "step": 24053 + }, + { + "epoch": 0.5638636403593593, + "grad_norm": 0.5802904963493347, + "learning_rate": 8.010144490849199e-05, + "loss": 0.1074, + "step": 24054 + }, + { + "epoch": 0.5638870819341644, + "grad_norm": 0.3733402192592621, + "learning_rate": 8.0094226109955e-05, + "loss": 0.0783, + "step": 24055 + }, + { + "epoch": 0.5639105235089693, + "grad_norm": 0.3096669316291809, + "learning_rate": 8.008700741942733e-05, + "loss": 0.0886, + "step": 24056 + }, + { + "epoch": 0.5639339650837744, + "grad_norm": 0.4748309254646301, + "learning_rate": 8.007978883694809e-05, + "loss": 0.0921, + "step": 24057 + }, + { + "epoch": 0.5639574066585793, + "grad_norm": 0.4357396364212036, + "learning_rate": 8.00725703625565e-05, + "loss": 0.0423, + "step": 24058 + }, + { + "epoch": 0.5639808482333843, + "grad_norm": 0.2863847613334656, + "learning_rate": 8.00653519962917e-05, + "loss": 0.0346, + "step": 24059 + }, + { + "epoch": 0.5640042898081893, + "grad_norm": 0.27871859073638916, + "learning_rate": 8.005813373819284e-05, + "loss": 0.0389, + "step": 24060 + }, + { + "epoch": 0.5640277313829943, + "grad_norm": 0.5513253808021545, + "learning_rate": 8.005091558829913e-05, + "loss": 0.6463, + "step": 24061 + }, + { + "epoch": 0.5640511729577993, + "grad_norm": 0.2261485755443573, + "learning_rate": 8.00436975466497e-05, + "loss": 0.0289, + "step": 24062 + }, + { + "epoch": 0.5640746145326043, + "grad_norm": 0.7436267137527466, + "learning_rate": 8.003647961328373e-05, + "loss": 0.1254, + "step": 24063 + }, + { + "epoch": 0.5640980561074093, + "grad_norm": 0.4724296033382416, + "learning_rate": 8.002926178824035e-05, + "loss": 0.0982, + "step": 24064 + }, + { + "epoch": 0.5641214976822143, + "grad_norm": 0.25680503249168396, + "learning_rate": 8.00220440715588e-05, + "loss": 0.0558, + "step": 24065 + }, + { + "epoch": 0.5641449392570193, + "grad_norm": 0.5142934322357178, + "learning_rate": 8.001482646327821e-05, + "loss": 0.0231, + "step": 24066 + }, + { + "epoch": 0.5641683808318243, + "grad_norm": 0.4257785677909851, + "learning_rate": 8.000760896343768e-05, + "loss": 0.0548, + "step": 24067 + }, + { + "epoch": 0.5641918224066292, + "grad_norm": 0.568554162979126, + "learning_rate": 8.000039157207649e-05, + "loss": 0.1428, + "step": 24068 + }, + { + "epoch": 0.5642152639814343, + "grad_norm": 0.7799170613288879, + "learning_rate": 7.999317428923371e-05, + "loss": 0.1003, + "step": 24069 + }, + { + "epoch": 0.5642387055562392, + "grad_norm": 0.45673125982284546, + "learning_rate": 7.998595711494852e-05, + "loss": 0.0466, + "step": 24070 + }, + { + "epoch": 0.5642621471310443, + "grad_norm": 0.19208312034606934, + "learning_rate": 7.99787400492601e-05, + "loss": 0.0159, + "step": 24071 + }, + { + "epoch": 0.5642855887058492, + "grad_norm": 0.3863498270511627, + "learning_rate": 7.99715230922076e-05, + "loss": 0.0998, + "step": 24072 + }, + { + "epoch": 0.5643090302806543, + "grad_norm": 0.26170599460601807, + "learning_rate": 7.996430624383016e-05, + "loss": 0.0362, + "step": 24073 + }, + { + "epoch": 0.5643324718554592, + "grad_norm": 0.2896185517311096, + "learning_rate": 7.995708950416698e-05, + "loss": 0.0291, + "step": 24074 + }, + { + "epoch": 0.5643559134302643, + "grad_norm": 0.5468422770500183, + "learning_rate": 7.99498728732572e-05, + "loss": 0.0992, + "step": 24075 + }, + { + "epoch": 0.5643793550050692, + "grad_norm": 0.35116297006607056, + "learning_rate": 7.994265635113996e-05, + "loss": 0.0817, + "step": 24076 + }, + { + "epoch": 0.5644027965798742, + "grad_norm": 0.1711767017841339, + "learning_rate": 7.993543993785439e-05, + "loss": 0.0274, + "step": 24077 + }, + { + "epoch": 0.5644262381546792, + "grad_norm": 0.3989582061767578, + "learning_rate": 7.992822363343975e-05, + "loss": 0.0466, + "step": 24078 + }, + { + "epoch": 0.5644496797294842, + "grad_norm": 0.4107147455215454, + "learning_rate": 7.992100743793511e-05, + "loss": 0.0871, + "step": 24079 + }, + { + "epoch": 0.5644731213042892, + "grad_norm": 0.4933415651321411, + "learning_rate": 7.991379135137964e-05, + "loss": 0.0716, + "step": 24080 + }, + { + "epoch": 0.5644965628790942, + "grad_norm": 0.3571830689907074, + "learning_rate": 7.990657537381252e-05, + "loss": 0.3111, + "step": 24081 + }, + { + "epoch": 0.5645200044538992, + "grad_norm": 0.44078752398490906, + "learning_rate": 7.98993595052729e-05, + "loss": 0.0669, + "step": 24082 + }, + { + "epoch": 0.5645434460287042, + "grad_norm": 0.17250105738639832, + "learning_rate": 7.989214374579989e-05, + "loss": 0.0374, + "step": 24083 + }, + { + "epoch": 0.5645668876035091, + "grad_norm": 0.39043787121772766, + "learning_rate": 7.988492809543268e-05, + "loss": 0.0719, + "step": 24084 + }, + { + "epoch": 0.5645903291783142, + "grad_norm": 0.40970832109451294, + "learning_rate": 7.987771255421044e-05, + "loss": 0.0593, + "step": 24085 + }, + { + "epoch": 0.5646137707531192, + "grad_norm": 0.1236988976597786, + "learning_rate": 7.987049712217226e-05, + "loss": 0.0298, + "step": 24086 + }, + { + "epoch": 0.5646372123279242, + "grad_norm": 0.5044126510620117, + "learning_rate": 7.986328179935736e-05, + "loss": 0.0713, + "step": 24087 + }, + { + "epoch": 0.5646606539027292, + "grad_norm": 0.4874870777130127, + "learning_rate": 7.985606658580485e-05, + "loss": 0.0468, + "step": 24088 + }, + { + "epoch": 0.5646840954775342, + "grad_norm": 0.46660539507865906, + "learning_rate": 7.984885148155386e-05, + "loss": 0.0558, + "step": 24089 + }, + { + "epoch": 0.5647075370523392, + "grad_norm": 0.12302099168300629, + "learning_rate": 7.984163648664358e-05, + "loss": 0.0176, + "step": 24090 + }, + { + "epoch": 0.5647309786271442, + "grad_norm": 0.13287578523159027, + "learning_rate": 7.983442160111316e-05, + "loss": 0.0276, + "step": 24091 + }, + { + "epoch": 0.5647544202019492, + "grad_norm": 0.17661698162555695, + "learning_rate": 7.982720682500175e-05, + "loss": 0.0446, + "step": 24092 + }, + { + "epoch": 0.5647778617767542, + "grad_norm": 0.1778423935174942, + "learning_rate": 7.981999215834846e-05, + "loss": 0.0265, + "step": 24093 + }, + { + "epoch": 0.5648013033515592, + "grad_norm": 1.2217127084732056, + "learning_rate": 7.981277760119248e-05, + "loss": 0.2316, + "step": 24094 + }, + { + "epoch": 0.5648247449263641, + "grad_norm": 0.3941537141799927, + "learning_rate": 7.98055631535729e-05, + "loss": 0.0793, + "step": 24095 + }, + { + "epoch": 0.5648481865011692, + "grad_norm": 0.5470201969146729, + "learning_rate": 7.979834881552895e-05, + "loss": 0.1444, + "step": 24096 + }, + { + "epoch": 0.5648716280759741, + "grad_norm": 0.7094164490699768, + "learning_rate": 7.979113458709971e-05, + "loss": 1.0196, + "step": 24097 + }, + { + "epoch": 0.5648950696507792, + "grad_norm": 0.348380446434021, + "learning_rate": 7.978392046832433e-05, + "loss": 0.0231, + "step": 24098 + }, + { + "epoch": 0.5649185112255841, + "grad_norm": 0.3978787362575531, + "learning_rate": 7.977670645924197e-05, + "loss": 0.0488, + "step": 24099 + }, + { + "epoch": 0.5649419528003892, + "grad_norm": 0.6166402101516724, + "learning_rate": 7.976949255989177e-05, + "loss": 0.7205, + "step": 24100 + }, + { + "epoch": 0.5649653943751941, + "grad_norm": 0.6085366606712341, + "learning_rate": 7.976227877031285e-05, + "loss": 0.5435, + "step": 24101 + }, + { + "epoch": 0.5649888359499992, + "grad_norm": 0.37308692932128906, + "learning_rate": 7.975506509054439e-05, + "loss": 0.0335, + "step": 24102 + }, + { + "epoch": 0.5650122775248041, + "grad_norm": 0.4644087851047516, + "learning_rate": 7.974785152062554e-05, + "loss": 0.0803, + "step": 24103 + }, + { + "epoch": 0.5650357190996091, + "grad_norm": 0.10723181068897247, + "learning_rate": 7.97406380605954e-05, + "loss": 0.0091, + "step": 24104 + }, + { + "epoch": 0.5650591606744141, + "grad_norm": 0.4966326057910919, + "learning_rate": 7.973342471049312e-05, + "loss": 0.0752, + "step": 24105 + }, + { + "epoch": 0.5650826022492191, + "grad_norm": 0.09987346082925797, + "learning_rate": 7.972621147035786e-05, + "loss": 0.0095, + "step": 24106 + }, + { + "epoch": 0.5651060438240241, + "grad_norm": 0.6076420545578003, + "learning_rate": 7.971899834022876e-05, + "loss": 0.1708, + "step": 24107 + }, + { + "epoch": 0.5651294853988291, + "grad_norm": 0.5664852857589722, + "learning_rate": 7.971178532014492e-05, + "loss": 0.0775, + "step": 24108 + }, + { + "epoch": 0.5651529269736341, + "grad_norm": 0.3875032663345337, + "learning_rate": 7.970457241014552e-05, + "loss": 0.0435, + "step": 24109 + }, + { + "epoch": 0.5651763685484391, + "grad_norm": 0.14913682639598846, + "learning_rate": 7.969735961026967e-05, + "loss": 0.0406, + "step": 24110 + }, + { + "epoch": 0.565199810123244, + "grad_norm": 0.468273401260376, + "learning_rate": 7.969014692055652e-05, + "loss": 0.0669, + "step": 24111 + }, + { + "epoch": 0.5652232516980491, + "grad_norm": 0.40113723278045654, + "learning_rate": 7.968293434104521e-05, + "loss": 0.0978, + "step": 24112 + }, + { + "epoch": 0.565246693272854, + "grad_norm": 0.41250917315483093, + "learning_rate": 7.967572187177484e-05, + "loss": 0.0619, + "step": 24113 + }, + { + "epoch": 0.5652701348476591, + "grad_norm": 1.3087024688720703, + "learning_rate": 7.96685095127846e-05, + "loss": 0.1514, + "step": 24114 + }, + { + "epoch": 0.565293576422464, + "grad_norm": 0.5630614161491394, + "learning_rate": 7.966129726411357e-05, + "loss": 0.1389, + "step": 24115 + }, + { + "epoch": 0.5653170179972691, + "grad_norm": 0.31829407811164856, + "learning_rate": 7.965408512580096e-05, + "loss": 0.0725, + "step": 24116 + }, + { + "epoch": 0.565340459572074, + "grad_norm": 0.46315997838974, + "learning_rate": 7.964687309788585e-05, + "loss": 0.0597, + "step": 24117 + }, + { + "epoch": 0.5653639011468791, + "grad_norm": 0.3911101520061493, + "learning_rate": 7.963966118040734e-05, + "loss": 0.0765, + "step": 24118 + }, + { + "epoch": 0.565387342721684, + "grad_norm": 0.20436476171016693, + "learning_rate": 7.963244937340464e-05, + "loss": 0.0398, + "step": 24119 + }, + { + "epoch": 0.565410784296489, + "grad_norm": 0.4539022147655487, + "learning_rate": 7.962523767691684e-05, + "loss": 0.1175, + "step": 24120 + }, + { + "epoch": 0.565434225871294, + "grad_norm": 0.7376295328140259, + "learning_rate": 7.961802609098305e-05, + "loss": 0.1024, + "step": 24121 + }, + { + "epoch": 0.565457667446099, + "grad_norm": 0.6689136028289795, + "learning_rate": 7.961081461564244e-05, + "loss": 0.1401, + "step": 24122 + }, + { + "epoch": 0.565481109020904, + "grad_norm": 0.33664390444755554, + "learning_rate": 7.960360325093413e-05, + "loss": 0.047, + "step": 24123 + }, + { + "epoch": 0.565504550595709, + "grad_norm": 0.3985865116119385, + "learning_rate": 7.959639199689721e-05, + "loss": 0.0775, + "step": 24124 + }, + { + "epoch": 0.565527992170514, + "grad_norm": 0.12090334296226501, + "learning_rate": 7.958918085357083e-05, + "loss": 0.0153, + "step": 24125 + }, + { + "epoch": 0.565551433745319, + "grad_norm": 0.39156201481819153, + "learning_rate": 7.958196982099418e-05, + "loss": 0.0707, + "step": 24126 + }, + { + "epoch": 0.565574875320124, + "grad_norm": 0.4098607897758484, + "learning_rate": 7.957475889920632e-05, + "loss": 0.0619, + "step": 24127 + }, + { + "epoch": 0.565598316894929, + "grad_norm": 0.4833936393260956, + "learning_rate": 7.956754808824638e-05, + "loss": 0.1099, + "step": 24128 + }, + { + "epoch": 0.565621758469734, + "grad_norm": 0.577497661113739, + "learning_rate": 7.956033738815351e-05, + "loss": 0.0779, + "step": 24129 + }, + { + "epoch": 0.565645200044539, + "grad_norm": 0.4037444591522217, + "learning_rate": 7.955312679896682e-05, + "loss": 0.0358, + "step": 24130 + }, + { + "epoch": 0.5656686416193439, + "grad_norm": 0.5154720544815063, + "learning_rate": 7.954591632072543e-05, + "loss": 0.0689, + "step": 24131 + }, + { + "epoch": 0.565692083194149, + "grad_norm": 0.30449578166007996, + "learning_rate": 7.95387059534685e-05, + "loss": 0.069, + "step": 24132 + }, + { + "epoch": 0.5657155247689539, + "grad_norm": 0.6387523412704468, + "learning_rate": 7.95314956972351e-05, + "loss": 0.1221, + "step": 24133 + }, + { + "epoch": 0.565738966343759, + "grad_norm": 0.24735307693481445, + "learning_rate": 7.95242855520644e-05, + "loss": 0.2702, + "step": 24134 + }, + { + "epoch": 0.5657624079185639, + "grad_norm": 0.4112478494644165, + "learning_rate": 7.95170755179955e-05, + "loss": 0.0641, + "step": 24135 + }, + { + "epoch": 0.565785849493369, + "grad_norm": 0.7014451622962952, + "learning_rate": 7.95098655950675e-05, + "loss": 0.8923, + "step": 24136 + }, + { + "epoch": 0.565809291068174, + "grad_norm": 0.462078332901001, + "learning_rate": 7.950265578331955e-05, + "loss": 0.1229, + "step": 24137 + }, + { + "epoch": 0.565832732642979, + "grad_norm": 0.12296032905578613, + "learning_rate": 7.949544608279077e-05, + "loss": 0.0244, + "step": 24138 + }, + { + "epoch": 0.565856174217784, + "grad_norm": 0.29246267676353455, + "learning_rate": 7.948823649352029e-05, + "loss": 0.0637, + "step": 24139 + }, + { + "epoch": 0.5658796157925889, + "grad_norm": 0.30137720704078674, + "learning_rate": 7.94810270155472e-05, + "loss": 0.0588, + "step": 24140 + }, + { + "epoch": 0.565903057367394, + "grad_norm": 0.11718083918094635, + "learning_rate": 7.947381764891066e-05, + "loss": 0.0197, + "step": 24141 + }, + { + "epoch": 0.5659264989421989, + "grad_norm": 0.5705751180648804, + "learning_rate": 7.946660839364975e-05, + "loss": 0.0662, + "step": 24142 + }, + { + "epoch": 0.565949940517004, + "grad_norm": 0.6076368093490601, + "learning_rate": 7.945939924980357e-05, + "loss": 0.1461, + "step": 24143 + }, + { + "epoch": 0.5659733820918089, + "grad_norm": 0.0753263533115387, + "learning_rate": 7.945219021741132e-05, + "loss": 0.0097, + "step": 24144 + }, + { + "epoch": 0.565996823666614, + "grad_norm": 0.5220951437950134, + "learning_rate": 7.944498129651203e-05, + "loss": 0.0952, + "step": 24145 + }, + { + "epoch": 0.5660202652414189, + "grad_norm": 0.38679924607276917, + "learning_rate": 7.943777248714485e-05, + "loss": 0.0986, + "step": 24146 + }, + { + "epoch": 0.566043706816224, + "grad_norm": 0.10806816071271896, + "learning_rate": 7.943056378934891e-05, + "loss": 0.018, + "step": 24147 + }, + { + "epoch": 0.5660671483910289, + "grad_norm": 0.6082363724708557, + "learning_rate": 7.942335520316331e-05, + "loss": 0.0658, + "step": 24148 + }, + { + "epoch": 0.566090589965834, + "grad_norm": 0.3657292425632477, + "learning_rate": 7.941614672862712e-05, + "loss": 0.089, + "step": 24149 + }, + { + "epoch": 0.5661140315406389, + "grad_norm": 0.43630874156951904, + "learning_rate": 7.940893836577951e-05, + "loss": 0.0808, + "step": 24150 + }, + { + "epoch": 0.5661374731154439, + "grad_norm": 0.6003715395927429, + "learning_rate": 7.940173011465963e-05, + "loss": 0.6867, + "step": 24151 + }, + { + "epoch": 0.5661609146902489, + "grad_norm": 0.38210368156433105, + "learning_rate": 7.939452197530652e-05, + "loss": 0.0415, + "step": 24152 + }, + { + "epoch": 0.5661843562650539, + "grad_norm": 0.584606409072876, + "learning_rate": 7.938731394775927e-05, + "loss": 0.1126, + "step": 24153 + }, + { + "epoch": 0.5662077978398589, + "grad_norm": 0.8037856221199036, + "learning_rate": 7.938010603205709e-05, + "loss": 0.276, + "step": 24154 + }, + { + "epoch": 0.5662312394146639, + "grad_norm": 0.25314342975616455, + "learning_rate": 7.937289822823902e-05, + "loss": 0.0338, + "step": 24155 + }, + { + "epoch": 0.5662546809894689, + "grad_norm": 0.705482542514801, + "learning_rate": 7.936569053634417e-05, + "loss": 0.1479, + "step": 24156 + }, + { + "epoch": 0.5662781225642739, + "grad_norm": 0.6048014163970947, + "learning_rate": 7.935848295641169e-05, + "loss": 0.5905, + "step": 24157 + }, + { + "epoch": 0.5663015641390788, + "grad_norm": 0.1153358444571495, + "learning_rate": 7.935127548848064e-05, + "loss": 0.0162, + "step": 24158 + }, + { + "epoch": 0.5663250057138839, + "grad_norm": 0.870203971862793, + "learning_rate": 7.934406813259014e-05, + "loss": 0.0901, + "step": 24159 + }, + { + "epoch": 0.5663484472886888, + "grad_norm": 0.6427084803581238, + "learning_rate": 7.933686088877933e-05, + "loss": 0.1678, + "step": 24160 + }, + { + "epoch": 0.5663718888634939, + "grad_norm": 0.8027223348617554, + "learning_rate": 7.93296537570873e-05, + "loss": 0.1586, + "step": 24161 + }, + { + "epoch": 0.5663953304382988, + "grad_norm": 0.3988144099712372, + "learning_rate": 7.932244673755309e-05, + "loss": 0.0639, + "step": 24162 + }, + { + "epoch": 0.5664187720131039, + "grad_norm": 0.7967145442962646, + "learning_rate": 7.931523983021589e-05, + "loss": 0.1326, + "step": 24163 + }, + { + "epoch": 0.5664422135879088, + "grad_norm": 0.652249276638031, + "learning_rate": 7.930803303511481e-05, + "loss": 0.1006, + "step": 24164 + }, + { + "epoch": 0.5664656551627139, + "grad_norm": 0.13256052136421204, + "learning_rate": 7.930082635228893e-05, + "loss": 0.0269, + "step": 24165 + }, + { + "epoch": 0.5664890967375188, + "grad_norm": 0.3796399235725403, + "learning_rate": 7.929361978177732e-05, + "loss": 0.0495, + "step": 24166 + }, + { + "epoch": 0.5665125383123238, + "grad_norm": 0.41439327597618103, + "learning_rate": 7.928641332361912e-05, + "loss": 0.052, + "step": 24167 + }, + { + "epoch": 0.5665359798871288, + "grad_norm": 0.18155330419540405, + "learning_rate": 7.927920697785345e-05, + "loss": 0.0347, + "step": 24168 + }, + { + "epoch": 0.5665594214619338, + "grad_norm": 0.41204091906547546, + "learning_rate": 7.927200074451935e-05, + "loss": 0.06, + "step": 24169 + }, + { + "epoch": 0.5665828630367388, + "grad_norm": 0.1210220605134964, + "learning_rate": 7.926479462365597e-05, + "loss": 0.0346, + "step": 24170 + }, + { + "epoch": 0.5666063046115438, + "grad_norm": 0.12405480444431305, + "learning_rate": 7.925758861530238e-05, + "loss": 0.0117, + "step": 24171 + }, + { + "epoch": 0.5666297461863488, + "grad_norm": 0.3413962423801422, + "learning_rate": 7.925038271949772e-05, + "loss": 0.0387, + "step": 24172 + }, + { + "epoch": 0.5666531877611538, + "grad_norm": 0.775536060333252, + "learning_rate": 7.924317693628107e-05, + "loss": 0.1006, + "step": 24173 + }, + { + "epoch": 0.5666766293359587, + "grad_norm": 0.6664732098579407, + "learning_rate": 7.923597126569149e-05, + "loss": 0.0679, + "step": 24174 + }, + { + "epoch": 0.5667000709107638, + "grad_norm": 0.5348911881446838, + "learning_rate": 7.922876570776813e-05, + "loss": 0.0617, + "step": 24175 + }, + { + "epoch": 0.5667235124855687, + "grad_norm": 1.0966455936431885, + "learning_rate": 7.922156026255009e-05, + "loss": 0.1032, + "step": 24176 + }, + { + "epoch": 0.5667469540603738, + "grad_norm": 0.28416603803634644, + "learning_rate": 7.921435493007645e-05, + "loss": 0.07, + "step": 24177 + }, + { + "epoch": 0.5667703956351787, + "grad_norm": 0.13710545003414154, + "learning_rate": 7.920714971038628e-05, + "loss": 0.0239, + "step": 24178 + }, + { + "epoch": 0.5667938372099838, + "grad_norm": 0.5631291270256042, + "learning_rate": 7.919994460351872e-05, + "loss": 0.0991, + "step": 24179 + }, + { + "epoch": 0.5668172787847887, + "grad_norm": 0.18886198103427887, + "learning_rate": 7.919273960951285e-05, + "loss": 0.02, + "step": 24180 + }, + { + "epoch": 0.5668407203595938, + "grad_norm": 0.5164439678192139, + "learning_rate": 7.918553472840773e-05, + "loss": 0.1037, + "step": 24181 + }, + { + "epoch": 0.5668641619343987, + "grad_norm": 0.18290339410305023, + "learning_rate": 7.917832996024253e-05, + "loss": 0.0275, + "step": 24182 + }, + { + "epoch": 0.5668876035092038, + "grad_norm": 0.1561315655708313, + "learning_rate": 7.917112530505627e-05, + "loss": 0.0332, + "step": 24183 + }, + { + "epoch": 0.5669110450840087, + "grad_norm": 0.18454372882843018, + "learning_rate": 7.916392076288805e-05, + "loss": 0.0378, + "step": 24184 + }, + { + "epoch": 0.5669344866588137, + "grad_norm": 0.4634994864463806, + "learning_rate": 7.915671633377702e-05, + "loss": 0.0653, + "step": 24185 + }, + { + "epoch": 0.5669579282336187, + "grad_norm": 0.5397464632987976, + "learning_rate": 7.914951201776217e-05, + "loss": 0.0802, + "step": 24186 + }, + { + "epoch": 0.5669813698084237, + "grad_norm": 0.50345778465271, + "learning_rate": 7.914230781488272e-05, + "loss": 0.0847, + "step": 24187 + }, + { + "epoch": 0.5670048113832288, + "grad_norm": 0.5330321192741394, + "learning_rate": 7.913510372517766e-05, + "loss": 0.1276, + "step": 24188 + }, + { + "epoch": 0.5670282529580337, + "grad_norm": 0.5702099204063416, + "learning_rate": 7.912789974868614e-05, + "loss": 0.422, + "step": 24189 + }, + { + "epoch": 0.5670516945328388, + "grad_norm": 0.19153062999248505, + "learning_rate": 7.912069588544722e-05, + "loss": 0.0337, + "step": 24190 + }, + { + "epoch": 0.5670751361076437, + "grad_norm": 0.2145940512418747, + "learning_rate": 7.911349213549997e-05, + "loss": 0.031, + "step": 24191 + }, + { + "epoch": 0.5670985776824488, + "grad_norm": 0.5716456770896912, + "learning_rate": 7.910628849888351e-05, + "loss": 0.1658, + "step": 24192 + }, + { + "epoch": 0.5671220192572537, + "grad_norm": 0.4976380169391632, + "learning_rate": 7.909908497563693e-05, + "loss": 0.5013, + "step": 24193 + }, + { + "epoch": 0.5671454608320587, + "grad_norm": 0.442149817943573, + "learning_rate": 7.909188156579927e-05, + "loss": 0.0759, + "step": 24194 + }, + { + "epoch": 0.5671689024068637, + "grad_norm": 0.6350727081298828, + "learning_rate": 7.908467826940969e-05, + "loss": 0.0842, + "step": 24195 + }, + { + "epoch": 0.5671923439816687, + "grad_norm": 0.33119863271713257, + "learning_rate": 7.907747508650722e-05, + "loss": 0.0498, + "step": 24196 + }, + { + "epoch": 0.5672157855564737, + "grad_norm": 0.6066350936889648, + "learning_rate": 7.907027201713093e-05, + "loss": 0.1006, + "step": 24197 + }, + { + "epoch": 0.5672392271312787, + "grad_norm": 0.22750072181224823, + "learning_rate": 7.906306906131992e-05, + "loss": 0.0395, + "step": 24198 + }, + { + "epoch": 0.5672626687060837, + "grad_norm": 0.618281364440918, + "learning_rate": 7.905586621911333e-05, + "loss": 0.0828, + "step": 24199 + }, + { + "epoch": 0.5672861102808887, + "grad_norm": 0.3110274076461792, + "learning_rate": 7.904866349055019e-05, + "loss": 0.0281, + "step": 24200 + }, + { + "epoch": 0.5673095518556937, + "grad_norm": 0.5793966054916382, + "learning_rate": 7.904146087566958e-05, + "loss": 0.155, + "step": 24201 + }, + { + "epoch": 0.5673329934304987, + "grad_norm": 0.7484996914863586, + "learning_rate": 7.90342583745106e-05, + "loss": 0.2008, + "step": 24202 + }, + { + "epoch": 0.5673564350053036, + "grad_norm": 0.2959727942943573, + "learning_rate": 7.902705598711234e-05, + "loss": 0.0506, + "step": 24203 + }, + { + "epoch": 0.5673798765801087, + "grad_norm": 0.09572110325098038, + "learning_rate": 7.901985371351383e-05, + "loss": 0.0065, + "step": 24204 + }, + { + "epoch": 0.5674033181549136, + "grad_norm": 0.5768560171127319, + "learning_rate": 7.901265155375421e-05, + "loss": 0.0782, + "step": 24205 + }, + { + "epoch": 0.5674267597297187, + "grad_norm": 0.9865442514419556, + "learning_rate": 7.900544950787254e-05, + "loss": 0.1384, + "step": 24206 + }, + { + "epoch": 0.5674502013045236, + "grad_norm": 0.9124630093574524, + "learning_rate": 7.899824757590787e-05, + "loss": 0.1061, + "step": 24207 + }, + { + "epoch": 0.5674736428793287, + "grad_norm": 0.7190462946891785, + "learning_rate": 7.89910457578993e-05, + "loss": 0.8303, + "step": 24208 + }, + { + "epoch": 0.5674970844541336, + "grad_norm": 0.30216312408447266, + "learning_rate": 7.898384405388592e-05, + "loss": 0.0595, + "step": 24209 + }, + { + "epoch": 0.5675205260289387, + "grad_norm": 0.3997548818588257, + "learning_rate": 7.897664246390677e-05, + "loss": 0.076, + "step": 24210 + }, + { + "epoch": 0.5675439676037436, + "grad_norm": 0.3492443561553955, + "learning_rate": 7.896944098800094e-05, + "loss": 0.0573, + "step": 24211 + }, + { + "epoch": 0.5675674091785486, + "grad_norm": 0.585982084274292, + "learning_rate": 7.896223962620754e-05, + "loss": 0.6529, + "step": 24212 + }, + { + "epoch": 0.5675908507533536, + "grad_norm": 0.4468802213668823, + "learning_rate": 7.895503837856562e-05, + "loss": 0.0437, + "step": 24213 + }, + { + "epoch": 0.5676142923281586, + "grad_norm": 0.07924781739711761, + "learning_rate": 7.894783724511425e-05, + "loss": 0.0106, + "step": 24214 + }, + { + "epoch": 0.5676377339029636, + "grad_norm": 0.9751037955284119, + "learning_rate": 7.894063622589253e-05, + "loss": 0.0429, + "step": 24215 + }, + { + "epoch": 0.5676611754777686, + "grad_norm": 0.14489680528640747, + "learning_rate": 7.893343532093948e-05, + "loss": 0.0119, + "step": 24216 + }, + { + "epoch": 0.5676846170525736, + "grad_norm": 0.32423996925354004, + "learning_rate": 7.89262345302942e-05, + "loss": 0.0568, + "step": 24217 + }, + { + "epoch": 0.5677080586273786, + "grad_norm": 0.46983590722084045, + "learning_rate": 7.89190338539958e-05, + "loss": 0.1244, + "step": 24218 + }, + { + "epoch": 0.5677315002021835, + "grad_norm": 0.5828786492347717, + "learning_rate": 7.891183329208328e-05, + "loss": 0.0545, + "step": 24219 + }, + { + "epoch": 0.5677549417769886, + "grad_norm": 0.1789000928401947, + "learning_rate": 7.890463284459575e-05, + "loss": 0.0337, + "step": 24220 + }, + { + "epoch": 0.5677783833517935, + "grad_norm": 0.5314632058143616, + "learning_rate": 7.889743251157229e-05, + "loss": 0.0918, + "step": 24221 + }, + { + "epoch": 0.5678018249265986, + "grad_norm": 0.6587109565734863, + "learning_rate": 7.889023229305191e-05, + "loss": 0.1616, + "step": 24222 + }, + { + "epoch": 0.5678252665014035, + "grad_norm": 1.075946569442749, + "learning_rate": 7.888303218907376e-05, + "loss": 0.2582, + "step": 24223 + }, + { + "epoch": 0.5678487080762086, + "grad_norm": 0.552689790725708, + "learning_rate": 7.887583219967687e-05, + "loss": 0.5284, + "step": 24224 + }, + { + "epoch": 0.5678721496510135, + "grad_norm": 0.6839324235916138, + "learning_rate": 7.886863232490032e-05, + "loss": 0.0991, + "step": 24225 + }, + { + "epoch": 0.5678955912258186, + "grad_norm": 0.20126672089099884, + "learning_rate": 7.886143256478314e-05, + "loss": 0.0225, + "step": 24226 + }, + { + "epoch": 0.5679190328006235, + "grad_norm": 0.38098716735839844, + "learning_rate": 7.885423291936445e-05, + "loss": 0.0805, + "step": 24227 + }, + { + "epoch": 0.5679424743754286, + "grad_norm": 0.1531953066587448, + "learning_rate": 7.884703338868327e-05, + "loss": 0.0323, + "step": 24228 + }, + { + "epoch": 0.5679659159502335, + "grad_norm": 0.4139905273914337, + "learning_rate": 7.883983397277868e-05, + "loss": 0.0807, + "step": 24229 + }, + { + "epoch": 0.5679893575250385, + "grad_norm": 0.4311801791191101, + "learning_rate": 7.883263467168975e-05, + "loss": 0.0772, + "step": 24230 + }, + { + "epoch": 0.5680127990998435, + "grad_norm": 0.30098390579223633, + "learning_rate": 7.882543548545555e-05, + "loss": 0.0725, + "step": 24231 + }, + { + "epoch": 0.5680362406746485, + "grad_norm": 0.43034687638282776, + "learning_rate": 7.88182364141151e-05, + "loss": 0.0331, + "step": 24232 + }, + { + "epoch": 0.5680596822494535, + "grad_norm": 0.8764131665229797, + "learning_rate": 7.881103745770752e-05, + "loss": 0.1134, + "step": 24233 + }, + { + "epoch": 0.5680831238242585, + "grad_norm": 0.33312422037124634, + "learning_rate": 7.880383861627185e-05, + "loss": 0.0474, + "step": 24234 + }, + { + "epoch": 0.5681065653990635, + "grad_norm": 0.5605358481407166, + "learning_rate": 7.879663988984712e-05, + "loss": 0.1148, + "step": 24235 + }, + { + "epoch": 0.5681300069738685, + "grad_norm": 0.3307230770587921, + "learning_rate": 7.878944127847242e-05, + "loss": 0.0551, + "step": 24236 + }, + { + "epoch": 0.5681534485486734, + "grad_norm": 0.1954333335161209, + "learning_rate": 7.878224278218683e-05, + "loss": 0.0438, + "step": 24237 + }, + { + "epoch": 0.5681768901234785, + "grad_norm": 0.5008290410041809, + "learning_rate": 7.877504440102939e-05, + "loss": 0.1095, + "step": 24238 + }, + { + "epoch": 0.5682003316982835, + "grad_norm": 0.2395717203617096, + "learning_rate": 7.876784613503914e-05, + "loss": 0.0342, + "step": 24239 + }, + { + "epoch": 0.5682237732730885, + "grad_norm": 0.7731165289878845, + "learning_rate": 7.876064798425517e-05, + "loss": 0.1195, + "step": 24240 + }, + { + "epoch": 0.5682472148478935, + "grad_norm": 0.5509139895439148, + "learning_rate": 7.875344994871652e-05, + "loss": 0.1139, + "step": 24241 + }, + { + "epoch": 0.5682706564226985, + "grad_norm": 0.11393580585718155, + "learning_rate": 7.874625202846224e-05, + "loss": 0.0259, + "step": 24242 + }, + { + "epoch": 0.5682940979975035, + "grad_norm": 0.4431701600551605, + "learning_rate": 7.873905422353139e-05, + "loss": 0.1043, + "step": 24243 + }, + { + "epoch": 0.5683175395723085, + "grad_norm": 0.462645024061203, + "learning_rate": 7.873185653396303e-05, + "loss": 0.1003, + "step": 24244 + }, + { + "epoch": 0.5683409811471135, + "grad_norm": 0.5365416407585144, + "learning_rate": 7.872465895979622e-05, + "loss": 0.0632, + "step": 24245 + }, + { + "epoch": 0.5683644227219184, + "grad_norm": 0.5549293160438538, + "learning_rate": 7.871746150107002e-05, + "loss": 0.0854, + "step": 24246 + }, + { + "epoch": 0.5683878642967235, + "grad_norm": 0.8925472497940063, + "learning_rate": 7.871026415782343e-05, + "loss": 0.1009, + "step": 24247 + }, + { + "epoch": 0.5684113058715284, + "grad_norm": 0.06411377340555191, + "learning_rate": 7.870306693009559e-05, + "loss": 0.0076, + "step": 24248 + }, + { + "epoch": 0.5684347474463335, + "grad_norm": 0.38226696848869324, + "learning_rate": 7.869586981792547e-05, + "loss": 0.0728, + "step": 24249 + }, + { + "epoch": 0.5684581890211384, + "grad_norm": 0.607117772102356, + "learning_rate": 7.86886728213522e-05, + "loss": 0.1017, + "step": 24250 + }, + { + "epoch": 0.5684816305959435, + "grad_norm": 0.38701435923576355, + "learning_rate": 7.868147594041477e-05, + "loss": 0.0791, + "step": 24251 + }, + { + "epoch": 0.5685050721707484, + "grad_norm": 0.097980797290802, + "learning_rate": 7.867427917515226e-05, + "loss": 0.0118, + "step": 24252 + }, + { + "epoch": 0.5685285137455535, + "grad_norm": 0.3685010075569153, + "learning_rate": 7.866708252560371e-05, + "loss": 0.0734, + "step": 24253 + }, + { + "epoch": 0.5685519553203584, + "grad_norm": 0.4338758885860443, + "learning_rate": 7.865988599180815e-05, + "loss": 0.0684, + "step": 24254 + }, + { + "epoch": 0.5685753968951635, + "grad_norm": 0.49196624755859375, + "learning_rate": 7.865268957380467e-05, + "loss": 0.1083, + "step": 24255 + }, + { + "epoch": 0.5685988384699684, + "grad_norm": 0.4422351121902466, + "learning_rate": 7.864549327163231e-05, + "loss": 0.0549, + "step": 24256 + }, + { + "epoch": 0.5686222800447734, + "grad_norm": 0.30681076645851135, + "learning_rate": 7.863829708533007e-05, + "loss": 0.0682, + "step": 24257 + }, + { + "epoch": 0.5686457216195784, + "grad_norm": 0.3565711975097656, + "learning_rate": 7.863110101493704e-05, + "loss": 0.2561, + "step": 24258 + }, + { + "epoch": 0.5686691631943834, + "grad_norm": 0.2989634871482849, + "learning_rate": 7.862390506049225e-05, + "loss": 0.0521, + "step": 24259 + }, + { + "epoch": 0.5686926047691884, + "grad_norm": 0.27267009019851685, + "learning_rate": 7.861670922203477e-05, + "loss": 0.0299, + "step": 24260 + }, + { + "epoch": 0.5687160463439934, + "grad_norm": 0.45548203587532043, + "learning_rate": 7.86095134996036e-05, + "loss": 0.077, + "step": 24261 + }, + { + "epoch": 0.5687394879187984, + "grad_norm": 0.3529399037361145, + "learning_rate": 7.860231789323785e-05, + "loss": 0.045, + "step": 24262 + }, + { + "epoch": 0.5687629294936034, + "grad_norm": 0.11533178389072418, + "learning_rate": 7.859512240297651e-05, + "loss": 0.0118, + "step": 24263 + }, + { + "epoch": 0.5687863710684083, + "grad_norm": 0.44170400500297546, + "learning_rate": 7.858792702885862e-05, + "loss": 0.0666, + "step": 24264 + }, + { + "epoch": 0.5688098126432134, + "grad_norm": 0.2004917412996292, + "learning_rate": 7.858073177092327e-05, + "loss": 0.0396, + "step": 24265 + }, + { + "epoch": 0.5688332542180183, + "grad_norm": 0.25253069400787354, + "learning_rate": 7.857353662920946e-05, + "loss": 0.0486, + "step": 24266 + }, + { + "epoch": 0.5688566957928234, + "grad_norm": 0.2869638204574585, + "learning_rate": 7.856634160375622e-05, + "loss": 0.0504, + "step": 24267 + }, + { + "epoch": 0.5688801373676283, + "grad_norm": 0.49881812930107117, + "learning_rate": 7.855914669460265e-05, + "loss": 0.1268, + "step": 24268 + }, + { + "epoch": 0.5689035789424334, + "grad_norm": 0.5755438208580017, + "learning_rate": 7.855195190178774e-05, + "loss": 0.6021, + "step": 24269 + }, + { + "epoch": 0.5689270205172383, + "grad_norm": 0.38781076669692993, + "learning_rate": 7.854475722535052e-05, + "loss": 0.0597, + "step": 24270 + }, + { + "epoch": 0.5689504620920434, + "grad_norm": 0.4610704183578491, + "learning_rate": 7.853756266533003e-05, + "loss": 0.4962, + "step": 24271 + }, + { + "epoch": 0.5689739036668483, + "grad_norm": 0.6524781584739685, + "learning_rate": 7.853036822176538e-05, + "loss": 0.0918, + "step": 24272 + }, + { + "epoch": 0.5689973452416534, + "grad_norm": 0.11078908294439316, + "learning_rate": 7.852317389469555e-05, + "loss": 0.0219, + "step": 24273 + }, + { + "epoch": 0.5690207868164583, + "grad_norm": 0.3036409318447113, + "learning_rate": 7.851597968415956e-05, + "loss": 0.5552, + "step": 24274 + }, + { + "epoch": 0.5690442283912633, + "grad_norm": 0.16143885254859924, + "learning_rate": 7.850878559019649e-05, + "loss": 0.0338, + "step": 24275 + }, + { + "epoch": 0.5690676699660683, + "grad_norm": 0.08939960598945618, + "learning_rate": 7.850159161284536e-05, + "loss": 0.0078, + "step": 24276 + }, + { + "epoch": 0.5690911115408733, + "grad_norm": 0.404825359582901, + "learning_rate": 7.849439775214518e-05, + "loss": 0.0638, + "step": 24277 + }, + { + "epoch": 0.5691145531156783, + "grad_norm": 0.4690052270889282, + "learning_rate": 7.8487204008135e-05, + "loss": 0.1196, + "step": 24278 + }, + { + "epoch": 0.5691379946904833, + "grad_norm": 0.5453909039497375, + "learning_rate": 7.848001038085389e-05, + "loss": 0.082, + "step": 24279 + }, + { + "epoch": 0.5691614362652883, + "grad_norm": 0.4959195554256439, + "learning_rate": 7.847281687034081e-05, + "loss": 0.0607, + "step": 24280 + }, + { + "epoch": 0.5691848778400933, + "grad_norm": 0.4204077422618866, + "learning_rate": 7.846562347663484e-05, + "loss": 0.0687, + "step": 24281 + }, + { + "epoch": 0.5692083194148982, + "grad_norm": 0.16046403348445892, + "learning_rate": 7.8458430199775e-05, + "loss": 0.0217, + "step": 24282 + }, + { + "epoch": 0.5692317609897033, + "grad_norm": 0.41142338514328003, + "learning_rate": 7.845123703980032e-05, + "loss": 0.0537, + "step": 24283 + }, + { + "epoch": 0.5692552025645082, + "grad_norm": 0.15411627292633057, + "learning_rate": 7.844404399674983e-05, + "loss": 0.0217, + "step": 24284 + }, + { + "epoch": 0.5692786441393133, + "grad_norm": 0.1074705421924591, + "learning_rate": 7.843685107066257e-05, + "loss": 0.0178, + "step": 24285 + }, + { + "epoch": 0.5693020857141182, + "grad_norm": 0.1575746238231659, + "learning_rate": 7.842965826157759e-05, + "loss": 0.0213, + "step": 24286 + }, + { + "epoch": 0.5693255272889233, + "grad_norm": 0.1723833531141281, + "learning_rate": 7.842246556953386e-05, + "loss": 0.0233, + "step": 24287 + }, + { + "epoch": 0.5693489688637282, + "grad_norm": 0.4413752853870392, + "learning_rate": 7.841527299457046e-05, + "loss": 0.0636, + "step": 24288 + }, + { + "epoch": 0.5693724104385333, + "grad_norm": 0.22284486889839172, + "learning_rate": 7.840808053672636e-05, + "loss": 0.0376, + "step": 24289 + }, + { + "epoch": 0.5693958520133383, + "grad_norm": 0.1620263010263443, + "learning_rate": 7.840088819604065e-05, + "loss": 0.026, + "step": 24290 + }, + { + "epoch": 0.5694192935881432, + "grad_norm": 0.44957900047302246, + "learning_rate": 7.839369597255234e-05, + "loss": 0.119, + "step": 24291 + }, + { + "epoch": 0.5694427351629483, + "grad_norm": 0.15467867255210876, + "learning_rate": 7.838650386630041e-05, + "loss": 0.0288, + "step": 24292 + }, + { + "epoch": 0.5694661767377532, + "grad_norm": 0.4746779501438141, + "learning_rate": 7.837931187732395e-05, + "loss": 0.0824, + "step": 24293 + }, + { + "epoch": 0.5694896183125583, + "grad_norm": 0.5232187509536743, + "learning_rate": 7.837212000566194e-05, + "loss": 0.1247, + "step": 24294 + }, + { + "epoch": 0.5695130598873632, + "grad_norm": 0.4111599922180176, + "learning_rate": 7.836492825135337e-05, + "loss": 0.0959, + "step": 24295 + }, + { + "epoch": 0.5695365014621683, + "grad_norm": 0.4952014982700348, + "learning_rate": 7.835773661443734e-05, + "loss": 0.0882, + "step": 24296 + }, + { + "epoch": 0.5695599430369732, + "grad_norm": 0.25115907192230225, + "learning_rate": 7.835054509495285e-05, + "loss": 0.0176, + "step": 24297 + }, + { + "epoch": 0.5695833846117783, + "grad_norm": 0.6042527556419373, + "learning_rate": 7.834335369293893e-05, + "loss": 0.1482, + "step": 24298 + }, + { + "epoch": 0.5696068261865832, + "grad_norm": 0.5386165380477905, + "learning_rate": 7.833616240843454e-05, + "loss": 0.0613, + "step": 24299 + }, + { + "epoch": 0.5696302677613883, + "grad_norm": 0.6449350714683533, + "learning_rate": 7.832897124147878e-05, + "loss": 0.1403, + "step": 24300 + }, + { + "epoch": 0.5696537093361932, + "grad_norm": 0.31994539499282837, + "learning_rate": 7.832178019211062e-05, + "loss": 0.0327, + "step": 24301 + }, + { + "epoch": 0.5696771509109982, + "grad_norm": 0.6580634117126465, + "learning_rate": 7.831458926036908e-05, + "loss": 0.1645, + "step": 24302 + }, + { + "epoch": 0.5697005924858032, + "grad_norm": 0.20333194732666016, + "learning_rate": 7.830739844629319e-05, + "loss": 0.0476, + "step": 24303 + }, + { + "epoch": 0.5697240340606082, + "grad_norm": 0.7105470895767212, + "learning_rate": 7.830020774992198e-05, + "loss": 0.6199, + "step": 24304 + }, + { + "epoch": 0.5697474756354132, + "grad_norm": 0.2740097641944885, + "learning_rate": 7.829301717129442e-05, + "loss": 0.0503, + "step": 24305 + }, + { + "epoch": 0.5697709172102182, + "grad_norm": 0.2832302153110504, + "learning_rate": 7.82858267104496e-05, + "loss": 0.0563, + "step": 24306 + }, + { + "epoch": 0.5697943587850232, + "grad_norm": 0.5251947045326233, + "learning_rate": 7.827863636742649e-05, + "loss": 0.1167, + "step": 24307 + }, + { + "epoch": 0.5698178003598282, + "grad_norm": 0.4614257216453552, + "learning_rate": 7.827144614226407e-05, + "loss": 0.0944, + "step": 24308 + }, + { + "epoch": 0.5698412419346331, + "grad_norm": 0.30607354640960693, + "learning_rate": 7.82642560350014e-05, + "loss": 0.0591, + "step": 24309 + }, + { + "epoch": 0.5698646835094382, + "grad_norm": 0.7408806681632996, + "learning_rate": 7.82570660456775e-05, + "loss": 0.1376, + "step": 24310 + }, + { + "epoch": 0.5698881250842431, + "grad_norm": 0.47073400020599365, + "learning_rate": 7.824987617433139e-05, + "loss": 0.0631, + "step": 24311 + }, + { + "epoch": 0.5699115666590482, + "grad_norm": 0.2740383446216583, + "learning_rate": 7.824268642100206e-05, + "loss": 0.0412, + "step": 24312 + }, + { + "epoch": 0.5699350082338531, + "grad_norm": 0.7119961977005005, + "learning_rate": 7.82354967857285e-05, + "loss": 0.0721, + "step": 24313 + }, + { + "epoch": 0.5699584498086582, + "grad_norm": 0.5484520196914673, + "learning_rate": 7.822830726854978e-05, + "loss": 0.1438, + "step": 24314 + }, + { + "epoch": 0.5699818913834631, + "grad_norm": 0.22724954783916473, + "learning_rate": 7.822111786950486e-05, + "loss": 0.0429, + "step": 24315 + }, + { + "epoch": 0.5700053329582682, + "grad_norm": 0.36502137780189514, + "learning_rate": 7.821392858863276e-05, + "loss": 0.0796, + "step": 24316 + }, + { + "epoch": 0.5700287745330731, + "grad_norm": 0.3034302890300751, + "learning_rate": 7.820673942597252e-05, + "loss": 0.0349, + "step": 24317 + }, + { + "epoch": 0.5700522161078782, + "grad_norm": 0.17609886825084686, + "learning_rate": 7.819955038156311e-05, + "loss": 0.0243, + "step": 24318 + }, + { + "epoch": 0.5700756576826831, + "grad_norm": 0.1599702537059784, + "learning_rate": 7.819236145544355e-05, + "loss": 0.0355, + "step": 24319 + }, + { + "epoch": 0.5700990992574881, + "grad_norm": 0.39710018038749695, + "learning_rate": 7.818517264765282e-05, + "loss": 0.053, + "step": 24320 + }, + { + "epoch": 0.5701225408322931, + "grad_norm": 0.675488293170929, + "learning_rate": 7.817798395822999e-05, + "loss": 0.1604, + "step": 24321 + }, + { + "epoch": 0.5701459824070981, + "grad_norm": 0.7048866152763367, + "learning_rate": 7.817079538721401e-05, + "loss": 0.3888, + "step": 24322 + }, + { + "epoch": 0.5701694239819031, + "grad_norm": 0.7109444737434387, + "learning_rate": 7.816360693464394e-05, + "loss": 0.0747, + "step": 24323 + }, + { + "epoch": 0.5701928655567081, + "grad_norm": 0.42280080914497375, + "learning_rate": 7.815641860055876e-05, + "loss": 0.0846, + "step": 24324 + }, + { + "epoch": 0.5702163071315131, + "grad_norm": 0.4548161029815674, + "learning_rate": 7.814923038499743e-05, + "loss": 0.5447, + "step": 24325 + }, + { + "epoch": 0.5702397487063181, + "grad_norm": 0.28442174196243286, + "learning_rate": 7.814204228799901e-05, + "loss": 0.0571, + "step": 24326 + }, + { + "epoch": 0.570263190281123, + "grad_norm": 1.2693262100219727, + "learning_rate": 7.813485430960248e-05, + "loss": 0.2826, + "step": 24327 + }, + { + "epoch": 0.5702866318559281, + "grad_norm": 0.17700393497943878, + "learning_rate": 7.812766644984685e-05, + "loss": 0.0335, + "step": 24328 + }, + { + "epoch": 0.570310073430733, + "grad_norm": 0.28476232290267944, + "learning_rate": 7.812047870877113e-05, + "loss": 0.023, + "step": 24329 + }, + { + "epoch": 0.5703335150055381, + "grad_norm": 0.4419505000114441, + "learning_rate": 7.811329108641428e-05, + "loss": 0.5409, + "step": 24330 + }, + { + "epoch": 0.570356956580343, + "grad_norm": 0.4003146290779114, + "learning_rate": 7.810610358281534e-05, + "loss": 0.0612, + "step": 24331 + }, + { + "epoch": 0.5703803981551481, + "grad_norm": 0.5808707475662231, + "learning_rate": 7.809891619801326e-05, + "loss": 0.0967, + "step": 24332 + }, + { + "epoch": 0.570403839729953, + "grad_norm": 0.26349446177482605, + "learning_rate": 7.809172893204712e-05, + "loss": 0.0381, + "step": 24333 + }, + { + "epoch": 0.5704272813047581, + "grad_norm": 0.35227710008621216, + "learning_rate": 7.808454178495585e-05, + "loss": 0.0796, + "step": 24334 + }, + { + "epoch": 0.570450722879563, + "grad_norm": 0.2240026295185089, + "learning_rate": 7.807735475677851e-05, + "loss": 0.0274, + "step": 24335 + }, + { + "epoch": 0.570474164454368, + "grad_norm": 0.6602413654327393, + "learning_rate": 7.807016784755404e-05, + "loss": 0.1267, + "step": 24336 + }, + { + "epoch": 0.570497606029173, + "grad_norm": 0.963501513004303, + "learning_rate": 7.806298105732145e-05, + "loss": 0.1511, + "step": 24337 + }, + { + "epoch": 0.570521047603978, + "grad_norm": 0.4445200264453888, + "learning_rate": 7.805579438611975e-05, + "loss": 0.0869, + "step": 24338 + }, + { + "epoch": 0.570544489178783, + "grad_norm": 0.19657820463180542, + "learning_rate": 7.804860783398794e-05, + "loss": 0.0396, + "step": 24339 + }, + { + "epoch": 0.570567930753588, + "grad_norm": 0.5035918354988098, + "learning_rate": 7.804142140096498e-05, + "loss": 0.091, + "step": 24340 + }, + { + "epoch": 0.5705913723283931, + "grad_norm": 0.5034584999084473, + "learning_rate": 7.80342350870899e-05, + "loss": 0.0841, + "step": 24341 + }, + { + "epoch": 0.570614813903198, + "grad_norm": 0.648186206817627, + "learning_rate": 7.802704889240166e-05, + "loss": 0.7006, + "step": 24342 + }, + { + "epoch": 0.5706382554780031, + "grad_norm": 0.33484166860580444, + "learning_rate": 7.801986281693927e-05, + "loss": 0.0597, + "step": 24343 + }, + { + "epoch": 0.570661697052808, + "grad_norm": 0.3943198025226593, + "learning_rate": 7.80126768607417e-05, + "loss": 0.1285, + "step": 24344 + }, + { + "epoch": 0.570685138627613, + "grad_norm": 0.12867362797260284, + "learning_rate": 7.8005491023848e-05, + "loss": 0.0197, + "step": 24345 + }, + { + "epoch": 0.570708580202418, + "grad_norm": 0.4447372853755951, + "learning_rate": 7.799830530629712e-05, + "loss": 0.1041, + "step": 24346 + }, + { + "epoch": 0.570732021777223, + "grad_norm": 0.3371398150920868, + "learning_rate": 7.799111970812805e-05, + "loss": 0.0269, + "step": 24347 + }, + { + "epoch": 0.570755463352028, + "grad_norm": 0.25755858421325684, + "learning_rate": 7.79839342293798e-05, + "loss": 0.0312, + "step": 24348 + }, + { + "epoch": 0.570778904926833, + "grad_norm": 0.2703699767589569, + "learning_rate": 7.797674887009133e-05, + "loss": 0.0388, + "step": 24349 + }, + { + "epoch": 0.570802346501638, + "grad_norm": 0.5510957837104797, + "learning_rate": 7.796956363030162e-05, + "loss": 0.0324, + "step": 24350 + }, + { + "epoch": 0.570825788076443, + "grad_norm": 0.38729676604270935, + "learning_rate": 7.79623785100497e-05, + "loss": 0.0607, + "step": 24351 + }, + { + "epoch": 0.570849229651248, + "grad_norm": 0.3901030123233795, + "learning_rate": 7.795519350937455e-05, + "loss": 0.0659, + "step": 24352 + }, + { + "epoch": 0.570872671226053, + "grad_norm": 0.48099783062934875, + "learning_rate": 7.794800862831511e-05, + "loss": 0.4795, + "step": 24353 + }, + { + "epoch": 0.570896112800858, + "grad_norm": 0.28325921297073364, + "learning_rate": 7.79408238669104e-05, + "loss": 0.0249, + "step": 24354 + }, + { + "epoch": 0.570919554375663, + "grad_norm": 0.281436562538147, + "learning_rate": 7.793363922519941e-05, + "loss": 0.0563, + "step": 24355 + }, + { + "epoch": 0.5709429959504679, + "grad_norm": 0.5726379156112671, + "learning_rate": 7.792645470322109e-05, + "loss": 0.0505, + "step": 24356 + }, + { + "epoch": 0.570966437525273, + "grad_norm": 0.4456551671028137, + "learning_rate": 7.791927030101445e-05, + "loss": 0.09, + "step": 24357 + }, + { + "epoch": 0.5709898791000779, + "grad_norm": 0.7166346311569214, + "learning_rate": 7.791208601861849e-05, + "loss": 0.1033, + "step": 24358 + }, + { + "epoch": 0.571013320674883, + "grad_norm": 0.4815327227115631, + "learning_rate": 7.790490185607218e-05, + "loss": 0.113, + "step": 24359 + }, + { + "epoch": 0.5710367622496879, + "grad_norm": 0.3685425817966461, + "learning_rate": 7.789771781341446e-05, + "loss": 0.086, + "step": 24360 + }, + { + "epoch": 0.571060203824493, + "grad_norm": 0.429334431886673, + "learning_rate": 7.789053389068438e-05, + "loss": 0.1109, + "step": 24361 + }, + { + "epoch": 0.5710836453992979, + "grad_norm": 0.3411233127117157, + "learning_rate": 7.788335008792086e-05, + "loss": 0.0434, + "step": 24362 + }, + { + "epoch": 0.571107086974103, + "grad_norm": 0.38808533549308777, + "learning_rate": 7.787616640516292e-05, + "loss": 0.0769, + "step": 24363 + }, + { + "epoch": 0.5711305285489079, + "grad_norm": 0.35644567012786865, + "learning_rate": 7.786898284244952e-05, + "loss": 0.0691, + "step": 24364 + }, + { + "epoch": 0.5711539701237129, + "grad_norm": 0.42145076394081116, + "learning_rate": 7.786179939981963e-05, + "loss": 0.0873, + "step": 24365 + }, + { + "epoch": 0.5711774116985179, + "grad_norm": 0.28439611196517944, + "learning_rate": 7.785461607731226e-05, + "loss": 0.0188, + "step": 24366 + }, + { + "epoch": 0.5712008532733229, + "grad_norm": 0.4954493045806885, + "learning_rate": 7.784743287496637e-05, + "loss": 0.5718, + "step": 24367 + }, + { + "epoch": 0.5712242948481279, + "grad_norm": 0.46196189522743225, + "learning_rate": 7.784024979282087e-05, + "loss": 0.6674, + "step": 24368 + }, + { + "epoch": 0.5712477364229329, + "grad_norm": 0.2784207761287689, + "learning_rate": 7.783306683091485e-05, + "loss": 0.0424, + "step": 24369 + }, + { + "epoch": 0.5712711779977379, + "grad_norm": 0.4797378480434418, + "learning_rate": 7.782588398928723e-05, + "loss": 0.0708, + "step": 24370 + }, + { + "epoch": 0.5712946195725429, + "grad_norm": 0.3933902978897095, + "learning_rate": 7.781870126797699e-05, + "loss": 0.4423, + "step": 24371 + }, + { + "epoch": 0.5713180611473478, + "grad_norm": 0.25838425755500793, + "learning_rate": 7.781151866702307e-05, + "loss": 0.0634, + "step": 24372 + }, + { + "epoch": 0.5713415027221529, + "grad_norm": 0.4655851125717163, + "learning_rate": 7.780433618646452e-05, + "loss": 0.0542, + "step": 24373 + }, + { + "epoch": 0.5713649442969578, + "grad_norm": 0.7778337597846985, + "learning_rate": 7.779715382634026e-05, + "loss": 0.0586, + "step": 24374 + }, + { + "epoch": 0.5713883858717629, + "grad_norm": 0.40939027070999146, + "learning_rate": 7.778997158668923e-05, + "loss": 0.1128, + "step": 24375 + }, + { + "epoch": 0.5714118274465678, + "grad_norm": 0.6075214147567749, + "learning_rate": 7.778278946755046e-05, + "loss": 0.1085, + "step": 24376 + }, + { + "epoch": 0.5714352690213729, + "grad_norm": 0.4054868519306183, + "learning_rate": 7.777560746896292e-05, + "loss": 0.501, + "step": 24377 + }, + { + "epoch": 0.5714587105961778, + "grad_norm": 1.7404518127441406, + "learning_rate": 7.776842559096552e-05, + "loss": 0.0699, + "step": 24378 + }, + { + "epoch": 0.5714821521709829, + "grad_norm": 0.23703902959823608, + "learning_rate": 7.776124383359729e-05, + "loss": 0.0319, + "step": 24379 + }, + { + "epoch": 0.5715055937457878, + "grad_norm": 0.08387057483196259, + "learning_rate": 7.775406219689714e-05, + "loss": 0.0183, + "step": 24380 + }, + { + "epoch": 0.5715290353205928, + "grad_norm": 0.37654560804367065, + "learning_rate": 7.774688068090412e-05, + "loss": 0.0981, + "step": 24381 + }, + { + "epoch": 0.5715524768953978, + "grad_norm": 0.5344024896621704, + "learning_rate": 7.773969928565713e-05, + "loss": 0.0798, + "step": 24382 + }, + { + "epoch": 0.5715759184702028, + "grad_norm": 0.6548934578895569, + "learning_rate": 7.773251801119517e-05, + "loss": 0.0872, + "step": 24383 + }, + { + "epoch": 0.5715993600450078, + "grad_norm": 0.31537339091300964, + "learning_rate": 7.77253368575572e-05, + "loss": 0.0531, + "step": 24384 + }, + { + "epoch": 0.5716228016198128, + "grad_norm": 0.49702540040016174, + "learning_rate": 7.771815582478214e-05, + "loss": 0.6118, + "step": 24385 + }, + { + "epoch": 0.5716462431946178, + "grad_norm": 0.21911819279193878, + "learning_rate": 7.771097491290904e-05, + "loss": 0.0285, + "step": 24386 + }, + { + "epoch": 0.5716696847694228, + "grad_norm": 0.5430512428283691, + "learning_rate": 7.77037941219768e-05, + "loss": 0.0776, + "step": 24387 + }, + { + "epoch": 0.5716931263442278, + "grad_norm": 0.4872431755065918, + "learning_rate": 7.769661345202439e-05, + "loss": 0.0917, + "step": 24388 + }, + { + "epoch": 0.5717165679190328, + "grad_norm": 0.33104783296585083, + "learning_rate": 7.76894329030908e-05, + "loss": 0.0632, + "step": 24389 + }, + { + "epoch": 0.5717400094938377, + "grad_norm": 0.3633274435997009, + "learning_rate": 7.768225247521496e-05, + "loss": 0.0523, + "step": 24390 + }, + { + "epoch": 0.5717634510686428, + "grad_norm": 0.434717059135437, + "learning_rate": 7.767507216843585e-05, + "loss": 0.0485, + "step": 24391 + }, + { + "epoch": 0.5717868926434478, + "grad_norm": 0.7223032116889954, + "learning_rate": 7.766789198279242e-05, + "loss": 0.0748, + "step": 24392 + }, + { + "epoch": 0.5718103342182528, + "grad_norm": 0.3686584234237671, + "learning_rate": 7.766071191832362e-05, + "loss": 0.0678, + "step": 24393 + }, + { + "epoch": 0.5718337757930578, + "grad_norm": 0.13878020644187927, + "learning_rate": 7.765353197506844e-05, + "loss": 0.0157, + "step": 24394 + }, + { + "epoch": 0.5718572173678628, + "grad_norm": 0.4642404019832611, + "learning_rate": 7.764635215306584e-05, + "loss": 0.0931, + "step": 24395 + }, + { + "epoch": 0.5718806589426678, + "grad_norm": 0.18840347230434418, + "learning_rate": 7.763917245235474e-05, + "loss": 0.028, + "step": 24396 + }, + { + "epoch": 0.5719041005174728, + "grad_norm": 0.6593840718269348, + "learning_rate": 7.763199287297416e-05, + "loss": 0.1121, + "step": 24397 + }, + { + "epoch": 0.5719275420922778, + "grad_norm": 0.41156071424484253, + "learning_rate": 7.762481341496298e-05, + "loss": 0.0569, + "step": 24398 + }, + { + "epoch": 0.5719509836670827, + "grad_norm": 0.06698479503393173, + "learning_rate": 7.76176340783602e-05, + "loss": 0.0096, + "step": 24399 + }, + { + "epoch": 0.5719744252418878, + "grad_norm": 0.3708235025405884, + "learning_rate": 7.761045486320476e-05, + "loss": 0.043, + "step": 24400 + }, + { + "epoch": 0.5719978668166927, + "grad_norm": 0.6321961283683777, + "learning_rate": 7.760327576953562e-05, + "loss": 0.1016, + "step": 24401 + }, + { + "epoch": 0.5720213083914978, + "grad_norm": 0.35196682810783386, + "learning_rate": 7.759609679739175e-05, + "loss": 0.0418, + "step": 24402 + }, + { + "epoch": 0.5720447499663027, + "grad_norm": 0.41479718685150146, + "learning_rate": 7.758891794681207e-05, + "loss": 0.0524, + "step": 24403 + }, + { + "epoch": 0.5720681915411078, + "grad_norm": 0.4008217751979828, + "learning_rate": 7.758173921783557e-05, + "loss": 0.0874, + "step": 24404 + }, + { + "epoch": 0.5720916331159127, + "grad_norm": 0.9211625456809998, + "learning_rate": 7.757456061050115e-05, + "loss": 0.2422, + "step": 24405 + }, + { + "epoch": 0.5721150746907178, + "grad_norm": 0.1067894846200943, + "learning_rate": 7.756738212484782e-05, + "loss": 0.0188, + "step": 24406 + }, + { + "epoch": 0.5721385162655227, + "grad_norm": 0.3650861084461212, + "learning_rate": 7.75602037609145e-05, + "loss": 0.0766, + "step": 24407 + }, + { + "epoch": 0.5721619578403278, + "grad_norm": 0.6178014278411865, + "learning_rate": 7.755302551874015e-05, + "loss": 0.6771, + "step": 24408 + }, + { + "epoch": 0.5721853994151327, + "grad_norm": 0.6184912919998169, + "learning_rate": 7.754584739836372e-05, + "loss": 0.1334, + "step": 24409 + }, + { + "epoch": 0.5722088409899377, + "grad_norm": 0.1451316773891449, + "learning_rate": 7.753866939982415e-05, + "loss": 0.0298, + "step": 24410 + }, + { + "epoch": 0.5722322825647427, + "grad_norm": 0.22067244350910187, + "learning_rate": 7.753149152316039e-05, + "loss": 0.0477, + "step": 24411 + }, + { + "epoch": 0.5722557241395477, + "grad_norm": 0.39496374130249023, + "learning_rate": 7.752431376841141e-05, + "loss": 0.0234, + "step": 24412 + }, + { + "epoch": 0.5722791657143527, + "grad_norm": 0.6548299193382263, + "learning_rate": 7.75171361356161e-05, + "loss": 0.1471, + "step": 24413 + }, + { + "epoch": 0.5723026072891577, + "grad_norm": 0.5470117926597595, + "learning_rate": 7.750995862481346e-05, + "loss": 0.1622, + "step": 24414 + }, + { + "epoch": 0.5723260488639627, + "grad_norm": 0.5434409379959106, + "learning_rate": 7.750278123604241e-05, + "loss": 0.1303, + "step": 24415 + }, + { + "epoch": 0.5723494904387677, + "grad_norm": 0.26536089181900024, + "learning_rate": 7.74956039693419e-05, + "loss": 0.0328, + "step": 24416 + }, + { + "epoch": 0.5723729320135726, + "grad_norm": 0.2933516204357147, + "learning_rate": 7.748842682475085e-05, + "loss": 0.0466, + "step": 24417 + }, + { + "epoch": 0.5723963735883777, + "grad_norm": 0.12140864133834839, + "learning_rate": 7.748124980230826e-05, + "loss": 0.0176, + "step": 24418 + }, + { + "epoch": 0.5724198151631826, + "grad_norm": 0.4985012114048004, + "learning_rate": 7.747407290205306e-05, + "loss": 0.0907, + "step": 24419 + }, + { + "epoch": 0.5724432567379877, + "grad_norm": 0.2773365378379822, + "learning_rate": 7.746689612402414e-05, + "loss": 0.019, + "step": 24420 + }, + { + "epoch": 0.5724666983127926, + "grad_norm": 0.27923962473869324, + "learning_rate": 7.74597194682605e-05, + "loss": 0.0242, + "step": 24421 + }, + { + "epoch": 0.5724901398875977, + "grad_norm": 0.5614469647407532, + "learning_rate": 7.745254293480106e-05, + "loss": 0.5283, + "step": 24422 + }, + { + "epoch": 0.5725135814624026, + "grad_norm": 0.5725875496864319, + "learning_rate": 7.744536652368472e-05, + "loss": 0.143, + "step": 24423 + }, + { + "epoch": 0.5725370230372077, + "grad_norm": 0.1426505148410797, + "learning_rate": 7.743819023495049e-05, + "loss": 0.0364, + "step": 24424 + }, + { + "epoch": 0.5725604646120126, + "grad_norm": 0.6061326265335083, + "learning_rate": 7.743101406863726e-05, + "loss": 0.1516, + "step": 24425 + }, + { + "epoch": 0.5725839061868176, + "grad_norm": 0.8297286033630371, + "learning_rate": 7.742383802478397e-05, + "loss": 0.1094, + "step": 24426 + }, + { + "epoch": 0.5726073477616226, + "grad_norm": 0.5706685185432434, + "learning_rate": 7.741666210342961e-05, + "loss": 0.1134, + "step": 24427 + }, + { + "epoch": 0.5726307893364276, + "grad_norm": 0.3727448582649231, + "learning_rate": 7.740948630461304e-05, + "loss": 0.0346, + "step": 24428 + }, + { + "epoch": 0.5726542309112326, + "grad_norm": 0.5178359150886536, + "learning_rate": 7.740231062837322e-05, + "loss": 0.5857, + "step": 24429 + }, + { + "epoch": 0.5726776724860376, + "grad_norm": 0.5704081654548645, + "learning_rate": 7.73951350747491e-05, + "loss": 0.1236, + "step": 24430 + }, + { + "epoch": 0.5727011140608426, + "grad_norm": 0.5124754905700684, + "learning_rate": 7.738795964377965e-05, + "loss": 0.0935, + "step": 24431 + }, + { + "epoch": 0.5727245556356476, + "grad_norm": 0.38866469264030457, + "learning_rate": 7.738078433550375e-05, + "loss": 0.106, + "step": 24432 + }, + { + "epoch": 0.5727479972104526, + "grad_norm": 0.5179601311683655, + "learning_rate": 7.737360914996035e-05, + "loss": 0.0891, + "step": 24433 + }, + { + "epoch": 0.5727714387852576, + "grad_norm": 0.4398076832294464, + "learning_rate": 7.736643408718837e-05, + "loss": 0.0525, + "step": 24434 + }, + { + "epoch": 0.5727948803600625, + "grad_norm": 0.376661092042923, + "learning_rate": 7.735925914722679e-05, + "loss": 0.075, + "step": 24435 + }, + { + "epoch": 0.5728183219348676, + "grad_norm": 0.5147467851638794, + "learning_rate": 7.735208433011447e-05, + "loss": 0.126, + "step": 24436 + }, + { + "epoch": 0.5728417635096725, + "grad_norm": 0.4323757588863373, + "learning_rate": 7.734490963589041e-05, + "loss": 0.0949, + "step": 24437 + }, + { + "epoch": 0.5728652050844776, + "grad_norm": 0.40323659777641296, + "learning_rate": 7.733773506459347e-05, + "loss": 0.0622, + "step": 24438 + }, + { + "epoch": 0.5728886466592825, + "grad_norm": 0.9284783601760864, + "learning_rate": 7.733056061626265e-05, + "loss": 0.138, + "step": 24439 + }, + { + "epoch": 0.5729120882340876, + "grad_norm": 0.3636539876461029, + "learning_rate": 7.732338629093682e-05, + "loss": 0.0491, + "step": 24440 + }, + { + "epoch": 0.5729355298088925, + "grad_norm": 0.4130566120147705, + "learning_rate": 7.731621208865492e-05, + "loss": 0.0923, + "step": 24441 + }, + { + "epoch": 0.5729589713836976, + "grad_norm": 0.5929756164550781, + "learning_rate": 7.730903800945592e-05, + "loss": 0.1364, + "step": 24442 + }, + { + "epoch": 0.5729824129585026, + "grad_norm": 0.3106389045715332, + "learning_rate": 7.730186405337872e-05, + "loss": 0.0289, + "step": 24443 + }, + { + "epoch": 0.5730058545333075, + "grad_norm": 0.45772480964660645, + "learning_rate": 7.729469022046223e-05, + "loss": 0.0496, + "step": 24444 + }, + { + "epoch": 0.5730292961081126, + "grad_norm": 0.7452641129493713, + "learning_rate": 7.72875165107454e-05, + "loss": 0.0558, + "step": 24445 + }, + { + "epoch": 0.5730527376829175, + "grad_norm": 0.06734584271907806, + "learning_rate": 7.728034292426714e-05, + "loss": 0.0098, + "step": 24446 + }, + { + "epoch": 0.5730761792577226, + "grad_norm": 0.8248334527015686, + "learning_rate": 7.727316946106639e-05, + "loss": 0.163, + "step": 24447 + }, + { + "epoch": 0.5730996208325275, + "grad_norm": 0.3275294303894043, + "learning_rate": 7.726599612118204e-05, + "loss": 0.1051, + "step": 24448 + }, + { + "epoch": 0.5731230624073326, + "grad_norm": 0.5414006114006042, + "learning_rate": 7.725882290465305e-05, + "loss": 0.0638, + "step": 24449 + }, + { + "epoch": 0.5731465039821375, + "grad_norm": 0.15136808156967163, + "learning_rate": 7.725164981151832e-05, + "loss": 0.0389, + "step": 24450 + }, + { + "epoch": 0.5731699455569426, + "grad_norm": 0.564315676689148, + "learning_rate": 7.724447684181677e-05, + "loss": 0.1355, + "step": 24451 + }, + { + "epoch": 0.5731933871317475, + "grad_norm": 0.37154635787010193, + "learning_rate": 7.723730399558735e-05, + "loss": 0.0643, + "step": 24452 + }, + { + "epoch": 0.5732168287065526, + "grad_norm": 0.15742772817611694, + "learning_rate": 7.723013127286891e-05, + "loss": 0.0213, + "step": 24453 + }, + { + "epoch": 0.5732402702813575, + "grad_norm": 0.825268566608429, + "learning_rate": 7.722295867370047e-05, + "loss": 0.1222, + "step": 24454 + }, + { + "epoch": 0.5732637118561625, + "grad_norm": 0.08105988800525665, + "learning_rate": 7.721578619812086e-05, + "loss": 0.0137, + "step": 24455 + }, + { + "epoch": 0.5732871534309675, + "grad_norm": 0.4314638376235962, + "learning_rate": 7.720861384616907e-05, + "loss": 0.0727, + "step": 24456 + }, + { + "epoch": 0.5733105950057725, + "grad_norm": 0.4850996434688568, + "learning_rate": 7.720144161788397e-05, + "loss": 0.1112, + "step": 24457 + }, + { + "epoch": 0.5733340365805775, + "grad_norm": 0.43246495723724365, + "learning_rate": 7.719426951330448e-05, + "loss": 0.0932, + "step": 24458 + }, + { + "epoch": 0.5733574781553825, + "grad_norm": 0.4501231610774994, + "learning_rate": 7.718709753246953e-05, + "loss": 0.0812, + "step": 24459 + }, + { + "epoch": 0.5733809197301875, + "grad_norm": 0.2154063880443573, + "learning_rate": 7.717992567541805e-05, + "loss": 0.0359, + "step": 24460 + }, + { + "epoch": 0.5734043613049925, + "grad_norm": 0.6018111109733582, + "learning_rate": 7.71727539421889e-05, + "loss": 0.1266, + "step": 24461 + }, + { + "epoch": 0.5734278028797974, + "grad_norm": 0.4220755398273468, + "learning_rate": 7.716558233282105e-05, + "loss": 0.0807, + "step": 24462 + }, + { + "epoch": 0.5734512444546025, + "grad_norm": 0.07732902467250824, + "learning_rate": 7.715841084735341e-05, + "loss": 0.0084, + "step": 24463 + }, + { + "epoch": 0.5734746860294074, + "grad_norm": 0.3714860677719116, + "learning_rate": 7.715123948582483e-05, + "loss": 0.0448, + "step": 24464 + }, + { + "epoch": 0.5734981276042125, + "grad_norm": 0.38511136174201965, + "learning_rate": 7.714406824827429e-05, + "loss": 0.059, + "step": 24465 + }, + { + "epoch": 0.5735215691790174, + "grad_norm": 0.5020738840103149, + "learning_rate": 7.713689713474064e-05, + "loss": 0.0575, + "step": 24466 + }, + { + "epoch": 0.5735450107538225, + "grad_norm": 0.4700830578804016, + "learning_rate": 7.712972614526288e-05, + "loss": 0.1255, + "step": 24467 + }, + { + "epoch": 0.5735684523286274, + "grad_norm": 0.469382643699646, + "learning_rate": 7.712255527987984e-05, + "loss": 0.1008, + "step": 24468 + }, + { + "epoch": 0.5735918939034325, + "grad_norm": 0.5127705931663513, + "learning_rate": 7.711538453863048e-05, + "loss": 0.1475, + "step": 24469 + }, + { + "epoch": 0.5736153354782374, + "grad_norm": 0.4854515492916107, + "learning_rate": 7.71082139215537e-05, + "loss": 0.1069, + "step": 24470 + }, + { + "epoch": 0.5736387770530424, + "grad_norm": 0.17096364498138428, + "learning_rate": 7.710104342868836e-05, + "loss": 0.0511, + "step": 24471 + }, + { + "epoch": 0.5736622186278474, + "grad_norm": 0.4531182646751404, + "learning_rate": 7.709387306007343e-05, + "loss": 0.0717, + "step": 24472 + }, + { + "epoch": 0.5736856602026524, + "grad_norm": 0.4366861581802368, + "learning_rate": 7.708670281574778e-05, + "loss": 0.0646, + "step": 24473 + }, + { + "epoch": 0.5737091017774574, + "grad_norm": 0.4691055715084076, + "learning_rate": 7.707953269575032e-05, + "loss": 0.1299, + "step": 24474 + }, + { + "epoch": 0.5737325433522624, + "grad_norm": 0.3776562809944153, + "learning_rate": 7.707236270011996e-05, + "loss": 0.0624, + "step": 24475 + }, + { + "epoch": 0.5737559849270674, + "grad_norm": 0.6284355521202087, + "learning_rate": 7.70651928288956e-05, + "loss": 0.1138, + "step": 24476 + }, + { + "epoch": 0.5737794265018724, + "grad_norm": 0.3799660801887512, + "learning_rate": 7.705802308211615e-05, + "loss": 0.0557, + "step": 24477 + }, + { + "epoch": 0.5738028680766774, + "grad_norm": 0.2620916962623596, + "learning_rate": 7.705085345982049e-05, + "loss": 0.0507, + "step": 24478 + }, + { + "epoch": 0.5738263096514824, + "grad_norm": 0.4911750257015228, + "learning_rate": 7.704368396204758e-05, + "loss": 0.11, + "step": 24479 + }, + { + "epoch": 0.5738497512262873, + "grad_norm": 0.30513453483581543, + "learning_rate": 7.703651458883628e-05, + "loss": 0.0296, + "step": 24480 + }, + { + "epoch": 0.5738731928010924, + "grad_norm": 0.4983021318912506, + "learning_rate": 7.702934534022549e-05, + "loss": 0.06, + "step": 24481 + }, + { + "epoch": 0.5738966343758973, + "grad_norm": 0.455268919467926, + "learning_rate": 7.702217621625414e-05, + "loss": 0.101, + "step": 24482 + }, + { + "epoch": 0.5739200759507024, + "grad_norm": 0.5221459269523621, + "learning_rate": 7.701500721696107e-05, + "loss": 0.447, + "step": 24483 + }, + { + "epoch": 0.5739435175255073, + "grad_norm": 1.1073766946792603, + "learning_rate": 7.700783834238525e-05, + "loss": 0.1396, + "step": 24484 + }, + { + "epoch": 0.5739669591003124, + "grad_norm": 0.36543911695480347, + "learning_rate": 7.700066959256553e-05, + "loss": 0.0724, + "step": 24485 + }, + { + "epoch": 0.5739904006751173, + "grad_norm": 0.4981340765953064, + "learning_rate": 7.699350096754082e-05, + "loss": 0.0873, + "step": 24486 + }, + { + "epoch": 0.5740138422499224, + "grad_norm": 0.6724082231521606, + "learning_rate": 7.698633246735005e-05, + "loss": 0.4437, + "step": 24487 + }, + { + "epoch": 0.5740372838247273, + "grad_norm": 0.39293909072875977, + "learning_rate": 7.697916409203206e-05, + "loss": 0.0755, + "step": 24488 + }, + { + "epoch": 0.5740607253995323, + "grad_norm": 0.46076762676239014, + "learning_rate": 7.697199584162576e-05, + "loss": 0.6524, + "step": 24489 + }, + { + "epoch": 0.5740841669743373, + "grad_norm": 0.5160092115402222, + "learning_rate": 7.696482771617005e-05, + "loss": 0.1486, + "step": 24490 + }, + { + "epoch": 0.5741076085491423, + "grad_norm": 0.4620484411716461, + "learning_rate": 7.695765971570386e-05, + "loss": 0.1461, + "step": 24491 + }, + { + "epoch": 0.5741310501239473, + "grad_norm": 0.5410926342010498, + "learning_rate": 7.695049184026607e-05, + "loss": 0.1165, + "step": 24492 + }, + { + "epoch": 0.5741544916987523, + "grad_norm": 0.3930913209915161, + "learning_rate": 7.694332408989553e-05, + "loss": 0.0664, + "step": 24493 + }, + { + "epoch": 0.5741779332735574, + "grad_norm": 0.25940924882888794, + "learning_rate": 7.693615646463119e-05, + "loss": 0.0581, + "step": 24494 + }, + { + "epoch": 0.5742013748483623, + "grad_norm": 0.5985508561134338, + "learning_rate": 7.69289889645119e-05, + "loss": 0.1606, + "step": 24495 + }, + { + "epoch": 0.5742248164231674, + "grad_norm": 0.2656039893627167, + "learning_rate": 7.692182158957655e-05, + "loss": 0.0347, + "step": 24496 + }, + { + "epoch": 0.5742482579979723, + "grad_norm": 0.2538638412952423, + "learning_rate": 7.691465433986407e-05, + "loss": 0.0721, + "step": 24497 + }, + { + "epoch": 0.5742716995727773, + "grad_norm": 0.42161044478416443, + "learning_rate": 7.690748721541332e-05, + "loss": 0.0822, + "step": 24498 + }, + { + "epoch": 0.5742951411475823, + "grad_norm": 0.1234259381890297, + "learning_rate": 7.690032021626319e-05, + "loss": 0.0192, + "step": 24499 + }, + { + "epoch": 0.5743185827223873, + "grad_norm": 0.3486366868019104, + "learning_rate": 7.689315334245257e-05, + "loss": 0.0631, + "step": 24500 + }, + { + "epoch": 0.5743420242971923, + "grad_norm": 0.1227995902299881, + "learning_rate": 7.688598659402035e-05, + "loss": 0.0299, + "step": 24501 + }, + { + "epoch": 0.5743654658719973, + "grad_norm": 0.476998895406723, + "learning_rate": 7.687881997100539e-05, + "loss": 0.1005, + "step": 24502 + }, + { + "epoch": 0.5743889074468023, + "grad_norm": 0.32461017370224, + "learning_rate": 7.687165347344661e-05, + "loss": 0.0985, + "step": 24503 + }, + { + "epoch": 0.5744123490216073, + "grad_norm": 0.383089542388916, + "learning_rate": 7.686448710138293e-05, + "loss": 0.0634, + "step": 24504 + }, + { + "epoch": 0.5744357905964123, + "grad_norm": 0.07904452830553055, + "learning_rate": 7.685732085485317e-05, + "loss": 0.0177, + "step": 24505 + }, + { + "epoch": 0.5744592321712173, + "grad_norm": 0.5960322022438049, + "learning_rate": 7.685015473389623e-05, + "loss": 0.1224, + "step": 24506 + }, + { + "epoch": 0.5744826737460222, + "grad_norm": 0.4144431948661804, + "learning_rate": 7.684298873855102e-05, + "loss": 0.0481, + "step": 24507 + }, + { + "epoch": 0.5745061153208273, + "grad_norm": 0.4106033444404602, + "learning_rate": 7.68358228688564e-05, + "loss": 0.0898, + "step": 24508 + }, + { + "epoch": 0.5745295568956322, + "grad_norm": 0.5041722059249878, + "learning_rate": 7.682865712485123e-05, + "loss": 0.0806, + "step": 24509 + }, + { + "epoch": 0.5745529984704373, + "grad_norm": 0.17997659742832184, + "learning_rate": 7.682149150657446e-05, + "loss": 0.0468, + "step": 24510 + }, + { + "epoch": 0.5745764400452422, + "grad_norm": 0.31720590591430664, + "learning_rate": 7.68143260140649e-05, + "loss": 0.0466, + "step": 24511 + }, + { + "epoch": 0.5745998816200473, + "grad_norm": 0.5408863425254822, + "learning_rate": 7.680716064736146e-05, + "loss": 0.1168, + "step": 24512 + }, + { + "epoch": 0.5746233231948522, + "grad_norm": 0.3340722322463989, + "learning_rate": 7.679999540650303e-05, + "loss": 0.0741, + "step": 24513 + }, + { + "epoch": 0.5746467647696573, + "grad_norm": 0.3522290885448456, + "learning_rate": 7.679283029152844e-05, + "loss": 0.0682, + "step": 24514 + }, + { + "epoch": 0.5746702063444622, + "grad_norm": 0.40662312507629395, + "learning_rate": 7.678566530247665e-05, + "loss": 0.4143, + "step": 24515 + }, + { + "epoch": 0.5746936479192672, + "grad_norm": 0.5302191376686096, + "learning_rate": 7.677850043938645e-05, + "loss": 0.1036, + "step": 24516 + }, + { + "epoch": 0.5747170894940722, + "grad_norm": 0.5128749012947083, + "learning_rate": 7.677133570229681e-05, + "loss": 0.1196, + "step": 24517 + }, + { + "epoch": 0.5747405310688772, + "grad_norm": 0.46467694640159607, + "learning_rate": 7.676417109124653e-05, + "loss": 0.0841, + "step": 24518 + }, + { + "epoch": 0.5747639726436822, + "grad_norm": 0.17620769143104553, + "learning_rate": 7.675700660627451e-05, + "loss": 0.0354, + "step": 24519 + }, + { + "epoch": 0.5747874142184872, + "grad_norm": 0.7150804400444031, + "learning_rate": 7.674984224741966e-05, + "loss": 0.1621, + "step": 24520 + }, + { + "epoch": 0.5748108557932922, + "grad_norm": 0.9525052309036255, + "learning_rate": 7.674267801472077e-05, + "loss": 0.1539, + "step": 24521 + }, + { + "epoch": 0.5748342973680972, + "grad_norm": 0.12188417464494705, + "learning_rate": 7.673551390821679e-05, + "loss": 0.0197, + "step": 24522 + }, + { + "epoch": 0.5748577389429022, + "grad_norm": 0.4809810221195221, + "learning_rate": 7.672834992794658e-05, + "loss": 0.1355, + "step": 24523 + }, + { + "epoch": 0.5748811805177072, + "grad_norm": 1.0414073467254639, + "learning_rate": 7.672118607394896e-05, + "loss": 0.0788, + "step": 24524 + }, + { + "epoch": 0.5749046220925121, + "grad_norm": 0.6071691513061523, + "learning_rate": 7.671402234626286e-05, + "loss": 0.1375, + "step": 24525 + }, + { + "epoch": 0.5749280636673172, + "grad_norm": 0.09170139580965042, + "learning_rate": 7.670685874492709e-05, + "loss": 0.0141, + "step": 24526 + }, + { + "epoch": 0.5749515052421221, + "grad_norm": 0.44256600737571716, + "learning_rate": 7.66996952699806e-05, + "loss": 0.1066, + "step": 24527 + }, + { + "epoch": 0.5749749468169272, + "grad_norm": 0.1416858285665512, + "learning_rate": 7.66925319214622e-05, + "loss": 0.0177, + "step": 24528 + }, + { + "epoch": 0.5749983883917321, + "grad_norm": 0.42672714591026306, + "learning_rate": 7.668536869941081e-05, + "loss": 0.042, + "step": 24529 + }, + { + "epoch": 0.5750218299665372, + "grad_norm": 0.1345185786485672, + "learning_rate": 7.667820560386525e-05, + "loss": 0.0353, + "step": 24530 + }, + { + "epoch": 0.5750452715413421, + "grad_norm": 0.4660937488079071, + "learning_rate": 7.667104263486439e-05, + "loss": 0.0584, + "step": 24531 + }, + { + "epoch": 0.5750687131161472, + "grad_norm": 0.5808982253074646, + "learning_rate": 7.666387979244713e-05, + "loss": 0.6035, + "step": 24532 + }, + { + "epoch": 0.5750921546909521, + "grad_norm": 0.42961233854293823, + "learning_rate": 7.665671707665231e-05, + "loss": 0.1454, + "step": 24533 + }, + { + "epoch": 0.5751155962657571, + "grad_norm": 0.4108881652355194, + "learning_rate": 7.664955448751878e-05, + "loss": 0.0854, + "step": 24534 + }, + { + "epoch": 0.5751390378405621, + "grad_norm": 0.47001922130584717, + "learning_rate": 7.664239202508544e-05, + "loss": 0.0943, + "step": 24535 + }, + { + "epoch": 0.5751624794153671, + "grad_norm": 0.5153661966323853, + "learning_rate": 7.663522968939114e-05, + "loss": 0.137, + "step": 24536 + }, + { + "epoch": 0.5751859209901721, + "grad_norm": 0.3802528977394104, + "learning_rate": 7.662806748047473e-05, + "loss": 0.0688, + "step": 24537 + }, + { + "epoch": 0.5752093625649771, + "grad_norm": 0.5278369784355164, + "learning_rate": 7.662090539837508e-05, + "loss": 0.0667, + "step": 24538 + }, + { + "epoch": 0.5752328041397821, + "grad_norm": 0.5324403047561646, + "learning_rate": 7.661374344313104e-05, + "loss": 0.1131, + "step": 24539 + }, + { + "epoch": 0.5752562457145871, + "grad_norm": 0.26700538396835327, + "learning_rate": 7.660658161478151e-05, + "loss": 0.0499, + "step": 24540 + }, + { + "epoch": 0.575279687289392, + "grad_norm": 0.45559296011924744, + "learning_rate": 7.659941991336531e-05, + "loss": 0.1619, + "step": 24541 + }, + { + "epoch": 0.5753031288641971, + "grad_norm": 0.5221758484840393, + "learning_rate": 7.659225833892134e-05, + "loss": 0.1595, + "step": 24542 + }, + { + "epoch": 0.575326570439002, + "grad_norm": 0.5557602047920227, + "learning_rate": 7.658509689148843e-05, + "loss": 0.1049, + "step": 24543 + }, + { + "epoch": 0.5753500120138071, + "grad_norm": 0.18139676749706268, + "learning_rate": 7.657793557110542e-05, + "loss": 0.035, + "step": 24544 + }, + { + "epoch": 0.5753734535886121, + "grad_norm": 0.39683422446250916, + "learning_rate": 7.657077437781122e-05, + "loss": 0.0677, + "step": 24545 + }, + { + "epoch": 0.5753968951634171, + "grad_norm": 0.4022623300552368, + "learning_rate": 7.656361331164464e-05, + "loss": 0.0643, + "step": 24546 + }, + { + "epoch": 0.5754203367382221, + "grad_norm": 0.335915207862854, + "learning_rate": 7.655645237264456e-05, + "loss": 0.0516, + "step": 24547 + }, + { + "epoch": 0.5754437783130271, + "grad_norm": 0.8609957098960876, + "learning_rate": 7.654929156084982e-05, + "loss": 0.0554, + "step": 24548 + }, + { + "epoch": 0.5754672198878321, + "grad_norm": 0.770340085029602, + "learning_rate": 7.654213087629927e-05, + "loss": 0.1716, + "step": 24549 + }, + { + "epoch": 0.575490661462637, + "grad_norm": 0.5171152949333191, + "learning_rate": 7.653497031903181e-05, + "loss": 0.0751, + "step": 24550 + }, + { + "epoch": 0.5755141030374421, + "grad_norm": 0.5001634359359741, + "learning_rate": 7.652780988908621e-05, + "loss": 0.0827, + "step": 24551 + }, + { + "epoch": 0.575537544612247, + "grad_norm": 0.14954699575901031, + "learning_rate": 7.652064958650142e-05, + "loss": 0.0238, + "step": 24552 + }, + { + "epoch": 0.5755609861870521, + "grad_norm": 0.29227063059806824, + "learning_rate": 7.651348941131625e-05, + "loss": 0.0506, + "step": 24553 + }, + { + "epoch": 0.575584427761857, + "grad_norm": 0.4752928912639618, + "learning_rate": 7.65063293635695e-05, + "loss": 0.0705, + "step": 24554 + }, + { + "epoch": 0.5756078693366621, + "grad_norm": 0.6064231991767883, + "learning_rate": 7.649916944330012e-05, + "loss": 0.0827, + "step": 24555 + }, + { + "epoch": 0.575631310911467, + "grad_norm": 1.10816490650177, + "learning_rate": 7.649200965054687e-05, + "loss": 0.0539, + "step": 24556 + }, + { + "epoch": 0.5756547524862721, + "grad_norm": 0.13578781485557556, + "learning_rate": 7.648484998534866e-05, + "loss": 0.0262, + "step": 24557 + }, + { + "epoch": 0.575678194061077, + "grad_norm": 0.9476429224014282, + "learning_rate": 7.647769044774433e-05, + "loss": 0.1426, + "step": 24558 + }, + { + "epoch": 0.5757016356358821, + "grad_norm": 0.19401027262210846, + "learning_rate": 7.647053103777267e-05, + "loss": 0.0376, + "step": 24559 + }, + { + "epoch": 0.575725077210687, + "grad_norm": 0.5145561695098877, + "learning_rate": 7.64633717554726e-05, + "loss": 0.0882, + "step": 24560 + }, + { + "epoch": 0.575748518785492, + "grad_norm": 0.4850272834300995, + "learning_rate": 7.645621260088293e-05, + "loss": 0.0839, + "step": 24561 + }, + { + "epoch": 0.575771960360297, + "grad_norm": 0.16107097268104553, + "learning_rate": 7.64490535740425e-05, + "loss": 0.0319, + "step": 24562 + }, + { + "epoch": 0.575795401935102, + "grad_norm": 0.5953372120857239, + "learning_rate": 7.644189467499015e-05, + "loss": 0.58, + "step": 24563 + }, + { + "epoch": 0.575818843509907, + "grad_norm": 0.35064712166786194, + "learning_rate": 7.643473590376477e-05, + "loss": 0.111, + "step": 24564 + }, + { + "epoch": 0.575842285084712, + "grad_norm": 0.3819553256034851, + "learning_rate": 7.642757726040519e-05, + "loss": 0.0985, + "step": 24565 + }, + { + "epoch": 0.575865726659517, + "grad_norm": 0.35607030987739563, + "learning_rate": 7.64204187449502e-05, + "loss": 0.0516, + "step": 24566 + }, + { + "epoch": 0.575889168234322, + "grad_norm": 0.3679428994655609, + "learning_rate": 7.641326035743872e-05, + "loss": 0.0558, + "step": 24567 + }, + { + "epoch": 0.575912609809127, + "grad_norm": 0.665184497833252, + "learning_rate": 7.640610209790955e-05, + "loss": 0.6419, + "step": 24568 + }, + { + "epoch": 0.575936051383932, + "grad_norm": 0.28953632712364197, + "learning_rate": 7.63989439664015e-05, + "loss": 0.0383, + "step": 24569 + }, + { + "epoch": 0.5759594929587369, + "grad_norm": 0.6365277767181396, + "learning_rate": 7.639178596295347e-05, + "loss": 0.1064, + "step": 24570 + }, + { + "epoch": 0.575982934533542, + "grad_norm": 0.4023969769477844, + "learning_rate": 7.638462808760428e-05, + "loss": 0.0625, + "step": 24571 + }, + { + "epoch": 0.5760063761083469, + "grad_norm": 0.4705401659011841, + "learning_rate": 7.637747034039275e-05, + "loss": 0.0643, + "step": 24572 + }, + { + "epoch": 0.576029817683152, + "grad_norm": 0.08071931451559067, + "learning_rate": 7.637031272135772e-05, + "loss": 0.0177, + "step": 24573 + }, + { + "epoch": 0.5760532592579569, + "grad_norm": 0.19225333631038666, + "learning_rate": 7.636315523053806e-05, + "loss": 0.0277, + "step": 24574 + }, + { + "epoch": 0.576076700832762, + "grad_norm": 0.5907712578773499, + "learning_rate": 7.635599786797253e-05, + "loss": 0.6081, + "step": 24575 + }, + { + "epoch": 0.5761001424075669, + "grad_norm": 0.5347496867179871, + "learning_rate": 7.634884063370007e-05, + "loss": 0.1149, + "step": 24576 + }, + { + "epoch": 0.576123583982372, + "grad_norm": 0.6239833235740662, + "learning_rate": 7.634168352775945e-05, + "loss": 0.1479, + "step": 24577 + }, + { + "epoch": 0.5761470255571769, + "grad_norm": 0.5339388847351074, + "learning_rate": 7.633452655018955e-05, + "loss": 0.5293, + "step": 24578 + }, + { + "epoch": 0.576170467131982, + "grad_norm": 0.26687943935394287, + "learning_rate": 7.632736970102914e-05, + "loss": 0.0519, + "step": 24579 + }, + { + "epoch": 0.5761939087067869, + "grad_norm": 0.4654081463813782, + "learning_rate": 7.632021298031711e-05, + "loss": 0.1053, + "step": 24580 + }, + { + "epoch": 0.5762173502815919, + "grad_norm": 0.5344995856285095, + "learning_rate": 7.631305638809226e-05, + "loss": 0.1243, + "step": 24581 + }, + { + "epoch": 0.5762407918563969, + "grad_norm": 0.3170737624168396, + "learning_rate": 7.630589992439342e-05, + "loss": 0.0945, + "step": 24582 + }, + { + "epoch": 0.5762642334312019, + "grad_norm": 0.11159223318099976, + "learning_rate": 7.629874358925946e-05, + "loss": 0.0215, + "step": 24583 + }, + { + "epoch": 0.5762876750060069, + "grad_norm": 0.45032230019569397, + "learning_rate": 7.629158738272919e-05, + "loss": 0.0818, + "step": 24584 + }, + { + "epoch": 0.5763111165808119, + "grad_norm": 0.5527994632720947, + "learning_rate": 7.628443130484139e-05, + "loss": 0.0929, + "step": 24585 + }, + { + "epoch": 0.5763345581556168, + "grad_norm": 0.15477347373962402, + "learning_rate": 7.627727535563496e-05, + "loss": 0.0251, + "step": 24586 + }, + { + "epoch": 0.5763579997304219, + "grad_norm": 0.49961018562316895, + "learning_rate": 7.627011953514867e-05, + "loss": 0.1291, + "step": 24587 + }, + { + "epoch": 0.5763814413052268, + "grad_norm": 0.6851866841316223, + "learning_rate": 7.626296384342141e-05, + "loss": 0.5761, + "step": 24588 + }, + { + "epoch": 0.5764048828800319, + "grad_norm": 0.910254180431366, + "learning_rate": 7.625580828049194e-05, + "loss": 0.1502, + "step": 24589 + }, + { + "epoch": 0.5764283244548368, + "grad_norm": 0.43782615661621094, + "learning_rate": 7.624865284639917e-05, + "loss": 0.0639, + "step": 24590 + }, + { + "epoch": 0.5764517660296419, + "grad_norm": 0.4169093370437622, + "learning_rate": 7.624149754118187e-05, + "loss": 0.6045, + "step": 24591 + }, + { + "epoch": 0.5764752076044468, + "grad_norm": 0.44717782735824585, + "learning_rate": 7.623434236487885e-05, + "loss": 0.1025, + "step": 24592 + }, + { + "epoch": 0.5764986491792519, + "grad_norm": 0.3649124205112457, + "learning_rate": 7.622718731752897e-05, + "loss": 0.0957, + "step": 24593 + }, + { + "epoch": 0.5765220907540568, + "grad_norm": 0.5764058828353882, + "learning_rate": 7.622003239917103e-05, + "loss": 0.092, + "step": 24594 + }, + { + "epoch": 0.5765455323288619, + "grad_norm": 0.21670441329479218, + "learning_rate": 7.621287760984388e-05, + "loss": 0.0381, + "step": 24595 + }, + { + "epoch": 0.5765689739036669, + "grad_norm": 0.20315444469451904, + "learning_rate": 7.620572294958634e-05, + "loss": 0.0472, + "step": 24596 + }, + { + "epoch": 0.5765924154784718, + "grad_norm": 0.43352124094963074, + "learning_rate": 7.619856841843717e-05, + "loss": 0.0527, + "step": 24597 + }, + { + "epoch": 0.5766158570532769, + "grad_norm": 0.633392870426178, + "learning_rate": 7.619141401643527e-05, + "loss": 0.0832, + "step": 24598 + }, + { + "epoch": 0.5766392986280818, + "grad_norm": 0.1725105494260788, + "learning_rate": 7.618425974361938e-05, + "loss": 0.0447, + "step": 24599 + }, + { + "epoch": 0.5766627402028869, + "grad_norm": 0.4778499901294708, + "learning_rate": 7.617710560002842e-05, + "loss": 0.5813, + "step": 24600 + }, + { + "epoch": 0.5766861817776918, + "grad_norm": 0.2914709746837616, + "learning_rate": 7.616995158570113e-05, + "loss": 0.052, + "step": 24601 + }, + { + "epoch": 0.5767096233524969, + "grad_norm": 0.5359567999839783, + "learning_rate": 7.616279770067635e-05, + "loss": 0.1272, + "step": 24602 + }, + { + "epoch": 0.5767330649273018, + "grad_norm": 0.3036316931247711, + "learning_rate": 7.615564394499293e-05, + "loss": 0.0457, + "step": 24603 + }, + { + "epoch": 0.5767565065021069, + "grad_norm": 0.32081684470176697, + "learning_rate": 7.614849031868962e-05, + "loss": 0.0835, + "step": 24604 + }, + { + "epoch": 0.5767799480769118, + "grad_norm": 0.2719431221485138, + "learning_rate": 7.61413368218053e-05, + "loss": 0.0384, + "step": 24605 + }, + { + "epoch": 0.5768033896517168, + "grad_norm": 0.32233163714408875, + "learning_rate": 7.613418345437876e-05, + "loss": 0.045, + "step": 24606 + }, + { + "epoch": 0.5768268312265218, + "grad_norm": 0.40898364782333374, + "learning_rate": 7.612703021644878e-05, + "loss": 0.0946, + "step": 24607 + }, + { + "epoch": 0.5768502728013268, + "grad_norm": 0.36341753602027893, + "learning_rate": 7.611987710805424e-05, + "loss": 0.0454, + "step": 24608 + }, + { + "epoch": 0.5768737143761318, + "grad_norm": 0.5389441847801208, + "learning_rate": 7.61127241292339e-05, + "loss": 0.0855, + "step": 24609 + }, + { + "epoch": 0.5768971559509368, + "grad_norm": 0.29091277718544006, + "learning_rate": 7.610557128002658e-05, + "loss": 0.0487, + "step": 24610 + }, + { + "epoch": 0.5769205975257418, + "grad_norm": 0.46717965602874756, + "learning_rate": 7.609841856047109e-05, + "loss": 0.1352, + "step": 24611 + }, + { + "epoch": 0.5769440391005468, + "grad_norm": 0.5082519054412842, + "learning_rate": 7.609126597060628e-05, + "loss": 0.428, + "step": 24612 + }, + { + "epoch": 0.5769674806753518, + "grad_norm": 0.15458084642887115, + "learning_rate": 7.608411351047093e-05, + "loss": 0.0199, + "step": 24613 + }, + { + "epoch": 0.5769909222501568, + "grad_norm": 0.8312186598777771, + "learning_rate": 7.607696118010386e-05, + "loss": 0.1651, + "step": 24614 + }, + { + "epoch": 0.5770143638249617, + "grad_norm": 0.07400254160165787, + "learning_rate": 7.606980897954387e-05, + "loss": 0.0177, + "step": 24615 + }, + { + "epoch": 0.5770378053997668, + "grad_norm": 0.5401853919029236, + "learning_rate": 7.606265690882976e-05, + "loss": 0.1036, + "step": 24616 + }, + { + "epoch": 0.5770612469745717, + "grad_norm": 0.4181540310382843, + "learning_rate": 7.605550496800035e-05, + "loss": 0.0838, + "step": 24617 + }, + { + "epoch": 0.5770846885493768, + "grad_norm": 0.516316831111908, + "learning_rate": 7.604835315709444e-05, + "loss": 0.1287, + "step": 24618 + }, + { + "epoch": 0.5771081301241817, + "grad_norm": 0.2302757352590561, + "learning_rate": 7.604120147615085e-05, + "loss": 0.0463, + "step": 24619 + }, + { + "epoch": 0.5771315716989868, + "grad_norm": 0.15591886639595032, + "learning_rate": 7.603404992520836e-05, + "loss": 0.0222, + "step": 24620 + }, + { + "epoch": 0.5771550132737917, + "grad_norm": 0.43921831250190735, + "learning_rate": 7.60268985043058e-05, + "loss": 0.0898, + "step": 24621 + }, + { + "epoch": 0.5771784548485968, + "grad_norm": 0.2972275912761688, + "learning_rate": 7.601974721348197e-05, + "loss": 0.05, + "step": 24622 + }, + { + "epoch": 0.5772018964234017, + "grad_norm": 0.43550023436546326, + "learning_rate": 7.601259605277565e-05, + "loss": 0.0413, + "step": 24623 + }, + { + "epoch": 0.5772253379982067, + "grad_norm": 0.1386755406856537, + "learning_rate": 7.600544502222561e-05, + "loss": 0.0272, + "step": 24624 + }, + { + "epoch": 0.5772487795730117, + "grad_norm": 0.24758046865463257, + "learning_rate": 7.599829412187077e-05, + "loss": 0.0946, + "step": 24625 + }, + { + "epoch": 0.5772722211478167, + "grad_norm": 0.38359320163726807, + "learning_rate": 7.599114335174986e-05, + "loss": 0.0414, + "step": 24626 + }, + { + "epoch": 0.5772956627226217, + "grad_norm": 0.46171385049819946, + "learning_rate": 7.598399271190165e-05, + "loss": 0.0641, + "step": 24627 + }, + { + "epoch": 0.5773191042974267, + "grad_norm": 0.7158359885215759, + "learning_rate": 7.5976842202365e-05, + "loss": 0.1603, + "step": 24628 + }, + { + "epoch": 0.5773425458722317, + "grad_norm": 0.5621856451034546, + "learning_rate": 7.596969182317866e-05, + "loss": 0.089, + "step": 24629 + }, + { + "epoch": 0.5773659874470367, + "grad_norm": 0.5189867615699768, + "learning_rate": 7.596254157438145e-05, + "loss": 0.1277, + "step": 24630 + }, + { + "epoch": 0.5773894290218416, + "grad_norm": 0.33054983615875244, + "learning_rate": 7.595539145601218e-05, + "loss": 0.047, + "step": 24631 + }, + { + "epoch": 0.5774128705966467, + "grad_norm": 0.4891555905342102, + "learning_rate": 7.59482414681096e-05, + "loss": 0.1132, + "step": 24632 + }, + { + "epoch": 0.5774363121714516, + "grad_norm": 0.5552834272384644, + "learning_rate": 7.594109161071258e-05, + "loss": 0.0479, + "step": 24633 + }, + { + "epoch": 0.5774597537462567, + "grad_norm": 0.1661355197429657, + "learning_rate": 7.593394188385987e-05, + "loss": 0.0193, + "step": 24634 + }, + { + "epoch": 0.5774831953210616, + "grad_norm": 0.6536083221435547, + "learning_rate": 7.592679228759022e-05, + "loss": 0.0749, + "step": 24635 + }, + { + "epoch": 0.5775066368958667, + "grad_norm": 0.19311828911304474, + "learning_rate": 7.591964282194247e-05, + "loss": 0.0166, + "step": 24636 + }, + { + "epoch": 0.5775300784706716, + "grad_norm": 0.3451695442199707, + "learning_rate": 7.591249348695546e-05, + "loss": 0.2911, + "step": 24637 + }, + { + "epoch": 0.5775535200454767, + "grad_norm": 0.1397494077682495, + "learning_rate": 7.590534428266794e-05, + "loss": 0.022, + "step": 24638 + }, + { + "epoch": 0.5775769616202816, + "grad_norm": 0.12628677487373352, + "learning_rate": 7.589819520911866e-05, + "loss": 0.0142, + "step": 24639 + }, + { + "epoch": 0.5776004031950867, + "grad_norm": 0.34963107109069824, + "learning_rate": 7.589104626634649e-05, + "loss": 0.0778, + "step": 24640 + }, + { + "epoch": 0.5776238447698916, + "grad_norm": 0.46394699811935425, + "learning_rate": 7.588389745439017e-05, + "loss": 0.0932, + "step": 24641 + }, + { + "epoch": 0.5776472863446966, + "grad_norm": 0.4558960497379303, + "learning_rate": 7.587674877328849e-05, + "loss": 0.5687, + "step": 24642 + }, + { + "epoch": 0.5776707279195016, + "grad_norm": 0.1286415159702301, + "learning_rate": 7.586960022308026e-05, + "loss": 0.0227, + "step": 24643 + }, + { + "epoch": 0.5776941694943066, + "grad_norm": 0.48252740502357483, + "learning_rate": 7.586245180380428e-05, + "loss": 0.0954, + "step": 24644 + }, + { + "epoch": 0.5777176110691116, + "grad_norm": 0.30644556879997253, + "learning_rate": 7.585530351549926e-05, + "loss": 0.0815, + "step": 24645 + }, + { + "epoch": 0.5777410526439166, + "grad_norm": 0.2134048491716385, + "learning_rate": 7.584815535820408e-05, + "loss": 0.0522, + "step": 24646 + }, + { + "epoch": 0.5777644942187217, + "grad_norm": 0.4600210189819336, + "learning_rate": 7.584100733195749e-05, + "loss": 0.0472, + "step": 24647 + }, + { + "epoch": 0.5777879357935266, + "grad_norm": 0.460714727640152, + "learning_rate": 7.583385943679822e-05, + "loss": 0.1255, + "step": 24648 + }, + { + "epoch": 0.5778113773683317, + "grad_norm": 0.16698551177978516, + "learning_rate": 7.582671167276514e-05, + "loss": 0.0234, + "step": 24649 + }, + { + "epoch": 0.5778348189431366, + "grad_norm": 0.2743982672691345, + "learning_rate": 7.581956403989702e-05, + "loss": 0.044, + "step": 24650 + }, + { + "epoch": 0.5778582605179416, + "grad_norm": 0.7604656219482422, + "learning_rate": 7.581241653823263e-05, + "loss": 0.1208, + "step": 24651 + }, + { + "epoch": 0.5778817020927466, + "grad_norm": 0.49545279145240784, + "learning_rate": 7.580526916781072e-05, + "loss": 0.1317, + "step": 24652 + }, + { + "epoch": 0.5779051436675516, + "grad_norm": 0.2940872311592102, + "learning_rate": 7.579812192867013e-05, + "loss": 0.0595, + "step": 24653 + }, + { + "epoch": 0.5779285852423566, + "grad_norm": 0.7018158435821533, + "learning_rate": 7.57909748208496e-05, + "loss": 0.1212, + "step": 24654 + }, + { + "epoch": 0.5779520268171616, + "grad_norm": 0.2875232696533203, + "learning_rate": 7.57838278443879e-05, + "loss": 0.0584, + "step": 24655 + }, + { + "epoch": 0.5779754683919666, + "grad_norm": 0.4180147647857666, + "learning_rate": 7.577668099932387e-05, + "loss": 0.0793, + "step": 24656 + }, + { + "epoch": 0.5779989099667716, + "grad_norm": 0.19428615272045135, + "learning_rate": 7.576953428569624e-05, + "loss": 0.0278, + "step": 24657 + }, + { + "epoch": 0.5780223515415766, + "grad_norm": 0.1369451880455017, + "learning_rate": 7.576238770354377e-05, + "loss": 0.0179, + "step": 24658 + }, + { + "epoch": 0.5780457931163816, + "grad_norm": 0.3504458963871002, + "learning_rate": 7.575524125290529e-05, + "loss": 0.0866, + "step": 24659 + }, + { + "epoch": 0.5780692346911865, + "grad_norm": 1.0217034816741943, + "learning_rate": 7.574809493381951e-05, + "loss": 0.1407, + "step": 24660 + }, + { + "epoch": 0.5780926762659916, + "grad_norm": 0.2598113715648651, + "learning_rate": 7.57409487463253e-05, + "loss": 0.0217, + "step": 24661 + }, + { + "epoch": 0.5781161178407965, + "grad_norm": 0.12730038166046143, + "learning_rate": 7.573380269046136e-05, + "loss": 0.0291, + "step": 24662 + }, + { + "epoch": 0.5781395594156016, + "grad_norm": 0.6598436236381531, + "learning_rate": 7.57266567662665e-05, + "loss": 0.1082, + "step": 24663 + }, + { + "epoch": 0.5781630009904065, + "grad_norm": 0.3170018196105957, + "learning_rate": 7.571951097377948e-05, + "loss": 0.0784, + "step": 24664 + }, + { + "epoch": 0.5781864425652116, + "grad_norm": 0.0963081493973732, + "learning_rate": 7.571236531303907e-05, + "loss": 0.0113, + "step": 24665 + }, + { + "epoch": 0.5782098841400165, + "grad_norm": 0.26504814624786377, + "learning_rate": 7.570521978408407e-05, + "loss": 0.0435, + "step": 24666 + }, + { + "epoch": 0.5782333257148216, + "grad_norm": 0.21043726801872253, + "learning_rate": 7.569807438695323e-05, + "loss": 0.0511, + "step": 24667 + }, + { + "epoch": 0.5782567672896265, + "grad_norm": 0.46654677391052246, + "learning_rate": 7.569092912168529e-05, + "loss": 0.1073, + "step": 24668 + }, + { + "epoch": 0.5782802088644315, + "grad_norm": 0.1933831125497818, + "learning_rate": 7.568378398831908e-05, + "loss": 0.0409, + "step": 24669 + }, + { + "epoch": 0.5783036504392365, + "grad_norm": 0.590798556804657, + "learning_rate": 7.567663898689332e-05, + "loss": 0.0944, + "step": 24670 + }, + { + "epoch": 0.5783270920140415, + "grad_norm": 0.224823996424675, + "learning_rate": 7.566949411744681e-05, + "loss": 0.0248, + "step": 24671 + }, + { + "epoch": 0.5783505335888465, + "grad_norm": 0.4957246482372284, + "learning_rate": 7.566234938001828e-05, + "loss": 0.0883, + "step": 24672 + }, + { + "epoch": 0.5783739751636515, + "grad_norm": 0.5168253183364868, + "learning_rate": 7.565520477464657e-05, + "loss": 0.0842, + "step": 24673 + }, + { + "epoch": 0.5783974167384565, + "grad_norm": 0.6683458089828491, + "learning_rate": 7.564806030137037e-05, + "loss": 0.1445, + "step": 24674 + }, + { + "epoch": 0.5784208583132615, + "grad_norm": 0.5522992014884949, + "learning_rate": 7.564091596022851e-05, + "loss": 0.0715, + "step": 24675 + }, + { + "epoch": 0.5784442998880664, + "grad_norm": 0.39064955711364746, + "learning_rate": 7.563377175125973e-05, + "loss": 0.0948, + "step": 24676 + }, + { + "epoch": 0.5784677414628715, + "grad_norm": 0.5457354784011841, + "learning_rate": 7.562662767450276e-05, + "loss": 0.1017, + "step": 24677 + }, + { + "epoch": 0.5784911830376764, + "grad_norm": 0.446569561958313, + "learning_rate": 7.561948372999642e-05, + "loss": 0.0903, + "step": 24678 + }, + { + "epoch": 0.5785146246124815, + "grad_norm": 0.2631627321243286, + "learning_rate": 7.561233991777945e-05, + "loss": 0.0523, + "step": 24679 + }, + { + "epoch": 0.5785380661872864, + "grad_norm": 0.1441318839788437, + "learning_rate": 7.560519623789058e-05, + "loss": 0.0247, + "step": 24680 + }, + { + "epoch": 0.5785615077620915, + "grad_norm": 0.29972755908966064, + "learning_rate": 7.559805269036862e-05, + "loss": 0.0193, + "step": 24681 + }, + { + "epoch": 0.5785849493368964, + "grad_norm": 0.3611668050289154, + "learning_rate": 7.559090927525231e-05, + "loss": 0.0778, + "step": 24682 + }, + { + "epoch": 0.5786083909117015, + "grad_norm": 0.1547095775604248, + "learning_rate": 7.558376599258041e-05, + "loss": 0.0402, + "step": 24683 + }, + { + "epoch": 0.5786318324865064, + "grad_norm": 0.7106608152389526, + "learning_rate": 7.557662284239164e-05, + "loss": 0.0689, + "step": 24684 + }, + { + "epoch": 0.5786552740613115, + "grad_norm": 0.511314332485199, + "learning_rate": 7.556947982472487e-05, + "loss": 0.0869, + "step": 24685 + }, + { + "epoch": 0.5786787156361164, + "grad_norm": 0.5192530751228333, + "learning_rate": 7.556233693961876e-05, + "loss": 0.0631, + "step": 24686 + }, + { + "epoch": 0.5787021572109214, + "grad_norm": 0.5849196910858154, + "learning_rate": 7.555519418711209e-05, + "loss": 0.0599, + "step": 24687 + }, + { + "epoch": 0.5787255987857264, + "grad_norm": 0.6375990509986877, + "learning_rate": 7.554805156724366e-05, + "loss": 0.1043, + "step": 24688 + }, + { + "epoch": 0.5787490403605314, + "grad_norm": 0.20855212211608887, + "learning_rate": 7.554090908005216e-05, + "loss": 0.0357, + "step": 24689 + }, + { + "epoch": 0.5787724819353364, + "grad_norm": 0.2807692587375641, + "learning_rate": 7.553376672557636e-05, + "loss": 0.0498, + "step": 24690 + }, + { + "epoch": 0.5787959235101414, + "grad_norm": 0.37084904313087463, + "learning_rate": 7.552662450385506e-05, + "loss": 0.0339, + "step": 24691 + }, + { + "epoch": 0.5788193650849464, + "grad_norm": 0.09444868564605713, + "learning_rate": 7.551948241492698e-05, + "loss": 0.0132, + "step": 24692 + }, + { + "epoch": 0.5788428066597514, + "grad_norm": 0.6614250540733337, + "learning_rate": 7.551234045883086e-05, + "loss": 0.0793, + "step": 24693 + }, + { + "epoch": 0.5788662482345563, + "grad_norm": 0.5651975870132446, + "learning_rate": 7.550519863560548e-05, + "loss": 0.0723, + "step": 24694 + }, + { + "epoch": 0.5788896898093614, + "grad_norm": 0.40550145506858826, + "learning_rate": 7.54980569452896e-05, + "loss": 0.0711, + "step": 24695 + }, + { + "epoch": 0.5789131313841663, + "grad_norm": 0.528843879699707, + "learning_rate": 7.549091538792191e-05, + "loss": 0.0944, + "step": 24696 + }, + { + "epoch": 0.5789365729589714, + "grad_norm": 0.5247714519500732, + "learning_rate": 7.548377396354119e-05, + "loss": 0.1065, + "step": 24697 + }, + { + "epoch": 0.5789600145337763, + "grad_norm": 0.17508679628372192, + "learning_rate": 7.547663267218624e-05, + "loss": 0.0248, + "step": 24698 + }, + { + "epoch": 0.5789834561085814, + "grad_norm": 0.49665457010269165, + "learning_rate": 7.546949151389578e-05, + "loss": 0.1083, + "step": 24699 + }, + { + "epoch": 0.5790068976833864, + "grad_norm": 0.18238584697246552, + "learning_rate": 7.546235048870852e-05, + "loss": 0.0154, + "step": 24700 + }, + { + "epoch": 0.5790303392581914, + "grad_norm": 0.49477270245552063, + "learning_rate": 7.545520959666326e-05, + "loss": 0.0674, + "step": 24701 + }, + { + "epoch": 0.5790537808329964, + "grad_norm": 1.0243974924087524, + "learning_rate": 7.544806883779873e-05, + "loss": 0.1189, + "step": 24702 + }, + { + "epoch": 0.5790772224078014, + "grad_norm": 0.3514276146888733, + "learning_rate": 7.544092821215363e-05, + "loss": 0.0438, + "step": 24703 + }, + { + "epoch": 0.5791006639826064, + "grad_norm": 0.6141146421432495, + "learning_rate": 7.543378771976678e-05, + "loss": 0.1403, + "step": 24704 + }, + { + "epoch": 0.5791241055574113, + "grad_norm": 0.31497129797935486, + "learning_rate": 7.542664736067686e-05, + "loss": 0.0564, + "step": 24705 + }, + { + "epoch": 0.5791475471322164, + "grad_norm": 0.5286357998847961, + "learning_rate": 7.541950713492267e-05, + "loss": 0.0861, + "step": 24706 + }, + { + "epoch": 0.5791709887070213, + "grad_norm": 0.11078224331140518, + "learning_rate": 7.541236704254291e-05, + "loss": 0.0202, + "step": 24707 + }, + { + "epoch": 0.5791944302818264, + "grad_norm": 0.5230551958084106, + "learning_rate": 7.540522708357634e-05, + "loss": 0.1491, + "step": 24708 + }, + { + "epoch": 0.5792178718566313, + "grad_norm": 0.44351881742477417, + "learning_rate": 7.539808725806168e-05, + "loss": 0.0732, + "step": 24709 + }, + { + "epoch": 0.5792413134314364, + "grad_norm": 0.11123175919055939, + "learning_rate": 7.53909475660377e-05, + "loss": 0.0212, + "step": 24710 + }, + { + "epoch": 0.5792647550062413, + "grad_norm": 0.4385516345500946, + "learning_rate": 7.538380800754315e-05, + "loss": 0.1118, + "step": 24711 + }, + { + "epoch": 0.5792881965810464, + "grad_norm": 0.6809009909629822, + "learning_rate": 7.537666858261674e-05, + "loss": 0.1691, + "step": 24712 + }, + { + "epoch": 0.5793116381558513, + "grad_norm": 0.40288493037223816, + "learning_rate": 7.536952929129722e-05, + "loss": 0.1218, + "step": 24713 + }, + { + "epoch": 0.5793350797306563, + "grad_norm": 0.720177948474884, + "learning_rate": 7.536239013362334e-05, + "loss": 0.0923, + "step": 24714 + }, + { + "epoch": 0.5793585213054613, + "grad_norm": 0.5666874051094055, + "learning_rate": 7.535525110963379e-05, + "loss": 0.2166, + "step": 24715 + }, + { + "epoch": 0.5793819628802663, + "grad_norm": 0.30449432134628296, + "learning_rate": 7.534811221936737e-05, + "loss": 0.0435, + "step": 24716 + }, + { + "epoch": 0.5794054044550713, + "grad_norm": 0.1573067009449005, + "learning_rate": 7.534097346286279e-05, + "loss": 0.0176, + "step": 24717 + }, + { + "epoch": 0.5794288460298763, + "grad_norm": 0.4056069850921631, + "learning_rate": 7.533383484015876e-05, + "loss": 0.0842, + "step": 24718 + }, + { + "epoch": 0.5794522876046813, + "grad_norm": 0.35600170493125916, + "learning_rate": 7.532669635129404e-05, + "loss": 0.0817, + "step": 24719 + }, + { + "epoch": 0.5794757291794863, + "grad_norm": 0.5543004274368286, + "learning_rate": 7.531955799630738e-05, + "loss": 0.154, + "step": 24720 + }, + { + "epoch": 0.5794991707542912, + "grad_norm": 0.21183215081691742, + "learning_rate": 7.531241977523744e-05, + "loss": 0.0465, + "step": 24721 + }, + { + "epoch": 0.5795226123290963, + "grad_norm": 0.20162075757980347, + "learning_rate": 7.530528168812303e-05, + "loss": 0.0277, + "step": 24722 + }, + { + "epoch": 0.5795460539039012, + "grad_norm": 0.2611235976219177, + "learning_rate": 7.529814373500287e-05, + "loss": 0.0307, + "step": 24723 + }, + { + "epoch": 0.5795694954787063, + "grad_norm": 0.4147534668445587, + "learning_rate": 7.529100591591567e-05, + "loss": 0.0625, + "step": 24724 + }, + { + "epoch": 0.5795929370535112, + "grad_norm": 0.37755128741264343, + "learning_rate": 7.528386823090017e-05, + "loss": 0.0821, + "step": 24725 + }, + { + "epoch": 0.5796163786283163, + "grad_norm": 0.4105544984340668, + "learning_rate": 7.52767306799951e-05, + "loss": 0.0708, + "step": 24726 + }, + { + "epoch": 0.5796398202031212, + "grad_norm": 0.5915250778198242, + "learning_rate": 7.52695932632392e-05, + "loss": 0.5478, + "step": 24727 + }, + { + "epoch": 0.5796632617779263, + "grad_norm": 0.4946793019771576, + "learning_rate": 7.526245598067115e-05, + "loss": 0.0554, + "step": 24728 + }, + { + "epoch": 0.5796867033527312, + "grad_norm": 0.0754314735531807, + "learning_rate": 7.525531883232973e-05, + "loss": 0.0114, + "step": 24729 + }, + { + "epoch": 0.5797101449275363, + "grad_norm": 0.4581564962863922, + "learning_rate": 7.524818181825364e-05, + "loss": 0.3486, + "step": 24730 + }, + { + "epoch": 0.5797335865023412, + "grad_norm": 0.34474384784698486, + "learning_rate": 7.524104493848159e-05, + "loss": 0.0315, + "step": 24731 + }, + { + "epoch": 0.5797570280771462, + "grad_norm": 1.0747548341751099, + "learning_rate": 7.523390819305235e-05, + "loss": 0.1485, + "step": 24732 + }, + { + "epoch": 0.5797804696519512, + "grad_norm": 0.6840550899505615, + "learning_rate": 7.522677158200459e-05, + "loss": 0.075, + "step": 24733 + }, + { + "epoch": 0.5798039112267562, + "grad_norm": 0.42669758200645447, + "learning_rate": 7.52196351053771e-05, + "loss": 0.0715, + "step": 24734 + }, + { + "epoch": 0.5798273528015612, + "grad_norm": 0.13862508535385132, + "learning_rate": 7.521249876320853e-05, + "loss": 0.0303, + "step": 24735 + }, + { + "epoch": 0.5798507943763662, + "grad_norm": 0.45735403895378113, + "learning_rate": 7.520536255553767e-05, + "loss": 0.0306, + "step": 24736 + }, + { + "epoch": 0.5798742359511712, + "grad_norm": 0.7761565446853638, + "learning_rate": 7.519822648240322e-05, + "loss": 0.8122, + "step": 24737 + }, + { + "epoch": 0.5798976775259762, + "grad_norm": 0.4298253059387207, + "learning_rate": 7.519109054384387e-05, + "loss": 0.0914, + "step": 24738 + }, + { + "epoch": 0.5799211191007811, + "grad_norm": 0.5862213969230652, + "learning_rate": 7.518395473989836e-05, + "loss": 0.1116, + "step": 24739 + }, + { + "epoch": 0.5799445606755862, + "grad_norm": 0.5147178769111633, + "learning_rate": 7.517681907060543e-05, + "loss": 0.0825, + "step": 24740 + }, + { + "epoch": 0.5799680022503911, + "grad_norm": 0.38755252957344055, + "learning_rate": 7.516968353600376e-05, + "loss": 0.0942, + "step": 24741 + }, + { + "epoch": 0.5799914438251962, + "grad_norm": 0.4011104106903076, + "learning_rate": 7.516254813613209e-05, + "loss": 0.0587, + "step": 24742 + }, + { + "epoch": 0.5800148854000011, + "grad_norm": 1.0238234996795654, + "learning_rate": 7.515541287102912e-05, + "loss": 0.0985, + "step": 24743 + }, + { + "epoch": 0.5800383269748062, + "grad_norm": 0.486695796251297, + "learning_rate": 7.514827774073359e-05, + "loss": 0.1539, + "step": 24744 + }, + { + "epoch": 0.5800617685496111, + "grad_norm": 0.3219977617263794, + "learning_rate": 7.514114274528417e-05, + "loss": 0.0456, + "step": 24745 + }, + { + "epoch": 0.5800852101244162, + "grad_norm": 0.37601983547210693, + "learning_rate": 7.513400788471965e-05, + "loss": 0.0456, + "step": 24746 + }, + { + "epoch": 0.5801086516992211, + "grad_norm": 0.25316664576530457, + "learning_rate": 7.512687315907869e-05, + "loss": 0.0281, + "step": 24747 + }, + { + "epoch": 0.5801320932740262, + "grad_norm": 0.7169998288154602, + "learning_rate": 7.511973856840001e-05, + "loss": 0.1045, + "step": 24748 + }, + { + "epoch": 0.5801555348488311, + "grad_norm": 0.10274633020162582, + "learning_rate": 7.511260411272235e-05, + "loss": 0.007, + "step": 24749 + }, + { + "epoch": 0.5801789764236361, + "grad_norm": 0.150557741522789, + "learning_rate": 7.510546979208436e-05, + "loss": 0.0126, + "step": 24750 + }, + { + "epoch": 0.5802024179984412, + "grad_norm": 0.09877502173185349, + "learning_rate": 7.509833560652483e-05, + "loss": 0.0144, + "step": 24751 + }, + { + "epoch": 0.5802258595732461, + "grad_norm": 0.6700955033302307, + "learning_rate": 7.509120155608244e-05, + "loss": 0.1701, + "step": 24752 + }, + { + "epoch": 0.5802493011480512, + "grad_norm": 0.6353602409362793, + "learning_rate": 7.508406764079585e-05, + "loss": 0.2233, + "step": 24753 + }, + { + "epoch": 0.5802727427228561, + "grad_norm": 0.4212462604045868, + "learning_rate": 7.507693386070384e-05, + "loss": 0.0975, + "step": 24754 + }, + { + "epoch": 0.5802961842976612, + "grad_norm": 0.44536617398262024, + "learning_rate": 7.50698002158451e-05, + "loss": 0.092, + "step": 24755 + }, + { + "epoch": 0.5803196258724661, + "grad_norm": 0.8062998652458191, + "learning_rate": 7.506266670625828e-05, + "loss": 0.1328, + "step": 24756 + }, + { + "epoch": 0.5803430674472712, + "grad_norm": 0.5121866464614868, + "learning_rate": 7.505553333198211e-05, + "loss": 0.0783, + "step": 24757 + }, + { + "epoch": 0.5803665090220761, + "grad_norm": 0.4889107346534729, + "learning_rate": 7.504840009305537e-05, + "loss": 0.1224, + "step": 24758 + }, + { + "epoch": 0.5803899505968811, + "grad_norm": 0.327802836894989, + "learning_rate": 7.50412669895167e-05, + "loss": 0.0825, + "step": 24759 + }, + { + "epoch": 0.5804133921716861, + "grad_norm": 0.44871607422828674, + "learning_rate": 7.503413402140483e-05, + "loss": 0.0842, + "step": 24760 + }, + { + "epoch": 0.5804368337464911, + "grad_norm": 0.5856935381889343, + "learning_rate": 7.502700118875844e-05, + "loss": 0.6022, + "step": 24761 + }, + { + "epoch": 0.5804602753212961, + "grad_norm": 0.2948419451713562, + "learning_rate": 7.501986849161625e-05, + "loss": 0.0429, + "step": 24762 + }, + { + "epoch": 0.5804837168961011, + "grad_norm": 0.7035495042800903, + "learning_rate": 7.501273593001694e-05, + "loss": 0.115, + "step": 24763 + }, + { + "epoch": 0.5805071584709061, + "grad_norm": 0.4408547580242157, + "learning_rate": 7.500560350399925e-05, + "loss": 0.0651, + "step": 24764 + }, + { + "epoch": 0.5805306000457111, + "grad_norm": 0.4444155991077423, + "learning_rate": 7.499847121360185e-05, + "loss": 0.0365, + "step": 24765 + }, + { + "epoch": 0.580554041620516, + "grad_norm": 0.5643434524536133, + "learning_rate": 7.499133905886342e-05, + "loss": 0.0774, + "step": 24766 + }, + { + "epoch": 0.5805774831953211, + "grad_norm": 0.41621842980384827, + "learning_rate": 7.498420703982274e-05, + "loss": 0.0674, + "step": 24767 + }, + { + "epoch": 0.580600924770126, + "grad_norm": 0.6520595550537109, + "learning_rate": 7.497707515651843e-05, + "loss": 0.6865, + "step": 24768 + }, + { + "epoch": 0.5806243663449311, + "grad_norm": 0.5113492608070374, + "learning_rate": 7.49699434089892e-05, + "loss": 0.0681, + "step": 24769 + }, + { + "epoch": 0.580647807919736, + "grad_norm": 0.4739570617675781, + "learning_rate": 7.496281179727374e-05, + "loss": 0.0419, + "step": 24770 + }, + { + "epoch": 0.5806712494945411, + "grad_norm": 0.6798104047775269, + "learning_rate": 7.495568032141081e-05, + "loss": 0.0666, + "step": 24771 + }, + { + "epoch": 0.580694691069346, + "grad_norm": 0.5008869171142578, + "learning_rate": 7.494854898143906e-05, + "loss": 0.1112, + "step": 24772 + }, + { + "epoch": 0.5807181326441511, + "grad_norm": 0.49678587913513184, + "learning_rate": 7.494141777739716e-05, + "loss": 0.0645, + "step": 24773 + }, + { + "epoch": 0.580741574218956, + "grad_norm": 0.5382492542266846, + "learning_rate": 7.493428670932387e-05, + "loss": 0.6018, + "step": 24774 + }, + { + "epoch": 0.580765015793761, + "grad_norm": 0.16236107051372528, + "learning_rate": 7.492715577725783e-05, + "loss": 0.0216, + "step": 24775 + }, + { + "epoch": 0.580788457368566, + "grad_norm": 0.6681550145149231, + "learning_rate": 7.492002498123774e-05, + "loss": 0.8193, + "step": 24776 + }, + { + "epoch": 0.580811898943371, + "grad_norm": 0.7192928194999695, + "learning_rate": 7.49128943213023e-05, + "loss": 0.1217, + "step": 24777 + }, + { + "epoch": 0.580835340518176, + "grad_norm": 0.5904276371002197, + "learning_rate": 7.490576379749021e-05, + "loss": 0.0625, + "step": 24778 + }, + { + "epoch": 0.580858782092981, + "grad_norm": 0.3552895486354828, + "learning_rate": 7.489863340984014e-05, + "loss": 0.0629, + "step": 24779 + }, + { + "epoch": 0.580882223667786, + "grad_norm": 0.28420376777648926, + "learning_rate": 7.48915031583908e-05, + "loss": 0.0669, + "step": 24780 + }, + { + "epoch": 0.580905665242591, + "grad_norm": 0.5467908978462219, + "learning_rate": 7.488437304318086e-05, + "loss": 0.0995, + "step": 24781 + }, + { + "epoch": 0.580929106817396, + "grad_norm": 0.1433703899383545, + "learning_rate": 7.487724306424902e-05, + "loss": 0.0212, + "step": 24782 + }, + { + "epoch": 0.580952548392201, + "grad_norm": 0.2506195306777954, + "learning_rate": 7.487011322163395e-05, + "loss": 0.0375, + "step": 24783 + }, + { + "epoch": 0.580975989967006, + "grad_norm": 0.6275935173034668, + "learning_rate": 7.486298351537437e-05, + "loss": 0.1293, + "step": 24784 + }, + { + "epoch": 0.580999431541811, + "grad_norm": 0.5704739093780518, + "learning_rate": 7.485585394550896e-05, + "loss": 0.7237, + "step": 24785 + }, + { + "epoch": 0.5810228731166159, + "grad_norm": 0.1776757687330246, + "learning_rate": 7.484872451207639e-05, + "loss": 0.0294, + "step": 24786 + }, + { + "epoch": 0.581046314691421, + "grad_norm": 0.35376816987991333, + "learning_rate": 7.484159521511535e-05, + "loss": 0.0575, + "step": 24787 + }, + { + "epoch": 0.5810697562662259, + "grad_norm": 0.4584321677684784, + "learning_rate": 7.483446605466449e-05, + "loss": 0.0601, + "step": 24788 + }, + { + "epoch": 0.581093197841031, + "grad_norm": 0.18544290959835052, + "learning_rate": 7.482733703076257e-05, + "loss": 0.0371, + "step": 24789 + }, + { + "epoch": 0.5811166394158359, + "grad_norm": 0.4207074046134949, + "learning_rate": 7.482020814344822e-05, + "loss": 0.0941, + "step": 24790 + }, + { + "epoch": 0.581140080990641, + "grad_norm": 0.3966217041015625, + "learning_rate": 7.481307939276008e-05, + "loss": 0.036, + "step": 24791 + }, + { + "epoch": 0.5811635225654459, + "grad_norm": 0.46470513939857483, + "learning_rate": 7.480595077873694e-05, + "loss": 0.0664, + "step": 24792 + }, + { + "epoch": 0.581186964140251, + "grad_norm": 0.586839497089386, + "learning_rate": 7.47988223014174e-05, + "loss": 0.1445, + "step": 24793 + }, + { + "epoch": 0.5812104057150559, + "grad_norm": 0.5969697833061218, + "learning_rate": 7.479169396084011e-05, + "loss": 0.1339, + "step": 24794 + }, + { + "epoch": 0.5812338472898609, + "grad_norm": 0.4114648997783661, + "learning_rate": 7.478456575704383e-05, + "loss": 0.4866, + "step": 24795 + }, + { + "epoch": 0.5812572888646659, + "grad_norm": 0.7140516638755798, + "learning_rate": 7.477743769006722e-05, + "loss": 0.5732, + "step": 24796 + }, + { + "epoch": 0.5812807304394709, + "grad_norm": 0.12327536195516586, + "learning_rate": 7.477030975994894e-05, + "loss": 0.0203, + "step": 24797 + }, + { + "epoch": 0.5813041720142759, + "grad_norm": 0.19344525039196014, + "learning_rate": 7.476318196672767e-05, + "loss": 0.038, + "step": 24798 + }, + { + "epoch": 0.5813276135890809, + "grad_norm": 0.42441436648368835, + "learning_rate": 7.475605431044208e-05, + "loss": 0.0734, + "step": 24799 + }, + { + "epoch": 0.5813510551638859, + "grad_norm": 0.5189234614372253, + "learning_rate": 7.474892679113086e-05, + "loss": 0.1054, + "step": 24800 + }, + { + "epoch": 0.5813744967386909, + "grad_norm": 0.5537474155426025, + "learning_rate": 7.474179940883265e-05, + "loss": 0.0977, + "step": 24801 + }, + { + "epoch": 0.581397938313496, + "grad_norm": 0.2951497733592987, + "learning_rate": 7.473467216358616e-05, + "loss": 0.0594, + "step": 24802 + }, + { + "epoch": 0.5814213798883009, + "grad_norm": 0.1331014335155487, + "learning_rate": 7.472754505543008e-05, + "loss": 0.0176, + "step": 24803 + }, + { + "epoch": 0.5814448214631059, + "grad_norm": 0.6721105575561523, + "learning_rate": 7.472041808440298e-05, + "loss": 0.1401, + "step": 24804 + }, + { + "epoch": 0.5814682630379109, + "grad_norm": 0.6953251361846924, + "learning_rate": 7.471329125054366e-05, + "loss": 0.6096, + "step": 24805 + }, + { + "epoch": 0.5814917046127159, + "grad_norm": 0.28846436738967896, + "learning_rate": 7.470616455389068e-05, + "loss": 0.0439, + "step": 24806 + }, + { + "epoch": 0.5815151461875209, + "grad_norm": 0.4487740397453308, + "learning_rate": 7.469903799448282e-05, + "loss": 0.5269, + "step": 24807 + }, + { + "epoch": 0.5815385877623259, + "grad_norm": 0.2408013492822647, + "learning_rate": 7.469191157235865e-05, + "loss": 0.0316, + "step": 24808 + }, + { + "epoch": 0.5815620293371309, + "grad_norm": 0.35238364338874817, + "learning_rate": 7.46847852875569e-05, + "loss": 0.0398, + "step": 24809 + }, + { + "epoch": 0.5815854709119359, + "grad_norm": 0.7929766774177551, + "learning_rate": 7.467765914011622e-05, + "loss": 0.1226, + "step": 24810 + }, + { + "epoch": 0.5816089124867408, + "grad_norm": 0.19885265827178955, + "learning_rate": 7.467053313007526e-05, + "loss": 0.0292, + "step": 24811 + }, + { + "epoch": 0.5816323540615459, + "grad_norm": 0.48205745220184326, + "learning_rate": 7.466340725747271e-05, + "loss": 0.5184, + "step": 24812 + }, + { + "epoch": 0.5816557956363508, + "grad_norm": 0.29225608706474304, + "learning_rate": 7.465628152234722e-05, + "loss": 0.0594, + "step": 24813 + }, + { + "epoch": 0.5816792372111559, + "grad_norm": 0.4689408242702484, + "learning_rate": 7.464915592473746e-05, + "loss": 0.6962, + "step": 24814 + }, + { + "epoch": 0.5817026787859608, + "grad_norm": 0.5109186172485352, + "learning_rate": 7.46420304646821e-05, + "loss": 0.0565, + "step": 24815 + }, + { + "epoch": 0.5817261203607659, + "grad_norm": 0.1790589690208435, + "learning_rate": 7.46349051422198e-05, + "loss": 0.0313, + "step": 24816 + }, + { + "epoch": 0.5817495619355708, + "grad_norm": 0.4114072024822235, + "learning_rate": 7.462777995738919e-05, + "loss": 0.0629, + "step": 24817 + }, + { + "epoch": 0.5817730035103759, + "grad_norm": 0.21865485608577728, + "learning_rate": 7.462065491022894e-05, + "loss": 0.0178, + "step": 24818 + }, + { + "epoch": 0.5817964450851808, + "grad_norm": 0.3822425901889801, + "learning_rate": 7.461353000077777e-05, + "loss": 0.063, + "step": 24819 + }, + { + "epoch": 0.5818198866599859, + "grad_norm": 0.31085634231567383, + "learning_rate": 7.460640522907432e-05, + "loss": 0.0859, + "step": 24820 + }, + { + "epoch": 0.5818433282347908, + "grad_norm": 0.29266905784606934, + "learning_rate": 7.45992805951572e-05, + "loss": 0.0325, + "step": 24821 + }, + { + "epoch": 0.5818667698095958, + "grad_norm": 0.35807356238365173, + "learning_rate": 7.45921560990651e-05, + "loss": 0.0727, + "step": 24822 + }, + { + "epoch": 0.5818902113844008, + "grad_norm": 0.3721694350242615, + "learning_rate": 7.458503174083666e-05, + "loss": 0.0631, + "step": 24823 + }, + { + "epoch": 0.5819136529592058, + "grad_norm": 0.18357735872268677, + "learning_rate": 7.457790752051059e-05, + "loss": 0.0145, + "step": 24824 + }, + { + "epoch": 0.5819370945340108, + "grad_norm": 0.2618117034435272, + "learning_rate": 7.457078343812549e-05, + "loss": 0.0649, + "step": 24825 + }, + { + "epoch": 0.5819605361088158, + "grad_norm": 0.2914334833621979, + "learning_rate": 7.456365949372003e-05, + "loss": 0.0624, + "step": 24826 + }, + { + "epoch": 0.5819839776836208, + "grad_norm": 0.49638548493385315, + "learning_rate": 7.455653568733288e-05, + "loss": 0.1182, + "step": 24827 + }, + { + "epoch": 0.5820074192584258, + "grad_norm": 0.7410662174224854, + "learning_rate": 7.454941201900268e-05, + "loss": 0.0406, + "step": 24828 + }, + { + "epoch": 0.5820308608332307, + "grad_norm": 0.4963183104991913, + "learning_rate": 7.454228848876808e-05, + "loss": 0.0478, + "step": 24829 + }, + { + "epoch": 0.5820543024080358, + "grad_norm": 0.5111039280891418, + "learning_rate": 7.453516509666769e-05, + "loss": 0.0604, + "step": 24830 + }, + { + "epoch": 0.5820777439828407, + "grad_norm": 0.08933233469724655, + "learning_rate": 7.452804184274027e-05, + "loss": 0.0189, + "step": 24831 + }, + { + "epoch": 0.5821011855576458, + "grad_norm": 0.18670925498008728, + "learning_rate": 7.452091872702441e-05, + "loss": 0.034, + "step": 24832 + }, + { + "epoch": 0.5821246271324507, + "grad_norm": 0.41883978247642517, + "learning_rate": 7.451379574955874e-05, + "loss": 0.0463, + "step": 24833 + }, + { + "epoch": 0.5821480687072558, + "grad_norm": 0.5048714876174927, + "learning_rate": 7.450667291038196e-05, + "loss": 0.0556, + "step": 24834 + }, + { + "epoch": 0.5821715102820607, + "grad_norm": 0.5322712063789368, + "learning_rate": 7.449955020953267e-05, + "loss": 0.0971, + "step": 24835 + }, + { + "epoch": 0.5821949518568658, + "grad_norm": 1.096714973449707, + "learning_rate": 7.449242764704953e-05, + "loss": 0.0928, + "step": 24836 + }, + { + "epoch": 0.5822183934316707, + "grad_norm": 0.1914166808128357, + "learning_rate": 7.448530522297121e-05, + "loss": 0.0368, + "step": 24837 + }, + { + "epoch": 0.5822418350064757, + "grad_norm": 0.4891294240951538, + "learning_rate": 7.447818293733635e-05, + "loss": 0.0478, + "step": 24838 + }, + { + "epoch": 0.5822652765812807, + "grad_norm": 0.33859771490097046, + "learning_rate": 7.447106079018357e-05, + "loss": 0.04, + "step": 24839 + }, + { + "epoch": 0.5822887181560857, + "grad_norm": 0.5576847791671753, + "learning_rate": 7.446393878155153e-05, + "loss": 0.0746, + "step": 24840 + }, + { + "epoch": 0.5823121597308907, + "grad_norm": 0.1333848237991333, + "learning_rate": 7.44568169114789e-05, + "loss": 0.0234, + "step": 24841 + }, + { + "epoch": 0.5823356013056957, + "grad_norm": 0.3408651053905487, + "learning_rate": 7.444969518000424e-05, + "loss": 0.0638, + "step": 24842 + }, + { + "epoch": 0.5823590428805007, + "grad_norm": 0.672029435634613, + "learning_rate": 7.444257358716629e-05, + "loss": 0.1105, + "step": 24843 + }, + { + "epoch": 0.5823824844553057, + "grad_norm": 0.6880014538764954, + "learning_rate": 7.443545213300367e-05, + "loss": 0.1622, + "step": 24844 + }, + { + "epoch": 0.5824059260301107, + "grad_norm": 0.6258700489997864, + "learning_rate": 7.4428330817555e-05, + "loss": 0.1903, + "step": 24845 + }, + { + "epoch": 0.5824293676049157, + "grad_norm": 0.40100064873695374, + "learning_rate": 7.44212096408589e-05, + "loss": 0.0645, + "step": 24846 + }, + { + "epoch": 0.5824528091797206, + "grad_norm": 0.2913656234741211, + "learning_rate": 7.441408860295407e-05, + "loss": 0.0275, + "step": 24847 + }, + { + "epoch": 0.5824762507545257, + "grad_norm": 0.4795089364051819, + "learning_rate": 7.440696770387911e-05, + "loss": 0.0912, + "step": 24848 + }, + { + "epoch": 0.5824996923293306, + "grad_norm": 0.3417850732803345, + "learning_rate": 7.439984694367264e-05, + "loss": 0.0564, + "step": 24849 + }, + { + "epoch": 0.5825231339041357, + "grad_norm": 0.7318497896194458, + "learning_rate": 7.439272632237334e-05, + "loss": 0.1195, + "step": 24850 + }, + { + "epoch": 0.5825465754789406, + "grad_norm": 0.5513720512390137, + "learning_rate": 7.438560584001983e-05, + "loss": 0.084, + "step": 24851 + }, + { + "epoch": 0.5825700170537457, + "grad_norm": 0.5736691355705261, + "learning_rate": 7.437848549665072e-05, + "loss": 0.1011, + "step": 24852 + }, + { + "epoch": 0.5825934586285507, + "grad_norm": 0.5413625836372375, + "learning_rate": 7.43713652923047e-05, + "loss": 0.0863, + "step": 24853 + }, + { + "epoch": 0.5826169002033557, + "grad_norm": 0.215537428855896, + "learning_rate": 7.436424522702035e-05, + "loss": 0.0241, + "step": 24854 + }, + { + "epoch": 0.5826403417781607, + "grad_norm": 0.38783103227615356, + "learning_rate": 7.43571253008363e-05, + "loss": 0.0409, + "step": 24855 + }, + { + "epoch": 0.5826637833529656, + "grad_norm": 0.5553156733512878, + "learning_rate": 7.435000551379124e-05, + "loss": 0.1388, + "step": 24856 + }, + { + "epoch": 0.5826872249277707, + "grad_norm": 0.3266468644142151, + "learning_rate": 7.434288586592378e-05, + "loss": 0.0365, + "step": 24857 + }, + { + "epoch": 0.5827106665025756, + "grad_norm": 0.3719116747379303, + "learning_rate": 7.433576635727254e-05, + "loss": 0.0468, + "step": 24858 + }, + { + "epoch": 0.5827341080773807, + "grad_norm": 0.2883957326412201, + "learning_rate": 7.432864698787611e-05, + "loss": 0.0555, + "step": 24859 + }, + { + "epoch": 0.5827575496521856, + "grad_norm": 0.11277155578136444, + "learning_rate": 7.432152775777323e-05, + "loss": 0.0186, + "step": 24860 + }, + { + "epoch": 0.5827809912269907, + "grad_norm": 0.4702254831790924, + "learning_rate": 7.431440866700241e-05, + "loss": 0.5384, + "step": 24861 + }, + { + "epoch": 0.5828044328017956, + "grad_norm": 0.6209413409233093, + "learning_rate": 7.430728971560237e-05, + "loss": 0.1756, + "step": 24862 + }, + { + "epoch": 0.5828278743766007, + "grad_norm": 0.6464095711708069, + "learning_rate": 7.430017090361167e-05, + "loss": 0.1923, + "step": 24863 + }, + { + "epoch": 0.5828513159514056, + "grad_norm": 0.6983405351638794, + "learning_rate": 7.429305223106897e-05, + "loss": 0.1304, + "step": 24864 + }, + { + "epoch": 0.5828747575262107, + "grad_norm": 0.17494986951351166, + "learning_rate": 7.42859336980129e-05, + "loss": 0.0293, + "step": 24865 + }, + { + "epoch": 0.5828981991010156, + "grad_norm": 0.8152849078178406, + "learning_rate": 7.42788153044821e-05, + "loss": 0.1824, + "step": 24866 + }, + { + "epoch": 0.5829216406758206, + "grad_norm": 0.41600343585014343, + "learning_rate": 7.427169705051511e-05, + "loss": 0.5649, + "step": 24867 + }, + { + "epoch": 0.5829450822506256, + "grad_norm": 0.6327637434005737, + "learning_rate": 7.426457893615064e-05, + "loss": 0.5959, + "step": 24868 + }, + { + "epoch": 0.5829685238254306, + "grad_norm": 0.4097307324409485, + "learning_rate": 7.42574609614273e-05, + "loss": 0.389, + "step": 24869 + }, + { + "epoch": 0.5829919654002356, + "grad_norm": 0.6073242425918579, + "learning_rate": 7.425034312638371e-05, + "loss": 0.7291, + "step": 24870 + }, + { + "epoch": 0.5830154069750406, + "grad_norm": 0.5915375351905823, + "learning_rate": 7.424322543105846e-05, + "loss": 0.0876, + "step": 24871 + }, + { + "epoch": 0.5830388485498456, + "grad_norm": 0.4464901387691498, + "learning_rate": 7.423610787549022e-05, + "loss": 0.0756, + "step": 24872 + }, + { + "epoch": 0.5830622901246506, + "grad_norm": 0.4497491121292114, + "learning_rate": 7.422899045971759e-05, + "loss": 0.0593, + "step": 24873 + }, + { + "epoch": 0.5830857316994555, + "grad_norm": 0.259011834859848, + "learning_rate": 7.422187318377913e-05, + "loss": 0.033, + "step": 24874 + }, + { + "epoch": 0.5831091732742606, + "grad_norm": 0.7217587232589722, + "learning_rate": 7.421475604771355e-05, + "loss": 0.1395, + "step": 24875 + }, + { + "epoch": 0.5831326148490655, + "grad_norm": 0.6837345361709595, + "learning_rate": 7.420763905155944e-05, + "loss": 0.1055, + "step": 24876 + }, + { + "epoch": 0.5831560564238706, + "grad_norm": 0.7150965929031372, + "learning_rate": 7.420052219535537e-05, + "loss": 0.0706, + "step": 24877 + }, + { + "epoch": 0.5831794979986755, + "grad_norm": 0.2551358640193939, + "learning_rate": 7.419340547914002e-05, + "loss": 0.0316, + "step": 24878 + }, + { + "epoch": 0.5832029395734806, + "grad_norm": 0.20265895128250122, + "learning_rate": 7.418628890295193e-05, + "loss": 0.0376, + "step": 24879 + }, + { + "epoch": 0.5832263811482855, + "grad_norm": 0.4352792203426361, + "learning_rate": 7.417917246682981e-05, + "loss": 0.0619, + "step": 24880 + }, + { + "epoch": 0.5832498227230906, + "grad_norm": 0.10797303915023804, + "learning_rate": 7.41720561708122e-05, + "loss": 0.0151, + "step": 24881 + }, + { + "epoch": 0.5832732642978955, + "grad_norm": 0.7724851369857788, + "learning_rate": 7.416494001493777e-05, + "loss": 0.1114, + "step": 24882 + }, + { + "epoch": 0.5832967058727005, + "grad_norm": 0.6779680252075195, + "learning_rate": 7.415782399924509e-05, + "loss": 0.1504, + "step": 24883 + }, + { + "epoch": 0.5833201474475055, + "grad_norm": 0.4550195634365082, + "learning_rate": 7.415070812377277e-05, + "loss": 0.0903, + "step": 24884 + }, + { + "epoch": 0.5833435890223105, + "grad_norm": 0.07013681530952454, + "learning_rate": 7.414359238855944e-05, + "loss": 0.0084, + "step": 24885 + }, + { + "epoch": 0.5833670305971155, + "grad_norm": 0.2662709057331085, + "learning_rate": 7.413647679364373e-05, + "loss": 0.0256, + "step": 24886 + }, + { + "epoch": 0.5833904721719205, + "grad_norm": 0.280824214220047, + "learning_rate": 7.412936133906419e-05, + "loss": 0.0601, + "step": 24887 + }, + { + "epoch": 0.5834139137467255, + "grad_norm": 0.27112051844596863, + "learning_rate": 7.412224602485948e-05, + "loss": 0.0388, + "step": 24888 + }, + { + "epoch": 0.5834373553215305, + "grad_norm": 0.5732932686805725, + "learning_rate": 7.41151308510682e-05, + "loss": 0.1262, + "step": 24889 + }, + { + "epoch": 0.5834607968963355, + "grad_norm": 0.34671351313591003, + "learning_rate": 7.41080158177289e-05, + "loss": 0.0575, + "step": 24890 + }, + { + "epoch": 0.5834842384711405, + "grad_norm": 0.16995297372341156, + "learning_rate": 7.410090092488024e-05, + "loss": 0.0258, + "step": 24891 + }, + { + "epoch": 0.5835076800459454, + "grad_norm": 0.17353089153766632, + "learning_rate": 7.409378617256086e-05, + "loss": 0.026, + "step": 24892 + }, + { + "epoch": 0.5835311216207505, + "grad_norm": 0.5681018233299255, + "learning_rate": 7.408667156080933e-05, + "loss": 0.0994, + "step": 24893 + }, + { + "epoch": 0.5835545631955554, + "grad_norm": 0.5516337156295776, + "learning_rate": 7.407955708966421e-05, + "loss": 0.0564, + "step": 24894 + }, + { + "epoch": 0.5835780047703605, + "grad_norm": 0.1047019213438034, + "learning_rate": 7.407244275916418e-05, + "loss": 0.0178, + "step": 24895 + }, + { + "epoch": 0.5836014463451654, + "grad_norm": 0.29113513231277466, + "learning_rate": 7.406532856934779e-05, + "loss": 0.034, + "step": 24896 + }, + { + "epoch": 0.5836248879199705, + "grad_norm": 0.25810572504997253, + "learning_rate": 7.405821452025364e-05, + "loss": 0.0435, + "step": 24897 + }, + { + "epoch": 0.5836483294947754, + "grad_norm": 0.23934531211853027, + "learning_rate": 7.405110061192038e-05, + "loss": 0.0414, + "step": 24898 + }, + { + "epoch": 0.5836717710695805, + "grad_norm": 0.5689123868942261, + "learning_rate": 7.404398684438653e-05, + "loss": 0.0841, + "step": 24899 + }, + { + "epoch": 0.5836952126443854, + "grad_norm": 0.09818612784147263, + "learning_rate": 7.403687321769079e-05, + "loss": 0.0063, + "step": 24900 + }, + { + "epoch": 0.5837186542191904, + "grad_norm": 0.35247159004211426, + "learning_rate": 7.40297597318717e-05, + "loss": 0.0841, + "step": 24901 + }, + { + "epoch": 0.5837420957939954, + "grad_norm": 0.9131850600242615, + "learning_rate": 7.402264638696783e-05, + "loss": 0.0928, + "step": 24902 + }, + { + "epoch": 0.5837655373688004, + "grad_norm": 0.3796807527542114, + "learning_rate": 7.40155331830178e-05, + "loss": 0.0765, + "step": 24903 + }, + { + "epoch": 0.5837889789436055, + "grad_norm": 0.38720986247062683, + "learning_rate": 7.400842012006024e-05, + "loss": 0.0419, + "step": 24904 + }, + { + "epoch": 0.5838124205184104, + "grad_norm": 0.3551895022392273, + "learning_rate": 7.400130719813374e-05, + "loss": 0.0504, + "step": 24905 + }, + { + "epoch": 0.5838358620932155, + "grad_norm": 0.6282287240028381, + "learning_rate": 7.399419441727686e-05, + "loss": 0.0738, + "step": 24906 + }, + { + "epoch": 0.5838593036680204, + "grad_norm": 0.26056671142578125, + "learning_rate": 7.398708177752823e-05, + "loss": 0.0506, + "step": 24907 + }, + { + "epoch": 0.5838827452428255, + "grad_norm": 0.4646547734737396, + "learning_rate": 7.397996927892641e-05, + "loss": 0.0553, + "step": 24908 + }, + { + "epoch": 0.5839061868176304, + "grad_norm": 0.6087744832038879, + "learning_rate": 7.397285692151e-05, + "loss": 0.1431, + "step": 24909 + }, + { + "epoch": 0.5839296283924355, + "grad_norm": 0.3156956732273102, + "learning_rate": 7.396574470531764e-05, + "loss": 0.4351, + "step": 24910 + }, + { + "epoch": 0.5839530699672404, + "grad_norm": 0.7863449454307556, + "learning_rate": 7.395863263038785e-05, + "loss": 0.1379, + "step": 24911 + }, + { + "epoch": 0.5839765115420454, + "grad_norm": 0.33866214752197266, + "learning_rate": 7.395152069675924e-05, + "loss": 0.0716, + "step": 24912 + }, + { + "epoch": 0.5839999531168504, + "grad_norm": 0.7139955759048462, + "learning_rate": 7.394440890447043e-05, + "loss": 0.109, + "step": 24913 + }, + { + "epoch": 0.5840233946916554, + "grad_norm": 0.20738281309604645, + "learning_rate": 7.393729725356e-05, + "loss": 0.0416, + "step": 24914 + }, + { + "epoch": 0.5840468362664604, + "grad_norm": 0.3625697195529938, + "learning_rate": 7.393018574406648e-05, + "loss": 0.0929, + "step": 24915 + }, + { + "epoch": 0.5840702778412654, + "grad_norm": 0.13206282258033752, + "learning_rate": 7.392307437602852e-05, + "loss": 0.0166, + "step": 24916 + }, + { + "epoch": 0.5840937194160704, + "grad_norm": 0.5500635504722595, + "learning_rate": 7.391596314948473e-05, + "loss": 0.1044, + "step": 24917 + }, + { + "epoch": 0.5841171609908754, + "grad_norm": 0.09210038185119629, + "learning_rate": 7.390885206447365e-05, + "loss": 0.019, + "step": 24918 + }, + { + "epoch": 0.5841406025656803, + "grad_norm": 0.2606131434440613, + "learning_rate": 7.390174112103386e-05, + "loss": 0.0206, + "step": 24919 + }, + { + "epoch": 0.5841640441404854, + "grad_norm": 0.5263400673866272, + "learning_rate": 7.389463031920396e-05, + "loss": 0.1285, + "step": 24920 + }, + { + "epoch": 0.5841874857152903, + "grad_norm": 0.16536904871463776, + "learning_rate": 7.388751965902252e-05, + "loss": 0.0484, + "step": 24921 + }, + { + "epoch": 0.5842109272900954, + "grad_norm": 0.3860887289047241, + "learning_rate": 7.388040914052814e-05, + "loss": 0.0301, + "step": 24922 + }, + { + "epoch": 0.5842343688649003, + "grad_norm": 0.2624201476573944, + "learning_rate": 7.387329876375941e-05, + "loss": 0.0392, + "step": 24923 + }, + { + "epoch": 0.5842578104397054, + "grad_norm": 0.5371412038803101, + "learning_rate": 7.386618852875489e-05, + "loss": 0.0778, + "step": 24924 + }, + { + "epoch": 0.5842812520145103, + "grad_norm": 0.6810188889503479, + "learning_rate": 7.385907843555314e-05, + "loss": 0.0374, + "step": 24925 + }, + { + "epoch": 0.5843046935893154, + "grad_norm": 0.2488088458776474, + "learning_rate": 7.385196848419278e-05, + "loss": 0.0479, + "step": 24926 + }, + { + "epoch": 0.5843281351641203, + "grad_norm": 0.3954738974571228, + "learning_rate": 7.38448586747124e-05, + "loss": 0.0525, + "step": 24927 + }, + { + "epoch": 0.5843515767389253, + "grad_norm": 0.6132223010063171, + "learning_rate": 7.383774900715048e-05, + "loss": 0.6214, + "step": 24928 + }, + { + "epoch": 0.5843750183137303, + "grad_norm": 0.3619402348995209, + "learning_rate": 7.383063948154572e-05, + "loss": 0.0538, + "step": 24929 + }, + { + "epoch": 0.5843984598885353, + "grad_norm": 0.4252009391784668, + "learning_rate": 7.382353009793664e-05, + "loss": 0.084, + "step": 24930 + }, + { + "epoch": 0.5844219014633403, + "grad_norm": 0.146770179271698, + "learning_rate": 7.381642085636183e-05, + "loss": 0.022, + "step": 24931 + }, + { + "epoch": 0.5844453430381453, + "grad_norm": 0.18692848086357117, + "learning_rate": 7.380931175685984e-05, + "loss": 0.0262, + "step": 24932 + }, + { + "epoch": 0.5844687846129503, + "grad_norm": 0.46662357449531555, + "learning_rate": 7.380220279946928e-05, + "loss": 0.0485, + "step": 24933 + }, + { + "epoch": 0.5844922261877553, + "grad_norm": 0.5180263519287109, + "learning_rate": 7.379509398422872e-05, + "loss": 0.0522, + "step": 24934 + }, + { + "epoch": 0.5845156677625603, + "grad_norm": 0.4689978063106537, + "learning_rate": 7.378798531117667e-05, + "loss": 0.0867, + "step": 24935 + }, + { + "epoch": 0.5845391093373653, + "grad_norm": 0.3383706212043762, + "learning_rate": 7.378087678035179e-05, + "loss": 0.0766, + "step": 24936 + }, + { + "epoch": 0.5845625509121702, + "grad_norm": 1.3597497940063477, + "learning_rate": 7.377376839179258e-05, + "loss": 0.1622, + "step": 24937 + }, + { + "epoch": 0.5845859924869753, + "grad_norm": 0.3319617807865143, + "learning_rate": 7.376666014553765e-05, + "loss": 0.0646, + "step": 24938 + }, + { + "epoch": 0.5846094340617802, + "grad_norm": 0.5217804312705994, + "learning_rate": 7.375955204162558e-05, + "loss": 0.1115, + "step": 24939 + }, + { + "epoch": 0.5846328756365853, + "grad_norm": 0.700107216835022, + "learning_rate": 7.375244408009486e-05, + "loss": 0.181, + "step": 24940 + }, + { + "epoch": 0.5846563172113902, + "grad_norm": 0.23140498995780945, + "learning_rate": 7.374533626098414e-05, + "loss": 0.0322, + "step": 24941 + }, + { + "epoch": 0.5846797587861953, + "grad_norm": 0.3215612769126892, + "learning_rate": 7.3738228584332e-05, + "loss": 0.0639, + "step": 24942 + }, + { + "epoch": 0.5847032003610002, + "grad_norm": 1.0105268955230713, + "learning_rate": 7.373112105017695e-05, + "loss": 0.15, + "step": 24943 + }, + { + "epoch": 0.5847266419358053, + "grad_norm": 0.15481136739253998, + "learning_rate": 7.372401365855756e-05, + "loss": 0.0207, + "step": 24944 + }, + { + "epoch": 0.5847500835106102, + "grad_norm": 0.3300027847290039, + "learning_rate": 7.371690640951245e-05, + "loss": 0.0298, + "step": 24945 + }, + { + "epoch": 0.5847735250854152, + "grad_norm": 0.4253430962562561, + "learning_rate": 7.370979930308013e-05, + "loss": 0.077, + "step": 24946 + }, + { + "epoch": 0.5847969666602202, + "grad_norm": 0.5913168787956238, + "learning_rate": 7.370269233929917e-05, + "loss": 0.5569, + "step": 24947 + }, + { + "epoch": 0.5848204082350252, + "grad_norm": 0.423122763633728, + "learning_rate": 7.369558551820814e-05, + "loss": 0.0581, + "step": 24948 + }, + { + "epoch": 0.5848438498098302, + "grad_norm": 0.21301551163196564, + "learning_rate": 7.368847883984561e-05, + "loss": 0.03, + "step": 24949 + }, + { + "epoch": 0.5848672913846352, + "grad_norm": 0.4510739743709564, + "learning_rate": 7.368137230425012e-05, + "loss": 0.0593, + "step": 24950 + }, + { + "epoch": 0.5848907329594402, + "grad_norm": 0.28956955671310425, + "learning_rate": 7.367426591146025e-05, + "loss": 0.0695, + "step": 24951 + }, + { + "epoch": 0.5849141745342452, + "grad_norm": 0.5716171264648438, + "learning_rate": 7.366715966151452e-05, + "loss": 0.6941, + "step": 24952 + }, + { + "epoch": 0.5849376161090502, + "grad_norm": 0.16321644186973572, + "learning_rate": 7.366005355445157e-05, + "loss": 0.0187, + "step": 24953 + }, + { + "epoch": 0.5849610576838552, + "grad_norm": 0.38179343938827515, + "learning_rate": 7.365294759030988e-05, + "loss": 0.0499, + "step": 24954 + }, + { + "epoch": 0.5849844992586603, + "grad_norm": 0.7009036540985107, + "learning_rate": 7.364584176912807e-05, + "loss": 0.5848, + "step": 24955 + }, + { + "epoch": 0.5850079408334652, + "grad_norm": 0.4253583550453186, + "learning_rate": 7.363873609094464e-05, + "loss": 0.0479, + "step": 24956 + }, + { + "epoch": 0.5850313824082702, + "grad_norm": 0.6539217829704285, + "learning_rate": 7.363163055579818e-05, + "loss": 0.1391, + "step": 24957 + }, + { + "epoch": 0.5850548239830752, + "grad_norm": 0.4008200168609619, + "learning_rate": 7.362452516372724e-05, + "loss": 0.0731, + "step": 24958 + }, + { + "epoch": 0.5850782655578802, + "grad_norm": 0.6107693314552307, + "learning_rate": 7.361741991477036e-05, + "loss": 0.4719, + "step": 24959 + }, + { + "epoch": 0.5851017071326852, + "grad_norm": 0.41791224479675293, + "learning_rate": 7.36103148089661e-05, + "loss": 0.0742, + "step": 24960 + }, + { + "epoch": 0.5851251487074902, + "grad_norm": 0.18270327150821686, + "learning_rate": 7.3603209846353e-05, + "loss": 0.0192, + "step": 24961 + }, + { + "epoch": 0.5851485902822952, + "grad_norm": 0.20785710215568542, + "learning_rate": 7.359610502696967e-05, + "loss": 0.0378, + "step": 24962 + }, + { + "epoch": 0.5851720318571002, + "grad_norm": 0.621990442276001, + "learning_rate": 7.358900035085456e-05, + "loss": 0.1994, + "step": 24963 + }, + { + "epoch": 0.5851954734319051, + "grad_norm": 0.46651124954223633, + "learning_rate": 7.358189581804627e-05, + "loss": 0.0949, + "step": 24964 + }, + { + "epoch": 0.5852189150067102, + "grad_norm": 0.5439582467079163, + "learning_rate": 7.35747914285834e-05, + "loss": 0.2804, + "step": 24965 + }, + { + "epoch": 0.5852423565815151, + "grad_norm": 0.20231181383132935, + "learning_rate": 7.356768718250446e-05, + "loss": 0.0165, + "step": 24966 + }, + { + "epoch": 0.5852657981563202, + "grad_norm": 0.5477743148803711, + "learning_rate": 7.356058307984796e-05, + "loss": 0.0924, + "step": 24967 + }, + { + "epoch": 0.5852892397311251, + "grad_norm": 0.5444834232330322, + "learning_rate": 7.35534791206525e-05, + "loss": 0.1333, + "step": 24968 + }, + { + "epoch": 0.5853126813059302, + "grad_norm": 0.45567071437835693, + "learning_rate": 7.35463753049566e-05, + "loss": 0.0861, + "step": 24969 + }, + { + "epoch": 0.5853361228807351, + "grad_norm": 0.2783116102218628, + "learning_rate": 7.353927163279882e-05, + "loss": 0.1233, + "step": 24970 + }, + { + "epoch": 0.5853595644555402, + "grad_norm": 0.4363009035587311, + "learning_rate": 7.35321681042177e-05, + "loss": 0.0598, + "step": 24971 + }, + { + "epoch": 0.5853830060303451, + "grad_norm": 0.3445284366607666, + "learning_rate": 7.352506471925177e-05, + "loss": 0.0353, + "step": 24972 + }, + { + "epoch": 0.5854064476051501, + "grad_norm": 0.26082509756088257, + "learning_rate": 7.351796147793956e-05, + "loss": 0.0314, + "step": 24973 + }, + { + "epoch": 0.5854298891799551, + "grad_norm": 0.27016374468803406, + "learning_rate": 7.351085838031968e-05, + "loss": 0.0392, + "step": 24974 + }, + { + "epoch": 0.5854533307547601, + "grad_norm": 0.32038357853889465, + "learning_rate": 7.350375542643059e-05, + "loss": 0.0691, + "step": 24975 + }, + { + "epoch": 0.5854767723295651, + "grad_norm": 0.3587420880794525, + "learning_rate": 7.349665261631086e-05, + "loss": 0.0464, + "step": 24976 + }, + { + "epoch": 0.5855002139043701, + "grad_norm": 0.3476995527744293, + "learning_rate": 7.348954994999906e-05, + "loss": 0.0673, + "step": 24977 + }, + { + "epoch": 0.5855236554791751, + "grad_norm": 1.0029776096343994, + "learning_rate": 7.348244742753372e-05, + "loss": 0.058, + "step": 24978 + }, + { + "epoch": 0.5855470970539801, + "grad_norm": 0.20858074724674225, + "learning_rate": 7.347534504895334e-05, + "loss": 0.0212, + "step": 24979 + }, + { + "epoch": 0.585570538628785, + "grad_norm": 0.68896484375, + "learning_rate": 7.346824281429651e-05, + "loss": 0.1084, + "step": 24980 + }, + { + "epoch": 0.5855939802035901, + "grad_norm": 0.3522385358810425, + "learning_rate": 7.346114072360172e-05, + "loss": 0.0652, + "step": 24981 + }, + { + "epoch": 0.585617421778395, + "grad_norm": 0.43245577812194824, + "learning_rate": 7.345403877690753e-05, + "loss": 0.0715, + "step": 24982 + }, + { + "epoch": 0.5856408633532001, + "grad_norm": 0.11636054515838623, + "learning_rate": 7.344693697425249e-05, + "loss": 0.0219, + "step": 24983 + }, + { + "epoch": 0.585664304928005, + "grad_norm": 0.05363144353032112, + "learning_rate": 7.343983531567511e-05, + "loss": 0.0041, + "step": 24984 + }, + { + "epoch": 0.5856877465028101, + "grad_norm": 0.5079225897789001, + "learning_rate": 7.343273380121391e-05, + "loss": 0.1448, + "step": 24985 + }, + { + "epoch": 0.585711188077615, + "grad_norm": 0.3937511444091797, + "learning_rate": 7.342563243090747e-05, + "loss": 0.055, + "step": 24986 + }, + { + "epoch": 0.5857346296524201, + "grad_norm": 0.6859532594680786, + "learning_rate": 7.34185312047943e-05, + "loss": 0.5972, + "step": 24987 + }, + { + "epoch": 0.585758071227225, + "grad_norm": 0.4885041415691376, + "learning_rate": 7.341143012291288e-05, + "loss": 0.083, + "step": 24988 + }, + { + "epoch": 0.5857815128020301, + "grad_norm": 0.5518720149993896, + "learning_rate": 7.34043291853018e-05, + "loss": 0.1062, + "step": 24989 + }, + { + "epoch": 0.585804954376835, + "grad_norm": 0.5492385029792786, + "learning_rate": 7.339722839199961e-05, + "loss": 0.1668, + "step": 24990 + }, + { + "epoch": 0.58582839595164, + "grad_norm": 0.17791172862052917, + "learning_rate": 7.33901277430448e-05, + "loss": 0.0311, + "step": 24991 + }, + { + "epoch": 0.585851837526445, + "grad_norm": 0.31183695793151855, + "learning_rate": 7.338302723847588e-05, + "loss": 0.0449, + "step": 24992 + }, + { + "epoch": 0.58587527910125, + "grad_norm": 0.17100532352924347, + "learning_rate": 7.337592687833143e-05, + "loss": 0.0279, + "step": 24993 + }, + { + "epoch": 0.585898720676055, + "grad_norm": 0.4808107614517212, + "learning_rate": 7.336882666264995e-05, + "loss": 0.0714, + "step": 24994 + }, + { + "epoch": 0.58592216225086, + "grad_norm": 0.35102200508117676, + "learning_rate": 7.336172659146994e-05, + "loss": 0.0598, + "step": 24995 + }, + { + "epoch": 0.585945603825665, + "grad_norm": 0.7938622236251831, + "learning_rate": 7.335462666482998e-05, + "loss": 0.0317, + "step": 24996 + }, + { + "epoch": 0.58596904540047, + "grad_norm": 0.3952620029449463, + "learning_rate": 7.334752688276856e-05, + "loss": 0.5134, + "step": 24997 + }, + { + "epoch": 0.585992486975275, + "grad_norm": 0.6016265153884888, + "learning_rate": 7.334042724532419e-05, + "loss": 0.1261, + "step": 24998 + }, + { + "epoch": 0.58601592855008, + "grad_norm": 0.626784086227417, + "learning_rate": 7.333332775253542e-05, + "loss": 0.1037, + "step": 24999 + }, + { + "epoch": 0.5860393701248849, + "grad_norm": 0.5814661979675293, + "learning_rate": 7.332622840444077e-05, + "loss": 0.0955, + "step": 25000 + }, + { + "epoch": 0.58606281169969, + "grad_norm": 0.4978327751159668, + "learning_rate": 7.331912920107872e-05, + "loss": 0.0986, + "step": 25001 + }, + { + "epoch": 0.5860862532744949, + "grad_norm": 0.07793442159891129, + "learning_rate": 7.331203014248784e-05, + "loss": 0.0155, + "step": 25002 + }, + { + "epoch": 0.5861096948493, + "grad_norm": 0.4962467849254608, + "learning_rate": 7.330493122870666e-05, + "loss": 0.559, + "step": 25003 + }, + { + "epoch": 0.5861331364241049, + "grad_norm": 0.168314591050148, + "learning_rate": 7.329783245977367e-05, + "loss": 0.0362, + "step": 25004 + }, + { + "epoch": 0.58615657799891, + "grad_norm": 0.31078484654426575, + "learning_rate": 7.329073383572737e-05, + "loss": 0.069, + "step": 25005 + }, + { + "epoch": 0.586180019573715, + "grad_norm": 0.7342271208763123, + "learning_rate": 7.328363535660631e-05, + "loss": 0.1547, + "step": 25006 + }, + { + "epoch": 0.58620346114852, + "grad_norm": 0.3813354969024658, + "learning_rate": 7.327653702244902e-05, + "loss": 0.0987, + "step": 25007 + }, + { + "epoch": 0.586226902723325, + "grad_norm": 0.36721011996269226, + "learning_rate": 7.326943883329396e-05, + "loss": 0.0708, + "step": 25008 + }, + { + "epoch": 0.5862503442981299, + "grad_norm": 0.5066884756088257, + "learning_rate": 7.326234078917969e-05, + "loss": 0.0793, + "step": 25009 + }, + { + "epoch": 0.586273785872935, + "grad_norm": 0.40936607122421265, + "learning_rate": 7.32552428901447e-05, + "loss": 0.0953, + "step": 25010 + }, + { + "epoch": 0.5862972274477399, + "grad_norm": 0.41555139422416687, + "learning_rate": 7.324814513622753e-05, + "loss": 0.059, + "step": 25011 + }, + { + "epoch": 0.586320669022545, + "grad_norm": 0.676504909992218, + "learning_rate": 7.324104752746668e-05, + "loss": 0.1105, + "step": 25012 + }, + { + "epoch": 0.5863441105973499, + "grad_norm": 0.3279550075531006, + "learning_rate": 7.323395006390062e-05, + "loss": 0.0153, + "step": 25013 + }, + { + "epoch": 0.586367552172155, + "grad_norm": 0.5219894051551819, + "learning_rate": 7.322685274556793e-05, + "loss": 0.6444, + "step": 25014 + }, + { + "epoch": 0.5863909937469599, + "grad_norm": 0.1261426955461502, + "learning_rate": 7.321975557250706e-05, + "loss": 0.0383, + "step": 25015 + }, + { + "epoch": 0.586414435321765, + "grad_norm": 0.2970173954963684, + "learning_rate": 7.32126585447566e-05, + "loss": 0.0476, + "step": 25016 + }, + { + "epoch": 0.5864378768965699, + "grad_norm": 0.13613595068454742, + "learning_rate": 7.320556166235497e-05, + "loss": 0.0209, + "step": 25017 + }, + { + "epoch": 0.586461318471375, + "grad_norm": 0.35126015543937683, + "learning_rate": 7.319846492534073e-05, + "loss": 0.054, + "step": 25018 + }, + { + "epoch": 0.5864847600461799, + "grad_norm": 0.335330992937088, + "learning_rate": 7.319136833375239e-05, + "loss": 0.0503, + "step": 25019 + }, + { + "epoch": 0.5865082016209849, + "grad_norm": 0.43682220578193665, + "learning_rate": 7.31842718876284e-05, + "loss": 0.0793, + "step": 25020 + }, + { + "epoch": 0.5865316431957899, + "grad_norm": 0.3488593101501465, + "learning_rate": 7.317717558700734e-05, + "loss": 0.0494, + "step": 25021 + }, + { + "epoch": 0.5865550847705949, + "grad_norm": 0.5633865594863892, + "learning_rate": 7.317007943192768e-05, + "loss": 0.1241, + "step": 25022 + }, + { + "epoch": 0.5865785263453999, + "grad_norm": 0.6415566205978394, + "learning_rate": 7.316298342242788e-05, + "loss": 0.1046, + "step": 25023 + }, + { + "epoch": 0.5866019679202049, + "grad_norm": 0.42346543073654175, + "learning_rate": 7.315588755854654e-05, + "loss": 0.0836, + "step": 25024 + }, + { + "epoch": 0.5866254094950099, + "grad_norm": 0.28819793462753296, + "learning_rate": 7.314879184032204e-05, + "loss": 0.3097, + "step": 25025 + }, + { + "epoch": 0.5866488510698149, + "grad_norm": 0.45091381669044495, + "learning_rate": 7.314169626779301e-05, + "loss": 0.0722, + "step": 25026 + }, + { + "epoch": 0.5866722926446198, + "grad_norm": 0.2962220311164856, + "learning_rate": 7.313460084099785e-05, + "loss": 0.0423, + "step": 25027 + }, + { + "epoch": 0.5866957342194249, + "grad_norm": 0.38083547353744507, + "learning_rate": 7.312750555997516e-05, + "loss": 0.074, + "step": 25028 + }, + { + "epoch": 0.5867191757942298, + "grad_norm": 0.1543821394443512, + "learning_rate": 7.312041042476334e-05, + "loss": 0.0225, + "step": 25029 + }, + { + "epoch": 0.5867426173690349, + "grad_norm": 0.3995073139667511, + "learning_rate": 7.311331543540093e-05, + "loss": 0.059, + "step": 25030 + }, + { + "epoch": 0.5867660589438398, + "grad_norm": 0.5919252038002014, + "learning_rate": 7.310622059192644e-05, + "loss": 0.0607, + "step": 25031 + }, + { + "epoch": 0.5867895005186449, + "grad_norm": 0.8246886134147644, + "learning_rate": 7.309912589437835e-05, + "loss": 0.1139, + "step": 25032 + }, + { + "epoch": 0.5868129420934498, + "grad_norm": 0.2703867256641388, + "learning_rate": 7.309203134279515e-05, + "loss": 0.0662, + "step": 25033 + }, + { + "epoch": 0.5868363836682549, + "grad_norm": 0.5733035802841187, + "learning_rate": 7.308493693721536e-05, + "loss": 0.1586, + "step": 25034 + }, + { + "epoch": 0.5868598252430598, + "grad_norm": 1.1064159870147705, + "learning_rate": 7.307784267767745e-05, + "loss": 0.1884, + "step": 25035 + }, + { + "epoch": 0.5868832668178648, + "grad_norm": 0.5080379247665405, + "learning_rate": 7.307074856421992e-05, + "loss": 0.1092, + "step": 25036 + }, + { + "epoch": 0.5869067083926698, + "grad_norm": 0.417738676071167, + "learning_rate": 7.306365459688123e-05, + "loss": 0.0977, + "step": 25037 + }, + { + "epoch": 0.5869301499674748, + "grad_norm": 0.377091646194458, + "learning_rate": 7.305656077569995e-05, + "loss": 0.1032, + "step": 25038 + }, + { + "epoch": 0.5869535915422798, + "grad_norm": 0.4235248565673828, + "learning_rate": 7.304946710071454e-05, + "loss": 0.1032, + "step": 25039 + }, + { + "epoch": 0.5869770331170848, + "grad_norm": 0.20798856019973755, + "learning_rate": 7.304237357196344e-05, + "loss": 0.0432, + "step": 25040 + }, + { + "epoch": 0.5870004746918898, + "grad_norm": 0.104228675365448, + "learning_rate": 7.303528018948523e-05, + "loss": 0.02, + "step": 25041 + }, + { + "epoch": 0.5870239162666948, + "grad_norm": 0.17212466895580292, + "learning_rate": 7.302818695331832e-05, + "loss": 0.0278, + "step": 25042 + }, + { + "epoch": 0.5870473578414998, + "grad_norm": 0.8084529638290405, + "learning_rate": 7.30210938635012e-05, + "loss": 0.0942, + "step": 25043 + }, + { + "epoch": 0.5870707994163048, + "grad_norm": 0.4713974595069885, + "learning_rate": 7.301400092007242e-05, + "loss": 0.3943, + "step": 25044 + }, + { + "epoch": 0.5870942409911097, + "grad_norm": 0.49117565155029297, + "learning_rate": 7.300690812307043e-05, + "loss": 0.1204, + "step": 25045 + }, + { + "epoch": 0.5871176825659148, + "grad_norm": 0.996286928653717, + "learning_rate": 7.29998154725337e-05, + "loss": 0.6378, + "step": 25046 + }, + { + "epoch": 0.5871411241407197, + "grad_norm": 0.28172123432159424, + "learning_rate": 7.299272296850075e-05, + "loss": 0.0527, + "step": 25047 + }, + { + "epoch": 0.5871645657155248, + "grad_norm": 0.4533732235431671, + "learning_rate": 7.298563061101e-05, + "loss": 0.0695, + "step": 25048 + }, + { + "epoch": 0.5871880072903297, + "grad_norm": 0.4991893172264099, + "learning_rate": 7.297853840010001e-05, + "loss": 0.1035, + "step": 25049 + }, + { + "epoch": 0.5872114488651348, + "grad_norm": 0.11384889483451843, + "learning_rate": 7.297144633580922e-05, + "loss": 0.0174, + "step": 25050 + }, + { + "epoch": 0.5872348904399397, + "grad_norm": 0.668378472328186, + "learning_rate": 7.296435441817612e-05, + "loss": 0.0994, + "step": 25051 + }, + { + "epoch": 0.5872583320147448, + "grad_norm": 0.3168965280056, + "learning_rate": 7.295726264723922e-05, + "loss": 0.0675, + "step": 25052 + }, + { + "epoch": 0.5872817735895497, + "grad_norm": 0.3864434063434601, + "learning_rate": 7.295017102303694e-05, + "loss": 0.0733, + "step": 25053 + }, + { + "epoch": 0.5873052151643547, + "grad_norm": 0.296645849943161, + "learning_rate": 7.294307954560782e-05, + "loss": 0.0376, + "step": 25054 + }, + { + "epoch": 0.5873286567391597, + "grad_norm": 0.527580201625824, + "learning_rate": 7.293598821499029e-05, + "loss": 0.147, + "step": 25055 + }, + { + "epoch": 0.5873520983139647, + "grad_norm": 0.5530155897140503, + "learning_rate": 7.292889703122287e-05, + "loss": 0.1075, + "step": 25056 + }, + { + "epoch": 0.5873755398887698, + "grad_norm": 0.49166008830070496, + "learning_rate": 7.292180599434401e-05, + "loss": 0.1193, + "step": 25057 + }, + { + "epoch": 0.5873989814635747, + "grad_norm": 0.43570661544799805, + "learning_rate": 7.291471510439218e-05, + "loss": 0.0685, + "step": 25058 + }, + { + "epoch": 0.5874224230383798, + "grad_norm": 0.3220524787902832, + "learning_rate": 7.290762436140588e-05, + "loss": 0.0337, + "step": 25059 + }, + { + "epoch": 0.5874458646131847, + "grad_norm": 0.1666877418756485, + "learning_rate": 7.290053376542357e-05, + "loss": 0.0143, + "step": 25060 + }, + { + "epoch": 0.5874693061879898, + "grad_norm": 0.10179275274276733, + "learning_rate": 7.289344331648368e-05, + "loss": 0.0189, + "step": 25061 + }, + { + "epoch": 0.5874927477627947, + "grad_norm": 0.12828762829303741, + "learning_rate": 7.288635301462477e-05, + "loss": 0.0322, + "step": 25062 + }, + { + "epoch": 0.5875161893375997, + "grad_norm": 0.6595523953437805, + "learning_rate": 7.287926285988528e-05, + "loss": 0.2167, + "step": 25063 + }, + { + "epoch": 0.5875396309124047, + "grad_norm": 0.4732966423034668, + "learning_rate": 7.287217285230367e-05, + "loss": 0.0513, + "step": 25064 + }, + { + "epoch": 0.5875630724872097, + "grad_norm": 0.3130308985710144, + "learning_rate": 7.286508299191839e-05, + "loss": 0.0742, + "step": 25065 + }, + { + "epoch": 0.5875865140620147, + "grad_norm": 0.08834938704967499, + "learning_rate": 7.285799327876796e-05, + "loss": 0.0077, + "step": 25066 + }, + { + "epoch": 0.5876099556368197, + "grad_norm": 0.12081120163202286, + "learning_rate": 7.285090371289083e-05, + "loss": 0.0231, + "step": 25067 + }, + { + "epoch": 0.5876333972116247, + "grad_norm": 0.5486874580383301, + "learning_rate": 7.284381429432542e-05, + "loss": 0.0681, + "step": 25068 + }, + { + "epoch": 0.5876568387864297, + "grad_norm": 0.13885152339935303, + "learning_rate": 7.283672502311027e-05, + "loss": 0.0376, + "step": 25069 + }, + { + "epoch": 0.5876802803612347, + "grad_norm": 0.6178752183914185, + "learning_rate": 7.28296358992838e-05, + "loss": 0.0888, + "step": 25070 + }, + { + "epoch": 0.5877037219360397, + "grad_norm": 0.3298492133617401, + "learning_rate": 7.282254692288448e-05, + "loss": 0.062, + "step": 25071 + }, + { + "epoch": 0.5877271635108446, + "grad_norm": 0.5389052629470825, + "learning_rate": 7.281545809395081e-05, + "loss": 0.0982, + "step": 25072 + }, + { + "epoch": 0.5877506050856497, + "grad_norm": 0.24989061057567596, + "learning_rate": 7.280836941252119e-05, + "loss": 0.0596, + "step": 25073 + }, + { + "epoch": 0.5877740466604546, + "grad_norm": 0.3996117413043976, + "learning_rate": 7.280128087863414e-05, + "loss": 0.0945, + "step": 25074 + }, + { + "epoch": 0.5877974882352597, + "grad_norm": 0.350734680891037, + "learning_rate": 7.27941924923281e-05, + "loss": 0.082, + "step": 25075 + }, + { + "epoch": 0.5878209298100646, + "grad_norm": 0.3827790319919586, + "learning_rate": 7.278710425364157e-05, + "loss": 0.0851, + "step": 25076 + }, + { + "epoch": 0.5878443713848697, + "grad_norm": 0.5207915902137756, + "learning_rate": 7.278001616261296e-05, + "loss": 0.2926, + "step": 25077 + }, + { + "epoch": 0.5878678129596746, + "grad_norm": 0.7171714901924133, + "learning_rate": 7.277292821928072e-05, + "loss": 0.1404, + "step": 25078 + }, + { + "epoch": 0.5878912545344797, + "grad_norm": 0.3165576457977295, + "learning_rate": 7.276584042368338e-05, + "loss": 0.0595, + "step": 25079 + }, + { + "epoch": 0.5879146961092846, + "grad_norm": 0.6085639595985413, + "learning_rate": 7.275875277585934e-05, + "loss": 0.1662, + "step": 25080 + }, + { + "epoch": 0.5879381376840896, + "grad_norm": 0.13614802062511444, + "learning_rate": 7.275166527584706e-05, + "loss": 0.0306, + "step": 25081 + }, + { + "epoch": 0.5879615792588946, + "grad_norm": 0.1650858223438263, + "learning_rate": 7.274457792368503e-05, + "loss": 0.0339, + "step": 25082 + }, + { + "epoch": 0.5879850208336996, + "grad_norm": 0.2059798687696457, + "learning_rate": 7.273749071941167e-05, + "loss": 0.0265, + "step": 25083 + }, + { + "epoch": 0.5880084624085046, + "grad_norm": 0.20308442413806915, + "learning_rate": 7.273040366306545e-05, + "loss": 0.0466, + "step": 25084 + }, + { + "epoch": 0.5880319039833096, + "grad_norm": 0.11970391869544983, + "learning_rate": 7.272331675468485e-05, + "loss": 0.0196, + "step": 25085 + }, + { + "epoch": 0.5880553455581146, + "grad_norm": 0.46491894125938416, + "learning_rate": 7.271622999430825e-05, + "loss": 0.0814, + "step": 25086 + }, + { + "epoch": 0.5880787871329196, + "grad_norm": 0.31781965494155884, + "learning_rate": 7.270914338197419e-05, + "loss": 0.0507, + "step": 25087 + }, + { + "epoch": 0.5881022287077246, + "grad_norm": 0.577869176864624, + "learning_rate": 7.270205691772107e-05, + "loss": 0.158, + "step": 25088 + }, + { + "epoch": 0.5881256702825296, + "grad_norm": 0.9627847075462341, + "learning_rate": 7.269497060158739e-05, + "loss": 0.1663, + "step": 25089 + }, + { + "epoch": 0.5881491118573345, + "grad_norm": 0.4124187231063843, + "learning_rate": 7.268788443361156e-05, + "loss": 0.0886, + "step": 25090 + }, + { + "epoch": 0.5881725534321396, + "grad_norm": 0.49707141518592834, + "learning_rate": 7.2680798413832e-05, + "loss": 0.077, + "step": 25091 + }, + { + "epoch": 0.5881959950069445, + "grad_norm": 0.6172505617141724, + "learning_rate": 7.267371254228723e-05, + "loss": 0.1946, + "step": 25092 + }, + { + "epoch": 0.5882194365817496, + "grad_norm": 0.24148842692375183, + "learning_rate": 7.266662681901564e-05, + "loss": 0.0614, + "step": 25093 + }, + { + "epoch": 0.5882428781565545, + "grad_norm": 0.47334814071655273, + "learning_rate": 7.265954124405573e-05, + "loss": 0.1107, + "step": 25094 + }, + { + "epoch": 0.5882663197313596, + "grad_norm": 2.248992919921875, + "learning_rate": 7.26524558174459e-05, + "loss": 0.0778, + "step": 25095 + }, + { + "epoch": 0.5882897613061645, + "grad_norm": 0.4754330515861511, + "learning_rate": 7.264537053922462e-05, + "loss": 0.028, + "step": 25096 + }, + { + "epoch": 0.5883132028809696, + "grad_norm": 0.5422973036766052, + "learning_rate": 7.263828540943033e-05, + "loss": 0.119, + "step": 25097 + }, + { + "epoch": 0.5883366444557745, + "grad_norm": 0.42850786447525024, + "learning_rate": 7.263120042810143e-05, + "loss": 0.1323, + "step": 25098 + }, + { + "epoch": 0.5883600860305795, + "grad_norm": 0.5214918851852417, + "learning_rate": 7.262411559527646e-05, + "loss": 0.0641, + "step": 25099 + }, + { + "epoch": 0.5883835276053845, + "grad_norm": 0.1369025558233261, + "learning_rate": 7.261703091099378e-05, + "loss": 0.0235, + "step": 25100 + }, + { + "epoch": 0.5884069691801895, + "grad_norm": 0.8783965110778809, + "learning_rate": 7.260994637529188e-05, + "loss": 0.0738, + "step": 25101 + }, + { + "epoch": 0.5884304107549945, + "grad_norm": 0.3978388011455536, + "learning_rate": 7.260286198820918e-05, + "loss": 0.0703, + "step": 25102 + }, + { + "epoch": 0.5884538523297995, + "grad_norm": 0.6828683614730835, + "learning_rate": 7.259577774978411e-05, + "loss": 0.1597, + "step": 25103 + }, + { + "epoch": 0.5884772939046045, + "grad_norm": 0.6253385543823242, + "learning_rate": 7.258869366005514e-05, + "loss": 0.0591, + "step": 25104 + }, + { + "epoch": 0.5885007354794095, + "grad_norm": 1.596490740776062, + "learning_rate": 7.258160971906067e-05, + "loss": 0.3251, + "step": 25105 + }, + { + "epoch": 0.5885241770542144, + "grad_norm": 0.6401325464248657, + "learning_rate": 7.257452592683915e-05, + "loss": 0.1303, + "step": 25106 + }, + { + "epoch": 0.5885476186290195, + "grad_norm": 0.697819173336029, + "learning_rate": 7.256744228342905e-05, + "loss": 0.0953, + "step": 25107 + }, + { + "epoch": 0.5885710602038245, + "grad_norm": 0.7069211602210999, + "learning_rate": 7.256035878886878e-05, + "loss": 0.1379, + "step": 25108 + }, + { + "epoch": 0.5885945017786295, + "grad_norm": 0.31652143597602844, + "learning_rate": 7.255327544319674e-05, + "loss": 0.4578, + "step": 25109 + }, + { + "epoch": 0.5886179433534345, + "grad_norm": 0.2532896101474762, + "learning_rate": 7.254619224645139e-05, + "loss": 0.0435, + "step": 25110 + }, + { + "epoch": 0.5886413849282395, + "grad_norm": 0.44380584359169006, + "learning_rate": 7.25391091986712e-05, + "loss": 0.5057, + "step": 25111 + }, + { + "epoch": 0.5886648265030445, + "grad_norm": 0.4416024386882782, + "learning_rate": 7.25320262998946e-05, + "loss": 0.4431, + "step": 25112 + }, + { + "epoch": 0.5886882680778495, + "grad_norm": 0.5418993234634399, + "learning_rate": 7.252494355015995e-05, + "loss": 0.1012, + "step": 25113 + }, + { + "epoch": 0.5887117096526545, + "grad_norm": 0.4839094877243042, + "learning_rate": 7.251786094950576e-05, + "loss": 0.0894, + "step": 25114 + }, + { + "epoch": 0.5887351512274595, + "grad_norm": 0.30444663763046265, + "learning_rate": 7.251077849797043e-05, + "loss": 0.0561, + "step": 25115 + }, + { + "epoch": 0.5887585928022645, + "grad_norm": 0.4862600862979889, + "learning_rate": 7.250369619559238e-05, + "loss": 0.0642, + "step": 25116 + }, + { + "epoch": 0.5887820343770694, + "grad_norm": 0.7674750089645386, + "learning_rate": 7.249661404241006e-05, + "loss": 0.1268, + "step": 25117 + }, + { + "epoch": 0.5888054759518745, + "grad_norm": 0.4101819396018982, + "learning_rate": 7.248953203846187e-05, + "loss": 0.055, + "step": 25118 + }, + { + "epoch": 0.5888289175266794, + "grad_norm": 0.4410962164402008, + "learning_rate": 7.248245018378626e-05, + "loss": 0.0791, + "step": 25119 + }, + { + "epoch": 0.5888523591014845, + "grad_norm": 0.12306905537843704, + "learning_rate": 7.247536847842164e-05, + "loss": 0.0172, + "step": 25120 + }, + { + "epoch": 0.5888758006762894, + "grad_norm": 0.43646642565727234, + "learning_rate": 7.246828692240647e-05, + "loss": 0.0545, + "step": 25121 + }, + { + "epoch": 0.5888992422510945, + "grad_norm": 0.7035192251205444, + "learning_rate": 7.24612055157791e-05, + "loss": 0.0941, + "step": 25122 + }, + { + "epoch": 0.5889226838258994, + "grad_norm": 0.2682832181453705, + "learning_rate": 7.245412425857802e-05, + "loss": 0.038, + "step": 25123 + }, + { + "epoch": 0.5889461254007045, + "grad_norm": 0.20457081496715546, + "learning_rate": 7.244704315084166e-05, + "loss": 0.1452, + "step": 25124 + }, + { + "epoch": 0.5889695669755094, + "grad_norm": 0.9006316065788269, + "learning_rate": 7.243996219260841e-05, + "loss": 0.1111, + "step": 25125 + }, + { + "epoch": 0.5889930085503144, + "grad_norm": 0.8475857377052307, + "learning_rate": 7.243288138391668e-05, + "loss": 0.1156, + "step": 25126 + }, + { + "epoch": 0.5890164501251194, + "grad_norm": 0.10071080923080444, + "learning_rate": 7.242580072480494e-05, + "loss": 0.016, + "step": 25127 + }, + { + "epoch": 0.5890398916999244, + "grad_norm": 0.3478618860244751, + "learning_rate": 7.241872021531157e-05, + "loss": 0.0756, + "step": 25128 + }, + { + "epoch": 0.5890633332747294, + "grad_norm": 0.13766130805015564, + "learning_rate": 7.2411639855475e-05, + "loss": 0.0251, + "step": 25129 + }, + { + "epoch": 0.5890867748495344, + "grad_norm": 0.3353310823440552, + "learning_rate": 7.240455964533366e-05, + "loss": 0.042, + "step": 25130 + }, + { + "epoch": 0.5891102164243394, + "grad_norm": 0.13576635718345642, + "learning_rate": 7.239747958492594e-05, + "loss": 0.0166, + "step": 25131 + }, + { + "epoch": 0.5891336579991444, + "grad_norm": 0.2296515256166458, + "learning_rate": 7.239039967429026e-05, + "loss": 0.0222, + "step": 25132 + }, + { + "epoch": 0.5891570995739493, + "grad_norm": 0.16170205175876617, + "learning_rate": 7.238331991346508e-05, + "loss": 0.0137, + "step": 25133 + }, + { + "epoch": 0.5891805411487544, + "grad_norm": 0.2723069190979004, + "learning_rate": 7.237624030248872e-05, + "loss": 0.0697, + "step": 25134 + }, + { + "epoch": 0.5892039827235593, + "grad_norm": 0.7265157103538513, + "learning_rate": 7.236916084139971e-05, + "loss": 0.1402, + "step": 25135 + }, + { + "epoch": 0.5892274242983644, + "grad_norm": 0.5526536703109741, + "learning_rate": 7.23620815302364e-05, + "loss": 0.0778, + "step": 25136 + }, + { + "epoch": 0.5892508658731693, + "grad_norm": 1.2399286031723022, + "learning_rate": 7.235500236903723e-05, + "loss": 0.1809, + "step": 25137 + }, + { + "epoch": 0.5892743074479744, + "grad_norm": 0.2885729670524597, + "learning_rate": 7.234792335784055e-05, + "loss": 0.0597, + "step": 25138 + }, + { + "epoch": 0.5892977490227793, + "grad_norm": 0.672995924949646, + "learning_rate": 7.234084449668485e-05, + "loss": 0.0807, + "step": 25139 + }, + { + "epoch": 0.5893211905975844, + "grad_norm": 0.5250492095947266, + "learning_rate": 7.233376578560852e-05, + "loss": 0.1295, + "step": 25140 + }, + { + "epoch": 0.5893446321723893, + "grad_norm": 0.2803446352481842, + "learning_rate": 7.232668722464992e-05, + "loss": 0.0353, + "step": 25141 + }, + { + "epoch": 0.5893680737471944, + "grad_norm": 0.18914319574832916, + "learning_rate": 7.231960881384752e-05, + "loss": 0.0377, + "step": 25142 + }, + { + "epoch": 0.5893915153219993, + "grad_norm": 0.2467014193534851, + "learning_rate": 7.23125305532397e-05, + "loss": 0.0269, + "step": 25143 + }, + { + "epoch": 0.5894149568968043, + "grad_norm": 0.6964625716209412, + "learning_rate": 7.230545244286484e-05, + "loss": 0.1664, + "step": 25144 + }, + { + "epoch": 0.5894383984716093, + "grad_norm": 0.12864847481250763, + "learning_rate": 7.22983744827614e-05, + "loss": 0.0159, + "step": 25145 + }, + { + "epoch": 0.5894618400464143, + "grad_norm": 0.5552235841751099, + "learning_rate": 7.229129667296772e-05, + "loss": 0.1419, + "step": 25146 + }, + { + "epoch": 0.5894852816212193, + "grad_norm": 0.1326669603586197, + "learning_rate": 7.228421901352227e-05, + "loss": 0.0177, + "step": 25147 + }, + { + "epoch": 0.5895087231960243, + "grad_norm": 0.1696786880493164, + "learning_rate": 7.227714150446343e-05, + "loss": 0.0342, + "step": 25148 + }, + { + "epoch": 0.5895321647708293, + "grad_norm": 0.5761035084724426, + "learning_rate": 7.22700641458296e-05, + "loss": 0.0605, + "step": 25149 + }, + { + "epoch": 0.5895556063456343, + "grad_norm": 0.13712197542190552, + "learning_rate": 7.226298693765918e-05, + "loss": 0.0254, + "step": 25150 + }, + { + "epoch": 0.5895790479204392, + "grad_norm": 0.41096124053001404, + "learning_rate": 7.225590987999057e-05, + "loss": 0.093, + "step": 25151 + }, + { + "epoch": 0.5896024894952443, + "grad_norm": 0.4299663305282593, + "learning_rate": 7.224883297286218e-05, + "loss": 0.0858, + "step": 25152 + }, + { + "epoch": 0.5896259310700492, + "grad_norm": 0.17094239592552185, + "learning_rate": 7.224175621631242e-05, + "loss": 0.0125, + "step": 25153 + }, + { + "epoch": 0.5896493726448543, + "grad_norm": 0.5365868806838989, + "learning_rate": 7.223467961037962e-05, + "loss": 0.1244, + "step": 25154 + }, + { + "epoch": 0.5896728142196592, + "grad_norm": 0.59989994764328, + "learning_rate": 7.222760315510229e-05, + "loss": 0.0839, + "step": 25155 + }, + { + "epoch": 0.5896962557944643, + "grad_norm": 0.39170730113983154, + "learning_rate": 7.222052685051873e-05, + "loss": 0.069, + "step": 25156 + }, + { + "epoch": 0.5897196973692692, + "grad_norm": 0.44305962324142456, + "learning_rate": 7.221345069666736e-05, + "loss": 0.0721, + "step": 25157 + }, + { + "epoch": 0.5897431389440743, + "grad_norm": 0.12266446650028229, + "learning_rate": 7.220637469358662e-05, + "loss": 0.02, + "step": 25158 + }, + { + "epoch": 0.5897665805188793, + "grad_norm": 0.6182718873023987, + "learning_rate": 7.219929884131482e-05, + "loss": 0.1523, + "step": 25159 + }, + { + "epoch": 0.5897900220936843, + "grad_norm": 0.6086711883544922, + "learning_rate": 7.219222313989047e-05, + "loss": 0.1547, + "step": 25160 + }, + { + "epoch": 0.5898134636684893, + "grad_norm": 0.49150747060775757, + "learning_rate": 7.218514758935186e-05, + "loss": 0.0574, + "step": 25161 + }, + { + "epoch": 0.5898369052432942, + "grad_norm": 0.29637300968170166, + "learning_rate": 7.217807218973745e-05, + "loss": 0.0263, + "step": 25162 + }, + { + "epoch": 0.5898603468180993, + "grad_norm": 0.4738989770412445, + "learning_rate": 7.217099694108557e-05, + "loss": 0.1488, + "step": 25163 + }, + { + "epoch": 0.5898837883929042, + "grad_norm": 0.18465451896190643, + "learning_rate": 7.216392184343466e-05, + "loss": 0.0454, + "step": 25164 + }, + { + "epoch": 0.5899072299677093, + "grad_norm": 0.6710749268531799, + "learning_rate": 7.21568468968231e-05, + "loss": 0.4924, + "step": 25165 + }, + { + "epoch": 0.5899306715425142, + "grad_norm": 0.277951717376709, + "learning_rate": 7.214977210128927e-05, + "loss": 0.0808, + "step": 25166 + }, + { + "epoch": 0.5899541131173193, + "grad_norm": 0.5286220908164978, + "learning_rate": 7.214269745687154e-05, + "loss": 0.0706, + "step": 25167 + }, + { + "epoch": 0.5899775546921242, + "grad_norm": 0.5092403292655945, + "learning_rate": 7.213562296360834e-05, + "loss": 0.108, + "step": 25168 + }, + { + "epoch": 0.5900009962669293, + "grad_norm": 0.5073137283325195, + "learning_rate": 7.212854862153802e-05, + "loss": 0.0911, + "step": 25169 + }, + { + "epoch": 0.5900244378417342, + "grad_norm": 0.12164202332496643, + "learning_rate": 7.212147443069899e-05, + "loss": 0.0267, + "step": 25170 + }, + { + "epoch": 0.5900478794165392, + "grad_norm": 0.41565707325935364, + "learning_rate": 7.211440039112959e-05, + "loss": 0.1444, + "step": 25171 + }, + { + "epoch": 0.5900713209913442, + "grad_norm": 0.4015485942363739, + "learning_rate": 7.210732650286827e-05, + "loss": 0.071, + "step": 25172 + }, + { + "epoch": 0.5900947625661492, + "grad_norm": 0.1964193731546402, + "learning_rate": 7.210025276595335e-05, + "loss": 0.0259, + "step": 25173 + }, + { + "epoch": 0.5901182041409542, + "grad_norm": 0.4961263835430145, + "learning_rate": 7.209317918042328e-05, + "loss": 0.0352, + "step": 25174 + }, + { + "epoch": 0.5901416457157592, + "grad_norm": 0.4229739308357239, + "learning_rate": 7.208610574631641e-05, + "loss": 0.072, + "step": 25175 + }, + { + "epoch": 0.5901650872905642, + "grad_norm": 0.40657907724380493, + "learning_rate": 7.207903246367109e-05, + "loss": 0.0883, + "step": 25176 + }, + { + "epoch": 0.5901885288653692, + "grad_norm": 0.43398886919021606, + "learning_rate": 7.207195933252574e-05, + "loss": 0.1363, + "step": 25177 + }, + { + "epoch": 0.5902119704401741, + "grad_norm": 0.4991288483142853, + "learning_rate": 7.206488635291872e-05, + "loss": 0.0692, + "step": 25178 + }, + { + "epoch": 0.5902354120149792, + "grad_norm": 0.19696851074695587, + "learning_rate": 7.205781352488841e-05, + "loss": 0.0433, + "step": 25179 + }, + { + "epoch": 0.5902588535897841, + "grad_norm": 0.3267340362071991, + "learning_rate": 7.205074084847319e-05, + "loss": 0.0382, + "step": 25180 + }, + { + "epoch": 0.5902822951645892, + "grad_norm": 0.09634804725646973, + "learning_rate": 7.204366832371144e-05, + "loss": 0.029, + "step": 25181 + }, + { + "epoch": 0.5903057367393941, + "grad_norm": 0.6005579829216003, + "learning_rate": 7.203659595064152e-05, + "loss": 0.1499, + "step": 25182 + }, + { + "epoch": 0.5903291783141992, + "grad_norm": 0.4506874978542328, + "learning_rate": 7.202952372930179e-05, + "loss": 0.0822, + "step": 25183 + }, + { + "epoch": 0.5903526198890041, + "grad_norm": 0.5611375570297241, + "learning_rate": 7.202245165973069e-05, + "loss": 0.1139, + "step": 25184 + }, + { + "epoch": 0.5903760614638092, + "grad_norm": 0.45990216732025146, + "learning_rate": 7.201537974196654e-05, + "loss": 0.0753, + "step": 25185 + }, + { + "epoch": 0.5903995030386141, + "grad_norm": 0.3607358932495117, + "learning_rate": 7.200830797604775e-05, + "loss": 0.0708, + "step": 25186 + }, + { + "epoch": 0.5904229446134192, + "grad_norm": 0.3154371380805969, + "learning_rate": 7.200123636201263e-05, + "loss": 0.3087, + "step": 25187 + }, + { + "epoch": 0.5904463861882241, + "grad_norm": 0.595673143863678, + "learning_rate": 7.199416489989963e-05, + "loss": 0.0874, + "step": 25188 + }, + { + "epoch": 0.5904698277630291, + "grad_norm": 0.2491367608308792, + "learning_rate": 7.198709358974705e-05, + "loss": 0.031, + "step": 25189 + }, + { + "epoch": 0.5904932693378341, + "grad_norm": 0.7136392593383789, + "learning_rate": 7.19800224315933e-05, + "loss": 0.5926, + "step": 25190 + }, + { + "epoch": 0.5905167109126391, + "grad_norm": 0.3443622887134552, + "learning_rate": 7.197295142547675e-05, + "loss": 0.0608, + "step": 25191 + }, + { + "epoch": 0.5905401524874441, + "grad_norm": 0.5523302555084229, + "learning_rate": 7.196588057143572e-05, + "loss": 0.3876, + "step": 25192 + }, + { + "epoch": 0.5905635940622491, + "grad_norm": 0.30027979612350464, + "learning_rate": 7.195880986950863e-05, + "loss": 0.0482, + "step": 25193 + }, + { + "epoch": 0.5905870356370541, + "grad_norm": 0.07851491868495941, + "learning_rate": 7.195173931973384e-05, + "loss": 0.0061, + "step": 25194 + }, + { + "epoch": 0.5906104772118591, + "grad_norm": 0.41695839166641235, + "learning_rate": 7.194466892214965e-05, + "loss": 0.5534, + "step": 25195 + }, + { + "epoch": 0.590633918786664, + "grad_norm": 0.6714498400688171, + "learning_rate": 7.19375986767945e-05, + "loss": 0.4679, + "step": 25196 + }, + { + "epoch": 0.5906573603614691, + "grad_norm": 0.3453350365161896, + "learning_rate": 7.193052858370675e-05, + "loss": 0.096, + "step": 25197 + }, + { + "epoch": 0.590680801936274, + "grad_norm": 0.539923906326294, + "learning_rate": 7.192345864292474e-05, + "loss": 0.0336, + "step": 25198 + }, + { + "epoch": 0.5907042435110791, + "grad_norm": 0.4251694679260254, + "learning_rate": 7.19163888544868e-05, + "loss": 0.0573, + "step": 25199 + }, + { + "epoch": 0.590727685085884, + "grad_norm": 0.1704942286014557, + "learning_rate": 7.190931921843136e-05, + "loss": 0.0288, + "step": 25200 + }, + { + "epoch": 0.5907511266606891, + "grad_norm": 0.5196975469589233, + "learning_rate": 7.190224973479673e-05, + "loss": 0.1305, + "step": 25201 + }, + { + "epoch": 0.590774568235494, + "grad_norm": 0.5168681740760803, + "learning_rate": 7.189518040362127e-05, + "loss": 0.106, + "step": 25202 + }, + { + "epoch": 0.5907980098102991, + "grad_norm": 0.4089916944503784, + "learning_rate": 7.188811122494337e-05, + "loss": 0.0663, + "step": 25203 + }, + { + "epoch": 0.590821451385104, + "grad_norm": 0.46352773904800415, + "learning_rate": 7.188104219880136e-05, + "loss": 0.102, + "step": 25204 + }, + { + "epoch": 0.590844892959909, + "grad_norm": 0.2602554261684418, + "learning_rate": 7.187397332523361e-05, + "loss": 0.0352, + "step": 25205 + }, + { + "epoch": 0.590868334534714, + "grad_norm": 0.43676257133483887, + "learning_rate": 7.186690460427847e-05, + "loss": 0.1225, + "step": 25206 + }, + { + "epoch": 0.590891776109519, + "grad_norm": 0.33836397528648376, + "learning_rate": 7.185983603597429e-05, + "loss": 0.0753, + "step": 25207 + }, + { + "epoch": 0.590915217684324, + "grad_norm": 0.307892382144928, + "learning_rate": 7.185276762035942e-05, + "loss": 0.0486, + "step": 25208 + }, + { + "epoch": 0.590938659259129, + "grad_norm": 0.1865272969007492, + "learning_rate": 7.184569935747224e-05, + "loss": 0.0237, + "step": 25209 + }, + { + "epoch": 0.5909621008339341, + "grad_norm": 0.3497069180011749, + "learning_rate": 7.18386312473511e-05, + "loss": 0.0571, + "step": 25210 + }, + { + "epoch": 0.590985542408739, + "grad_norm": 0.30083727836608887, + "learning_rate": 7.183156329003431e-05, + "loss": 0.0554, + "step": 25211 + }, + { + "epoch": 0.5910089839835441, + "grad_norm": 0.3458321690559387, + "learning_rate": 7.182449548556028e-05, + "loss": 0.0521, + "step": 25212 + }, + { + "epoch": 0.591032425558349, + "grad_norm": 0.25349652767181396, + "learning_rate": 7.181742783396732e-05, + "loss": 0.0271, + "step": 25213 + }, + { + "epoch": 0.5910558671331541, + "grad_norm": 0.5316063761711121, + "learning_rate": 7.181036033529378e-05, + "loss": 0.0632, + "step": 25214 + }, + { + "epoch": 0.591079308707959, + "grad_norm": 0.3256372809410095, + "learning_rate": 7.180329298957804e-05, + "loss": 0.0217, + "step": 25215 + }, + { + "epoch": 0.591102750282764, + "grad_norm": 0.4007304310798645, + "learning_rate": 7.179622579685841e-05, + "loss": 0.0686, + "step": 25216 + }, + { + "epoch": 0.591126191857569, + "grad_norm": 0.37007009983062744, + "learning_rate": 7.178915875717323e-05, + "loss": 0.0734, + "step": 25217 + }, + { + "epoch": 0.591149633432374, + "grad_norm": 0.6497409343719482, + "learning_rate": 7.178209187056091e-05, + "loss": 0.1072, + "step": 25218 + }, + { + "epoch": 0.591173075007179, + "grad_norm": 0.19367483258247375, + "learning_rate": 7.17750251370597e-05, + "loss": 0.0415, + "step": 25219 + }, + { + "epoch": 0.591196516581984, + "grad_norm": 0.4597693085670471, + "learning_rate": 7.176795855670803e-05, + "loss": 0.0846, + "step": 25220 + }, + { + "epoch": 0.591219958156789, + "grad_norm": 0.10796021670103073, + "learning_rate": 7.176089212954419e-05, + "loss": 0.0106, + "step": 25221 + }, + { + "epoch": 0.591243399731594, + "grad_norm": 0.46504929661750793, + "learning_rate": 7.175382585560657e-05, + "loss": 0.0566, + "step": 25222 + }, + { + "epoch": 0.591266841306399, + "grad_norm": 0.40786483883857727, + "learning_rate": 7.17467597349335e-05, + "loss": 0.4716, + "step": 25223 + }, + { + "epoch": 0.591290282881204, + "grad_norm": 0.38884779810905457, + "learning_rate": 7.173969376756327e-05, + "loss": 0.0576, + "step": 25224 + }, + { + "epoch": 0.5913137244560089, + "grad_norm": 0.21625016629695892, + "learning_rate": 7.173262795353427e-05, + "loss": 0.0303, + "step": 25225 + }, + { + "epoch": 0.591337166030814, + "grad_norm": 0.18765243887901306, + "learning_rate": 7.172556229288483e-05, + "loss": 0.0397, + "step": 25226 + }, + { + "epoch": 0.5913606076056189, + "grad_norm": 0.6341661810874939, + "learning_rate": 7.171849678565327e-05, + "loss": 0.101, + "step": 25227 + }, + { + "epoch": 0.591384049180424, + "grad_norm": 0.5232693552970886, + "learning_rate": 7.171143143187795e-05, + "loss": 0.08, + "step": 25228 + }, + { + "epoch": 0.5914074907552289, + "grad_norm": 0.995400607585907, + "learning_rate": 7.170436623159722e-05, + "loss": 0.1378, + "step": 25229 + }, + { + "epoch": 0.591430932330034, + "grad_norm": 0.18008442223072052, + "learning_rate": 7.169730118484936e-05, + "loss": 0.0318, + "step": 25230 + }, + { + "epoch": 0.5914543739048389, + "grad_norm": 0.26247549057006836, + "learning_rate": 7.169023629167275e-05, + "loss": 0.0152, + "step": 25231 + }, + { + "epoch": 0.591477815479644, + "grad_norm": 0.9236041307449341, + "learning_rate": 7.168317155210569e-05, + "loss": 0.2108, + "step": 25232 + }, + { + "epoch": 0.5915012570544489, + "grad_norm": 0.3162188231945038, + "learning_rate": 7.167610696618657e-05, + "loss": 0.0435, + "step": 25233 + }, + { + "epoch": 0.5915246986292539, + "grad_norm": 0.12137982249259949, + "learning_rate": 7.166904253395368e-05, + "loss": 0.0226, + "step": 25234 + }, + { + "epoch": 0.5915481402040589, + "grad_norm": 0.20569048821926117, + "learning_rate": 7.166197825544538e-05, + "loss": 0.045, + "step": 25235 + }, + { + "epoch": 0.5915715817788639, + "grad_norm": 0.4769582450389862, + "learning_rate": 7.165491413069997e-05, + "loss": 0.0488, + "step": 25236 + }, + { + "epoch": 0.5915950233536689, + "grad_norm": 0.7380474209785461, + "learning_rate": 7.164785015975579e-05, + "loss": 0.09, + "step": 25237 + }, + { + "epoch": 0.5916184649284739, + "grad_norm": 0.6148788928985596, + "learning_rate": 7.164078634265116e-05, + "loss": 0.5574, + "step": 25238 + }, + { + "epoch": 0.5916419065032789, + "grad_norm": 0.2758338749408722, + "learning_rate": 7.163372267942445e-05, + "loss": 0.0625, + "step": 25239 + }, + { + "epoch": 0.5916653480780839, + "grad_norm": 0.5225195288658142, + "learning_rate": 7.162665917011393e-05, + "loss": 0.1116, + "step": 25240 + }, + { + "epoch": 0.5916887896528888, + "grad_norm": 0.4444088637828827, + "learning_rate": 7.161959581475798e-05, + "loss": 0.388, + "step": 25241 + }, + { + "epoch": 0.5917122312276939, + "grad_norm": 0.3569474220275879, + "learning_rate": 7.16125326133949e-05, + "loss": 0.429, + "step": 25242 + }, + { + "epoch": 0.5917356728024988, + "grad_norm": 0.530925989151001, + "learning_rate": 7.1605469566063e-05, + "loss": 0.8611, + "step": 25243 + }, + { + "epoch": 0.5917591143773039, + "grad_norm": 0.4768519103527069, + "learning_rate": 7.159840667280059e-05, + "loss": 0.0808, + "step": 25244 + }, + { + "epoch": 0.5917825559521088, + "grad_norm": 0.3433018624782562, + "learning_rate": 7.159134393364606e-05, + "loss": 0.0745, + "step": 25245 + }, + { + "epoch": 0.5918059975269139, + "grad_norm": 0.8991045951843262, + "learning_rate": 7.158428134863769e-05, + "loss": 0.2584, + "step": 25246 + }, + { + "epoch": 0.5918294391017188, + "grad_norm": 0.23896124958992004, + "learning_rate": 7.157721891781383e-05, + "loss": 0.0449, + "step": 25247 + }, + { + "epoch": 0.5918528806765239, + "grad_norm": 0.4581235349178314, + "learning_rate": 7.157015664121277e-05, + "loss": 0.0955, + "step": 25248 + }, + { + "epoch": 0.5918763222513288, + "grad_norm": 0.4549141228199005, + "learning_rate": 7.156309451887282e-05, + "loss": 0.0833, + "step": 25249 + }, + { + "epoch": 0.5918997638261339, + "grad_norm": 0.26283153891563416, + "learning_rate": 7.155603255083234e-05, + "loss": 0.0163, + "step": 25250 + }, + { + "epoch": 0.5919232054009388, + "grad_norm": 0.470379501581192, + "learning_rate": 7.154897073712962e-05, + "loss": 0.0746, + "step": 25251 + }, + { + "epoch": 0.5919466469757438, + "grad_norm": 0.4729973077774048, + "learning_rate": 7.154190907780298e-05, + "loss": 0.0648, + "step": 25252 + }, + { + "epoch": 0.5919700885505488, + "grad_norm": 0.15832968056201935, + "learning_rate": 7.153484757289075e-05, + "loss": 0.0325, + "step": 25253 + }, + { + "epoch": 0.5919935301253538, + "grad_norm": 0.6500359773635864, + "learning_rate": 7.152778622243124e-05, + "loss": 0.1323, + "step": 25254 + }, + { + "epoch": 0.5920169717001588, + "grad_norm": 0.31454116106033325, + "learning_rate": 7.152072502646276e-05, + "loss": 0.0283, + "step": 25255 + }, + { + "epoch": 0.5920404132749638, + "grad_norm": 0.28833362460136414, + "learning_rate": 7.151366398502358e-05, + "loss": 0.0357, + "step": 25256 + }, + { + "epoch": 0.5920638548497688, + "grad_norm": 0.13664229214191437, + "learning_rate": 7.150660309815211e-05, + "loss": 0.0181, + "step": 25257 + }, + { + "epoch": 0.5920872964245738, + "grad_norm": 0.23230652511119843, + "learning_rate": 7.149954236588662e-05, + "loss": 0.0494, + "step": 25258 + }, + { + "epoch": 0.5921107379993787, + "grad_norm": 0.15872378647327423, + "learning_rate": 7.149248178826538e-05, + "loss": 0.0269, + "step": 25259 + }, + { + "epoch": 0.5921341795741838, + "grad_norm": 0.07852023094892502, + "learning_rate": 7.148542136532677e-05, + "loss": 0.0073, + "step": 25260 + }, + { + "epoch": 0.5921576211489888, + "grad_norm": 0.2490461766719818, + "learning_rate": 7.147836109710906e-05, + "loss": 0.0424, + "step": 25261 + }, + { + "epoch": 0.5921810627237938, + "grad_norm": 0.2419515997171402, + "learning_rate": 7.147130098365054e-05, + "loss": 0.0365, + "step": 25262 + }, + { + "epoch": 0.5922045042985988, + "grad_norm": 0.38284242153167725, + "learning_rate": 7.146424102498956e-05, + "loss": 0.5927, + "step": 25263 + }, + { + "epoch": 0.5922279458734038, + "grad_norm": 1.192826271057129, + "learning_rate": 7.145718122116442e-05, + "loss": 0.0956, + "step": 25264 + }, + { + "epoch": 0.5922513874482088, + "grad_norm": 0.37040993571281433, + "learning_rate": 7.14501215722134e-05, + "loss": 0.0453, + "step": 25265 + }, + { + "epoch": 0.5922748290230138, + "grad_norm": 0.32799777388572693, + "learning_rate": 7.144306207817484e-05, + "loss": 0.0537, + "step": 25266 + }, + { + "epoch": 0.5922982705978188, + "grad_norm": 0.5218610167503357, + "learning_rate": 7.143600273908703e-05, + "loss": 0.1688, + "step": 25267 + }, + { + "epoch": 0.5923217121726237, + "grad_norm": 0.41248345375061035, + "learning_rate": 7.142894355498823e-05, + "loss": 0.0688, + "step": 25268 + }, + { + "epoch": 0.5923451537474288, + "grad_norm": 0.5854545831680298, + "learning_rate": 7.14218845259168e-05, + "loss": 0.7466, + "step": 25269 + }, + { + "epoch": 0.5923685953222337, + "grad_norm": 0.184282124042511, + "learning_rate": 7.141482565191105e-05, + "loss": 0.0363, + "step": 25270 + }, + { + "epoch": 0.5923920368970388, + "grad_norm": 0.37601277232170105, + "learning_rate": 7.140776693300926e-05, + "loss": 0.0802, + "step": 25271 + }, + { + "epoch": 0.5924154784718437, + "grad_norm": 0.07719281315803528, + "learning_rate": 7.140070836924972e-05, + "loss": 0.0073, + "step": 25272 + }, + { + "epoch": 0.5924389200466488, + "grad_norm": 0.16927726566791534, + "learning_rate": 7.139364996067075e-05, + "loss": 0.0397, + "step": 25273 + }, + { + "epoch": 0.5924623616214537, + "grad_norm": 0.4469565153121948, + "learning_rate": 7.138659170731064e-05, + "loss": 0.0824, + "step": 25274 + }, + { + "epoch": 0.5924858031962588, + "grad_norm": 0.5270856022834778, + "learning_rate": 7.137953360920768e-05, + "loss": 0.0972, + "step": 25275 + }, + { + "epoch": 0.5925092447710637, + "grad_norm": 0.6747596859931946, + "learning_rate": 7.137247566640017e-05, + "loss": 0.5129, + "step": 25276 + }, + { + "epoch": 0.5925326863458688, + "grad_norm": 0.4380663335323334, + "learning_rate": 7.136541787892643e-05, + "loss": 0.1169, + "step": 25277 + }, + { + "epoch": 0.5925561279206737, + "grad_norm": 0.5501932501792908, + "learning_rate": 7.135836024682472e-05, + "loss": 0.6393, + "step": 25278 + }, + { + "epoch": 0.5925795694954787, + "grad_norm": 0.25469478964805603, + "learning_rate": 7.135130277013336e-05, + "loss": 0.0541, + "step": 25279 + }, + { + "epoch": 0.5926030110702837, + "grad_norm": 0.3267759084701538, + "learning_rate": 7.134424544889061e-05, + "loss": 0.072, + "step": 25280 + }, + { + "epoch": 0.5926264526450887, + "grad_norm": 0.3438290059566498, + "learning_rate": 7.133718828313482e-05, + "loss": 0.0513, + "step": 25281 + }, + { + "epoch": 0.5926498942198937, + "grad_norm": 0.4086448550224304, + "learning_rate": 7.133013127290425e-05, + "loss": 0.0937, + "step": 25282 + }, + { + "epoch": 0.5926733357946987, + "grad_norm": 0.11493897438049316, + "learning_rate": 7.13230744182372e-05, + "loss": 0.0161, + "step": 25283 + }, + { + "epoch": 0.5926967773695037, + "grad_norm": 0.26026418805122375, + "learning_rate": 7.131601771917193e-05, + "loss": 0.0356, + "step": 25284 + }, + { + "epoch": 0.5927202189443087, + "grad_norm": 0.16986523568630219, + "learning_rate": 7.130896117574678e-05, + "loss": 0.043, + "step": 25285 + }, + { + "epoch": 0.5927436605191136, + "grad_norm": 0.12162752449512482, + "learning_rate": 7.1301904788e-05, + "loss": 0.0177, + "step": 25286 + }, + { + "epoch": 0.5927671020939187, + "grad_norm": 0.6002218723297119, + "learning_rate": 7.12948485559699e-05, + "loss": 0.069, + "step": 25287 + }, + { + "epoch": 0.5927905436687236, + "grad_norm": 0.3214830458164215, + "learning_rate": 7.128779247969477e-05, + "loss": 0.0459, + "step": 25288 + }, + { + "epoch": 0.5928139852435287, + "grad_norm": 0.5185866951942444, + "learning_rate": 7.128073655921287e-05, + "loss": 0.091, + "step": 25289 + }, + { + "epoch": 0.5928374268183336, + "grad_norm": 0.42087632417678833, + "learning_rate": 7.12736807945625e-05, + "loss": 0.048, + "step": 25290 + }, + { + "epoch": 0.5928608683931387, + "grad_norm": 0.631451427936554, + "learning_rate": 7.126662518578195e-05, + "loss": 0.1427, + "step": 25291 + }, + { + "epoch": 0.5928843099679436, + "grad_norm": 0.3747531473636627, + "learning_rate": 7.125956973290945e-05, + "loss": 0.0319, + "step": 25292 + }, + { + "epoch": 0.5929077515427487, + "grad_norm": 0.13885565102100372, + "learning_rate": 7.12525144359834e-05, + "loss": 0.0204, + "step": 25293 + }, + { + "epoch": 0.5929311931175536, + "grad_norm": 0.11200293153524399, + "learning_rate": 7.124545929504198e-05, + "loss": 0.0141, + "step": 25294 + }, + { + "epoch": 0.5929546346923587, + "grad_norm": 0.22926735877990723, + "learning_rate": 7.123840431012353e-05, + "loss": 0.0478, + "step": 25295 + }, + { + "epoch": 0.5929780762671636, + "grad_norm": 0.6572253108024597, + "learning_rate": 7.123134948126631e-05, + "loss": 0.1904, + "step": 25296 + }, + { + "epoch": 0.5930015178419686, + "grad_norm": 0.38654521107673645, + "learning_rate": 7.122429480850856e-05, + "loss": 0.0646, + "step": 25297 + }, + { + "epoch": 0.5930249594167736, + "grad_norm": 0.16248048841953278, + "learning_rate": 7.121724029188863e-05, + "loss": 0.0224, + "step": 25298 + }, + { + "epoch": 0.5930484009915786, + "grad_norm": 0.4728693962097168, + "learning_rate": 7.121018593144478e-05, + "loss": 0.0429, + "step": 25299 + }, + { + "epoch": 0.5930718425663836, + "grad_norm": 0.1542728990316391, + "learning_rate": 7.120313172721522e-05, + "loss": 0.0257, + "step": 25300 + }, + { + "epoch": 0.5930952841411886, + "grad_norm": 0.46151983737945557, + "learning_rate": 7.11960776792383e-05, + "loss": 0.0731, + "step": 25301 + }, + { + "epoch": 0.5931187257159936, + "grad_norm": 0.6798660755157471, + "learning_rate": 7.11890237875523e-05, + "loss": 0.1414, + "step": 25302 + }, + { + "epoch": 0.5931421672907986, + "grad_norm": 0.5780606269836426, + "learning_rate": 7.118197005219543e-05, + "loss": 0.0884, + "step": 25303 + }, + { + "epoch": 0.5931656088656035, + "grad_norm": 0.506621241569519, + "learning_rate": 7.117491647320597e-05, + "loss": 0.1202, + "step": 25304 + }, + { + "epoch": 0.5931890504404086, + "grad_norm": 0.5947623252868652, + "learning_rate": 7.116786305062228e-05, + "loss": 0.081, + "step": 25305 + }, + { + "epoch": 0.5932124920152135, + "grad_norm": 0.9154940843582153, + "learning_rate": 7.116080978448257e-05, + "loss": 0.2627, + "step": 25306 + }, + { + "epoch": 0.5932359335900186, + "grad_norm": 0.4395977258682251, + "learning_rate": 7.11537566748251e-05, + "loss": 0.0751, + "step": 25307 + }, + { + "epoch": 0.5932593751648235, + "grad_norm": 0.5216322541236877, + "learning_rate": 7.114670372168817e-05, + "loss": 0.0796, + "step": 25308 + }, + { + "epoch": 0.5932828167396286, + "grad_norm": 0.47119230031967163, + "learning_rate": 7.113965092511003e-05, + "loss": 0.076, + "step": 25309 + }, + { + "epoch": 0.5933062583144335, + "grad_norm": 0.49421247839927673, + "learning_rate": 7.113259828512895e-05, + "loss": 0.1258, + "step": 25310 + }, + { + "epoch": 0.5933296998892386, + "grad_norm": 0.41972121596336365, + "learning_rate": 7.112554580178321e-05, + "loss": 0.056, + "step": 25311 + }, + { + "epoch": 0.5933531414640436, + "grad_norm": 0.1546507477760315, + "learning_rate": 7.111849347511108e-05, + "loss": 0.0297, + "step": 25312 + }, + { + "epoch": 0.5933765830388485, + "grad_norm": 0.19008713960647583, + "learning_rate": 7.111144130515079e-05, + "loss": 0.0272, + "step": 25313 + }, + { + "epoch": 0.5934000246136536, + "grad_norm": 0.4285849630832672, + "learning_rate": 7.110438929194067e-05, + "loss": 0.0573, + "step": 25314 + }, + { + "epoch": 0.5934234661884585, + "grad_norm": 0.1483493596315384, + "learning_rate": 7.109733743551894e-05, + "loss": 0.027, + "step": 25315 + }, + { + "epoch": 0.5934469077632636, + "grad_norm": 0.3483582139015198, + "learning_rate": 7.109028573592384e-05, + "loss": 0.0816, + "step": 25316 + }, + { + "epoch": 0.5934703493380685, + "grad_norm": 0.41630733013153076, + "learning_rate": 7.108323419319364e-05, + "loss": 0.0781, + "step": 25317 + }, + { + "epoch": 0.5934937909128736, + "grad_norm": 0.6408789157867432, + "learning_rate": 7.107618280736668e-05, + "loss": 0.577, + "step": 25318 + }, + { + "epoch": 0.5935172324876785, + "grad_norm": 0.31665024161338806, + "learning_rate": 7.106913157848115e-05, + "loss": 0.0275, + "step": 25319 + }, + { + "epoch": 0.5935406740624836, + "grad_norm": 0.21646948158740997, + "learning_rate": 7.106208050657532e-05, + "loss": 0.0516, + "step": 25320 + }, + { + "epoch": 0.5935641156372885, + "grad_norm": 0.7741660475730896, + "learning_rate": 7.105502959168746e-05, + "loss": 0.8071, + "step": 25321 + }, + { + "epoch": 0.5935875572120936, + "grad_norm": 0.37383806705474854, + "learning_rate": 7.10479788338558e-05, + "loss": 0.0621, + "step": 25322 + }, + { + "epoch": 0.5936109987868985, + "grad_norm": 0.27540603280067444, + "learning_rate": 7.104092823311865e-05, + "loss": 0.0358, + "step": 25323 + }, + { + "epoch": 0.5936344403617035, + "grad_norm": 0.354832261800766, + "learning_rate": 7.103387778951425e-05, + "loss": 0.0705, + "step": 25324 + }, + { + "epoch": 0.5936578819365085, + "grad_norm": 0.139883890748024, + "learning_rate": 7.102682750308081e-05, + "loss": 0.0333, + "step": 25325 + }, + { + "epoch": 0.5936813235113135, + "grad_norm": 0.48228368163108826, + "learning_rate": 7.101977737385663e-05, + "loss": 0.7095, + "step": 25326 + }, + { + "epoch": 0.5937047650861185, + "grad_norm": 0.6532460451126099, + "learning_rate": 7.101272740187996e-05, + "loss": 0.0903, + "step": 25327 + }, + { + "epoch": 0.5937282066609235, + "grad_norm": 0.5859694480895996, + "learning_rate": 7.100567758718904e-05, + "loss": 0.1757, + "step": 25328 + }, + { + "epoch": 0.5937516482357285, + "grad_norm": 0.3013959527015686, + "learning_rate": 7.09986279298221e-05, + "loss": 0.0636, + "step": 25329 + }, + { + "epoch": 0.5937750898105335, + "grad_norm": 0.2677820026874542, + "learning_rate": 7.099157842981746e-05, + "loss": 0.0251, + "step": 25330 + }, + { + "epoch": 0.5937985313853384, + "grad_norm": 0.24966369569301605, + "learning_rate": 7.098452908721333e-05, + "loss": 0.0633, + "step": 25331 + }, + { + "epoch": 0.5938219729601435, + "grad_norm": 0.4310985207557678, + "learning_rate": 7.097747990204793e-05, + "loss": 0.0304, + "step": 25332 + }, + { + "epoch": 0.5938454145349484, + "grad_norm": 0.7236248850822449, + "learning_rate": 7.097043087435957e-05, + "loss": 0.1957, + "step": 25333 + }, + { + "epoch": 0.5938688561097535, + "grad_norm": 0.21700482070446014, + "learning_rate": 7.096338200418648e-05, + "loss": 0.0288, + "step": 25334 + }, + { + "epoch": 0.5938922976845584, + "grad_norm": 0.3936397433280945, + "learning_rate": 7.095633329156686e-05, + "loss": 0.0844, + "step": 25335 + }, + { + "epoch": 0.5939157392593635, + "grad_norm": 0.1537836790084839, + "learning_rate": 7.094928473653901e-05, + "loss": 0.0331, + "step": 25336 + }, + { + "epoch": 0.5939391808341684, + "grad_norm": 0.36061397194862366, + "learning_rate": 7.094223633914117e-05, + "loss": 0.0567, + "step": 25337 + }, + { + "epoch": 0.5939626224089735, + "grad_norm": 0.35107889771461487, + "learning_rate": 7.093518809941156e-05, + "loss": 0.0695, + "step": 25338 + }, + { + "epoch": 0.5939860639837784, + "grad_norm": 0.49789366126060486, + "learning_rate": 7.092814001738844e-05, + "loss": 0.0832, + "step": 25339 + }, + { + "epoch": 0.5940095055585835, + "grad_norm": 0.21426044404506683, + "learning_rate": 7.092109209311004e-05, + "loss": 0.0404, + "step": 25340 + }, + { + "epoch": 0.5940329471333884, + "grad_norm": 0.11342902481555939, + "learning_rate": 7.09140443266146e-05, + "loss": 0.0177, + "step": 25341 + }, + { + "epoch": 0.5940563887081934, + "grad_norm": 0.07543619722127914, + "learning_rate": 7.090699671794036e-05, + "loss": 0.0159, + "step": 25342 + }, + { + "epoch": 0.5940798302829984, + "grad_norm": 0.5102484822273254, + "learning_rate": 7.089994926712563e-05, + "loss": 0.0973, + "step": 25343 + }, + { + "epoch": 0.5941032718578034, + "grad_norm": 0.3650941252708435, + "learning_rate": 7.089290197420856e-05, + "loss": 0.0463, + "step": 25344 + }, + { + "epoch": 0.5941267134326084, + "grad_norm": 1.007900357246399, + "learning_rate": 7.088585483922742e-05, + "loss": 0.1941, + "step": 25345 + }, + { + "epoch": 0.5941501550074134, + "grad_norm": 0.1368914246559143, + "learning_rate": 7.087880786222047e-05, + "loss": 0.0234, + "step": 25346 + }, + { + "epoch": 0.5941735965822184, + "grad_norm": 0.631321132183075, + "learning_rate": 7.087176104322592e-05, + "loss": 0.1037, + "step": 25347 + }, + { + "epoch": 0.5941970381570234, + "grad_norm": 0.47565826773643494, + "learning_rate": 7.086471438228201e-05, + "loss": 0.1934, + "step": 25348 + }, + { + "epoch": 0.5942204797318283, + "grad_norm": 0.16972443461418152, + "learning_rate": 7.085766787942698e-05, + "loss": 0.0191, + "step": 25349 + }, + { + "epoch": 0.5942439213066334, + "grad_norm": 0.6948354840278625, + "learning_rate": 7.085062153469908e-05, + "loss": 0.1519, + "step": 25350 + }, + { + "epoch": 0.5942673628814383, + "grad_norm": 0.9233744144439697, + "learning_rate": 7.084357534813651e-05, + "loss": 0.1054, + "step": 25351 + }, + { + "epoch": 0.5942908044562434, + "grad_norm": 0.42655882239341736, + "learning_rate": 7.083652931977751e-05, + "loss": 0.0881, + "step": 25352 + }, + { + "epoch": 0.5943142460310483, + "grad_norm": 0.48330599069595337, + "learning_rate": 7.08294834496603e-05, + "loss": 0.1036, + "step": 25353 + }, + { + "epoch": 0.5943376876058534, + "grad_norm": 0.49989545345306396, + "learning_rate": 7.082243773782318e-05, + "loss": 0.46, + "step": 25354 + }, + { + "epoch": 0.5943611291806583, + "grad_norm": 0.19222703576087952, + "learning_rate": 7.081539218430431e-05, + "loss": 0.0235, + "step": 25355 + }, + { + "epoch": 0.5943845707554634, + "grad_norm": 0.4026462137699127, + "learning_rate": 7.080834678914196e-05, + "loss": 0.0714, + "step": 25356 + }, + { + "epoch": 0.5944080123302683, + "grad_norm": 0.4445631206035614, + "learning_rate": 7.080130155237434e-05, + "loss": 0.1087, + "step": 25357 + }, + { + "epoch": 0.5944314539050733, + "grad_norm": 0.3516829013824463, + "learning_rate": 7.079425647403965e-05, + "loss": 0.095, + "step": 25358 + }, + { + "epoch": 0.5944548954798783, + "grad_norm": 0.571363091468811, + "learning_rate": 7.078721155417618e-05, + "loss": 0.5338, + "step": 25359 + }, + { + "epoch": 0.5944783370546833, + "grad_norm": 0.3015291094779968, + "learning_rate": 7.07801667928221e-05, + "loss": 0.0267, + "step": 25360 + }, + { + "epoch": 0.5945017786294883, + "grad_norm": 0.38750895857810974, + "learning_rate": 7.077312219001567e-05, + "loss": 0.0632, + "step": 25361 + }, + { + "epoch": 0.5945252202042933, + "grad_norm": 0.5063624978065491, + "learning_rate": 7.07660777457951e-05, + "loss": 0.0951, + "step": 25362 + }, + { + "epoch": 0.5945486617790984, + "grad_norm": 0.12952002882957458, + "learning_rate": 7.07590334601986e-05, + "loss": 0.033, + "step": 25363 + }, + { + "epoch": 0.5945721033539033, + "grad_norm": 0.6452958583831787, + "learning_rate": 7.075198933326443e-05, + "loss": 0.484, + "step": 25364 + }, + { + "epoch": 0.5945955449287084, + "grad_norm": 0.20049847662448883, + "learning_rate": 7.074494536503075e-05, + "loss": 0.0546, + "step": 25365 + }, + { + "epoch": 0.5946189865035133, + "grad_norm": 0.3915305435657501, + "learning_rate": 7.073790155553584e-05, + "loss": 0.0593, + "step": 25366 + }, + { + "epoch": 0.5946424280783184, + "grad_norm": 0.2652563750743866, + "learning_rate": 7.07308579048179e-05, + "loss": 0.0657, + "step": 25367 + }, + { + "epoch": 0.5946658696531233, + "grad_norm": 0.35161590576171875, + "learning_rate": 7.072381441291515e-05, + "loss": 0.0654, + "step": 25368 + }, + { + "epoch": 0.5946893112279283, + "grad_norm": 0.1326351761817932, + "learning_rate": 7.071677107986583e-05, + "loss": 0.0085, + "step": 25369 + }, + { + "epoch": 0.5947127528027333, + "grad_norm": 0.592134416103363, + "learning_rate": 7.070972790570808e-05, + "loss": 0.4586, + "step": 25370 + }, + { + "epoch": 0.5947361943775383, + "grad_norm": 0.14779992401599884, + "learning_rate": 7.070268489048021e-05, + "loss": 0.0282, + "step": 25371 + }, + { + "epoch": 0.5947596359523433, + "grad_norm": 0.2790604829788208, + "learning_rate": 7.069564203422041e-05, + "loss": 0.0518, + "step": 25372 + }, + { + "epoch": 0.5947830775271483, + "grad_norm": 0.357357919216156, + "learning_rate": 7.068859933696683e-05, + "loss": 0.0484, + "step": 25373 + }, + { + "epoch": 0.5948065191019533, + "grad_norm": 0.3766433894634247, + "learning_rate": 7.068155679875779e-05, + "loss": 0.0318, + "step": 25374 + }, + { + "epoch": 0.5948299606767583, + "grad_norm": 0.04253068193793297, + "learning_rate": 7.067451441963143e-05, + "loss": 0.0064, + "step": 25375 + }, + { + "epoch": 0.5948534022515632, + "grad_norm": 0.5096739530563354, + "learning_rate": 7.066747219962594e-05, + "loss": 0.046, + "step": 25376 + }, + { + "epoch": 0.5948768438263683, + "grad_norm": 0.13031725585460663, + "learning_rate": 7.066043013877957e-05, + "loss": 0.0375, + "step": 25377 + }, + { + "epoch": 0.5949002854011732, + "grad_norm": 0.3748100996017456, + "learning_rate": 7.065338823713057e-05, + "loss": 0.3105, + "step": 25378 + }, + { + "epoch": 0.5949237269759783, + "grad_norm": 0.4901840090751648, + "learning_rate": 7.06463464947171e-05, + "loss": 0.1212, + "step": 25379 + }, + { + "epoch": 0.5949471685507832, + "grad_norm": 0.3570377826690674, + "learning_rate": 7.063930491157736e-05, + "loss": 0.0441, + "step": 25380 + }, + { + "epoch": 0.5949706101255883, + "grad_norm": 0.22698524594306946, + "learning_rate": 7.06322634877496e-05, + "loss": 0.057, + "step": 25381 + }, + { + "epoch": 0.5949940517003932, + "grad_norm": 0.38262301683425903, + "learning_rate": 7.062522222327202e-05, + "loss": 0.0661, + "step": 25382 + }, + { + "epoch": 0.5950174932751983, + "grad_norm": 0.25614893436431885, + "learning_rate": 7.061818111818277e-05, + "loss": 0.0861, + "step": 25383 + }, + { + "epoch": 0.5950409348500032, + "grad_norm": 0.21629004180431366, + "learning_rate": 7.061114017252014e-05, + "loss": 0.0385, + "step": 25384 + }, + { + "epoch": 0.5950643764248083, + "grad_norm": 0.4056874215602875, + "learning_rate": 7.060409938632227e-05, + "loss": 0.1026, + "step": 25385 + }, + { + "epoch": 0.5950878179996132, + "grad_norm": 0.529845118522644, + "learning_rate": 7.059705875962736e-05, + "loss": 0.7693, + "step": 25386 + }, + { + "epoch": 0.5951112595744182, + "grad_norm": 0.5875534415245056, + "learning_rate": 7.059001829247367e-05, + "loss": 0.0699, + "step": 25387 + }, + { + "epoch": 0.5951347011492232, + "grad_norm": 0.3467232286930084, + "learning_rate": 7.058297798489935e-05, + "loss": 0.0594, + "step": 25388 + }, + { + "epoch": 0.5951581427240282, + "grad_norm": 0.2698882520198822, + "learning_rate": 7.057593783694261e-05, + "loss": 0.0382, + "step": 25389 + }, + { + "epoch": 0.5951815842988332, + "grad_norm": 0.7456487417221069, + "learning_rate": 7.056889784864163e-05, + "loss": 0.1261, + "step": 25390 + }, + { + "epoch": 0.5952050258736382, + "grad_norm": 0.43237659335136414, + "learning_rate": 7.05618580200347e-05, + "loss": 0.0477, + "step": 25391 + }, + { + "epoch": 0.5952284674484432, + "grad_norm": 0.2664048969745636, + "learning_rate": 7.055481835115994e-05, + "loss": 0.0269, + "step": 25392 + }, + { + "epoch": 0.5952519090232482, + "grad_norm": 0.18482664227485657, + "learning_rate": 7.054777884205555e-05, + "loss": 0.0173, + "step": 25393 + }, + { + "epoch": 0.5952753505980531, + "grad_norm": 0.4800603985786438, + "learning_rate": 7.054073949275975e-05, + "loss": 0.0492, + "step": 25394 + }, + { + "epoch": 0.5952987921728582, + "grad_norm": 0.5789722800254822, + "learning_rate": 7.053370030331073e-05, + "loss": 0.0603, + "step": 25395 + }, + { + "epoch": 0.5953222337476631, + "grad_norm": 0.4012768566608429, + "learning_rate": 7.052666127374665e-05, + "loss": 0.0806, + "step": 25396 + }, + { + "epoch": 0.5953456753224682, + "grad_norm": 0.5478939414024353, + "learning_rate": 7.051962240410577e-05, + "loss": 0.0711, + "step": 25397 + }, + { + "epoch": 0.5953691168972731, + "grad_norm": 0.6728467345237732, + "learning_rate": 7.051258369442623e-05, + "loss": 0.1546, + "step": 25398 + }, + { + "epoch": 0.5953925584720782, + "grad_norm": 0.46788081526756287, + "learning_rate": 7.050554514474626e-05, + "loss": 0.0947, + "step": 25399 + }, + { + "epoch": 0.5954160000468831, + "grad_norm": 0.6148096919059753, + "learning_rate": 7.049850675510401e-05, + "loss": 0.0663, + "step": 25400 + }, + { + "epoch": 0.5954394416216882, + "grad_norm": 0.29285165667533875, + "learning_rate": 7.04914685255377e-05, + "loss": 0.0687, + "step": 25401 + }, + { + "epoch": 0.5954628831964931, + "grad_norm": 0.6867154240608215, + "learning_rate": 7.048443045608546e-05, + "loss": 0.1051, + "step": 25402 + }, + { + "epoch": 0.5954863247712981, + "grad_norm": 0.4122805595397949, + "learning_rate": 7.04773925467856e-05, + "loss": 0.0984, + "step": 25403 + }, + { + "epoch": 0.5955097663461031, + "grad_norm": 0.11539885401725769, + "learning_rate": 7.047035479767623e-05, + "loss": 0.0271, + "step": 25404 + }, + { + "epoch": 0.5955332079209081, + "grad_norm": 0.09674682468175888, + "learning_rate": 7.04633172087955e-05, + "loss": 0.0131, + "step": 25405 + }, + { + "epoch": 0.5955566494957131, + "grad_norm": 0.15713073313236237, + "learning_rate": 7.04562797801817e-05, + "loss": 0.0261, + "step": 25406 + }, + { + "epoch": 0.5955800910705181, + "grad_norm": 0.5686013102531433, + "learning_rate": 7.044924251187294e-05, + "loss": 0.7356, + "step": 25407 + }, + { + "epoch": 0.5956035326453231, + "grad_norm": 0.1965024769306183, + "learning_rate": 7.04422054039074e-05, + "loss": 0.0509, + "step": 25408 + }, + { + "epoch": 0.5956269742201281, + "grad_norm": 0.11007481813430786, + "learning_rate": 7.043516845632331e-05, + "loss": 0.0154, + "step": 25409 + }, + { + "epoch": 0.595650415794933, + "grad_norm": 0.5180685520172119, + "learning_rate": 7.042813166915882e-05, + "loss": 0.1009, + "step": 25410 + }, + { + "epoch": 0.5956738573697381, + "grad_norm": 0.5006284117698669, + "learning_rate": 7.042109504245211e-05, + "loss": 0.0633, + "step": 25411 + }, + { + "epoch": 0.595697298944543, + "grad_norm": 0.11162486672401428, + "learning_rate": 7.041405857624137e-05, + "loss": 0.0118, + "step": 25412 + }, + { + "epoch": 0.5957207405193481, + "grad_norm": 0.5803369283676147, + "learning_rate": 7.040702227056479e-05, + "loss": 0.1229, + "step": 25413 + }, + { + "epoch": 0.595744182094153, + "grad_norm": 0.6185377240180969, + "learning_rate": 7.03999861254605e-05, + "loss": 0.09, + "step": 25414 + }, + { + "epoch": 0.5957676236689581, + "grad_norm": 0.562828540802002, + "learning_rate": 7.039295014096675e-05, + "loss": 0.1029, + "step": 25415 + }, + { + "epoch": 0.5957910652437631, + "grad_norm": 0.42103543877601624, + "learning_rate": 7.038591431712169e-05, + "loss": 0.0539, + "step": 25416 + }, + { + "epoch": 0.5958145068185681, + "grad_norm": 0.2774687707424164, + "learning_rate": 7.037887865396348e-05, + "loss": 0.0587, + "step": 25417 + }, + { + "epoch": 0.5958379483933731, + "grad_norm": 0.09114957600831985, + "learning_rate": 7.037184315153032e-05, + "loss": 0.0177, + "step": 25418 + }, + { + "epoch": 0.5958613899681781, + "grad_norm": 0.47920796275138855, + "learning_rate": 7.036480780986036e-05, + "loss": 0.0933, + "step": 25419 + }, + { + "epoch": 0.5958848315429831, + "grad_norm": 0.3144846558570862, + "learning_rate": 7.035777262899181e-05, + "loss": 0.0467, + "step": 25420 + }, + { + "epoch": 0.595908273117788, + "grad_norm": 0.8634315729141235, + "learning_rate": 7.035073760896279e-05, + "loss": 0.1328, + "step": 25421 + }, + { + "epoch": 0.5959317146925931, + "grad_norm": 0.7723269462585449, + "learning_rate": 7.034370274981152e-05, + "loss": 0.1066, + "step": 25422 + }, + { + "epoch": 0.595955156267398, + "grad_norm": 0.25775444507598877, + "learning_rate": 7.033666805157616e-05, + "loss": 0.0413, + "step": 25423 + }, + { + "epoch": 0.5959785978422031, + "grad_norm": 0.20925115048885345, + "learning_rate": 7.032963351429484e-05, + "loss": 0.0396, + "step": 25424 + }, + { + "epoch": 0.596002039417008, + "grad_norm": 0.5163790583610535, + "learning_rate": 7.032259913800579e-05, + "loss": 0.093, + "step": 25425 + }, + { + "epoch": 0.5960254809918131, + "grad_norm": 0.2505832314491272, + "learning_rate": 7.031556492274711e-05, + "loss": 0.0437, + "step": 25426 + }, + { + "epoch": 0.596048922566618, + "grad_norm": 0.3602164685726166, + "learning_rate": 7.030853086855706e-05, + "loss": 0.0613, + "step": 25427 + }, + { + "epoch": 0.5960723641414231, + "grad_norm": 0.40689608454704285, + "learning_rate": 7.030149697547373e-05, + "loss": 0.0805, + "step": 25428 + }, + { + "epoch": 0.596095805716228, + "grad_norm": 0.5960780382156372, + "learning_rate": 7.029446324353533e-05, + "loss": 0.0777, + "step": 25429 + }, + { + "epoch": 0.596119247291033, + "grad_norm": 0.4827140271663666, + "learning_rate": 7.028742967278001e-05, + "loss": 0.1142, + "step": 25430 + }, + { + "epoch": 0.596142688865838, + "grad_norm": 0.4354664981365204, + "learning_rate": 7.02803962632459e-05, + "loss": 0.0988, + "step": 25431 + }, + { + "epoch": 0.596166130440643, + "grad_norm": 0.5785381197929382, + "learning_rate": 7.027336301497123e-05, + "loss": 0.7123, + "step": 25432 + }, + { + "epoch": 0.596189572015448, + "grad_norm": 0.6273946166038513, + "learning_rate": 7.026632992799411e-05, + "loss": 0.1768, + "step": 25433 + }, + { + "epoch": 0.596213013590253, + "grad_norm": 0.3641176223754883, + "learning_rate": 7.025929700235272e-05, + "loss": 0.0895, + "step": 25434 + }, + { + "epoch": 0.596236455165058, + "grad_norm": 0.12467993050813675, + "learning_rate": 7.025226423808525e-05, + "loss": 0.0148, + "step": 25435 + }, + { + "epoch": 0.596259896739863, + "grad_norm": 0.4971535801887512, + "learning_rate": 7.024523163522978e-05, + "loss": 0.0469, + "step": 25436 + }, + { + "epoch": 0.596283338314668, + "grad_norm": 0.12876106798648834, + "learning_rate": 7.023819919382455e-05, + "loss": 0.0144, + "step": 25437 + }, + { + "epoch": 0.596306779889473, + "grad_norm": 0.5848972201347351, + "learning_rate": 7.023116691390765e-05, + "loss": 0.8322, + "step": 25438 + }, + { + "epoch": 0.5963302214642779, + "grad_norm": 0.6269803643226624, + "learning_rate": 7.022413479551733e-05, + "loss": 0.1382, + "step": 25439 + }, + { + "epoch": 0.596353663039083, + "grad_norm": 1.5540343523025513, + "learning_rate": 7.021710283869165e-05, + "loss": 0.0964, + "step": 25440 + }, + { + "epoch": 0.5963771046138879, + "grad_norm": 1.2046030759811401, + "learning_rate": 7.021007104346884e-05, + "loss": 0.1075, + "step": 25441 + }, + { + "epoch": 0.596400546188693, + "grad_norm": 0.5081012845039368, + "learning_rate": 7.020303940988702e-05, + "loss": 0.0731, + "step": 25442 + }, + { + "epoch": 0.5964239877634979, + "grad_norm": 0.3970206081867218, + "learning_rate": 7.019600793798433e-05, + "loss": 0.0654, + "step": 25443 + }, + { + "epoch": 0.596447429338303, + "grad_norm": 0.5034318566322327, + "learning_rate": 7.018897662779895e-05, + "loss": 0.1329, + "step": 25444 + }, + { + "epoch": 0.5964708709131079, + "grad_norm": 0.1715829074382782, + "learning_rate": 7.018194547936903e-05, + "loss": 0.0338, + "step": 25445 + }, + { + "epoch": 0.596494312487913, + "grad_norm": 0.6957589983940125, + "learning_rate": 7.017491449273268e-05, + "loss": 0.1591, + "step": 25446 + }, + { + "epoch": 0.5965177540627179, + "grad_norm": 0.3265746831893921, + "learning_rate": 7.016788366792812e-05, + "loss": 0.075, + "step": 25447 + }, + { + "epoch": 0.596541195637523, + "grad_norm": 0.13739633560180664, + "learning_rate": 7.016085300499344e-05, + "loss": 0.0266, + "step": 25448 + }, + { + "epoch": 0.5965646372123279, + "grad_norm": 0.2560543715953827, + "learning_rate": 7.015382250396681e-05, + "loss": 0.0537, + "step": 25449 + }, + { + "epoch": 0.5965880787871329, + "grad_norm": 0.19857525825500488, + "learning_rate": 7.014679216488634e-05, + "loss": 0.0167, + "step": 25450 + }, + { + "epoch": 0.5966115203619379, + "grad_norm": 0.44327256083488464, + "learning_rate": 7.013976198779026e-05, + "loss": 0.0846, + "step": 25451 + }, + { + "epoch": 0.5966349619367429, + "grad_norm": 0.5896260142326355, + "learning_rate": 7.013273197271668e-05, + "loss": 0.1016, + "step": 25452 + }, + { + "epoch": 0.5966584035115479, + "grad_norm": 0.5013416409492493, + "learning_rate": 7.012570211970372e-05, + "loss": 0.0617, + "step": 25453 + }, + { + "epoch": 0.5966818450863529, + "grad_norm": 0.33125001192092896, + "learning_rate": 7.011867242878955e-05, + "loss": 0.0613, + "step": 25454 + }, + { + "epoch": 0.5967052866611579, + "grad_norm": 0.17887549102306366, + "learning_rate": 7.011164290001231e-05, + "loss": 0.0329, + "step": 25455 + }, + { + "epoch": 0.5967287282359629, + "grad_norm": 0.3178160786628723, + "learning_rate": 7.010461353341012e-05, + "loss": 0.0428, + "step": 25456 + }, + { + "epoch": 0.5967521698107678, + "grad_norm": 0.4938317835330963, + "learning_rate": 7.009758432902114e-05, + "loss": 0.1521, + "step": 25457 + }, + { + "epoch": 0.5967756113855729, + "grad_norm": 0.4335218071937561, + "learning_rate": 7.009055528688353e-05, + "loss": 0.0674, + "step": 25458 + }, + { + "epoch": 0.5967990529603778, + "grad_norm": 0.4753839671611786, + "learning_rate": 7.008352640703537e-05, + "loss": 0.0644, + "step": 25459 + }, + { + "epoch": 0.5968224945351829, + "grad_norm": 0.5733703374862671, + "learning_rate": 7.007649768951487e-05, + "loss": 0.6508, + "step": 25460 + }, + { + "epoch": 0.5968459361099878, + "grad_norm": 0.13697673380374908, + "learning_rate": 7.006946913436013e-05, + "loss": 0.017, + "step": 25461 + }, + { + "epoch": 0.5968693776847929, + "grad_norm": 0.4393606185913086, + "learning_rate": 7.006244074160926e-05, + "loss": 0.0395, + "step": 25462 + }, + { + "epoch": 0.5968928192595978, + "grad_norm": 0.3103506565093994, + "learning_rate": 7.00554125113004e-05, + "loss": 0.0608, + "step": 25463 + }, + { + "epoch": 0.5969162608344029, + "grad_norm": 0.437386155128479, + "learning_rate": 7.00483844434718e-05, + "loss": 0.0651, + "step": 25464 + }, + { + "epoch": 0.5969397024092078, + "grad_norm": 0.6787919402122498, + "learning_rate": 7.004135653816148e-05, + "loss": 0.5543, + "step": 25465 + }, + { + "epoch": 0.5969631439840128, + "grad_norm": 0.172132670879364, + "learning_rate": 7.003432879540757e-05, + "loss": 0.0153, + "step": 25466 + }, + { + "epoch": 0.5969865855588179, + "grad_norm": 0.25670361518859863, + "learning_rate": 7.002730121524827e-05, + "loss": 0.0324, + "step": 25467 + }, + { + "epoch": 0.5970100271336228, + "grad_norm": 0.4917672872543335, + "learning_rate": 7.002027379772167e-05, + "loss": 0.0952, + "step": 25468 + }, + { + "epoch": 0.5970334687084279, + "grad_norm": 0.49002090096473694, + "learning_rate": 7.00132465428659e-05, + "loss": 0.0793, + "step": 25469 + }, + { + "epoch": 0.5970569102832328, + "grad_norm": 0.11482325941324234, + "learning_rate": 7.00062194507191e-05, + "loss": 0.0106, + "step": 25470 + }, + { + "epoch": 0.5970803518580379, + "grad_norm": 0.5280939936637878, + "learning_rate": 6.99991925213194e-05, + "loss": 0.1562, + "step": 25471 + }, + { + "epoch": 0.5971037934328428, + "grad_norm": 0.12436441332101822, + "learning_rate": 6.999216575470491e-05, + "loss": 0.0197, + "step": 25472 + }, + { + "epoch": 0.5971272350076479, + "grad_norm": 0.46222376823425293, + "learning_rate": 6.99851391509138e-05, + "loss": 0.0918, + "step": 25473 + }, + { + "epoch": 0.5971506765824528, + "grad_norm": 0.47236907482147217, + "learning_rate": 6.997811270998413e-05, + "loss": 0.0823, + "step": 25474 + }, + { + "epoch": 0.5971741181572578, + "grad_norm": 0.2758052945137024, + "learning_rate": 6.99710864319541e-05, + "loss": 0.0104, + "step": 25475 + }, + { + "epoch": 0.5971975597320628, + "grad_norm": 0.3001839518547058, + "learning_rate": 6.996406031686176e-05, + "loss": 0.0529, + "step": 25476 + }, + { + "epoch": 0.5972210013068678, + "grad_norm": 0.16147220134735107, + "learning_rate": 6.995703436474531e-05, + "loss": 0.0282, + "step": 25477 + }, + { + "epoch": 0.5972444428816728, + "grad_norm": 0.6559911966323853, + "learning_rate": 6.99500085756428e-05, + "loss": 0.131, + "step": 25478 + }, + { + "epoch": 0.5972678844564778, + "grad_norm": 0.4328368604183197, + "learning_rate": 6.994298294959242e-05, + "loss": 0.0486, + "step": 25479 + }, + { + "epoch": 0.5972913260312828, + "grad_norm": 0.527616560459137, + "learning_rate": 6.993595748663227e-05, + "loss": 0.0854, + "step": 25480 + }, + { + "epoch": 0.5973147676060878, + "grad_norm": 0.3401888608932495, + "learning_rate": 6.992893218680043e-05, + "loss": 0.0717, + "step": 25481 + }, + { + "epoch": 0.5973382091808928, + "grad_norm": 0.5177536606788635, + "learning_rate": 6.992190705013509e-05, + "loss": 0.1094, + "step": 25482 + }, + { + "epoch": 0.5973616507556978, + "grad_norm": 0.5304074883460999, + "learning_rate": 6.99148820766743e-05, + "loss": 0.089, + "step": 25483 + }, + { + "epoch": 0.5973850923305027, + "grad_norm": 0.6592394709587097, + "learning_rate": 6.99078572664562e-05, + "loss": 0.0688, + "step": 25484 + }, + { + "epoch": 0.5974085339053078, + "grad_norm": 0.5370627045631409, + "learning_rate": 6.990083261951892e-05, + "loss": 0.0952, + "step": 25485 + }, + { + "epoch": 0.5974319754801127, + "grad_norm": 0.3030398190021515, + "learning_rate": 6.989380813590058e-05, + "loss": 0.0569, + "step": 25486 + }, + { + "epoch": 0.5974554170549178, + "grad_norm": 0.5568791627883911, + "learning_rate": 6.988678381563924e-05, + "loss": 0.1747, + "step": 25487 + }, + { + "epoch": 0.5974788586297227, + "grad_norm": 0.44835469126701355, + "learning_rate": 6.987975965877309e-05, + "loss": 0.1132, + "step": 25488 + }, + { + "epoch": 0.5975023002045278, + "grad_norm": 0.12759467959403992, + "learning_rate": 6.987273566534021e-05, + "loss": 0.0231, + "step": 25489 + }, + { + "epoch": 0.5975257417793327, + "grad_norm": 0.3969469666481018, + "learning_rate": 6.986571183537872e-05, + "loss": 0.0282, + "step": 25490 + }, + { + "epoch": 0.5975491833541378, + "grad_norm": 0.36366379261016846, + "learning_rate": 6.985868816892672e-05, + "loss": 0.0429, + "step": 25491 + }, + { + "epoch": 0.5975726249289427, + "grad_norm": 0.4777145981788635, + "learning_rate": 6.985166466602233e-05, + "loss": 0.1301, + "step": 25492 + }, + { + "epoch": 0.5975960665037477, + "grad_norm": 0.33464953303337097, + "learning_rate": 6.984464132670365e-05, + "loss": 0.0442, + "step": 25493 + }, + { + "epoch": 0.5976195080785527, + "grad_norm": 0.34536102414131165, + "learning_rate": 6.983761815100878e-05, + "loss": 0.0627, + "step": 25494 + }, + { + "epoch": 0.5976429496533577, + "grad_norm": 0.30612221360206604, + "learning_rate": 6.983059513897588e-05, + "loss": 0.0493, + "step": 25495 + }, + { + "epoch": 0.5976663912281627, + "grad_norm": 0.40320873260498047, + "learning_rate": 6.982357229064299e-05, + "loss": 0.0621, + "step": 25496 + }, + { + "epoch": 0.5976898328029677, + "grad_norm": 0.19720181822776794, + "learning_rate": 6.981654960604824e-05, + "loss": 0.0437, + "step": 25497 + }, + { + "epoch": 0.5977132743777727, + "grad_norm": 0.6089397072792053, + "learning_rate": 6.980952708522976e-05, + "loss": 0.1007, + "step": 25498 + }, + { + "epoch": 0.5977367159525777, + "grad_norm": 0.2467879056930542, + "learning_rate": 6.980250472822558e-05, + "loss": 0.0345, + "step": 25499 + }, + { + "epoch": 0.5977601575273827, + "grad_norm": 0.2400353103876114, + "learning_rate": 6.979548253507391e-05, + "loss": 0.1822, + "step": 25500 + }, + { + "epoch": 0.5977835991021877, + "grad_norm": 0.3281504213809967, + "learning_rate": 6.978846050581281e-05, + "loss": 0.0352, + "step": 25501 + }, + { + "epoch": 0.5978070406769926, + "grad_norm": 0.10954324156045914, + "learning_rate": 6.978143864048034e-05, + "loss": 0.0247, + "step": 25502 + }, + { + "epoch": 0.5978304822517977, + "grad_norm": 0.46790996193885803, + "learning_rate": 6.977441693911468e-05, + "loss": 0.147, + "step": 25503 + }, + { + "epoch": 0.5978539238266026, + "grad_norm": 0.49202072620391846, + "learning_rate": 6.976739540175383e-05, + "loss": 0.0629, + "step": 25504 + }, + { + "epoch": 0.5978773654014077, + "grad_norm": 0.17736805975437164, + "learning_rate": 6.976037402843599e-05, + "loss": 0.0283, + "step": 25505 + }, + { + "epoch": 0.5979008069762126, + "grad_norm": 0.25610122084617615, + "learning_rate": 6.97533528191992e-05, + "loss": 0.1333, + "step": 25506 + }, + { + "epoch": 0.5979242485510177, + "grad_norm": 0.5761252641677856, + "learning_rate": 6.974633177408153e-05, + "loss": 0.5346, + "step": 25507 + }, + { + "epoch": 0.5979476901258226, + "grad_norm": 0.11772308498620987, + "learning_rate": 6.973931089312116e-05, + "loss": 0.0118, + "step": 25508 + }, + { + "epoch": 0.5979711317006277, + "grad_norm": 0.15620601177215576, + "learning_rate": 6.973229017635609e-05, + "loss": 0.0241, + "step": 25509 + }, + { + "epoch": 0.5979945732754326, + "grad_norm": 0.3788142800331116, + "learning_rate": 6.97252696238245e-05, + "loss": 0.0849, + "step": 25510 + }, + { + "epoch": 0.5980180148502376, + "grad_norm": 0.7327344417572021, + "learning_rate": 6.971824923556442e-05, + "loss": 0.1831, + "step": 25511 + }, + { + "epoch": 0.5980414564250426, + "grad_norm": 0.5732836723327637, + "learning_rate": 6.9711229011614e-05, + "loss": 0.083, + "step": 25512 + }, + { + "epoch": 0.5980648979998476, + "grad_norm": 0.3239598572254181, + "learning_rate": 6.97042089520113e-05, + "loss": 0.0713, + "step": 25513 + }, + { + "epoch": 0.5980883395746526, + "grad_norm": 0.42189696431159973, + "learning_rate": 6.969718905679441e-05, + "loss": 0.067, + "step": 25514 + }, + { + "epoch": 0.5981117811494576, + "grad_norm": 0.2055761218070984, + "learning_rate": 6.969016932600143e-05, + "loss": 0.0311, + "step": 25515 + }, + { + "epoch": 0.5981352227242626, + "grad_norm": 0.6220351457595825, + "learning_rate": 6.968314975967043e-05, + "loss": 0.0871, + "step": 25516 + }, + { + "epoch": 0.5981586642990676, + "grad_norm": 0.7024778127670288, + "learning_rate": 6.967613035783953e-05, + "loss": 0.1344, + "step": 25517 + }, + { + "epoch": 0.5981821058738727, + "grad_norm": 0.40980231761932373, + "learning_rate": 6.96691111205468e-05, + "loss": 0.111, + "step": 25518 + }, + { + "epoch": 0.5982055474486776, + "grad_norm": 0.7245493531227112, + "learning_rate": 6.966209204783031e-05, + "loss": 0.0897, + "step": 25519 + }, + { + "epoch": 0.5982289890234826, + "grad_norm": 0.07684198766946793, + "learning_rate": 6.965507313972818e-05, + "loss": 0.0142, + "step": 25520 + }, + { + "epoch": 0.5982524305982876, + "grad_norm": 0.4270578622817993, + "learning_rate": 6.964805439627846e-05, + "loss": 0.0697, + "step": 25521 + }, + { + "epoch": 0.5982758721730926, + "grad_norm": 0.14445602893829346, + "learning_rate": 6.964103581751924e-05, + "loss": 0.0217, + "step": 25522 + }, + { + "epoch": 0.5982993137478976, + "grad_norm": 0.5199670791625977, + "learning_rate": 6.963401740348861e-05, + "loss": 0.1243, + "step": 25523 + }, + { + "epoch": 0.5983227553227026, + "grad_norm": 0.321361243724823, + "learning_rate": 6.962699915422468e-05, + "loss": 0.0463, + "step": 25524 + }, + { + "epoch": 0.5983461968975076, + "grad_norm": 0.48250845074653625, + "learning_rate": 6.96199810697655e-05, + "loss": 0.0557, + "step": 25525 + }, + { + "epoch": 0.5983696384723126, + "grad_norm": 0.3807079493999481, + "learning_rate": 6.961296315014915e-05, + "loss": 0.0769, + "step": 25526 + }, + { + "epoch": 0.5983930800471176, + "grad_norm": 0.21448315680027008, + "learning_rate": 6.960594539541374e-05, + "loss": 0.0169, + "step": 25527 + }, + { + "epoch": 0.5984165216219226, + "grad_norm": 0.36539146304130554, + "learning_rate": 6.959892780559732e-05, + "loss": 0.0632, + "step": 25528 + }, + { + "epoch": 0.5984399631967275, + "grad_norm": 0.46640121936798096, + "learning_rate": 6.959191038073794e-05, + "loss": 0.7498, + "step": 25529 + }, + { + "epoch": 0.5984634047715326, + "grad_norm": 0.20017693936824799, + "learning_rate": 6.958489312087374e-05, + "loss": 0.0281, + "step": 25530 + }, + { + "epoch": 0.5984868463463375, + "grad_norm": 0.3136792778968811, + "learning_rate": 6.957787602604275e-05, + "loss": 0.0534, + "step": 25531 + }, + { + "epoch": 0.5985102879211426, + "grad_norm": 0.3730877637863159, + "learning_rate": 6.957085909628307e-05, + "loss": 0.0498, + "step": 25532 + }, + { + "epoch": 0.5985337294959475, + "grad_norm": 0.3969966471195221, + "learning_rate": 6.956384233163276e-05, + "loss": 0.0681, + "step": 25533 + }, + { + "epoch": 0.5985571710707526, + "grad_norm": 0.5964147448539734, + "learning_rate": 6.955682573212991e-05, + "loss": 0.6662, + "step": 25534 + }, + { + "epoch": 0.5985806126455575, + "grad_norm": 0.7033746242523193, + "learning_rate": 6.954980929781253e-05, + "loss": 0.1413, + "step": 25535 + }, + { + "epoch": 0.5986040542203626, + "grad_norm": 0.1465989202260971, + "learning_rate": 6.954279302871876e-05, + "loss": 0.0225, + "step": 25536 + }, + { + "epoch": 0.5986274957951675, + "grad_norm": 0.11893251538276672, + "learning_rate": 6.953577692488668e-05, + "loss": 0.0192, + "step": 25537 + }, + { + "epoch": 0.5986509373699725, + "grad_norm": 0.111994668841362, + "learning_rate": 6.952876098635432e-05, + "loss": 0.015, + "step": 25538 + }, + { + "epoch": 0.5986743789447775, + "grad_norm": 0.45083191990852356, + "learning_rate": 6.952174521315975e-05, + "loss": 0.0595, + "step": 25539 + }, + { + "epoch": 0.5986978205195825, + "grad_norm": 0.28342190384864807, + "learning_rate": 6.951472960534107e-05, + "loss": 0.0338, + "step": 25540 + }, + { + "epoch": 0.5987212620943875, + "grad_norm": 0.4198765158653259, + "learning_rate": 6.950771416293632e-05, + "loss": 0.0785, + "step": 25541 + }, + { + "epoch": 0.5987447036691925, + "grad_norm": 0.8575053215026855, + "learning_rate": 6.950069888598355e-05, + "loss": 0.139, + "step": 25542 + }, + { + "epoch": 0.5987681452439975, + "grad_norm": 0.503522515296936, + "learning_rate": 6.949368377452087e-05, + "loss": 0.0751, + "step": 25543 + }, + { + "epoch": 0.5987915868188025, + "grad_norm": 0.43610838055610657, + "learning_rate": 6.948666882858632e-05, + "loss": 0.0942, + "step": 25544 + }, + { + "epoch": 0.5988150283936075, + "grad_norm": 0.3843676745891571, + "learning_rate": 6.947965404821792e-05, + "loss": 0.0688, + "step": 25545 + }, + { + "epoch": 0.5988384699684125, + "grad_norm": 0.18735626339912415, + "learning_rate": 6.947263943345382e-05, + "loss": 0.027, + "step": 25546 + }, + { + "epoch": 0.5988619115432174, + "grad_norm": 0.44490736722946167, + "learning_rate": 6.946562498433201e-05, + "loss": 0.043, + "step": 25547 + }, + { + "epoch": 0.5988853531180225, + "grad_norm": 0.444197416305542, + "learning_rate": 6.945861070089059e-05, + "loss": 0.0873, + "step": 25548 + }, + { + "epoch": 0.5989087946928274, + "grad_norm": 0.40820467472076416, + "learning_rate": 6.945159658316757e-05, + "loss": 0.0536, + "step": 25549 + }, + { + "epoch": 0.5989322362676325, + "grad_norm": 1.0736445188522339, + "learning_rate": 6.94445826312011e-05, + "loss": 0.1153, + "step": 25550 + }, + { + "epoch": 0.5989556778424374, + "grad_norm": 0.20812375843524933, + "learning_rate": 6.943756884502918e-05, + "loss": 0.0127, + "step": 25551 + }, + { + "epoch": 0.5989791194172425, + "grad_norm": 0.3096397817134857, + "learning_rate": 6.943055522468984e-05, + "loss": 0.0659, + "step": 25552 + }, + { + "epoch": 0.5990025609920474, + "grad_norm": 0.33673274517059326, + "learning_rate": 6.942354177022119e-05, + "loss": 0.27, + "step": 25553 + }, + { + "epoch": 0.5990260025668525, + "grad_norm": 0.33141669631004333, + "learning_rate": 6.941652848166125e-05, + "loss": 0.0471, + "step": 25554 + }, + { + "epoch": 0.5990494441416574, + "grad_norm": 0.32199960947036743, + "learning_rate": 6.940951535904809e-05, + "loss": 0.0191, + "step": 25555 + }, + { + "epoch": 0.5990728857164624, + "grad_norm": 0.2559594213962555, + "learning_rate": 6.940250240241978e-05, + "loss": 0.0594, + "step": 25556 + }, + { + "epoch": 0.5990963272912674, + "grad_norm": 0.5691043734550476, + "learning_rate": 6.939548961181433e-05, + "loss": 0.1464, + "step": 25557 + }, + { + "epoch": 0.5991197688660724, + "grad_norm": 0.3586761951446533, + "learning_rate": 6.938847698726981e-05, + "loss": 0.0619, + "step": 25558 + }, + { + "epoch": 0.5991432104408774, + "grad_norm": 0.5871264338493347, + "learning_rate": 6.93814645288243e-05, + "loss": 0.0736, + "step": 25559 + }, + { + "epoch": 0.5991666520156824, + "grad_norm": 0.6008265018463135, + "learning_rate": 6.937445223651578e-05, + "loss": 0.127, + "step": 25560 + }, + { + "epoch": 0.5991900935904874, + "grad_norm": 0.4998217523097992, + "learning_rate": 6.936744011038234e-05, + "loss": 0.1448, + "step": 25561 + }, + { + "epoch": 0.5992135351652924, + "grad_norm": 0.2985686957836151, + "learning_rate": 6.936042815046208e-05, + "loss": 0.067, + "step": 25562 + }, + { + "epoch": 0.5992369767400973, + "grad_norm": 0.3950730860233307, + "learning_rate": 6.935341635679298e-05, + "loss": 0.1084, + "step": 25563 + }, + { + "epoch": 0.5992604183149024, + "grad_norm": 0.2941477298736572, + "learning_rate": 6.93464047294131e-05, + "loss": 0.06, + "step": 25564 + }, + { + "epoch": 0.5992838598897073, + "grad_norm": 0.13305029273033142, + "learning_rate": 6.93393932683605e-05, + "loss": 0.026, + "step": 25565 + }, + { + "epoch": 0.5993073014645124, + "grad_norm": 0.31175369024276733, + "learning_rate": 6.933238197367322e-05, + "loss": 0.0181, + "step": 25566 + }, + { + "epoch": 0.5993307430393173, + "grad_norm": 0.38812974095344543, + "learning_rate": 6.932537084538926e-05, + "loss": 0.0776, + "step": 25567 + }, + { + "epoch": 0.5993541846141224, + "grad_norm": 0.4340515732765198, + "learning_rate": 6.931835988354674e-05, + "loss": 0.1093, + "step": 25568 + }, + { + "epoch": 0.5993776261889274, + "grad_norm": 0.29525312781333923, + "learning_rate": 6.931134908818367e-05, + "loss": 0.0495, + "step": 25569 + }, + { + "epoch": 0.5994010677637324, + "grad_norm": 0.1208394318819046, + "learning_rate": 6.930433845933804e-05, + "loss": 0.0151, + "step": 25570 + }, + { + "epoch": 0.5994245093385374, + "grad_norm": 0.174562007188797, + "learning_rate": 6.929732799704796e-05, + "loss": 0.0157, + "step": 25571 + }, + { + "epoch": 0.5994479509133424, + "grad_norm": 0.45503872632980347, + "learning_rate": 6.929031770135139e-05, + "loss": 0.0426, + "step": 25572 + }, + { + "epoch": 0.5994713924881474, + "grad_norm": 0.7576184272766113, + "learning_rate": 6.928330757228648e-05, + "loss": 0.1126, + "step": 25573 + }, + { + "epoch": 0.5994948340629523, + "grad_norm": 0.12420141696929932, + "learning_rate": 6.927629760989118e-05, + "loss": 0.0109, + "step": 25574 + }, + { + "epoch": 0.5995182756377574, + "grad_norm": 0.9254049062728882, + "learning_rate": 6.926928781420357e-05, + "loss": 0.1357, + "step": 25575 + }, + { + "epoch": 0.5995417172125623, + "grad_norm": 0.5549609661102295, + "learning_rate": 6.926227818526168e-05, + "loss": 0.0776, + "step": 25576 + }, + { + "epoch": 0.5995651587873674, + "grad_norm": 0.3456721603870392, + "learning_rate": 6.92552687231035e-05, + "loss": 0.4473, + "step": 25577 + }, + { + "epoch": 0.5995886003621723, + "grad_norm": 0.39867398142814636, + "learning_rate": 6.924825942776711e-05, + "loss": 0.1168, + "step": 25578 + }, + { + "epoch": 0.5996120419369774, + "grad_norm": 0.3654765486717224, + "learning_rate": 6.924125029929055e-05, + "loss": 0.0595, + "step": 25579 + }, + { + "epoch": 0.5996354835117823, + "grad_norm": 0.7297887802124023, + "learning_rate": 6.92342413377118e-05, + "loss": 0.104, + "step": 25580 + }, + { + "epoch": 0.5996589250865874, + "grad_norm": 0.4335060715675354, + "learning_rate": 6.922723254306894e-05, + "loss": 0.0804, + "step": 25581 + }, + { + "epoch": 0.5996823666613923, + "grad_norm": 0.7069634795188904, + "learning_rate": 6.922022391539998e-05, + "loss": 0.1387, + "step": 25582 + }, + { + "epoch": 0.5997058082361973, + "grad_norm": 0.2689400017261505, + "learning_rate": 6.921321545474294e-05, + "loss": 0.074, + "step": 25583 + }, + { + "epoch": 0.5997292498110023, + "grad_norm": 0.10564091801643372, + "learning_rate": 6.920620716113582e-05, + "loss": 0.0159, + "step": 25584 + }, + { + "epoch": 0.5997526913858073, + "grad_norm": 0.395862340927124, + "learning_rate": 6.919919903461675e-05, + "loss": 0.0552, + "step": 25585 + }, + { + "epoch": 0.5997761329606123, + "grad_norm": 0.21427768468856812, + "learning_rate": 6.919219107522367e-05, + "loss": 0.0423, + "step": 25586 + }, + { + "epoch": 0.5997995745354173, + "grad_norm": 0.3140915036201477, + "learning_rate": 6.918518328299463e-05, + "loss": 0.082, + "step": 25587 + }, + { + "epoch": 0.5998230161102223, + "grad_norm": 0.1889011263847351, + "learning_rate": 6.917817565796764e-05, + "loss": 0.0256, + "step": 25588 + }, + { + "epoch": 0.5998464576850273, + "grad_norm": 0.6933404803276062, + "learning_rate": 6.917116820018075e-05, + "loss": 0.1334, + "step": 25589 + }, + { + "epoch": 0.5998698992598323, + "grad_norm": 0.1850273460149765, + "learning_rate": 6.916416090967197e-05, + "loss": 0.0297, + "step": 25590 + }, + { + "epoch": 0.5998933408346373, + "grad_norm": 0.3679185211658478, + "learning_rate": 6.915715378647933e-05, + "loss": 0.0898, + "step": 25591 + }, + { + "epoch": 0.5999167824094422, + "grad_norm": 0.6609131097793579, + "learning_rate": 6.91501468306408e-05, + "loss": 0.0769, + "step": 25592 + }, + { + "epoch": 0.5999402239842473, + "grad_norm": 0.5192911624908447, + "learning_rate": 6.914314004219449e-05, + "loss": 0.1031, + "step": 25593 + }, + { + "epoch": 0.5999636655590522, + "grad_norm": 3.3754186630249023, + "learning_rate": 6.913613342117834e-05, + "loss": 0.2479, + "step": 25594 + }, + { + "epoch": 0.5999871071338573, + "grad_norm": 0.711341381072998, + "learning_rate": 6.912912696763041e-05, + "loss": 0.1093, + "step": 25595 + }, + { + "epoch": 0.6000105487086622, + "grad_norm": 0.33551207184791565, + "learning_rate": 6.912212068158866e-05, + "loss": 0.0461, + "step": 25596 + }, + { + "epoch": 0.6000339902834673, + "grad_norm": 0.2882371246814728, + "learning_rate": 6.911511456309118e-05, + "loss": 0.0821, + "step": 25597 + }, + { + "epoch": 0.6000574318582722, + "grad_norm": 0.7572864294052124, + "learning_rate": 6.910810861217599e-05, + "loss": 0.1277, + "step": 25598 + }, + { + "epoch": 0.6000808734330773, + "grad_norm": 0.2663973569869995, + "learning_rate": 6.910110282888103e-05, + "loss": 0.0509, + "step": 25599 + }, + { + "epoch": 0.6001043150078822, + "grad_norm": 0.4715081453323364, + "learning_rate": 6.909409721324438e-05, + "loss": 0.0911, + "step": 25600 + }, + { + "epoch": 0.6001277565826872, + "grad_norm": 0.3066955506801605, + "learning_rate": 6.908709176530404e-05, + "loss": 0.071, + "step": 25601 + }, + { + "epoch": 0.6001511981574922, + "grad_norm": 0.6761054992675781, + "learning_rate": 6.908008648509798e-05, + "loss": 0.1194, + "step": 25602 + }, + { + "epoch": 0.6001746397322972, + "grad_norm": 0.5726126432418823, + "learning_rate": 6.907308137266426e-05, + "loss": 0.0696, + "step": 25603 + }, + { + "epoch": 0.6001980813071022, + "grad_norm": 0.3429984748363495, + "learning_rate": 6.906607642804086e-05, + "loss": 0.0921, + "step": 25604 + }, + { + "epoch": 0.6002215228819072, + "grad_norm": 0.17024216055870056, + "learning_rate": 6.905907165126579e-05, + "loss": 0.0257, + "step": 25605 + }, + { + "epoch": 0.6002449644567122, + "grad_norm": 0.09659513086080551, + "learning_rate": 6.905206704237709e-05, + "loss": 0.0092, + "step": 25606 + }, + { + "epoch": 0.6002684060315172, + "grad_norm": 0.28996920585632324, + "learning_rate": 6.904506260141274e-05, + "loss": 0.0544, + "step": 25607 + }, + { + "epoch": 0.6002918476063221, + "grad_norm": 0.24803122878074646, + "learning_rate": 6.90380583284107e-05, + "loss": 0.0379, + "step": 25608 + }, + { + "epoch": 0.6003152891811272, + "grad_norm": 0.5457252860069275, + "learning_rate": 6.903105422340906e-05, + "loss": 0.5854, + "step": 25609 + }, + { + "epoch": 0.6003387307559321, + "grad_norm": 0.5175703167915344, + "learning_rate": 6.90240502864458e-05, + "loss": 0.1128, + "step": 25610 + }, + { + "epoch": 0.6003621723307372, + "grad_norm": 0.3771432638168335, + "learning_rate": 6.901704651755892e-05, + "loss": 0.0689, + "step": 25611 + }, + { + "epoch": 0.6003856139055421, + "grad_norm": 0.3957904577255249, + "learning_rate": 6.901004291678641e-05, + "loss": 0.0509, + "step": 25612 + }, + { + "epoch": 0.6004090554803472, + "grad_norm": 0.23975956439971924, + "learning_rate": 6.900303948416629e-05, + "loss": 0.0449, + "step": 25613 + }, + { + "epoch": 0.6004324970551521, + "grad_norm": 0.2895929217338562, + "learning_rate": 6.899603621973653e-05, + "loss": 0.0552, + "step": 25614 + }, + { + "epoch": 0.6004559386299572, + "grad_norm": 0.2645431458950043, + "learning_rate": 6.898903312353515e-05, + "loss": 0.0652, + "step": 25615 + }, + { + "epoch": 0.6004793802047621, + "grad_norm": 0.12877170741558075, + "learning_rate": 6.898203019560017e-05, + "loss": 0.0214, + "step": 25616 + }, + { + "epoch": 0.6005028217795672, + "grad_norm": 0.397965669631958, + "learning_rate": 6.897502743596955e-05, + "loss": 0.1221, + "step": 25617 + }, + { + "epoch": 0.6005262633543721, + "grad_norm": 0.15980809926986694, + "learning_rate": 6.89680248446813e-05, + "loss": 0.0171, + "step": 25618 + }, + { + "epoch": 0.6005497049291771, + "grad_norm": 0.29683607816696167, + "learning_rate": 6.896102242177343e-05, + "loss": 0.0444, + "step": 25619 + }, + { + "epoch": 0.6005731465039822, + "grad_norm": 0.13726070523262024, + "learning_rate": 6.895402016728393e-05, + "loss": 0.0351, + "step": 25620 + }, + { + "epoch": 0.6005965880787871, + "grad_norm": 0.5110905766487122, + "learning_rate": 6.894701808125075e-05, + "loss": 0.1109, + "step": 25621 + }, + { + "epoch": 0.6006200296535922, + "grad_norm": 0.22365683317184448, + "learning_rate": 6.894001616371193e-05, + "loss": 0.0317, + "step": 25622 + }, + { + "epoch": 0.6006434712283971, + "grad_norm": 0.4650903046131134, + "learning_rate": 6.893301441470548e-05, + "loss": 0.0681, + "step": 25623 + }, + { + "epoch": 0.6006669128032022, + "grad_norm": 0.5046815276145935, + "learning_rate": 6.892601283426938e-05, + "loss": 0.0974, + "step": 25624 + }, + { + "epoch": 0.6006903543780071, + "grad_norm": 0.5814077854156494, + "learning_rate": 6.891901142244159e-05, + "loss": 0.6539, + "step": 25625 + }, + { + "epoch": 0.6007137959528122, + "grad_norm": 0.1616372913122177, + "learning_rate": 6.891201017926012e-05, + "loss": 0.0326, + "step": 25626 + }, + { + "epoch": 0.6007372375276171, + "grad_norm": 0.19394342601299286, + "learning_rate": 6.890500910476294e-05, + "loss": 0.0326, + "step": 25627 + }, + { + "epoch": 0.6007606791024221, + "grad_norm": 0.37057146430015564, + "learning_rate": 6.889800819898808e-05, + "loss": 0.0419, + "step": 25628 + }, + { + "epoch": 0.6007841206772271, + "grad_norm": 0.40239229798316956, + "learning_rate": 6.88910074619735e-05, + "loss": 0.0753, + "step": 25629 + }, + { + "epoch": 0.6008075622520321, + "grad_norm": 0.08198689669370651, + "learning_rate": 6.888400689375718e-05, + "loss": 0.0163, + "step": 25630 + }, + { + "epoch": 0.6008310038268371, + "grad_norm": 0.5202480554580688, + "learning_rate": 6.887700649437711e-05, + "loss": 0.204, + "step": 25631 + }, + { + "epoch": 0.6008544454016421, + "grad_norm": 0.24719572067260742, + "learning_rate": 6.88700062638713e-05, + "loss": 0.0362, + "step": 25632 + }, + { + "epoch": 0.6008778869764471, + "grad_norm": 0.5721733570098877, + "learning_rate": 6.886300620227767e-05, + "loss": 0.8101, + "step": 25633 + }, + { + "epoch": 0.6009013285512521, + "grad_norm": 0.0820956751704216, + "learning_rate": 6.885600630963424e-05, + "loss": 0.0131, + "step": 25634 + }, + { + "epoch": 0.600924770126057, + "grad_norm": 0.5066181421279907, + "learning_rate": 6.884900658597905e-05, + "loss": 0.097, + "step": 25635 + }, + { + "epoch": 0.6009482117008621, + "grad_norm": 0.4290255606174469, + "learning_rate": 6.884200703135e-05, + "loss": 0.0638, + "step": 25636 + }, + { + "epoch": 0.600971653275667, + "grad_norm": 0.32197290658950806, + "learning_rate": 6.883500764578509e-05, + "loss": 0.0888, + "step": 25637 + }, + { + "epoch": 0.6009950948504721, + "grad_norm": 0.546482503414154, + "learning_rate": 6.882800842932232e-05, + "loss": 0.0369, + "step": 25638 + }, + { + "epoch": 0.601018536425277, + "grad_norm": 0.9803746342658997, + "learning_rate": 6.882100938199965e-05, + "loss": 0.0813, + "step": 25639 + }, + { + "epoch": 0.6010419780000821, + "grad_norm": 0.5247982144355774, + "learning_rate": 6.881401050385505e-05, + "loss": 0.0922, + "step": 25640 + }, + { + "epoch": 0.601065419574887, + "grad_norm": 0.7122521996498108, + "learning_rate": 6.88070117949265e-05, + "loss": 0.0866, + "step": 25641 + }, + { + "epoch": 0.6010888611496921, + "grad_norm": 0.3177236020565033, + "learning_rate": 6.8800013255252e-05, + "loss": 0.0418, + "step": 25642 + }, + { + "epoch": 0.601112302724497, + "grad_norm": 0.4000081419944763, + "learning_rate": 6.879301488486949e-05, + "loss": 0.3234, + "step": 25643 + }, + { + "epoch": 0.601135744299302, + "grad_norm": 0.3359471261501312, + "learning_rate": 6.878601668381698e-05, + "loss": 0.0809, + "step": 25644 + }, + { + "epoch": 0.601159185874107, + "grad_norm": 0.5990534424781799, + "learning_rate": 6.877901865213237e-05, + "loss": 0.1317, + "step": 25645 + }, + { + "epoch": 0.601182627448912, + "grad_norm": 0.5485382676124573, + "learning_rate": 6.877202078985374e-05, + "loss": 0.1059, + "step": 25646 + }, + { + "epoch": 0.601206069023717, + "grad_norm": 0.37457969784736633, + "learning_rate": 6.876502309701895e-05, + "loss": 0.0749, + "step": 25647 + }, + { + "epoch": 0.601229510598522, + "grad_norm": 0.4040386974811554, + "learning_rate": 6.875802557366608e-05, + "loss": 0.0583, + "step": 25648 + }, + { + "epoch": 0.601252952173327, + "grad_norm": 0.4123164415359497, + "learning_rate": 6.875102821983301e-05, + "loss": 0.066, + "step": 25649 + }, + { + "epoch": 0.601276393748132, + "grad_norm": 0.6900220513343811, + "learning_rate": 6.874403103555774e-05, + "loss": 0.1485, + "step": 25650 + }, + { + "epoch": 0.601299835322937, + "grad_norm": 0.4886768162250519, + "learning_rate": 6.873703402087825e-05, + "loss": 0.5668, + "step": 25651 + }, + { + "epoch": 0.601323276897742, + "grad_norm": 0.3573405146598816, + "learning_rate": 6.87300371758325e-05, + "loss": 0.0513, + "step": 25652 + }, + { + "epoch": 0.601346718472547, + "grad_norm": 0.4285476803779602, + "learning_rate": 6.872304050045841e-05, + "loss": 0.0757, + "step": 25653 + }, + { + "epoch": 0.601370160047352, + "grad_norm": 0.42709770798683167, + "learning_rate": 6.871604399479402e-05, + "loss": 0.0688, + "step": 25654 + }, + { + "epoch": 0.6013936016221569, + "grad_norm": 0.4470697343349457, + "learning_rate": 6.870904765887725e-05, + "loss": 0.0902, + "step": 25655 + }, + { + "epoch": 0.601417043196962, + "grad_norm": 0.46216779947280884, + "learning_rate": 6.870205149274604e-05, + "loss": 0.5993, + "step": 25656 + }, + { + "epoch": 0.6014404847717669, + "grad_norm": 0.29258185625076294, + "learning_rate": 6.869505549643837e-05, + "loss": 0.0513, + "step": 25657 + }, + { + "epoch": 0.601463926346572, + "grad_norm": 0.6176170706748962, + "learning_rate": 6.868805966999223e-05, + "loss": 0.1084, + "step": 25658 + }, + { + "epoch": 0.6014873679213769, + "grad_norm": 0.11547476053237915, + "learning_rate": 6.868106401344559e-05, + "loss": 0.0219, + "step": 25659 + }, + { + "epoch": 0.601510809496182, + "grad_norm": 0.29008015990257263, + "learning_rate": 6.867406852683633e-05, + "loss": 0.076, + "step": 25660 + }, + { + "epoch": 0.6015342510709869, + "grad_norm": 0.49210324883461, + "learning_rate": 6.866707321020248e-05, + "loss": 0.067, + "step": 25661 + }, + { + "epoch": 0.601557692645792, + "grad_norm": 0.2721143960952759, + "learning_rate": 6.866007806358196e-05, + "loss": 0.0458, + "step": 25662 + }, + { + "epoch": 0.6015811342205969, + "grad_norm": 0.27411961555480957, + "learning_rate": 6.865308308701274e-05, + "loss": 0.0452, + "step": 25663 + }, + { + "epoch": 0.6016045757954019, + "grad_norm": 0.12603791058063507, + "learning_rate": 6.864608828053279e-05, + "loss": 0.0158, + "step": 25664 + }, + { + "epoch": 0.6016280173702069, + "grad_norm": 0.5609967708587646, + "learning_rate": 6.863909364418002e-05, + "loss": 0.1025, + "step": 25665 + }, + { + "epoch": 0.6016514589450119, + "grad_norm": 0.5758389830589294, + "learning_rate": 6.863209917799244e-05, + "loss": 0.7416, + "step": 25666 + }, + { + "epoch": 0.6016749005198169, + "grad_norm": 0.10445394366979599, + "learning_rate": 6.862510488200795e-05, + "loss": 0.0156, + "step": 25667 + }, + { + "epoch": 0.6016983420946219, + "grad_norm": 0.3964124023914337, + "learning_rate": 6.861811075626452e-05, + "loss": 0.0378, + "step": 25668 + }, + { + "epoch": 0.6017217836694269, + "grad_norm": 0.41071540117263794, + "learning_rate": 6.86111168008001e-05, + "loss": 0.081, + "step": 25669 + }, + { + "epoch": 0.6017452252442319, + "grad_norm": 0.3188021779060364, + "learning_rate": 6.860412301565265e-05, + "loss": 0.0585, + "step": 25670 + }, + { + "epoch": 0.601768666819037, + "grad_norm": 0.22386640310287476, + "learning_rate": 6.859712940086011e-05, + "loss": 0.0596, + "step": 25671 + }, + { + "epoch": 0.6017921083938419, + "grad_norm": 0.5028513669967651, + "learning_rate": 6.859013595646044e-05, + "loss": 0.1052, + "step": 25672 + }, + { + "epoch": 0.601815549968647, + "grad_norm": 0.36657679080963135, + "learning_rate": 6.858314268249157e-05, + "loss": 0.0509, + "step": 25673 + }, + { + "epoch": 0.6018389915434519, + "grad_norm": 0.5803035497665405, + "learning_rate": 6.857614957899146e-05, + "loss": 0.1027, + "step": 25674 + }, + { + "epoch": 0.6018624331182569, + "grad_norm": 1.0070606470108032, + "learning_rate": 6.856915664599803e-05, + "loss": 0.048, + "step": 25675 + }, + { + "epoch": 0.6018858746930619, + "grad_norm": 0.3094078302383423, + "learning_rate": 6.856216388354925e-05, + "loss": 0.2778, + "step": 25676 + }, + { + "epoch": 0.6019093162678669, + "grad_norm": 0.19411100447177887, + "learning_rate": 6.855517129168305e-05, + "loss": 0.022, + "step": 25677 + }, + { + "epoch": 0.6019327578426719, + "grad_norm": 0.5735670924186707, + "learning_rate": 6.854817887043736e-05, + "loss": 0.0852, + "step": 25678 + }, + { + "epoch": 0.6019561994174769, + "grad_norm": 0.18064048886299133, + "learning_rate": 6.854118661985017e-05, + "loss": 0.0555, + "step": 25679 + }, + { + "epoch": 0.6019796409922819, + "grad_norm": 0.2655056118965149, + "learning_rate": 6.853419453995937e-05, + "loss": 0.0374, + "step": 25680 + }, + { + "epoch": 0.6020030825670869, + "grad_norm": 0.37464439868927, + "learning_rate": 6.852720263080287e-05, + "loss": 0.0596, + "step": 25681 + }, + { + "epoch": 0.6020265241418918, + "grad_norm": 0.3084242045879364, + "learning_rate": 6.85202108924187e-05, + "loss": 0.0491, + "step": 25682 + }, + { + "epoch": 0.6020499657166969, + "grad_norm": 0.14786411821842194, + "learning_rate": 6.851321932484474e-05, + "loss": 0.0144, + "step": 25683 + }, + { + "epoch": 0.6020734072915018, + "grad_norm": 0.12369675189256668, + "learning_rate": 6.850622792811896e-05, + "loss": 0.0118, + "step": 25684 + }, + { + "epoch": 0.6020968488663069, + "grad_norm": 0.22111265361309052, + "learning_rate": 6.849923670227925e-05, + "loss": 0.0266, + "step": 25685 + }, + { + "epoch": 0.6021202904411118, + "grad_norm": 0.5792317986488342, + "learning_rate": 6.849224564736357e-05, + "loss": 0.0672, + "step": 25686 + }, + { + "epoch": 0.6021437320159169, + "grad_norm": 0.3874890208244324, + "learning_rate": 6.848525476340986e-05, + "loss": 0.0655, + "step": 25687 + }, + { + "epoch": 0.6021671735907218, + "grad_norm": 0.09368466585874557, + "learning_rate": 6.847826405045604e-05, + "loss": 0.0142, + "step": 25688 + }, + { + "epoch": 0.6021906151655269, + "grad_norm": 0.2771117389202118, + "learning_rate": 6.847127350854005e-05, + "loss": 0.0446, + "step": 25689 + }, + { + "epoch": 0.6022140567403318, + "grad_norm": 0.07851795852184296, + "learning_rate": 6.846428313769981e-05, + "loss": 0.0069, + "step": 25690 + }, + { + "epoch": 0.6022374983151368, + "grad_norm": 0.2221153974533081, + "learning_rate": 6.845729293797325e-05, + "loss": 0.0557, + "step": 25691 + }, + { + "epoch": 0.6022609398899418, + "grad_norm": 0.17313091456890106, + "learning_rate": 6.845030290939831e-05, + "loss": 0.0481, + "step": 25692 + }, + { + "epoch": 0.6022843814647468, + "grad_norm": 0.39131131768226624, + "learning_rate": 6.844331305201292e-05, + "loss": 0.0431, + "step": 25693 + }, + { + "epoch": 0.6023078230395518, + "grad_norm": 0.34285905957221985, + "learning_rate": 6.843632336585497e-05, + "loss": 0.0508, + "step": 25694 + }, + { + "epoch": 0.6023312646143568, + "grad_norm": 0.4264238476753235, + "learning_rate": 6.842933385096244e-05, + "loss": 0.3747, + "step": 25695 + }, + { + "epoch": 0.6023547061891618, + "grad_norm": 0.4728371798992157, + "learning_rate": 6.842234450737323e-05, + "loss": 0.082, + "step": 25696 + }, + { + "epoch": 0.6023781477639668, + "grad_norm": 0.4308444857597351, + "learning_rate": 6.841535533512528e-05, + "loss": 0.103, + "step": 25697 + }, + { + "epoch": 0.6024015893387717, + "grad_norm": 0.39548739790916443, + "learning_rate": 6.84083663342565e-05, + "loss": 0.067, + "step": 25698 + }, + { + "epoch": 0.6024250309135768, + "grad_norm": 0.47249773144721985, + "learning_rate": 6.84013775048048e-05, + "loss": 0.1007, + "step": 25699 + }, + { + "epoch": 0.6024484724883817, + "grad_norm": 0.17316709458827972, + "learning_rate": 6.839438884680813e-05, + "loss": 0.017, + "step": 25700 + }, + { + "epoch": 0.6024719140631868, + "grad_norm": 0.4033241271972656, + "learning_rate": 6.838740036030437e-05, + "loss": 0.1004, + "step": 25701 + }, + { + "epoch": 0.6024953556379917, + "grad_norm": 0.6090607643127441, + "learning_rate": 6.838041204533148e-05, + "loss": 0.6082, + "step": 25702 + }, + { + "epoch": 0.6025187972127968, + "grad_norm": 0.27554500102996826, + "learning_rate": 6.837342390192736e-05, + "loss": 0.0183, + "step": 25703 + }, + { + "epoch": 0.6025422387876017, + "grad_norm": 0.746654748916626, + "learning_rate": 6.836643593012993e-05, + "loss": 0.1573, + "step": 25704 + }, + { + "epoch": 0.6025656803624068, + "grad_norm": 0.22374564409255981, + "learning_rate": 6.835944812997712e-05, + "loss": 0.0487, + "step": 25705 + }, + { + "epoch": 0.6025891219372117, + "grad_norm": 0.20187681913375854, + "learning_rate": 6.835246050150679e-05, + "loss": 0.0222, + "step": 25706 + }, + { + "epoch": 0.6026125635120168, + "grad_norm": 0.6038612723350525, + "learning_rate": 6.834547304475691e-05, + "loss": 0.0994, + "step": 25707 + }, + { + "epoch": 0.6026360050868217, + "grad_norm": 0.6132184267044067, + "learning_rate": 6.833848575976543e-05, + "loss": 0.0561, + "step": 25708 + }, + { + "epoch": 0.6026594466616267, + "grad_norm": 0.5417194962501526, + "learning_rate": 6.833149864657021e-05, + "loss": 0.0921, + "step": 25709 + }, + { + "epoch": 0.6026828882364317, + "grad_norm": 0.42334502935409546, + "learning_rate": 6.832451170520912e-05, + "loss": 0.0556, + "step": 25710 + }, + { + "epoch": 0.6027063298112367, + "grad_norm": 0.13920417428016663, + "learning_rate": 6.831752493572017e-05, + "loss": 0.0245, + "step": 25711 + }, + { + "epoch": 0.6027297713860417, + "grad_norm": 0.12118013203144073, + "learning_rate": 6.831053833814121e-05, + "loss": 0.0135, + "step": 25712 + }, + { + "epoch": 0.6027532129608467, + "grad_norm": 0.6221813559532166, + "learning_rate": 6.830355191251015e-05, + "loss": 0.651, + "step": 25713 + }, + { + "epoch": 0.6027766545356517, + "grad_norm": 0.5308594107627869, + "learning_rate": 6.829656565886491e-05, + "loss": 0.1213, + "step": 25714 + }, + { + "epoch": 0.6028000961104567, + "grad_norm": 0.5096186995506287, + "learning_rate": 6.82895795772434e-05, + "loss": 0.5934, + "step": 25715 + }, + { + "epoch": 0.6028235376852616, + "grad_norm": 0.33699220418930054, + "learning_rate": 6.828259366768352e-05, + "loss": 0.0385, + "step": 25716 + }, + { + "epoch": 0.6028469792600667, + "grad_norm": 0.4747380316257477, + "learning_rate": 6.827560793022319e-05, + "loss": 0.0735, + "step": 25717 + }, + { + "epoch": 0.6028704208348716, + "grad_norm": 0.4558665156364441, + "learning_rate": 6.826862236490025e-05, + "loss": 0.1026, + "step": 25718 + }, + { + "epoch": 0.6028938624096767, + "grad_norm": 0.4382794499397278, + "learning_rate": 6.82616369717527e-05, + "loss": 0.0863, + "step": 25719 + }, + { + "epoch": 0.6029173039844816, + "grad_norm": 0.646117091178894, + "learning_rate": 6.82546517508184e-05, + "loss": 0.0744, + "step": 25720 + }, + { + "epoch": 0.6029407455592867, + "grad_norm": 0.30457428097724915, + "learning_rate": 6.824766670213524e-05, + "loss": 0.0593, + "step": 25721 + }, + { + "epoch": 0.6029641871340917, + "grad_norm": 0.4285163879394531, + "learning_rate": 6.824068182574117e-05, + "loss": 0.0844, + "step": 25722 + }, + { + "epoch": 0.6029876287088967, + "grad_norm": 0.594782829284668, + "learning_rate": 6.823369712167402e-05, + "loss": 0.145, + "step": 25723 + }, + { + "epoch": 0.6030110702837017, + "grad_norm": 0.46176677942276, + "learning_rate": 6.822671258997174e-05, + "loss": 0.0895, + "step": 25724 + }, + { + "epoch": 0.6030345118585066, + "grad_norm": 0.41139382123947144, + "learning_rate": 6.82197282306722e-05, + "loss": 0.0633, + "step": 25725 + }, + { + "epoch": 0.6030579534333117, + "grad_norm": 0.2412862330675125, + "learning_rate": 6.82127440438133e-05, + "loss": 0.0578, + "step": 25726 + }, + { + "epoch": 0.6030813950081166, + "grad_norm": 0.3800292909145355, + "learning_rate": 6.820576002943297e-05, + "loss": 0.0724, + "step": 25727 + }, + { + "epoch": 0.6031048365829217, + "grad_norm": 0.3351784944534302, + "learning_rate": 6.819877618756907e-05, + "loss": 0.0365, + "step": 25728 + }, + { + "epoch": 0.6031282781577266, + "grad_norm": 0.8423696756362915, + "learning_rate": 6.819179251825949e-05, + "loss": 0.1491, + "step": 25729 + }, + { + "epoch": 0.6031517197325317, + "grad_norm": 0.5526975393295288, + "learning_rate": 6.818480902154212e-05, + "loss": 0.1171, + "step": 25730 + }, + { + "epoch": 0.6031751613073366, + "grad_norm": 0.473514199256897, + "learning_rate": 6.81778256974549e-05, + "loss": 0.1228, + "step": 25731 + }, + { + "epoch": 0.6031986028821417, + "grad_norm": 0.6788147687911987, + "learning_rate": 6.81708425460357e-05, + "loss": 0.1207, + "step": 25732 + }, + { + "epoch": 0.6032220444569466, + "grad_norm": 0.25329095125198364, + "learning_rate": 6.816385956732238e-05, + "loss": 0.0308, + "step": 25733 + }, + { + "epoch": 0.6032454860317517, + "grad_norm": 0.2808264493942261, + "learning_rate": 6.815687676135287e-05, + "loss": 0.0268, + "step": 25734 + }, + { + "epoch": 0.6032689276065566, + "grad_norm": 0.473285436630249, + "learning_rate": 6.814989412816506e-05, + "loss": 0.0759, + "step": 25735 + }, + { + "epoch": 0.6032923691813616, + "grad_norm": 0.2711651921272278, + "learning_rate": 6.814291166779678e-05, + "loss": 0.0369, + "step": 25736 + }, + { + "epoch": 0.6033158107561666, + "grad_norm": 0.5726067423820496, + "learning_rate": 6.813592938028598e-05, + "loss": 0.5215, + "step": 25737 + }, + { + "epoch": 0.6033392523309716, + "grad_norm": 0.27529376745224, + "learning_rate": 6.812894726567053e-05, + "loss": 0.037, + "step": 25738 + }, + { + "epoch": 0.6033626939057766, + "grad_norm": 0.5805978178977966, + "learning_rate": 6.812196532398829e-05, + "loss": 0.1451, + "step": 25739 + }, + { + "epoch": 0.6033861354805816, + "grad_norm": 0.5449054837226868, + "learning_rate": 6.811498355527717e-05, + "loss": 0.0985, + "step": 25740 + }, + { + "epoch": 0.6034095770553866, + "grad_norm": 0.3262108564376831, + "learning_rate": 6.810800195957505e-05, + "loss": 0.0824, + "step": 25741 + }, + { + "epoch": 0.6034330186301916, + "grad_norm": 0.8303923010826111, + "learning_rate": 6.810102053691978e-05, + "loss": 0.174, + "step": 25742 + }, + { + "epoch": 0.6034564602049965, + "grad_norm": 0.6510986685752869, + "learning_rate": 6.809403928734928e-05, + "loss": 0.0384, + "step": 25743 + }, + { + "epoch": 0.6034799017798016, + "grad_norm": 0.33030375838279724, + "learning_rate": 6.808705821090145e-05, + "loss": 0.0435, + "step": 25744 + }, + { + "epoch": 0.6035033433546065, + "grad_norm": 0.12064597010612488, + "learning_rate": 6.808007730761411e-05, + "loss": 0.0267, + "step": 25745 + }, + { + "epoch": 0.6035267849294116, + "grad_norm": 0.6929631233215332, + "learning_rate": 6.807309657752519e-05, + "loss": 0.1385, + "step": 25746 + }, + { + "epoch": 0.6035502265042165, + "grad_norm": 0.6196721792221069, + "learning_rate": 6.806611602067254e-05, + "loss": 0.0835, + "step": 25747 + }, + { + "epoch": 0.6035736680790216, + "grad_norm": 0.3789442479610443, + "learning_rate": 6.805913563709402e-05, + "loss": 0.0647, + "step": 25748 + }, + { + "epoch": 0.6035971096538265, + "grad_norm": 0.6302769184112549, + "learning_rate": 6.805215542682757e-05, + "loss": 0.0827, + "step": 25749 + }, + { + "epoch": 0.6036205512286316, + "grad_norm": 1.052610993385315, + "learning_rate": 6.8045175389911e-05, + "loss": 0.1616, + "step": 25750 + }, + { + "epoch": 0.6036439928034365, + "grad_norm": 0.3467239737510681, + "learning_rate": 6.80381955263822e-05, + "loss": 0.0542, + "step": 25751 + }, + { + "epoch": 0.6036674343782416, + "grad_norm": 0.640860378742218, + "learning_rate": 6.803121583627907e-05, + "loss": 0.1537, + "step": 25752 + }, + { + "epoch": 0.6036908759530465, + "grad_norm": 0.4791470468044281, + "learning_rate": 6.802423631963946e-05, + "loss": 0.0819, + "step": 25753 + }, + { + "epoch": 0.6037143175278515, + "grad_norm": 0.3253745138645172, + "learning_rate": 6.801725697650119e-05, + "loss": 0.0737, + "step": 25754 + }, + { + "epoch": 0.6037377591026565, + "grad_norm": 0.0755709856748581, + "learning_rate": 6.80102778069022e-05, + "loss": 0.013, + "step": 25755 + }, + { + "epoch": 0.6037612006774615, + "grad_norm": 0.34117624163627625, + "learning_rate": 6.800329881088038e-05, + "loss": 0.0344, + "step": 25756 + }, + { + "epoch": 0.6037846422522665, + "grad_norm": 0.38988760113716125, + "learning_rate": 6.799631998847356e-05, + "loss": 0.0511, + "step": 25757 + }, + { + "epoch": 0.6038080838270715, + "grad_norm": 0.4983859360218048, + "learning_rate": 6.798934133971958e-05, + "loss": 0.0938, + "step": 25758 + }, + { + "epoch": 0.6038315254018765, + "grad_norm": 0.09000337868928909, + "learning_rate": 6.798236286465636e-05, + "loss": 0.011, + "step": 25759 + }, + { + "epoch": 0.6038549669766815, + "grad_norm": 0.55528324842453, + "learning_rate": 6.797538456332174e-05, + "loss": 0.1437, + "step": 25760 + }, + { + "epoch": 0.6038784085514864, + "grad_norm": 0.4625735282897949, + "learning_rate": 6.796840643575355e-05, + "loss": 0.1532, + "step": 25761 + }, + { + "epoch": 0.6039018501262915, + "grad_norm": 0.42693158984184265, + "learning_rate": 6.796142848198973e-05, + "loss": 0.0539, + "step": 25762 + }, + { + "epoch": 0.6039252917010964, + "grad_norm": 0.4443672001361847, + "learning_rate": 6.795445070206807e-05, + "loss": 0.1003, + "step": 25763 + }, + { + "epoch": 0.6039487332759015, + "grad_norm": 0.13120779395103455, + "learning_rate": 6.794747309602647e-05, + "loss": 0.0296, + "step": 25764 + }, + { + "epoch": 0.6039721748507064, + "grad_norm": 0.49263083934783936, + "learning_rate": 6.794049566390279e-05, + "loss": 0.0562, + "step": 25765 + }, + { + "epoch": 0.6039956164255115, + "grad_norm": 0.5823320746421814, + "learning_rate": 6.793351840573483e-05, + "loss": 0.152, + "step": 25766 + }, + { + "epoch": 0.6040190580003164, + "grad_norm": 0.2760554254055023, + "learning_rate": 6.792654132156056e-05, + "loss": 0.0212, + "step": 25767 + }, + { + "epoch": 0.6040424995751215, + "grad_norm": 0.5049606561660767, + "learning_rate": 6.791956441141776e-05, + "loss": 0.6625, + "step": 25768 + }, + { + "epoch": 0.6040659411499264, + "grad_norm": 0.3314017355442047, + "learning_rate": 6.791258767534432e-05, + "loss": 0.0869, + "step": 25769 + }, + { + "epoch": 0.6040893827247314, + "grad_norm": 0.16264423727989197, + "learning_rate": 6.790561111337808e-05, + "loss": 0.0261, + "step": 25770 + }, + { + "epoch": 0.6041128242995364, + "grad_norm": 0.23581932485103607, + "learning_rate": 6.789863472555687e-05, + "loss": 0.0328, + "step": 25771 + }, + { + "epoch": 0.6041362658743414, + "grad_norm": 0.2790631949901581, + "learning_rate": 6.789165851191858e-05, + "loss": 0.035, + "step": 25772 + }, + { + "epoch": 0.6041597074491465, + "grad_norm": 0.20000053942203522, + "learning_rate": 6.788468247250108e-05, + "loss": 0.0248, + "step": 25773 + }, + { + "epoch": 0.6041831490239514, + "grad_norm": 0.274607390165329, + "learning_rate": 6.787770660734216e-05, + "loss": 0.0482, + "step": 25774 + }, + { + "epoch": 0.6042065905987565, + "grad_norm": 0.24552832543849945, + "learning_rate": 6.787073091647974e-05, + "loss": 0.0287, + "step": 25775 + }, + { + "epoch": 0.6042300321735614, + "grad_norm": 0.12353482097387314, + "learning_rate": 6.786375539995161e-05, + "loss": 0.0279, + "step": 25776 + }, + { + "epoch": 0.6042534737483665, + "grad_norm": 0.4128018319606781, + "learning_rate": 6.785678005779567e-05, + "loss": 0.0825, + "step": 25777 + }, + { + "epoch": 0.6042769153231714, + "grad_norm": 0.4424181580543518, + "learning_rate": 6.784980489004975e-05, + "loss": 0.0428, + "step": 25778 + }, + { + "epoch": 0.6043003568979765, + "grad_norm": 0.34988802671432495, + "learning_rate": 6.784282989675164e-05, + "loss": 0.0511, + "step": 25779 + }, + { + "epoch": 0.6043237984727814, + "grad_norm": 0.3315696716308594, + "learning_rate": 6.783585507793929e-05, + "loss": 0.0593, + "step": 25780 + }, + { + "epoch": 0.6043472400475864, + "grad_norm": 0.11317574232816696, + "learning_rate": 6.782888043365045e-05, + "loss": 0.019, + "step": 25781 + }, + { + "epoch": 0.6043706816223914, + "grad_norm": 0.26174336671829224, + "learning_rate": 6.782190596392306e-05, + "loss": 0.0519, + "step": 25782 + }, + { + "epoch": 0.6043941231971964, + "grad_norm": 0.49503830075263977, + "learning_rate": 6.781493166879487e-05, + "loss": 0.076, + "step": 25783 + }, + { + "epoch": 0.6044175647720014, + "grad_norm": 0.6828860640525818, + "learning_rate": 6.78079575483038e-05, + "loss": 0.1023, + "step": 25784 + }, + { + "epoch": 0.6044410063468064, + "grad_norm": 0.24256479740142822, + "learning_rate": 6.780098360248766e-05, + "loss": 0.0111, + "step": 25785 + }, + { + "epoch": 0.6044644479216114, + "grad_norm": 0.5843859910964966, + "learning_rate": 6.779400983138426e-05, + "loss": 0.0779, + "step": 25786 + }, + { + "epoch": 0.6044878894964164, + "grad_norm": 0.45493626594543457, + "learning_rate": 6.778703623503149e-05, + "loss": 0.418, + "step": 25787 + }, + { + "epoch": 0.6045113310712213, + "grad_norm": 0.5941781401634216, + "learning_rate": 6.778006281346717e-05, + "loss": 0.1509, + "step": 25788 + }, + { + "epoch": 0.6045347726460264, + "grad_norm": 0.5379448533058167, + "learning_rate": 6.777308956672912e-05, + "loss": 0.0834, + "step": 25789 + }, + { + "epoch": 0.6045582142208313, + "grad_norm": 0.34077903628349304, + "learning_rate": 6.77661164948552e-05, + "loss": 0.106, + "step": 25790 + }, + { + "epoch": 0.6045816557956364, + "grad_norm": 0.3558430075645447, + "learning_rate": 6.77591435978832e-05, + "loss": 0.0474, + "step": 25791 + }, + { + "epoch": 0.6046050973704413, + "grad_norm": 0.48670583963394165, + "learning_rate": 6.775217087585104e-05, + "loss": 0.1029, + "step": 25792 + }, + { + "epoch": 0.6046285389452464, + "grad_norm": 0.5113636255264282, + "learning_rate": 6.774519832879648e-05, + "loss": 0.1465, + "step": 25793 + }, + { + "epoch": 0.6046519805200513, + "grad_norm": 0.6510811448097229, + "learning_rate": 6.773822595675741e-05, + "loss": 0.1663, + "step": 25794 + }, + { + "epoch": 0.6046754220948564, + "grad_norm": 0.43565163016319275, + "learning_rate": 6.773125375977162e-05, + "loss": 0.0878, + "step": 25795 + }, + { + "epoch": 0.6046988636696613, + "grad_norm": 0.506804347038269, + "learning_rate": 6.772428173787694e-05, + "loss": 0.0961, + "step": 25796 + }, + { + "epoch": 0.6047223052444664, + "grad_norm": 0.07574092596769333, + "learning_rate": 6.771730989111124e-05, + "loss": 0.0104, + "step": 25797 + }, + { + "epoch": 0.6047457468192713, + "grad_norm": 0.34017571806907654, + "learning_rate": 6.771033821951232e-05, + "loss": 0.0411, + "step": 25798 + }, + { + "epoch": 0.6047691883940763, + "grad_norm": 0.1802942007780075, + "learning_rate": 6.7703366723118e-05, + "loss": 0.0288, + "step": 25799 + }, + { + "epoch": 0.6047926299688813, + "grad_norm": 0.21898990869522095, + "learning_rate": 6.769639540196613e-05, + "loss": 0.0515, + "step": 25800 + }, + { + "epoch": 0.6048160715436863, + "grad_norm": 0.3167240619659424, + "learning_rate": 6.768942425609454e-05, + "loss": 0.1488, + "step": 25801 + }, + { + "epoch": 0.6048395131184913, + "grad_norm": 0.32559460401535034, + "learning_rate": 6.768245328554101e-05, + "loss": 0.0426, + "step": 25802 + }, + { + "epoch": 0.6048629546932963, + "grad_norm": 0.32630905508995056, + "learning_rate": 6.767548249034339e-05, + "loss": 0.0219, + "step": 25803 + }, + { + "epoch": 0.6048863962681013, + "grad_norm": 0.18986767530441284, + "learning_rate": 6.766851187053953e-05, + "loss": 0.0277, + "step": 25804 + }, + { + "epoch": 0.6049098378429063, + "grad_norm": 0.2762645184993744, + "learning_rate": 6.766154142616725e-05, + "loss": 0.0634, + "step": 25805 + }, + { + "epoch": 0.6049332794177112, + "grad_norm": 0.609143853187561, + "learning_rate": 6.765457115726432e-05, + "loss": 0.5066, + "step": 25806 + }, + { + "epoch": 0.6049567209925163, + "grad_norm": 0.5588138699531555, + "learning_rate": 6.764760106386864e-05, + "loss": 0.1104, + "step": 25807 + }, + { + "epoch": 0.6049801625673212, + "grad_norm": 0.6477574706077576, + "learning_rate": 6.764063114601797e-05, + "loss": 0.1255, + "step": 25808 + }, + { + "epoch": 0.6050036041421263, + "grad_norm": 0.837151288986206, + "learning_rate": 6.763366140375014e-05, + "loss": 0.2525, + "step": 25809 + }, + { + "epoch": 0.6050270457169312, + "grad_norm": 0.35372257232666016, + "learning_rate": 6.762669183710298e-05, + "loss": 0.0435, + "step": 25810 + }, + { + "epoch": 0.6050504872917363, + "grad_norm": 0.05645241588354111, + "learning_rate": 6.76197224461143e-05, + "loss": 0.0102, + "step": 25811 + }, + { + "epoch": 0.6050739288665412, + "grad_norm": 0.39315515756607056, + "learning_rate": 6.76127532308219e-05, + "loss": 0.0548, + "step": 25812 + }, + { + "epoch": 0.6050973704413463, + "grad_norm": 0.4193496108055115, + "learning_rate": 6.760578419126362e-05, + "loss": 0.1118, + "step": 25813 + }, + { + "epoch": 0.6051208120161512, + "grad_norm": 0.13082437217235565, + "learning_rate": 6.759881532747727e-05, + "loss": 0.0217, + "step": 25814 + }, + { + "epoch": 0.6051442535909562, + "grad_norm": 0.7487642765045166, + "learning_rate": 6.759184663950065e-05, + "loss": 0.1481, + "step": 25815 + }, + { + "epoch": 0.6051676951657612, + "grad_norm": 0.5611490607261658, + "learning_rate": 6.758487812737157e-05, + "loss": 0.1112, + "step": 25816 + }, + { + "epoch": 0.6051911367405662, + "grad_norm": 0.40409189462661743, + "learning_rate": 6.757790979112789e-05, + "loss": 0.1156, + "step": 25817 + }, + { + "epoch": 0.6052145783153712, + "grad_norm": 0.5407857298851013, + "learning_rate": 6.757094163080736e-05, + "loss": 0.1349, + "step": 25818 + }, + { + "epoch": 0.6052380198901762, + "grad_norm": 0.17034320533275604, + "learning_rate": 6.75639736464478e-05, + "loss": 0.0317, + "step": 25819 + }, + { + "epoch": 0.6052614614649812, + "grad_norm": 0.3051338791847229, + "learning_rate": 6.755700583808706e-05, + "loss": 0.0824, + "step": 25820 + }, + { + "epoch": 0.6052849030397862, + "grad_norm": 0.38345518708229065, + "learning_rate": 6.755003820576288e-05, + "loss": 0.0843, + "step": 25821 + }, + { + "epoch": 0.6053083446145912, + "grad_norm": 0.574579656124115, + "learning_rate": 6.754307074951315e-05, + "loss": 0.5666, + "step": 25822 + }, + { + "epoch": 0.6053317861893962, + "grad_norm": 0.37167152762413025, + "learning_rate": 6.753610346937562e-05, + "loss": 0.0785, + "step": 25823 + }, + { + "epoch": 0.6053552277642013, + "grad_norm": 0.355490505695343, + "learning_rate": 6.75291363653881e-05, + "loss": 0.0956, + "step": 25824 + }, + { + "epoch": 0.6053786693390062, + "grad_norm": 0.8375073075294495, + "learning_rate": 6.75221694375884e-05, + "loss": 0.6816, + "step": 25825 + }, + { + "epoch": 0.6054021109138112, + "grad_norm": 0.7856320738792419, + "learning_rate": 6.751520268601433e-05, + "loss": 0.1282, + "step": 25826 + }, + { + "epoch": 0.6054255524886162, + "grad_norm": 0.36297571659088135, + "learning_rate": 6.750823611070364e-05, + "loss": 0.064, + "step": 25827 + }, + { + "epoch": 0.6054489940634212, + "grad_norm": 0.2614783048629761, + "learning_rate": 6.75012697116942e-05, + "loss": 0.0239, + "step": 25828 + }, + { + "epoch": 0.6054724356382262, + "grad_norm": 0.8463103175163269, + "learning_rate": 6.749430348902379e-05, + "loss": 0.4347, + "step": 25829 + }, + { + "epoch": 0.6054958772130312, + "grad_norm": 0.3798221945762634, + "learning_rate": 6.748733744273021e-05, + "loss": 0.0466, + "step": 25830 + }, + { + "epoch": 0.6055193187878362, + "grad_norm": 0.6773107647895813, + "learning_rate": 6.748037157285123e-05, + "loss": 0.0539, + "step": 25831 + }, + { + "epoch": 0.6055427603626412, + "grad_norm": 0.48123377561569214, + "learning_rate": 6.74734058794247e-05, + "loss": 0.1445, + "step": 25832 + }, + { + "epoch": 0.6055662019374461, + "grad_norm": 0.5439791679382324, + "learning_rate": 6.746644036248838e-05, + "loss": 0.1179, + "step": 25833 + }, + { + "epoch": 0.6055896435122512, + "grad_norm": 0.25635388493537903, + "learning_rate": 6.745947502208004e-05, + "loss": 0.0362, + "step": 25834 + }, + { + "epoch": 0.6056130850870561, + "grad_norm": 0.3221745789051056, + "learning_rate": 6.745250985823753e-05, + "loss": 0.07, + "step": 25835 + }, + { + "epoch": 0.6056365266618612, + "grad_norm": 0.7877113819122314, + "learning_rate": 6.744554487099862e-05, + "loss": 0.0837, + "step": 25836 + }, + { + "epoch": 0.6056599682366661, + "grad_norm": 0.4068511724472046, + "learning_rate": 6.743858006040107e-05, + "loss": 0.1112, + "step": 25837 + }, + { + "epoch": 0.6056834098114712, + "grad_norm": 0.6926597356796265, + "learning_rate": 6.743161542648274e-05, + "loss": 0.1626, + "step": 25838 + }, + { + "epoch": 0.6057068513862761, + "grad_norm": 0.14470350742340088, + "learning_rate": 6.742465096928132e-05, + "loss": 0.0214, + "step": 25839 + }, + { + "epoch": 0.6057302929610812, + "grad_norm": 0.9532350301742554, + "learning_rate": 6.74176866888347e-05, + "loss": 0.1628, + "step": 25840 + }, + { + "epoch": 0.6057537345358861, + "grad_norm": 0.2606896460056305, + "learning_rate": 6.741072258518062e-05, + "loss": 0.0474, + "step": 25841 + }, + { + "epoch": 0.6057771761106912, + "grad_norm": 0.16879726946353912, + "learning_rate": 6.740375865835689e-05, + "loss": 0.0285, + "step": 25842 + }, + { + "epoch": 0.6058006176854961, + "grad_norm": 0.7435389757156372, + "learning_rate": 6.739679490840127e-05, + "loss": 0.0832, + "step": 25843 + }, + { + "epoch": 0.6058240592603011, + "grad_norm": 0.030208447948098183, + "learning_rate": 6.738983133535155e-05, + "loss": 0.0037, + "step": 25844 + }, + { + "epoch": 0.6058475008351061, + "grad_norm": 0.29395848512649536, + "learning_rate": 6.738286793924554e-05, + "loss": 0.0422, + "step": 25845 + }, + { + "epoch": 0.6058709424099111, + "grad_norm": 0.2767643630504608, + "learning_rate": 6.7375904720121e-05, + "loss": 0.0725, + "step": 25846 + }, + { + "epoch": 0.6058943839847161, + "grad_norm": 0.35106974840164185, + "learning_rate": 6.736894167801571e-05, + "loss": 0.067, + "step": 25847 + }, + { + "epoch": 0.6059178255595211, + "grad_norm": 0.3513224720954895, + "learning_rate": 6.736197881296748e-05, + "loss": 0.0486, + "step": 25848 + }, + { + "epoch": 0.6059412671343261, + "grad_norm": 0.6319400072097778, + "learning_rate": 6.735501612501407e-05, + "loss": 0.6253, + "step": 25849 + }, + { + "epoch": 0.6059647087091311, + "grad_norm": 0.1057383120059967, + "learning_rate": 6.734805361419322e-05, + "loss": 0.0209, + "step": 25850 + }, + { + "epoch": 0.605988150283936, + "grad_norm": 0.6445247530937195, + "learning_rate": 6.73410912805428e-05, + "loss": 0.1226, + "step": 25851 + }, + { + "epoch": 0.6060115918587411, + "grad_norm": 0.3287045955657959, + "learning_rate": 6.733412912410048e-05, + "loss": 0.0697, + "step": 25852 + }, + { + "epoch": 0.606035033433546, + "grad_norm": 0.22051867842674255, + "learning_rate": 6.732716714490413e-05, + "loss": 0.0433, + "step": 25853 + }, + { + "epoch": 0.6060584750083511, + "grad_norm": 0.30672985315322876, + "learning_rate": 6.732020534299148e-05, + "loss": 0.056, + "step": 25854 + }, + { + "epoch": 0.606081916583156, + "grad_norm": 0.4442969858646393, + "learning_rate": 6.731324371840033e-05, + "loss": 0.0954, + "step": 25855 + }, + { + "epoch": 0.6061053581579611, + "grad_norm": 0.2849842607975006, + "learning_rate": 6.730628227116844e-05, + "loss": 0.0476, + "step": 25856 + }, + { + "epoch": 0.606128799732766, + "grad_norm": 0.250514417886734, + "learning_rate": 6.729932100133355e-05, + "loss": 0.0772, + "step": 25857 + }, + { + "epoch": 0.6061522413075711, + "grad_norm": 0.4909043610095978, + "learning_rate": 6.72923599089335e-05, + "loss": 0.1008, + "step": 25858 + }, + { + "epoch": 0.606175682882376, + "grad_norm": 0.3077731728553772, + "learning_rate": 6.7285398994006e-05, + "loss": 0.0405, + "step": 25859 + }, + { + "epoch": 0.606199124457181, + "grad_norm": 0.37024182081222534, + "learning_rate": 6.727843825658886e-05, + "loss": 0.0882, + "step": 25860 + }, + { + "epoch": 0.606222566031986, + "grad_norm": 0.373647540807724, + "learning_rate": 6.727147769671984e-05, + "loss": 0.087, + "step": 25861 + }, + { + "epoch": 0.606246007606791, + "grad_norm": 0.4084595739841461, + "learning_rate": 6.726451731443669e-05, + "loss": 0.0914, + "step": 25862 + }, + { + "epoch": 0.606269449181596, + "grad_norm": 0.28185251355171204, + "learning_rate": 6.72575571097772e-05, + "loss": 0.0524, + "step": 25863 + }, + { + "epoch": 0.606292890756401, + "grad_norm": 0.5806524753570557, + "learning_rate": 6.725059708277907e-05, + "loss": 0.0835, + "step": 25864 + }, + { + "epoch": 0.606316332331206, + "grad_norm": 0.4261913597583771, + "learning_rate": 6.724363723348018e-05, + "loss": 0.038, + "step": 25865 + }, + { + "epoch": 0.606339773906011, + "grad_norm": 0.21063949167728424, + "learning_rate": 6.723667756191821e-05, + "loss": 0.0483, + "step": 25866 + }, + { + "epoch": 0.606363215480816, + "grad_norm": 0.7676092982292175, + "learning_rate": 6.722971806813097e-05, + "loss": 0.1266, + "step": 25867 + }, + { + "epoch": 0.606386657055621, + "grad_norm": 0.3837578296661377, + "learning_rate": 6.722275875215621e-05, + "loss": 0.0512, + "step": 25868 + }, + { + "epoch": 0.6064100986304259, + "grad_norm": 0.3769058883190155, + "learning_rate": 6.721579961403166e-05, + "loss": 0.0596, + "step": 25869 + }, + { + "epoch": 0.606433540205231, + "grad_norm": 0.3462214767932892, + "learning_rate": 6.720884065379511e-05, + "loss": 0.0467, + "step": 25870 + }, + { + "epoch": 0.6064569817800359, + "grad_norm": 0.361853688955307, + "learning_rate": 6.720188187148433e-05, + "loss": 0.0297, + "step": 25871 + }, + { + "epoch": 0.606480423354841, + "grad_norm": 0.1797589659690857, + "learning_rate": 6.719492326713703e-05, + "loss": 0.0242, + "step": 25872 + }, + { + "epoch": 0.6065038649296459, + "grad_norm": 0.46639537811279297, + "learning_rate": 6.718796484079102e-05, + "loss": 0.0903, + "step": 25873 + }, + { + "epoch": 0.606527306504451, + "grad_norm": 0.31768789887428284, + "learning_rate": 6.718100659248402e-05, + "loss": 0.0658, + "step": 25874 + }, + { + "epoch": 0.606550748079256, + "grad_norm": 0.36095336079597473, + "learning_rate": 6.71740485222538e-05, + "loss": 0.0725, + "step": 25875 + }, + { + "epoch": 0.606574189654061, + "grad_norm": 0.4941612482070923, + "learning_rate": 6.71670906301381e-05, + "loss": 0.0994, + "step": 25876 + }, + { + "epoch": 0.606597631228866, + "grad_norm": 0.5050368905067444, + "learning_rate": 6.716013291617472e-05, + "loss": 0.0758, + "step": 25877 + }, + { + "epoch": 0.606621072803671, + "grad_norm": 0.5728288888931274, + "learning_rate": 6.715317538040138e-05, + "loss": 0.0766, + "step": 25878 + }, + { + "epoch": 0.606644514378476, + "grad_norm": 0.4472147226333618, + "learning_rate": 6.714621802285583e-05, + "loss": 0.0559, + "step": 25879 + }, + { + "epoch": 0.6066679559532809, + "grad_norm": 0.17714901268482208, + "learning_rate": 6.713926084357582e-05, + "loss": 0.0458, + "step": 25880 + }, + { + "epoch": 0.606691397528086, + "grad_norm": 0.6144035458564758, + "learning_rate": 6.713230384259912e-05, + "loss": 0.0931, + "step": 25881 + }, + { + "epoch": 0.6067148391028909, + "grad_norm": 0.22668452560901642, + "learning_rate": 6.712534701996344e-05, + "loss": 0.0218, + "step": 25882 + }, + { + "epoch": 0.606738280677696, + "grad_norm": 0.38772568106651306, + "learning_rate": 6.711839037570658e-05, + "loss": 0.0451, + "step": 25883 + }, + { + "epoch": 0.6067617222525009, + "grad_norm": 0.5086352229118347, + "learning_rate": 6.711143390986625e-05, + "loss": 0.0695, + "step": 25884 + }, + { + "epoch": 0.606785163827306, + "grad_norm": 0.5095759034156799, + "learning_rate": 6.710447762248018e-05, + "loss": 0.0613, + "step": 25885 + }, + { + "epoch": 0.6068086054021109, + "grad_norm": 0.22249113023281097, + "learning_rate": 6.709752151358617e-05, + "loss": 0.0297, + "step": 25886 + }, + { + "epoch": 0.606832046976916, + "grad_norm": 0.5927579402923584, + "learning_rate": 6.709056558322192e-05, + "loss": 0.0776, + "step": 25887 + }, + { + "epoch": 0.6068554885517209, + "grad_norm": 0.32552969455718994, + "learning_rate": 6.708360983142515e-05, + "loss": 0.0627, + "step": 25888 + }, + { + "epoch": 0.6068789301265259, + "grad_norm": 0.3139696419239044, + "learning_rate": 6.707665425823366e-05, + "loss": 0.0556, + "step": 25889 + }, + { + "epoch": 0.6069023717013309, + "grad_norm": 0.468821257352829, + "learning_rate": 6.706969886368519e-05, + "loss": 0.0545, + "step": 25890 + }, + { + "epoch": 0.6069258132761359, + "grad_norm": 0.4862147569656372, + "learning_rate": 6.706274364781747e-05, + "loss": 0.1107, + "step": 25891 + }, + { + "epoch": 0.6069492548509409, + "grad_norm": 0.5849016904830933, + "learning_rate": 6.70557886106682e-05, + "loss": 0.1029, + "step": 25892 + }, + { + "epoch": 0.6069726964257459, + "grad_norm": 0.36328575015068054, + "learning_rate": 6.704883375227517e-05, + "loss": 0.0705, + "step": 25893 + }, + { + "epoch": 0.6069961380005509, + "grad_norm": 0.9306491613388062, + "learning_rate": 6.70418790726761e-05, + "loss": 0.118, + "step": 25894 + }, + { + "epoch": 0.6070195795753559, + "grad_norm": 0.3988306224346161, + "learning_rate": 6.70349245719087e-05, + "loss": 0.0522, + "step": 25895 + }, + { + "epoch": 0.6070430211501608, + "grad_norm": 0.5017435550689697, + "learning_rate": 6.702797025001073e-05, + "loss": 0.0421, + "step": 25896 + }, + { + "epoch": 0.6070664627249659, + "grad_norm": 0.5999549627304077, + "learning_rate": 6.702101610701991e-05, + "loss": 0.7721, + "step": 25897 + }, + { + "epoch": 0.6070899042997708, + "grad_norm": 0.5147703289985657, + "learning_rate": 6.701406214297401e-05, + "loss": 0.0581, + "step": 25898 + }, + { + "epoch": 0.6071133458745759, + "grad_norm": 0.7382416725158691, + "learning_rate": 6.700710835791073e-05, + "loss": 0.0822, + "step": 25899 + }, + { + "epoch": 0.6071367874493808, + "grad_norm": 0.07449036091566086, + "learning_rate": 6.700015475186776e-05, + "loss": 0.0136, + "step": 25900 + }, + { + "epoch": 0.6071602290241859, + "grad_norm": 0.6036822199821472, + "learning_rate": 6.699320132488291e-05, + "loss": 0.1281, + "step": 25901 + }, + { + "epoch": 0.6071836705989908, + "grad_norm": 0.370412677526474, + "learning_rate": 6.69862480769939e-05, + "loss": 0.0321, + "step": 25902 + }, + { + "epoch": 0.6072071121737959, + "grad_norm": 0.2755577862262726, + "learning_rate": 6.697929500823843e-05, + "loss": 0.0464, + "step": 25903 + }, + { + "epoch": 0.6072305537486008, + "grad_norm": 0.41887155175209045, + "learning_rate": 6.697234211865422e-05, + "loss": 0.0523, + "step": 25904 + }, + { + "epoch": 0.6072539953234058, + "grad_norm": 0.32416367530822754, + "learning_rate": 6.696538940827901e-05, + "loss": 0.0318, + "step": 25905 + }, + { + "epoch": 0.6072774368982108, + "grad_norm": 0.41442668437957764, + "learning_rate": 6.695843687715055e-05, + "loss": 0.0759, + "step": 25906 + }, + { + "epoch": 0.6073008784730158, + "grad_norm": 0.47160714864730835, + "learning_rate": 6.69514845253065e-05, + "loss": 0.1318, + "step": 25907 + }, + { + "epoch": 0.6073243200478208, + "grad_norm": 0.11970765143632889, + "learning_rate": 6.694453235278466e-05, + "loss": 0.0219, + "step": 25908 + }, + { + "epoch": 0.6073477616226258, + "grad_norm": 0.21730320155620575, + "learning_rate": 6.693758035962272e-05, + "loss": 0.0274, + "step": 25909 + }, + { + "epoch": 0.6073712031974308, + "grad_norm": 0.29900047183036804, + "learning_rate": 6.693062854585837e-05, + "loss": 0.041, + "step": 25910 + }, + { + "epoch": 0.6073946447722358, + "grad_norm": 0.7322565913200378, + "learning_rate": 6.692367691152938e-05, + "loss": 0.5968, + "step": 25911 + }, + { + "epoch": 0.6074180863470408, + "grad_norm": 0.12903347611427307, + "learning_rate": 6.69167254566734e-05, + "loss": 0.0119, + "step": 25912 + }, + { + "epoch": 0.6074415279218458, + "grad_norm": 0.203377828001976, + "learning_rate": 6.690977418132825e-05, + "loss": 0.0299, + "step": 25913 + }, + { + "epoch": 0.6074649694966507, + "grad_norm": 0.06749095022678375, + "learning_rate": 6.690282308553158e-05, + "loss": 0.0044, + "step": 25914 + }, + { + "epoch": 0.6074884110714558, + "grad_norm": 0.2556385397911072, + "learning_rate": 6.689587216932114e-05, + "loss": 0.0561, + "step": 25915 + }, + { + "epoch": 0.6075118526462607, + "grad_norm": 0.30744311213493347, + "learning_rate": 6.688892143273462e-05, + "loss": 0.0488, + "step": 25916 + }, + { + "epoch": 0.6075352942210658, + "grad_norm": 0.1295100450515747, + "learning_rate": 6.688197087580972e-05, + "loss": 0.0277, + "step": 25917 + }, + { + "epoch": 0.6075587357958707, + "grad_norm": 0.4388842284679413, + "learning_rate": 6.68750204985842e-05, + "loss": 0.089, + "step": 25918 + }, + { + "epoch": 0.6075821773706758, + "grad_norm": 0.7307022213935852, + "learning_rate": 6.686807030109577e-05, + "loss": 0.1866, + "step": 25919 + }, + { + "epoch": 0.6076056189454807, + "grad_norm": 0.6061185002326965, + "learning_rate": 6.686112028338207e-05, + "loss": 0.0888, + "step": 25920 + }, + { + "epoch": 0.6076290605202858, + "grad_norm": 0.4019118547439575, + "learning_rate": 6.685417044548091e-05, + "loss": 0.0516, + "step": 25921 + }, + { + "epoch": 0.6076525020950907, + "grad_norm": 0.6681090593338013, + "learning_rate": 6.684722078742994e-05, + "loss": 0.6295, + "step": 25922 + }, + { + "epoch": 0.6076759436698957, + "grad_norm": 0.3138735890388489, + "learning_rate": 6.684027130926687e-05, + "loss": 0.0226, + "step": 25923 + }, + { + "epoch": 0.6076993852447007, + "grad_norm": 0.1749674677848816, + "learning_rate": 6.683332201102943e-05, + "loss": 0.0172, + "step": 25924 + }, + { + "epoch": 0.6077228268195057, + "grad_norm": 0.44399523735046387, + "learning_rate": 6.682637289275529e-05, + "loss": 0.4634, + "step": 25925 + }, + { + "epoch": 0.6077462683943108, + "grad_norm": 0.7219593524932861, + "learning_rate": 6.681942395448222e-05, + "loss": 0.1663, + "step": 25926 + }, + { + "epoch": 0.6077697099691157, + "grad_norm": 0.564159095287323, + "learning_rate": 6.681247519624785e-05, + "loss": 0.0697, + "step": 25927 + }, + { + "epoch": 0.6077931515439208, + "grad_norm": 0.5580727458000183, + "learning_rate": 6.680552661808996e-05, + "loss": 0.0598, + "step": 25928 + }, + { + "epoch": 0.6078165931187257, + "grad_norm": 0.3262938857078552, + "learning_rate": 6.679857822004623e-05, + "loss": 0.0526, + "step": 25929 + }, + { + "epoch": 0.6078400346935308, + "grad_norm": 0.2361421436071396, + "learning_rate": 6.679163000215429e-05, + "loss": 0.0417, + "step": 25930 + }, + { + "epoch": 0.6078634762683357, + "grad_norm": 0.4769335985183716, + "learning_rate": 6.678468196445195e-05, + "loss": 0.1337, + "step": 25931 + }, + { + "epoch": 0.6078869178431408, + "grad_norm": 0.20560158789157867, + "learning_rate": 6.677773410697683e-05, + "loss": 0.0434, + "step": 25932 + }, + { + "epoch": 0.6079103594179457, + "grad_norm": 0.05100120231509209, + "learning_rate": 6.677078642976669e-05, + "loss": 0.0087, + "step": 25933 + }, + { + "epoch": 0.6079338009927507, + "grad_norm": 0.5317437052726746, + "learning_rate": 6.676383893285917e-05, + "loss": 0.0387, + "step": 25934 + }, + { + "epoch": 0.6079572425675557, + "grad_norm": 0.7351654171943665, + "learning_rate": 6.6756891616292e-05, + "loss": 0.4774, + "step": 25935 + }, + { + "epoch": 0.6079806841423607, + "grad_norm": 0.35014644265174866, + "learning_rate": 6.674994448010287e-05, + "loss": 0.063, + "step": 25936 + }, + { + "epoch": 0.6080041257171657, + "grad_norm": 0.39483824372291565, + "learning_rate": 6.674299752432946e-05, + "loss": 0.05, + "step": 25937 + }, + { + "epoch": 0.6080275672919707, + "grad_norm": 0.2144414484500885, + "learning_rate": 6.67360507490095e-05, + "loss": 0.0381, + "step": 25938 + }, + { + "epoch": 0.6080510088667757, + "grad_norm": 0.40180331468582153, + "learning_rate": 6.672910415418066e-05, + "loss": 0.031, + "step": 25939 + }, + { + "epoch": 0.6080744504415807, + "grad_norm": 0.05671749264001846, + "learning_rate": 6.672215773988064e-05, + "loss": 0.0066, + "step": 25940 + }, + { + "epoch": 0.6080978920163856, + "grad_norm": 0.6459435224533081, + "learning_rate": 6.671521150614715e-05, + "loss": 0.1792, + "step": 25941 + }, + { + "epoch": 0.6081213335911907, + "grad_norm": 0.4991580843925476, + "learning_rate": 6.670826545301783e-05, + "loss": 0.1068, + "step": 25942 + }, + { + "epoch": 0.6081447751659956, + "grad_norm": 0.2853560447692871, + "learning_rate": 6.670131958053041e-05, + "loss": 0.055, + "step": 25943 + }, + { + "epoch": 0.6081682167408007, + "grad_norm": 0.2224910706281662, + "learning_rate": 6.669437388872256e-05, + "loss": 0.0468, + "step": 25944 + }, + { + "epoch": 0.6081916583156056, + "grad_norm": 0.45151108503341675, + "learning_rate": 6.668742837763197e-05, + "loss": 0.0633, + "step": 25945 + }, + { + "epoch": 0.6082150998904107, + "grad_norm": 0.34036314487457275, + "learning_rate": 6.668048304729634e-05, + "loss": 0.0745, + "step": 25946 + }, + { + "epoch": 0.6082385414652156, + "grad_norm": 0.6732414364814758, + "learning_rate": 6.667353789775335e-05, + "loss": 0.0838, + "step": 25947 + }, + { + "epoch": 0.6082619830400207, + "grad_norm": 0.23492905497550964, + "learning_rate": 6.666659292904065e-05, + "loss": 0.0461, + "step": 25948 + }, + { + "epoch": 0.6082854246148256, + "grad_norm": 0.21959662437438965, + "learning_rate": 6.665964814119595e-05, + "loss": 0.0439, + "step": 25949 + }, + { + "epoch": 0.6083088661896306, + "grad_norm": 0.4042263925075531, + "learning_rate": 6.665270353425697e-05, + "loss": 0.0558, + "step": 25950 + }, + { + "epoch": 0.6083323077644356, + "grad_norm": 0.19327493011951447, + "learning_rate": 6.664575910826137e-05, + "loss": 0.0298, + "step": 25951 + }, + { + "epoch": 0.6083557493392406, + "grad_norm": 0.26135075092315674, + "learning_rate": 6.663881486324677e-05, + "loss": 0.038, + "step": 25952 + }, + { + "epoch": 0.6083791909140456, + "grad_norm": 0.18365855515003204, + "learning_rate": 6.663187079925094e-05, + "loss": 0.0303, + "step": 25953 + }, + { + "epoch": 0.6084026324888506, + "grad_norm": 0.5872632265090942, + "learning_rate": 6.662492691631152e-05, + "loss": 0.067, + "step": 25954 + }, + { + "epoch": 0.6084260740636556, + "grad_norm": 0.35274800658226013, + "learning_rate": 6.661798321446615e-05, + "loss": 0.0767, + "step": 25955 + }, + { + "epoch": 0.6084495156384606, + "grad_norm": 0.5485057234764099, + "learning_rate": 6.661103969375257e-05, + "loss": 0.1431, + "step": 25956 + }, + { + "epoch": 0.6084729572132656, + "grad_norm": 0.5440756678581238, + "learning_rate": 6.660409635420844e-05, + "loss": 0.1113, + "step": 25957 + }, + { + "epoch": 0.6084963987880706, + "grad_norm": 0.22734159231185913, + "learning_rate": 6.659715319587138e-05, + "loss": 0.0534, + "step": 25958 + }, + { + "epoch": 0.6085198403628755, + "grad_norm": 0.499379962682724, + "learning_rate": 6.659021021877914e-05, + "loss": 0.0601, + "step": 25959 + }, + { + "epoch": 0.6085432819376806, + "grad_norm": 0.43556737899780273, + "learning_rate": 6.658326742296934e-05, + "loss": 0.4256, + "step": 25960 + }, + { + "epoch": 0.6085667235124855, + "grad_norm": 0.35014602541923523, + "learning_rate": 6.657632480847965e-05, + "loss": 0.0604, + "step": 25961 + }, + { + "epoch": 0.6085901650872906, + "grad_norm": 0.39402005076408386, + "learning_rate": 6.656938237534778e-05, + "loss": 0.0624, + "step": 25962 + }, + { + "epoch": 0.6086136066620955, + "grad_norm": 0.48446041345596313, + "learning_rate": 6.656244012361139e-05, + "loss": 0.1356, + "step": 25963 + }, + { + "epoch": 0.6086370482369006, + "grad_norm": 0.29565587639808655, + "learning_rate": 6.655549805330815e-05, + "loss": 0.0738, + "step": 25964 + }, + { + "epoch": 0.6086604898117055, + "grad_norm": 0.45267385244369507, + "learning_rate": 6.65485561644757e-05, + "loss": 0.1014, + "step": 25965 + }, + { + "epoch": 0.6086839313865106, + "grad_norm": 0.5915699005126953, + "learning_rate": 6.654161445715173e-05, + "loss": 0.5781, + "step": 25966 + }, + { + "epoch": 0.6087073729613155, + "grad_norm": 0.26076430082321167, + "learning_rate": 6.653467293137392e-05, + "loss": 0.0344, + "step": 25967 + }, + { + "epoch": 0.6087308145361205, + "grad_norm": 0.20694179832935333, + "learning_rate": 6.652773158717988e-05, + "loss": 0.0287, + "step": 25968 + }, + { + "epoch": 0.6087542561109255, + "grad_norm": 0.15321867167949677, + "learning_rate": 6.652079042460734e-05, + "loss": 0.0366, + "step": 25969 + }, + { + "epoch": 0.6087776976857305, + "grad_norm": 0.6725318431854248, + "learning_rate": 6.65138494436939e-05, + "loss": 0.1128, + "step": 25970 + }, + { + "epoch": 0.6088011392605355, + "grad_norm": 0.5392751693725586, + "learning_rate": 6.650690864447729e-05, + "loss": 0.0538, + "step": 25971 + }, + { + "epoch": 0.6088245808353405, + "grad_norm": 0.6975038647651672, + "learning_rate": 6.649996802699513e-05, + "loss": 0.1049, + "step": 25972 + }, + { + "epoch": 0.6088480224101455, + "grad_norm": 0.32115358114242554, + "learning_rate": 6.649302759128504e-05, + "loss": 0.0465, + "step": 25973 + }, + { + "epoch": 0.6088714639849505, + "grad_norm": 0.5329757332801819, + "learning_rate": 6.648608733738477e-05, + "loss": 0.1112, + "step": 25974 + }, + { + "epoch": 0.6088949055597555, + "grad_norm": 0.34388554096221924, + "learning_rate": 6.647914726533191e-05, + "loss": 0.0742, + "step": 25975 + }, + { + "epoch": 0.6089183471345605, + "grad_norm": 0.3889452815055847, + "learning_rate": 6.647220737516416e-05, + "loss": 0.0699, + "step": 25976 + }, + { + "epoch": 0.6089417887093656, + "grad_norm": 0.3788398206233978, + "learning_rate": 6.646526766691914e-05, + "loss": 0.0544, + "step": 25977 + }, + { + "epoch": 0.6089652302841705, + "grad_norm": 0.26876941323280334, + "learning_rate": 6.645832814063453e-05, + "loss": 0.0326, + "step": 25978 + }, + { + "epoch": 0.6089886718589755, + "grad_norm": 0.29974788427352905, + "learning_rate": 6.645138879634799e-05, + "loss": 0.0383, + "step": 25979 + }, + { + "epoch": 0.6090121134337805, + "grad_norm": 0.285138875246048, + "learning_rate": 6.644444963409712e-05, + "loss": 0.0413, + "step": 25980 + }, + { + "epoch": 0.6090355550085855, + "grad_norm": 0.6213381290435791, + "learning_rate": 6.643751065391964e-05, + "loss": 0.0958, + "step": 25981 + }, + { + "epoch": 0.6090589965833905, + "grad_norm": 0.10467272251844406, + "learning_rate": 6.643057185585317e-05, + "loss": 0.0234, + "step": 25982 + }, + { + "epoch": 0.6090824381581955, + "grad_norm": 0.6264420747756958, + "learning_rate": 6.642363323993534e-05, + "loss": 0.1422, + "step": 25983 + }, + { + "epoch": 0.6091058797330005, + "grad_norm": 0.5084954500198364, + "learning_rate": 6.641669480620383e-05, + "loss": 0.4756, + "step": 25984 + }, + { + "epoch": 0.6091293213078055, + "grad_norm": 0.5004844665527344, + "learning_rate": 6.640975655469626e-05, + "loss": 0.0863, + "step": 25985 + }, + { + "epoch": 0.6091527628826104, + "grad_norm": 0.5117517709732056, + "learning_rate": 6.640281848545032e-05, + "loss": 0.1069, + "step": 25986 + }, + { + "epoch": 0.6091762044574155, + "grad_norm": 0.4748938977718353, + "learning_rate": 6.639588059850361e-05, + "loss": 0.1193, + "step": 25987 + }, + { + "epoch": 0.6091996460322204, + "grad_norm": 0.32289671897888184, + "learning_rate": 6.638894289389382e-05, + "loss": 0.0594, + "step": 25988 + }, + { + "epoch": 0.6092230876070255, + "grad_norm": 0.24476049840450287, + "learning_rate": 6.638200537165858e-05, + "loss": 0.0514, + "step": 25989 + }, + { + "epoch": 0.6092465291818304, + "grad_norm": 0.4751192629337311, + "learning_rate": 6.63750680318355e-05, + "loss": 0.0921, + "step": 25990 + }, + { + "epoch": 0.6092699707566355, + "grad_norm": 0.0707332119345665, + "learning_rate": 6.636813087446225e-05, + "loss": 0.011, + "step": 25991 + }, + { + "epoch": 0.6092934123314404, + "grad_norm": 0.46633538603782654, + "learning_rate": 6.636119389957649e-05, + "loss": 0.0492, + "step": 25992 + }, + { + "epoch": 0.6093168539062455, + "grad_norm": 0.3339546322822571, + "learning_rate": 6.635425710721582e-05, + "loss": 0.0751, + "step": 25993 + }, + { + "epoch": 0.6093402954810504, + "grad_norm": 0.1305597722530365, + "learning_rate": 6.634732049741789e-05, + "loss": 0.0302, + "step": 25994 + }, + { + "epoch": 0.6093637370558554, + "grad_norm": 0.4324108064174652, + "learning_rate": 6.634038407022036e-05, + "loss": 0.1237, + "step": 25995 + }, + { + "epoch": 0.6093871786306604, + "grad_norm": 0.4654635190963745, + "learning_rate": 6.633344782566085e-05, + "loss": 0.1049, + "step": 25996 + }, + { + "epoch": 0.6094106202054654, + "grad_norm": 0.4172046482563019, + "learning_rate": 6.632651176377696e-05, + "loss": 0.0755, + "step": 25997 + }, + { + "epoch": 0.6094340617802704, + "grad_norm": 0.5784053206443787, + "learning_rate": 6.63195758846064e-05, + "loss": 0.1045, + "step": 25998 + }, + { + "epoch": 0.6094575033550754, + "grad_norm": 0.6109050512313843, + "learning_rate": 6.631264018818678e-05, + "loss": 0.0532, + "step": 25999 + }, + { + "epoch": 0.6094809449298804, + "grad_norm": 0.5939928889274597, + "learning_rate": 6.630570467455568e-05, + "loss": 0.4392, + "step": 26000 + }, + { + "epoch": 0.6095043865046854, + "grad_norm": 0.5481806397438049, + "learning_rate": 6.629876934375081e-05, + "loss": 0.3775, + "step": 26001 + }, + { + "epoch": 0.6095278280794904, + "grad_norm": 0.05197207257151604, + "learning_rate": 6.629183419580978e-05, + "loss": 0.006, + "step": 26002 + }, + { + "epoch": 0.6095512696542954, + "grad_norm": 0.18754881620407104, + "learning_rate": 6.628489923077016e-05, + "loss": 0.0206, + "step": 26003 + }, + { + "epoch": 0.6095747112291003, + "grad_norm": 0.44071483612060547, + "learning_rate": 6.627796444866965e-05, + "loss": 0.1131, + "step": 26004 + }, + { + "epoch": 0.6095981528039054, + "grad_norm": 0.3916849195957184, + "learning_rate": 6.627102984954588e-05, + "loss": 0.1109, + "step": 26005 + }, + { + "epoch": 0.6096215943787103, + "grad_norm": 0.5000531077384949, + "learning_rate": 6.62640954334364e-05, + "loss": 0.1288, + "step": 26006 + }, + { + "epoch": 0.6096450359535154, + "grad_norm": 0.3520033061504364, + "learning_rate": 6.62571612003789e-05, + "loss": 0.0803, + "step": 26007 + }, + { + "epoch": 0.6096684775283203, + "grad_norm": 0.25872084498405457, + "learning_rate": 6.6250227150411e-05, + "loss": 0.0338, + "step": 26008 + }, + { + "epoch": 0.6096919191031254, + "grad_norm": 0.13196898996829987, + "learning_rate": 6.624329328357032e-05, + "loss": 0.0268, + "step": 26009 + }, + { + "epoch": 0.6097153606779303, + "grad_norm": 0.25246790051460266, + "learning_rate": 6.623635959989445e-05, + "loss": 0.0533, + "step": 26010 + }, + { + "epoch": 0.6097388022527354, + "grad_norm": 0.44993463158607483, + "learning_rate": 6.622942609942107e-05, + "loss": 0.0633, + "step": 26011 + }, + { + "epoch": 0.6097622438275403, + "grad_norm": 0.13253086805343628, + "learning_rate": 6.622249278218777e-05, + "loss": 0.0333, + "step": 26012 + }, + { + "epoch": 0.6097856854023453, + "grad_norm": 0.14777375757694244, + "learning_rate": 6.621555964823218e-05, + "loss": 0.0386, + "step": 26013 + }, + { + "epoch": 0.6098091269771503, + "grad_norm": 0.4862077832221985, + "learning_rate": 6.620862669759192e-05, + "loss": 0.1131, + "step": 26014 + }, + { + "epoch": 0.6098325685519553, + "grad_norm": 0.15832577645778656, + "learning_rate": 6.620169393030456e-05, + "loss": 0.0358, + "step": 26015 + }, + { + "epoch": 0.6098560101267603, + "grad_norm": 0.08605484664440155, + "learning_rate": 6.61947613464078e-05, + "loss": 0.0161, + "step": 26016 + }, + { + "epoch": 0.6098794517015653, + "grad_norm": 0.45361030101776123, + "learning_rate": 6.618782894593923e-05, + "loss": 0.0398, + "step": 26017 + }, + { + "epoch": 0.6099028932763703, + "grad_norm": 0.19002430140972137, + "learning_rate": 6.618089672893639e-05, + "loss": 0.0235, + "step": 26018 + }, + { + "epoch": 0.6099263348511753, + "grad_norm": 0.18809038400650024, + "learning_rate": 6.617396469543701e-05, + "loss": 0.0165, + "step": 26019 + }, + { + "epoch": 0.6099497764259803, + "grad_norm": 0.5449823141098022, + "learning_rate": 6.616703284547863e-05, + "loss": 0.0721, + "step": 26020 + }, + { + "epoch": 0.6099732180007853, + "grad_norm": 0.5232678055763245, + "learning_rate": 6.616010117909888e-05, + "loss": 0.0736, + "step": 26021 + }, + { + "epoch": 0.6099966595755902, + "grad_norm": 1.4256047010421753, + "learning_rate": 6.615316969633534e-05, + "loss": 0.0586, + "step": 26022 + }, + { + "epoch": 0.6100201011503953, + "grad_norm": 0.20081526041030884, + "learning_rate": 6.61462383972257e-05, + "loss": 0.0435, + "step": 26023 + }, + { + "epoch": 0.6100435427252002, + "grad_norm": 0.3894270360469818, + "learning_rate": 6.61393072818075e-05, + "loss": 0.0617, + "step": 26024 + }, + { + "epoch": 0.6100669843000053, + "grad_norm": 0.4522853493690491, + "learning_rate": 6.613237635011837e-05, + "loss": 0.0842, + "step": 26025 + }, + { + "epoch": 0.6100904258748102, + "grad_norm": 0.655810534954071, + "learning_rate": 6.612544560219594e-05, + "loss": 0.1786, + "step": 26026 + }, + { + "epoch": 0.6101138674496153, + "grad_norm": 0.3753197491168976, + "learning_rate": 6.611851503807779e-05, + "loss": 0.0804, + "step": 26027 + }, + { + "epoch": 0.6101373090244203, + "grad_norm": 1.2317428588867188, + "learning_rate": 6.61115846578015e-05, + "loss": 0.1588, + "step": 26028 + }, + { + "epoch": 0.6101607505992253, + "grad_norm": 0.23718486726284027, + "learning_rate": 6.610465446140475e-05, + "loss": 0.065, + "step": 26029 + }, + { + "epoch": 0.6101841921740303, + "grad_norm": 0.3934721052646637, + "learning_rate": 6.609772444892508e-05, + "loss": 0.0323, + "step": 26030 + }, + { + "epoch": 0.6102076337488352, + "grad_norm": 0.4998803734779358, + "learning_rate": 6.609079462040009e-05, + "loss": 0.4792, + "step": 26031 + }, + { + "epoch": 0.6102310753236403, + "grad_norm": 0.2862521708011627, + "learning_rate": 6.608386497586743e-05, + "loss": 0.0365, + "step": 26032 + }, + { + "epoch": 0.6102545168984452, + "grad_norm": 0.4801344871520996, + "learning_rate": 6.607693551536466e-05, + "loss": 0.136, + "step": 26033 + }, + { + "epoch": 0.6102779584732503, + "grad_norm": 0.3986625373363495, + "learning_rate": 6.607000623892936e-05, + "loss": 0.0628, + "step": 26034 + }, + { + "epoch": 0.6103014000480552, + "grad_norm": 0.7776373028755188, + "learning_rate": 6.606307714659918e-05, + "loss": 0.1032, + "step": 26035 + }, + { + "epoch": 0.6103248416228603, + "grad_norm": 0.2924145460128784, + "learning_rate": 6.60561482384117e-05, + "loss": 0.0794, + "step": 26036 + }, + { + "epoch": 0.6103482831976652, + "grad_norm": 0.388813316822052, + "learning_rate": 6.604921951440454e-05, + "loss": 0.073, + "step": 26037 + }, + { + "epoch": 0.6103717247724703, + "grad_norm": 0.3854733109474182, + "learning_rate": 6.604229097461524e-05, + "loss": 0.0312, + "step": 26038 + }, + { + "epoch": 0.6103951663472752, + "grad_norm": 0.4652620255947113, + "learning_rate": 6.603536261908145e-05, + "loss": 0.0632, + "step": 26039 + }, + { + "epoch": 0.6104186079220802, + "grad_norm": 0.4094516932964325, + "learning_rate": 6.602843444784072e-05, + "loss": 0.0587, + "step": 26040 + }, + { + "epoch": 0.6104420494968852, + "grad_norm": 1.0447678565979004, + "learning_rate": 6.602150646093066e-05, + "loss": 0.135, + "step": 26041 + }, + { + "epoch": 0.6104654910716902, + "grad_norm": 0.1515606790781021, + "learning_rate": 6.601457865838887e-05, + "loss": 0.0388, + "step": 26042 + }, + { + "epoch": 0.6104889326464952, + "grad_norm": 0.7142705321311951, + "learning_rate": 6.600765104025294e-05, + "loss": 0.1146, + "step": 26043 + }, + { + "epoch": 0.6105123742213002, + "grad_norm": 0.14731785655021667, + "learning_rate": 6.600072360656042e-05, + "loss": 0.0121, + "step": 26044 + }, + { + "epoch": 0.6105358157961052, + "grad_norm": 0.4794316589832306, + "learning_rate": 6.599379635734896e-05, + "loss": 0.094, + "step": 26045 + }, + { + "epoch": 0.6105592573709102, + "grad_norm": 0.48496416211128235, + "learning_rate": 6.598686929265608e-05, + "loss": 0.0995, + "step": 26046 + }, + { + "epoch": 0.6105826989457152, + "grad_norm": 0.2595074474811554, + "learning_rate": 6.597994241251943e-05, + "loss": 0.0519, + "step": 26047 + }, + { + "epoch": 0.6106061405205202, + "grad_norm": 0.6195632815361023, + "learning_rate": 6.597301571697655e-05, + "loss": 0.1024, + "step": 26048 + }, + { + "epoch": 0.6106295820953251, + "grad_norm": 0.43139833211898804, + "learning_rate": 6.596608920606506e-05, + "loss": 0.0832, + "step": 26049 + }, + { + "epoch": 0.6106530236701302, + "grad_norm": 0.22141209244728088, + "learning_rate": 6.595916287982253e-05, + "loss": 0.111, + "step": 26050 + }, + { + "epoch": 0.6106764652449351, + "grad_norm": 0.3135945796966553, + "learning_rate": 6.595223673828655e-05, + "loss": 0.0404, + "step": 26051 + }, + { + "epoch": 0.6106999068197402, + "grad_norm": 0.39431318640708923, + "learning_rate": 6.594531078149468e-05, + "loss": 0.3964, + "step": 26052 + }, + { + "epoch": 0.6107233483945451, + "grad_norm": 0.40651243925094604, + "learning_rate": 6.593838500948449e-05, + "loss": 0.0531, + "step": 26053 + }, + { + "epoch": 0.6107467899693502, + "grad_norm": 0.29068058729171753, + "learning_rate": 6.593145942229361e-05, + "loss": 0.069, + "step": 26054 + }, + { + "epoch": 0.6107702315441551, + "grad_norm": 0.22927717864513397, + "learning_rate": 6.592453401995959e-05, + "loss": 0.0347, + "step": 26055 + }, + { + "epoch": 0.6107936731189602, + "grad_norm": 0.6094481348991394, + "learning_rate": 6.591760880251998e-05, + "loss": 0.6825, + "step": 26056 + }, + { + "epoch": 0.6108171146937651, + "grad_norm": 0.484868586063385, + "learning_rate": 6.591068377001241e-05, + "loss": 0.0669, + "step": 26057 + }, + { + "epoch": 0.6108405562685701, + "grad_norm": 0.30602577328681946, + "learning_rate": 6.590375892247438e-05, + "loss": 0.0641, + "step": 26058 + }, + { + "epoch": 0.6108639978433751, + "grad_norm": 0.3827296495437622, + "learning_rate": 6.589683425994355e-05, + "loss": 0.0613, + "step": 26059 + }, + { + "epoch": 0.6108874394181801, + "grad_norm": 0.09540089964866638, + "learning_rate": 6.588990978245745e-05, + "loss": 0.0207, + "step": 26060 + }, + { + "epoch": 0.6109108809929851, + "grad_norm": 0.39743685722351074, + "learning_rate": 6.588298549005367e-05, + "loss": 0.0522, + "step": 26061 + }, + { + "epoch": 0.6109343225677901, + "grad_norm": 0.24315641820430756, + "learning_rate": 6.587606138276979e-05, + "loss": 0.0485, + "step": 26062 + }, + { + "epoch": 0.6109577641425951, + "grad_norm": 0.22445785999298096, + "learning_rate": 6.586913746064332e-05, + "loss": 0.0604, + "step": 26063 + }, + { + "epoch": 0.6109812057174001, + "grad_norm": 0.5277338624000549, + "learning_rate": 6.586221372371188e-05, + "loss": 0.1066, + "step": 26064 + }, + { + "epoch": 0.611004647292205, + "grad_norm": 0.47897031903266907, + "learning_rate": 6.585529017201304e-05, + "loss": 0.0919, + "step": 26065 + }, + { + "epoch": 0.6110280888670101, + "grad_norm": 0.2530364990234375, + "learning_rate": 6.584836680558434e-05, + "loss": 0.0256, + "step": 26066 + }, + { + "epoch": 0.611051530441815, + "grad_norm": 0.565274715423584, + "learning_rate": 6.584144362446337e-05, + "loss": 0.0906, + "step": 26067 + }, + { + "epoch": 0.6110749720166201, + "grad_norm": 0.46074047684669495, + "learning_rate": 6.58345206286877e-05, + "loss": 0.0876, + "step": 26068 + }, + { + "epoch": 0.611098413591425, + "grad_norm": 0.5675301551818848, + "learning_rate": 6.582759781829486e-05, + "loss": 0.5074, + "step": 26069 + }, + { + "epoch": 0.6111218551662301, + "grad_norm": 0.5795090198516846, + "learning_rate": 6.58206751933224e-05, + "loss": 0.0857, + "step": 26070 + }, + { + "epoch": 0.611145296741035, + "grad_norm": 0.26685115694999695, + "learning_rate": 6.581375275380798e-05, + "loss": 0.0455, + "step": 26071 + }, + { + "epoch": 0.6111687383158401, + "grad_norm": 0.4781448245048523, + "learning_rate": 6.580683049978908e-05, + "loss": 0.0953, + "step": 26072 + }, + { + "epoch": 0.611192179890645, + "grad_norm": 0.37491488456726074, + "learning_rate": 6.579990843130328e-05, + "loss": 0.0725, + "step": 26073 + }, + { + "epoch": 0.61121562146545, + "grad_norm": 0.39678263664245605, + "learning_rate": 6.579298654838814e-05, + "loss": 0.0436, + "step": 26074 + }, + { + "epoch": 0.611239063040255, + "grad_norm": 0.5601609945297241, + "learning_rate": 6.578606485108123e-05, + "loss": 0.0833, + "step": 26075 + }, + { + "epoch": 0.61126250461506, + "grad_norm": 0.6445022225379944, + "learning_rate": 6.577914333942005e-05, + "loss": 0.1014, + "step": 26076 + }, + { + "epoch": 0.611285946189865, + "grad_norm": 0.4881713390350342, + "learning_rate": 6.577222201344224e-05, + "loss": 0.0745, + "step": 26077 + }, + { + "epoch": 0.61130938776467, + "grad_norm": 0.07188998907804489, + "learning_rate": 6.576530087318533e-05, + "loss": 0.0115, + "step": 26078 + }, + { + "epoch": 0.6113328293394751, + "grad_norm": 0.18715016543865204, + "learning_rate": 6.57583799186868e-05, + "loss": 0.0328, + "step": 26079 + }, + { + "epoch": 0.61135627091428, + "grad_norm": 0.1875247210264206, + "learning_rate": 6.575145914998432e-05, + "loss": 0.0301, + "step": 26080 + }, + { + "epoch": 0.6113797124890851, + "grad_norm": 0.42418453097343445, + "learning_rate": 6.574453856711538e-05, + "loss": 0.0639, + "step": 26081 + }, + { + "epoch": 0.61140315406389, + "grad_norm": 0.49503371119499207, + "learning_rate": 6.57376181701175e-05, + "loss": 0.684, + "step": 26082 + }, + { + "epoch": 0.6114265956386951, + "grad_norm": 0.573725700378418, + "learning_rate": 6.573069795902827e-05, + "loss": 0.5256, + "step": 26083 + }, + { + "epoch": 0.6114500372135, + "grad_norm": 0.20074479281902313, + "learning_rate": 6.572377793388525e-05, + "loss": 0.0133, + "step": 26084 + }, + { + "epoch": 0.611473478788305, + "grad_norm": 0.5488497614860535, + "learning_rate": 6.571685809472599e-05, + "loss": 0.1057, + "step": 26085 + }, + { + "epoch": 0.61149692036311, + "grad_norm": 0.2890622913837433, + "learning_rate": 6.570993844158799e-05, + "loss": 0.0417, + "step": 26086 + }, + { + "epoch": 0.611520361937915, + "grad_norm": 0.39364156126976013, + "learning_rate": 6.570301897450887e-05, + "loss": 0.0685, + "step": 26087 + }, + { + "epoch": 0.61154380351272, + "grad_norm": 0.08676113933324814, + "learning_rate": 6.56960996935261e-05, + "loss": 0.0132, + "step": 26088 + }, + { + "epoch": 0.611567245087525, + "grad_norm": 0.42752400040626526, + "learning_rate": 6.568918059867728e-05, + "loss": 0.0494, + "step": 26089 + }, + { + "epoch": 0.61159068666233, + "grad_norm": 0.36586278676986694, + "learning_rate": 6.568226168999992e-05, + "loss": 0.0654, + "step": 26090 + }, + { + "epoch": 0.611614128237135, + "grad_norm": 0.17543818056583405, + "learning_rate": 6.567534296753157e-05, + "loss": 0.027, + "step": 26091 + }, + { + "epoch": 0.61163756981194, + "grad_norm": 0.4735441207885742, + "learning_rate": 6.566842443130978e-05, + "loss": 0.1175, + "step": 26092 + }, + { + "epoch": 0.611661011386745, + "grad_norm": 0.21326693892478943, + "learning_rate": 6.566150608137209e-05, + "loss": 0.0398, + "step": 26093 + }, + { + "epoch": 0.6116844529615499, + "grad_norm": 0.260345458984375, + "learning_rate": 6.565458791775603e-05, + "loss": 0.0166, + "step": 26094 + }, + { + "epoch": 0.611707894536355, + "grad_norm": 0.1820545643568039, + "learning_rate": 6.564766994049909e-05, + "loss": 0.0223, + "step": 26095 + }, + { + "epoch": 0.6117313361111599, + "grad_norm": 0.12405206263065338, + "learning_rate": 6.564075214963892e-05, + "loss": 0.0216, + "step": 26096 + }, + { + "epoch": 0.611754777685965, + "grad_norm": 0.6025298237800598, + "learning_rate": 6.563383454521299e-05, + "loss": 0.1081, + "step": 26097 + }, + { + "epoch": 0.6117782192607699, + "grad_norm": 0.5844202041625977, + "learning_rate": 6.562691712725882e-05, + "loss": 0.1131, + "step": 26098 + }, + { + "epoch": 0.611801660835575, + "grad_norm": 0.6481610536575317, + "learning_rate": 6.561999989581398e-05, + "loss": 0.1251, + "step": 26099 + }, + { + "epoch": 0.6118251024103799, + "grad_norm": 0.2707633376121521, + "learning_rate": 6.5613082850916e-05, + "loss": 0.0452, + "step": 26100 + }, + { + "epoch": 0.611848543985185, + "grad_norm": 0.8203204274177551, + "learning_rate": 6.560616599260236e-05, + "loss": 0.1304, + "step": 26101 + }, + { + "epoch": 0.6118719855599899, + "grad_norm": 0.4836406111717224, + "learning_rate": 6.559924932091066e-05, + "loss": 0.056, + "step": 26102 + }, + { + "epoch": 0.611895427134795, + "grad_norm": 0.12504883110523224, + "learning_rate": 6.559233283587841e-05, + "loss": 0.0231, + "step": 26103 + }, + { + "epoch": 0.6119188687095999, + "grad_norm": 0.5178942680358887, + "learning_rate": 6.55854165375431e-05, + "loss": 0.1161, + "step": 26104 + }, + { + "epoch": 0.6119423102844049, + "grad_norm": 0.6171276569366455, + "learning_rate": 6.557850042594232e-05, + "loss": 0.7488, + "step": 26105 + }, + { + "epoch": 0.6119657518592099, + "grad_norm": 0.11559702455997467, + "learning_rate": 6.557158450111354e-05, + "loss": 0.0129, + "step": 26106 + }, + { + "epoch": 0.6119891934340149, + "grad_norm": 0.33505311608314514, + "learning_rate": 6.556466876309428e-05, + "loss": 0.0447, + "step": 26107 + }, + { + "epoch": 0.6120126350088199, + "grad_norm": 0.38664501905441284, + "learning_rate": 6.555775321192212e-05, + "loss": 0.0343, + "step": 26108 + }, + { + "epoch": 0.6120360765836249, + "grad_norm": 0.365350604057312, + "learning_rate": 6.555083784763458e-05, + "loss": 0.0548, + "step": 26109 + }, + { + "epoch": 0.6120595181584298, + "grad_norm": 0.1445893794298172, + "learning_rate": 6.554392267026916e-05, + "loss": 0.028, + "step": 26110 + }, + { + "epoch": 0.6120829597332349, + "grad_norm": 0.5404203534126282, + "learning_rate": 6.553700767986337e-05, + "loss": 0.0959, + "step": 26111 + }, + { + "epoch": 0.6121064013080398, + "grad_norm": 0.26781266927719116, + "learning_rate": 6.553009287645475e-05, + "loss": 0.0526, + "step": 26112 + }, + { + "epoch": 0.6121298428828449, + "grad_norm": 0.6349937319755554, + "learning_rate": 6.552317826008083e-05, + "loss": 0.1202, + "step": 26113 + }, + { + "epoch": 0.6121532844576498, + "grad_norm": 0.10296736657619476, + "learning_rate": 6.551626383077908e-05, + "loss": 0.0176, + "step": 26114 + }, + { + "epoch": 0.6121767260324549, + "grad_norm": 0.6229033470153809, + "learning_rate": 6.55093495885871e-05, + "loss": 0.0627, + "step": 26115 + }, + { + "epoch": 0.6122001676072598, + "grad_norm": 0.2319263219833374, + "learning_rate": 6.550243553354233e-05, + "loss": 0.0384, + "step": 26116 + }, + { + "epoch": 0.6122236091820649, + "grad_norm": 0.6436220407485962, + "learning_rate": 6.549552166568231e-05, + "loss": 0.1339, + "step": 26117 + }, + { + "epoch": 0.6122470507568698, + "grad_norm": 0.5927790403366089, + "learning_rate": 6.548860798504459e-05, + "loss": 0.1279, + "step": 26118 + }, + { + "epoch": 0.6122704923316749, + "grad_norm": 0.3496629297733307, + "learning_rate": 6.548169449166659e-05, + "loss": 0.0862, + "step": 26119 + }, + { + "epoch": 0.6122939339064798, + "grad_norm": 0.24501103162765503, + "learning_rate": 6.547478118558595e-05, + "loss": 0.039, + "step": 26120 + }, + { + "epoch": 0.6123173754812848, + "grad_norm": 0.32464712858200073, + "learning_rate": 6.546786806684006e-05, + "loss": 0.0569, + "step": 26121 + }, + { + "epoch": 0.6123408170560898, + "grad_norm": 0.4739351272583008, + "learning_rate": 6.546095513546654e-05, + "loss": 0.0717, + "step": 26122 + }, + { + "epoch": 0.6123642586308948, + "grad_norm": 0.19368694722652435, + "learning_rate": 6.545404239150284e-05, + "loss": 0.018, + "step": 26123 + }, + { + "epoch": 0.6123877002056998, + "grad_norm": 0.3128199279308319, + "learning_rate": 6.544712983498647e-05, + "loss": 0.0644, + "step": 26124 + }, + { + "epoch": 0.6124111417805048, + "grad_norm": 0.5549314618110657, + "learning_rate": 6.544021746595495e-05, + "loss": 0.1262, + "step": 26125 + }, + { + "epoch": 0.6124345833553098, + "grad_norm": 0.4713408648967743, + "learning_rate": 6.543330528444576e-05, + "loss": 0.0993, + "step": 26126 + }, + { + "epoch": 0.6124580249301148, + "grad_norm": 0.9993085265159607, + "learning_rate": 6.542639329049645e-05, + "loss": 0.2017, + "step": 26127 + }, + { + "epoch": 0.6124814665049197, + "grad_norm": 0.2250470072031021, + "learning_rate": 6.541948148414451e-05, + "loss": 0.0455, + "step": 26128 + }, + { + "epoch": 0.6125049080797248, + "grad_norm": 0.11180825531482697, + "learning_rate": 6.54125698654274e-05, + "loss": 0.0169, + "step": 26129 + }, + { + "epoch": 0.6125283496545298, + "grad_norm": 0.5578884482383728, + "learning_rate": 6.540565843438269e-05, + "loss": 0.6304, + "step": 26130 + }, + { + "epoch": 0.6125517912293348, + "grad_norm": 0.527312695980072, + "learning_rate": 6.539874719104782e-05, + "loss": 0.0586, + "step": 26131 + }, + { + "epoch": 0.6125752328041398, + "grad_norm": 0.4630059599876404, + "learning_rate": 6.539183613546033e-05, + "loss": 0.0712, + "step": 26132 + }, + { + "epoch": 0.6125986743789448, + "grad_norm": 0.29678648710250854, + "learning_rate": 6.538492526765771e-05, + "loss": 0.0561, + "step": 26133 + }, + { + "epoch": 0.6126221159537498, + "grad_norm": 0.5655714869499207, + "learning_rate": 6.53780145876775e-05, + "loss": 0.411, + "step": 26134 + }, + { + "epoch": 0.6126455575285548, + "grad_norm": 0.3713870048522949, + "learning_rate": 6.537110409555712e-05, + "loss": 0.0377, + "step": 26135 + }, + { + "epoch": 0.6126689991033598, + "grad_norm": 0.16963636875152588, + "learning_rate": 6.536419379133409e-05, + "loss": 0.04, + "step": 26136 + }, + { + "epoch": 0.6126924406781648, + "grad_norm": 0.31440722942352295, + "learning_rate": 6.535728367504597e-05, + "loss": 0.0638, + "step": 26137 + }, + { + "epoch": 0.6127158822529698, + "grad_norm": 0.2781258821487427, + "learning_rate": 6.535037374673017e-05, + "loss": 0.0315, + "step": 26138 + }, + { + "epoch": 0.6127393238277747, + "grad_norm": 0.32468074560165405, + "learning_rate": 6.53434640064242e-05, + "loss": 0.0656, + "step": 26139 + }, + { + "epoch": 0.6127627654025798, + "grad_norm": 0.23146241903305054, + "learning_rate": 6.53365544541656e-05, + "loss": 0.022, + "step": 26140 + }, + { + "epoch": 0.6127862069773847, + "grad_norm": 0.3425482511520386, + "learning_rate": 6.532964508999182e-05, + "loss": 0.0735, + "step": 26141 + }, + { + "epoch": 0.6128096485521898, + "grad_norm": 0.3764853775501251, + "learning_rate": 6.532273591394035e-05, + "loss": 0.0871, + "step": 26142 + }, + { + "epoch": 0.6128330901269947, + "grad_norm": 0.5819620490074158, + "learning_rate": 6.531582692604865e-05, + "loss": 0.0786, + "step": 26143 + }, + { + "epoch": 0.6128565317017998, + "grad_norm": 0.42556217312812805, + "learning_rate": 6.530891812635431e-05, + "loss": 0.0645, + "step": 26144 + }, + { + "epoch": 0.6128799732766047, + "grad_norm": 0.31317803263664246, + "learning_rate": 6.530200951489475e-05, + "loss": 0.0397, + "step": 26145 + }, + { + "epoch": 0.6129034148514098, + "grad_norm": 0.8223060965538025, + "learning_rate": 6.529510109170745e-05, + "loss": 0.2195, + "step": 26146 + }, + { + "epoch": 0.6129268564262147, + "grad_norm": 0.14260917901992798, + "learning_rate": 6.528819285682991e-05, + "loss": 0.0118, + "step": 26147 + }, + { + "epoch": 0.6129502980010197, + "grad_norm": 0.07727418839931488, + "learning_rate": 6.528128481029963e-05, + "loss": 0.0096, + "step": 26148 + }, + { + "epoch": 0.6129737395758247, + "grad_norm": 0.6052612066268921, + "learning_rate": 6.527437695215403e-05, + "loss": 0.5383, + "step": 26149 + }, + { + "epoch": 0.6129971811506297, + "grad_norm": 0.18879276514053345, + "learning_rate": 6.526746928243069e-05, + "loss": 0.0518, + "step": 26150 + }, + { + "epoch": 0.6130206227254347, + "grad_norm": 0.1961657702922821, + "learning_rate": 6.5260561801167e-05, + "loss": 0.0255, + "step": 26151 + }, + { + "epoch": 0.6130440643002397, + "grad_norm": 0.47314614057540894, + "learning_rate": 6.525365450840049e-05, + "loss": 0.0697, + "step": 26152 + }, + { + "epoch": 0.6130675058750447, + "grad_norm": 0.4985019862651825, + "learning_rate": 6.524674740416863e-05, + "loss": 0.0304, + "step": 26153 + }, + { + "epoch": 0.6130909474498497, + "grad_norm": 0.29654282331466675, + "learning_rate": 6.523984048850891e-05, + "loss": 0.0469, + "step": 26154 + }, + { + "epoch": 0.6131143890246546, + "grad_norm": 0.49093133211135864, + "learning_rate": 6.523293376145875e-05, + "loss": 0.0641, + "step": 26155 + }, + { + "epoch": 0.6131378305994597, + "grad_norm": 0.06151563674211502, + "learning_rate": 6.522602722305567e-05, + "loss": 0.0145, + "step": 26156 + }, + { + "epoch": 0.6131612721742646, + "grad_norm": 0.09380003064870834, + "learning_rate": 6.521912087333716e-05, + "loss": 0.021, + "step": 26157 + }, + { + "epoch": 0.6131847137490697, + "grad_norm": 0.6922360062599182, + "learning_rate": 6.521221471234072e-05, + "loss": 0.1474, + "step": 26158 + }, + { + "epoch": 0.6132081553238746, + "grad_norm": 0.15643522143363953, + "learning_rate": 6.520530874010374e-05, + "loss": 0.0267, + "step": 26159 + }, + { + "epoch": 0.6132315968986797, + "grad_norm": 0.5828675627708435, + "learning_rate": 6.519840295666373e-05, + "loss": 0.1488, + "step": 26160 + }, + { + "epoch": 0.6132550384734846, + "grad_norm": 0.32334065437316895, + "learning_rate": 6.519149736205819e-05, + "loss": 0.0519, + "step": 26161 + }, + { + "epoch": 0.6132784800482897, + "grad_norm": 0.124756820499897, + "learning_rate": 6.518459195632454e-05, + "loss": 0.0202, + "step": 26162 + }, + { + "epoch": 0.6133019216230946, + "grad_norm": 0.14058537781238556, + "learning_rate": 6.517768673950028e-05, + "loss": 0.0244, + "step": 26163 + }, + { + "epoch": 0.6133253631978997, + "grad_norm": 0.2057933360338211, + "learning_rate": 6.517078171162287e-05, + "loss": 0.0328, + "step": 26164 + }, + { + "epoch": 0.6133488047727046, + "grad_norm": 0.18955805897712708, + "learning_rate": 6.51638768727298e-05, + "loss": 0.0336, + "step": 26165 + }, + { + "epoch": 0.6133722463475096, + "grad_norm": 0.5651907920837402, + "learning_rate": 6.515697222285849e-05, + "loss": 0.1035, + "step": 26166 + }, + { + "epoch": 0.6133956879223146, + "grad_norm": 0.19584335386753082, + "learning_rate": 6.515006776204642e-05, + "loss": 0.0408, + "step": 26167 + }, + { + "epoch": 0.6134191294971196, + "grad_norm": 0.3800911605358124, + "learning_rate": 6.514316349033105e-05, + "loss": 0.0665, + "step": 26168 + }, + { + "epoch": 0.6134425710719246, + "grad_norm": 0.2716951370239258, + "learning_rate": 6.513625940774987e-05, + "loss": 0.0505, + "step": 26169 + }, + { + "epoch": 0.6134660126467296, + "grad_norm": 0.6362848281860352, + "learning_rate": 6.512935551434036e-05, + "loss": 0.1797, + "step": 26170 + }, + { + "epoch": 0.6134894542215346, + "grad_norm": 0.5477259755134583, + "learning_rate": 6.512245181013992e-05, + "loss": 0.0686, + "step": 26171 + }, + { + "epoch": 0.6135128957963396, + "grad_norm": 0.4501727521419525, + "learning_rate": 6.511554829518604e-05, + "loss": 0.0324, + "step": 26172 + }, + { + "epoch": 0.6135363373711445, + "grad_norm": 0.4225015938282013, + "learning_rate": 6.510864496951619e-05, + "loss": 0.0874, + "step": 26173 + }, + { + "epoch": 0.6135597789459496, + "grad_norm": 0.16012834012508392, + "learning_rate": 6.51017418331678e-05, + "loss": 0.02, + "step": 26174 + }, + { + "epoch": 0.6135832205207545, + "grad_norm": 0.32324695587158203, + "learning_rate": 6.509483888617833e-05, + "loss": 0.0781, + "step": 26175 + }, + { + "epoch": 0.6136066620955596, + "grad_norm": 0.2913019359111786, + "learning_rate": 6.508793612858528e-05, + "loss": 0.0376, + "step": 26176 + }, + { + "epoch": 0.6136301036703645, + "grad_norm": 0.16859975457191467, + "learning_rate": 6.508103356042603e-05, + "loss": 0.0339, + "step": 26177 + }, + { + "epoch": 0.6136535452451696, + "grad_norm": 0.3293060064315796, + "learning_rate": 6.50741311817381e-05, + "loss": 0.0655, + "step": 26178 + }, + { + "epoch": 0.6136769868199745, + "grad_norm": 0.4849894940853119, + "learning_rate": 6.506722899255892e-05, + "loss": 0.0748, + "step": 26179 + }, + { + "epoch": 0.6137004283947796, + "grad_norm": 0.561943769454956, + "learning_rate": 6.506032699292588e-05, + "loss": 0.124, + "step": 26180 + }, + { + "epoch": 0.6137238699695845, + "grad_norm": 0.4981536567211151, + "learning_rate": 6.505342518287653e-05, + "loss": 0.1006, + "step": 26181 + }, + { + "epoch": 0.6137473115443896, + "grad_norm": 0.48554089665412903, + "learning_rate": 6.504652356244829e-05, + "loss": 0.0719, + "step": 26182 + }, + { + "epoch": 0.6137707531191946, + "grad_norm": 0.4319089353084564, + "learning_rate": 6.50396221316786e-05, + "loss": 0.0599, + "step": 26183 + }, + { + "epoch": 0.6137941946939995, + "grad_norm": 0.23862873017787933, + "learning_rate": 6.503272089060486e-05, + "loss": 0.0439, + "step": 26184 + }, + { + "epoch": 0.6138176362688046, + "grad_norm": 0.5599182844161987, + "learning_rate": 6.50258198392646e-05, + "loss": 0.1219, + "step": 26185 + }, + { + "epoch": 0.6138410778436095, + "grad_norm": 0.13484768569469452, + "learning_rate": 6.501891897769523e-05, + "loss": 0.016, + "step": 26186 + }, + { + "epoch": 0.6138645194184146, + "grad_norm": 0.08672585338354111, + "learning_rate": 6.501201830593415e-05, + "loss": 0.0097, + "step": 26187 + }, + { + "epoch": 0.6138879609932195, + "grad_norm": 0.43580910563468933, + "learning_rate": 6.500511782401888e-05, + "loss": 0.0846, + "step": 26188 + }, + { + "epoch": 0.6139114025680246, + "grad_norm": 0.3848266303539276, + "learning_rate": 6.499821753198683e-05, + "loss": 0.088, + "step": 26189 + }, + { + "epoch": 0.6139348441428295, + "grad_norm": 0.18330639600753784, + "learning_rate": 6.499131742987538e-05, + "loss": 0.0329, + "step": 26190 + }, + { + "epoch": 0.6139582857176346, + "grad_norm": 0.07339102774858475, + "learning_rate": 6.498441751772208e-05, + "loss": 0.0158, + "step": 26191 + }, + { + "epoch": 0.6139817272924395, + "grad_norm": 0.3925187587738037, + "learning_rate": 6.497751779556426e-05, + "loss": 0.1115, + "step": 26192 + }, + { + "epoch": 0.6140051688672445, + "grad_norm": 0.25200486183166504, + "learning_rate": 6.497061826343946e-05, + "loss": 0.0306, + "step": 26193 + }, + { + "epoch": 0.6140286104420495, + "grad_norm": 0.28865712881088257, + "learning_rate": 6.496371892138506e-05, + "loss": 0.0441, + "step": 26194 + }, + { + "epoch": 0.6140520520168545, + "grad_norm": 0.19314280152320862, + "learning_rate": 6.495681976943851e-05, + "loss": 0.0282, + "step": 26195 + }, + { + "epoch": 0.6140754935916595, + "grad_norm": 0.4838317632675171, + "learning_rate": 6.494992080763724e-05, + "loss": 0.0896, + "step": 26196 + }, + { + "epoch": 0.6140989351664645, + "grad_norm": 0.9550197720527649, + "learning_rate": 6.494302203601867e-05, + "loss": 0.0644, + "step": 26197 + }, + { + "epoch": 0.6141223767412695, + "grad_norm": 0.1103113666176796, + "learning_rate": 6.493612345462028e-05, + "loss": 0.0172, + "step": 26198 + }, + { + "epoch": 0.6141458183160745, + "grad_norm": 0.5626174807548523, + "learning_rate": 6.492922506347946e-05, + "loss": 0.639, + "step": 26199 + }, + { + "epoch": 0.6141692598908794, + "grad_norm": 0.14920015633106232, + "learning_rate": 6.492232686263363e-05, + "loss": 0.0178, + "step": 26200 + }, + { + "epoch": 0.6141927014656845, + "grad_norm": 0.4138741195201874, + "learning_rate": 6.491542885212026e-05, + "loss": 0.0376, + "step": 26201 + }, + { + "epoch": 0.6142161430404894, + "grad_norm": 0.2774561643600464, + "learning_rate": 6.490853103197674e-05, + "loss": 0.0285, + "step": 26202 + }, + { + "epoch": 0.6142395846152945, + "grad_norm": 0.3479493260383606, + "learning_rate": 6.490163340224055e-05, + "loss": 0.0656, + "step": 26203 + }, + { + "epoch": 0.6142630261900994, + "grad_norm": 0.6536412835121155, + "learning_rate": 6.489473596294902e-05, + "loss": 0.0903, + "step": 26204 + }, + { + "epoch": 0.6142864677649045, + "grad_norm": 0.7939274311065674, + "learning_rate": 6.48878387141397e-05, + "loss": 0.0379, + "step": 26205 + }, + { + "epoch": 0.6143099093397094, + "grad_norm": 0.3297821581363678, + "learning_rate": 6.488094165584992e-05, + "loss": 0.0564, + "step": 26206 + }, + { + "epoch": 0.6143333509145145, + "grad_norm": 0.17400023341178894, + "learning_rate": 6.487404478811717e-05, + "loss": 0.02, + "step": 26207 + }, + { + "epoch": 0.6143567924893194, + "grad_norm": 0.2349105179309845, + "learning_rate": 6.486714811097883e-05, + "loss": 0.0367, + "step": 26208 + }, + { + "epoch": 0.6143802340641245, + "grad_norm": 0.5380275249481201, + "learning_rate": 6.486025162447233e-05, + "loss": 0.6805, + "step": 26209 + }, + { + "epoch": 0.6144036756389294, + "grad_norm": 0.6637479066848755, + "learning_rate": 6.48533553286351e-05, + "loss": 0.1131, + "step": 26210 + }, + { + "epoch": 0.6144271172137344, + "grad_norm": 0.5505032539367676, + "learning_rate": 6.484645922350454e-05, + "loss": 0.6661, + "step": 26211 + }, + { + "epoch": 0.6144505587885394, + "grad_norm": 0.4858758747577667, + "learning_rate": 6.483956330911808e-05, + "loss": 0.0664, + "step": 26212 + }, + { + "epoch": 0.6144740003633444, + "grad_norm": 0.2496519833803177, + "learning_rate": 6.483266758551315e-05, + "loss": 0.0414, + "step": 26213 + }, + { + "epoch": 0.6144974419381494, + "grad_norm": 0.3852343261241913, + "learning_rate": 6.482577205272715e-05, + "loss": 0.0884, + "step": 26214 + }, + { + "epoch": 0.6145208835129544, + "grad_norm": 0.1594112515449524, + "learning_rate": 6.48188767107975e-05, + "loss": 0.0312, + "step": 26215 + }, + { + "epoch": 0.6145443250877594, + "grad_norm": 0.1017339676618576, + "learning_rate": 6.481198155976157e-05, + "loss": 0.0118, + "step": 26216 + }, + { + "epoch": 0.6145677666625644, + "grad_norm": 0.1288374364376068, + "learning_rate": 6.480508659965688e-05, + "loss": 0.0232, + "step": 26217 + }, + { + "epoch": 0.6145912082373693, + "grad_norm": 0.10790453851222992, + "learning_rate": 6.479819183052076e-05, + "loss": 0.0224, + "step": 26218 + }, + { + "epoch": 0.6146146498121744, + "grad_norm": 0.4603927433490753, + "learning_rate": 6.479129725239061e-05, + "loss": 0.0787, + "step": 26219 + }, + { + "epoch": 0.6146380913869793, + "grad_norm": 0.3358626365661621, + "learning_rate": 6.478440286530392e-05, + "loss": 0.0691, + "step": 26220 + }, + { + "epoch": 0.6146615329617844, + "grad_norm": 0.119735486805439, + "learning_rate": 6.477750866929804e-05, + "loss": 0.0295, + "step": 26221 + }, + { + "epoch": 0.6146849745365893, + "grad_norm": 0.09205815196037292, + "learning_rate": 6.477061466441037e-05, + "loss": 0.011, + "step": 26222 + }, + { + "epoch": 0.6147084161113944, + "grad_norm": 0.4453812539577484, + "learning_rate": 6.476372085067835e-05, + "loss": 0.0582, + "step": 26223 + }, + { + "epoch": 0.6147318576861993, + "grad_norm": 0.6191111207008362, + "learning_rate": 6.475682722813938e-05, + "loss": 0.0606, + "step": 26224 + }, + { + "epoch": 0.6147552992610044, + "grad_norm": 0.22441643476486206, + "learning_rate": 6.474993379683083e-05, + "loss": 0.0446, + "step": 26225 + }, + { + "epoch": 0.6147787408358093, + "grad_norm": 0.11248888075351715, + "learning_rate": 6.474304055679014e-05, + "loss": 0.0251, + "step": 26226 + }, + { + "epoch": 0.6148021824106144, + "grad_norm": 0.313780277967453, + "learning_rate": 6.473614750805472e-05, + "loss": 0.0548, + "step": 26227 + }, + { + "epoch": 0.6148256239854193, + "grad_norm": 0.34743013978004456, + "learning_rate": 6.472925465066191e-05, + "loss": 0.0536, + "step": 26228 + }, + { + "epoch": 0.6148490655602243, + "grad_norm": 0.5131170749664307, + "learning_rate": 6.472236198464919e-05, + "loss": 0.0675, + "step": 26229 + }, + { + "epoch": 0.6148725071350293, + "grad_norm": 0.3765373229980469, + "learning_rate": 6.471546951005392e-05, + "loss": 0.0392, + "step": 26230 + }, + { + "epoch": 0.6148959487098343, + "grad_norm": 0.3147365152835846, + "learning_rate": 6.470857722691352e-05, + "loss": 0.047, + "step": 26231 + }, + { + "epoch": 0.6149193902846393, + "grad_norm": 0.3822207748889923, + "learning_rate": 6.470168513526535e-05, + "loss": 0.0488, + "step": 26232 + }, + { + "epoch": 0.6149428318594443, + "grad_norm": 0.47282639145851135, + "learning_rate": 6.469479323514684e-05, + "loss": 0.0714, + "step": 26233 + }, + { + "epoch": 0.6149662734342494, + "grad_norm": 0.14918296039104462, + "learning_rate": 6.468790152659539e-05, + "loss": 0.0202, + "step": 26234 + }, + { + "epoch": 0.6149897150090543, + "grad_norm": 0.3574073314666748, + "learning_rate": 6.468101000964834e-05, + "loss": 0.0668, + "step": 26235 + }, + { + "epoch": 0.6150131565838594, + "grad_norm": 0.28805357217788696, + "learning_rate": 6.467411868434315e-05, + "loss": 0.0195, + "step": 26236 + }, + { + "epoch": 0.6150365981586643, + "grad_norm": 0.6697033047676086, + "learning_rate": 6.466722755071717e-05, + "loss": 0.1838, + "step": 26237 + }, + { + "epoch": 0.6150600397334693, + "grad_norm": 0.3592507541179657, + "learning_rate": 6.466033660880784e-05, + "loss": 0.0754, + "step": 26238 + }, + { + "epoch": 0.6150834813082743, + "grad_norm": 0.4900186061859131, + "learning_rate": 6.46534458586525e-05, + "loss": 0.0735, + "step": 26239 + }, + { + "epoch": 0.6151069228830793, + "grad_norm": 0.37992820143699646, + "learning_rate": 6.464655530028853e-05, + "loss": 0.0572, + "step": 26240 + }, + { + "epoch": 0.6151303644578843, + "grad_norm": 0.5564332008361816, + "learning_rate": 6.463966493375335e-05, + "loss": 0.1378, + "step": 26241 + }, + { + "epoch": 0.6151538060326893, + "grad_norm": 0.19782425463199615, + "learning_rate": 6.463277475908436e-05, + "loss": 0.0333, + "step": 26242 + }, + { + "epoch": 0.6151772476074943, + "grad_norm": 0.2681114077568054, + "learning_rate": 6.462588477631893e-05, + "loss": 0.0468, + "step": 26243 + }, + { + "epoch": 0.6152006891822993, + "grad_norm": 0.3741176426410675, + "learning_rate": 6.461899498549442e-05, + "loss": 0.0637, + "step": 26244 + }, + { + "epoch": 0.6152241307571042, + "grad_norm": 0.23455381393432617, + "learning_rate": 6.461210538664829e-05, + "loss": 0.035, + "step": 26245 + }, + { + "epoch": 0.6152475723319093, + "grad_norm": 0.11231329292058945, + "learning_rate": 6.460521597981783e-05, + "loss": 0.0238, + "step": 26246 + }, + { + "epoch": 0.6152710139067142, + "grad_norm": 0.5458969473838806, + "learning_rate": 6.459832676504048e-05, + "loss": 0.0745, + "step": 26247 + }, + { + "epoch": 0.6152944554815193, + "grad_norm": 0.4893178939819336, + "learning_rate": 6.459143774235359e-05, + "loss": 0.0504, + "step": 26248 + }, + { + "epoch": 0.6153178970563242, + "grad_norm": 0.6189354658126831, + "learning_rate": 6.458454891179458e-05, + "loss": 0.0439, + "step": 26249 + }, + { + "epoch": 0.6153413386311293, + "grad_norm": 0.527657687664032, + "learning_rate": 6.457766027340078e-05, + "loss": 0.0941, + "step": 26250 + }, + { + "epoch": 0.6153647802059342, + "grad_norm": 0.6861836910247803, + "learning_rate": 6.457077182720958e-05, + "loss": 0.1306, + "step": 26251 + }, + { + "epoch": 0.6153882217807393, + "grad_norm": 0.24486225843429565, + "learning_rate": 6.456388357325841e-05, + "loss": 0.0413, + "step": 26252 + }, + { + "epoch": 0.6154116633555442, + "grad_norm": 0.505419135093689, + "learning_rate": 6.455699551158454e-05, + "loss": 0.0987, + "step": 26253 + }, + { + "epoch": 0.6154351049303493, + "grad_norm": 0.07207003980875015, + "learning_rate": 6.455010764222544e-05, + "loss": 0.0093, + "step": 26254 + }, + { + "epoch": 0.6154585465051542, + "grad_norm": 0.20413175225257874, + "learning_rate": 6.454321996521847e-05, + "loss": 0.0219, + "step": 26255 + }, + { + "epoch": 0.6154819880799592, + "grad_norm": 0.5826854705810547, + "learning_rate": 6.453633248060099e-05, + "loss": 0.0945, + "step": 26256 + }, + { + "epoch": 0.6155054296547642, + "grad_norm": 0.2990938127040863, + "learning_rate": 6.452944518841032e-05, + "loss": 0.0372, + "step": 26257 + }, + { + "epoch": 0.6155288712295692, + "grad_norm": 0.29058393836021423, + "learning_rate": 6.452255808868394e-05, + "loss": 0.0431, + "step": 26258 + }, + { + "epoch": 0.6155523128043742, + "grad_norm": 0.4110240638256073, + "learning_rate": 6.451567118145914e-05, + "loss": 0.05, + "step": 26259 + }, + { + "epoch": 0.6155757543791792, + "grad_norm": 0.07806085795164108, + "learning_rate": 6.450878446677328e-05, + "loss": 0.0121, + "step": 26260 + }, + { + "epoch": 0.6155991959539842, + "grad_norm": 0.38557031750679016, + "learning_rate": 6.450189794466377e-05, + "loss": 0.0837, + "step": 26261 + }, + { + "epoch": 0.6156226375287892, + "grad_norm": 0.4032316207885742, + "learning_rate": 6.449501161516797e-05, + "loss": 0.06, + "step": 26262 + }, + { + "epoch": 0.6156460791035941, + "grad_norm": 0.3337624967098236, + "learning_rate": 6.448812547832322e-05, + "loss": 0.0634, + "step": 26263 + }, + { + "epoch": 0.6156695206783992, + "grad_norm": 0.12637510895729065, + "learning_rate": 6.448123953416691e-05, + "loss": 0.0202, + "step": 26264 + }, + { + "epoch": 0.6156929622532041, + "grad_norm": 0.3823229968547821, + "learning_rate": 6.447435378273634e-05, + "loss": 0.0912, + "step": 26265 + }, + { + "epoch": 0.6157164038280092, + "grad_norm": 0.36423468589782715, + "learning_rate": 6.446746822406899e-05, + "loss": 0.039, + "step": 26266 + }, + { + "epoch": 0.6157398454028141, + "grad_norm": 0.4371761679649353, + "learning_rate": 6.446058285820212e-05, + "loss": 0.0672, + "step": 26267 + }, + { + "epoch": 0.6157632869776192, + "grad_norm": 0.5469321608543396, + "learning_rate": 6.445369768517316e-05, + "loss": 0.0692, + "step": 26268 + }, + { + "epoch": 0.6157867285524241, + "grad_norm": 0.1776985228061676, + "learning_rate": 6.444681270501942e-05, + "loss": 0.0117, + "step": 26269 + }, + { + "epoch": 0.6158101701272292, + "grad_norm": 0.2549796998500824, + "learning_rate": 6.443992791777827e-05, + "loss": 0.0551, + "step": 26270 + }, + { + "epoch": 0.6158336117020341, + "grad_norm": 0.49563339352607727, + "learning_rate": 6.443304332348708e-05, + "loss": 0.1221, + "step": 26271 + }, + { + "epoch": 0.6158570532768392, + "grad_norm": 0.38597366213798523, + "learning_rate": 6.44261589221832e-05, + "loss": 0.0846, + "step": 26272 + }, + { + "epoch": 0.6158804948516441, + "grad_norm": 0.37100374698638916, + "learning_rate": 6.441927471390396e-05, + "loss": 0.092, + "step": 26273 + }, + { + "epoch": 0.6159039364264491, + "grad_norm": 0.3882574141025543, + "learning_rate": 6.441239069868676e-05, + "loss": 0.0405, + "step": 26274 + }, + { + "epoch": 0.6159273780012541, + "grad_norm": 0.5036629438400269, + "learning_rate": 6.440550687656889e-05, + "loss": 0.1111, + "step": 26275 + }, + { + "epoch": 0.6159508195760591, + "grad_norm": 0.17685385048389435, + "learning_rate": 6.439862324758779e-05, + "loss": 0.0254, + "step": 26276 + }, + { + "epoch": 0.6159742611508641, + "grad_norm": 0.6388865113258362, + "learning_rate": 6.43917398117807e-05, + "loss": 0.3948, + "step": 26277 + }, + { + "epoch": 0.6159977027256691, + "grad_norm": 0.745193600654602, + "learning_rate": 6.438485656918509e-05, + "loss": 0.8518, + "step": 26278 + }, + { + "epoch": 0.616021144300474, + "grad_norm": 0.18187165260314941, + "learning_rate": 6.437797351983824e-05, + "loss": 0.0193, + "step": 26279 + }, + { + "epoch": 0.6160445858752791, + "grad_norm": 0.4343693256378174, + "learning_rate": 6.43710906637775e-05, + "loss": 0.0387, + "step": 26280 + }, + { + "epoch": 0.616068027450084, + "grad_norm": 0.5155010223388672, + "learning_rate": 6.436420800104022e-05, + "loss": 0.0808, + "step": 26281 + }, + { + "epoch": 0.6160914690248891, + "grad_norm": 0.8514944911003113, + "learning_rate": 6.435732553166374e-05, + "loss": 0.4908, + "step": 26282 + }, + { + "epoch": 0.616114910599694, + "grad_norm": 0.15299057960510254, + "learning_rate": 6.435044325568543e-05, + "loss": 0.0217, + "step": 26283 + }, + { + "epoch": 0.6161383521744991, + "grad_norm": 0.856839656829834, + "learning_rate": 6.434356117314261e-05, + "loss": 0.5059, + "step": 26284 + }, + { + "epoch": 0.6161617937493041, + "grad_norm": 0.6031740307807922, + "learning_rate": 6.433667928407264e-05, + "loss": 0.0937, + "step": 26285 + }, + { + "epoch": 0.6161852353241091, + "grad_norm": 0.1341046243906021, + "learning_rate": 6.432979758851285e-05, + "loss": 0.0152, + "step": 26286 + }, + { + "epoch": 0.6162086768989141, + "grad_norm": 0.573898196220398, + "learning_rate": 6.432291608650059e-05, + "loss": 0.6535, + "step": 26287 + }, + { + "epoch": 0.6162321184737191, + "grad_norm": 0.3607582449913025, + "learning_rate": 6.431603477807315e-05, + "loss": 0.0546, + "step": 26288 + }, + { + "epoch": 0.6162555600485241, + "grad_norm": 1.0121034383773804, + "learning_rate": 6.430915366326791e-05, + "loss": 0.1627, + "step": 26289 + }, + { + "epoch": 0.616279001623329, + "grad_norm": 0.4402260482311249, + "learning_rate": 6.430227274212223e-05, + "loss": 0.0527, + "step": 26290 + }, + { + "epoch": 0.6163024431981341, + "grad_norm": 0.6446419954299927, + "learning_rate": 6.429539201467343e-05, + "loss": 0.0901, + "step": 26291 + }, + { + "epoch": 0.616325884772939, + "grad_norm": 0.1634494662284851, + "learning_rate": 6.428851148095882e-05, + "loss": 0.0259, + "step": 26292 + }, + { + "epoch": 0.6163493263477441, + "grad_norm": 0.4045301079750061, + "learning_rate": 6.428163114101577e-05, + "loss": 0.0612, + "step": 26293 + }, + { + "epoch": 0.616372767922549, + "grad_norm": 0.6932044625282288, + "learning_rate": 6.427475099488159e-05, + "loss": 0.179, + "step": 26294 + }, + { + "epoch": 0.6163962094973541, + "grad_norm": 0.3001876771450043, + "learning_rate": 6.42678710425936e-05, + "loss": 0.073, + "step": 26295 + }, + { + "epoch": 0.616419651072159, + "grad_norm": 0.5406622886657715, + "learning_rate": 6.426099128418916e-05, + "loss": 0.1299, + "step": 26296 + }, + { + "epoch": 0.6164430926469641, + "grad_norm": 0.23871053755283356, + "learning_rate": 6.425411171970559e-05, + "loss": 0.0419, + "step": 26297 + }, + { + "epoch": 0.616466534221769, + "grad_norm": 0.3129439353942871, + "learning_rate": 6.424723234918019e-05, + "loss": 0.0544, + "step": 26298 + }, + { + "epoch": 0.616489975796574, + "grad_norm": 0.24101346731185913, + "learning_rate": 6.424035317265034e-05, + "loss": 0.0306, + "step": 26299 + }, + { + "epoch": 0.616513417371379, + "grad_norm": 0.16994568705558777, + "learning_rate": 6.423347419015334e-05, + "loss": 0.0446, + "step": 26300 + }, + { + "epoch": 0.616536858946184, + "grad_norm": 0.7445706725120544, + "learning_rate": 6.422659540172646e-05, + "loss": 0.1147, + "step": 26301 + }, + { + "epoch": 0.616560300520989, + "grad_norm": 0.7665511965751648, + "learning_rate": 6.421971680740711e-05, + "loss": 0.1072, + "step": 26302 + }, + { + "epoch": 0.616583742095794, + "grad_norm": 0.2288765162229538, + "learning_rate": 6.42128384072326e-05, + "loss": 0.0307, + "step": 26303 + }, + { + "epoch": 0.616607183670599, + "grad_norm": 0.30800119042396545, + "learning_rate": 6.420596020124023e-05, + "loss": 0.0492, + "step": 26304 + }, + { + "epoch": 0.616630625245404, + "grad_norm": 0.568828821182251, + "learning_rate": 6.419908218946732e-05, + "loss": 0.0911, + "step": 26305 + }, + { + "epoch": 0.616654066820209, + "grad_norm": 0.3983379006385803, + "learning_rate": 6.419220437195122e-05, + "loss": 0.0568, + "step": 26306 + }, + { + "epoch": 0.616677508395014, + "grad_norm": 0.23630324006080627, + "learning_rate": 6.418532674872921e-05, + "loss": 0.0897, + "step": 26307 + }, + { + "epoch": 0.616700949969819, + "grad_norm": 0.14064709842205048, + "learning_rate": 6.417844931983861e-05, + "loss": 0.03, + "step": 26308 + }, + { + "epoch": 0.616724391544624, + "grad_norm": 0.2668825089931488, + "learning_rate": 6.417157208531676e-05, + "loss": 0.0389, + "step": 26309 + }, + { + "epoch": 0.6167478331194289, + "grad_norm": 0.6204602718353271, + "learning_rate": 6.416469504520099e-05, + "loss": 0.1162, + "step": 26310 + }, + { + "epoch": 0.616771274694234, + "grad_norm": 0.4840851128101349, + "learning_rate": 6.415781819952856e-05, + "loss": 0.0691, + "step": 26311 + }, + { + "epoch": 0.6167947162690389, + "grad_norm": 1.004888653755188, + "learning_rate": 6.415094154833684e-05, + "loss": 0.1523, + "step": 26312 + }, + { + "epoch": 0.616818157843844, + "grad_norm": 0.40780144929885864, + "learning_rate": 6.414406509166309e-05, + "loss": 0.0573, + "step": 26313 + }, + { + "epoch": 0.6168415994186489, + "grad_norm": 0.14316391944885254, + "learning_rate": 6.413718882954466e-05, + "loss": 0.0266, + "step": 26314 + }, + { + "epoch": 0.616865040993454, + "grad_norm": 0.4157360792160034, + "learning_rate": 6.413031276201885e-05, + "loss": 0.0904, + "step": 26315 + }, + { + "epoch": 0.6168884825682589, + "grad_norm": 0.9252107739448547, + "learning_rate": 6.412343688912299e-05, + "loss": 0.0996, + "step": 26316 + }, + { + "epoch": 0.616911924143064, + "grad_norm": 0.4247440695762634, + "learning_rate": 6.411656121089437e-05, + "loss": 0.0479, + "step": 26317 + }, + { + "epoch": 0.6169353657178689, + "grad_norm": 0.19485361874103546, + "learning_rate": 6.410968572737028e-05, + "loss": 0.0258, + "step": 26318 + }, + { + "epoch": 0.6169588072926739, + "grad_norm": 0.37040698528289795, + "learning_rate": 6.410281043858806e-05, + "loss": 0.0791, + "step": 26319 + }, + { + "epoch": 0.6169822488674789, + "grad_norm": 0.8161930441856384, + "learning_rate": 6.409593534458499e-05, + "loss": 0.2616, + "step": 26320 + }, + { + "epoch": 0.6170056904422839, + "grad_norm": 0.7121196389198303, + "learning_rate": 6.408906044539839e-05, + "loss": 0.1049, + "step": 26321 + }, + { + "epoch": 0.6170291320170889, + "grad_norm": 0.3754960298538208, + "learning_rate": 6.408218574106557e-05, + "loss": 0.0513, + "step": 26322 + }, + { + "epoch": 0.6170525735918939, + "grad_norm": 0.193800151348114, + "learning_rate": 6.40753112316238e-05, + "loss": 0.0281, + "step": 26323 + }, + { + "epoch": 0.6170760151666989, + "grad_norm": 0.5029227137565613, + "learning_rate": 6.406843691711043e-05, + "loss": 0.1266, + "step": 26324 + }, + { + "epoch": 0.6170994567415039, + "grad_norm": 0.23578034341335297, + "learning_rate": 6.40615627975627e-05, + "loss": 0.0419, + "step": 26325 + }, + { + "epoch": 0.6171228983163088, + "grad_norm": 0.4358397424221039, + "learning_rate": 6.405468887301792e-05, + "loss": 0.0829, + "step": 26326 + }, + { + "epoch": 0.6171463398911139, + "grad_norm": 0.31611207127571106, + "learning_rate": 6.404781514351343e-05, + "loss": 0.0516, + "step": 26327 + }, + { + "epoch": 0.6171697814659188, + "grad_norm": 0.6240947246551514, + "learning_rate": 6.404094160908651e-05, + "loss": 0.0957, + "step": 26328 + }, + { + "epoch": 0.6171932230407239, + "grad_norm": 0.2537287175655365, + "learning_rate": 6.403406826977447e-05, + "loss": 0.042, + "step": 26329 + }, + { + "epoch": 0.6172166646155288, + "grad_norm": 0.6066444516181946, + "learning_rate": 6.402719512561457e-05, + "loss": 0.08, + "step": 26330 + }, + { + "epoch": 0.6172401061903339, + "grad_norm": 0.4792553186416626, + "learning_rate": 6.402032217664413e-05, + "loss": 0.0463, + "step": 26331 + }, + { + "epoch": 0.6172635477651388, + "grad_norm": 0.09123682230710983, + "learning_rate": 6.401344942290044e-05, + "loss": 0.0175, + "step": 26332 + }, + { + "epoch": 0.6172869893399439, + "grad_norm": 0.09754109382629395, + "learning_rate": 6.400657686442077e-05, + "loss": 0.0247, + "step": 26333 + }, + { + "epoch": 0.6173104309147488, + "grad_norm": 0.388476699590683, + "learning_rate": 6.399970450124243e-05, + "loss": 0.0668, + "step": 26334 + }, + { + "epoch": 0.6173338724895538, + "grad_norm": 0.409651517868042, + "learning_rate": 6.399283233340271e-05, + "loss": 0.0798, + "step": 26335 + }, + { + "epoch": 0.6173573140643589, + "grad_norm": 0.31080883741378784, + "learning_rate": 6.398596036093888e-05, + "loss": 0.0426, + "step": 26336 + }, + { + "epoch": 0.6173807556391638, + "grad_norm": 0.1244920939207077, + "learning_rate": 6.397908858388826e-05, + "loss": 0.0326, + "step": 26337 + }, + { + "epoch": 0.6174041972139689, + "grad_norm": 0.6707414984703064, + "learning_rate": 6.397221700228808e-05, + "loss": 0.1353, + "step": 26338 + }, + { + "epoch": 0.6174276387887738, + "grad_norm": 0.43137580156326294, + "learning_rate": 6.39653456161757e-05, + "loss": 0.0725, + "step": 26339 + }, + { + "epoch": 0.6174510803635789, + "grad_norm": 0.6545182466506958, + "learning_rate": 6.395847442558836e-05, + "loss": 0.7467, + "step": 26340 + }, + { + "epoch": 0.6174745219383838, + "grad_norm": 0.12903526425361633, + "learning_rate": 6.395160343056334e-05, + "loss": 0.0325, + "step": 26341 + }, + { + "epoch": 0.6174979635131889, + "grad_norm": 0.3766317665576935, + "learning_rate": 6.394473263113796e-05, + "loss": 0.0308, + "step": 26342 + }, + { + "epoch": 0.6175214050879938, + "grad_norm": 0.1074090376496315, + "learning_rate": 6.393786202734945e-05, + "loss": 0.0215, + "step": 26343 + }, + { + "epoch": 0.6175448466627989, + "grad_norm": 0.694389283657074, + "learning_rate": 6.393099161923514e-05, + "loss": 0.0873, + "step": 26344 + }, + { + "epoch": 0.6175682882376038, + "grad_norm": 0.1817886084318161, + "learning_rate": 6.392412140683227e-05, + "loss": 0.0208, + "step": 26345 + }, + { + "epoch": 0.6175917298124088, + "grad_norm": 0.810933530330658, + "learning_rate": 6.391725139017813e-05, + "loss": 0.0876, + "step": 26346 + }, + { + "epoch": 0.6176151713872138, + "grad_norm": 0.46180808544158936, + "learning_rate": 6.391038156930999e-05, + "loss": 0.0749, + "step": 26347 + }, + { + "epoch": 0.6176386129620188, + "grad_norm": 0.49738332629203796, + "learning_rate": 6.390351194426517e-05, + "loss": 0.1143, + "step": 26348 + }, + { + "epoch": 0.6176620545368238, + "grad_norm": 0.9602863788604736, + "learning_rate": 6.389664251508085e-05, + "loss": 0.102, + "step": 26349 + }, + { + "epoch": 0.6176854961116288, + "grad_norm": 0.49260807037353516, + "learning_rate": 6.388977328179438e-05, + "loss": 0.4406, + "step": 26350 + }, + { + "epoch": 0.6177089376864338, + "grad_norm": 0.4502430260181427, + "learning_rate": 6.388290424444303e-05, + "loss": 0.5335, + "step": 26351 + }, + { + "epoch": 0.6177323792612388, + "grad_norm": 0.09306270629167557, + "learning_rate": 6.387603540306406e-05, + "loss": 0.0122, + "step": 26352 + }, + { + "epoch": 0.6177558208360437, + "grad_norm": 0.6135888695716858, + "learning_rate": 6.386916675769472e-05, + "loss": 0.1082, + "step": 26353 + }, + { + "epoch": 0.6177792624108488, + "grad_norm": 0.32565027475357056, + "learning_rate": 6.386229830837233e-05, + "loss": 0.057, + "step": 26354 + }, + { + "epoch": 0.6178027039856537, + "grad_norm": 0.36555084586143494, + "learning_rate": 6.385543005513409e-05, + "loss": 0.0561, + "step": 26355 + }, + { + "epoch": 0.6178261455604588, + "grad_norm": 0.4515480399131775, + "learning_rate": 6.384856199801733e-05, + "loss": 0.0884, + "step": 26356 + }, + { + "epoch": 0.6178495871352637, + "grad_norm": 0.5435884594917297, + "learning_rate": 6.384169413705929e-05, + "loss": 0.1431, + "step": 26357 + }, + { + "epoch": 0.6178730287100688, + "grad_norm": 0.382502943277359, + "learning_rate": 6.383482647229721e-05, + "loss": 0.0555, + "step": 26358 + }, + { + "epoch": 0.6178964702848737, + "grad_norm": 0.35288071632385254, + "learning_rate": 6.38279590037684e-05, + "loss": 0.0487, + "step": 26359 + }, + { + "epoch": 0.6179199118596788, + "grad_norm": 0.3806936740875244, + "learning_rate": 6.38210917315101e-05, + "loss": 0.0627, + "step": 26360 + }, + { + "epoch": 0.6179433534344837, + "grad_norm": 0.5340033173561096, + "learning_rate": 6.381422465555955e-05, + "loss": 0.0979, + "step": 26361 + }, + { + "epoch": 0.6179667950092887, + "grad_norm": 0.26908835768699646, + "learning_rate": 6.380735777595404e-05, + "loss": 0.0443, + "step": 26362 + }, + { + "epoch": 0.6179902365840937, + "grad_norm": 0.3381352126598358, + "learning_rate": 6.380049109273083e-05, + "loss": 0.0624, + "step": 26363 + }, + { + "epoch": 0.6180136781588987, + "grad_norm": 0.19190971553325653, + "learning_rate": 6.379362460592719e-05, + "loss": 0.026, + "step": 26364 + }, + { + "epoch": 0.6180371197337037, + "grad_norm": 0.5314721465110779, + "learning_rate": 6.378675831558035e-05, + "loss": 0.0591, + "step": 26365 + }, + { + "epoch": 0.6180605613085087, + "grad_norm": 0.6012066006660461, + "learning_rate": 6.377989222172757e-05, + "loss": 0.0814, + "step": 26366 + }, + { + "epoch": 0.6180840028833137, + "grad_norm": 0.5190572738647461, + "learning_rate": 6.377302632440614e-05, + "loss": 0.6932, + "step": 26367 + }, + { + "epoch": 0.6181074444581187, + "grad_norm": 0.4737037718296051, + "learning_rate": 6.376616062365325e-05, + "loss": 0.1359, + "step": 26368 + }, + { + "epoch": 0.6181308860329237, + "grad_norm": 0.45876070857048035, + "learning_rate": 6.375929511950622e-05, + "loss": 0.1092, + "step": 26369 + }, + { + "epoch": 0.6181543276077287, + "grad_norm": 0.751839816570282, + "learning_rate": 6.375242981200227e-05, + "loss": 0.0867, + "step": 26370 + }, + { + "epoch": 0.6181777691825336, + "grad_norm": 0.1390056610107422, + "learning_rate": 6.374556470117863e-05, + "loss": 0.0215, + "step": 26371 + }, + { + "epoch": 0.6182012107573387, + "grad_norm": 0.23294243216514587, + "learning_rate": 6.373869978707261e-05, + "loss": 0.0224, + "step": 26372 + }, + { + "epoch": 0.6182246523321436, + "grad_norm": 0.398162305355072, + "learning_rate": 6.373183506972142e-05, + "loss": 0.082, + "step": 26373 + }, + { + "epoch": 0.6182480939069487, + "grad_norm": 0.2361094206571579, + "learning_rate": 6.372497054916227e-05, + "loss": 0.0489, + "step": 26374 + }, + { + "epoch": 0.6182715354817536, + "grad_norm": 1.0866725444793701, + "learning_rate": 6.371810622543246e-05, + "loss": 0.0996, + "step": 26375 + }, + { + "epoch": 0.6182949770565587, + "grad_norm": 0.08994098007678986, + "learning_rate": 6.371124209856927e-05, + "loss": 0.0226, + "step": 26376 + }, + { + "epoch": 0.6183184186313636, + "grad_norm": 0.2672719657421112, + "learning_rate": 6.370437816860988e-05, + "loss": 0.0539, + "step": 26377 + }, + { + "epoch": 0.6183418602061687, + "grad_norm": 0.6563417315483093, + "learning_rate": 6.369751443559153e-05, + "loss": 0.0981, + "step": 26378 + }, + { + "epoch": 0.6183653017809736, + "grad_norm": 0.24170811474323273, + "learning_rate": 6.369065089955152e-05, + "loss": 0.0352, + "step": 26379 + }, + { + "epoch": 0.6183887433557786, + "grad_norm": 0.4708726704120636, + "learning_rate": 6.368378756052707e-05, + "loss": 0.089, + "step": 26380 + }, + { + "epoch": 0.6184121849305836, + "grad_norm": 0.36875247955322266, + "learning_rate": 6.367692441855537e-05, + "loss": 0.0728, + "step": 26381 + }, + { + "epoch": 0.6184356265053886, + "grad_norm": 0.15133799612522125, + "learning_rate": 6.367006147367373e-05, + "loss": 0.0335, + "step": 26382 + }, + { + "epoch": 0.6184590680801936, + "grad_norm": 0.5010063052177429, + "learning_rate": 6.366319872591935e-05, + "loss": 0.0927, + "step": 26383 + }, + { + "epoch": 0.6184825096549986, + "grad_norm": 0.5681599378585815, + "learning_rate": 6.365633617532945e-05, + "loss": 0.1322, + "step": 26384 + }, + { + "epoch": 0.6185059512298036, + "grad_norm": 0.47509631514549255, + "learning_rate": 6.364947382194133e-05, + "loss": 0.4366, + "step": 26385 + }, + { + "epoch": 0.6185293928046086, + "grad_norm": 0.37597861886024475, + "learning_rate": 6.364261166579214e-05, + "loss": 0.0415, + "step": 26386 + }, + { + "epoch": 0.6185528343794137, + "grad_norm": 0.613855242729187, + "learning_rate": 6.36357497069192e-05, + "loss": 0.51, + "step": 26387 + }, + { + "epoch": 0.6185762759542186, + "grad_norm": 0.4833737909793854, + "learning_rate": 6.362888794535968e-05, + "loss": 0.5837, + "step": 26388 + }, + { + "epoch": 0.6185997175290237, + "grad_norm": 0.45076364278793335, + "learning_rate": 6.362202638115087e-05, + "loss": 0.1355, + "step": 26389 + }, + { + "epoch": 0.6186231591038286, + "grad_norm": 0.38032740354537964, + "learning_rate": 6.361516501432995e-05, + "loss": 0.483, + "step": 26390 + }, + { + "epoch": 0.6186466006786336, + "grad_norm": 0.5852817296981812, + "learning_rate": 6.360830384493416e-05, + "loss": 0.801, + "step": 26391 + }, + { + "epoch": 0.6186700422534386, + "grad_norm": 0.5234682559967041, + "learning_rate": 6.360144287300075e-05, + "loss": 0.0529, + "step": 26392 + }, + { + "epoch": 0.6186934838282436, + "grad_norm": 0.7066068053245544, + "learning_rate": 6.359458209856693e-05, + "loss": 0.0764, + "step": 26393 + }, + { + "epoch": 0.6187169254030486, + "grad_norm": 0.14198537170886993, + "learning_rate": 6.358772152166992e-05, + "loss": 0.0317, + "step": 26394 + }, + { + "epoch": 0.6187403669778536, + "grad_norm": 0.495910108089447, + "learning_rate": 6.358086114234698e-05, + "loss": 0.101, + "step": 26395 + }, + { + "epoch": 0.6187638085526586, + "grad_norm": 0.8817992210388184, + "learning_rate": 6.357400096063528e-05, + "loss": 0.1634, + "step": 26396 + }, + { + "epoch": 0.6187872501274636, + "grad_norm": 0.1762615293264389, + "learning_rate": 6.356714097657211e-05, + "loss": 0.0389, + "step": 26397 + }, + { + "epoch": 0.6188106917022685, + "grad_norm": 0.5339661836624146, + "learning_rate": 6.356028119019466e-05, + "loss": 0.0915, + "step": 26398 + }, + { + "epoch": 0.6188341332770736, + "grad_norm": 0.49386465549468994, + "learning_rate": 6.355342160154009e-05, + "loss": 0.0721, + "step": 26399 + }, + { + "epoch": 0.6188575748518785, + "grad_norm": 0.7047786712646484, + "learning_rate": 6.35465622106457e-05, + "loss": 0.1325, + "step": 26400 + }, + { + "epoch": 0.6188810164266836, + "grad_norm": 0.40007105469703674, + "learning_rate": 6.353970301754872e-05, + "loss": 0.0555, + "step": 26401 + }, + { + "epoch": 0.6189044580014885, + "grad_norm": 0.14202895760536194, + "learning_rate": 6.353284402228635e-05, + "loss": 0.0253, + "step": 26402 + }, + { + "epoch": 0.6189278995762936, + "grad_norm": 0.5744585990905762, + "learning_rate": 6.352598522489575e-05, + "loss": 0.5987, + "step": 26403 + }, + { + "epoch": 0.6189513411510985, + "grad_norm": 0.3239660859107971, + "learning_rate": 6.351912662541419e-05, + "loss": 0.0423, + "step": 26404 + }, + { + "epoch": 0.6189747827259036, + "grad_norm": 0.5262115001678467, + "learning_rate": 6.35122682238789e-05, + "loss": 0.0927, + "step": 26405 + }, + { + "epoch": 0.6189982243007085, + "grad_norm": 0.6434187293052673, + "learning_rate": 6.350541002032704e-05, + "loss": 0.0895, + "step": 26406 + }, + { + "epoch": 0.6190216658755135, + "grad_norm": 0.28463056683540344, + "learning_rate": 6.349855201479586e-05, + "loss": 0.031, + "step": 26407 + }, + { + "epoch": 0.6190451074503185, + "grad_norm": 0.3821094334125519, + "learning_rate": 6.349169420732259e-05, + "loss": 0.0798, + "step": 26408 + }, + { + "epoch": 0.6190685490251235, + "grad_norm": 0.05330096557736397, + "learning_rate": 6.348483659794438e-05, + "loss": 0.0061, + "step": 26409 + }, + { + "epoch": 0.6190919905999285, + "grad_norm": 0.11064455658197403, + "learning_rate": 6.347797918669848e-05, + "loss": 0.0177, + "step": 26410 + }, + { + "epoch": 0.6191154321747335, + "grad_norm": 0.29545357823371887, + "learning_rate": 6.347112197362206e-05, + "loss": 0.0176, + "step": 26411 + }, + { + "epoch": 0.6191388737495385, + "grad_norm": 0.40241941809654236, + "learning_rate": 6.34642649587524e-05, + "loss": 0.0814, + "step": 26412 + }, + { + "epoch": 0.6191623153243435, + "grad_norm": 0.2595226466655731, + "learning_rate": 6.345740814212665e-05, + "loss": 0.0464, + "step": 26413 + }, + { + "epoch": 0.6191857568991485, + "grad_norm": 0.5005647540092468, + "learning_rate": 6.345055152378206e-05, + "loss": 0.0912, + "step": 26414 + }, + { + "epoch": 0.6192091984739535, + "grad_norm": 0.19003860652446747, + "learning_rate": 6.344369510375579e-05, + "loss": 0.0258, + "step": 26415 + }, + { + "epoch": 0.6192326400487584, + "grad_norm": 0.43356549739837646, + "learning_rate": 6.343683888208502e-05, + "loss": 0.4335, + "step": 26416 + }, + { + "epoch": 0.6192560816235635, + "grad_norm": 0.41539573669433594, + "learning_rate": 6.342998285880704e-05, + "loss": 0.0558, + "step": 26417 + }, + { + "epoch": 0.6192795231983684, + "grad_norm": 0.620751142501831, + "learning_rate": 6.3423127033959e-05, + "loss": 0.1213, + "step": 26418 + }, + { + "epoch": 0.6193029647731735, + "grad_norm": 0.42402052879333496, + "learning_rate": 6.341627140757807e-05, + "loss": 0.0614, + "step": 26419 + }, + { + "epoch": 0.6193264063479784, + "grad_norm": 0.3811595141887665, + "learning_rate": 6.340941597970152e-05, + "loss": 0.0509, + "step": 26420 + }, + { + "epoch": 0.6193498479227835, + "grad_norm": 0.7467316389083862, + "learning_rate": 6.340256075036648e-05, + "loss": 0.1885, + "step": 26421 + }, + { + "epoch": 0.6193732894975884, + "grad_norm": 0.10205339640378952, + "learning_rate": 6.339570571961016e-05, + "loss": 0.0214, + "step": 26422 + }, + { + "epoch": 0.6193967310723935, + "grad_norm": 0.3910634517669678, + "learning_rate": 6.338885088746976e-05, + "loss": 0.0637, + "step": 26423 + }, + { + "epoch": 0.6194201726471984, + "grad_norm": 1.0657720565795898, + "learning_rate": 6.338199625398252e-05, + "loss": 0.2509, + "step": 26424 + }, + { + "epoch": 0.6194436142220034, + "grad_norm": 0.5489380359649658, + "learning_rate": 6.33751418191856e-05, + "loss": 0.0672, + "step": 26425 + }, + { + "epoch": 0.6194670557968084, + "grad_norm": 0.3574250638484955, + "learning_rate": 6.336828758311617e-05, + "loss": 0.3716, + "step": 26426 + }, + { + "epoch": 0.6194904973716134, + "grad_norm": 0.16867128014564514, + "learning_rate": 6.336143354581145e-05, + "loss": 0.0359, + "step": 26427 + }, + { + "epoch": 0.6195139389464184, + "grad_norm": 0.20170247554779053, + "learning_rate": 6.335457970730865e-05, + "loss": 0.0531, + "step": 26428 + }, + { + "epoch": 0.6195373805212234, + "grad_norm": 0.4629872739315033, + "learning_rate": 6.334772606764489e-05, + "loss": 0.0766, + "step": 26429 + }, + { + "epoch": 0.6195608220960284, + "grad_norm": 0.5153535604476929, + "learning_rate": 6.334087262685741e-05, + "loss": 0.0724, + "step": 26430 + }, + { + "epoch": 0.6195842636708334, + "grad_norm": 0.3268885016441345, + "learning_rate": 6.333401938498339e-05, + "loss": 0.0518, + "step": 26431 + }, + { + "epoch": 0.6196077052456384, + "grad_norm": 0.3880516290664673, + "learning_rate": 6.332716634206001e-05, + "loss": 0.055, + "step": 26432 + }, + { + "epoch": 0.6196311468204434, + "grad_norm": 0.29437336325645447, + "learning_rate": 6.332031349812448e-05, + "loss": 0.0402, + "step": 26433 + }, + { + "epoch": 0.6196545883952483, + "grad_norm": 0.21022309362888336, + "learning_rate": 6.331346085321393e-05, + "loss": 0.0305, + "step": 26434 + }, + { + "epoch": 0.6196780299700534, + "grad_norm": 0.3515067398548126, + "learning_rate": 6.330660840736556e-05, + "loss": 0.0471, + "step": 26435 + }, + { + "epoch": 0.6197014715448583, + "grad_norm": 0.18798592686653137, + "learning_rate": 6.329975616061658e-05, + "loss": 0.0278, + "step": 26436 + }, + { + "epoch": 0.6197249131196634, + "grad_norm": 0.44765788316726685, + "learning_rate": 6.329290411300417e-05, + "loss": 0.0734, + "step": 26437 + }, + { + "epoch": 0.6197483546944684, + "grad_norm": 0.5218333005905151, + "learning_rate": 6.328605226456548e-05, + "loss": 0.5949, + "step": 26438 + }, + { + "epoch": 0.6197717962692734, + "grad_norm": 0.6306186318397522, + "learning_rate": 6.327920061533771e-05, + "loss": 0.0912, + "step": 26439 + }, + { + "epoch": 0.6197952378440784, + "grad_norm": 0.43694910407066345, + "learning_rate": 6.327234916535803e-05, + "loss": 0.0777, + "step": 26440 + }, + { + "epoch": 0.6198186794188834, + "grad_norm": 0.5565556287765503, + "learning_rate": 6.326549791466361e-05, + "loss": 0.4695, + "step": 26441 + }, + { + "epoch": 0.6198421209936884, + "grad_norm": 0.31461966037750244, + "learning_rate": 6.325864686329163e-05, + "loss": 0.0733, + "step": 26442 + }, + { + "epoch": 0.6198655625684933, + "grad_norm": 0.25176018476486206, + "learning_rate": 6.325179601127928e-05, + "loss": 0.032, + "step": 26443 + }, + { + "epoch": 0.6198890041432984, + "grad_norm": 0.5684336423873901, + "learning_rate": 6.32449453586637e-05, + "loss": 0.1159, + "step": 26444 + }, + { + "epoch": 0.6199124457181033, + "grad_norm": 0.580595076084137, + "learning_rate": 6.323809490548211e-05, + "loss": 0.1188, + "step": 26445 + }, + { + "epoch": 0.6199358872929084, + "grad_norm": 0.46754640340805054, + "learning_rate": 6.323124465177164e-05, + "loss": 0.0652, + "step": 26446 + }, + { + "epoch": 0.6199593288677133, + "grad_norm": 0.5048360824584961, + "learning_rate": 6.322439459756941e-05, + "loss": 0.5484, + "step": 26447 + }, + { + "epoch": 0.6199827704425184, + "grad_norm": 0.32747724652290344, + "learning_rate": 6.32175447429127e-05, + "loss": 0.0599, + "step": 26448 + }, + { + "epoch": 0.6200062120173233, + "grad_norm": 0.36333775520324707, + "learning_rate": 6.321069508783863e-05, + "loss": 0.0336, + "step": 26449 + }, + { + "epoch": 0.6200296535921284, + "grad_norm": 0.41157975792884827, + "learning_rate": 6.320384563238436e-05, + "loss": 0.0361, + "step": 26450 + }, + { + "epoch": 0.6200530951669333, + "grad_norm": 0.46900883316993713, + "learning_rate": 6.319699637658705e-05, + "loss": 0.0465, + "step": 26451 + }, + { + "epoch": 0.6200765367417383, + "grad_norm": 0.9254416823387146, + "learning_rate": 6.319014732048388e-05, + "loss": 0.1718, + "step": 26452 + }, + { + "epoch": 0.6200999783165433, + "grad_norm": 0.5636146664619446, + "learning_rate": 6.318329846411202e-05, + "loss": 0.7036, + "step": 26453 + }, + { + "epoch": 0.6201234198913483, + "grad_norm": 0.1621074229478836, + "learning_rate": 6.317644980750859e-05, + "loss": 0.0421, + "step": 26454 + }, + { + "epoch": 0.6201468614661533, + "grad_norm": 1.0071468353271484, + "learning_rate": 6.31696013507108e-05, + "loss": 0.1225, + "step": 26455 + }, + { + "epoch": 0.6201703030409583, + "grad_norm": 0.37421727180480957, + "learning_rate": 6.316275309375579e-05, + "loss": 0.0652, + "step": 26456 + }, + { + "epoch": 0.6201937446157633, + "grad_norm": 0.7327738404273987, + "learning_rate": 6.315590503668069e-05, + "loss": 0.1259, + "step": 26457 + }, + { + "epoch": 0.6202171861905683, + "grad_norm": 0.2990381717681885, + "learning_rate": 6.314905717952271e-05, + "loss": 0.0504, + "step": 26458 + }, + { + "epoch": 0.6202406277653733, + "grad_norm": 0.5339720845222473, + "learning_rate": 6.314220952231895e-05, + "loss": 0.0973, + "step": 26459 + }, + { + "epoch": 0.6202640693401783, + "grad_norm": 0.474714994430542, + "learning_rate": 6.313536206510661e-05, + "loss": 0.0928, + "step": 26460 + }, + { + "epoch": 0.6202875109149832, + "grad_norm": 0.3921850323677063, + "learning_rate": 6.312851480792285e-05, + "loss": 0.0738, + "step": 26461 + }, + { + "epoch": 0.6203109524897883, + "grad_norm": 0.3120369613170624, + "learning_rate": 6.31216677508048e-05, + "loss": 0.077, + "step": 26462 + }, + { + "epoch": 0.6203343940645932, + "grad_norm": 0.39129433035850525, + "learning_rate": 6.311482089378962e-05, + "loss": 0.1104, + "step": 26463 + }, + { + "epoch": 0.6203578356393983, + "grad_norm": 0.5355682373046875, + "learning_rate": 6.310797423691445e-05, + "loss": 0.8087, + "step": 26464 + }, + { + "epoch": 0.6203812772142032, + "grad_norm": 0.12462306022644043, + "learning_rate": 6.310112778021647e-05, + "loss": 0.016, + "step": 26465 + }, + { + "epoch": 0.6204047187890083, + "grad_norm": 0.4216105341911316, + "learning_rate": 6.30942815237328e-05, + "loss": 0.0527, + "step": 26466 + }, + { + "epoch": 0.6204281603638132, + "grad_norm": 0.09897881746292114, + "learning_rate": 6.308743546750057e-05, + "loss": 0.0231, + "step": 26467 + }, + { + "epoch": 0.6204516019386183, + "grad_norm": 0.6957291960716248, + "learning_rate": 6.308058961155698e-05, + "loss": 0.6311, + "step": 26468 + }, + { + "epoch": 0.6204750435134232, + "grad_norm": 0.5825012922286987, + "learning_rate": 6.307374395593914e-05, + "loss": 0.0943, + "step": 26469 + }, + { + "epoch": 0.6204984850882282, + "grad_norm": 0.09130793809890747, + "learning_rate": 6.306689850068421e-05, + "loss": 0.0186, + "step": 26470 + }, + { + "epoch": 0.6205219266630332, + "grad_norm": 0.16196832060813904, + "learning_rate": 6.306005324582933e-05, + "loss": 0.0222, + "step": 26471 + }, + { + "epoch": 0.6205453682378382, + "grad_norm": 0.25833672285079956, + "learning_rate": 6.30532081914116e-05, + "loss": 0.0437, + "step": 26472 + }, + { + "epoch": 0.6205688098126432, + "grad_norm": 0.3633694052696228, + "learning_rate": 6.304636333746823e-05, + "loss": 0.035, + "step": 26473 + }, + { + "epoch": 0.6205922513874482, + "grad_norm": 0.6845635175704956, + "learning_rate": 6.303951868403633e-05, + "loss": 0.0459, + "step": 26474 + }, + { + "epoch": 0.6206156929622532, + "grad_norm": 0.3500077724456787, + "learning_rate": 6.303267423115305e-05, + "loss": 0.0558, + "step": 26475 + }, + { + "epoch": 0.6206391345370582, + "grad_norm": 0.6294918060302734, + "learning_rate": 6.302582997885551e-05, + "loss": 0.1322, + "step": 26476 + }, + { + "epoch": 0.6206625761118632, + "grad_norm": 0.5191760659217834, + "learning_rate": 6.301898592718086e-05, + "loss": 0.0901, + "step": 26477 + }, + { + "epoch": 0.6206860176866682, + "grad_norm": 0.37712523341178894, + "learning_rate": 6.301214207616623e-05, + "loss": 0.0907, + "step": 26478 + }, + { + "epoch": 0.6207094592614731, + "grad_norm": 0.5806485414505005, + "learning_rate": 6.300529842584877e-05, + "loss": 0.0927, + "step": 26479 + }, + { + "epoch": 0.6207329008362782, + "grad_norm": 0.5493894219398499, + "learning_rate": 6.29984549762656e-05, + "loss": 0.0285, + "step": 26480 + }, + { + "epoch": 0.6207563424110831, + "grad_norm": 0.44856131076812744, + "learning_rate": 6.299161172745384e-05, + "loss": 0.1157, + "step": 26481 + }, + { + "epoch": 0.6207797839858882, + "grad_norm": 0.3289484977722168, + "learning_rate": 6.298476867945062e-05, + "loss": 0.0576, + "step": 26482 + }, + { + "epoch": 0.6208032255606931, + "grad_norm": 0.37420347332954407, + "learning_rate": 6.297792583229312e-05, + "loss": 0.5203, + "step": 26483 + }, + { + "epoch": 0.6208266671354982, + "grad_norm": 0.7022780179977417, + "learning_rate": 6.297108318601838e-05, + "loss": 0.1387, + "step": 26484 + }, + { + "epoch": 0.6208501087103031, + "grad_norm": 0.41942423582077026, + "learning_rate": 6.296424074066365e-05, + "loss": 0.1369, + "step": 26485 + }, + { + "epoch": 0.6208735502851082, + "grad_norm": 0.3170134127140045, + "learning_rate": 6.295739849626595e-05, + "loss": 0.0603, + "step": 26486 + }, + { + "epoch": 0.6208969918599131, + "grad_norm": 0.43573155999183655, + "learning_rate": 6.295055645286248e-05, + "loss": 0.1064, + "step": 26487 + }, + { + "epoch": 0.6209204334347181, + "grad_norm": 0.37958717346191406, + "learning_rate": 6.294371461049031e-05, + "loss": 0.0765, + "step": 26488 + }, + { + "epoch": 0.6209438750095232, + "grad_norm": 0.7260890007019043, + "learning_rate": 6.293687296918657e-05, + "loss": 0.1476, + "step": 26489 + }, + { + "epoch": 0.6209673165843281, + "grad_norm": 0.5618662238121033, + "learning_rate": 6.293003152898844e-05, + "loss": 0.1479, + "step": 26490 + }, + { + "epoch": 0.6209907581591332, + "grad_norm": 0.1050349697470665, + "learning_rate": 6.2923190289933e-05, + "loss": 0.0153, + "step": 26491 + }, + { + "epoch": 0.6210141997339381, + "grad_norm": 0.34525778889656067, + "learning_rate": 6.291634925205733e-05, + "loss": 0.0622, + "step": 26492 + }, + { + "epoch": 0.6210376413087432, + "grad_norm": 0.20764008164405823, + "learning_rate": 6.290950841539864e-05, + "loss": 0.0273, + "step": 26493 + }, + { + "epoch": 0.6210610828835481, + "grad_norm": 0.37815606594085693, + "learning_rate": 6.290266777999398e-05, + "loss": 0.047, + "step": 26494 + }, + { + "epoch": 0.6210845244583532, + "grad_norm": 0.43671759963035583, + "learning_rate": 6.289582734588046e-05, + "loss": 0.07, + "step": 26495 + }, + { + "epoch": 0.6211079660331581, + "grad_norm": 0.20912592113018036, + "learning_rate": 6.288898711309523e-05, + "loss": 0.0384, + "step": 26496 + }, + { + "epoch": 0.6211314076079631, + "grad_norm": 0.1432778537273407, + "learning_rate": 6.288214708167542e-05, + "loss": 0.0238, + "step": 26497 + }, + { + "epoch": 0.6211548491827681, + "grad_norm": 0.40024739503860474, + "learning_rate": 6.287530725165813e-05, + "loss": 0.0786, + "step": 26498 + }, + { + "epoch": 0.6211782907575731, + "grad_norm": 0.4884229004383087, + "learning_rate": 6.286846762308046e-05, + "loss": 0.0629, + "step": 26499 + }, + { + "epoch": 0.6212017323323781, + "grad_norm": 0.5966379046440125, + "learning_rate": 6.286162819597952e-05, + "loss": 0.1656, + "step": 26500 + }, + { + "epoch": 0.6212251739071831, + "grad_norm": 0.30012911558151245, + "learning_rate": 6.285478897039245e-05, + "loss": 0.0399, + "step": 26501 + }, + { + "epoch": 0.6212486154819881, + "grad_norm": 0.3738211393356323, + "learning_rate": 6.284794994635631e-05, + "loss": 0.0471, + "step": 26502 + }, + { + "epoch": 0.6212720570567931, + "grad_norm": 0.07463362067937851, + "learning_rate": 6.284111112390827e-05, + "loss": 0.0052, + "step": 26503 + }, + { + "epoch": 0.621295498631598, + "grad_norm": 0.36667338013648987, + "learning_rate": 6.28342725030854e-05, + "loss": 0.0585, + "step": 26504 + }, + { + "epoch": 0.6213189402064031, + "grad_norm": 0.5267511010169983, + "learning_rate": 6.282743408392478e-05, + "loss": 0.7922, + "step": 26505 + }, + { + "epoch": 0.621342381781208, + "grad_norm": 0.3021036684513092, + "learning_rate": 6.282059586646358e-05, + "loss": 0.032, + "step": 26506 + }, + { + "epoch": 0.6213658233560131, + "grad_norm": 0.2196028232574463, + "learning_rate": 6.281375785073885e-05, + "loss": 0.0303, + "step": 26507 + }, + { + "epoch": 0.621389264930818, + "grad_norm": 0.3129883408546448, + "learning_rate": 6.280692003678772e-05, + "loss": 0.0826, + "step": 26508 + }, + { + "epoch": 0.6214127065056231, + "grad_norm": 0.6441774368286133, + "learning_rate": 6.280008242464728e-05, + "loss": 0.0699, + "step": 26509 + }, + { + "epoch": 0.621436148080428, + "grad_norm": 0.5498372316360474, + "learning_rate": 6.279324501435466e-05, + "loss": 0.6142, + "step": 26510 + }, + { + "epoch": 0.6214595896552331, + "grad_norm": 0.1968003660440445, + "learning_rate": 6.278640780594693e-05, + "loss": 0.0523, + "step": 26511 + }, + { + "epoch": 0.621483031230038, + "grad_norm": 0.5163787603378296, + "learning_rate": 6.277957079946121e-05, + "loss": 0.4631, + "step": 26512 + }, + { + "epoch": 0.6215064728048431, + "grad_norm": 0.3071964681148529, + "learning_rate": 6.277273399493458e-05, + "loss": 0.0414, + "step": 26513 + }, + { + "epoch": 0.621529914379648, + "grad_norm": 0.23421859741210938, + "learning_rate": 6.276589739240413e-05, + "loss": 0.0578, + "step": 26514 + }, + { + "epoch": 0.621553355954453, + "grad_norm": 0.2144661247730255, + "learning_rate": 6.275906099190699e-05, + "loss": 0.0142, + "step": 26515 + }, + { + "epoch": 0.621576797529258, + "grad_norm": 0.5516716837882996, + "learning_rate": 6.275222479348022e-05, + "loss": 0.0948, + "step": 26516 + }, + { + "epoch": 0.621600239104063, + "grad_norm": 0.434087872505188, + "learning_rate": 6.274538879716091e-05, + "loss": 0.0854, + "step": 26517 + }, + { + "epoch": 0.621623680678868, + "grad_norm": 0.2709273099899292, + "learning_rate": 6.273855300298618e-05, + "loss": 0.034, + "step": 26518 + }, + { + "epoch": 0.621647122253673, + "grad_norm": 0.5252123475074768, + "learning_rate": 6.273171741099313e-05, + "loss": 0.0592, + "step": 26519 + }, + { + "epoch": 0.621670563828478, + "grad_norm": 0.9749188423156738, + "learning_rate": 6.272488202121878e-05, + "loss": 0.1848, + "step": 26520 + }, + { + "epoch": 0.621694005403283, + "grad_norm": 0.0860094502568245, + "learning_rate": 6.271804683370028e-05, + "loss": 0.0115, + "step": 26521 + }, + { + "epoch": 0.621717446978088, + "grad_norm": 0.2414400577545166, + "learning_rate": 6.271121184847473e-05, + "loss": 0.0526, + "step": 26522 + }, + { + "epoch": 0.621740888552893, + "grad_norm": 0.12413597106933594, + "learning_rate": 6.27043770655792e-05, + "loss": 0.0347, + "step": 26523 + }, + { + "epoch": 0.6217643301276979, + "grad_norm": 0.29988425970077515, + "learning_rate": 6.269754248505074e-05, + "loss": 0.0388, + "step": 26524 + }, + { + "epoch": 0.621787771702503, + "grad_norm": 0.305421382188797, + "learning_rate": 6.269070810692649e-05, + "loss": 0.0462, + "step": 26525 + }, + { + "epoch": 0.6218112132773079, + "grad_norm": 0.6436859369277954, + "learning_rate": 6.26838739312435e-05, + "loss": 0.0985, + "step": 26526 + }, + { + "epoch": 0.621834654852113, + "grad_norm": 0.13023243844509125, + "learning_rate": 6.267703995803884e-05, + "loss": 0.0239, + "step": 26527 + }, + { + "epoch": 0.6218580964269179, + "grad_norm": 0.3742404282093048, + "learning_rate": 6.267020618734965e-05, + "loss": 0.0817, + "step": 26528 + }, + { + "epoch": 0.621881538001723, + "grad_norm": 0.20116077363491058, + "learning_rate": 6.266337261921294e-05, + "loss": 0.0257, + "step": 26529 + }, + { + "epoch": 0.6219049795765279, + "grad_norm": 0.2853694558143616, + "learning_rate": 6.265653925366581e-05, + "loss": 0.0661, + "step": 26530 + }, + { + "epoch": 0.621928421151333, + "grad_norm": 0.3191418945789337, + "learning_rate": 6.264970609074539e-05, + "loss": 0.036, + "step": 26531 + }, + { + "epoch": 0.6219518627261379, + "grad_norm": 0.5848738551139832, + "learning_rate": 6.264287313048865e-05, + "loss": 0.0596, + "step": 26532 + }, + { + "epoch": 0.6219753043009429, + "grad_norm": 0.1521402895450592, + "learning_rate": 6.263604037293278e-05, + "loss": 0.0306, + "step": 26533 + }, + { + "epoch": 0.6219987458757479, + "grad_norm": 0.2416313737630844, + "learning_rate": 6.26292078181148e-05, + "loss": 0.0373, + "step": 26534 + }, + { + "epoch": 0.6220221874505529, + "grad_norm": 0.12635332345962524, + "learning_rate": 6.262237546607178e-05, + "loss": 0.0116, + "step": 26535 + }, + { + "epoch": 0.6220456290253579, + "grad_norm": 0.19385665655136108, + "learning_rate": 6.261554331684081e-05, + "loss": 0.028, + "step": 26536 + }, + { + "epoch": 0.6220690706001629, + "grad_norm": 0.2047678828239441, + "learning_rate": 6.260871137045894e-05, + "loss": 0.0249, + "step": 26537 + }, + { + "epoch": 0.6220925121749679, + "grad_norm": 0.14193066954612732, + "learning_rate": 6.260187962696326e-05, + "loss": 0.036, + "step": 26538 + }, + { + "epoch": 0.6221159537497729, + "grad_norm": 0.31734079122543335, + "learning_rate": 6.259504808639085e-05, + "loss": 0.074, + "step": 26539 + }, + { + "epoch": 0.622139395324578, + "grad_norm": 0.24360911548137665, + "learning_rate": 6.258821674877875e-05, + "loss": 0.0257, + "step": 26540 + }, + { + "epoch": 0.6221628368993829, + "grad_norm": 0.45153623819351196, + "learning_rate": 6.258138561416403e-05, + "loss": 0.0946, + "step": 26541 + }, + { + "epoch": 0.622186278474188, + "grad_norm": 0.3534415364265442, + "learning_rate": 6.257455468258379e-05, + "loss": 0.0554, + "step": 26542 + }, + { + "epoch": 0.6222097200489929, + "grad_norm": 0.590760350227356, + "learning_rate": 6.256772395407503e-05, + "loss": 0.5299, + "step": 26543 + }, + { + "epoch": 0.6222331616237979, + "grad_norm": 0.24598567187786102, + "learning_rate": 6.256089342867488e-05, + "loss": 0.0615, + "step": 26544 + }, + { + "epoch": 0.6222566031986029, + "grad_norm": 0.5962978005409241, + "learning_rate": 6.255406310642033e-05, + "loss": 0.1505, + "step": 26545 + }, + { + "epoch": 0.6222800447734079, + "grad_norm": 0.4689931273460388, + "learning_rate": 6.254723298734853e-05, + "loss": 0.0744, + "step": 26546 + }, + { + "epoch": 0.6223034863482129, + "grad_norm": 0.5545377135276794, + "learning_rate": 6.254040307149646e-05, + "loss": 0.0949, + "step": 26547 + }, + { + "epoch": 0.6223269279230179, + "grad_norm": 0.17383906245231628, + "learning_rate": 6.253357335890127e-05, + "loss": 0.0323, + "step": 26548 + }, + { + "epoch": 0.6223503694978229, + "grad_norm": 0.12125943601131439, + "learning_rate": 6.252674384959992e-05, + "loss": 0.0208, + "step": 26549 + }, + { + "epoch": 0.6223738110726279, + "grad_norm": 0.3561282157897949, + "learning_rate": 6.251991454362955e-05, + "loss": 0.606, + "step": 26550 + }, + { + "epoch": 0.6223972526474328, + "grad_norm": 0.2481386363506317, + "learning_rate": 6.251308544102717e-05, + "loss": 0.0409, + "step": 26551 + }, + { + "epoch": 0.6224206942222379, + "grad_norm": 0.4857485592365265, + "learning_rate": 6.25062565418298e-05, + "loss": 0.0909, + "step": 26552 + }, + { + "epoch": 0.6224441357970428, + "grad_norm": 0.5109977126121521, + "learning_rate": 6.249942784607458e-05, + "loss": 0.0834, + "step": 26553 + }, + { + "epoch": 0.6224675773718479, + "grad_norm": 0.34037911891937256, + "learning_rate": 6.249259935379852e-05, + "loss": 0.0593, + "step": 26554 + }, + { + "epoch": 0.6224910189466528, + "grad_norm": 0.45378315448760986, + "learning_rate": 6.248577106503864e-05, + "loss": 0.0977, + "step": 26555 + }, + { + "epoch": 0.6225144605214579, + "grad_norm": 0.18743900954723358, + "learning_rate": 6.247894297983203e-05, + "loss": 0.0289, + "step": 26556 + }, + { + "epoch": 0.6225379020962628, + "grad_norm": 0.42129194736480713, + "learning_rate": 6.247211509821572e-05, + "loss": 0.056, + "step": 26557 + }, + { + "epoch": 0.6225613436710679, + "grad_norm": 0.14932169020175934, + "learning_rate": 6.24652874202268e-05, + "loss": 0.0175, + "step": 26558 + }, + { + "epoch": 0.6225847852458728, + "grad_norm": 0.3713931143283844, + "learning_rate": 6.245845994590225e-05, + "loss": 0.04, + "step": 26559 + }, + { + "epoch": 0.6226082268206778, + "grad_norm": 0.6788297295570374, + "learning_rate": 6.245163267527918e-05, + "loss": 0.7177, + "step": 26560 + }, + { + "epoch": 0.6226316683954828, + "grad_norm": 0.3501865565776825, + "learning_rate": 6.244480560839461e-05, + "loss": 0.58, + "step": 26561 + }, + { + "epoch": 0.6226551099702878, + "grad_norm": 0.7410675287246704, + "learning_rate": 6.243797874528557e-05, + "loss": 0.1018, + "step": 26562 + }, + { + "epoch": 0.6226785515450928, + "grad_norm": 0.2684433162212372, + "learning_rate": 6.24311520859891e-05, + "loss": 0.0406, + "step": 26563 + }, + { + "epoch": 0.6227019931198978, + "grad_norm": 0.3993340730667114, + "learning_rate": 6.242432563054229e-05, + "loss": 0.0767, + "step": 26564 + }, + { + "epoch": 0.6227254346947028, + "grad_norm": 0.4118175506591797, + "learning_rate": 6.241749937898212e-05, + "loss": 0.1023, + "step": 26565 + }, + { + "epoch": 0.6227488762695078, + "grad_norm": 0.43984201550483704, + "learning_rate": 6.241067333134567e-05, + "loss": 0.0877, + "step": 26566 + }, + { + "epoch": 0.6227723178443128, + "grad_norm": 0.5113315582275391, + "learning_rate": 6.240384748766996e-05, + "loss": 0.1104, + "step": 26567 + }, + { + "epoch": 0.6227957594191178, + "grad_norm": 0.3093615174293518, + "learning_rate": 6.239702184799201e-05, + "loss": 0.036, + "step": 26568 + }, + { + "epoch": 0.6228192009939227, + "grad_norm": 0.18207533657550812, + "learning_rate": 6.239019641234886e-05, + "loss": 0.0273, + "step": 26569 + }, + { + "epoch": 0.6228426425687278, + "grad_norm": 0.2261209636926651, + "learning_rate": 6.238337118077759e-05, + "loss": 0.0157, + "step": 26570 + }, + { + "epoch": 0.6228660841435327, + "grad_norm": 0.28168338537216187, + "learning_rate": 6.237654615331523e-05, + "loss": 0.0563, + "step": 26571 + }, + { + "epoch": 0.6228895257183378, + "grad_norm": 0.5915396809577942, + "learning_rate": 6.236972132999874e-05, + "loss": 0.1443, + "step": 26572 + }, + { + "epoch": 0.6229129672931427, + "grad_norm": 0.43117713928222656, + "learning_rate": 6.236289671086524e-05, + "loss": 0.0578, + "step": 26573 + }, + { + "epoch": 0.6229364088679478, + "grad_norm": 0.47104206681251526, + "learning_rate": 6.235607229595173e-05, + "loss": 0.11, + "step": 26574 + }, + { + "epoch": 0.6229598504427527, + "grad_norm": 0.11024858802556992, + "learning_rate": 6.234924808529518e-05, + "loss": 0.0174, + "step": 26575 + }, + { + "epoch": 0.6229832920175578, + "grad_norm": 0.42587655782699585, + "learning_rate": 6.234242407893272e-05, + "loss": 0.0667, + "step": 26576 + }, + { + "epoch": 0.6230067335923627, + "grad_norm": 0.08203493803739548, + "learning_rate": 6.233560027690131e-05, + "loss": 0.0118, + "step": 26577 + }, + { + "epoch": 0.6230301751671677, + "grad_norm": 0.2659605145454407, + "learning_rate": 6.232877667923796e-05, + "loss": 0.0323, + "step": 26578 + }, + { + "epoch": 0.6230536167419727, + "grad_norm": 0.10948534309864044, + "learning_rate": 6.232195328597977e-05, + "loss": 0.0285, + "step": 26579 + }, + { + "epoch": 0.6230770583167777, + "grad_norm": 0.3466523587703705, + "learning_rate": 6.231513009716369e-05, + "loss": 0.0617, + "step": 26580 + }, + { + "epoch": 0.6231004998915827, + "grad_norm": 0.24583032727241516, + "learning_rate": 6.230830711282677e-05, + "loss": 0.0544, + "step": 26581 + }, + { + "epoch": 0.6231239414663877, + "grad_norm": 0.12994064390659332, + "learning_rate": 6.230148433300605e-05, + "loss": 0.0182, + "step": 26582 + }, + { + "epoch": 0.6231473830411927, + "grad_norm": 0.6622911095619202, + "learning_rate": 6.229466175773853e-05, + "loss": 0.1028, + "step": 26583 + }, + { + "epoch": 0.6231708246159977, + "grad_norm": 0.24150650203227997, + "learning_rate": 6.228783938706125e-05, + "loss": 0.0219, + "step": 26584 + }, + { + "epoch": 0.6231942661908026, + "grad_norm": 0.3315672278404236, + "learning_rate": 6.22810172210112e-05, + "loss": 0.0282, + "step": 26585 + }, + { + "epoch": 0.6232177077656077, + "grad_norm": 0.4821951389312744, + "learning_rate": 6.227419525962544e-05, + "loss": 0.1022, + "step": 26586 + }, + { + "epoch": 0.6232411493404126, + "grad_norm": 0.7200678586959839, + "learning_rate": 6.22673735029409e-05, + "loss": 0.0748, + "step": 26587 + }, + { + "epoch": 0.6232645909152177, + "grad_norm": 0.34267207980155945, + "learning_rate": 6.22605519509947e-05, + "loss": 0.0488, + "step": 26588 + }, + { + "epoch": 0.6232880324900226, + "grad_norm": 0.2371111363172531, + "learning_rate": 6.22537306038238e-05, + "loss": 0.0275, + "step": 26589 + }, + { + "epoch": 0.6233114740648277, + "grad_norm": 0.4284898638725281, + "learning_rate": 6.22469094614652e-05, + "loss": 0.107, + "step": 26590 + }, + { + "epoch": 0.6233349156396327, + "grad_norm": 0.5129969120025635, + "learning_rate": 6.224008852395595e-05, + "loss": 0.5149, + "step": 26591 + }, + { + "epoch": 0.6233583572144377, + "grad_norm": 0.3696483075618744, + "learning_rate": 6.223326779133305e-05, + "loss": 0.0767, + "step": 26592 + }, + { + "epoch": 0.6233817987892427, + "grad_norm": 0.46497008204460144, + "learning_rate": 6.222644726363346e-05, + "loss": 0.4713, + "step": 26593 + }, + { + "epoch": 0.6234052403640477, + "grad_norm": 0.4713473916053772, + "learning_rate": 6.221962694089424e-05, + "loss": 0.0475, + "step": 26594 + }, + { + "epoch": 0.6234286819388527, + "grad_norm": 0.47506141662597656, + "learning_rate": 6.22128068231524e-05, + "loss": 0.5642, + "step": 26595 + }, + { + "epoch": 0.6234521235136576, + "grad_norm": 0.4987066388130188, + "learning_rate": 6.220598691044495e-05, + "loss": 0.0837, + "step": 26596 + }, + { + "epoch": 0.6234755650884627, + "grad_norm": 0.1939842849969864, + "learning_rate": 6.219916720280885e-05, + "loss": 0.0344, + "step": 26597 + }, + { + "epoch": 0.6234990066632676, + "grad_norm": 0.5527849793434143, + "learning_rate": 6.219234770028115e-05, + "loss": 0.0858, + "step": 26598 + }, + { + "epoch": 0.6235224482380727, + "grad_norm": 0.27249640226364136, + "learning_rate": 6.218552840289884e-05, + "loss": 0.0259, + "step": 26599 + }, + { + "epoch": 0.6235458898128776, + "grad_norm": 0.46219807863235474, + "learning_rate": 6.21787093106989e-05, + "loss": 0.0912, + "step": 26600 + }, + { + "epoch": 0.6235693313876827, + "grad_norm": 0.6415053009986877, + "learning_rate": 6.217189042371834e-05, + "loss": 0.0788, + "step": 26601 + }, + { + "epoch": 0.6235927729624876, + "grad_norm": 0.34658631682395935, + "learning_rate": 6.21650717419942e-05, + "loss": 0.3935, + "step": 26602 + }, + { + "epoch": 0.6236162145372927, + "grad_norm": 0.2993748188018799, + "learning_rate": 6.21582532655634e-05, + "loss": 0.0337, + "step": 26603 + }, + { + "epoch": 0.6236396561120976, + "grad_norm": 0.23571166396141052, + "learning_rate": 6.215143499446301e-05, + "loss": 0.0429, + "step": 26604 + }, + { + "epoch": 0.6236630976869026, + "grad_norm": 0.7597689032554626, + "learning_rate": 6.214461692872996e-05, + "loss": 0.6002, + "step": 26605 + }, + { + "epoch": 0.6236865392617076, + "grad_norm": 0.3964080810546875, + "learning_rate": 6.213779906840134e-05, + "loss": 0.0902, + "step": 26606 + }, + { + "epoch": 0.6237099808365126, + "grad_norm": 0.3785911500453949, + "learning_rate": 6.213098141351403e-05, + "loss": 0.0887, + "step": 26607 + }, + { + "epoch": 0.6237334224113176, + "grad_norm": 0.39468398690223694, + "learning_rate": 6.212416396410511e-05, + "loss": 0.0858, + "step": 26608 + }, + { + "epoch": 0.6237568639861226, + "grad_norm": 0.1823202222585678, + "learning_rate": 6.211734672021157e-05, + "loss": 0.0214, + "step": 26609 + }, + { + "epoch": 0.6237803055609276, + "grad_norm": 0.37039583921432495, + "learning_rate": 6.211052968187032e-05, + "loss": 0.0343, + "step": 26610 + }, + { + "epoch": 0.6238037471357326, + "grad_norm": 0.4659436345100403, + "learning_rate": 6.210371284911845e-05, + "loss": 0.1277, + "step": 26611 + }, + { + "epoch": 0.6238271887105376, + "grad_norm": 0.39340806007385254, + "learning_rate": 6.209689622199287e-05, + "loss": 0.0782, + "step": 26612 + }, + { + "epoch": 0.6238506302853426, + "grad_norm": 0.24578800797462463, + "learning_rate": 6.209007980053059e-05, + "loss": 0.0177, + "step": 26613 + }, + { + "epoch": 0.6238740718601475, + "grad_norm": 0.6055042147636414, + "learning_rate": 6.208326358476863e-05, + "loss": 0.0696, + "step": 26614 + }, + { + "epoch": 0.6238975134349526, + "grad_norm": 0.6456387639045715, + "learning_rate": 6.207644757474393e-05, + "loss": 0.0981, + "step": 26615 + }, + { + "epoch": 0.6239209550097575, + "grad_norm": 0.23545978963375092, + "learning_rate": 6.20696317704935e-05, + "loss": 0.0361, + "step": 26616 + }, + { + "epoch": 0.6239443965845626, + "grad_norm": 0.8098254799842834, + "learning_rate": 6.206281617205427e-05, + "loss": 0.1211, + "step": 26617 + }, + { + "epoch": 0.6239678381593675, + "grad_norm": 0.11719202250242233, + "learning_rate": 6.20560007794633e-05, + "loss": 0.0164, + "step": 26618 + }, + { + "epoch": 0.6239912797341726, + "grad_norm": 0.5895845890045166, + "learning_rate": 6.204918559275757e-05, + "loss": 0.0891, + "step": 26619 + }, + { + "epoch": 0.6240147213089775, + "grad_norm": 0.5093047022819519, + "learning_rate": 6.204237061197397e-05, + "loss": 0.075, + "step": 26620 + }, + { + "epoch": 0.6240381628837826, + "grad_norm": 0.7482693195343018, + "learning_rate": 6.203555583714957e-05, + "loss": 0.0886, + "step": 26621 + }, + { + "epoch": 0.6240616044585875, + "grad_norm": 0.1979951560497284, + "learning_rate": 6.202874126832133e-05, + "loss": 0.0206, + "step": 26622 + }, + { + "epoch": 0.6240850460333925, + "grad_norm": 0.23012162744998932, + "learning_rate": 6.202192690552615e-05, + "loss": 0.044, + "step": 26623 + }, + { + "epoch": 0.6241084876081975, + "grad_norm": 0.5588664412498474, + "learning_rate": 6.201511274880111e-05, + "loss": 0.6053, + "step": 26624 + }, + { + "epoch": 0.6241319291830025, + "grad_norm": 0.4598616063594818, + "learning_rate": 6.20082987981831e-05, + "loss": 0.1125, + "step": 26625 + }, + { + "epoch": 0.6241553707578075, + "grad_norm": 0.28712865710258484, + "learning_rate": 6.200148505370916e-05, + "loss": 0.0508, + "step": 26626 + }, + { + "epoch": 0.6241788123326125, + "grad_norm": 0.11286652833223343, + "learning_rate": 6.199467151541622e-05, + "loss": 0.0333, + "step": 26627 + }, + { + "epoch": 0.6242022539074175, + "grad_norm": 0.45249176025390625, + "learning_rate": 6.198785818334125e-05, + "loss": 0.0519, + "step": 26628 + }, + { + "epoch": 0.6242256954822225, + "grad_norm": 0.25784507393836975, + "learning_rate": 6.198104505752126e-05, + "loss": 0.0274, + "step": 26629 + }, + { + "epoch": 0.6242491370570274, + "grad_norm": 0.31010812520980835, + "learning_rate": 6.197423213799314e-05, + "loss": 0.0239, + "step": 26630 + }, + { + "epoch": 0.6242725786318325, + "grad_norm": 0.4348386526107788, + "learning_rate": 6.196741942479394e-05, + "loss": 0.0973, + "step": 26631 + }, + { + "epoch": 0.6242960202066374, + "grad_norm": 0.5147219896316528, + "learning_rate": 6.196060691796057e-05, + "loss": 0.0613, + "step": 26632 + }, + { + "epoch": 0.6243194617814425, + "grad_norm": 0.304171621799469, + "learning_rate": 6.195379461753004e-05, + "loss": 0.0682, + "step": 26633 + }, + { + "epoch": 0.6243429033562474, + "grad_norm": 0.4778212308883667, + "learning_rate": 6.19469825235393e-05, + "loss": 0.0576, + "step": 26634 + }, + { + "epoch": 0.6243663449310525, + "grad_norm": 0.1348559707403183, + "learning_rate": 6.194017063602527e-05, + "loss": 0.0193, + "step": 26635 + }, + { + "epoch": 0.6243897865058574, + "grad_norm": 0.3906407654285431, + "learning_rate": 6.193335895502497e-05, + "loss": 0.0625, + "step": 26636 + }, + { + "epoch": 0.6244132280806625, + "grad_norm": 0.46643027663230896, + "learning_rate": 6.192654748057534e-05, + "loss": 0.0655, + "step": 26637 + }, + { + "epoch": 0.6244366696554674, + "grad_norm": 0.39960139989852905, + "learning_rate": 6.19197362127133e-05, + "loss": 0.1134, + "step": 26638 + }, + { + "epoch": 0.6244601112302725, + "grad_norm": 0.45166003704071045, + "learning_rate": 6.191292515147587e-05, + "loss": 0.0758, + "step": 26639 + }, + { + "epoch": 0.6244835528050774, + "grad_norm": 0.6247228980064392, + "learning_rate": 6.190611429689996e-05, + "loss": 0.0671, + "step": 26640 + }, + { + "epoch": 0.6245069943798824, + "grad_norm": 0.16782911121845245, + "learning_rate": 6.189930364902255e-05, + "loss": 0.0192, + "step": 26641 + }, + { + "epoch": 0.6245304359546875, + "grad_norm": 0.4915284514427185, + "learning_rate": 6.189249320788057e-05, + "loss": 0.1012, + "step": 26642 + }, + { + "epoch": 0.6245538775294924, + "grad_norm": 0.49191468954086304, + "learning_rate": 6.188568297351103e-05, + "loss": 0.5665, + "step": 26643 + }, + { + "epoch": 0.6245773191042975, + "grad_norm": 0.4209316074848175, + "learning_rate": 6.187887294595083e-05, + "loss": 0.0712, + "step": 26644 + }, + { + "epoch": 0.6246007606791024, + "grad_norm": 0.5020423531532288, + "learning_rate": 6.187206312523694e-05, + "loss": 0.1326, + "step": 26645 + }, + { + "epoch": 0.6246242022539075, + "grad_norm": 0.4467982351779938, + "learning_rate": 6.18652535114063e-05, + "loss": 0.0775, + "step": 26646 + }, + { + "epoch": 0.6246476438287124, + "grad_norm": 0.23796534538269043, + "learning_rate": 6.18584441044959e-05, + "loss": 0.0401, + "step": 26647 + }, + { + "epoch": 0.6246710854035175, + "grad_norm": 0.12798023223876953, + "learning_rate": 6.185163490454261e-05, + "loss": 0.0259, + "step": 26648 + }, + { + "epoch": 0.6246945269783224, + "grad_norm": 0.1305743157863617, + "learning_rate": 6.184482591158345e-05, + "loss": 0.0249, + "step": 26649 + }, + { + "epoch": 0.6247179685531274, + "grad_norm": 0.3670911192893982, + "learning_rate": 6.183801712565533e-05, + "loss": 0.0667, + "step": 26650 + }, + { + "epoch": 0.6247414101279324, + "grad_norm": 0.27317139506340027, + "learning_rate": 6.183120854679519e-05, + "loss": 0.0605, + "step": 26651 + }, + { + "epoch": 0.6247648517027374, + "grad_norm": 0.6306323409080505, + "learning_rate": 6.182440017504e-05, + "loss": 0.6113, + "step": 26652 + }, + { + "epoch": 0.6247882932775424, + "grad_norm": 0.11319237947463989, + "learning_rate": 6.181759201042669e-05, + "loss": 0.017, + "step": 26653 + }, + { + "epoch": 0.6248117348523474, + "grad_norm": 0.10842818766832352, + "learning_rate": 6.181078405299216e-05, + "loss": 0.0166, + "step": 26654 + }, + { + "epoch": 0.6248351764271524, + "grad_norm": 0.3703317642211914, + "learning_rate": 6.180397630277343e-05, + "loss": 0.0778, + "step": 26655 + }, + { + "epoch": 0.6248586180019574, + "grad_norm": 0.1631317287683487, + "learning_rate": 6.179716875980739e-05, + "loss": 0.0255, + "step": 26656 + }, + { + "epoch": 0.6248820595767623, + "grad_norm": 0.4408304691314697, + "learning_rate": 6.1790361424131e-05, + "loss": 0.1828, + "step": 26657 + }, + { + "epoch": 0.6249055011515674, + "grad_norm": 0.7337767481803894, + "learning_rate": 6.178355429578115e-05, + "loss": 0.0989, + "step": 26658 + }, + { + "epoch": 0.6249289427263723, + "grad_norm": 0.2923574149608612, + "learning_rate": 6.177674737479485e-05, + "loss": 0.0863, + "step": 26659 + }, + { + "epoch": 0.6249523843011774, + "grad_norm": 0.14158207178115845, + "learning_rate": 6.176994066120898e-05, + "loss": 0.0354, + "step": 26660 + }, + { + "epoch": 0.6249758258759823, + "grad_norm": 1.0318048000335693, + "learning_rate": 6.176313415506047e-05, + "loss": 0.1321, + "step": 26661 + }, + { + "epoch": 0.6249992674507874, + "grad_norm": 0.3775378167629242, + "learning_rate": 6.17563278563863e-05, + "loss": 0.0808, + "step": 26662 + }, + { + "epoch": 0.6250227090255923, + "grad_norm": 0.45690107345581055, + "learning_rate": 6.174952176522333e-05, + "loss": 0.3941, + "step": 26663 + }, + { + "epoch": 0.6250461506003974, + "grad_norm": 0.3237793743610382, + "learning_rate": 6.174271588160856e-05, + "loss": 0.0647, + "step": 26664 + }, + { + "epoch": 0.6250695921752023, + "grad_norm": 0.9031160473823547, + "learning_rate": 6.173591020557888e-05, + "loss": 0.1819, + "step": 26665 + }, + { + "epoch": 0.6250930337500074, + "grad_norm": 0.37323224544525146, + "learning_rate": 6.17291047371712e-05, + "loss": 0.0553, + "step": 26666 + }, + { + "epoch": 0.6251164753248123, + "grad_norm": 0.3642179071903229, + "learning_rate": 6.172229947642249e-05, + "loss": 0.1277, + "step": 26667 + }, + { + "epoch": 0.6251399168996173, + "grad_norm": 0.5800895094871521, + "learning_rate": 6.171549442336968e-05, + "loss": 0.1381, + "step": 26668 + }, + { + "epoch": 0.6251633584744223, + "grad_norm": 0.47077780961990356, + "learning_rate": 6.170868957804968e-05, + "loss": 0.1339, + "step": 26669 + }, + { + "epoch": 0.6251868000492273, + "grad_norm": 0.5124191641807556, + "learning_rate": 6.170188494049938e-05, + "loss": 0.7064, + "step": 26670 + }, + { + "epoch": 0.6252102416240323, + "grad_norm": 0.241937056183815, + "learning_rate": 6.169508051075575e-05, + "loss": 0.0596, + "step": 26671 + }, + { + "epoch": 0.6252336831988373, + "grad_norm": 0.4765181541442871, + "learning_rate": 6.168827628885567e-05, + "loss": 0.0937, + "step": 26672 + }, + { + "epoch": 0.6252571247736423, + "grad_norm": 0.4801982343196869, + "learning_rate": 6.16814722748361e-05, + "loss": 0.7178, + "step": 26673 + }, + { + "epoch": 0.6252805663484473, + "grad_norm": 1.1848372220993042, + "learning_rate": 6.167466846873393e-05, + "loss": 0.1351, + "step": 26674 + }, + { + "epoch": 0.6253040079232522, + "grad_norm": 0.23602266609668732, + "learning_rate": 6.16678648705861e-05, + "loss": 0.0218, + "step": 26675 + }, + { + "epoch": 0.6253274494980573, + "grad_norm": 0.5265737771987915, + "learning_rate": 6.166106148042948e-05, + "loss": 0.0437, + "step": 26676 + }, + { + "epoch": 0.6253508910728622, + "grad_norm": 0.06739125400781631, + "learning_rate": 6.165425829830105e-05, + "loss": 0.0149, + "step": 26677 + }, + { + "epoch": 0.6253743326476673, + "grad_norm": 0.36985597014427185, + "learning_rate": 6.164745532423764e-05, + "loss": 0.0982, + "step": 26678 + }, + { + "epoch": 0.6253977742224722, + "grad_norm": 0.30443084239959717, + "learning_rate": 6.164065255827627e-05, + "loss": 0.0655, + "step": 26679 + }, + { + "epoch": 0.6254212157972773, + "grad_norm": 0.38299113512039185, + "learning_rate": 6.163385000045377e-05, + "loss": 0.0744, + "step": 26680 + }, + { + "epoch": 0.6254446573720822, + "grad_norm": 0.35211217403411865, + "learning_rate": 6.162704765080709e-05, + "loss": 0.0287, + "step": 26681 + }, + { + "epoch": 0.6254680989468873, + "grad_norm": 0.6288926005363464, + "learning_rate": 6.162024550937315e-05, + "loss": 0.1084, + "step": 26682 + }, + { + "epoch": 0.6254915405216922, + "grad_norm": 0.7246949076652527, + "learning_rate": 6.16134435761888e-05, + "loss": 0.1495, + "step": 26683 + }, + { + "epoch": 0.6255149820964973, + "grad_norm": 0.27821916341781616, + "learning_rate": 6.160664185129101e-05, + "loss": 0.0406, + "step": 26684 + }, + { + "epoch": 0.6255384236713022, + "grad_norm": 0.22339777648448944, + "learning_rate": 6.159984033471667e-05, + "loss": 0.0223, + "step": 26685 + }, + { + "epoch": 0.6255618652461072, + "grad_norm": 0.3384384512901306, + "learning_rate": 6.159303902650264e-05, + "loss": 0.0529, + "step": 26686 + }, + { + "epoch": 0.6255853068209122, + "grad_norm": 0.4403378963470459, + "learning_rate": 6.158623792668588e-05, + "loss": 0.0606, + "step": 26687 + }, + { + "epoch": 0.6256087483957172, + "grad_norm": 0.5113351941108704, + "learning_rate": 6.157943703530329e-05, + "loss": 0.1372, + "step": 26688 + }, + { + "epoch": 0.6256321899705222, + "grad_norm": 0.11285433918237686, + "learning_rate": 6.157263635239172e-05, + "loss": 0.0243, + "step": 26689 + }, + { + "epoch": 0.6256556315453272, + "grad_norm": 0.4523792266845703, + "learning_rate": 6.15658358779881e-05, + "loss": 0.1239, + "step": 26690 + }, + { + "epoch": 0.6256790731201322, + "grad_norm": 0.5737700462341309, + "learning_rate": 6.155903561212937e-05, + "loss": 0.0575, + "step": 26691 + }, + { + "epoch": 0.6257025146949372, + "grad_norm": 0.49397194385528564, + "learning_rate": 6.15522355548524e-05, + "loss": 0.538, + "step": 26692 + }, + { + "epoch": 0.6257259562697423, + "grad_norm": 0.36183062195777893, + "learning_rate": 6.154543570619403e-05, + "loss": 0.053, + "step": 26693 + }, + { + "epoch": 0.6257493978445472, + "grad_norm": 0.5877431631088257, + "learning_rate": 6.153863606619126e-05, + "loss": 0.0805, + "step": 26694 + }, + { + "epoch": 0.6257728394193522, + "grad_norm": 0.26716944575309753, + "learning_rate": 6.153183663488091e-05, + "loss": 0.0595, + "step": 26695 + }, + { + "epoch": 0.6257962809941572, + "grad_norm": 0.5845519304275513, + "learning_rate": 6.152503741229989e-05, + "loss": 0.1636, + "step": 26696 + }, + { + "epoch": 0.6258197225689622, + "grad_norm": 0.3516848087310791, + "learning_rate": 6.151823839848512e-05, + "loss": 0.0385, + "step": 26697 + }, + { + "epoch": 0.6258431641437672, + "grad_norm": 0.13864773511886597, + "learning_rate": 6.151143959347343e-05, + "loss": 0.0401, + "step": 26698 + }, + { + "epoch": 0.6258666057185722, + "grad_norm": 0.38371315598487854, + "learning_rate": 6.15046409973018e-05, + "loss": 0.0311, + "step": 26699 + }, + { + "epoch": 0.6258900472933772, + "grad_norm": 0.582004189491272, + "learning_rate": 6.149784261000705e-05, + "loss": 0.1427, + "step": 26700 + }, + { + "epoch": 0.6259134888681822, + "grad_norm": 1.1025986671447754, + "learning_rate": 6.149104443162607e-05, + "loss": 0.084, + "step": 26701 + }, + { + "epoch": 0.6259369304429871, + "grad_norm": 0.4672185480594635, + "learning_rate": 6.14842464621958e-05, + "loss": 0.0852, + "step": 26702 + }, + { + "epoch": 0.6259603720177922, + "grad_norm": 0.3588142693042755, + "learning_rate": 6.147744870175303e-05, + "loss": 0.0574, + "step": 26703 + }, + { + "epoch": 0.6259838135925971, + "grad_norm": 0.6200891137123108, + "learning_rate": 6.147065115033476e-05, + "loss": 0.1775, + "step": 26704 + }, + { + "epoch": 0.6260072551674022, + "grad_norm": 0.0981612280011177, + "learning_rate": 6.146385380797781e-05, + "loss": 0.0119, + "step": 26705 + }, + { + "epoch": 0.6260306967422071, + "grad_norm": 0.0974513366818428, + "learning_rate": 6.145705667471908e-05, + "loss": 0.0094, + "step": 26706 + }, + { + "epoch": 0.6260541383170122, + "grad_norm": 0.3424827754497528, + "learning_rate": 6.145025975059544e-05, + "loss": 0.0462, + "step": 26707 + }, + { + "epoch": 0.6260775798918171, + "grad_norm": 0.32588881254196167, + "learning_rate": 6.144346303564376e-05, + "loss": 0.0422, + "step": 26708 + }, + { + "epoch": 0.6261010214666222, + "grad_norm": 0.6435893774032593, + "learning_rate": 6.143666652990096e-05, + "loss": 0.1524, + "step": 26709 + }, + { + "epoch": 0.6261244630414271, + "grad_norm": 0.5020213723182678, + "learning_rate": 6.142987023340389e-05, + "loss": 0.056, + "step": 26710 + }, + { + "epoch": 0.6261479046162322, + "grad_norm": 0.2793407440185547, + "learning_rate": 6.14230741461894e-05, + "loss": 0.0726, + "step": 26711 + }, + { + "epoch": 0.6261713461910371, + "grad_norm": 0.5780156254768372, + "learning_rate": 6.141627826829442e-05, + "loss": 0.1113, + "step": 26712 + }, + { + "epoch": 0.6261947877658421, + "grad_norm": 0.4076666533946991, + "learning_rate": 6.14094825997558e-05, + "loss": 0.0933, + "step": 26713 + }, + { + "epoch": 0.6262182293406471, + "grad_norm": 0.4977710545063019, + "learning_rate": 6.14026871406104e-05, + "loss": 0.0878, + "step": 26714 + }, + { + "epoch": 0.6262416709154521, + "grad_norm": 0.485200971364975, + "learning_rate": 6.139589189089506e-05, + "loss": 0.0646, + "step": 26715 + }, + { + "epoch": 0.6262651124902571, + "grad_norm": 0.728722870349884, + "learning_rate": 6.138909685064676e-05, + "loss": 0.1231, + "step": 26716 + }, + { + "epoch": 0.6262885540650621, + "grad_norm": 0.09794988483190536, + "learning_rate": 6.13823020199023e-05, + "loss": 0.0335, + "step": 26717 + }, + { + "epoch": 0.6263119956398671, + "grad_norm": 0.6004647612571716, + "learning_rate": 6.137550739869852e-05, + "loss": 0.126, + "step": 26718 + }, + { + "epoch": 0.6263354372146721, + "grad_norm": 0.655705988407135, + "learning_rate": 6.136871298707236e-05, + "loss": 0.1684, + "step": 26719 + }, + { + "epoch": 0.626358878789477, + "grad_norm": 0.5491716265678406, + "learning_rate": 6.136191878506066e-05, + "loss": 0.0815, + "step": 26720 + }, + { + "epoch": 0.6263823203642821, + "grad_norm": 0.4282565712928772, + "learning_rate": 6.135512479270022e-05, + "loss": 0.0892, + "step": 26721 + }, + { + "epoch": 0.626405761939087, + "grad_norm": 0.23313534259796143, + "learning_rate": 6.134833101002802e-05, + "loss": 0.0483, + "step": 26722 + }, + { + "epoch": 0.6264292035138921, + "grad_norm": 0.12536601722240448, + "learning_rate": 6.134153743708084e-05, + "loss": 0.0234, + "step": 26723 + }, + { + "epoch": 0.626452645088697, + "grad_norm": 0.5700113773345947, + "learning_rate": 6.133474407389555e-05, + "loss": 0.067, + "step": 26724 + }, + { + "epoch": 0.6264760866635021, + "grad_norm": 0.5513464212417603, + "learning_rate": 6.132795092050902e-05, + "loss": 0.1297, + "step": 26725 + }, + { + "epoch": 0.626499528238307, + "grad_norm": 0.5105479955673218, + "learning_rate": 6.132115797695815e-05, + "loss": 0.1275, + "step": 26726 + }, + { + "epoch": 0.6265229698131121, + "grad_norm": 0.40557095408439636, + "learning_rate": 6.131436524327972e-05, + "loss": 0.1037, + "step": 26727 + }, + { + "epoch": 0.626546411387917, + "grad_norm": 0.36171889305114746, + "learning_rate": 6.130757271951064e-05, + "loss": 0.0326, + "step": 26728 + }, + { + "epoch": 0.626569852962722, + "grad_norm": 0.5529534816741943, + "learning_rate": 6.130078040568778e-05, + "loss": 0.1062, + "step": 26729 + }, + { + "epoch": 0.626593294537527, + "grad_norm": 0.5765427350997925, + "learning_rate": 6.129398830184798e-05, + "loss": 0.0964, + "step": 26730 + }, + { + "epoch": 0.626616736112332, + "grad_norm": 0.601536214351654, + "learning_rate": 6.128719640802806e-05, + "loss": 0.0744, + "step": 26731 + }, + { + "epoch": 0.626640177687137, + "grad_norm": 0.2670668363571167, + "learning_rate": 6.128040472426492e-05, + "loss": 0.0228, + "step": 26732 + }, + { + "epoch": 0.626663619261942, + "grad_norm": 0.7573511600494385, + "learning_rate": 6.127361325059541e-05, + "loss": 0.0819, + "step": 26733 + }, + { + "epoch": 0.626687060836747, + "grad_norm": 0.41866081953048706, + "learning_rate": 6.126682198705633e-05, + "loss": 0.0543, + "step": 26734 + }, + { + "epoch": 0.626710502411552, + "grad_norm": 0.793489396572113, + "learning_rate": 6.126003093368459e-05, + "loss": 0.0976, + "step": 26735 + }, + { + "epoch": 0.626733943986357, + "grad_norm": 0.3302801847457886, + "learning_rate": 6.125324009051698e-05, + "loss": 0.0556, + "step": 26736 + }, + { + "epoch": 0.626757385561162, + "grad_norm": 0.3751717507839203, + "learning_rate": 6.124644945759041e-05, + "loss": 0.0614, + "step": 26737 + }, + { + "epoch": 0.626780827135967, + "grad_norm": 0.7692760825157166, + "learning_rate": 6.123965903494169e-05, + "loss": 0.0948, + "step": 26738 + }, + { + "epoch": 0.626804268710772, + "grad_norm": 0.12639117240905762, + "learning_rate": 6.123286882260763e-05, + "loss": 0.0227, + "step": 26739 + }, + { + "epoch": 0.6268277102855769, + "grad_norm": 0.4511306583881378, + "learning_rate": 6.122607882062516e-05, + "loss": 0.0895, + "step": 26740 + }, + { + "epoch": 0.626851151860382, + "grad_norm": 0.33908677101135254, + "learning_rate": 6.121928902903104e-05, + "loss": 0.091, + "step": 26741 + }, + { + "epoch": 0.6268745934351869, + "grad_norm": 0.42252179980278015, + "learning_rate": 6.121249944786219e-05, + "loss": 0.0942, + "step": 26742 + }, + { + "epoch": 0.626898035009992, + "grad_norm": 0.42219558358192444, + "learning_rate": 6.120571007715537e-05, + "loss": 0.0693, + "step": 26743 + }, + { + "epoch": 0.626921476584797, + "grad_norm": 0.5355006456375122, + "learning_rate": 6.119892091694748e-05, + "loss": 0.6707, + "step": 26744 + }, + { + "epoch": 0.626944918159602, + "grad_norm": 0.5756266713142395, + "learning_rate": 6.119213196727534e-05, + "loss": 0.0556, + "step": 26745 + }, + { + "epoch": 0.626968359734407, + "grad_norm": 0.30274122953414917, + "learning_rate": 6.118534322817577e-05, + "loss": 0.0462, + "step": 26746 + }, + { + "epoch": 0.626991801309212, + "grad_norm": 0.21129338443279266, + "learning_rate": 6.117855469968563e-05, + "loss": 0.0386, + "step": 26747 + }, + { + "epoch": 0.627015242884017, + "grad_norm": 0.2299191802740097, + "learning_rate": 6.117176638184174e-05, + "loss": 0.0241, + "step": 26748 + }, + { + "epoch": 0.6270386844588219, + "grad_norm": 0.5677034258842468, + "learning_rate": 6.116497827468093e-05, + "loss": 0.602, + "step": 26749 + }, + { + "epoch": 0.627062126033627, + "grad_norm": 0.3963748514652252, + "learning_rate": 6.115819037824005e-05, + "loss": 0.0708, + "step": 26750 + }, + { + "epoch": 0.6270855676084319, + "grad_norm": 0.38990816473960876, + "learning_rate": 6.11514026925559e-05, + "loss": 0.422, + "step": 26751 + }, + { + "epoch": 0.627109009183237, + "grad_norm": 0.19792860746383667, + "learning_rate": 6.114461521766534e-05, + "loss": 0.0178, + "step": 26752 + }, + { + "epoch": 0.6271324507580419, + "grad_norm": 0.5289853811264038, + "learning_rate": 6.113782795360519e-05, + "loss": 0.1022, + "step": 26753 + }, + { + "epoch": 0.627155892332847, + "grad_norm": 0.47102952003479004, + "learning_rate": 6.113104090041231e-05, + "loss": 0.0744, + "step": 26754 + }, + { + "epoch": 0.6271793339076519, + "grad_norm": 0.446008563041687, + "learning_rate": 6.112425405812348e-05, + "loss": 0.0605, + "step": 26755 + }, + { + "epoch": 0.627202775482457, + "grad_norm": 0.0668923631310463, + "learning_rate": 6.111746742677551e-05, + "loss": 0.0059, + "step": 26756 + }, + { + "epoch": 0.6272262170572619, + "grad_norm": 0.6534887552261353, + "learning_rate": 6.11106810064053e-05, + "loss": 0.0963, + "step": 26757 + }, + { + "epoch": 0.6272496586320669, + "grad_norm": 0.5031149983406067, + "learning_rate": 6.11038947970496e-05, + "loss": 0.803, + "step": 26758 + }, + { + "epoch": 0.6272731002068719, + "grad_norm": 0.08923327177762985, + "learning_rate": 6.109710879874526e-05, + "loss": 0.0166, + "step": 26759 + }, + { + "epoch": 0.6272965417816769, + "grad_norm": 0.7670652270317078, + "learning_rate": 6.10903230115291e-05, + "loss": 0.5066, + "step": 26760 + }, + { + "epoch": 0.6273199833564819, + "grad_norm": 0.465378075838089, + "learning_rate": 6.108353743543796e-05, + "loss": 0.0595, + "step": 26761 + }, + { + "epoch": 0.6273434249312869, + "grad_norm": 0.3936851918697357, + "learning_rate": 6.107675207050862e-05, + "loss": 0.0787, + "step": 26762 + }, + { + "epoch": 0.6273668665060919, + "grad_norm": 0.29441338777542114, + "learning_rate": 6.10699669167779e-05, + "loss": 0.0417, + "step": 26763 + }, + { + "epoch": 0.6273903080808969, + "grad_norm": 0.08123283833265305, + "learning_rate": 6.106318197428267e-05, + "loss": 0.0164, + "step": 26764 + }, + { + "epoch": 0.6274137496557018, + "grad_norm": 0.19747976958751678, + "learning_rate": 6.105639724305971e-05, + "loss": 0.0235, + "step": 26765 + }, + { + "epoch": 0.6274371912305069, + "grad_norm": 0.870417594909668, + "learning_rate": 6.104961272314581e-05, + "loss": 0.1894, + "step": 26766 + }, + { + "epoch": 0.6274606328053118, + "grad_norm": 0.6442238688468933, + "learning_rate": 6.104282841457782e-05, + "loss": 0.1494, + "step": 26767 + }, + { + "epoch": 0.6274840743801169, + "grad_norm": 0.41302576661109924, + "learning_rate": 6.103604431739256e-05, + "loss": 0.0852, + "step": 26768 + }, + { + "epoch": 0.6275075159549218, + "grad_norm": 0.31640687584877014, + "learning_rate": 6.102926043162678e-05, + "loss": 0.0761, + "step": 26769 + }, + { + "epoch": 0.6275309575297269, + "grad_norm": 0.36587411165237427, + "learning_rate": 6.102247675731736e-05, + "loss": 0.0748, + "step": 26770 + }, + { + "epoch": 0.6275543991045318, + "grad_norm": 0.5003912448883057, + "learning_rate": 6.101569329450108e-05, + "loss": 0.1199, + "step": 26771 + }, + { + "epoch": 0.6275778406793369, + "grad_norm": 0.45252475142478943, + "learning_rate": 6.1008910043214705e-05, + "loss": 0.0968, + "step": 26772 + }, + { + "epoch": 0.6276012822541418, + "grad_norm": 0.3563704192638397, + "learning_rate": 6.100212700349511e-05, + "loss": 0.0799, + "step": 26773 + }, + { + "epoch": 0.6276247238289469, + "grad_norm": 0.609160304069519, + "learning_rate": 6.0995344175379055e-05, + "loss": 0.0803, + "step": 26774 + }, + { + "epoch": 0.6276481654037518, + "grad_norm": 0.20766814053058624, + "learning_rate": 6.098856155890338e-05, + "loss": 0.0176, + "step": 26775 + }, + { + "epoch": 0.6276716069785568, + "grad_norm": 0.3551354706287384, + "learning_rate": 6.098177915410481e-05, + "loss": 0.0761, + "step": 26776 + }, + { + "epoch": 0.6276950485533618, + "grad_norm": 0.16225433349609375, + "learning_rate": 6.097499696102027e-05, + "loss": 0.0241, + "step": 26777 + }, + { + "epoch": 0.6277184901281668, + "grad_norm": 0.43175312876701355, + "learning_rate": 6.096821497968648e-05, + "loss": 0.074, + "step": 26778 + }, + { + "epoch": 0.6277419317029718, + "grad_norm": 0.39099612832069397, + "learning_rate": 6.096143321014023e-05, + "loss": 0.0625, + "step": 26779 + }, + { + "epoch": 0.6277653732777768, + "grad_norm": 0.42953774333000183, + "learning_rate": 6.095465165241836e-05, + "loss": 0.1004, + "step": 26780 + }, + { + "epoch": 0.6277888148525818, + "grad_norm": 0.1262897551059723, + "learning_rate": 6.094787030655764e-05, + "loss": 0.0289, + "step": 26781 + }, + { + "epoch": 0.6278122564273868, + "grad_norm": 0.3484669327735901, + "learning_rate": 6.094108917259489e-05, + "loss": 0.0306, + "step": 26782 + }, + { + "epoch": 0.6278356980021917, + "grad_norm": 0.5481164455413818, + "learning_rate": 6.093430825056687e-05, + "loss": 0.1674, + "step": 26783 + }, + { + "epoch": 0.6278591395769968, + "grad_norm": 0.1701817363500595, + "learning_rate": 6.092752754051039e-05, + "loss": 0.0237, + "step": 26784 + }, + { + "epoch": 0.6278825811518017, + "grad_norm": 0.22744609415531158, + "learning_rate": 6.0920747042462266e-05, + "loss": 0.0466, + "step": 26785 + }, + { + "epoch": 0.6279060227266068, + "grad_norm": 0.39457568526268005, + "learning_rate": 6.091396675645926e-05, + "loss": 0.1283, + "step": 26786 + }, + { + "epoch": 0.6279294643014117, + "grad_norm": 0.31736522912979126, + "learning_rate": 6.090718668253815e-05, + "loss": 0.0446, + "step": 26787 + }, + { + "epoch": 0.6279529058762168, + "grad_norm": 0.45035645365715027, + "learning_rate": 6.0900406820735726e-05, + "loss": 0.059, + "step": 26788 + }, + { + "epoch": 0.6279763474510217, + "grad_norm": 0.23037783801555634, + "learning_rate": 6.089362717108884e-05, + "loss": 0.0336, + "step": 26789 + }, + { + "epoch": 0.6279997890258268, + "grad_norm": 0.3470064103603363, + "learning_rate": 6.088684773363422e-05, + "loss": 0.0898, + "step": 26790 + }, + { + "epoch": 0.6280232306006317, + "grad_norm": 0.46571242809295654, + "learning_rate": 6.088006850840865e-05, + "loss": 0.0859, + "step": 26791 + }, + { + "epoch": 0.6280466721754367, + "grad_norm": 0.2971831262111664, + "learning_rate": 6.0873289495448946e-05, + "loss": 0.0433, + "step": 26792 + }, + { + "epoch": 0.6280701137502417, + "grad_norm": 0.49734070897102356, + "learning_rate": 6.086651069479188e-05, + "loss": 0.1562, + "step": 26793 + }, + { + "epoch": 0.6280935553250467, + "grad_norm": 0.28160130977630615, + "learning_rate": 6.08597321064742e-05, + "loss": 0.0571, + "step": 26794 + }, + { + "epoch": 0.6281169968998518, + "grad_norm": 0.5414356589317322, + "learning_rate": 6.085295373053275e-05, + "loss": 0.566, + "step": 26795 + }, + { + "epoch": 0.6281404384746567, + "grad_norm": 0.12320496141910553, + "learning_rate": 6.0846175567004246e-05, + "loss": 0.0104, + "step": 26796 + }, + { + "epoch": 0.6281638800494618, + "grad_norm": 0.5566375255584717, + "learning_rate": 6.08393976159255e-05, + "loss": 0.0783, + "step": 26797 + }, + { + "epoch": 0.6281873216242667, + "grad_norm": 0.10431613773107529, + "learning_rate": 6.0832619877333284e-05, + "loss": 0.0215, + "step": 26798 + }, + { + "epoch": 0.6282107631990718, + "grad_norm": 0.36276981234550476, + "learning_rate": 6.0825842351264385e-05, + "loss": 0.0656, + "step": 26799 + }, + { + "epoch": 0.6282342047738767, + "grad_norm": 0.4889976680278778, + "learning_rate": 6.081906503775553e-05, + "loss": 0.068, + "step": 26800 + }, + { + "epoch": 0.6282576463486818, + "grad_norm": 0.2991228699684143, + "learning_rate": 6.081228793684355e-05, + "loss": 0.0374, + "step": 26801 + }, + { + "epoch": 0.6282810879234867, + "grad_norm": 0.6010429859161377, + "learning_rate": 6.08055110485652e-05, + "loss": 0.6707, + "step": 26802 + }, + { + "epoch": 0.6283045294982917, + "grad_norm": 0.4629882872104645, + "learning_rate": 6.0798734372957264e-05, + "loss": 0.093, + "step": 26803 + }, + { + "epoch": 0.6283279710730967, + "grad_norm": 0.22211511433124542, + "learning_rate": 6.0791957910056473e-05, + "loss": 0.033, + "step": 26804 + }, + { + "epoch": 0.6283514126479017, + "grad_norm": 0.5043394565582275, + "learning_rate": 6.0785181659899635e-05, + "loss": 0.1051, + "step": 26805 + }, + { + "epoch": 0.6283748542227067, + "grad_norm": 0.32211780548095703, + "learning_rate": 6.0778405622523525e-05, + "loss": 0.0649, + "step": 26806 + }, + { + "epoch": 0.6283982957975117, + "grad_norm": 0.34866863489151, + "learning_rate": 6.077162979796485e-05, + "loss": 0.0507, + "step": 26807 + }, + { + "epoch": 0.6284217373723167, + "grad_norm": 0.5633473992347717, + "learning_rate": 6.076485418626044e-05, + "loss": 0.0633, + "step": 26808 + }, + { + "epoch": 0.6284451789471217, + "grad_norm": 0.656114935874939, + "learning_rate": 6.0758078787447036e-05, + "loss": 0.1262, + "step": 26809 + }, + { + "epoch": 0.6284686205219266, + "grad_norm": 0.3963674008846283, + "learning_rate": 6.075130360156138e-05, + "loss": 0.0463, + "step": 26810 + }, + { + "epoch": 0.6284920620967317, + "grad_norm": 0.6269482374191284, + "learning_rate": 6.0744528628640265e-05, + "loss": 0.1082, + "step": 26811 + }, + { + "epoch": 0.6285155036715366, + "grad_norm": 0.3918965458869934, + "learning_rate": 6.073775386872042e-05, + "loss": 0.0576, + "step": 26812 + }, + { + "epoch": 0.6285389452463417, + "grad_norm": 0.6960907578468323, + "learning_rate": 6.0730979321838646e-05, + "loss": 0.1072, + "step": 26813 + }, + { + "epoch": 0.6285623868211466, + "grad_norm": 0.48083531856536865, + "learning_rate": 6.072420498803167e-05, + "loss": 0.6005, + "step": 26814 + }, + { + "epoch": 0.6285858283959517, + "grad_norm": 0.06333029270172119, + "learning_rate": 6.071743086733629e-05, + "loss": 0.0113, + "step": 26815 + }, + { + "epoch": 0.6286092699707566, + "grad_norm": 0.6223271489143372, + "learning_rate": 6.0710656959789215e-05, + "loss": 0.1065, + "step": 26816 + }, + { + "epoch": 0.6286327115455617, + "grad_norm": 0.12371890991926193, + "learning_rate": 6.070388326542722e-05, + "loss": 0.0232, + "step": 26817 + }, + { + "epoch": 0.6286561531203666, + "grad_norm": 0.4486226439476013, + "learning_rate": 6.0697109784287074e-05, + "loss": 0.0987, + "step": 26818 + }, + { + "epoch": 0.6286795946951717, + "grad_norm": 0.6242238879203796, + "learning_rate": 6.06903365164055e-05, + "loss": 0.0899, + "step": 26819 + }, + { + "epoch": 0.6287030362699766, + "grad_norm": 0.1442897766828537, + "learning_rate": 6.068356346181927e-05, + "loss": 0.02, + "step": 26820 + }, + { + "epoch": 0.6287264778447816, + "grad_norm": 0.14222630858421326, + "learning_rate": 6.067679062056515e-05, + "loss": 0.0214, + "step": 26821 + }, + { + "epoch": 0.6287499194195866, + "grad_norm": 0.46861565113067627, + "learning_rate": 6.067001799267984e-05, + "loss": 0.0923, + "step": 26822 + }, + { + "epoch": 0.6287733609943916, + "grad_norm": 0.6875134706497192, + "learning_rate": 6.0663245578200135e-05, + "loss": 0.0576, + "step": 26823 + }, + { + "epoch": 0.6287968025691966, + "grad_norm": 0.27927765250205994, + "learning_rate": 6.065647337716273e-05, + "loss": 0.0612, + "step": 26824 + }, + { + "epoch": 0.6288202441440016, + "grad_norm": 0.5330350399017334, + "learning_rate": 6.064970138960444e-05, + "loss": 0.0784, + "step": 26825 + }, + { + "epoch": 0.6288436857188066, + "grad_norm": 0.4416896402835846, + "learning_rate": 6.064292961556196e-05, + "loss": 0.0713, + "step": 26826 + }, + { + "epoch": 0.6288671272936116, + "grad_norm": 0.3707752525806427, + "learning_rate": 6.0636158055072075e-05, + "loss": 0.0918, + "step": 26827 + }, + { + "epoch": 0.6288905688684165, + "grad_norm": 0.16913548111915588, + "learning_rate": 6.062938670817151e-05, + "loss": 0.0266, + "step": 26828 + }, + { + "epoch": 0.6289140104432216, + "grad_norm": 0.26266855001449585, + "learning_rate": 6.062261557489697e-05, + "loss": 0.3071, + "step": 26829 + }, + { + "epoch": 0.6289374520180265, + "grad_norm": 0.2670489549636841, + "learning_rate": 6.061584465528524e-05, + "loss": 0.0519, + "step": 26830 + }, + { + "epoch": 0.6289608935928316, + "grad_norm": 0.2243759036064148, + "learning_rate": 6.0609073949373055e-05, + "loss": 0.0325, + "step": 26831 + }, + { + "epoch": 0.6289843351676365, + "grad_norm": 0.4025965929031372, + "learning_rate": 6.060230345719711e-05, + "loss": 0.0606, + "step": 26832 + }, + { + "epoch": 0.6290077767424416, + "grad_norm": 0.5093463063240051, + "learning_rate": 6.0595533178794205e-05, + "loss": 0.1307, + "step": 26833 + }, + { + "epoch": 0.6290312183172465, + "grad_norm": 0.33154216408729553, + "learning_rate": 6.058876311420103e-05, + "loss": 0.0945, + "step": 26834 + }, + { + "epoch": 0.6290546598920516, + "grad_norm": 0.09741096943616867, + "learning_rate": 6.058199326345433e-05, + "loss": 0.0135, + "step": 26835 + }, + { + "epoch": 0.6290781014668565, + "grad_norm": 0.6637623906135559, + "learning_rate": 6.057522362659082e-05, + "loss": 0.1264, + "step": 26836 + }, + { + "epoch": 0.6291015430416615, + "grad_norm": 0.6258143782615662, + "learning_rate": 6.0568454203647295e-05, + "loss": 0.1136, + "step": 26837 + }, + { + "epoch": 0.6291249846164665, + "grad_norm": 0.25885283946990967, + "learning_rate": 6.056168499466044e-05, + "loss": 0.0229, + "step": 26838 + }, + { + "epoch": 0.6291484261912715, + "grad_norm": 0.34121036529541016, + "learning_rate": 6.055491599966698e-05, + "loss": 0.0471, + "step": 26839 + }, + { + "epoch": 0.6291718677660765, + "grad_norm": 0.4229595363140106, + "learning_rate": 6.054814721870367e-05, + "loss": 0.1405, + "step": 26840 + }, + { + "epoch": 0.6291953093408815, + "grad_norm": 0.9134506583213806, + "learning_rate": 6.0541378651807224e-05, + "loss": 0.2564, + "step": 26841 + }, + { + "epoch": 0.6292187509156865, + "grad_norm": 0.47033438086509705, + "learning_rate": 6.053461029901435e-05, + "loss": 0.0411, + "step": 26842 + }, + { + "epoch": 0.6292421924904915, + "grad_norm": 0.4586160480976105, + "learning_rate": 6.0527842160361804e-05, + "loss": 0.0896, + "step": 26843 + }, + { + "epoch": 0.6292656340652965, + "grad_norm": 0.4025191366672516, + "learning_rate": 6.052107423588629e-05, + "loss": 0.0606, + "step": 26844 + }, + { + "epoch": 0.6292890756401015, + "grad_norm": 0.2602636218070984, + "learning_rate": 6.0514306525624534e-05, + "loss": 0.0546, + "step": 26845 + }, + { + "epoch": 0.6293125172149066, + "grad_norm": 0.49737638235092163, + "learning_rate": 6.050753902961327e-05, + "loss": 0.1257, + "step": 26846 + }, + { + "epoch": 0.6293359587897115, + "grad_norm": 0.21316993236541748, + "learning_rate": 6.050077174788922e-05, + "loss": 0.0561, + "step": 26847 + }, + { + "epoch": 0.6293594003645165, + "grad_norm": 0.38964587450027466, + "learning_rate": 6.0494004680489036e-05, + "loss": 0.0952, + "step": 26848 + }, + { + "epoch": 0.6293828419393215, + "grad_norm": 0.3542628586292267, + "learning_rate": 6.048723782744952e-05, + "loss": 0.0656, + "step": 26849 + }, + { + "epoch": 0.6294062835141265, + "grad_norm": 0.6057885885238647, + "learning_rate": 6.048047118880738e-05, + "loss": 0.1016, + "step": 26850 + }, + { + "epoch": 0.6294297250889315, + "grad_norm": 0.37013012170791626, + "learning_rate": 6.047370476459933e-05, + "loss": 0.0548, + "step": 26851 + }, + { + "epoch": 0.6294531666637365, + "grad_norm": 0.4517463743686676, + "learning_rate": 6.046693855486203e-05, + "loss": 0.0657, + "step": 26852 + }, + { + "epoch": 0.6294766082385415, + "grad_norm": 0.5337816476821899, + "learning_rate": 6.046017255963228e-05, + "loss": 0.0355, + "step": 26853 + }, + { + "epoch": 0.6295000498133465, + "grad_norm": 0.31338515877723694, + "learning_rate": 6.04534067789467e-05, + "loss": 0.0461, + "step": 26854 + }, + { + "epoch": 0.6295234913881514, + "grad_norm": 0.5108804106712341, + "learning_rate": 6.044664121284207e-05, + "loss": 0.1024, + "step": 26855 + }, + { + "epoch": 0.6295469329629565, + "grad_norm": 0.5105453133583069, + "learning_rate": 6.043987586135509e-05, + "loss": 0.1415, + "step": 26856 + }, + { + "epoch": 0.6295703745377614, + "grad_norm": 0.23806233704090118, + "learning_rate": 6.043311072452242e-05, + "loss": 0.0532, + "step": 26857 + }, + { + "epoch": 0.6295938161125665, + "grad_norm": 0.8387278318405151, + "learning_rate": 6.0426345802380824e-05, + "loss": 0.1675, + "step": 26858 + }, + { + "epoch": 0.6296172576873714, + "grad_norm": 0.39854857325553894, + "learning_rate": 6.0419581094967004e-05, + "loss": 0.0706, + "step": 26859 + }, + { + "epoch": 0.6296406992621765, + "grad_norm": 0.3774068355560303, + "learning_rate": 6.041281660231762e-05, + "loss": 0.0859, + "step": 26860 + }, + { + "epoch": 0.6296641408369814, + "grad_norm": 0.5628002285957336, + "learning_rate": 6.040605232446939e-05, + "loss": 0.1734, + "step": 26861 + }, + { + "epoch": 0.6296875824117865, + "grad_norm": 0.45376190543174744, + "learning_rate": 6.0399288261459064e-05, + "loss": 0.1489, + "step": 26862 + }, + { + "epoch": 0.6297110239865914, + "grad_norm": 0.2981502413749695, + "learning_rate": 6.039252441332333e-05, + "loss": 0.0579, + "step": 26863 + }, + { + "epoch": 0.6297344655613965, + "grad_norm": 0.4527492821216583, + "learning_rate": 6.038576078009883e-05, + "loss": 0.4934, + "step": 26864 + }, + { + "epoch": 0.6297579071362014, + "grad_norm": 0.25659623742103577, + "learning_rate": 6.037899736182235e-05, + "loss": 0.0489, + "step": 26865 + }, + { + "epoch": 0.6297813487110064, + "grad_norm": 0.15745659172534943, + "learning_rate": 6.037223415853053e-05, + "loss": 0.0229, + "step": 26866 + }, + { + "epoch": 0.6298047902858114, + "grad_norm": 0.2868969738483429, + "learning_rate": 6.036547117026006e-05, + "loss": 0.05, + "step": 26867 + }, + { + "epoch": 0.6298282318606164, + "grad_norm": 0.33549949526786804, + "learning_rate": 6.0358708397047694e-05, + "loss": 0.0471, + "step": 26868 + }, + { + "epoch": 0.6298516734354214, + "grad_norm": 0.20517019927501678, + "learning_rate": 6.035194583893007e-05, + "loss": 0.0208, + "step": 26869 + }, + { + "epoch": 0.6298751150102264, + "grad_norm": 0.11215697973966599, + "learning_rate": 6.034518349594389e-05, + "loss": 0.0151, + "step": 26870 + }, + { + "epoch": 0.6298985565850314, + "grad_norm": 0.09516613930463791, + "learning_rate": 6.033842136812589e-05, + "loss": 0.0084, + "step": 26871 + }, + { + "epoch": 0.6299219981598364, + "grad_norm": 0.5106312036514282, + "learning_rate": 6.0331659455512715e-05, + "loss": 0.1369, + "step": 26872 + }, + { + "epoch": 0.6299454397346413, + "grad_norm": 0.5806544423103333, + "learning_rate": 6.032489775814104e-05, + "loss": 0.1025, + "step": 26873 + }, + { + "epoch": 0.6299688813094464, + "grad_norm": 0.3576680123806, + "learning_rate": 6.031813627604761e-05, + "loss": 0.3866, + "step": 26874 + }, + { + "epoch": 0.6299923228842513, + "grad_norm": 0.3716905415058136, + "learning_rate": 6.0311375009269086e-05, + "loss": 0.0612, + "step": 26875 + }, + { + "epoch": 0.6300157644590564, + "grad_norm": 0.13979001343250275, + "learning_rate": 6.0304613957842184e-05, + "loss": 0.0257, + "step": 26876 + }, + { + "epoch": 0.6300392060338613, + "grad_norm": 0.4362089931964874, + "learning_rate": 6.029785312180353e-05, + "loss": 0.0851, + "step": 26877 + }, + { + "epoch": 0.6300626476086664, + "grad_norm": 0.2786291241645813, + "learning_rate": 6.029109250118986e-05, + "loss": 0.0398, + "step": 26878 + }, + { + "epoch": 0.6300860891834713, + "grad_norm": 0.35592392086982727, + "learning_rate": 6.0284332096037845e-05, + "loss": 0.0886, + "step": 26879 + }, + { + "epoch": 0.6301095307582764, + "grad_norm": 0.3291975259780884, + "learning_rate": 6.0277571906384134e-05, + "loss": 0.0443, + "step": 26880 + }, + { + "epoch": 0.6301329723330813, + "grad_norm": 0.19392599165439606, + "learning_rate": 6.027081193226546e-05, + "loss": 0.0193, + "step": 26881 + }, + { + "epoch": 0.6301564139078863, + "grad_norm": 0.3472140431404114, + "learning_rate": 6.0264052173718475e-05, + "loss": 0.1066, + "step": 26882 + }, + { + "epoch": 0.6301798554826913, + "grad_norm": 0.517835795879364, + "learning_rate": 6.025729263077984e-05, + "loss": 0.4361, + "step": 26883 + }, + { + "epoch": 0.6302032970574963, + "grad_norm": 0.5062897801399231, + "learning_rate": 6.025053330348628e-05, + "loss": 0.1061, + "step": 26884 + }, + { + "epoch": 0.6302267386323013, + "grad_norm": 0.1636422723531723, + "learning_rate": 6.024377419187439e-05, + "loss": 0.0195, + "step": 26885 + }, + { + "epoch": 0.6302501802071063, + "grad_norm": 0.23319457471370697, + "learning_rate": 6.0237015295980946e-05, + "loss": 0.057, + "step": 26886 + }, + { + "epoch": 0.6302736217819113, + "grad_norm": 0.8078904747962952, + "learning_rate": 6.023025661584256e-05, + "loss": 0.1363, + "step": 26887 + }, + { + "epoch": 0.6302970633567163, + "grad_norm": 0.18554943799972534, + "learning_rate": 6.022349815149593e-05, + "loss": 0.0254, + "step": 26888 + }, + { + "epoch": 0.6303205049315213, + "grad_norm": 0.3341228663921356, + "learning_rate": 6.021673990297773e-05, + "loss": 0.0406, + "step": 26889 + }, + { + "epoch": 0.6303439465063263, + "grad_norm": 0.4745998680591583, + "learning_rate": 6.020998187032458e-05, + "loss": 0.0954, + "step": 26890 + }, + { + "epoch": 0.6303673880811312, + "grad_norm": 0.22573328018188477, + "learning_rate": 6.020322405357323e-05, + "loss": 0.0508, + "step": 26891 + }, + { + "epoch": 0.6303908296559363, + "grad_norm": 0.5273569226264954, + "learning_rate": 6.019646645276027e-05, + "loss": 0.104, + "step": 26892 + }, + { + "epoch": 0.6304142712307412, + "grad_norm": 0.7240933179855347, + "learning_rate": 6.018970906792243e-05, + "loss": 0.1913, + "step": 26893 + }, + { + "epoch": 0.6304377128055463, + "grad_norm": 0.1993112862110138, + "learning_rate": 6.018295189909634e-05, + "loss": 0.0424, + "step": 26894 + }, + { + "epoch": 0.6304611543803512, + "grad_norm": 0.36835139989852905, + "learning_rate": 6.017619494631865e-05, + "loss": 0.0434, + "step": 26895 + }, + { + "epoch": 0.6304845959551563, + "grad_norm": 0.36441993713378906, + "learning_rate": 6.016943820962608e-05, + "loss": 0.0522, + "step": 26896 + }, + { + "epoch": 0.6305080375299613, + "grad_norm": 0.5249710083007812, + "learning_rate": 6.016268168905521e-05, + "loss": 0.1367, + "step": 26897 + }, + { + "epoch": 0.6305314791047663, + "grad_norm": 0.3268657326698303, + "learning_rate": 6.0155925384642784e-05, + "loss": 0.0772, + "step": 26898 + }, + { + "epoch": 0.6305549206795713, + "grad_norm": 0.3945034146308899, + "learning_rate": 6.014916929642541e-05, + "loss": 0.0705, + "step": 26899 + }, + { + "epoch": 0.6305783622543762, + "grad_norm": 0.18970173597335815, + "learning_rate": 6.014241342443979e-05, + "loss": 0.0275, + "step": 26900 + }, + { + "epoch": 0.6306018038291813, + "grad_norm": 0.18404914438724518, + "learning_rate": 6.013565776872255e-05, + "loss": 0.0354, + "step": 26901 + }, + { + "epoch": 0.6306252454039862, + "grad_norm": 0.31454944610595703, + "learning_rate": 6.012890232931034e-05, + "loss": 0.0624, + "step": 26902 + }, + { + "epoch": 0.6306486869787913, + "grad_norm": 0.24004264175891876, + "learning_rate": 6.012214710623983e-05, + "loss": 0.0454, + "step": 26903 + }, + { + "epoch": 0.6306721285535962, + "grad_norm": 0.2705572843551636, + "learning_rate": 6.0115392099547685e-05, + "loss": 0.0442, + "step": 26904 + }, + { + "epoch": 0.6306955701284013, + "grad_norm": 0.3437609374523163, + "learning_rate": 6.010863730927052e-05, + "loss": 0.0397, + "step": 26905 + }, + { + "epoch": 0.6307190117032062, + "grad_norm": 0.6552075147628784, + "learning_rate": 6.0101882735445036e-05, + "loss": 0.1182, + "step": 26906 + }, + { + "epoch": 0.6307424532780113, + "grad_norm": 0.11789724975824356, + "learning_rate": 6.009512837810785e-05, + "loss": 0.0197, + "step": 26907 + }, + { + "epoch": 0.6307658948528162, + "grad_norm": 0.36959776282310486, + "learning_rate": 6.008837423729561e-05, + "loss": 0.0489, + "step": 26908 + }, + { + "epoch": 0.6307893364276213, + "grad_norm": 0.37698066234588623, + "learning_rate": 6.008162031304494e-05, + "loss": 0.1082, + "step": 26909 + }, + { + "epoch": 0.6308127780024262, + "grad_norm": 0.7874116897583008, + "learning_rate": 6.0074866605392564e-05, + "loss": 0.0738, + "step": 26910 + }, + { + "epoch": 0.6308362195772312, + "grad_norm": 0.3096791207790375, + "learning_rate": 6.006811311437509e-05, + "loss": 0.0979, + "step": 26911 + }, + { + "epoch": 0.6308596611520362, + "grad_norm": 0.4202256500720978, + "learning_rate": 6.006135984002914e-05, + "loss": 0.1101, + "step": 26912 + }, + { + "epoch": 0.6308831027268412, + "grad_norm": 0.43092724680900574, + "learning_rate": 6.0054606782391386e-05, + "loss": 0.0778, + "step": 26913 + }, + { + "epoch": 0.6309065443016462, + "grad_norm": 0.554489254951477, + "learning_rate": 6.0047853941498456e-05, + "loss": 0.1106, + "step": 26914 + }, + { + "epoch": 0.6309299858764512, + "grad_norm": 0.37248316407203674, + "learning_rate": 6.0041101317386985e-05, + "loss": 0.0474, + "step": 26915 + }, + { + "epoch": 0.6309534274512562, + "grad_norm": 0.05362505838274956, + "learning_rate": 6.003434891009364e-05, + "loss": 0.0073, + "step": 26916 + }, + { + "epoch": 0.6309768690260612, + "grad_norm": 0.2480463832616806, + "learning_rate": 6.002759671965503e-05, + "loss": 0.0332, + "step": 26917 + }, + { + "epoch": 0.6310003106008661, + "grad_norm": 0.4074070453643799, + "learning_rate": 6.00208447461078e-05, + "loss": 0.4959, + "step": 26918 + }, + { + "epoch": 0.6310237521756712, + "grad_norm": 0.28855273127555847, + "learning_rate": 6.00140929894886e-05, + "loss": 0.0839, + "step": 26919 + }, + { + "epoch": 0.6310471937504761, + "grad_norm": 0.08518987894058228, + "learning_rate": 6.000734144983407e-05, + "loss": 0.0155, + "step": 26920 + }, + { + "epoch": 0.6310706353252812, + "grad_norm": 0.3481876850128174, + "learning_rate": 6.0000590127180766e-05, + "loss": 0.0672, + "step": 26921 + }, + { + "epoch": 0.6310940769000861, + "grad_norm": 0.382453590631485, + "learning_rate": 5.999383902156541e-05, + "loss": 0.0559, + "step": 26922 + }, + { + "epoch": 0.6311175184748912, + "grad_norm": 0.2452489584684372, + "learning_rate": 5.998708813302465e-05, + "loss": 0.0485, + "step": 26923 + }, + { + "epoch": 0.6311409600496961, + "grad_norm": 0.16850559413433075, + "learning_rate": 5.998033746159506e-05, + "loss": 0.0238, + "step": 26924 + }, + { + "epoch": 0.6311644016245012, + "grad_norm": 0.5000989437103271, + "learning_rate": 5.997358700731326e-05, + "loss": 0.0556, + "step": 26925 + }, + { + "epoch": 0.6311878431993061, + "grad_norm": 0.37164732813835144, + "learning_rate": 5.996683677021593e-05, + "loss": 0.0637, + "step": 26926 + }, + { + "epoch": 0.6312112847741111, + "grad_norm": 0.5155075788497925, + "learning_rate": 5.996008675033966e-05, + "loss": 0.0641, + "step": 26927 + }, + { + "epoch": 0.6312347263489161, + "grad_norm": 0.5409479141235352, + "learning_rate": 5.9953336947721074e-05, + "loss": 0.091, + "step": 26928 + }, + { + "epoch": 0.6312581679237211, + "grad_norm": 0.23104116320610046, + "learning_rate": 5.994658736239682e-05, + "loss": 0.0348, + "step": 26929 + }, + { + "epoch": 0.6312816094985261, + "grad_norm": 0.08977364003658295, + "learning_rate": 5.99398379944035e-05, + "loss": 0.0061, + "step": 26930 + }, + { + "epoch": 0.6313050510733311, + "grad_norm": 0.14998047053813934, + "learning_rate": 5.993308884377775e-05, + "loss": 0.0336, + "step": 26931 + }, + { + "epoch": 0.6313284926481361, + "grad_norm": 0.056085795164108276, + "learning_rate": 5.9926339910556206e-05, + "loss": 0.0088, + "step": 26932 + }, + { + "epoch": 0.6313519342229411, + "grad_norm": 0.14638091623783112, + "learning_rate": 5.991959119477544e-05, + "loss": 0.0186, + "step": 26933 + }, + { + "epoch": 0.631375375797746, + "grad_norm": 0.5900242328643799, + "learning_rate": 5.991284269647207e-05, + "loss": 0.1608, + "step": 26934 + }, + { + "epoch": 0.6313988173725511, + "grad_norm": 0.3089456260204315, + "learning_rate": 5.990609441568279e-05, + "loss": 0.0398, + "step": 26935 + }, + { + "epoch": 0.631422258947356, + "grad_norm": 0.6354747414588928, + "learning_rate": 5.9899346352444165e-05, + "loss": 0.1439, + "step": 26936 + }, + { + "epoch": 0.6314457005221611, + "grad_norm": 0.3804694414138794, + "learning_rate": 5.989259850679281e-05, + "loss": 0.0724, + "step": 26937 + }, + { + "epoch": 0.631469142096966, + "grad_norm": 0.3995286822319031, + "learning_rate": 5.988585087876535e-05, + "loss": 0.1431, + "step": 26938 + }, + { + "epoch": 0.6314925836717711, + "grad_norm": 0.12062391638755798, + "learning_rate": 5.987910346839839e-05, + "loss": 0.0136, + "step": 26939 + }, + { + "epoch": 0.631516025246576, + "grad_norm": 0.5805497765541077, + "learning_rate": 5.987235627572852e-05, + "loss": 0.0901, + "step": 26940 + }, + { + "epoch": 0.6315394668213811, + "grad_norm": 0.5005233883857727, + "learning_rate": 5.98656093007924e-05, + "loss": 0.1286, + "step": 26941 + }, + { + "epoch": 0.631562908396186, + "grad_norm": 0.3478303849697113, + "learning_rate": 5.985886254362661e-05, + "loss": 0.0806, + "step": 26942 + }, + { + "epoch": 0.6315863499709911, + "grad_norm": 0.31626826524734497, + "learning_rate": 5.9852116004267745e-05, + "loss": 0.0404, + "step": 26943 + }, + { + "epoch": 0.631609791545796, + "grad_norm": 0.23584309220314026, + "learning_rate": 5.984536968275245e-05, + "loss": 0.0197, + "step": 26944 + }, + { + "epoch": 0.631633233120601, + "grad_norm": 0.33547645807266235, + "learning_rate": 5.983862357911729e-05, + "loss": 0.04, + "step": 26945 + }, + { + "epoch": 0.631656674695406, + "grad_norm": 0.2744874060153961, + "learning_rate": 5.9831877693398866e-05, + "loss": 0.0557, + "step": 26946 + }, + { + "epoch": 0.631680116270211, + "grad_norm": 0.5757856965065002, + "learning_rate": 5.982513202563381e-05, + "loss": 0.1449, + "step": 26947 + }, + { + "epoch": 0.631703557845016, + "grad_norm": 0.1865862011909485, + "learning_rate": 5.981838657585874e-05, + "loss": 0.0222, + "step": 26948 + }, + { + "epoch": 0.631726999419821, + "grad_norm": 0.32557931542396545, + "learning_rate": 5.9811641344110246e-05, + "loss": 0.0264, + "step": 26949 + }, + { + "epoch": 0.6317504409946261, + "grad_norm": 0.6655750274658203, + "learning_rate": 5.9804896330424886e-05, + "loss": 0.0732, + "step": 26950 + }, + { + "epoch": 0.631773882569431, + "grad_norm": 0.3289265036582947, + "learning_rate": 5.979815153483932e-05, + "loss": 0.0321, + "step": 26951 + }, + { + "epoch": 0.6317973241442361, + "grad_norm": 0.19757753610610962, + "learning_rate": 5.979140695739009e-05, + "loss": 0.0194, + "step": 26952 + }, + { + "epoch": 0.631820765719041, + "grad_norm": 0.0854073315858841, + "learning_rate": 5.9784662598113816e-05, + "loss": 0.0161, + "step": 26953 + }, + { + "epoch": 0.631844207293846, + "grad_norm": 0.14673857390880585, + "learning_rate": 5.9777918457047125e-05, + "loss": 0.0221, + "step": 26954 + }, + { + "epoch": 0.631867648868651, + "grad_norm": 0.6670629978179932, + "learning_rate": 5.9771174534226556e-05, + "loss": 0.1198, + "step": 26955 + }, + { + "epoch": 0.631891090443456, + "grad_norm": 0.531995415687561, + "learning_rate": 5.976443082968871e-05, + "loss": 0.0999, + "step": 26956 + }, + { + "epoch": 0.631914532018261, + "grad_norm": 0.5257010459899902, + "learning_rate": 5.975768734347022e-05, + "loss": 0.1268, + "step": 26957 + }, + { + "epoch": 0.631937973593066, + "grad_norm": 0.17794786393642426, + "learning_rate": 5.9750944075607595e-05, + "loss": 0.0366, + "step": 26958 + }, + { + "epoch": 0.631961415167871, + "grad_norm": 0.1892293244600296, + "learning_rate": 5.974420102613754e-05, + "loss": 0.0254, + "step": 26959 + }, + { + "epoch": 0.631984856742676, + "grad_norm": 0.2982079088687897, + "learning_rate": 5.973745819509655e-05, + "loss": 0.0539, + "step": 26960 + }, + { + "epoch": 0.632008298317481, + "grad_norm": 0.37650710344314575, + "learning_rate": 5.9730715582521256e-05, + "loss": 0.0656, + "step": 26961 + }, + { + "epoch": 0.632031739892286, + "grad_norm": 0.5261242389678955, + "learning_rate": 5.972397318844825e-05, + "loss": 0.1068, + "step": 26962 + }, + { + "epoch": 0.6320551814670909, + "grad_norm": 0.2303164005279541, + "learning_rate": 5.9717231012914064e-05, + "loss": 0.051, + "step": 26963 + }, + { + "epoch": 0.632078623041896, + "grad_norm": 0.6249560713768005, + "learning_rate": 5.9710489055955356e-05, + "loss": 0.0967, + "step": 26964 + }, + { + "epoch": 0.6321020646167009, + "grad_norm": 1.0628881454467773, + "learning_rate": 5.9703747317608635e-05, + "loss": 0.1293, + "step": 26965 + }, + { + "epoch": 0.632125506191506, + "grad_norm": 0.6159721612930298, + "learning_rate": 5.969700579791052e-05, + "loss": 0.1188, + "step": 26966 + }, + { + "epoch": 0.6321489477663109, + "grad_norm": 0.3195740282535553, + "learning_rate": 5.9690264496897596e-05, + "loss": 0.0622, + "step": 26967 + }, + { + "epoch": 0.632172389341116, + "grad_norm": 0.3261503279209137, + "learning_rate": 5.9683523414606414e-05, + "loss": 0.0863, + "step": 26968 + }, + { + "epoch": 0.6321958309159209, + "grad_norm": 0.46962830424308777, + "learning_rate": 5.9676782551073586e-05, + "loss": 0.7279, + "step": 26969 + }, + { + "epoch": 0.632219272490726, + "grad_norm": 0.2471160590648651, + "learning_rate": 5.967004190633563e-05, + "loss": 0.0521, + "step": 26970 + }, + { + "epoch": 0.6322427140655309, + "grad_norm": 0.3709462881088257, + "learning_rate": 5.966330148042919e-05, + "loss": 0.0522, + "step": 26971 + }, + { + "epoch": 0.632266155640336, + "grad_norm": 0.6452446579933167, + "learning_rate": 5.96565612733908e-05, + "loss": 0.0873, + "step": 26972 + }, + { + "epoch": 0.6322895972151409, + "grad_norm": 0.19288776814937592, + "learning_rate": 5.964982128525707e-05, + "loss": 0.0454, + "step": 26973 + }, + { + "epoch": 0.6323130387899459, + "grad_norm": 0.2345106452703476, + "learning_rate": 5.9643081516064526e-05, + "loss": 0.0262, + "step": 26974 + }, + { + "epoch": 0.6323364803647509, + "grad_norm": 0.372359961271286, + "learning_rate": 5.963634196584975e-05, + "loss": 0.0875, + "step": 26975 + }, + { + "epoch": 0.6323599219395559, + "grad_norm": 0.1735285520553589, + "learning_rate": 5.962960263464933e-05, + "loss": 0.0285, + "step": 26976 + }, + { + "epoch": 0.6323833635143609, + "grad_norm": 0.33346858620643616, + "learning_rate": 5.9622863522499816e-05, + "loss": 0.0543, + "step": 26977 + }, + { + "epoch": 0.6324068050891659, + "grad_norm": 0.7683921456336975, + "learning_rate": 5.961612462943778e-05, + "loss": 0.0861, + "step": 26978 + }, + { + "epoch": 0.6324302466639709, + "grad_norm": 0.648972749710083, + "learning_rate": 5.960938595549978e-05, + "loss": 0.1376, + "step": 26979 + }, + { + "epoch": 0.6324536882387759, + "grad_norm": 0.24550195038318634, + "learning_rate": 5.960264750072242e-05, + "loss": 0.0562, + "step": 26980 + }, + { + "epoch": 0.6324771298135808, + "grad_norm": 0.41771939396858215, + "learning_rate": 5.959590926514219e-05, + "loss": 0.0696, + "step": 26981 + }, + { + "epoch": 0.6325005713883859, + "grad_norm": 0.3251994848251343, + "learning_rate": 5.9589171248795683e-05, + "loss": 0.0577, + "step": 26982 + }, + { + "epoch": 0.6325240129631908, + "grad_norm": 0.363831102848053, + "learning_rate": 5.95824334517195e-05, + "loss": 0.0282, + "step": 26983 + }, + { + "epoch": 0.6325474545379959, + "grad_norm": 0.1435777097940445, + "learning_rate": 5.957569587395016e-05, + "loss": 0.0214, + "step": 26984 + }, + { + "epoch": 0.6325708961128008, + "grad_norm": 0.19464291632175446, + "learning_rate": 5.956895851552423e-05, + "loss": 0.0356, + "step": 26985 + }, + { + "epoch": 0.6325943376876059, + "grad_norm": 0.43855899572372437, + "learning_rate": 5.9562221376478275e-05, + "loss": 0.0851, + "step": 26986 + }, + { + "epoch": 0.6326177792624108, + "grad_norm": 0.24871687591075897, + "learning_rate": 5.955548445684885e-05, + "loss": 0.0721, + "step": 26987 + }, + { + "epoch": 0.6326412208372159, + "grad_norm": 0.20129407942295074, + "learning_rate": 5.9548747756672485e-05, + "loss": 0.0458, + "step": 26988 + }, + { + "epoch": 0.6326646624120208, + "grad_norm": 0.28146812319755554, + "learning_rate": 5.9542011275985774e-05, + "loss": 0.0555, + "step": 26989 + }, + { + "epoch": 0.6326881039868258, + "grad_norm": 0.7367722988128662, + "learning_rate": 5.953527501482524e-05, + "loss": 0.0995, + "step": 26990 + }, + { + "epoch": 0.6327115455616308, + "grad_norm": 0.23604241013526917, + "learning_rate": 5.9528538973227434e-05, + "loss": 0.0418, + "step": 26991 + }, + { + "epoch": 0.6327349871364358, + "grad_norm": 0.15576528012752533, + "learning_rate": 5.9521803151228915e-05, + "loss": 0.0331, + "step": 26992 + }, + { + "epoch": 0.6327584287112408, + "grad_norm": 0.10577838867902756, + "learning_rate": 5.951506754886624e-05, + "loss": 0.0185, + "step": 26993 + }, + { + "epoch": 0.6327818702860458, + "grad_norm": 0.5022614598274231, + "learning_rate": 5.9508332166175906e-05, + "loss": 0.6286, + "step": 26994 + }, + { + "epoch": 0.6328053118608508, + "grad_norm": 0.29289278388023376, + "learning_rate": 5.950159700319452e-05, + "loss": 0.0385, + "step": 26995 + }, + { + "epoch": 0.6328287534356558, + "grad_norm": 0.25448256731033325, + "learning_rate": 5.9494862059958624e-05, + "loss": 0.0417, + "step": 26996 + }, + { + "epoch": 0.6328521950104607, + "grad_norm": 0.34776678681373596, + "learning_rate": 5.948812733650475e-05, + "loss": 0.0479, + "step": 26997 + }, + { + "epoch": 0.6328756365852658, + "grad_norm": 0.20912295579910278, + "learning_rate": 5.948139283286942e-05, + "loss": 0.0537, + "step": 26998 + }, + { + "epoch": 0.6328990781600707, + "grad_norm": 0.41100984811782837, + "learning_rate": 5.94746585490892e-05, + "loss": 0.1215, + "step": 26999 + }, + { + "epoch": 0.6329225197348758, + "grad_norm": 0.6307018995285034, + "learning_rate": 5.9467924485200644e-05, + "loss": 0.0824, + "step": 27000 + }, + { + "epoch": 0.6329459613096808, + "grad_norm": 0.7017253041267395, + "learning_rate": 5.9461190641240226e-05, + "loss": 0.1244, + "step": 27001 + }, + { + "epoch": 0.6329694028844858, + "grad_norm": 0.4953352212905884, + "learning_rate": 5.945445701724456e-05, + "loss": 0.0506, + "step": 27002 + }, + { + "epoch": 0.6329928444592908, + "grad_norm": 0.22991588711738586, + "learning_rate": 5.9447723613250125e-05, + "loss": 0.0313, + "step": 27003 + }, + { + "epoch": 0.6330162860340958, + "grad_norm": 0.6386599540710449, + "learning_rate": 5.94409904292935e-05, + "loss": 0.1711, + "step": 27004 + }, + { + "epoch": 0.6330397276089008, + "grad_norm": 0.3328632414340973, + "learning_rate": 5.94342574654112e-05, + "loss": 0.0559, + "step": 27005 + }, + { + "epoch": 0.6330631691837058, + "grad_norm": 0.5185007452964783, + "learning_rate": 5.942752472163975e-05, + "loss": 0.1189, + "step": 27006 + }, + { + "epoch": 0.6330866107585108, + "grad_norm": 0.9415832161903381, + "learning_rate": 5.9420792198015686e-05, + "loss": 0.1993, + "step": 27007 + }, + { + "epoch": 0.6331100523333157, + "grad_norm": 0.5017757415771484, + "learning_rate": 5.941405989457555e-05, + "loss": 0.1061, + "step": 27008 + }, + { + "epoch": 0.6331334939081208, + "grad_norm": 0.5571178197860718, + "learning_rate": 5.9407327811355876e-05, + "loss": 0.703, + "step": 27009 + }, + { + "epoch": 0.6331569354829257, + "grad_norm": 0.12635740637779236, + "learning_rate": 5.940059594839317e-05, + "loss": 0.0141, + "step": 27010 + }, + { + "epoch": 0.6331803770577308, + "grad_norm": 0.10803849250078201, + "learning_rate": 5.939386430572399e-05, + "loss": 0.0085, + "step": 27011 + }, + { + "epoch": 0.6332038186325357, + "grad_norm": 0.1633174568414688, + "learning_rate": 5.9387132883384844e-05, + "loss": 0.0166, + "step": 27012 + }, + { + "epoch": 0.6332272602073408, + "grad_norm": 0.19167330861091614, + "learning_rate": 5.938040168141224e-05, + "loss": 0.0275, + "step": 27013 + }, + { + "epoch": 0.6332507017821457, + "grad_norm": 0.10508021712303162, + "learning_rate": 5.937367069984273e-05, + "loss": 0.0187, + "step": 27014 + }, + { + "epoch": 0.6332741433569508, + "grad_norm": 0.29188427329063416, + "learning_rate": 5.936693993871283e-05, + "loss": 0.3885, + "step": 27015 + }, + { + "epoch": 0.6332975849317557, + "grad_norm": 0.8150923252105713, + "learning_rate": 5.936020939805904e-05, + "loss": 0.1228, + "step": 27016 + }, + { + "epoch": 0.6333210265065607, + "grad_norm": 0.609635055065155, + "learning_rate": 5.9353479077917906e-05, + "loss": 0.1637, + "step": 27017 + }, + { + "epoch": 0.6333444680813657, + "grad_norm": 0.4354948401451111, + "learning_rate": 5.934674897832595e-05, + "loss": 0.3839, + "step": 27018 + }, + { + "epoch": 0.6333679096561707, + "grad_norm": 0.24756547808647156, + "learning_rate": 5.934001909931963e-05, + "loss": 0.0213, + "step": 27019 + }, + { + "epoch": 0.6333913512309757, + "grad_norm": 0.294614315032959, + "learning_rate": 5.933328944093552e-05, + "loss": 0.0587, + "step": 27020 + }, + { + "epoch": 0.6334147928057807, + "grad_norm": 0.25969740748405457, + "learning_rate": 5.9326560003210166e-05, + "loss": 0.0409, + "step": 27021 + }, + { + "epoch": 0.6334382343805857, + "grad_norm": 0.27576059103012085, + "learning_rate": 5.931983078618003e-05, + "loss": 0.0332, + "step": 27022 + }, + { + "epoch": 0.6334616759553907, + "grad_norm": 0.503441572189331, + "learning_rate": 5.931310178988162e-05, + "loss": 0.059, + "step": 27023 + }, + { + "epoch": 0.6334851175301957, + "grad_norm": 0.3225933015346527, + "learning_rate": 5.930637301435148e-05, + "loss": 0.0752, + "step": 27024 + }, + { + "epoch": 0.6335085591050007, + "grad_norm": 0.12494121491909027, + "learning_rate": 5.92996444596261e-05, + "loss": 0.0195, + "step": 27025 + }, + { + "epoch": 0.6335320006798056, + "grad_norm": 0.3667072653770447, + "learning_rate": 5.929291612574198e-05, + "loss": 0.0733, + "step": 27026 + }, + { + "epoch": 0.6335554422546107, + "grad_norm": 0.4559006989002228, + "learning_rate": 5.9286188012735664e-05, + "loss": 0.1342, + "step": 27027 + }, + { + "epoch": 0.6335788838294156, + "grad_norm": 0.21807508170604706, + "learning_rate": 5.9279460120643625e-05, + "loss": 0.0309, + "step": 27028 + }, + { + "epoch": 0.6336023254042207, + "grad_norm": 0.29906532168388367, + "learning_rate": 5.9272732449502374e-05, + "loss": 0.0447, + "step": 27029 + }, + { + "epoch": 0.6336257669790256, + "grad_norm": 0.5248858332633972, + "learning_rate": 5.9266004999348434e-05, + "loss": 0.0594, + "step": 27030 + }, + { + "epoch": 0.6336492085538307, + "grad_norm": 0.9344056248664856, + "learning_rate": 5.925927777021826e-05, + "loss": 0.113, + "step": 27031 + }, + { + "epoch": 0.6336726501286356, + "grad_norm": 1.3290550708770752, + "learning_rate": 5.925255076214843e-05, + "loss": 0.3166, + "step": 27032 + }, + { + "epoch": 0.6336960917034407, + "grad_norm": 0.45832040905952454, + "learning_rate": 5.9245823975175375e-05, + "loss": 0.0842, + "step": 27033 + }, + { + "epoch": 0.6337195332782456, + "grad_norm": 0.4443831443786621, + "learning_rate": 5.923909740933566e-05, + "loss": 0.0467, + "step": 27034 + }, + { + "epoch": 0.6337429748530506, + "grad_norm": 0.16865722835063934, + "learning_rate": 5.923237106466575e-05, + "loss": 0.0242, + "step": 27035 + }, + { + "epoch": 0.6337664164278556, + "grad_norm": 0.5846245884895325, + "learning_rate": 5.9225644941202104e-05, + "loss": 0.1287, + "step": 27036 + }, + { + "epoch": 0.6337898580026606, + "grad_norm": 0.3770008981227875, + "learning_rate": 5.921891903898129e-05, + "loss": 0.0482, + "step": 27037 + }, + { + "epoch": 0.6338132995774656, + "grad_norm": 0.16663973033428192, + "learning_rate": 5.921219335803976e-05, + "loss": 0.0247, + "step": 27038 + }, + { + "epoch": 0.6338367411522706, + "grad_norm": 0.39973199367523193, + "learning_rate": 5.920546789841399e-05, + "loss": 0.0411, + "step": 27039 + }, + { + "epoch": 0.6338601827270756, + "grad_norm": 0.45107072591781616, + "learning_rate": 5.9198742660140536e-05, + "loss": 0.0673, + "step": 27040 + }, + { + "epoch": 0.6338836243018806, + "grad_norm": 0.5609005093574524, + "learning_rate": 5.919201764325581e-05, + "loss": 0.7118, + "step": 27041 + }, + { + "epoch": 0.6339070658766855, + "grad_norm": 0.44209277629852295, + "learning_rate": 5.918529284779639e-05, + "loss": 0.0398, + "step": 27042 + }, + { + "epoch": 0.6339305074514906, + "grad_norm": 0.656520664691925, + "learning_rate": 5.9178568273798666e-05, + "loss": 0.071, + "step": 27043 + }, + { + "epoch": 0.6339539490262955, + "grad_norm": 0.40454885363578796, + "learning_rate": 5.917184392129921e-05, + "loss": 0.0471, + "step": 27044 + }, + { + "epoch": 0.6339773906011006, + "grad_norm": 0.5932125449180603, + "learning_rate": 5.916511979033448e-05, + "loss": 0.0918, + "step": 27045 + }, + { + "epoch": 0.6340008321759055, + "grad_norm": 0.11659060418605804, + "learning_rate": 5.915839588094094e-05, + "loss": 0.0123, + "step": 27046 + }, + { + "epoch": 0.6340242737507106, + "grad_norm": 0.18812522292137146, + "learning_rate": 5.915167219315512e-05, + "loss": 0.0389, + "step": 27047 + }, + { + "epoch": 0.6340477153255155, + "grad_norm": 0.5808197855949402, + "learning_rate": 5.914494872701345e-05, + "loss": 0.0856, + "step": 27048 + }, + { + "epoch": 0.6340711569003206, + "grad_norm": 0.3093724548816681, + "learning_rate": 5.913822548255246e-05, + "loss": 0.0623, + "step": 27049 + }, + { + "epoch": 0.6340945984751255, + "grad_norm": 0.4915456175804138, + "learning_rate": 5.9131502459808596e-05, + "loss": 0.0436, + "step": 27050 + }, + { + "epoch": 0.6341180400499306, + "grad_norm": 0.6766019463539124, + "learning_rate": 5.912477965881833e-05, + "loss": 0.0925, + "step": 27051 + }, + { + "epoch": 0.6341414816247356, + "grad_norm": 0.3734784424304962, + "learning_rate": 5.911805707961819e-05, + "loss": 0.0649, + "step": 27052 + }, + { + "epoch": 0.6341649231995405, + "grad_norm": 0.4567638635635376, + "learning_rate": 5.9111334722244615e-05, + "loss": 0.071, + "step": 27053 + }, + { + "epoch": 0.6341883647743456, + "grad_norm": 0.35824668407440186, + "learning_rate": 5.9104612586734074e-05, + "loss": 0.0716, + "step": 27054 + }, + { + "epoch": 0.6342118063491505, + "grad_norm": 0.6663918495178223, + "learning_rate": 5.909789067312304e-05, + "loss": 0.1138, + "step": 27055 + }, + { + "epoch": 0.6342352479239556, + "grad_norm": 0.40690913796424866, + "learning_rate": 5.9091168981448016e-05, + "loss": 0.069, + "step": 27056 + }, + { + "epoch": 0.6342586894987605, + "grad_norm": 0.5341787934303284, + "learning_rate": 5.908444751174548e-05, + "loss": 0.1292, + "step": 27057 + }, + { + "epoch": 0.6342821310735656, + "grad_norm": 0.1916477531194687, + "learning_rate": 5.907772626405186e-05, + "loss": 0.0242, + "step": 27058 + }, + { + "epoch": 0.6343055726483705, + "grad_norm": 0.5167561173439026, + "learning_rate": 5.907100523840366e-05, + "loss": 0.0295, + "step": 27059 + }, + { + "epoch": 0.6343290142231756, + "grad_norm": 0.32174062728881836, + "learning_rate": 5.906428443483734e-05, + "loss": 0.0568, + "step": 27060 + }, + { + "epoch": 0.6343524557979805, + "grad_norm": 0.36319243907928467, + "learning_rate": 5.905756385338934e-05, + "loss": 0.3094, + "step": 27061 + }, + { + "epoch": 0.6343758973727855, + "grad_norm": 0.17169086635112762, + "learning_rate": 5.905084349409618e-05, + "loss": 0.0237, + "step": 27062 + }, + { + "epoch": 0.6343993389475905, + "grad_norm": 0.6368560791015625, + "learning_rate": 5.9044123356994296e-05, + "loss": 0.0857, + "step": 27063 + }, + { + "epoch": 0.6344227805223955, + "grad_norm": 0.1805488020181656, + "learning_rate": 5.903740344212012e-05, + "loss": 0.0245, + "step": 27064 + }, + { + "epoch": 0.6344462220972005, + "grad_norm": 0.2791651487350464, + "learning_rate": 5.903068374951017e-05, + "loss": 0.0464, + "step": 27065 + }, + { + "epoch": 0.6344696636720055, + "grad_norm": 0.24668937921524048, + "learning_rate": 5.902396427920088e-05, + "loss": 0.039, + "step": 27066 + }, + { + "epoch": 0.6344931052468105, + "grad_norm": 0.4911058843135834, + "learning_rate": 5.901724503122867e-05, + "loss": 0.095, + "step": 27067 + }, + { + "epoch": 0.6345165468216155, + "grad_norm": 0.5510976910591125, + "learning_rate": 5.901052600563006e-05, + "loss": 0.0548, + "step": 27068 + }, + { + "epoch": 0.6345399883964205, + "grad_norm": 0.11873436719179153, + "learning_rate": 5.9003807202441516e-05, + "loss": 0.0154, + "step": 27069 + }, + { + "epoch": 0.6345634299712255, + "grad_norm": 0.3749217689037323, + "learning_rate": 5.899708862169946e-05, + "loss": 0.0825, + "step": 27070 + }, + { + "epoch": 0.6345868715460304, + "grad_norm": 0.26013055443763733, + "learning_rate": 5.899037026344034e-05, + "loss": 0.0426, + "step": 27071 + }, + { + "epoch": 0.6346103131208355, + "grad_norm": 0.5875470042228699, + "learning_rate": 5.8983652127700636e-05, + "loss": 0.0808, + "step": 27072 + }, + { + "epoch": 0.6346337546956404, + "grad_norm": 0.9388850331306458, + "learning_rate": 5.89769342145168e-05, + "loss": 0.7763, + "step": 27073 + }, + { + "epoch": 0.6346571962704455, + "grad_norm": 0.15168967843055725, + "learning_rate": 5.897021652392525e-05, + "loss": 0.0203, + "step": 27074 + }, + { + "epoch": 0.6346806378452504, + "grad_norm": 0.3170040249824524, + "learning_rate": 5.896349905596247e-05, + "loss": 0.0604, + "step": 27075 + }, + { + "epoch": 0.6347040794200555, + "grad_norm": 0.8021150827407837, + "learning_rate": 5.895678181066491e-05, + "loss": 0.1301, + "step": 27076 + }, + { + "epoch": 0.6347275209948604, + "grad_norm": 0.33060532808303833, + "learning_rate": 5.895006478806897e-05, + "loss": 0.0469, + "step": 27077 + }, + { + "epoch": 0.6347509625696655, + "grad_norm": 0.6539751887321472, + "learning_rate": 5.8943347988211165e-05, + "loss": 0.0754, + "step": 27078 + }, + { + "epoch": 0.6347744041444704, + "grad_norm": 0.5168921947479248, + "learning_rate": 5.8936631411127864e-05, + "loss": 0.0686, + "step": 27079 + }, + { + "epoch": 0.6347978457192754, + "grad_norm": 0.6468834280967712, + "learning_rate": 5.892991505685559e-05, + "loss": 0.1195, + "step": 27080 + }, + { + "epoch": 0.6348212872940804, + "grad_norm": 0.6875078678131104, + "learning_rate": 5.8923198925430746e-05, + "loss": 0.1307, + "step": 27081 + }, + { + "epoch": 0.6348447288688854, + "grad_norm": 0.15995006263256073, + "learning_rate": 5.891648301688979e-05, + "loss": 0.0394, + "step": 27082 + }, + { + "epoch": 0.6348681704436904, + "grad_norm": 0.7815718650817871, + "learning_rate": 5.890976733126915e-05, + "loss": 0.2326, + "step": 27083 + }, + { + "epoch": 0.6348916120184954, + "grad_norm": 0.22816714644432068, + "learning_rate": 5.890305186860525e-05, + "loss": 0.0636, + "step": 27084 + }, + { + "epoch": 0.6349150535933004, + "grad_norm": 0.4993482530117035, + "learning_rate": 5.8896336628934566e-05, + "loss": 0.0508, + "step": 27085 + }, + { + "epoch": 0.6349384951681054, + "grad_norm": 0.4554744064807892, + "learning_rate": 5.88896216122935e-05, + "loss": 0.0877, + "step": 27086 + }, + { + "epoch": 0.6349619367429103, + "grad_norm": 0.3795296847820282, + "learning_rate": 5.888290681871852e-05, + "loss": 0.04, + "step": 27087 + }, + { + "epoch": 0.6349853783177154, + "grad_norm": 0.5124405026435852, + "learning_rate": 5.887619224824604e-05, + "loss": 0.6498, + "step": 27088 + }, + { + "epoch": 0.6350088198925203, + "grad_norm": 0.3106321096420288, + "learning_rate": 5.886947790091248e-05, + "loss": 0.0339, + "step": 27089 + }, + { + "epoch": 0.6350322614673254, + "grad_norm": 0.6626781225204468, + "learning_rate": 5.8862763776754306e-05, + "loss": 0.0755, + "step": 27090 + }, + { + "epoch": 0.6350557030421303, + "grad_norm": 0.2915557622909546, + "learning_rate": 5.885604987580794e-05, + "loss": 0.0707, + "step": 27091 + }, + { + "epoch": 0.6350791446169354, + "grad_norm": 0.3190003037452698, + "learning_rate": 5.884933619810976e-05, + "loss": 0.0522, + "step": 27092 + }, + { + "epoch": 0.6351025861917403, + "grad_norm": 0.32827261090278625, + "learning_rate": 5.884262274369626e-05, + "loss": 0.058, + "step": 27093 + }, + { + "epoch": 0.6351260277665454, + "grad_norm": 0.3150852620601654, + "learning_rate": 5.883590951260387e-05, + "loss": 0.0538, + "step": 27094 + }, + { + "epoch": 0.6351494693413503, + "grad_norm": 0.1018272340297699, + "learning_rate": 5.882919650486898e-05, + "loss": 0.0249, + "step": 27095 + }, + { + "epoch": 0.6351729109161554, + "grad_norm": 0.13435161113739014, + "learning_rate": 5.882248372052802e-05, + "loss": 0.0281, + "step": 27096 + }, + { + "epoch": 0.6351963524909603, + "grad_norm": 0.3752656579017639, + "learning_rate": 5.881577115961744e-05, + "loss": 0.091, + "step": 27097 + }, + { + "epoch": 0.6352197940657653, + "grad_norm": 0.33328792452812195, + "learning_rate": 5.880905882217363e-05, + "loss": 0.0716, + "step": 27098 + }, + { + "epoch": 0.6352432356405703, + "grad_norm": 0.1871107965707779, + "learning_rate": 5.880234670823302e-05, + "loss": 0.0391, + "step": 27099 + }, + { + "epoch": 0.6352666772153753, + "grad_norm": 0.20338274538516998, + "learning_rate": 5.879563481783206e-05, + "loss": 0.0363, + "step": 27100 + }, + { + "epoch": 0.6352901187901803, + "grad_norm": 0.1057959645986557, + "learning_rate": 5.878892315100713e-05, + "loss": 0.0266, + "step": 27101 + }, + { + "epoch": 0.6353135603649853, + "grad_norm": 0.4732954204082489, + "learning_rate": 5.878221170779464e-05, + "loss": 0.1182, + "step": 27102 + }, + { + "epoch": 0.6353370019397904, + "grad_norm": 0.22912991046905518, + "learning_rate": 5.877550048823106e-05, + "loss": 0.0167, + "step": 27103 + }, + { + "epoch": 0.6353604435145953, + "grad_norm": 0.8921862244606018, + "learning_rate": 5.876878949235273e-05, + "loss": 0.6685, + "step": 27104 + }, + { + "epoch": 0.6353838850894004, + "grad_norm": 0.8468694686889648, + "learning_rate": 5.8762078720196144e-05, + "loss": 0.1648, + "step": 27105 + }, + { + "epoch": 0.6354073266642053, + "grad_norm": 0.41685789823532104, + "learning_rate": 5.875536817179765e-05, + "loss": 0.0607, + "step": 27106 + }, + { + "epoch": 0.6354307682390103, + "grad_norm": 0.4826701283454895, + "learning_rate": 5.8748657847193725e-05, + "loss": 0.0944, + "step": 27107 + }, + { + "epoch": 0.6354542098138153, + "grad_norm": 0.3343803882598877, + "learning_rate": 5.874194774642074e-05, + "loss": 0.0537, + "step": 27108 + }, + { + "epoch": 0.6354776513886203, + "grad_norm": 0.5660693049430847, + "learning_rate": 5.873523786951507e-05, + "loss": 0.1321, + "step": 27109 + }, + { + "epoch": 0.6355010929634253, + "grad_norm": 0.48483729362487793, + "learning_rate": 5.8728528216513204e-05, + "loss": 0.1423, + "step": 27110 + }, + { + "epoch": 0.6355245345382303, + "grad_norm": 0.390941321849823, + "learning_rate": 5.872181878745149e-05, + "loss": 0.0367, + "step": 27111 + }, + { + "epoch": 0.6355479761130353, + "grad_norm": 0.24267996847629547, + "learning_rate": 5.871510958236632e-05, + "loss": 0.0567, + "step": 27112 + }, + { + "epoch": 0.6355714176878403, + "grad_norm": 0.4632917642593384, + "learning_rate": 5.870840060129417e-05, + "loss": 0.498, + "step": 27113 + }, + { + "epoch": 0.6355948592626453, + "grad_norm": 0.18810874223709106, + "learning_rate": 5.870169184427138e-05, + "loss": 0.0273, + "step": 27114 + }, + { + "epoch": 0.6356183008374503, + "grad_norm": 0.49852749705314636, + "learning_rate": 5.869498331133436e-05, + "loss": 0.0775, + "step": 27115 + }, + { + "epoch": 0.6356417424122552, + "grad_norm": 0.3425953686237335, + "learning_rate": 5.86882750025195e-05, + "loss": 0.0923, + "step": 27116 + }, + { + "epoch": 0.6356651839870603, + "grad_norm": 0.6634798645973206, + "learning_rate": 5.868156691786325e-05, + "loss": 0.6652, + "step": 27117 + }, + { + "epoch": 0.6356886255618652, + "grad_norm": 0.5940936207771301, + "learning_rate": 5.8674859057402e-05, + "loss": 0.1145, + "step": 27118 + }, + { + "epoch": 0.6357120671366703, + "grad_norm": 0.407697468996048, + "learning_rate": 5.866815142117209e-05, + "loss": 0.08, + "step": 27119 + }, + { + "epoch": 0.6357355087114752, + "grad_norm": 0.23250725865364075, + "learning_rate": 5.866144400920997e-05, + "loss": 0.054, + "step": 27120 + }, + { + "epoch": 0.6357589502862803, + "grad_norm": 0.4995424747467041, + "learning_rate": 5.8654736821552026e-05, + "loss": 0.0851, + "step": 27121 + }, + { + "epoch": 0.6357823918610852, + "grad_norm": 0.4961957633495331, + "learning_rate": 5.864802985823463e-05, + "loss": 0.0724, + "step": 27122 + }, + { + "epoch": 0.6358058334358903, + "grad_norm": 0.49536430835723877, + "learning_rate": 5.86413231192942e-05, + "loss": 0.0655, + "step": 27123 + }, + { + "epoch": 0.6358292750106952, + "grad_norm": 0.5091007351875305, + "learning_rate": 5.863461660476709e-05, + "loss": 0.1342, + "step": 27124 + }, + { + "epoch": 0.6358527165855002, + "grad_norm": 0.42028579115867615, + "learning_rate": 5.862791031468974e-05, + "loss": 0.0631, + "step": 27125 + }, + { + "epoch": 0.6358761581603052, + "grad_norm": 0.4672226905822754, + "learning_rate": 5.86212042490985e-05, + "loss": 0.0532, + "step": 27126 + }, + { + "epoch": 0.6358995997351102, + "grad_norm": 0.10943581908941269, + "learning_rate": 5.8614498408029764e-05, + "loss": 0.0425, + "step": 27127 + }, + { + "epoch": 0.6359230413099152, + "grad_norm": 0.4056168794631958, + "learning_rate": 5.860779279151989e-05, + "loss": 0.087, + "step": 27128 + }, + { + "epoch": 0.6359464828847202, + "grad_norm": 0.45819979906082153, + "learning_rate": 5.8601087399605345e-05, + "loss": 0.0675, + "step": 27129 + }, + { + "epoch": 0.6359699244595252, + "grad_norm": 0.647514820098877, + "learning_rate": 5.859438223232246e-05, + "loss": 0.1033, + "step": 27130 + }, + { + "epoch": 0.6359933660343302, + "grad_norm": 0.3126155734062195, + "learning_rate": 5.858767728970761e-05, + "loss": 0.052, + "step": 27131 + }, + { + "epoch": 0.6360168076091351, + "grad_norm": 0.33145835995674133, + "learning_rate": 5.85809725717972e-05, + "loss": 0.0722, + "step": 27132 + }, + { + "epoch": 0.6360402491839402, + "grad_norm": 0.4795321524143219, + "learning_rate": 5.85742680786276e-05, + "loss": 0.1398, + "step": 27133 + }, + { + "epoch": 0.6360636907587451, + "grad_norm": 0.4700043499469757, + "learning_rate": 5.856756381023516e-05, + "loss": 0.0942, + "step": 27134 + }, + { + "epoch": 0.6360871323335502, + "grad_norm": 0.6162118911743164, + "learning_rate": 5.856085976665631e-05, + "loss": 0.1117, + "step": 27135 + }, + { + "epoch": 0.6361105739083551, + "grad_norm": 0.565345823764801, + "learning_rate": 5.855415594792739e-05, + "loss": 0.7041, + "step": 27136 + }, + { + "epoch": 0.6361340154831602, + "grad_norm": 0.5015655755996704, + "learning_rate": 5.854745235408479e-05, + "loss": 0.1046, + "step": 27137 + }, + { + "epoch": 0.6361574570579651, + "grad_norm": 0.44423091411590576, + "learning_rate": 5.8540748985164876e-05, + "loss": 0.126, + "step": 27138 + }, + { + "epoch": 0.6361808986327702, + "grad_norm": 0.28895992040634155, + "learning_rate": 5.8534045841204035e-05, + "loss": 0.0468, + "step": 27139 + }, + { + "epoch": 0.6362043402075751, + "grad_norm": 0.27403461933135986, + "learning_rate": 5.852734292223857e-05, + "loss": 0.0663, + "step": 27140 + }, + { + "epoch": 0.6362277817823802, + "grad_norm": 0.25560906529426575, + "learning_rate": 5.8520640228304944e-05, + "loss": 0.0434, + "step": 27141 + }, + { + "epoch": 0.6362512233571851, + "grad_norm": 0.5219300985336304, + "learning_rate": 5.85139377594395e-05, + "loss": 0.0738, + "step": 27142 + }, + { + "epoch": 0.6362746649319901, + "grad_norm": 0.6478857398033142, + "learning_rate": 5.85072355156786e-05, + "loss": 0.1017, + "step": 27143 + }, + { + "epoch": 0.6362981065067951, + "grad_norm": 0.3666751980781555, + "learning_rate": 5.8500533497058595e-05, + "loss": 0.0761, + "step": 27144 + }, + { + "epoch": 0.6363215480816001, + "grad_norm": 0.7095387578010559, + "learning_rate": 5.849383170361589e-05, + "loss": 0.1554, + "step": 27145 + }, + { + "epoch": 0.6363449896564051, + "grad_norm": 0.6849818229675293, + "learning_rate": 5.8487130135386805e-05, + "loss": 0.1403, + "step": 27146 + }, + { + "epoch": 0.6363684312312101, + "grad_norm": 0.48976415395736694, + "learning_rate": 5.848042879240769e-05, + "loss": 0.6534, + "step": 27147 + }, + { + "epoch": 0.6363918728060151, + "grad_norm": 0.646484911441803, + "learning_rate": 5.847372767471497e-05, + "loss": 0.0616, + "step": 27148 + }, + { + "epoch": 0.6364153143808201, + "grad_norm": 0.3445233106613159, + "learning_rate": 5.846702678234497e-05, + "loss": 0.0701, + "step": 27149 + }, + { + "epoch": 0.636438755955625, + "grad_norm": 0.6204058527946472, + "learning_rate": 5.846032611533403e-05, + "loss": 0.1157, + "step": 27150 + }, + { + "epoch": 0.6364621975304301, + "grad_norm": 0.06602124869823456, + "learning_rate": 5.845362567371855e-05, + "loss": 0.0096, + "step": 27151 + }, + { + "epoch": 0.636485639105235, + "grad_norm": 0.22302000224590302, + "learning_rate": 5.844692545753483e-05, + "loss": 0.0405, + "step": 27152 + }, + { + "epoch": 0.6365090806800401, + "grad_norm": 0.23191677033901215, + "learning_rate": 5.844022546681929e-05, + "loss": 0.0256, + "step": 27153 + }, + { + "epoch": 0.6365325222548451, + "grad_norm": 0.6354242563247681, + "learning_rate": 5.8433525701608236e-05, + "loss": 0.1437, + "step": 27154 + }, + { + "epoch": 0.6365559638296501, + "grad_norm": 0.4752725064754486, + "learning_rate": 5.842682616193806e-05, + "loss": 0.684, + "step": 27155 + }, + { + "epoch": 0.6365794054044551, + "grad_norm": 0.19868767261505127, + "learning_rate": 5.84201268478451e-05, + "loss": 0.0313, + "step": 27156 + }, + { + "epoch": 0.6366028469792601, + "grad_norm": 0.3240823745727539, + "learning_rate": 5.841342775936568e-05, + "loss": 0.0828, + "step": 27157 + }, + { + "epoch": 0.6366262885540651, + "grad_norm": 0.4591093957424164, + "learning_rate": 5.840672889653619e-05, + "loss": 0.1108, + "step": 27158 + }, + { + "epoch": 0.63664973012887, + "grad_norm": 0.3797561228275299, + "learning_rate": 5.8400030259392936e-05, + "loss": 0.0227, + "step": 27159 + }, + { + "epoch": 0.6366731717036751, + "grad_norm": 0.22497305274009705, + "learning_rate": 5.839333184797231e-05, + "loss": 0.0307, + "step": 27160 + }, + { + "epoch": 0.63669661327848, + "grad_norm": 0.35995569825172424, + "learning_rate": 5.838663366231063e-05, + "loss": 0.0557, + "step": 27161 + }, + { + "epoch": 0.6367200548532851, + "grad_norm": 0.45988965034484863, + "learning_rate": 5.8379935702444256e-05, + "loss": 0.1093, + "step": 27162 + }, + { + "epoch": 0.63674349642809, + "grad_norm": 0.6870259642601013, + "learning_rate": 5.837323796840952e-05, + "loss": 0.1432, + "step": 27163 + }, + { + "epoch": 0.6367669380028951, + "grad_norm": 0.18109993636608124, + "learning_rate": 5.836654046024271e-05, + "loss": 0.0436, + "step": 27164 + }, + { + "epoch": 0.6367903795777, + "grad_norm": 0.36962831020355225, + "learning_rate": 5.8359843177980244e-05, + "loss": 0.0425, + "step": 27165 + }, + { + "epoch": 0.6368138211525051, + "grad_norm": 0.8990886211395264, + "learning_rate": 5.835314612165846e-05, + "loss": 0.2527, + "step": 27166 + }, + { + "epoch": 0.63683726272731, + "grad_norm": 0.3267107903957367, + "learning_rate": 5.834644929131369e-05, + "loss": 0.0365, + "step": 27167 + }, + { + "epoch": 0.636860704302115, + "grad_norm": 0.28245648741722107, + "learning_rate": 5.833975268698224e-05, + "loss": 0.0413, + "step": 27168 + }, + { + "epoch": 0.63688414587692, + "grad_norm": 0.21133220195770264, + "learning_rate": 5.8333056308700485e-05, + "loss": 0.0215, + "step": 27169 + }, + { + "epoch": 0.636907587451725, + "grad_norm": 0.5514341592788696, + "learning_rate": 5.832636015650472e-05, + "loss": 0.0941, + "step": 27170 + }, + { + "epoch": 0.63693102902653, + "grad_norm": 0.46340304613113403, + "learning_rate": 5.831966423043127e-05, + "loss": 0.0789, + "step": 27171 + }, + { + "epoch": 0.636954470601335, + "grad_norm": 0.4820004999637604, + "learning_rate": 5.831296853051653e-05, + "loss": 0.6499, + "step": 27172 + }, + { + "epoch": 0.63697791217614, + "grad_norm": 0.18476814031600952, + "learning_rate": 5.830627305679679e-05, + "loss": 0.0493, + "step": 27173 + }, + { + "epoch": 0.637001353750945, + "grad_norm": 0.5472596287727356, + "learning_rate": 5.8299577809308394e-05, + "loss": 0.5818, + "step": 27174 + }, + { + "epoch": 0.63702479532575, + "grad_norm": 0.33685556054115295, + "learning_rate": 5.8292882788087656e-05, + "loss": 0.0795, + "step": 27175 + }, + { + "epoch": 0.637048236900555, + "grad_norm": 0.8168653249740601, + "learning_rate": 5.828618799317087e-05, + "loss": 0.0999, + "step": 27176 + }, + { + "epoch": 0.63707167847536, + "grad_norm": 0.3236829340457916, + "learning_rate": 5.827949342459443e-05, + "loss": 0.2842, + "step": 27177 + }, + { + "epoch": 0.637095120050165, + "grad_norm": 0.32967609167099, + "learning_rate": 5.82727990823946e-05, + "loss": 0.0581, + "step": 27178 + }, + { + "epoch": 0.6371185616249699, + "grad_norm": 0.3977374732494354, + "learning_rate": 5.826610496660778e-05, + "loss": 0.1186, + "step": 27179 + }, + { + "epoch": 0.637142003199775, + "grad_norm": 0.36886659264564514, + "learning_rate": 5.8259411077270245e-05, + "loss": 0.0508, + "step": 27180 + }, + { + "epoch": 0.6371654447745799, + "grad_norm": 0.22299142181873322, + "learning_rate": 5.8252717414418315e-05, + "loss": 0.0578, + "step": 27181 + }, + { + "epoch": 0.637188886349385, + "grad_norm": 0.3143042027950287, + "learning_rate": 5.824602397808832e-05, + "loss": 0.0478, + "step": 27182 + }, + { + "epoch": 0.6372123279241899, + "grad_norm": 0.37452876567840576, + "learning_rate": 5.8239330768316534e-05, + "loss": 0.0453, + "step": 27183 + }, + { + "epoch": 0.637235769498995, + "grad_norm": 0.39048832654953003, + "learning_rate": 5.8232637785139345e-05, + "loss": 0.0998, + "step": 27184 + }, + { + "epoch": 0.6372592110737999, + "grad_norm": 0.45626312494277954, + "learning_rate": 5.8225945028593045e-05, + "loss": 0.1072, + "step": 27185 + }, + { + "epoch": 0.637282652648605, + "grad_norm": 0.11471989005804062, + "learning_rate": 5.821925249871394e-05, + "loss": 0.014, + "step": 27186 + }, + { + "epoch": 0.6373060942234099, + "grad_norm": 0.3200225234031677, + "learning_rate": 5.8212560195538345e-05, + "loss": 0.0519, + "step": 27187 + }, + { + "epoch": 0.6373295357982149, + "grad_norm": 0.25651392340660095, + "learning_rate": 5.820586811910258e-05, + "loss": 0.0583, + "step": 27188 + }, + { + "epoch": 0.6373529773730199, + "grad_norm": 0.3896055221557617, + "learning_rate": 5.819917626944289e-05, + "loss": 0.0638, + "step": 27189 + }, + { + "epoch": 0.6373764189478249, + "grad_norm": 0.5343098044395447, + "learning_rate": 5.819248464659566e-05, + "loss": 0.0993, + "step": 27190 + }, + { + "epoch": 0.6373998605226299, + "grad_norm": 0.40553736686706543, + "learning_rate": 5.8185793250597234e-05, + "loss": 0.0592, + "step": 27191 + }, + { + "epoch": 0.6374233020974349, + "grad_norm": 0.0974951684474945, + "learning_rate": 5.817910208148385e-05, + "loss": 0.0154, + "step": 27192 + }, + { + "epoch": 0.6374467436722399, + "grad_norm": 0.5947311520576477, + "learning_rate": 5.817241113929184e-05, + "loss": 0.1255, + "step": 27193 + }, + { + "epoch": 0.6374701852470449, + "grad_norm": 0.1403120905160904, + "learning_rate": 5.81657204240575e-05, + "loss": 0.0175, + "step": 27194 + }, + { + "epoch": 0.6374936268218498, + "grad_norm": 0.3992743492126465, + "learning_rate": 5.815902993581714e-05, + "loss": 0.0483, + "step": 27195 + }, + { + "epoch": 0.6375170683966549, + "grad_norm": 0.8021008372306824, + "learning_rate": 5.8152339674607026e-05, + "loss": 0.0626, + "step": 27196 + }, + { + "epoch": 0.6375405099714598, + "grad_norm": 0.4903500974178314, + "learning_rate": 5.814564964046354e-05, + "loss": 0.112, + "step": 27197 + }, + { + "epoch": 0.6375639515462649, + "grad_norm": 0.9005741477012634, + "learning_rate": 5.8138959833422925e-05, + "loss": 0.1942, + "step": 27198 + }, + { + "epoch": 0.6375873931210698, + "grad_norm": 0.27793556451797485, + "learning_rate": 5.8132270253521485e-05, + "loss": 0.0661, + "step": 27199 + }, + { + "epoch": 0.6376108346958749, + "grad_norm": 0.6403080224990845, + "learning_rate": 5.812558090079554e-05, + "loss": 0.1272, + "step": 27200 + }, + { + "epoch": 0.6376342762706798, + "grad_norm": 0.5979264378547668, + "learning_rate": 5.8118891775281314e-05, + "loss": 0.0852, + "step": 27201 + }, + { + "epoch": 0.6376577178454849, + "grad_norm": 0.3106088936328888, + "learning_rate": 5.81122028770152e-05, + "loss": 0.043, + "step": 27202 + }, + { + "epoch": 0.6376811594202898, + "grad_norm": 0.3048093020915985, + "learning_rate": 5.810551420603342e-05, + "loss": 0.0562, + "step": 27203 + }, + { + "epoch": 0.6377046009950949, + "grad_norm": 0.12213219702243805, + "learning_rate": 5.809882576237233e-05, + "loss": 0.0327, + "step": 27204 + }, + { + "epoch": 0.6377280425698999, + "grad_norm": 0.2978383004665375, + "learning_rate": 5.8092137546068196e-05, + "loss": 0.0656, + "step": 27205 + }, + { + "epoch": 0.6377514841447048, + "grad_norm": 0.40556496381759644, + "learning_rate": 5.80854495571573e-05, + "loss": 0.0768, + "step": 27206 + }, + { + "epoch": 0.6377749257195099, + "grad_norm": 0.4058302640914917, + "learning_rate": 5.807876179567593e-05, + "loss": 0.065, + "step": 27207 + }, + { + "epoch": 0.6377983672943148, + "grad_norm": 0.44014972448349, + "learning_rate": 5.8072074261660395e-05, + "loss": 0.0684, + "step": 27208 + }, + { + "epoch": 0.6378218088691199, + "grad_norm": 0.670780599117279, + "learning_rate": 5.8065386955146894e-05, + "loss": 0.0413, + "step": 27209 + }, + { + "epoch": 0.6378452504439248, + "grad_norm": 0.3813888728618622, + "learning_rate": 5.8058699876171854e-05, + "loss": 0.0817, + "step": 27210 + }, + { + "epoch": 0.6378686920187299, + "grad_norm": 0.6807593703269958, + "learning_rate": 5.805201302477146e-05, + "loss": 0.1444, + "step": 27211 + }, + { + "epoch": 0.6378921335935348, + "grad_norm": 0.571686863899231, + "learning_rate": 5.804532640098204e-05, + "loss": 0.1373, + "step": 27212 + }, + { + "epoch": 0.6379155751683399, + "grad_norm": 0.6778916716575623, + "learning_rate": 5.8038640004839804e-05, + "loss": 0.053, + "step": 27213 + }, + { + "epoch": 0.6379390167431448, + "grad_norm": 0.5102030038833618, + "learning_rate": 5.8031953836381136e-05, + "loss": 0.1549, + "step": 27214 + }, + { + "epoch": 0.6379624583179498, + "grad_norm": 0.4560664892196655, + "learning_rate": 5.802526789564227e-05, + "loss": 0.6139, + "step": 27215 + }, + { + "epoch": 0.6379858998927548, + "grad_norm": 0.4562012851238251, + "learning_rate": 5.801858218265944e-05, + "loss": 0.0676, + "step": 27216 + }, + { + "epoch": 0.6380093414675598, + "grad_norm": 0.5251767039299011, + "learning_rate": 5.8011896697469e-05, + "loss": 0.1045, + "step": 27217 + }, + { + "epoch": 0.6380327830423648, + "grad_norm": 0.5311900973320007, + "learning_rate": 5.80052114401072e-05, + "loss": 0.0776, + "step": 27218 + }, + { + "epoch": 0.6380562246171698, + "grad_norm": 0.18857911229133606, + "learning_rate": 5.7998526410610275e-05, + "loss": 0.0271, + "step": 27219 + }, + { + "epoch": 0.6380796661919748, + "grad_norm": 0.5016832947731018, + "learning_rate": 5.799184160901455e-05, + "loss": 0.1361, + "step": 27220 + }, + { + "epoch": 0.6381031077667798, + "grad_norm": 0.3554057478904724, + "learning_rate": 5.798515703535623e-05, + "loss": 0.0568, + "step": 27221 + }, + { + "epoch": 0.6381265493415847, + "grad_norm": 0.282500296831131, + "learning_rate": 5.797847268967167e-05, + "loss": 0.0511, + "step": 27222 + }, + { + "epoch": 0.6381499909163898, + "grad_norm": 0.42742377519607544, + "learning_rate": 5.79717885719971e-05, + "loss": 0.0405, + "step": 27223 + }, + { + "epoch": 0.6381734324911947, + "grad_norm": 0.5023899674415588, + "learning_rate": 5.7965104682368776e-05, + "loss": 0.0991, + "step": 27224 + }, + { + "epoch": 0.6381968740659998, + "grad_norm": 0.5196191072463989, + "learning_rate": 5.795842102082295e-05, + "loss": 0.1369, + "step": 27225 + }, + { + "epoch": 0.6382203156408047, + "grad_norm": 0.48237431049346924, + "learning_rate": 5.795173758739595e-05, + "loss": 0.1492, + "step": 27226 + }, + { + "epoch": 0.6382437572156098, + "grad_norm": 0.18170931935310364, + "learning_rate": 5.7945054382123996e-05, + "loss": 0.0351, + "step": 27227 + }, + { + "epoch": 0.6382671987904147, + "grad_norm": 0.37607109546661377, + "learning_rate": 5.7938371405043323e-05, + "loss": 0.079, + "step": 27228 + }, + { + "epoch": 0.6382906403652198, + "grad_norm": 0.476338654756546, + "learning_rate": 5.793168865619028e-05, + "loss": 0.0763, + "step": 27229 + }, + { + "epoch": 0.6383140819400247, + "grad_norm": 0.5022688508033752, + "learning_rate": 5.792500613560106e-05, + "loss": 0.1169, + "step": 27230 + }, + { + "epoch": 0.6383375235148298, + "grad_norm": 0.5026535391807556, + "learning_rate": 5.791832384331195e-05, + "loss": 0.0791, + "step": 27231 + }, + { + "epoch": 0.6383609650896347, + "grad_norm": 0.3074202835559845, + "learning_rate": 5.791164177935919e-05, + "loss": 0.0391, + "step": 27232 + }, + { + "epoch": 0.6383844066644397, + "grad_norm": 0.41667184233665466, + "learning_rate": 5.7904959943779055e-05, + "loss": 0.0851, + "step": 27233 + }, + { + "epoch": 0.6384078482392447, + "grad_norm": 0.6187949180603027, + "learning_rate": 5.789827833660775e-05, + "loss": 0.0455, + "step": 27234 + }, + { + "epoch": 0.6384312898140497, + "grad_norm": 0.2388347089290619, + "learning_rate": 5.789159695788161e-05, + "loss": 0.0231, + "step": 27235 + }, + { + "epoch": 0.6384547313888547, + "grad_norm": 0.36897581815719604, + "learning_rate": 5.788491580763684e-05, + "loss": 0.0611, + "step": 27236 + }, + { + "epoch": 0.6384781729636597, + "grad_norm": 0.1992652714252472, + "learning_rate": 5.787823488590971e-05, + "loss": 0.0396, + "step": 27237 + }, + { + "epoch": 0.6385016145384647, + "grad_norm": 0.5214138627052307, + "learning_rate": 5.787155419273642e-05, + "loss": 0.098, + "step": 27238 + }, + { + "epoch": 0.6385250561132697, + "grad_norm": 0.22316251695156097, + "learning_rate": 5.7864873728153305e-05, + "loss": 0.0408, + "step": 27239 + }, + { + "epoch": 0.6385484976880746, + "grad_norm": 0.33858242630958557, + "learning_rate": 5.785819349219656e-05, + "loss": 0.0949, + "step": 27240 + }, + { + "epoch": 0.6385719392628797, + "grad_norm": 0.26205530762672424, + "learning_rate": 5.785151348490241e-05, + "loss": 0.0538, + "step": 27241 + }, + { + "epoch": 0.6385953808376846, + "grad_norm": 0.5445172190666199, + "learning_rate": 5.784483370630718e-05, + "loss": 0.0789, + "step": 27242 + }, + { + "epoch": 0.6386188224124897, + "grad_norm": 0.2040841281414032, + "learning_rate": 5.783815415644705e-05, + "loss": 0.0386, + "step": 27243 + }, + { + "epoch": 0.6386422639872946, + "grad_norm": 0.5231845378875732, + "learning_rate": 5.783147483535829e-05, + "loss": 0.0945, + "step": 27244 + }, + { + "epoch": 0.6386657055620997, + "grad_norm": 0.6024438142776489, + "learning_rate": 5.782479574307713e-05, + "loss": 0.1218, + "step": 27245 + }, + { + "epoch": 0.6386891471369046, + "grad_norm": 0.6390947103500366, + "learning_rate": 5.7818116879639824e-05, + "loss": 0.1298, + "step": 27246 + }, + { + "epoch": 0.6387125887117097, + "grad_norm": 0.553510308265686, + "learning_rate": 5.781143824508255e-05, + "loss": 0.0908, + "step": 27247 + }, + { + "epoch": 0.6387360302865146, + "grad_norm": 0.49148210883140564, + "learning_rate": 5.780475983944165e-05, + "loss": 0.1136, + "step": 27248 + }, + { + "epoch": 0.6387594718613196, + "grad_norm": 0.1832801252603531, + "learning_rate": 5.77980816627533e-05, + "loss": 0.0283, + "step": 27249 + }, + { + "epoch": 0.6387829134361246, + "grad_norm": 1.0064493417739868, + "learning_rate": 5.779140371505371e-05, + "loss": 0.1339, + "step": 27250 + }, + { + "epoch": 0.6388063550109296, + "grad_norm": 0.39645329117774963, + "learning_rate": 5.7784725996379185e-05, + "loss": 0.0596, + "step": 27251 + }, + { + "epoch": 0.6388297965857346, + "grad_norm": 0.6822864413261414, + "learning_rate": 5.7778048506765934e-05, + "loss": 0.1563, + "step": 27252 + }, + { + "epoch": 0.6388532381605396, + "grad_norm": 0.13947482407093048, + "learning_rate": 5.777137124625016e-05, + "loss": 0.0253, + "step": 27253 + }, + { + "epoch": 0.6388766797353446, + "grad_norm": 0.556485652923584, + "learning_rate": 5.77646942148681e-05, + "loss": 0.6406, + "step": 27254 + }, + { + "epoch": 0.6389001213101496, + "grad_norm": 0.31621232628822327, + "learning_rate": 5.7758017412656005e-05, + "loss": 0.0279, + "step": 27255 + }, + { + "epoch": 0.6389235628849547, + "grad_norm": 0.31416013836860657, + "learning_rate": 5.775134083965013e-05, + "loss": 0.0564, + "step": 27256 + }, + { + "epoch": 0.6389470044597596, + "grad_norm": 0.6505727767944336, + "learning_rate": 5.774466449588665e-05, + "loss": 0.1081, + "step": 27257 + }, + { + "epoch": 0.6389704460345647, + "grad_norm": 0.3564033508300781, + "learning_rate": 5.7737988381401795e-05, + "loss": 0.5129, + "step": 27258 + }, + { + "epoch": 0.6389938876093696, + "grad_norm": 0.2165457159280777, + "learning_rate": 5.773131249623178e-05, + "loss": 0.0547, + "step": 27259 + }, + { + "epoch": 0.6390173291841746, + "grad_norm": 0.7327407598495483, + "learning_rate": 5.7724636840412895e-05, + "loss": 0.1581, + "step": 27260 + }, + { + "epoch": 0.6390407707589796, + "grad_norm": 0.7315300703048706, + "learning_rate": 5.7717961413981314e-05, + "loss": 0.0644, + "step": 27261 + }, + { + "epoch": 0.6390642123337846, + "grad_norm": 0.36944252252578735, + "learning_rate": 5.7711286216973215e-05, + "loss": 0.0478, + "step": 27262 + }, + { + "epoch": 0.6390876539085896, + "grad_norm": 0.4844552278518677, + "learning_rate": 5.770461124942491e-05, + "loss": 0.0725, + "step": 27263 + }, + { + "epoch": 0.6391110954833946, + "grad_norm": 0.4852673411369324, + "learning_rate": 5.769793651137259e-05, + "loss": 0.0968, + "step": 27264 + }, + { + "epoch": 0.6391345370581996, + "grad_norm": 0.5658544301986694, + "learning_rate": 5.769126200285243e-05, + "loss": 0.1215, + "step": 27265 + }, + { + "epoch": 0.6391579786330046, + "grad_norm": 0.31958335638046265, + "learning_rate": 5.768458772390064e-05, + "loss": 0.0436, + "step": 27266 + }, + { + "epoch": 0.6391814202078095, + "grad_norm": 0.39716246724128723, + "learning_rate": 5.767791367455352e-05, + "loss": 0.07, + "step": 27267 + }, + { + "epoch": 0.6392048617826146, + "grad_norm": 0.359679639339447, + "learning_rate": 5.767123985484722e-05, + "loss": 0.021, + "step": 27268 + }, + { + "epoch": 0.6392283033574195, + "grad_norm": 0.4047558903694153, + "learning_rate": 5.766456626481796e-05, + "loss": 0.0822, + "step": 27269 + }, + { + "epoch": 0.6392517449322246, + "grad_norm": 0.21630534529685974, + "learning_rate": 5.7657892904501965e-05, + "loss": 0.0281, + "step": 27270 + }, + { + "epoch": 0.6392751865070295, + "grad_norm": 0.2985183894634247, + "learning_rate": 5.765121977393541e-05, + "loss": 0.0386, + "step": 27271 + }, + { + "epoch": 0.6392986280818346, + "grad_norm": 0.39578261971473694, + "learning_rate": 5.764454687315453e-05, + "loss": 0.4712, + "step": 27272 + }, + { + "epoch": 0.6393220696566395, + "grad_norm": 0.3253885805606842, + "learning_rate": 5.763787420219553e-05, + "loss": 0.0519, + "step": 27273 + }, + { + "epoch": 0.6393455112314446, + "grad_norm": 0.35422730445861816, + "learning_rate": 5.763120176109459e-05, + "loss": 0.0737, + "step": 27274 + }, + { + "epoch": 0.6393689528062495, + "grad_norm": 0.4412185847759247, + "learning_rate": 5.762452954988797e-05, + "loss": 0.0898, + "step": 27275 + }, + { + "epoch": 0.6393923943810546, + "grad_norm": 0.24900995194911957, + "learning_rate": 5.761785756861187e-05, + "loss": 0.0325, + "step": 27276 + }, + { + "epoch": 0.6394158359558595, + "grad_norm": 0.5046765208244324, + "learning_rate": 5.761118581730245e-05, + "loss": 0.0739, + "step": 27277 + }, + { + "epoch": 0.6394392775306645, + "grad_norm": 0.2583377957344055, + "learning_rate": 5.7604514295995926e-05, + "loss": 0.0131, + "step": 27278 + }, + { + "epoch": 0.6394627191054695, + "grad_norm": 0.2522551715373993, + "learning_rate": 5.759784300472846e-05, + "loss": 0.0379, + "step": 27279 + }, + { + "epoch": 0.6394861606802745, + "grad_norm": 0.29562908411026, + "learning_rate": 5.759117194353634e-05, + "loss": 0.0378, + "step": 27280 + }, + { + "epoch": 0.6395096022550795, + "grad_norm": 0.46964728832244873, + "learning_rate": 5.75845011124557e-05, + "loss": 0.0798, + "step": 27281 + }, + { + "epoch": 0.6395330438298845, + "grad_norm": 0.5213167667388916, + "learning_rate": 5.757783051152275e-05, + "loss": 0.0552, + "step": 27282 + }, + { + "epoch": 0.6395564854046895, + "grad_norm": 0.4303351044654846, + "learning_rate": 5.7571160140773683e-05, + "loss": 0.0846, + "step": 27283 + }, + { + "epoch": 0.6395799269794945, + "grad_norm": 0.3236737549304962, + "learning_rate": 5.756449000024471e-05, + "loss": 0.0264, + "step": 27284 + }, + { + "epoch": 0.6396033685542994, + "grad_norm": 0.5086342096328735, + "learning_rate": 5.7557820089971956e-05, + "loss": 0.0381, + "step": 27285 + }, + { + "epoch": 0.6396268101291045, + "grad_norm": 0.587239146232605, + "learning_rate": 5.7551150409991663e-05, + "loss": 0.0624, + "step": 27286 + }, + { + "epoch": 0.6396502517039094, + "grad_norm": 0.47282618284225464, + "learning_rate": 5.7544480960340067e-05, + "loss": 0.1184, + "step": 27287 + }, + { + "epoch": 0.6396736932787145, + "grad_norm": 0.12131664901971817, + "learning_rate": 5.75378117410533e-05, + "loss": 0.0133, + "step": 27288 + }, + { + "epoch": 0.6396971348535194, + "grad_norm": 0.3679350018501282, + "learning_rate": 5.753114275216758e-05, + "loss": 0.0398, + "step": 27289 + }, + { + "epoch": 0.6397205764283245, + "grad_norm": 0.7334392666816711, + "learning_rate": 5.752447399371905e-05, + "loss": 0.1047, + "step": 27290 + }, + { + "epoch": 0.6397440180031294, + "grad_norm": 0.4120940566062927, + "learning_rate": 5.7517805465743946e-05, + "loss": 0.0486, + "step": 27291 + }, + { + "epoch": 0.6397674595779345, + "grad_norm": 0.5505932569503784, + "learning_rate": 5.751113716827836e-05, + "loss": 0.1177, + "step": 27292 + }, + { + "epoch": 0.6397909011527394, + "grad_norm": 0.3916744589805603, + "learning_rate": 5.7504469101358585e-05, + "loss": 0.0613, + "step": 27293 + }, + { + "epoch": 0.6398143427275444, + "grad_norm": 0.39758867025375366, + "learning_rate": 5.749780126502076e-05, + "loss": 0.0857, + "step": 27294 + }, + { + "epoch": 0.6398377843023494, + "grad_norm": 0.4118970036506653, + "learning_rate": 5.749113365930106e-05, + "loss": 0.0919, + "step": 27295 + }, + { + "epoch": 0.6398612258771544, + "grad_norm": 0.46491536498069763, + "learning_rate": 5.748446628423566e-05, + "loss": 0.0605, + "step": 27296 + }, + { + "epoch": 0.6398846674519594, + "grad_norm": 0.32953572273254395, + "learning_rate": 5.74777991398607e-05, + "loss": 0.0525, + "step": 27297 + }, + { + "epoch": 0.6399081090267644, + "grad_norm": 0.18040437996387482, + "learning_rate": 5.747113222621244e-05, + "loss": 0.0314, + "step": 27298 + }, + { + "epoch": 0.6399315506015694, + "grad_norm": 0.12386038154363632, + "learning_rate": 5.7464465543326976e-05, + "loss": 0.0129, + "step": 27299 + }, + { + "epoch": 0.6399549921763744, + "grad_norm": 0.5731300115585327, + "learning_rate": 5.745779909124056e-05, + "loss": 0.619, + "step": 27300 + }, + { + "epoch": 0.6399784337511794, + "grad_norm": 0.13364988565444946, + "learning_rate": 5.745113286998931e-05, + "loss": 0.0276, + "step": 27301 + }, + { + "epoch": 0.6400018753259844, + "grad_norm": 0.49607929587364197, + "learning_rate": 5.744446687960942e-05, + "loss": 0.0443, + "step": 27302 + }, + { + "epoch": 0.6400253169007893, + "grad_norm": 0.6264023780822754, + "learning_rate": 5.743780112013704e-05, + "loss": 0.1482, + "step": 27303 + }, + { + "epoch": 0.6400487584755944, + "grad_norm": 0.39881908893585205, + "learning_rate": 5.743113559160831e-05, + "loss": 0.0894, + "step": 27304 + }, + { + "epoch": 0.6400722000503993, + "grad_norm": 0.5774019956588745, + "learning_rate": 5.7424470294059485e-05, + "loss": 0.125, + "step": 27305 + }, + { + "epoch": 0.6400956416252044, + "grad_norm": 0.1482519805431366, + "learning_rate": 5.7417805227526675e-05, + "loss": 0.0264, + "step": 27306 + }, + { + "epoch": 0.6401190832000094, + "grad_norm": 0.42170587182044983, + "learning_rate": 5.741114039204605e-05, + "loss": 0.0879, + "step": 27307 + }, + { + "epoch": 0.6401425247748144, + "grad_norm": 0.4526015520095825, + "learning_rate": 5.740447578765378e-05, + "loss": 0.101, + "step": 27308 + }, + { + "epoch": 0.6401659663496194, + "grad_norm": 0.3442271053791046, + "learning_rate": 5.739781141438602e-05, + "loss": 0.0536, + "step": 27309 + }, + { + "epoch": 0.6401894079244244, + "grad_norm": 0.3253409266471863, + "learning_rate": 5.739114727227889e-05, + "loss": 0.0454, + "step": 27310 + }, + { + "epoch": 0.6402128494992294, + "grad_norm": 0.20062296092510223, + "learning_rate": 5.738448336136859e-05, + "loss": 0.0216, + "step": 27311 + }, + { + "epoch": 0.6402362910740343, + "grad_norm": 0.6268566250801086, + "learning_rate": 5.737781968169133e-05, + "loss": 0.0821, + "step": 27312 + }, + { + "epoch": 0.6402597326488394, + "grad_norm": 0.4693603217601776, + "learning_rate": 5.737115623328322e-05, + "loss": 0.5873, + "step": 27313 + }, + { + "epoch": 0.6402831742236443, + "grad_norm": 0.24582147598266602, + "learning_rate": 5.7364493016180406e-05, + "loss": 0.0256, + "step": 27314 + }, + { + "epoch": 0.6403066157984494, + "grad_norm": 0.1892797350883484, + "learning_rate": 5.735783003041906e-05, + "loss": 0.0388, + "step": 27315 + }, + { + "epoch": 0.6403300573732543, + "grad_norm": 0.23074786365032196, + "learning_rate": 5.735116727603531e-05, + "loss": 0.0455, + "step": 27316 + }, + { + "epoch": 0.6403534989480594, + "grad_norm": 0.5311577320098877, + "learning_rate": 5.73445047530653e-05, + "loss": 0.0713, + "step": 27317 + }, + { + "epoch": 0.6403769405228643, + "grad_norm": 0.4619847238063812, + "learning_rate": 5.7337842461545244e-05, + "loss": 0.0715, + "step": 27318 + }, + { + "epoch": 0.6404003820976694, + "grad_norm": 0.4816126823425293, + "learning_rate": 5.7331180401511265e-05, + "loss": 0.0632, + "step": 27319 + }, + { + "epoch": 0.6404238236724743, + "grad_norm": 0.7269614338874817, + "learning_rate": 5.732451857299947e-05, + "loss": 0.1138, + "step": 27320 + }, + { + "epoch": 0.6404472652472794, + "grad_norm": 0.5247974991798401, + "learning_rate": 5.731785697604607e-05, + "loss": 0.1041, + "step": 27321 + }, + { + "epoch": 0.6404707068220843, + "grad_norm": 0.4509384334087372, + "learning_rate": 5.731119561068716e-05, + "loss": 0.0844, + "step": 27322 + }, + { + "epoch": 0.6404941483968893, + "grad_norm": 0.37486732006073, + "learning_rate": 5.730453447695887e-05, + "loss": 0.2718, + "step": 27323 + }, + { + "epoch": 0.6405175899716943, + "grad_norm": 0.3073325455188751, + "learning_rate": 5.7297873574897374e-05, + "loss": 0.0546, + "step": 27324 + }, + { + "epoch": 0.6405410315464993, + "grad_norm": 0.4059765934944153, + "learning_rate": 5.729121290453887e-05, + "loss": 0.1234, + "step": 27325 + }, + { + "epoch": 0.6405644731213043, + "grad_norm": 0.3900620639324188, + "learning_rate": 5.728455246591943e-05, + "loss": 0.0677, + "step": 27326 + }, + { + "epoch": 0.6405879146961093, + "grad_norm": 0.15905475616455078, + "learning_rate": 5.7277892259075206e-05, + "loss": 0.031, + "step": 27327 + }, + { + "epoch": 0.6406113562709143, + "grad_norm": 0.12501421570777893, + "learning_rate": 5.727123228404235e-05, + "loss": 0.0153, + "step": 27328 + }, + { + "epoch": 0.6406347978457193, + "grad_norm": 0.3596917390823364, + "learning_rate": 5.726457254085699e-05, + "loss": 0.0837, + "step": 27329 + }, + { + "epoch": 0.6406582394205242, + "grad_norm": 0.507879376411438, + "learning_rate": 5.725791302955521e-05, + "loss": 0.1628, + "step": 27330 + }, + { + "epoch": 0.6406816809953293, + "grad_norm": 0.24446648359298706, + "learning_rate": 5.725125375017325e-05, + "loss": 0.294, + "step": 27331 + }, + { + "epoch": 0.6407051225701342, + "grad_norm": 0.32205167412757874, + "learning_rate": 5.724459470274718e-05, + "loss": 0.0425, + "step": 27332 + }, + { + "epoch": 0.6407285641449393, + "grad_norm": 0.11354387551546097, + "learning_rate": 5.7237935887313144e-05, + "loss": 0.0225, + "step": 27333 + }, + { + "epoch": 0.6407520057197442, + "grad_norm": 0.5670445561408997, + "learning_rate": 5.723127730390727e-05, + "loss": 0.141, + "step": 27334 + }, + { + "epoch": 0.6407754472945493, + "grad_norm": 0.5036522746086121, + "learning_rate": 5.7224618952565646e-05, + "loss": 0.0658, + "step": 27335 + }, + { + "epoch": 0.6407988888693542, + "grad_norm": 0.5420361161231995, + "learning_rate": 5.721796083332445e-05, + "loss": 0.523, + "step": 27336 + }, + { + "epoch": 0.6408223304441593, + "grad_norm": 0.3313114643096924, + "learning_rate": 5.721130294621984e-05, + "loss": 0.0388, + "step": 27337 + }, + { + "epoch": 0.6408457720189642, + "grad_norm": 0.09626567363739014, + "learning_rate": 5.720464529128788e-05, + "loss": 0.0158, + "step": 27338 + }, + { + "epoch": 0.6408692135937692, + "grad_norm": 0.19498415291309357, + "learning_rate": 5.719798786856474e-05, + "loss": 0.0348, + "step": 27339 + }, + { + "epoch": 0.6408926551685742, + "grad_norm": 0.3713233172893524, + "learning_rate": 5.719133067808652e-05, + "loss": 0.0703, + "step": 27340 + }, + { + "epoch": 0.6409160967433792, + "grad_norm": 0.4809226095676422, + "learning_rate": 5.718467371988934e-05, + "loss": 0.0564, + "step": 27341 + }, + { + "epoch": 0.6409395383181842, + "grad_norm": 0.14249108731746674, + "learning_rate": 5.717801699400929e-05, + "loss": 0.0296, + "step": 27342 + }, + { + "epoch": 0.6409629798929892, + "grad_norm": 0.183047354221344, + "learning_rate": 5.717136050048255e-05, + "loss": 0.0331, + "step": 27343 + }, + { + "epoch": 0.6409864214677942, + "grad_norm": 1.3313974142074585, + "learning_rate": 5.7164704239345235e-05, + "loss": 0.269, + "step": 27344 + }, + { + "epoch": 0.6410098630425992, + "grad_norm": 0.48485201597213745, + "learning_rate": 5.715804821063342e-05, + "loss": 0.1059, + "step": 27345 + }, + { + "epoch": 0.6410333046174042, + "grad_norm": 0.30681484937667847, + "learning_rate": 5.7151392414383244e-05, + "loss": 0.0304, + "step": 27346 + }, + { + "epoch": 0.6410567461922092, + "grad_norm": 0.6976965665817261, + "learning_rate": 5.714473685063078e-05, + "loss": 0.1781, + "step": 27347 + }, + { + "epoch": 0.6410801877670141, + "grad_norm": 0.3918188810348511, + "learning_rate": 5.713808151941222e-05, + "loss": 0.1114, + "step": 27348 + }, + { + "epoch": 0.6411036293418192, + "grad_norm": 0.23738045990467072, + "learning_rate": 5.713142642076359e-05, + "loss": 0.045, + "step": 27349 + }, + { + "epoch": 0.6411270709166241, + "grad_norm": 0.07484840601682663, + "learning_rate": 5.712477155472109e-05, + "loss": 0.0123, + "step": 27350 + }, + { + "epoch": 0.6411505124914292, + "grad_norm": 0.10914284735918045, + "learning_rate": 5.711811692132079e-05, + "loss": 0.0247, + "step": 27351 + }, + { + "epoch": 0.6411739540662341, + "grad_norm": 0.1419045329093933, + "learning_rate": 5.711146252059878e-05, + "loss": 0.033, + "step": 27352 + }, + { + "epoch": 0.6411973956410392, + "grad_norm": 0.6860780119895935, + "learning_rate": 5.710480835259119e-05, + "loss": 0.1456, + "step": 27353 + }, + { + "epoch": 0.6412208372158441, + "grad_norm": 0.39175236225128174, + "learning_rate": 5.7098154417334105e-05, + "loss": 0.0744, + "step": 27354 + }, + { + "epoch": 0.6412442787906492, + "grad_norm": 0.36496222019195557, + "learning_rate": 5.709150071486361e-05, + "loss": 0.0493, + "step": 27355 + }, + { + "epoch": 0.6412677203654541, + "grad_norm": 0.6724371910095215, + "learning_rate": 5.708484724521588e-05, + "loss": 0.1048, + "step": 27356 + }, + { + "epoch": 0.6412911619402591, + "grad_norm": 0.4180815815925598, + "learning_rate": 5.707819400842698e-05, + "loss": 0.0703, + "step": 27357 + }, + { + "epoch": 0.6413146035150642, + "grad_norm": 0.32469823956489563, + "learning_rate": 5.707154100453299e-05, + "loss": 0.0711, + "step": 27358 + }, + { + "epoch": 0.6413380450898691, + "grad_norm": 0.5301365256309509, + "learning_rate": 5.706488823356999e-05, + "loss": 0.1211, + "step": 27359 + }, + { + "epoch": 0.6413614866646742, + "grad_norm": 0.38625073432922363, + "learning_rate": 5.705823569557416e-05, + "loss": 0.0536, + "step": 27360 + }, + { + "epoch": 0.6413849282394791, + "grad_norm": 0.573398768901825, + "learning_rate": 5.705158339058155e-05, + "loss": 0.1345, + "step": 27361 + }, + { + "epoch": 0.6414083698142842, + "grad_norm": 0.4512713551521301, + "learning_rate": 5.704493131862821e-05, + "loss": 0.0731, + "step": 27362 + }, + { + "epoch": 0.6414318113890891, + "grad_norm": 0.2383349984884262, + "learning_rate": 5.7038279479750335e-05, + "loss": 0.0585, + "step": 27363 + }, + { + "epoch": 0.6414552529638942, + "grad_norm": 0.5795689821243286, + "learning_rate": 5.703162787398395e-05, + "loss": 0.0722, + "step": 27364 + }, + { + "epoch": 0.6414786945386991, + "grad_norm": 0.48975735902786255, + "learning_rate": 5.7024976501365156e-05, + "loss": 0.1293, + "step": 27365 + }, + { + "epoch": 0.6415021361135042, + "grad_norm": 0.3670252561569214, + "learning_rate": 5.701832536193006e-05, + "loss": 0.0982, + "step": 27366 + }, + { + "epoch": 0.6415255776883091, + "grad_norm": 0.43306779861450195, + "learning_rate": 5.7011674455714734e-05, + "loss": 0.6564, + "step": 27367 + }, + { + "epoch": 0.6415490192631141, + "grad_norm": 0.4132343828678131, + "learning_rate": 5.700502378275523e-05, + "loss": 0.1314, + "step": 27368 + }, + { + "epoch": 0.6415724608379191, + "grad_norm": 0.4780244529247284, + "learning_rate": 5.699837334308773e-05, + "loss": 0.0676, + "step": 27369 + }, + { + "epoch": 0.6415959024127241, + "grad_norm": 0.31846362352371216, + "learning_rate": 5.699172313674825e-05, + "loss": 0.0518, + "step": 27370 + }, + { + "epoch": 0.6416193439875291, + "grad_norm": 0.15957707166671753, + "learning_rate": 5.698507316377285e-05, + "loss": 0.0287, + "step": 27371 + }, + { + "epoch": 0.6416427855623341, + "grad_norm": 0.5994426608085632, + "learning_rate": 5.6978423424197704e-05, + "loss": 0.1094, + "step": 27372 + }, + { + "epoch": 0.6416662271371391, + "grad_norm": 0.3246506154537201, + "learning_rate": 5.697177391805882e-05, + "loss": 0.0456, + "step": 27373 + }, + { + "epoch": 0.6416896687119441, + "grad_norm": 0.22894617915153503, + "learning_rate": 5.6965124645392275e-05, + "loss": 0.0433, + "step": 27374 + }, + { + "epoch": 0.641713110286749, + "grad_norm": 0.49850693345069885, + "learning_rate": 5.695847560623422e-05, + "loss": 0.0972, + "step": 27375 + }, + { + "epoch": 0.6417365518615541, + "grad_norm": 0.4342837333679199, + "learning_rate": 5.6951826800620675e-05, + "loss": 0.0603, + "step": 27376 + }, + { + "epoch": 0.641759993436359, + "grad_norm": 0.27000677585601807, + "learning_rate": 5.6945178228587736e-05, + "loss": 0.0364, + "step": 27377 + }, + { + "epoch": 0.6417834350111641, + "grad_norm": 0.49003422260284424, + "learning_rate": 5.6938529890171464e-05, + "loss": 0.1288, + "step": 27378 + }, + { + "epoch": 0.641806876585969, + "grad_norm": 0.34100398421287537, + "learning_rate": 5.693188178540794e-05, + "loss": 0.0806, + "step": 27379 + }, + { + "epoch": 0.6418303181607741, + "grad_norm": 0.6010288000106812, + "learning_rate": 5.6925233914333197e-05, + "loss": 0.1093, + "step": 27380 + }, + { + "epoch": 0.641853759735579, + "grad_norm": 0.3263186514377594, + "learning_rate": 5.691858627698339e-05, + "loss": 0.0572, + "step": 27381 + }, + { + "epoch": 0.6418772013103841, + "grad_norm": 0.1638457030057907, + "learning_rate": 5.691193887339454e-05, + "loss": 0.0185, + "step": 27382 + }, + { + "epoch": 0.641900642885189, + "grad_norm": 0.051996566355228424, + "learning_rate": 5.690529170360269e-05, + "loss": 0.0034, + "step": 27383 + }, + { + "epoch": 0.641924084459994, + "grad_norm": 0.11111189424991608, + "learning_rate": 5.689864476764397e-05, + "loss": 0.0154, + "step": 27384 + }, + { + "epoch": 0.641947526034799, + "grad_norm": 0.5387281179428101, + "learning_rate": 5.689199806555442e-05, + "loss": 0.0947, + "step": 27385 + }, + { + "epoch": 0.641970967609604, + "grad_norm": 0.47190243005752563, + "learning_rate": 5.6885351597370096e-05, + "loss": 0.1016, + "step": 27386 + }, + { + "epoch": 0.641994409184409, + "grad_norm": 0.40346574783325195, + "learning_rate": 5.6878705363127024e-05, + "loss": 0.0645, + "step": 27387 + }, + { + "epoch": 0.642017850759214, + "grad_norm": 0.27464038133621216, + "learning_rate": 5.6872059362861374e-05, + "loss": 0.0339, + "step": 27388 + }, + { + "epoch": 0.642041292334019, + "grad_norm": 0.08807676285505295, + "learning_rate": 5.686541359660911e-05, + "loss": 0.0108, + "step": 27389 + }, + { + "epoch": 0.642064733908824, + "grad_norm": 0.5148056149482727, + "learning_rate": 5.685876806440634e-05, + "loss": 0.1156, + "step": 27390 + }, + { + "epoch": 0.642088175483629, + "grad_norm": 0.17138178646564484, + "learning_rate": 5.685212276628911e-05, + "loss": 0.03, + "step": 27391 + }, + { + "epoch": 0.642111617058434, + "grad_norm": 0.45010092854499817, + "learning_rate": 5.684547770229347e-05, + "loss": 0.062, + "step": 27392 + }, + { + "epoch": 0.6421350586332389, + "grad_norm": 0.20382548868656158, + "learning_rate": 5.683883287245544e-05, + "loss": 0.0346, + "step": 27393 + }, + { + "epoch": 0.642158500208044, + "grad_norm": 0.4069600999355316, + "learning_rate": 5.683218827681116e-05, + "loss": 0.0747, + "step": 27394 + }, + { + "epoch": 0.6421819417828489, + "grad_norm": 0.3708013594150543, + "learning_rate": 5.682554391539664e-05, + "loss": 0.0741, + "step": 27395 + }, + { + "epoch": 0.642205383357654, + "grad_norm": 0.5505343675613403, + "learning_rate": 5.6818899788247895e-05, + "loss": 0.0675, + "step": 27396 + }, + { + "epoch": 0.6422288249324589, + "grad_norm": 0.5399264693260193, + "learning_rate": 5.6812255895401044e-05, + "loss": 0.0813, + "step": 27397 + }, + { + "epoch": 0.642252266507264, + "grad_norm": 0.5416821837425232, + "learning_rate": 5.680561223689211e-05, + "loss": 0.0966, + "step": 27398 + }, + { + "epoch": 0.6422757080820689, + "grad_norm": 0.6406545639038086, + "learning_rate": 5.6798968812757144e-05, + "loss": 0.149, + "step": 27399 + }, + { + "epoch": 0.642299149656874, + "grad_norm": 0.6079881191253662, + "learning_rate": 5.679232562303215e-05, + "loss": 0.1948, + "step": 27400 + }, + { + "epoch": 0.6423225912316789, + "grad_norm": 0.2713991105556488, + "learning_rate": 5.6785682667753245e-05, + "loss": 0.0291, + "step": 27401 + }, + { + "epoch": 0.642346032806484, + "grad_norm": 0.41046062111854553, + "learning_rate": 5.677903994695645e-05, + "loss": 0.055, + "step": 27402 + }, + { + "epoch": 0.6423694743812889, + "grad_norm": 0.3249278664588928, + "learning_rate": 5.67723974606778e-05, + "loss": 0.0894, + "step": 27403 + }, + { + "epoch": 0.6423929159560939, + "grad_norm": 0.37083858251571655, + "learning_rate": 5.676575520895332e-05, + "loss": 0.0756, + "step": 27404 + }, + { + "epoch": 0.6424163575308989, + "grad_norm": 0.6496335864067078, + "learning_rate": 5.675911319181903e-05, + "loss": 0.1071, + "step": 27405 + }, + { + "epoch": 0.6424397991057039, + "grad_norm": 0.21958310902118683, + "learning_rate": 5.675247140931106e-05, + "loss": 0.0479, + "step": 27406 + }, + { + "epoch": 0.6424632406805089, + "grad_norm": 0.2341429740190506, + "learning_rate": 5.674582986146539e-05, + "loss": 0.0196, + "step": 27407 + }, + { + "epoch": 0.6424866822553139, + "grad_norm": 0.2765853703022003, + "learning_rate": 5.673918854831803e-05, + "loss": 0.0264, + "step": 27408 + }, + { + "epoch": 0.642510123830119, + "grad_norm": 0.6338503956794739, + "learning_rate": 5.673254746990508e-05, + "loss": 0.0671, + "step": 27409 + }, + { + "epoch": 0.6425335654049239, + "grad_norm": 0.49631884694099426, + "learning_rate": 5.6725906626262546e-05, + "loss": 0.0811, + "step": 27410 + }, + { + "epoch": 0.642557006979729, + "grad_norm": 0.07546015083789825, + "learning_rate": 5.671926601742647e-05, + "loss": 0.0114, + "step": 27411 + }, + { + "epoch": 0.6425804485545339, + "grad_norm": 0.5851303935050964, + "learning_rate": 5.671262564343283e-05, + "loss": 0.1138, + "step": 27412 + }, + { + "epoch": 0.6426038901293389, + "grad_norm": 0.48811256885528564, + "learning_rate": 5.6705985504317726e-05, + "loss": 0.1092, + "step": 27413 + }, + { + "epoch": 0.6426273317041439, + "grad_norm": 0.3256678283214569, + "learning_rate": 5.669934560011717e-05, + "loss": 0.085, + "step": 27414 + }, + { + "epoch": 0.6426507732789489, + "grad_norm": 0.1545756608247757, + "learning_rate": 5.669270593086719e-05, + "loss": 0.0291, + "step": 27415 + }, + { + "epoch": 0.6426742148537539, + "grad_norm": 0.3973415791988373, + "learning_rate": 5.6686066496603795e-05, + "loss": 0.0567, + "step": 27416 + }, + { + "epoch": 0.6426976564285589, + "grad_norm": 0.0416884571313858, + "learning_rate": 5.667942729736302e-05, + "loss": 0.0052, + "step": 27417 + }, + { + "epoch": 0.6427210980033639, + "grad_norm": 0.40352335572242737, + "learning_rate": 5.667278833318086e-05, + "loss": 0.0604, + "step": 27418 + }, + { + "epoch": 0.6427445395781689, + "grad_norm": 0.4317137897014618, + "learning_rate": 5.66661496040934e-05, + "loss": 0.0616, + "step": 27419 + }, + { + "epoch": 0.6427679811529738, + "grad_norm": 0.10196401178836823, + "learning_rate": 5.6659511110136596e-05, + "loss": 0.0223, + "step": 27420 + }, + { + "epoch": 0.6427914227277789, + "grad_norm": 0.3277888894081116, + "learning_rate": 5.665287285134655e-05, + "loss": 0.0592, + "step": 27421 + }, + { + "epoch": 0.6428148643025838, + "grad_norm": 0.07358302175998688, + "learning_rate": 5.664623482775923e-05, + "loss": 0.0125, + "step": 27422 + }, + { + "epoch": 0.6428383058773889, + "grad_norm": 0.15107735991477966, + "learning_rate": 5.663959703941065e-05, + "loss": 0.0178, + "step": 27423 + }, + { + "epoch": 0.6428617474521938, + "grad_norm": 0.5856153964996338, + "learning_rate": 5.663295948633685e-05, + "loss": 0.6095, + "step": 27424 + }, + { + "epoch": 0.6428851890269989, + "grad_norm": 0.18822723627090454, + "learning_rate": 5.662632216857378e-05, + "loss": 0.0357, + "step": 27425 + }, + { + "epoch": 0.6429086306018038, + "grad_norm": 0.1269839107990265, + "learning_rate": 5.661968508615754e-05, + "loss": 0.0271, + "step": 27426 + }, + { + "epoch": 0.6429320721766089, + "grad_norm": 0.4229382574558258, + "learning_rate": 5.661304823912412e-05, + "loss": 0.0712, + "step": 27427 + }, + { + "epoch": 0.6429555137514138, + "grad_norm": 0.4760405719280243, + "learning_rate": 5.660641162750952e-05, + "loss": 0.0704, + "step": 27428 + }, + { + "epoch": 0.6429789553262188, + "grad_norm": 0.3132598102092743, + "learning_rate": 5.659977525134974e-05, + "loss": 0.0672, + "step": 27429 + }, + { + "epoch": 0.6430023969010238, + "grad_norm": 0.4115520715713501, + "learning_rate": 5.659313911068082e-05, + "loss": 0.0729, + "step": 27430 + }, + { + "epoch": 0.6430258384758288, + "grad_norm": 0.05742393061518669, + "learning_rate": 5.658650320553869e-05, + "loss": 0.0074, + "step": 27431 + }, + { + "epoch": 0.6430492800506338, + "grad_norm": 0.6019082069396973, + "learning_rate": 5.657986753595942e-05, + "loss": 0.7552, + "step": 27432 + }, + { + "epoch": 0.6430727216254388, + "grad_norm": 0.6900784373283386, + "learning_rate": 5.657323210197904e-05, + "loss": 0.0804, + "step": 27433 + }, + { + "epoch": 0.6430961632002438, + "grad_norm": 0.5262513160705566, + "learning_rate": 5.656659690363354e-05, + "loss": 0.1232, + "step": 27434 + }, + { + "epoch": 0.6431196047750488, + "grad_norm": 0.23502470552921295, + "learning_rate": 5.655996194095889e-05, + "loss": 0.0253, + "step": 27435 + }, + { + "epoch": 0.6431430463498538, + "grad_norm": 0.9619262218475342, + "learning_rate": 5.6553327213991124e-05, + "loss": 0.1158, + "step": 27436 + }, + { + "epoch": 0.6431664879246588, + "grad_norm": 0.2439403235912323, + "learning_rate": 5.6546692722766224e-05, + "loss": 0.0375, + "step": 27437 + }, + { + "epoch": 0.6431899294994637, + "grad_norm": 0.05610530078411102, + "learning_rate": 5.654005846732015e-05, + "loss": 0.0077, + "step": 27438 + }, + { + "epoch": 0.6432133710742688, + "grad_norm": 0.43036743998527527, + "learning_rate": 5.6533424447688967e-05, + "loss": 0.084, + "step": 27439 + }, + { + "epoch": 0.6432368126490737, + "grad_norm": 0.34807950258255005, + "learning_rate": 5.652679066390866e-05, + "loss": 0.0942, + "step": 27440 + }, + { + "epoch": 0.6432602542238788, + "grad_norm": 0.4547298848628998, + "learning_rate": 5.6520157116015206e-05, + "loss": 0.0712, + "step": 27441 + }, + { + "epoch": 0.6432836957986837, + "grad_norm": 0.2877745032310486, + "learning_rate": 5.65135238040446e-05, + "loss": 0.0437, + "step": 27442 + }, + { + "epoch": 0.6433071373734888, + "grad_norm": 0.32363736629486084, + "learning_rate": 5.650689072803279e-05, + "loss": 0.0664, + "step": 27443 + }, + { + "epoch": 0.6433305789482937, + "grad_norm": 0.44405078887939453, + "learning_rate": 5.650025788801587e-05, + "loss": 0.0563, + "step": 27444 + }, + { + "epoch": 0.6433540205230988, + "grad_norm": 0.1679149717092514, + "learning_rate": 5.649362528402973e-05, + "loss": 0.0298, + "step": 27445 + }, + { + "epoch": 0.6433774620979037, + "grad_norm": 0.6309844255447388, + "learning_rate": 5.648699291611042e-05, + "loss": 0.098, + "step": 27446 + }, + { + "epoch": 0.6434009036727087, + "grad_norm": 0.2678754925727844, + "learning_rate": 5.648036078429393e-05, + "loss": 0.0191, + "step": 27447 + }, + { + "epoch": 0.6434243452475137, + "grad_norm": 0.5528305768966675, + "learning_rate": 5.647372888861623e-05, + "loss": 0.7612, + "step": 27448 + }, + { + "epoch": 0.6434477868223187, + "grad_norm": 0.3247356414794922, + "learning_rate": 5.64670972291133e-05, + "loss": 0.0624, + "step": 27449 + }, + { + "epoch": 0.6434712283971237, + "grad_norm": 0.8558300733566284, + "learning_rate": 5.646046580582107e-05, + "loss": 0.1132, + "step": 27450 + }, + { + "epoch": 0.6434946699719287, + "grad_norm": 0.1963406652212143, + "learning_rate": 5.6453834618775624e-05, + "loss": 0.0429, + "step": 27451 + }, + { + "epoch": 0.6435181115467337, + "grad_norm": 0.4652354121208191, + "learning_rate": 5.6447203668012905e-05, + "loss": 0.1008, + "step": 27452 + }, + { + "epoch": 0.6435415531215387, + "grad_norm": 0.2508060038089752, + "learning_rate": 5.644057295356887e-05, + "loss": 0.0415, + "step": 27453 + }, + { + "epoch": 0.6435649946963437, + "grad_norm": 0.42672765254974365, + "learning_rate": 5.6433942475479516e-05, + "loss": 0.726, + "step": 27454 + }, + { + "epoch": 0.6435884362711487, + "grad_norm": 0.315228670835495, + "learning_rate": 5.6427312233780816e-05, + "loss": 0.312, + "step": 27455 + }, + { + "epoch": 0.6436118778459536, + "grad_norm": 0.5172624588012695, + "learning_rate": 5.6420682228508704e-05, + "loss": 0.1572, + "step": 27456 + }, + { + "epoch": 0.6436353194207587, + "grad_norm": 0.35618171095848083, + "learning_rate": 5.6414052459699195e-05, + "loss": 0.0773, + "step": 27457 + }, + { + "epoch": 0.6436587609955636, + "grad_norm": 0.1962878257036209, + "learning_rate": 5.64074229273883e-05, + "loss": 0.0175, + "step": 27458 + }, + { + "epoch": 0.6436822025703687, + "grad_norm": 0.5302063822746277, + "learning_rate": 5.640079363161197e-05, + "loss": 0.0715, + "step": 27459 + }, + { + "epoch": 0.6437056441451737, + "grad_norm": 0.28941646218299866, + "learning_rate": 5.639416457240614e-05, + "loss": 0.0492, + "step": 27460 + }, + { + "epoch": 0.6437290857199787, + "grad_norm": 0.40165597200393677, + "learning_rate": 5.638753574980681e-05, + "loss": 0.0802, + "step": 27461 + }, + { + "epoch": 0.6437525272947837, + "grad_norm": 0.3585779666900635, + "learning_rate": 5.638090716384994e-05, + "loss": 0.0462, + "step": 27462 + }, + { + "epoch": 0.6437759688695887, + "grad_norm": 0.313135027885437, + "learning_rate": 5.6374278814571446e-05, + "loss": 0.4832, + "step": 27463 + }, + { + "epoch": 0.6437994104443937, + "grad_norm": 0.3776996433734894, + "learning_rate": 5.636765070200738e-05, + "loss": 0.0495, + "step": 27464 + }, + { + "epoch": 0.6438228520191986, + "grad_norm": 0.11983965337276459, + "learning_rate": 5.6361022826193665e-05, + "loss": 0.023, + "step": 27465 + }, + { + "epoch": 0.6438462935940037, + "grad_norm": 0.41966819763183594, + "learning_rate": 5.635439518716628e-05, + "loss": 0.0572, + "step": 27466 + }, + { + "epoch": 0.6438697351688086, + "grad_norm": 0.1677660048007965, + "learning_rate": 5.634776778496117e-05, + "loss": 0.0305, + "step": 27467 + }, + { + "epoch": 0.6438931767436137, + "grad_norm": 0.2742869257926941, + "learning_rate": 5.6341140619614295e-05, + "loss": 0.0554, + "step": 27468 + }, + { + "epoch": 0.6439166183184186, + "grad_norm": 0.5210973620414734, + "learning_rate": 5.633451369116158e-05, + "loss": 0.6243, + "step": 27469 + }, + { + "epoch": 0.6439400598932237, + "grad_norm": 0.1339719146490097, + "learning_rate": 5.6327886999639026e-05, + "loss": 0.0181, + "step": 27470 + }, + { + "epoch": 0.6439635014680286, + "grad_norm": 0.30360686779022217, + "learning_rate": 5.632126054508262e-05, + "loss": 0.0391, + "step": 27471 + }, + { + "epoch": 0.6439869430428337, + "grad_norm": 0.4490404427051544, + "learning_rate": 5.631463432752828e-05, + "loss": 0.0507, + "step": 27472 + }, + { + "epoch": 0.6440103846176386, + "grad_norm": 0.3713727593421936, + "learning_rate": 5.6308008347011976e-05, + "loss": 0.0502, + "step": 27473 + }, + { + "epoch": 0.6440338261924436, + "grad_norm": 0.6103556156158447, + "learning_rate": 5.630138260356963e-05, + "loss": 0.5739, + "step": 27474 + }, + { + "epoch": 0.6440572677672486, + "grad_norm": 0.4551681876182556, + "learning_rate": 5.629475709723722e-05, + "loss": 0.0769, + "step": 27475 + }, + { + "epoch": 0.6440807093420536, + "grad_norm": 0.07004958391189575, + "learning_rate": 5.628813182805064e-05, + "loss": 0.0102, + "step": 27476 + }, + { + "epoch": 0.6441041509168586, + "grad_norm": 0.13198387622833252, + "learning_rate": 5.628150679604594e-05, + "loss": 0.0233, + "step": 27477 + }, + { + "epoch": 0.6441275924916636, + "grad_norm": 0.30498582124710083, + "learning_rate": 5.627488200125901e-05, + "loss": 0.0316, + "step": 27478 + }, + { + "epoch": 0.6441510340664686, + "grad_norm": 0.13891662657260895, + "learning_rate": 5.626825744372579e-05, + "loss": 0.0131, + "step": 27479 + }, + { + "epoch": 0.6441744756412736, + "grad_norm": 0.5194779634475708, + "learning_rate": 5.626163312348224e-05, + "loss": 0.0964, + "step": 27480 + }, + { + "epoch": 0.6441979172160786, + "grad_norm": 0.6038376688957214, + "learning_rate": 5.625500904056425e-05, + "loss": 0.1341, + "step": 27481 + }, + { + "epoch": 0.6442213587908836, + "grad_norm": 0.2233111411333084, + "learning_rate": 5.6248385195007856e-05, + "loss": 0.0171, + "step": 27482 + }, + { + "epoch": 0.6442448003656885, + "grad_norm": 0.6035875678062439, + "learning_rate": 5.624176158684893e-05, + "loss": 0.1021, + "step": 27483 + }, + { + "epoch": 0.6442682419404936, + "grad_norm": 0.4712222218513489, + "learning_rate": 5.623513821612345e-05, + "loss": 0.1211, + "step": 27484 + }, + { + "epoch": 0.6442916835152985, + "grad_norm": 0.35527652502059937, + "learning_rate": 5.622851508286735e-05, + "loss": 0.0407, + "step": 27485 + }, + { + "epoch": 0.6443151250901036, + "grad_norm": 0.41742727160453796, + "learning_rate": 5.6221892187116574e-05, + "loss": 0.0413, + "step": 27486 + }, + { + "epoch": 0.6443385666649085, + "grad_norm": 0.49448445439338684, + "learning_rate": 5.621526952890702e-05, + "loss": 0.0769, + "step": 27487 + }, + { + "epoch": 0.6443620082397136, + "grad_norm": 0.3329450786113739, + "learning_rate": 5.620864710827463e-05, + "loss": 0.0268, + "step": 27488 + }, + { + "epoch": 0.6443854498145185, + "grad_norm": 0.23400846123695374, + "learning_rate": 5.620202492525537e-05, + "loss": 0.043, + "step": 27489 + }, + { + "epoch": 0.6444088913893236, + "grad_norm": 0.2959321439266205, + "learning_rate": 5.6195402979885145e-05, + "loss": 0.066, + "step": 27490 + }, + { + "epoch": 0.6444323329641285, + "grad_norm": 0.41532135009765625, + "learning_rate": 5.618878127219992e-05, + "loss": 0.0657, + "step": 27491 + }, + { + "epoch": 0.6444557745389335, + "grad_norm": 0.4163137376308441, + "learning_rate": 5.618215980223559e-05, + "loss": 0.0648, + "step": 27492 + }, + { + "epoch": 0.6444792161137385, + "grad_norm": 0.5704114437103271, + "learning_rate": 5.6175538570028056e-05, + "loss": 0.0883, + "step": 27493 + }, + { + "epoch": 0.6445026576885435, + "grad_norm": 0.31348657608032227, + "learning_rate": 5.616891757561331e-05, + "loss": 0.0535, + "step": 27494 + }, + { + "epoch": 0.6445260992633485, + "grad_norm": 0.3234443664550781, + "learning_rate": 5.6162296819027224e-05, + "loss": 0.0485, + "step": 27495 + }, + { + "epoch": 0.6445495408381535, + "grad_norm": 0.4450068473815918, + "learning_rate": 5.615567630030578e-05, + "loss": 0.0953, + "step": 27496 + }, + { + "epoch": 0.6445729824129585, + "grad_norm": 0.22445757687091827, + "learning_rate": 5.614905601948488e-05, + "loss": 0.0518, + "step": 27497 + }, + { + "epoch": 0.6445964239877635, + "grad_norm": 0.15296359360218048, + "learning_rate": 5.6142435976600415e-05, + "loss": 0.0273, + "step": 27498 + }, + { + "epoch": 0.6446198655625685, + "grad_norm": 0.28083521127700806, + "learning_rate": 5.613581617168835e-05, + "loss": 0.0285, + "step": 27499 + }, + { + "epoch": 0.6446433071373735, + "grad_norm": 0.479789137840271, + "learning_rate": 5.6129196604784554e-05, + "loss": 0.0647, + "step": 27500 + }, + { + "epoch": 0.6446667487121784, + "grad_norm": 0.5331899523735046, + "learning_rate": 5.612257727592495e-05, + "loss": 0.0619, + "step": 27501 + }, + { + "epoch": 0.6446901902869835, + "grad_norm": 0.3098272979259491, + "learning_rate": 5.61159581851455e-05, + "loss": 0.0229, + "step": 27502 + }, + { + "epoch": 0.6447136318617884, + "grad_norm": 0.5352256298065186, + "learning_rate": 5.6109339332482104e-05, + "loss": 0.0864, + "step": 27503 + }, + { + "epoch": 0.6447370734365935, + "grad_norm": 0.22545410692691803, + "learning_rate": 5.610272071797067e-05, + "loss": 0.0364, + "step": 27504 + }, + { + "epoch": 0.6447605150113984, + "grad_norm": 0.6172199249267578, + "learning_rate": 5.609610234164706e-05, + "loss": 0.1824, + "step": 27505 + }, + { + "epoch": 0.6447839565862035, + "grad_norm": 0.20285046100616455, + "learning_rate": 5.608948420354729e-05, + "loss": 0.0396, + "step": 27506 + }, + { + "epoch": 0.6448073981610084, + "grad_norm": 0.4156607687473297, + "learning_rate": 5.6082866303707206e-05, + "loss": 0.0669, + "step": 27507 + }, + { + "epoch": 0.6448308397358135, + "grad_norm": 0.4627551734447479, + "learning_rate": 5.6076248642162675e-05, + "loss": 0.053, + "step": 27508 + }, + { + "epoch": 0.6448542813106184, + "grad_norm": 0.29292088747024536, + "learning_rate": 5.606963121894969e-05, + "loss": 0.034, + "step": 27509 + }, + { + "epoch": 0.6448777228854234, + "grad_norm": 0.1382034420967102, + "learning_rate": 5.6063014034104135e-05, + "loss": 0.012, + "step": 27510 + }, + { + "epoch": 0.6449011644602285, + "grad_norm": 0.2557089924812317, + "learning_rate": 5.605639708766189e-05, + "loss": 0.0392, + "step": 27511 + }, + { + "epoch": 0.6449246060350334, + "grad_norm": 0.1525169461965561, + "learning_rate": 5.604978037965888e-05, + "loss": 0.0305, + "step": 27512 + }, + { + "epoch": 0.6449480476098385, + "grad_norm": 0.7905251383781433, + "learning_rate": 5.604316391013099e-05, + "loss": 0.3782, + "step": 27513 + }, + { + "epoch": 0.6449714891846434, + "grad_norm": 0.41441065073013306, + "learning_rate": 5.603654767911409e-05, + "loss": 0.1013, + "step": 27514 + }, + { + "epoch": 0.6449949307594485, + "grad_norm": 0.18534071743488312, + "learning_rate": 5.6029931686644164e-05, + "loss": 0.0369, + "step": 27515 + }, + { + "epoch": 0.6450183723342534, + "grad_norm": 0.6165931224822998, + "learning_rate": 5.6023315932757056e-05, + "loss": 0.1521, + "step": 27516 + }, + { + "epoch": 0.6450418139090585, + "grad_norm": 0.6066078543663025, + "learning_rate": 5.6016700417488643e-05, + "loss": 0.1045, + "step": 27517 + }, + { + "epoch": 0.6450652554838634, + "grad_norm": 0.3986835181713104, + "learning_rate": 5.601008514087489e-05, + "loss": 0.1348, + "step": 27518 + }, + { + "epoch": 0.6450886970586684, + "grad_norm": 0.4147108793258667, + "learning_rate": 5.600347010295164e-05, + "loss": 0.0947, + "step": 27519 + }, + { + "epoch": 0.6451121386334734, + "grad_norm": 0.5993915796279907, + "learning_rate": 5.59968553037548e-05, + "loss": 0.1073, + "step": 27520 + }, + { + "epoch": 0.6451355802082784, + "grad_norm": 0.6195726990699768, + "learning_rate": 5.5990240743320235e-05, + "loss": 0.0759, + "step": 27521 + }, + { + "epoch": 0.6451590217830834, + "grad_norm": 0.23896458745002747, + "learning_rate": 5.598362642168389e-05, + "loss": 0.0556, + "step": 27522 + }, + { + "epoch": 0.6451824633578884, + "grad_norm": 0.3824180066585541, + "learning_rate": 5.5977012338881615e-05, + "loss": 0.0772, + "step": 27523 + }, + { + "epoch": 0.6452059049326934, + "grad_norm": 0.30336859822273254, + "learning_rate": 5.597039849494933e-05, + "loss": 0.0568, + "step": 27524 + }, + { + "epoch": 0.6452293465074984, + "grad_norm": 0.3887976109981537, + "learning_rate": 5.5963784889922885e-05, + "loss": 0.0604, + "step": 27525 + }, + { + "epoch": 0.6452527880823034, + "grad_norm": 0.2072279006242752, + "learning_rate": 5.595717152383815e-05, + "loss": 0.0387, + "step": 27526 + }, + { + "epoch": 0.6452762296571084, + "grad_norm": 0.5377342104911804, + "learning_rate": 5.595055839673109e-05, + "loss": 0.0984, + "step": 27527 + }, + { + "epoch": 0.6452996712319133, + "grad_norm": 0.10104741156101227, + "learning_rate": 5.594394550863752e-05, + "loss": 0.0101, + "step": 27528 + }, + { + "epoch": 0.6453231128067184, + "grad_norm": 0.2849179208278656, + "learning_rate": 5.5937332859593306e-05, + "loss": 0.0206, + "step": 27529 + }, + { + "epoch": 0.6453465543815233, + "grad_norm": 0.759828507900238, + "learning_rate": 5.5930720449634414e-05, + "loss": 0.1644, + "step": 27530 + }, + { + "epoch": 0.6453699959563284, + "grad_norm": 0.39317235350608826, + "learning_rate": 5.5924108278796663e-05, + "loss": 0.0731, + "step": 27531 + }, + { + "epoch": 0.6453934375311333, + "grad_norm": 0.2886253893375397, + "learning_rate": 5.591749634711594e-05, + "loss": 0.0503, + "step": 27532 + }, + { + "epoch": 0.6454168791059384, + "grad_norm": 0.408220499753952, + "learning_rate": 5.591088465462807e-05, + "loss": 0.0674, + "step": 27533 + }, + { + "epoch": 0.6454403206807433, + "grad_norm": 0.5245789289474487, + "learning_rate": 5.590427320136904e-05, + "loss": 0.1008, + "step": 27534 + }, + { + "epoch": 0.6454637622555484, + "grad_norm": 0.4705588221549988, + "learning_rate": 5.589766198737465e-05, + "loss": 0.0984, + "step": 27535 + }, + { + "epoch": 0.6454872038303533, + "grad_norm": 0.292557030916214, + "learning_rate": 5.58910510126808e-05, + "loss": 0.0552, + "step": 27536 + }, + { + "epoch": 0.6455106454051583, + "grad_norm": 0.11635120958089828, + "learning_rate": 5.588444027732332e-05, + "loss": 0.0217, + "step": 27537 + }, + { + "epoch": 0.6455340869799633, + "grad_norm": 0.149903804063797, + "learning_rate": 5.587782978133813e-05, + "loss": 0.0239, + "step": 27538 + }, + { + "epoch": 0.6455575285547683, + "grad_norm": 0.18418681621551514, + "learning_rate": 5.5871219524761034e-05, + "loss": 0.0326, + "step": 27539 + }, + { + "epoch": 0.6455809701295733, + "grad_norm": 0.17678527534008026, + "learning_rate": 5.5864609507627975e-05, + "loss": 0.0365, + "step": 27540 + }, + { + "epoch": 0.6456044117043783, + "grad_norm": 0.6644841432571411, + "learning_rate": 5.585799972997474e-05, + "loss": 0.0772, + "step": 27541 + }, + { + "epoch": 0.6456278532791833, + "grad_norm": 0.5412521958351135, + "learning_rate": 5.5851390191837294e-05, + "loss": 0.1119, + "step": 27542 + }, + { + "epoch": 0.6456512948539883, + "grad_norm": 0.6087445616722107, + "learning_rate": 5.5844780893251445e-05, + "loss": 0.0727, + "step": 27543 + }, + { + "epoch": 0.6456747364287933, + "grad_norm": 0.39037206768989563, + "learning_rate": 5.583817183425306e-05, + "loss": 0.049, + "step": 27544 + }, + { + "epoch": 0.6456981780035983, + "grad_norm": 0.5207681059837341, + "learning_rate": 5.583156301487797e-05, + "loss": 0.5305, + "step": 27545 + }, + { + "epoch": 0.6457216195784032, + "grad_norm": 0.415737122297287, + "learning_rate": 5.582495443516205e-05, + "loss": 0.0712, + "step": 27546 + }, + { + "epoch": 0.6457450611532083, + "grad_norm": 0.1568770706653595, + "learning_rate": 5.581834609514121e-05, + "loss": 0.0242, + "step": 27547 + }, + { + "epoch": 0.6457685027280132, + "grad_norm": 0.5724811553955078, + "learning_rate": 5.581173799485124e-05, + "loss": 0.0533, + "step": 27548 + }, + { + "epoch": 0.6457919443028183, + "grad_norm": 0.5716189742088318, + "learning_rate": 5.580513013432804e-05, + "loss": 0.0653, + "step": 27549 + }, + { + "epoch": 0.6458153858776232, + "grad_norm": 0.08468977361917496, + "learning_rate": 5.5798522513607445e-05, + "loss": 0.0183, + "step": 27550 + }, + { + "epoch": 0.6458388274524283, + "grad_norm": 0.09597382694482803, + "learning_rate": 5.579191513272532e-05, + "loss": 0.0157, + "step": 27551 + }, + { + "epoch": 0.6458622690272332, + "grad_norm": 0.8536069989204407, + "learning_rate": 5.578530799171745e-05, + "loss": 0.2435, + "step": 27552 + }, + { + "epoch": 0.6458857106020383, + "grad_norm": 0.6287335753440857, + "learning_rate": 5.577870109061978e-05, + "loss": 0.1215, + "step": 27553 + }, + { + "epoch": 0.6459091521768432, + "grad_norm": 0.2072010636329651, + "learning_rate": 5.577209442946809e-05, + "loss": 0.0417, + "step": 27554 + }, + { + "epoch": 0.6459325937516482, + "grad_norm": 0.444853812456131, + "learning_rate": 5.576548800829831e-05, + "loss": 0.1121, + "step": 27555 + }, + { + "epoch": 0.6459560353264532, + "grad_norm": 0.06261401623487473, + "learning_rate": 5.575888182714621e-05, + "loss": 0.0062, + "step": 27556 + }, + { + "epoch": 0.6459794769012582, + "grad_norm": 0.5777972340583801, + "learning_rate": 5.575227588604768e-05, + "loss": 0.6439, + "step": 27557 + }, + { + "epoch": 0.6460029184760632, + "grad_norm": 0.4758620262145996, + "learning_rate": 5.574567018503854e-05, + "loss": 0.0941, + "step": 27558 + }, + { + "epoch": 0.6460263600508682, + "grad_norm": 0.5659769177436829, + "learning_rate": 5.57390647241546e-05, + "loss": 0.0783, + "step": 27559 + }, + { + "epoch": 0.6460498016256732, + "grad_norm": 0.5043002963066101, + "learning_rate": 5.573245950343178e-05, + "loss": 0.0836, + "step": 27560 + }, + { + "epoch": 0.6460732432004782, + "grad_norm": 0.17973124980926514, + "learning_rate": 5.572585452290587e-05, + "loss": 0.0197, + "step": 27561 + }, + { + "epoch": 0.6460966847752833, + "grad_norm": 0.37463369965553284, + "learning_rate": 5.571924978261274e-05, + "loss": 0.4719, + "step": 27562 + }, + { + "epoch": 0.6461201263500882, + "grad_norm": 0.3680790066719055, + "learning_rate": 5.5712645282588184e-05, + "loss": 0.0703, + "step": 27563 + }, + { + "epoch": 0.6461435679248932, + "grad_norm": 0.16915515065193176, + "learning_rate": 5.570604102286803e-05, + "loss": 0.0388, + "step": 27564 + }, + { + "epoch": 0.6461670094996982, + "grad_norm": 0.3855760395526886, + "learning_rate": 5.5699437003488175e-05, + "loss": 0.0578, + "step": 27565 + }, + { + "epoch": 0.6461904510745032, + "grad_norm": 0.568419873714447, + "learning_rate": 5.569283322448438e-05, + "loss": 0.508, + "step": 27566 + }, + { + "epoch": 0.6462138926493082, + "grad_norm": 0.22531567513942719, + "learning_rate": 5.568622968589257e-05, + "loss": 0.0233, + "step": 27567 + }, + { + "epoch": 0.6462373342241132, + "grad_norm": 0.15226627886295319, + "learning_rate": 5.567962638774851e-05, + "loss": 0.0219, + "step": 27568 + }, + { + "epoch": 0.6462607757989182, + "grad_norm": 0.2304355800151825, + "learning_rate": 5.567302333008806e-05, + "loss": 0.0445, + "step": 27569 + }, + { + "epoch": 0.6462842173737232, + "grad_norm": 0.33749857544898987, + "learning_rate": 5.566642051294701e-05, + "loss": 0.0593, + "step": 27570 + }, + { + "epoch": 0.6463076589485282, + "grad_norm": 0.5942836403846741, + "learning_rate": 5.565981793636119e-05, + "loss": 0.1202, + "step": 27571 + }, + { + "epoch": 0.6463311005233332, + "grad_norm": 0.28227174282073975, + "learning_rate": 5.565321560036647e-05, + "loss": 0.0526, + "step": 27572 + }, + { + "epoch": 0.6463545420981381, + "grad_norm": 0.1341894119977951, + "learning_rate": 5.564661350499864e-05, + "loss": 0.0233, + "step": 27573 + }, + { + "epoch": 0.6463779836729432, + "grad_norm": 0.32221314311027527, + "learning_rate": 5.564001165029355e-05, + "loss": 0.0764, + "step": 27574 + }, + { + "epoch": 0.6464014252477481, + "grad_norm": 0.23904967308044434, + "learning_rate": 5.5633410036287016e-05, + "loss": 0.0382, + "step": 27575 + }, + { + "epoch": 0.6464248668225532, + "grad_norm": 0.6321885585784912, + "learning_rate": 5.562680866301482e-05, + "loss": 0.116, + "step": 27576 + }, + { + "epoch": 0.6464483083973581, + "grad_norm": 0.281642347574234, + "learning_rate": 5.562020753051278e-05, + "loss": 0.3262, + "step": 27577 + }, + { + "epoch": 0.6464717499721632, + "grad_norm": 0.18098674714565277, + "learning_rate": 5.561360663881674e-05, + "loss": 0.0319, + "step": 27578 + }, + { + "epoch": 0.6464951915469681, + "grad_norm": 0.3063374161720276, + "learning_rate": 5.560700598796256e-05, + "loss": 0.0556, + "step": 27579 + }, + { + "epoch": 0.6465186331217732, + "grad_norm": 0.8821703195571899, + "learning_rate": 5.560040557798603e-05, + "loss": 0.174, + "step": 27580 + }, + { + "epoch": 0.6465420746965781, + "grad_norm": 0.4468432664871216, + "learning_rate": 5.559380540892292e-05, + "loss": 0.1133, + "step": 27581 + }, + { + "epoch": 0.6465655162713831, + "grad_norm": 0.3981032073497772, + "learning_rate": 5.558720548080908e-05, + "loss": 0.0522, + "step": 27582 + }, + { + "epoch": 0.6465889578461881, + "grad_norm": 0.3532926142215729, + "learning_rate": 5.5580605793680315e-05, + "loss": 0.0391, + "step": 27583 + }, + { + "epoch": 0.6466123994209931, + "grad_norm": 0.18723630905151367, + "learning_rate": 5.557400634757239e-05, + "loss": 0.0296, + "step": 27584 + }, + { + "epoch": 0.6466358409957981, + "grad_norm": 0.6106835007667542, + "learning_rate": 5.55674071425212e-05, + "loss": 0.1167, + "step": 27585 + }, + { + "epoch": 0.6466592825706031, + "grad_norm": 0.5036324858665466, + "learning_rate": 5.5560808178562494e-05, + "loss": 0.0766, + "step": 27586 + }, + { + "epoch": 0.6466827241454081, + "grad_norm": 1.2775193452835083, + "learning_rate": 5.555420945573211e-05, + "loss": 0.1334, + "step": 27587 + }, + { + "epoch": 0.6467061657202131, + "grad_norm": 0.1521187275648117, + "learning_rate": 5.554761097406582e-05, + "loss": 0.0219, + "step": 27588 + }, + { + "epoch": 0.646729607295018, + "grad_norm": 0.539705216884613, + "learning_rate": 5.554101273359945e-05, + "loss": 0.5806, + "step": 27589 + }, + { + "epoch": 0.6467530488698231, + "grad_norm": 0.12657833099365234, + "learning_rate": 5.5534414734368756e-05, + "loss": 0.0179, + "step": 27590 + }, + { + "epoch": 0.646776490444628, + "grad_norm": 0.3145582377910614, + "learning_rate": 5.5527816976409584e-05, + "loss": 0.0447, + "step": 27591 + }, + { + "epoch": 0.6467999320194331, + "grad_norm": 0.23410557210445404, + "learning_rate": 5.552121945975776e-05, + "loss": 0.0804, + "step": 27592 + }, + { + "epoch": 0.646823373594238, + "grad_norm": 0.337912917137146, + "learning_rate": 5.551462218444905e-05, + "loss": 0.0259, + "step": 27593 + }, + { + "epoch": 0.6468468151690431, + "grad_norm": 0.7030194401741028, + "learning_rate": 5.550802515051925e-05, + "loss": 0.103, + "step": 27594 + }, + { + "epoch": 0.646870256743848, + "grad_norm": 0.4178074300289154, + "learning_rate": 5.550142835800416e-05, + "loss": 0.0988, + "step": 27595 + }, + { + "epoch": 0.6468936983186531, + "grad_norm": 0.36183589696884155, + "learning_rate": 5.5494831806939565e-05, + "loss": 0.0749, + "step": 27596 + }, + { + "epoch": 0.646917139893458, + "grad_norm": 0.33274412155151367, + "learning_rate": 5.548823549736122e-05, + "loss": 0.0403, + "step": 27597 + }, + { + "epoch": 0.646940581468263, + "grad_norm": 0.649251401424408, + "learning_rate": 5.548163942930501e-05, + "loss": 0.175, + "step": 27598 + }, + { + "epoch": 0.646964023043068, + "grad_norm": 0.4672039747238159, + "learning_rate": 5.5475043602806675e-05, + "loss": 0.0764, + "step": 27599 + }, + { + "epoch": 0.646987464617873, + "grad_norm": 0.5686002969741821, + "learning_rate": 5.546844801790201e-05, + "loss": 0.0813, + "step": 27600 + }, + { + "epoch": 0.647010906192678, + "grad_norm": 0.5352548956871033, + "learning_rate": 5.546185267462679e-05, + "loss": 0.0707, + "step": 27601 + }, + { + "epoch": 0.647034347767483, + "grad_norm": 0.9630313515663147, + "learning_rate": 5.545525757301677e-05, + "loss": 0.0818, + "step": 27602 + }, + { + "epoch": 0.647057789342288, + "grad_norm": 0.3660872280597687, + "learning_rate": 5.544866271310782e-05, + "loss": 0.0434, + "step": 27603 + }, + { + "epoch": 0.647081230917093, + "grad_norm": 0.21537898480892181, + "learning_rate": 5.544206809493564e-05, + "loss": 0.0492, + "step": 27604 + }, + { + "epoch": 0.647104672491898, + "grad_norm": 0.23533040285110474, + "learning_rate": 5.543547371853608e-05, + "loss": 0.0335, + "step": 27605 + }, + { + "epoch": 0.647128114066703, + "grad_norm": 0.24195915460586548, + "learning_rate": 5.54288795839449e-05, + "loss": 0.0449, + "step": 27606 + }, + { + "epoch": 0.647151555641508, + "grad_norm": 0.2443806231021881, + "learning_rate": 5.542228569119787e-05, + "loss": 0.064, + "step": 27607 + }, + { + "epoch": 0.647174997216313, + "grad_norm": 0.17290371656417847, + "learning_rate": 5.541569204033079e-05, + "loss": 0.0344, + "step": 27608 + }, + { + "epoch": 0.6471984387911179, + "grad_norm": 0.2642686367034912, + "learning_rate": 5.5409098631379354e-05, + "loss": 0.0481, + "step": 27609 + }, + { + "epoch": 0.647221880365923, + "grad_norm": 0.05360671505331993, + "learning_rate": 5.540250546437946e-05, + "loss": 0.0062, + "step": 27610 + }, + { + "epoch": 0.6472453219407279, + "grad_norm": 0.4585127830505371, + "learning_rate": 5.539591253936681e-05, + "loss": 0.0706, + "step": 27611 + }, + { + "epoch": 0.647268763515533, + "grad_norm": 0.25596001744270325, + "learning_rate": 5.53893198563772e-05, + "loss": 0.0459, + "step": 27612 + }, + { + "epoch": 0.647292205090338, + "grad_norm": 0.5201974511146545, + "learning_rate": 5.538272741544639e-05, + "loss": 0.0841, + "step": 27613 + }, + { + "epoch": 0.647315646665143, + "grad_norm": 0.33198487758636475, + "learning_rate": 5.5376135216610125e-05, + "loss": 0.072, + "step": 27614 + }, + { + "epoch": 0.647339088239948, + "grad_norm": 0.12828819453716278, + "learning_rate": 5.5369543259904243e-05, + "loss": 0.0387, + "step": 27615 + }, + { + "epoch": 0.647362529814753, + "grad_norm": 0.18393276631832123, + "learning_rate": 5.536295154536443e-05, + "loss": 0.0313, + "step": 27616 + }, + { + "epoch": 0.647385971389558, + "grad_norm": 0.4702511131763458, + "learning_rate": 5.5356360073026534e-05, + "loss": 0.0521, + "step": 27617 + }, + { + "epoch": 0.6474094129643629, + "grad_norm": 0.17760658264160156, + "learning_rate": 5.534976884292628e-05, + "loss": 0.0398, + "step": 27618 + }, + { + "epoch": 0.647432854539168, + "grad_norm": 0.44296011328697205, + "learning_rate": 5.534317785509946e-05, + "loss": 0.0946, + "step": 27619 + }, + { + "epoch": 0.6474562961139729, + "grad_norm": 0.7147495746612549, + "learning_rate": 5.533658710958178e-05, + "loss": 0.7655, + "step": 27620 + }, + { + "epoch": 0.647479737688778, + "grad_norm": 0.728201150894165, + "learning_rate": 5.532999660640905e-05, + "loss": 0.0921, + "step": 27621 + }, + { + "epoch": 0.6475031792635829, + "grad_norm": 0.456119179725647, + "learning_rate": 5.532340634561699e-05, + "loss": 0.0801, + "step": 27622 + }, + { + "epoch": 0.647526620838388, + "grad_norm": 0.1354345828294754, + "learning_rate": 5.53168163272414e-05, + "loss": 0.0248, + "step": 27623 + }, + { + "epoch": 0.6475500624131929, + "grad_norm": 0.19267575442790985, + "learning_rate": 5.5310226551318033e-05, + "loss": 0.0253, + "step": 27624 + }, + { + "epoch": 0.647573503987998, + "grad_norm": 0.5591666102409363, + "learning_rate": 5.5303637017882625e-05, + "loss": 0.0647, + "step": 27625 + }, + { + "epoch": 0.6475969455628029, + "grad_norm": 0.19795924425125122, + "learning_rate": 5.529704772697095e-05, + "loss": 0.105, + "step": 27626 + }, + { + "epoch": 0.647620387137608, + "grad_norm": 0.7235272526741028, + "learning_rate": 5.529045867861871e-05, + "loss": 0.1308, + "step": 27627 + }, + { + "epoch": 0.6476438287124129, + "grad_norm": 0.6491577625274658, + "learning_rate": 5.5283869872861716e-05, + "loss": 0.4572, + "step": 27628 + }, + { + "epoch": 0.6476672702872179, + "grad_norm": 0.45539382100105286, + "learning_rate": 5.5277281309735685e-05, + "loss": 0.3879, + "step": 27629 + }, + { + "epoch": 0.6476907118620229, + "grad_norm": 0.49375656247138977, + "learning_rate": 5.527069298927642e-05, + "loss": 0.0597, + "step": 27630 + }, + { + "epoch": 0.6477141534368279, + "grad_norm": 0.16228041052818298, + "learning_rate": 5.5264104911519635e-05, + "loss": 0.0254, + "step": 27631 + }, + { + "epoch": 0.6477375950116329, + "grad_norm": 0.38387179374694824, + "learning_rate": 5.5257517076501065e-05, + "loss": 0.0472, + "step": 27632 + }, + { + "epoch": 0.6477610365864379, + "grad_norm": 0.6109052896499634, + "learning_rate": 5.5250929484256463e-05, + "loss": 0.087, + "step": 27633 + }, + { + "epoch": 0.6477844781612428, + "grad_norm": 0.5302616357803345, + "learning_rate": 5.524434213482158e-05, + "loss": 0.0761, + "step": 27634 + }, + { + "epoch": 0.6478079197360479, + "grad_norm": 0.27829769253730774, + "learning_rate": 5.523775502823212e-05, + "loss": 0.0282, + "step": 27635 + }, + { + "epoch": 0.6478313613108528, + "grad_norm": 0.4288775622844696, + "learning_rate": 5.523116816452389e-05, + "loss": 0.1109, + "step": 27636 + }, + { + "epoch": 0.6478548028856579, + "grad_norm": 0.5089483261108398, + "learning_rate": 5.5224581543732615e-05, + "loss": 0.1165, + "step": 27637 + }, + { + "epoch": 0.6478782444604628, + "grad_norm": 0.15218767523765564, + "learning_rate": 5.521799516589401e-05, + "loss": 0.0235, + "step": 27638 + }, + { + "epoch": 0.6479016860352679, + "grad_norm": 0.6352003812789917, + "learning_rate": 5.521140903104378e-05, + "loss": 0.0863, + "step": 27639 + }, + { + "epoch": 0.6479251276100728, + "grad_norm": 0.3308069407939911, + "learning_rate": 5.520482313921775e-05, + "loss": 0.0981, + "step": 27640 + }, + { + "epoch": 0.6479485691848779, + "grad_norm": 0.07112134248018265, + "learning_rate": 5.519823749045156e-05, + "loss": 0.009, + "step": 27641 + }, + { + "epoch": 0.6479720107596828, + "grad_norm": 0.6434394717216492, + "learning_rate": 5.519165208478104e-05, + "loss": 0.0891, + "step": 27642 + }, + { + "epoch": 0.6479954523344879, + "grad_norm": 0.29544344544410706, + "learning_rate": 5.518506692224188e-05, + "loss": 0.0273, + "step": 27643 + }, + { + "epoch": 0.6480188939092928, + "grad_norm": 0.4067216217517853, + "learning_rate": 5.51784820028698e-05, + "loss": 0.0626, + "step": 27644 + }, + { + "epoch": 0.6480423354840978, + "grad_norm": 0.3659784495830536, + "learning_rate": 5.517189732670054e-05, + "loss": 0.0868, + "step": 27645 + }, + { + "epoch": 0.6480657770589028, + "grad_norm": 0.27469223737716675, + "learning_rate": 5.5165312893769824e-05, + "loss": 0.0642, + "step": 27646 + }, + { + "epoch": 0.6480892186337078, + "grad_norm": 0.27606698870658875, + "learning_rate": 5.5158728704113336e-05, + "loss": 0.0469, + "step": 27647 + }, + { + "epoch": 0.6481126602085128, + "grad_norm": 0.47706133127212524, + "learning_rate": 5.51521447577669e-05, + "loss": 0.0763, + "step": 27648 + }, + { + "epoch": 0.6481361017833178, + "grad_norm": 0.21276848018169403, + "learning_rate": 5.514556105476618e-05, + "loss": 0.0319, + "step": 27649 + }, + { + "epoch": 0.6481595433581228, + "grad_norm": 0.8337386250495911, + "learning_rate": 5.513897759514691e-05, + "loss": 0.1612, + "step": 27650 + }, + { + "epoch": 0.6481829849329278, + "grad_norm": 0.5479716062545776, + "learning_rate": 5.513239437894477e-05, + "loss": 0.1048, + "step": 27651 + }, + { + "epoch": 0.6482064265077327, + "grad_norm": 0.42559948563575745, + "learning_rate": 5.512581140619555e-05, + "loss": 0.6555, + "step": 27652 + }, + { + "epoch": 0.6482298680825378, + "grad_norm": 0.3751361668109894, + "learning_rate": 5.511922867693495e-05, + "loss": 0.0367, + "step": 27653 + }, + { + "epoch": 0.6482533096573427, + "grad_norm": 0.703685998916626, + "learning_rate": 5.5112646191198624e-05, + "loss": 0.1423, + "step": 27654 + }, + { + "epoch": 0.6482767512321478, + "grad_norm": 0.32388895750045776, + "learning_rate": 5.51060639490224e-05, + "loss": 0.0542, + "step": 27655 + }, + { + "epoch": 0.6483001928069527, + "grad_norm": 0.33982187509536743, + "learning_rate": 5.5099481950441926e-05, + "loss": 0.0306, + "step": 27656 + }, + { + "epoch": 0.6483236343817578, + "grad_norm": 0.376297265291214, + "learning_rate": 5.5092900195492916e-05, + "loss": 0.0488, + "step": 27657 + }, + { + "epoch": 0.6483470759565627, + "grad_norm": 0.1014944389462471, + "learning_rate": 5.50863186842111e-05, + "loss": 0.0271, + "step": 27658 + }, + { + "epoch": 0.6483705175313678, + "grad_norm": 0.20326313376426697, + "learning_rate": 5.507973741663219e-05, + "loss": 0.0212, + "step": 27659 + }, + { + "epoch": 0.6483939591061727, + "grad_norm": 0.10772424936294556, + "learning_rate": 5.507315639279184e-05, + "loss": 0.019, + "step": 27660 + }, + { + "epoch": 0.6484174006809778, + "grad_norm": 0.4323863983154297, + "learning_rate": 5.5066575612725844e-05, + "loss": 0.0564, + "step": 27661 + }, + { + "epoch": 0.6484408422557827, + "grad_norm": 0.17618946731090546, + "learning_rate": 5.5059995076469875e-05, + "loss": 0.024, + "step": 27662 + }, + { + "epoch": 0.6484642838305877, + "grad_norm": 0.38933128118515015, + "learning_rate": 5.50534147840596e-05, + "loss": 0.0553, + "step": 27663 + }, + { + "epoch": 0.6484877254053928, + "grad_norm": 0.28922948241233826, + "learning_rate": 5.504683473553079e-05, + "loss": 0.0385, + "step": 27664 + }, + { + "epoch": 0.6485111669801977, + "grad_norm": 0.3993701934814453, + "learning_rate": 5.5040254930919136e-05, + "loss": 0.0961, + "step": 27665 + }, + { + "epoch": 0.6485346085550028, + "grad_norm": 0.1443602591753006, + "learning_rate": 5.503367537026031e-05, + "loss": 0.0239, + "step": 27666 + }, + { + "epoch": 0.6485580501298077, + "grad_norm": 0.5445528626441956, + "learning_rate": 5.5027096053589975e-05, + "loss": 0.0642, + "step": 27667 + }, + { + "epoch": 0.6485814917046128, + "grad_norm": 0.3396929204463959, + "learning_rate": 5.502051698094395e-05, + "loss": 0.049, + "step": 27668 + }, + { + "epoch": 0.6486049332794177, + "grad_norm": 0.9477288722991943, + "learning_rate": 5.5013938152357845e-05, + "loss": 0.1698, + "step": 27669 + }, + { + "epoch": 0.6486283748542228, + "grad_norm": 0.5070169568061829, + "learning_rate": 5.500735956786738e-05, + "loss": 0.0683, + "step": 27670 + }, + { + "epoch": 0.6486518164290277, + "grad_norm": 0.6062853932380676, + "learning_rate": 5.5000781227508244e-05, + "loss": 0.0961, + "step": 27671 + }, + { + "epoch": 0.6486752580038327, + "grad_norm": 0.4025551676750183, + "learning_rate": 5.499420313131614e-05, + "loss": 0.0563, + "step": 27672 + }, + { + "epoch": 0.6486986995786377, + "grad_norm": 0.5124267339706421, + "learning_rate": 5.49876252793267e-05, + "loss": 0.0855, + "step": 27673 + }, + { + "epoch": 0.6487221411534427, + "grad_norm": 0.16552935540676117, + "learning_rate": 5.498104767157573e-05, + "loss": 0.0241, + "step": 27674 + }, + { + "epoch": 0.6487455827282477, + "grad_norm": 0.13193637132644653, + "learning_rate": 5.4974470308098814e-05, + "loss": 0.0196, + "step": 27675 + }, + { + "epoch": 0.6487690243030527, + "grad_norm": 0.2913009524345398, + "learning_rate": 5.496789318893174e-05, + "loss": 0.0338, + "step": 27676 + }, + { + "epoch": 0.6487924658778577, + "grad_norm": 0.3555962145328522, + "learning_rate": 5.496131631411012e-05, + "loss": 0.0287, + "step": 27677 + }, + { + "epoch": 0.6488159074526627, + "grad_norm": 0.25641846656799316, + "learning_rate": 5.4954739683669685e-05, + "loss": 0.0369, + "step": 27678 + }, + { + "epoch": 0.6488393490274676, + "grad_norm": 0.5804362893104553, + "learning_rate": 5.494816329764605e-05, + "loss": 0.0345, + "step": 27679 + }, + { + "epoch": 0.6488627906022727, + "grad_norm": 0.7188860774040222, + "learning_rate": 5.494158715607499e-05, + "loss": 0.2949, + "step": 27680 + }, + { + "epoch": 0.6488862321770776, + "grad_norm": 0.1307939738035202, + "learning_rate": 5.493501125899214e-05, + "loss": 0.0156, + "step": 27681 + }, + { + "epoch": 0.6489096737518827, + "grad_norm": 0.5030290484428406, + "learning_rate": 5.4928435606433194e-05, + "loss": 0.0957, + "step": 27682 + }, + { + "epoch": 0.6489331153266876, + "grad_norm": 0.37138646841049194, + "learning_rate": 5.4921860198433816e-05, + "loss": 0.0634, + "step": 27683 + }, + { + "epoch": 0.6489565569014927, + "grad_norm": 0.20129655301570892, + "learning_rate": 5.4915285035029696e-05, + "loss": 0.0232, + "step": 27684 + }, + { + "epoch": 0.6489799984762976, + "grad_norm": 0.15543055534362793, + "learning_rate": 5.490871011625647e-05, + "loss": 0.0261, + "step": 27685 + }, + { + "epoch": 0.6490034400511027, + "grad_norm": 0.5475608706474304, + "learning_rate": 5.4902135442149884e-05, + "loss": 0.085, + "step": 27686 + }, + { + "epoch": 0.6490268816259076, + "grad_norm": 0.5076895952224731, + "learning_rate": 5.4895561012745545e-05, + "loss": 0.1642, + "step": 27687 + }, + { + "epoch": 0.6490503232007127, + "grad_norm": 0.2973371148109436, + "learning_rate": 5.4888986828079193e-05, + "loss": 0.018, + "step": 27688 + }, + { + "epoch": 0.6490737647755176, + "grad_norm": 0.23559384047985077, + "learning_rate": 5.488241288818647e-05, + "loss": 0.0574, + "step": 27689 + }, + { + "epoch": 0.6490972063503226, + "grad_norm": 0.4114783704280853, + "learning_rate": 5.487583919310305e-05, + "loss": 0.0608, + "step": 27690 + }, + { + "epoch": 0.6491206479251276, + "grad_norm": 0.18126046657562256, + "learning_rate": 5.4869265742864596e-05, + "loss": 0.0374, + "step": 27691 + }, + { + "epoch": 0.6491440894999326, + "grad_norm": 0.5863878726959229, + "learning_rate": 5.486269253750673e-05, + "loss": 0.6115, + "step": 27692 + }, + { + "epoch": 0.6491675310747376, + "grad_norm": 0.5080339312553406, + "learning_rate": 5.48561195770652e-05, + "loss": 0.1187, + "step": 27693 + }, + { + "epoch": 0.6491909726495426, + "grad_norm": 0.4636751711368561, + "learning_rate": 5.484954686157564e-05, + "loss": 0.0569, + "step": 27694 + }, + { + "epoch": 0.6492144142243476, + "grad_norm": 0.5747190117835999, + "learning_rate": 5.484297439107371e-05, + "loss": 0.1007, + "step": 27695 + }, + { + "epoch": 0.6492378557991526, + "grad_norm": 0.5491540431976318, + "learning_rate": 5.483640216559506e-05, + "loss": 0.1261, + "step": 27696 + }, + { + "epoch": 0.6492612973739575, + "grad_norm": 0.29354941844940186, + "learning_rate": 5.482983018517536e-05, + "loss": 0.0461, + "step": 27697 + }, + { + "epoch": 0.6492847389487626, + "grad_norm": 0.3983078598976135, + "learning_rate": 5.4823258449850246e-05, + "loss": 0.0567, + "step": 27698 + }, + { + "epoch": 0.6493081805235675, + "grad_norm": 0.682677686214447, + "learning_rate": 5.4816686959655425e-05, + "loss": 0.1843, + "step": 27699 + }, + { + "epoch": 0.6493316220983726, + "grad_norm": 0.6844687461853027, + "learning_rate": 5.481011571462651e-05, + "loss": 0.5248, + "step": 27700 + }, + { + "epoch": 0.6493550636731775, + "grad_norm": 0.2590080499649048, + "learning_rate": 5.4803544714799203e-05, + "loss": 0.0658, + "step": 27701 + }, + { + "epoch": 0.6493785052479826, + "grad_norm": 0.5668297410011292, + "learning_rate": 5.479697396020915e-05, + "loss": 0.0926, + "step": 27702 + }, + { + "epoch": 0.6494019468227875, + "grad_norm": 0.40737804770469666, + "learning_rate": 5.4790403450891967e-05, + "loss": 0.0521, + "step": 27703 + }, + { + "epoch": 0.6494253883975926, + "grad_norm": 0.3031606674194336, + "learning_rate": 5.478383318688334e-05, + "loss": 0.0651, + "step": 27704 + }, + { + "epoch": 0.6494488299723975, + "grad_norm": 0.24652525782585144, + "learning_rate": 5.477726316821886e-05, + "loss": 0.0339, + "step": 27705 + }, + { + "epoch": 0.6494722715472026, + "grad_norm": 0.3072455823421478, + "learning_rate": 5.4770693394934256e-05, + "loss": 0.0692, + "step": 27706 + }, + { + "epoch": 0.6494957131220075, + "grad_norm": 0.561962366104126, + "learning_rate": 5.476412386706515e-05, + "loss": 0.0448, + "step": 27707 + }, + { + "epoch": 0.6495191546968125, + "grad_norm": 0.41565847396850586, + "learning_rate": 5.4757554584647185e-05, + "loss": 0.0863, + "step": 27708 + }, + { + "epoch": 0.6495425962716175, + "grad_norm": 0.32147324085235596, + "learning_rate": 5.4750985547716e-05, + "loss": 0.0724, + "step": 27709 + }, + { + "epoch": 0.6495660378464225, + "grad_norm": 0.2524997889995575, + "learning_rate": 5.474441675630723e-05, + "loss": 0.0344, + "step": 27710 + }, + { + "epoch": 0.6495894794212275, + "grad_norm": 0.13371595740318298, + "learning_rate": 5.473784821045649e-05, + "loss": 0.0185, + "step": 27711 + }, + { + "epoch": 0.6496129209960325, + "grad_norm": 0.20926494896411896, + "learning_rate": 5.4731279910199465e-05, + "loss": 0.0507, + "step": 27712 + }, + { + "epoch": 0.6496363625708375, + "grad_norm": 0.180466428399086, + "learning_rate": 5.472471185557182e-05, + "loss": 0.0352, + "step": 27713 + }, + { + "epoch": 0.6496598041456425, + "grad_norm": 0.3635503649711609, + "learning_rate": 5.4718144046609154e-05, + "loss": 0.0446, + "step": 27714 + }, + { + "epoch": 0.6496832457204474, + "grad_norm": 0.1795368641614914, + "learning_rate": 5.471157648334713e-05, + "loss": 0.0354, + "step": 27715 + }, + { + "epoch": 0.6497066872952525, + "grad_norm": 0.5911720991134644, + "learning_rate": 5.470500916582135e-05, + "loss": 0.1172, + "step": 27716 + }, + { + "epoch": 0.6497301288700575, + "grad_norm": 0.3257904350757599, + "learning_rate": 5.4698442094067426e-05, + "loss": 0.0898, + "step": 27717 + }, + { + "epoch": 0.6497535704448625, + "grad_norm": 0.42842209339141846, + "learning_rate": 5.469187526812107e-05, + "loss": 0.0847, + "step": 27718 + }, + { + "epoch": 0.6497770120196675, + "grad_norm": 0.3220343589782715, + "learning_rate": 5.468530868801789e-05, + "loss": 0.0529, + "step": 27719 + }, + { + "epoch": 0.6498004535944725, + "grad_norm": 0.31093308329582214, + "learning_rate": 5.4678742353793475e-05, + "loss": 0.0505, + "step": 27720 + }, + { + "epoch": 0.6498238951692775, + "grad_norm": 0.5021570920944214, + "learning_rate": 5.4672176265483486e-05, + "loss": 0.08, + "step": 27721 + }, + { + "epoch": 0.6498473367440825, + "grad_norm": 0.44525593519210815, + "learning_rate": 5.4665610423123546e-05, + "loss": 0.0849, + "step": 27722 + }, + { + "epoch": 0.6498707783188875, + "grad_norm": 0.7629275321960449, + "learning_rate": 5.465904482674924e-05, + "loss": 0.0908, + "step": 27723 + }, + { + "epoch": 0.6498942198936924, + "grad_norm": 0.45670273900032043, + "learning_rate": 5.4652479476396234e-05, + "loss": 0.0585, + "step": 27724 + }, + { + "epoch": 0.6499176614684975, + "grad_norm": 0.2981577515602112, + "learning_rate": 5.464591437210018e-05, + "loss": 0.0383, + "step": 27725 + }, + { + "epoch": 0.6499411030433024, + "grad_norm": 0.25924038887023926, + "learning_rate": 5.463934951389669e-05, + "loss": 0.0426, + "step": 27726 + }, + { + "epoch": 0.6499645446181075, + "grad_norm": 0.4115352928638458, + "learning_rate": 5.4632784901821335e-05, + "loss": 0.0409, + "step": 27727 + }, + { + "epoch": 0.6499879861929124, + "grad_norm": 0.2618597447872162, + "learning_rate": 5.462622053590978e-05, + "loss": 0.1304, + "step": 27728 + }, + { + "epoch": 0.6500114277677175, + "grad_norm": 0.18064184486865997, + "learning_rate": 5.461965641619764e-05, + "loss": 0.0312, + "step": 27729 + }, + { + "epoch": 0.6500348693425224, + "grad_norm": 0.6180480122566223, + "learning_rate": 5.461309254272047e-05, + "loss": 0.0614, + "step": 27730 + }, + { + "epoch": 0.6500583109173275, + "grad_norm": 0.15525469183921814, + "learning_rate": 5.4606528915513965e-05, + "loss": 0.0333, + "step": 27731 + }, + { + "epoch": 0.6500817524921324, + "grad_norm": 0.3686022162437439, + "learning_rate": 5.459996553461372e-05, + "loss": 0.0617, + "step": 27732 + }, + { + "epoch": 0.6501051940669375, + "grad_norm": 0.42646825313568115, + "learning_rate": 5.459340240005534e-05, + "loss": 0.0774, + "step": 27733 + }, + { + "epoch": 0.6501286356417424, + "grad_norm": 0.4492097795009613, + "learning_rate": 5.4586839511874435e-05, + "loss": 0.059, + "step": 27734 + }, + { + "epoch": 0.6501520772165474, + "grad_norm": 0.47900891304016113, + "learning_rate": 5.45802768701066e-05, + "loss": 0.4375, + "step": 27735 + }, + { + "epoch": 0.6501755187913524, + "grad_norm": 0.396393746137619, + "learning_rate": 5.457371447478745e-05, + "loss": 0.0753, + "step": 27736 + }, + { + "epoch": 0.6501989603661574, + "grad_norm": 0.4162052571773529, + "learning_rate": 5.456715232595257e-05, + "loss": 0.4061, + "step": 27737 + }, + { + "epoch": 0.6502224019409624, + "grad_norm": 0.140027716755867, + "learning_rate": 5.456059042363766e-05, + "loss": 0.0298, + "step": 27738 + }, + { + "epoch": 0.6502458435157674, + "grad_norm": 0.155838280916214, + "learning_rate": 5.455402876787827e-05, + "loss": 0.0269, + "step": 27739 + }, + { + "epoch": 0.6502692850905724, + "grad_norm": 0.33666476607322693, + "learning_rate": 5.454746735870998e-05, + "loss": 0.0215, + "step": 27740 + }, + { + "epoch": 0.6502927266653774, + "grad_norm": 0.4299520254135132, + "learning_rate": 5.454090619616843e-05, + "loss": 0.0499, + "step": 27741 + }, + { + "epoch": 0.6503161682401823, + "grad_norm": 0.40010157227516174, + "learning_rate": 5.453434528028918e-05, + "loss": 0.0806, + "step": 27742 + }, + { + "epoch": 0.6503396098149874, + "grad_norm": 0.3785281777381897, + "learning_rate": 5.452778461110781e-05, + "loss": 0.0979, + "step": 27743 + }, + { + "epoch": 0.6503630513897923, + "grad_norm": 0.24706405401229858, + "learning_rate": 5.4521224188660016e-05, + "loss": 0.0389, + "step": 27744 + }, + { + "epoch": 0.6503864929645974, + "grad_norm": 0.7472342252731323, + "learning_rate": 5.451466401298133e-05, + "loss": 0.1207, + "step": 27745 + }, + { + "epoch": 0.6504099345394023, + "grad_norm": 0.5676332712173462, + "learning_rate": 5.450810408410736e-05, + "loss": 0.0622, + "step": 27746 + }, + { + "epoch": 0.6504333761142074, + "grad_norm": 0.45895054936408997, + "learning_rate": 5.450154440207369e-05, + "loss": 0.08, + "step": 27747 + }, + { + "epoch": 0.6504568176890123, + "grad_norm": 0.38682594895362854, + "learning_rate": 5.449498496691589e-05, + "loss": 0.0478, + "step": 27748 + }, + { + "epoch": 0.6504802592638174, + "grad_norm": 0.2136612981557846, + "learning_rate": 5.4488425778669616e-05, + "loss": 0.055, + "step": 27749 + }, + { + "epoch": 0.6505037008386223, + "grad_norm": 0.7874125838279724, + "learning_rate": 5.448186683737038e-05, + "loss": 0.1767, + "step": 27750 + }, + { + "epoch": 0.6505271424134274, + "grad_norm": 0.7962357997894287, + "learning_rate": 5.447530814305385e-05, + "loss": 0.6377, + "step": 27751 + }, + { + "epoch": 0.6505505839882323, + "grad_norm": 0.8848236799240112, + "learning_rate": 5.446874969575558e-05, + "loss": 0.1173, + "step": 27752 + }, + { + "epoch": 0.6505740255630373, + "grad_norm": 0.5596956610679626, + "learning_rate": 5.446219149551116e-05, + "loss": 0.1147, + "step": 27753 + }, + { + "epoch": 0.6505974671378423, + "grad_norm": 0.25009551644325256, + "learning_rate": 5.445563354235616e-05, + "loss": 0.0463, + "step": 27754 + }, + { + "epoch": 0.6506209087126473, + "grad_norm": 0.4605143070220947, + "learning_rate": 5.4449075836326144e-05, + "loss": 0.1299, + "step": 27755 + }, + { + "epoch": 0.6506443502874523, + "grad_norm": 0.8671454191207886, + "learning_rate": 5.444251837745676e-05, + "loss": 0.1661, + "step": 27756 + }, + { + "epoch": 0.6506677918622573, + "grad_norm": 0.30221718549728394, + "learning_rate": 5.443596116578356e-05, + "loss": 0.0375, + "step": 27757 + }, + { + "epoch": 0.6506912334370623, + "grad_norm": 0.35136157274246216, + "learning_rate": 5.4429404201342105e-05, + "loss": 0.0731, + "step": 27758 + }, + { + "epoch": 0.6507146750118673, + "grad_norm": 0.38551631569862366, + "learning_rate": 5.4422847484167995e-05, + "loss": 0.0807, + "step": 27759 + }, + { + "epoch": 0.6507381165866722, + "grad_norm": 0.5836222767829895, + "learning_rate": 5.441629101429674e-05, + "loss": 0.1385, + "step": 27760 + }, + { + "epoch": 0.6507615581614773, + "grad_norm": 0.13114260137081146, + "learning_rate": 5.440973479176402e-05, + "loss": 0.0243, + "step": 27761 + }, + { + "epoch": 0.6507849997362822, + "grad_norm": 0.3325056731700897, + "learning_rate": 5.440317881660533e-05, + "loss": 0.0667, + "step": 27762 + }, + { + "epoch": 0.6508084413110873, + "grad_norm": 0.36274823546409607, + "learning_rate": 5.4396623088856294e-05, + "loss": 0.3102, + "step": 27763 + }, + { + "epoch": 0.6508318828858922, + "grad_norm": 0.6799209117889404, + "learning_rate": 5.439006760855248e-05, + "loss": 0.1502, + "step": 27764 + }, + { + "epoch": 0.6508553244606973, + "grad_norm": 0.4590297341346741, + "learning_rate": 5.438351237572943e-05, + "loss": 0.1217, + "step": 27765 + }, + { + "epoch": 0.6508787660355022, + "grad_norm": 0.5984391570091248, + "learning_rate": 5.437695739042272e-05, + "loss": 0.522, + "step": 27766 + }, + { + "epoch": 0.6509022076103073, + "grad_norm": 0.38659238815307617, + "learning_rate": 5.437040265266794e-05, + "loss": 0.0659, + "step": 27767 + }, + { + "epoch": 0.6509256491851123, + "grad_norm": 0.5571165680885315, + "learning_rate": 5.4363848162500595e-05, + "loss": 0.1174, + "step": 27768 + }, + { + "epoch": 0.6509490907599172, + "grad_norm": 0.5867059826850891, + "learning_rate": 5.435729391995632e-05, + "loss": 0.0709, + "step": 27769 + }, + { + "epoch": 0.6509725323347223, + "grad_norm": 0.7239621877670288, + "learning_rate": 5.435073992507067e-05, + "loss": 0.0879, + "step": 27770 + }, + { + "epoch": 0.6509959739095272, + "grad_norm": 1.0811502933502197, + "learning_rate": 5.434418617787918e-05, + "loss": 0.0726, + "step": 27771 + }, + { + "epoch": 0.6510194154843323, + "grad_norm": 0.11094973981380463, + "learning_rate": 5.433763267841737e-05, + "loss": 0.0217, + "step": 27772 + }, + { + "epoch": 0.6510428570591372, + "grad_norm": 0.1345207393169403, + "learning_rate": 5.4331079426720896e-05, + "loss": 0.0283, + "step": 27773 + }, + { + "epoch": 0.6510662986339423, + "grad_norm": 0.3164036273956299, + "learning_rate": 5.432452642282527e-05, + "loss": 0.0525, + "step": 27774 + }, + { + "epoch": 0.6510897402087472, + "grad_norm": 0.49935173988342285, + "learning_rate": 5.431797366676601e-05, + "loss": 0.0711, + "step": 27775 + }, + { + "epoch": 0.6511131817835523, + "grad_norm": 0.4395046830177307, + "learning_rate": 5.4311421158578744e-05, + "loss": 0.0897, + "step": 27776 + }, + { + "epoch": 0.6511366233583572, + "grad_norm": 0.7000748515129089, + "learning_rate": 5.4304868898299e-05, + "loss": 0.0479, + "step": 27777 + }, + { + "epoch": 0.6511600649331623, + "grad_norm": 0.6298531889915466, + "learning_rate": 5.429831688596232e-05, + "loss": 0.0828, + "step": 27778 + }, + { + "epoch": 0.6511835065079672, + "grad_norm": 0.307953804731369, + "learning_rate": 5.4291765121604256e-05, + "loss": 0.0688, + "step": 27779 + }, + { + "epoch": 0.6512069480827722, + "grad_norm": 0.24165424704551697, + "learning_rate": 5.428521360526036e-05, + "loss": 0.0225, + "step": 27780 + }, + { + "epoch": 0.6512303896575772, + "grad_norm": 0.48748913407325745, + "learning_rate": 5.427866233696613e-05, + "loss": 0.6252, + "step": 27781 + }, + { + "epoch": 0.6512538312323822, + "grad_norm": 0.32275068759918213, + "learning_rate": 5.4272111316757215e-05, + "loss": 0.0299, + "step": 27782 + }, + { + "epoch": 0.6512772728071872, + "grad_norm": 0.4574848711490631, + "learning_rate": 5.42655605446691e-05, + "loss": 0.0875, + "step": 27783 + }, + { + "epoch": 0.6513007143819922, + "grad_norm": 0.20503303408622742, + "learning_rate": 5.425901002073734e-05, + "loss": 0.0343, + "step": 27784 + }, + { + "epoch": 0.6513241559567972, + "grad_norm": 0.6945239305496216, + "learning_rate": 5.425245974499745e-05, + "loss": 0.1159, + "step": 27785 + }, + { + "epoch": 0.6513475975316022, + "grad_norm": 0.3733011484146118, + "learning_rate": 5.4245909717485045e-05, + "loss": 0.3689, + "step": 27786 + }, + { + "epoch": 0.6513710391064071, + "grad_norm": 0.687828004360199, + "learning_rate": 5.42393599382356e-05, + "loss": 0.076, + "step": 27787 + }, + { + "epoch": 0.6513944806812122, + "grad_norm": 0.09743236005306244, + "learning_rate": 5.4232810407284653e-05, + "loss": 0.0133, + "step": 27788 + }, + { + "epoch": 0.6514179222560171, + "grad_norm": 0.20649226009845734, + "learning_rate": 5.4226261124667786e-05, + "loss": 0.0327, + "step": 27789 + }, + { + "epoch": 0.6514413638308222, + "grad_norm": 0.4381125867366791, + "learning_rate": 5.4219712090420514e-05, + "loss": 0.0953, + "step": 27790 + }, + { + "epoch": 0.6514648054056271, + "grad_norm": 0.6740872859954834, + "learning_rate": 5.421316330457839e-05, + "loss": 0.1414, + "step": 27791 + }, + { + "epoch": 0.6514882469804322, + "grad_norm": 0.18880978226661682, + "learning_rate": 5.420661476717692e-05, + "loss": 0.0227, + "step": 27792 + }, + { + "epoch": 0.6515116885552371, + "grad_norm": 0.25833946466445923, + "learning_rate": 5.42000664782516e-05, + "loss": 0.0555, + "step": 27793 + }, + { + "epoch": 0.6515351301300422, + "grad_norm": 0.3744635283946991, + "learning_rate": 5.419351843783805e-05, + "loss": 0.0418, + "step": 27794 + }, + { + "epoch": 0.6515585717048471, + "grad_norm": 0.46150246262550354, + "learning_rate": 5.418697064597176e-05, + "loss": 0.1339, + "step": 27795 + }, + { + "epoch": 0.6515820132796522, + "grad_norm": 0.298125684261322, + "learning_rate": 5.418042310268825e-05, + "loss": 0.0778, + "step": 27796 + }, + { + "epoch": 0.6516054548544571, + "grad_norm": 0.4346739649772644, + "learning_rate": 5.417387580802302e-05, + "loss": 0.0959, + "step": 27797 + }, + { + "epoch": 0.6516288964292621, + "grad_norm": 0.6140979528427124, + "learning_rate": 5.416732876201167e-05, + "loss": 0.1896, + "step": 27798 + }, + { + "epoch": 0.6516523380040671, + "grad_norm": 0.18901383876800537, + "learning_rate": 5.416078196468968e-05, + "loss": 0.0299, + "step": 27799 + }, + { + "epoch": 0.6516757795788721, + "grad_norm": 0.29523077607154846, + "learning_rate": 5.415423541609253e-05, + "loss": 0.0238, + "step": 27800 + }, + { + "epoch": 0.6516992211536771, + "grad_norm": 0.45620813965797424, + "learning_rate": 5.414768911625583e-05, + "loss": 0.0555, + "step": 27801 + }, + { + "epoch": 0.6517226627284821, + "grad_norm": 0.6099672913551331, + "learning_rate": 5.414114306521505e-05, + "loss": 0.0493, + "step": 27802 + }, + { + "epoch": 0.651746104303287, + "grad_norm": 0.4282570481300354, + "learning_rate": 5.4134597263005736e-05, + "loss": 0.0352, + "step": 27803 + }, + { + "epoch": 0.6517695458780921, + "grad_norm": 0.28842300176620483, + "learning_rate": 5.4128051709663374e-05, + "loss": 0.0453, + "step": 27804 + }, + { + "epoch": 0.651792987452897, + "grad_norm": 0.6553765535354614, + "learning_rate": 5.4121506405223486e-05, + "loss": 0.1637, + "step": 27805 + }, + { + "epoch": 0.6518164290277021, + "grad_norm": 0.5236862301826477, + "learning_rate": 5.4114961349721574e-05, + "loss": 0.6584, + "step": 27806 + }, + { + "epoch": 0.651839870602507, + "grad_norm": 0.3497430086135864, + "learning_rate": 5.4108416543193205e-05, + "loss": 0.0863, + "step": 27807 + }, + { + "epoch": 0.6518633121773121, + "grad_norm": 0.17156045138835907, + "learning_rate": 5.410187198567385e-05, + "loss": 0.0279, + "step": 27808 + }, + { + "epoch": 0.651886753752117, + "grad_norm": 0.5888344645500183, + "learning_rate": 5.4095327677199e-05, + "loss": 0.5172, + "step": 27809 + }, + { + "epoch": 0.6519101953269221, + "grad_norm": 0.2512569725513458, + "learning_rate": 5.4088783617804225e-05, + "loss": 0.0445, + "step": 27810 + }, + { + "epoch": 0.651933636901727, + "grad_norm": 0.1440783441066742, + "learning_rate": 5.4082239807525e-05, + "loss": 0.0373, + "step": 27811 + }, + { + "epoch": 0.6519570784765321, + "grad_norm": 0.4344354271888733, + "learning_rate": 5.407569624639683e-05, + "loss": 0.078, + "step": 27812 + }, + { + "epoch": 0.651980520051337, + "grad_norm": 0.3371862471103668, + "learning_rate": 5.40691529344552e-05, + "loss": 0.0739, + "step": 27813 + }, + { + "epoch": 0.652003961626142, + "grad_norm": 0.12079666554927826, + "learning_rate": 5.406260987173566e-05, + "loss": 0.0232, + "step": 27814 + }, + { + "epoch": 0.652027403200947, + "grad_norm": 0.22161999344825745, + "learning_rate": 5.405606705827371e-05, + "loss": 0.0224, + "step": 27815 + }, + { + "epoch": 0.652050844775752, + "grad_norm": 0.33148229122161865, + "learning_rate": 5.404952449410481e-05, + "loss": 0.0753, + "step": 27816 + }, + { + "epoch": 0.652074286350557, + "grad_norm": 0.36106938123703003, + "learning_rate": 5.404298217926449e-05, + "loss": 0.0489, + "step": 27817 + }, + { + "epoch": 0.652097727925362, + "grad_norm": 0.5222919583320618, + "learning_rate": 5.4036440113788244e-05, + "loss": 0.0555, + "step": 27818 + }, + { + "epoch": 0.6521211695001671, + "grad_norm": 0.34447550773620605, + "learning_rate": 5.4029898297711526e-05, + "loss": 0.1006, + "step": 27819 + }, + { + "epoch": 0.652144611074972, + "grad_norm": 0.5131435394287109, + "learning_rate": 5.402335673106991e-05, + "loss": 0.0842, + "step": 27820 + }, + { + "epoch": 0.6521680526497771, + "grad_norm": 0.3405489921569824, + "learning_rate": 5.4016815413898816e-05, + "loss": 0.0582, + "step": 27821 + }, + { + "epoch": 0.652191494224582, + "grad_norm": 0.14829683303833008, + "learning_rate": 5.40102743462338e-05, + "loss": 0.0141, + "step": 27822 + }, + { + "epoch": 0.652214935799387, + "grad_norm": 0.1351693719625473, + "learning_rate": 5.400373352811035e-05, + "loss": 0.0225, + "step": 27823 + }, + { + "epoch": 0.652238377374192, + "grad_norm": 0.34846654534339905, + "learning_rate": 5.399719295956393e-05, + "loss": 0.0559, + "step": 27824 + }, + { + "epoch": 0.652261818948997, + "grad_norm": 0.19883054494857788, + "learning_rate": 5.3990652640630025e-05, + "loss": 0.0257, + "step": 27825 + }, + { + "epoch": 0.652285260523802, + "grad_norm": 0.2040087878704071, + "learning_rate": 5.39841125713441e-05, + "loss": 0.0352, + "step": 27826 + }, + { + "epoch": 0.652308702098607, + "grad_norm": 0.2766905725002289, + "learning_rate": 5.397757275174171e-05, + "loss": 0.0448, + "step": 27827 + }, + { + "epoch": 0.652332143673412, + "grad_norm": 0.35644927620887756, + "learning_rate": 5.397103318185831e-05, + "loss": 0.0558, + "step": 27828 + }, + { + "epoch": 0.652355585248217, + "grad_norm": 0.292509526014328, + "learning_rate": 5.396449386172937e-05, + "loss": 0.0247, + "step": 27829 + }, + { + "epoch": 0.652379026823022, + "grad_norm": 0.3612540364265442, + "learning_rate": 5.3957954791390386e-05, + "loss": 0.0593, + "step": 27830 + }, + { + "epoch": 0.652402468397827, + "grad_norm": 0.6492856740951538, + "learning_rate": 5.3951415970876795e-05, + "loss": 0.0587, + "step": 27831 + }, + { + "epoch": 0.652425909972632, + "grad_norm": 0.49735498428344727, + "learning_rate": 5.394487740022416e-05, + "loss": 0.1263, + "step": 27832 + }, + { + "epoch": 0.652449351547437, + "grad_norm": 0.7612026929855347, + "learning_rate": 5.393833907946786e-05, + "loss": 0.0523, + "step": 27833 + }, + { + "epoch": 0.6524727931222419, + "grad_norm": 0.5566297173500061, + "learning_rate": 5.393180100864348e-05, + "loss": 0.1206, + "step": 27834 + }, + { + "epoch": 0.652496234697047, + "grad_norm": 0.22922413051128387, + "learning_rate": 5.3925263187786434e-05, + "loss": 0.032, + "step": 27835 + }, + { + "epoch": 0.6525196762718519, + "grad_norm": 0.7290578484535217, + "learning_rate": 5.391872561693221e-05, + "loss": 0.0702, + "step": 27836 + }, + { + "epoch": 0.652543117846657, + "grad_norm": 0.4029582142829895, + "learning_rate": 5.3912188296116283e-05, + "loss": 0.0849, + "step": 27837 + }, + { + "epoch": 0.6525665594214619, + "grad_norm": 0.4278336465358734, + "learning_rate": 5.3905651225374075e-05, + "loss": 0.065, + "step": 27838 + }, + { + "epoch": 0.652590000996267, + "grad_norm": 0.10853134095668793, + "learning_rate": 5.389911440474113e-05, + "loss": 0.0209, + "step": 27839 + }, + { + "epoch": 0.6526134425710719, + "grad_norm": 0.5878703594207764, + "learning_rate": 5.38925778342529e-05, + "loss": 0.0844, + "step": 27840 + }, + { + "epoch": 0.652636884145877, + "grad_norm": 0.14546208083629608, + "learning_rate": 5.388604151394483e-05, + "loss": 0.0375, + "step": 27841 + }, + { + "epoch": 0.6526603257206819, + "grad_norm": 0.11457506567239761, + "learning_rate": 5.3879505443852405e-05, + "loss": 0.0148, + "step": 27842 + }, + { + "epoch": 0.6526837672954869, + "grad_norm": 0.3837844431400299, + "learning_rate": 5.387296962401107e-05, + "loss": 0.0924, + "step": 27843 + }, + { + "epoch": 0.6527072088702919, + "grad_norm": 0.15244558453559875, + "learning_rate": 5.386643405445626e-05, + "loss": 0.0233, + "step": 27844 + }, + { + "epoch": 0.6527306504450969, + "grad_norm": 0.2999192774295807, + "learning_rate": 5.385989873522348e-05, + "loss": 0.059, + "step": 27845 + }, + { + "epoch": 0.6527540920199019, + "grad_norm": 0.6466639637947083, + "learning_rate": 5.385336366634822e-05, + "loss": 0.1536, + "step": 27846 + }, + { + "epoch": 0.6527775335947069, + "grad_norm": 0.42828208208084106, + "learning_rate": 5.384682884786592e-05, + "loss": 0.0457, + "step": 27847 + }, + { + "epoch": 0.6528009751695119, + "grad_norm": 0.11905539780855179, + "learning_rate": 5.384029427981201e-05, + "loss": 0.0165, + "step": 27848 + }, + { + "epoch": 0.6528244167443169, + "grad_norm": 0.2634156346321106, + "learning_rate": 5.383375996222196e-05, + "loss": 0.0271, + "step": 27849 + }, + { + "epoch": 0.6528478583191218, + "grad_norm": 0.07198372483253479, + "learning_rate": 5.382722589513124e-05, + "loss": 0.0111, + "step": 27850 + }, + { + "epoch": 0.6528712998939269, + "grad_norm": 0.12842877209186554, + "learning_rate": 5.3820692078575244e-05, + "loss": 0.0114, + "step": 27851 + }, + { + "epoch": 0.6528947414687318, + "grad_norm": 0.17054371535778046, + "learning_rate": 5.3814158512589505e-05, + "loss": 0.0409, + "step": 27852 + }, + { + "epoch": 0.6529181830435369, + "grad_norm": 0.26878541707992554, + "learning_rate": 5.3807625197209455e-05, + "loss": 0.0449, + "step": 27853 + }, + { + "epoch": 0.6529416246183418, + "grad_norm": 0.5699167251586914, + "learning_rate": 5.380109213247051e-05, + "loss": 0.0985, + "step": 27854 + }, + { + "epoch": 0.6529650661931469, + "grad_norm": 0.7405704855918884, + "learning_rate": 5.379455931840816e-05, + "loss": 0.1813, + "step": 27855 + }, + { + "epoch": 0.6529885077679518, + "grad_norm": 0.5313791632652283, + "learning_rate": 5.3788026755057805e-05, + "loss": 0.0828, + "step": 27856 + }, + { + "epoch": 0.6530119493427569, + "grad_norm": 0.545607328414917, + "learning_rate": 5.378149444245491e-05, + "loss": 0.0745, + "step": 27857 + }, + { + "epoch": 0.6530353909175618, + "grad_norm": 0.37929877638816833, + "learning_rate": 5.37749623806349e-05, + "loss": 0.0419, + "step": 27858 + }, + { + "epoch": 0.6530588324923668, + "grad_norm": 0.261892169713974, + "learning_rate": 5.376843056963329e-05, + "loss": 0.0427, + "step": 27859 + }, + { + "epoch": 0.6530822740671718, + "grad_norm": 0.17285774648189545, + "learning_rate": 5.376189900948547e-05, + "loss": 0.0294, + "step": 27860 + }, + { + "epoch": 0.6531057156419768, + "grad_norm": 0.6974220871925354, + "learning_rate": 5.375536770022689e-05, + "loss": 0.0442, + "step": 27861 + }, + { + "epoch": 0.6531291572167818, + "grad_norm": 0.6883987784385681, + "learning_rate": 5.374883664189299e-05, + "loss": 0.0689, + "step": 27862 + }, + { + "epoch": 0.6531525987915868, + "grad_norm": 0.3980990946292877, + "learning_rate": 5.3742305834519203e-05, + "loss": 0.0661, + "step": 27863 + }, + { + "epoch": 0.6531760403663918, + "grad_norm": 0.31650951504707336, + "learning_rate": 5.373577527814092e-05, + "loss": 0.0679, + "step": 27864 + }, + { + "epoch": 0.6531994819411968, + "grad_norm": 0.11305896937847137, + "learning_rate": 5.372924497279366e-05, + "loss": 0.0217, + "step": 27865 + }, + { + "epoch": 0.6532229235160018, + "grad_norm": 0.5538470149040222, + "learning_rate": 5.3722714918512816e-05, + "loss": 0.0806, + "step": 27866 + }, + { + "epoch": 0.6532463650908068, + "grad_norm": 0.3272162079811096, + "learning_rate": 5.371618511533384e-05, + "loss": 0.0441, + "step": 27867 + }, + { + "epoch": 0.6532698066656117, + "grad_norm": 0.208407923579216, + "learning_rate": 5.3709655563292125e-05, + "loss": 0.0198, + "step": 27868 + }, + { + "epoch": 0.6532932482404168, + "grad_norm": 0.3019541800022125, + "learning_rate": 5.370312626242309e-05, + "loss": 0.0408, + "step": 27869 + }, + { + "epoch": 0.6533166898152218, + "grad_norm": 0.7195683121681213, + "learning_rate": 5.369659721276224e-05, + "loss": 0.1741, + "step": 27870 + }, + { + "epoch": 0.6533401313900268, + "grad_norm": 0.48908087611198425, + "learning_rate": 5.36900684143449e-05, + "loss": 0.0593, + "step": 27871 + }, + { + "epoch": 0.6533635729648318, + "grad_norm": 0.2227831929922104, + "learning_rate": 5.36835398672066e-05, + "loss": 0.0218, + "step": 27872 + }, + { + "epoch": 0.6533870145396368, + "grad_norm": 0.6457773447036743, + "learning_rate": 5.367701157138271e-05, + "loss": 0.0607, + "step": 27873 + }, + { + "epoch": 0.6534104561144418, + "grad_norm": 0.27445247769355774, + "learning_rate": 5.367048352690866e-05, + "loss": 0.2118, + "step": 27874 + }, + { + "epoch": 0.6534338976892468, + "grad_norm": 0.8760389685630798, + "learning_rate": 5.366395573381987e-05, + "loss": 0.068, + "step": 27875 + }, + { + "epoch": 0.6534573392640518, + "grad_norm": 0.37471112608909607, + "learning_rate": 5.365742819215172e-05, + "loss": 0.0549, + "step": 27876 + }, + { + "epoch": 0.6534807808388567, + "grad_norm": 2.3325140476226807, + "learning_rate": 5.36509009019397e-05, + "loss": 0.122, + "step": 27877 + }, + { + "epoch": 0.6535042224136618, + "grad_norm": 0.08074858784675598, + "learning_rate": 5.364437386321921e-05, + "loss": 0.0154, + "step": 27878 + }, + { + "epoch": 0.6535276639884667, + "grad_norm": 0.3263956606388092, + "learning_rate": 5.363784707602564e-05, + "loss": 0.0945, + "step": 27879 + }, + { + "epoch": 0.6535511055632718, + "grad_norm": 0.4350654184818268, + "learning_rate": 5.3631320540394414e-05, + "loss": 0.0732, + "step": 27880 + }, + { + "epoch": 0.6535745471380767, + "grad_norm": 0.4377255141735077, + "learning_rate": 5.362479425636096e-05, + "loss": 0.0798, + "step": 27881 + }, + { + "epoch": 0.6535979887128818, + "grad_norm": 0.46612778306007385, + "learning_rate": 5.361826822396061e-05, + "loss": 0.0601, + "step": 27882 + }, + { + "epoch": 0.6536214302876867, + "grad_norm": 0.6239333748817444, + "learning_rate": 5.3611742443228863e-05, + "loss": 0.0995, + "step": 27883 + }, + { + "epoch": 0.6536448718624918, + "grad_norm": 0.4481213092803955, + "learning_rate": 5.360521691420114e-05, + "loss": 0.0401, + "step": 27884 + }, + { + "epoch": 0.6536683134372967, + "grad_norm": 0.2643950581550598, + "learning_rate": 5.359869163691281e-05, + "loss": 0.0626, + "step": 27885 + }, + { + "epoch": 0.6536917550121017, + "grad_norm": 0.32206588983535767, + "learning_rate": 5.3592166611399296e-05, + "loss": 0.0807, + "step": 27886 + }, + { + "epoch": 0.6537151965869067, + "grad_norm": 0.091537706553936, + "learning_rate": 5.358564183769598e-05, + "loss": 0.0139, + "step": 27887 + }, + { + "epoch": 0.6537386381617117, + "grad_norm": 0.4888019859790802, + "learning_rate": 5.357911731583827e-05, + "loss": 0.0404, + "step": 27888 + }, + { + "epoch": 0.6537620797365167, + "grad_norm": 0.5677840709686279, + "learning_rate": 5.357259304586155e-05, + "loss": 0.0891, + "step": 27889 + }, + { + "epoch": 0.6537855213113217, + "grad_norm": 0.4372645318508148, + "learning_rate": 5.356606902780128e-05, + "loss": 0.0595, + "step": 27890 + }, + { + "epoch": 0.6538089628861267, + "grad_norm": 0.061440277844667435, + "learning_rate": 5.3559545261692826e-05, + "loss": 0.0067, + "step": 27891 + }, + { + "epoch": 0.6538324044609317, + "grad_norm": 0.202329620718956, + "learning_rate": 5.355302174757157e-05, + "loss": 0.0416, + "step": 27892 + }, + { + "epoch": 0.6538558460357367, + "grad_norm": 1.1763335466384888, + "learning_rate": 5.354649848547295e-05, + "loss": 0.1813, + "step": 27893 + }, + { + "epoch": 0.6538792876105417, + "grad_norm": 0.6189087629318237, + "learning_rate": 5.353997547543228e-05, + "loss": 0.8931, + "step": 27894 + }, + { + "epoch": 0.6539027291853466, + "grad_norm": 0.4825475513935089, + "learning_rate": 5.353345271748505e-05, + "loss": 0.1284, + "step": 27895 + }, + { + "epoch": 0.6539261707601517, + "grad_norm": 0.2147561013698578, + "learning_rate": 5.352693021166658e-05, + "loss": 0.032, + "step": 27896 + }, + { + "epoch": 0.6539496123349566, + "grad_norm": 0.7712482213973999, + "learning_rate": 5.352040795801232e-05, + "loss": 0.0844, + "step": 27897 + }, + { + "epoch": 0.6539730539097617, + "grad_norm": 0.356969952583313, + "learning_rate": 5.351388595655764e-05, + "loss": 0.069, + "step": 27898 + }, + { + "epoch": 0.6539964954845666, + "grad_norm": 0.38297346234321594, + "learning_rate": 5.350736420733793e-05, + "loss": 0.0394, + "step": 27899 + }, + { + "epoch": 0.6540199370593717, + "grad_norm": 0.25624537467956543, + "learning_rate": 5.3500842710388565e-05, + "loss": 0.0348, + "step": 27900 + }, + { + "epoch": 0.6540433786341766, + "grad_norm": 1.0477102994918823, + "learning_rate": 5.349432146574493e-05, + "loss": 0.2609, + "step": 27901 + }, + { + "epoch": 0.6540668202089817, + "grad_norm": 0.14820925891399384, + "learning_rate": 5.348780047344237e-05, + "loss": 0.0211, + "step": 27902 + }, + { + "epoch": 0.6540902617837866, + "grad_norm": 0.21413570642471313, + "learning_rate": 5.3481279733516365e-05, + "loss": 0.0511, + "step": 27903 + }, + { + "epoch": 0.6541137033585916, + "grad_norm": 0.6391478180885315, + "learning_rate": 5.347475924600224e-05, + "loss": 0.0749, + "step": 27904 + }, + { + "epoch": 0.6541371449333966, + "grad_norm": 0.4509938061237335, + "learning_rate": 5.346823901093538e-05, + "loss": 0.0929, + "step": 27905 + }, + { + "epoch": 0.6541605865082016, + "grad_norm": 0.4694367051124573, + "learning_rate": 5.346171902835112e-05, + "loss": 0.0937, + "step": 27906 + }, + { + "epoch": 0.6541840280830066, + "grad_norm": 0.551567018032074, + "learning_rate": 5.345519929828492e-05, + "loss": 0.0687, + "step": 27907 + }, + { + "epoch": 0.6542074696578116, + "grad_norm": 0.5043877363204956, + "learning_rate": 5.344867982077212e-05, + "loss": 0.0749, + "step": 27908 + }, + { + "epoch": 0.6542309112326166, + "grad_norm": 0.3528675436973572, + "learning_rate": 5.3442160595848056e-05, + "loss": 0.066, + "step": 27909 + }, + { + "epoch": 0.6542543528074216, + "grad_norm": 0.1276991069316864, + "learning_rate": 5.343564162354817e-05, + "loss": 0.022, + "step": 27910 + }, + { + "epoch": 0.6542777943822266, + "grad_norm": 0.5394777059555054, + "learning_rate": 5.34291229039078e-05, + "loss": 0.0956, + "step": 27911 + }, + { + "epoch": 0.6543012359570316, + "grad_norm": 0.421700656414032, + "learning_rate": 5.342260443696232e-05, + "loss": 0.0838, + "step": 27912 + }, + { + "epoch": 0.6543246775318365, + "grad_norm": 0.3956778347492218, + "learning_rate": 5.341608622274711e-05, + "loss": 0.0334, + "step": 27913 + }, + { + "epoch": 0.6543481191066416, + "grad_norm": 0.5267444849014282, + "learning_rate": 5.3409568261297484e-05, + "loss": 0.1106, + "step": 27914 + }, + { + "epoch": 0.6543715606814465, + "grad_norm": 0.4488515555858612, + "learning_rate": 5.340305055264887e-05, + "loss": 0.0623, + "step": 27915 + }, + { + "epoch": 0.6543950022562516, + "grad_norm": 0.16335134208202362, + "learning_rate": 5.3396533096836635e-05, + "loss": 0.0415, + "step": 27916 + }, + { + "epoch": 0.6544184438310565, + "grad_norm": 0.24320079386234283, + "learning_rate": 5.339001589389612e-05, + "loss": 0.0311, + "step": 27917 + }, + { + "epoch": 0.6544418854058616, + "grad_norm": 0.27896425127983093, + "learning_rate": 5.338349894386264e-05, + "loss": 0.0402, + "step": 27918 + }, + { + "epoch": 0.6544653269806665, + "grad_norm": 0.15298107266426086, + "learning_rate": 5.3376982246771656e-05, + "loss": 0.0092, + "step": 27919 + }, + { + "epoch": 0.6544887685554716, + "grad_norm": 0.1278580278158188, + "learning_rate": 5.337046580265847e-05, + "loss": 0.0118, + "step": 27920 + }, + { + "epoch": 0.6545122101302766, + "grad_norm": 0.1425095945596695, + "learning_rate": 5.3363949611558404e-05, + "loss": 0.0239, + "step": 27921 + }, + { + "epoch": 0.6545356517050815, + "grad_norm": 0.1412239372730255, + "learning_rate": 5.335743367350689e-05, + "loss": 0.0379, + "step": 27922 + }, + { + "epoch": 0.6545590932798866, + "grad_norm": 0.6192347407341003, + "learning_rate": 5.3350917988539285e-05, + "loss": 0.0667, + "step": 27923 + }, + { + "epoch": 0.6545825348546915, + "grad_norm": 0.4383031129837036, + "learning_rate": 5.3344402556690886e-05, + "loss": 0.0862, + "step": 27924 + }, + { + "epoch": 0.6546059764294966, + "grad_norm": 0.40475553274154663, + "learning_rate": 5.333788737799707e-05, + "loss": 0.0838, + "step": 27925 + }, + { + "epoch": 0.6546294180043015, + "grad_norm": 0.5029616951942444, + "learning_rate": 5.33313724524932e-05, + "loss": 0.0742, + "step": 27926 + }, + { + "epoch": 0.6546528595791066, + "grad_norm": 0.2634128928184509, + "learning_rate": 5.332485778021457e-05, + "loss": 0.0307, + "step": 27927 + }, + { + "epoch": 0.6546763011539115, + "grad_norm": 0.46434593200683594, + "learning_rate": 5.331834336119661e-05, + "loss": 0.0614, + "step": 27928 + }, + { + "epoch": 0.6546997427287166, + "grad_norm": 0.302333801984787, + "learning_rate": 5.331182919547464e-05, + "loss": 0.0402, + "step": 27929 + }, + { + "epoch": 0.6547231843035215, + "grad_norm": 0.14208172261714935, + "learning_rate": 5.3305315283084e-05, + "loss": 0.031, + "step": 27930 + }, + { + "epoch": 0.6547466258783265, + "grad_norm": 0.16682057082653046, + "learning_rate": 5.329880162405998e-05, + "loss": 0.0161, + "step": 27931 + }, + { + "epoch": 0.6547700674531315, + "grad_norm": 0.4910765290260315, + "learning_rate": 5.329228821843801e-05, + "loss": 0.0749, + "step": 27932 + }, + { + "epoch": 0.6547935090279365, + "grad_norm": 0.26783469319343567, + "learning_rate": 5.3285775066253415e-05, + "loss": 0.0671, + "step": 27933 + }, + { + "epoch": 0.6548169506027415, + "grad_norm": 0.20278432965278625, + "learning_rate": 5.3279262167541476e-05, + "loss": 0.0502, + "step": 27934 + }, + { + "epoch": 0.6548403921775465, + "grad_norm": 0.6699065566062927, + "learning_rate": 5.327274952233761e-05, + "loss": 0.0707, + "step": 27935 + }, + { + "epoch": 0.6548638337523515, + "grad_norm": 1.2015330791473389, + "learning_rate": 5.326623713067711e-05, + "loss": 0.1323, + "step": 27936 + }, + { + "epoch": 0.6548872753271565, + "grad_norm": 0.4710773229598999, + "learning_rate": 5.3259724992595336e-05, + "loss": 0.111, + "step": 27937 + }, + { + "epoch": 0.6549107169019615, + "grad_norm": 0.6704079508781433, + "learning_rate": 5.325321310812761e-05, + "loss": 0.0739, + "step": 27938 + }, + { + "epoch": 0.6549341584767665, + "grad_norm": 0.5377764701843262, + "learning_rate": 5.3246701477309255e-05, + "loss": 0.0998, + "step": 27939 + }, + { + "epoch": 0.6549576000515714, + "grad_norm": 0.6685393452644348, + "learning_rate": 5.324019010017557e-05, + "loss": 0.1948, + "step": 27940 + }, + { + "epoch": 0.6549810416263765, + "grad_norm": 0.19452811777591705, + "learning_rate": 5.323367897676197e-05, + "loss": 0.0319, + "step": 27941 + }, + { + "epoch": 0.6550044832011814, + "grad_norm": 0.6555112600326538, + "learning_rate": 5.3227168107103755e-05, + "loss": 0.1236, + "step": 27942 + }, + { + "epoch": 0.6550279247759865, + "grad_norm": 0.1576927900314331, + "learning_rate": 5.322065749123618e-05, + "loss": 0.0091, + "step": 27943 + }, + { + "epoch": 0.6550513663507914, + "grad_norm": 0.7019968032836914, + "learning_rate": 5.321414712919468e-05, + "loss": 0.0979, + "step": 27944 + }, + { + "epoch": 0.6550748079255965, + "grad_norm": 0.5117219686508179, + "learning_rate": 5.3207637021014545e-05, + "loss": 0.0536, + "step": 27945 + }, + { + "epoch": 0.6550982495004014, + "grad_norm": 0.4798985719680786, + "learning_rate": 5.320112716673108e-05, + "loss": 0.0827, + "step": 27946 + }, + { + "epoch": 0.6551216910752065, + "grad_norm": 0.3483162224292755, + "learning_rate": 5.319461756637958e-05, + "loss": 0.0531, + "step": 27947 + }, + { + "epoch": 0.6551451326500114, + "grad_norm": 0.5364170074462891, + "learning_rate": 5.3188108219995424e-05, + "loss": 0.1586, + "step": 27948 + }, + { + "epoch": 0.6551685742248164, + "grad_norm": 0.5031477212905884, + "learning_rate": 5.3181599127613926e-05, + "loss": 0.0865, + "step": 27949 + }, + { + "epoch": 0.6551920157996214, + "grad_norm": 0.47716739773750305, + "learning_rate": 5.317509028927038e-05, + "loss": 0.0889, + "step": 27950 + }, + { + "epoch": 0.6552154573744264, + "grad_norm": 0.16826635599136353, + "learning_rate": 5.316858170500011e-05, + "loss": 0.0319, + "step": 27951 + }, + { + "epoch": 0.6552388989492314, + "grad_norm": 0.6212014555931091, + "learning_rate": 5.3162073374838395e-05, + "loss": 0.3242, + "step": 27952 + }, + { + "epoch": 0.6552623405240364, + "grad_norm": 0.7005054950714111, + "learning_rate": 5.3155565298820634e-05, + "loss": 0.7239, + "step": 27953 + }, + { + "epoch": 0.6552857820988414, + "grad_norm": 0.8503459692001343, + "learning_rate": 5.314905747698208e-05, + "loss": 0.1424, + "step": 27954 + }, + { + "epoch": 0.6553092236736464, + "grad_norm": 0.17879325151443481, + "learning_rate": 5.3142549909358034e-05, + "loss": 0.0298, + "step": 27955 + }, + { + "epoch": 0.6553326652484514, + "grad_norm": 0.2161979377269745, + "learning_rate": 5.313604259598385e-05, + "loss": 0.0442, + "step": 27956 + }, + { + "epoch": 0.6553561068232564, + "grad_norm": 0.3693031072616577, + "learning_rate": 5.3129535536894825e-05, + "loss": 0.0367, + "step": 27957 + }, + { + "epoch": 0.6553795483980613, + "grad_norm": 0.2545059323310852, + "learning_rate": 5.3123028732126256e-05, + "loss": 0.04, + "step": 27958 + }, + { + "epoch": 0.6554029899728664, + "grad_norm": 0.39327725768089294, + "learning_rate": 5.3116522181713415e-05, + "loss": 0.0766, + "step": 27959 + }, + { + "epoch": 0.6554264315476713, + "grad_norm": 0.521218478679657, + "learning_rate": 5.3110015885691686e-05, + "loss": 0.0901, + "step": 27960 + }, + { + "epoch": 0.6554498731224764, + "grad_norm": 0.3644111156463623, + "learning_rate": 5.310350984409632e-05, + "loss": 0.0577, + "step": 27961 + }, + { + "epoch": 0.6554733146972813, + "grad_norm": 0.259319931268692, + "learning_rate": 5.309700405696263e-05, + "loss": 0.0417, + "step": 27962 + }, + { + "epoch": 0.6554967562720864, + "grad_norm": 0.08086361736059189, + "learning_rate": 5.309049852432593e-05, + "loss": 0.0104, + "step": 27963 + }, + { + "epoch": 0.6555201978468913, + "grad_norm": 0.1462184637784958, + "learning_rate": 5.308399324622149e-05, + "loss": 0.0304, + "step": 27964 + }, + { + "epoch": 0.6555436394216964, + "grad_norm": 0.48072659969329834, + "learning_rate": 5.307748822268457e-05, + "loss": 0.0648, + "step": 27965 + }, + { + "epoch": 0.6555670809965013, + "grad_norm": 0.6572821736335754, + "learning_rate": 5.307098345375058e-05, + "loss": 0.1054, + "step": 27966 + }, + { + "epoch": 0.6555905225713063, + "grad_norm": 0.4777050316333771, + "learning_rate": 5.30644789394547e-05, + "loss": 0.0988, + "step": 27967 + }, + { + "epoch": 0.6556139641461113, + "grad_norm": 0.5873354077339172, + "learning_rate": 5.3057974679832324e-05, + "loss": 0.0976, + "step": 27968 + }, + { + "epoch": 0.6556374057209163, + "grad_norm": 0.4562912881374359, + "learning_rate": 5.305147067491869e-05, + "loss": 0.0799, + "step": 27969 + }, + { + "epoch": 0.6556608472957213, + "grad_norm": 0.5484372973442078, + "learning_rate": 5.3044966924749106e-05, + "loss": 0.0807, + "step": 27970 + }, + { + "epoch": 0.6556842888705263, + "grad_norm": 0.3413991928100586, + "learning_rate": 5.3038463429358856e-05, + "loss": 0.0648, + "step": 27971 + }, + { + "epoch": 0.6557077304453314, + "grad_norm": 0.3552965521812439, + "learning_rate": 5.3031960188783166e-05, + "loss": 0.0679, + "step": 27972 + }, + { + "epoch": 0.6557311720201363, + "grad_norm": 0.19160720705986023, + "learning_rate": 5.3025457203057424e-05, + "loss": 0.033, + "step": 27973 + }, + { + "epoch": 0.6557546135949414, + "grad_norm": 0.33493009209632874, + "learning_rate": 5.3018954472216886e-05, + "loss": 0.0826, + "step": 27974 + }, + { + "epoch": 0.6557780551697463, + "grad_norm": 0.47568097710609436, + "learning_rate": 5.301245199629681e-05, + "loss": 0.1325, + "step": 27975 + }, + { + "epoch": 0.6558014967445513, + "grad_norm": 0.5525836944580078, + "learning_rate": 5.30059497753325e-05, + "loss": 0.0983, + "step": 27976 + }, + { + "epoch": 0.6558249383193563, + "grad_norm": 0.7300435304641724, + "learning_rate": 5.2999447809359214e-05, + "loss": 0.0806, + "step": 27977 + }, + { + "epoch": 0.6558483798941613, + "grad_norm": 0.24480141699314117, + "learning_rate": 5.299294609841221e-05, + "loss": 0.0302, + "step": 27978 + }, + { + "epoch": 0.6558718214689663, + "grad_norm": 1.0097922086715698, + "learning_rate": 5.298644464252682e-05, + "loss": 0.086, + "step": 27979 + }, + { + "epoch": 0.6558952630437713, + "grad_norm": 0.7614080309867859, + "learning_rate": 5.297994344173832e-05, + "loss": 0.1963, + "step": 27980 + }, + { + "epoch": 0.6559187046185763, + "grad_norm": 0.6681447625160217, + "learning_rate": 5.297344249608198e-05, + "loss": 0.0993, + "step": 27981 + }, + { + "epoch": 0.6559421461933813, + "grad_norm": 0.6526390314102173, + "learning_rate": 5.296694180559306e-05, + "loss": 0.4424, + "step": 27982 + }, + { + "epoch": 0.6559655877681863, + "grad_norm": 0.40641647577285767, + "learning_rate": 5.296044137030685e-05, + "loss": 0.1215, + "step": 27983 + }, + { + "epoch": 0.6559890293429913, + "grad_norm": 0.9758399128913879, + "learning_rate": 5.295394119025855e-05, + "loss": 0.1379, + "step": 27984 + }, + { + "epoch": 0.6560124709177962, + "grad_norm": 0.843362033367157, + "learning_rate": 5.294744126548353e-05, + "loss": 0.1044, + "step": 27985 + }, + { + "epoch": 0.6560359124926013, + "grad_norm": 0.2997335195541382, + "learning_rate": 5.294094159601702e-05, + "loss": 0.0678, + "step": 27986 + }, + { + "epoch": 0.6560593540674062, + "grad_norm": 0.65966796875, + "learning_rate": 5.293444218189428e-05, + "loss": 0.1807, + "step": 27987 + }, + { + "epoch": 0.6560827956422113, + "grad_norm": 0.19521558284759521, + "learning_rate": 5.2927943023150586e-05, + "loss": 0.0347, + "step": 27988 + }, + { + "epoch": 0.6561062372170162, + "grad_norm": 0.4621328115463257, + "learning_rate": 5.2921444119821204e-05, + "loss": 0.0841, + "step": 27989 + }, + { + "epoch": 0.6561296787918213, + "grad_norm": 0.40875178575515747, + "learning_rate": 5.2914945471941334e-05, + "loss": 0.0647, + "step": 27990 + }, + { + "epoch": 0.6561531203666262, + "grad_norm": 0.32225605845451355, + "learning_rate": 5.29084470795463e-05, + "loss": 0.0374, + "step": 27991 + }, + { + "epoch": 0.6561765619414313, + "grad_norm": 0.6144921779632568, + "learning_rate": 5.29019489426714e-05, + "loss": 0.4165, + "step": 27992 + }, + { + "epoch": 0.6562000035162362, + "grad_norm": 0.39147844910621643, + "learning_rate": 5.289545106135185e-05, + "loss": 0.0667, + "step": 27993 + }, + { + "epoch": 0.6562234450910412, + "grad_norm": 0.07527860999107361, + "learning_rate": 5.28889534356229e-05, + "loss": 0.0141, + "step": 27994 + }, + { + "epoch": 0.6562468866658462, + "grad_norm": 0.526384711265564, + "learning_rate": 5.288245606551983e-05, + "loss": 0.4846, + "step": 27995 + }, + { + "epoch": 0.6562703282406512, + "grad_norm": 0.38064172863960266, + "learning_rate": 5.287595895107785e-05, + "loss": 0.0549, + "step": 27996 + }, + { + "epoch": 0.6562937698154562, + "grad_norm": 0.9965184926986694, + "learning_rate": 5.286946209233221e-05, + "loss": 0.0808, + "step": 27997 + }, + { + "epoch": 0.6563172113902612, + "grad_norm": 0.2886291742324829, + "learning_rate": 5.286296548931825e-05, + "loss": 0.0474, + "step": 27998 + }, + { + "epoch": 0.6563406529650662, + "grad_norm": 0.16718342900276184, + "learning_rate": 5.285646914207116e-05, + "loss": 0.0316, + "step": 27999 + }, + { + "epoch": 0.6563640945398712, + "grad_norm": 0.29729118943214417, + "learning_rate": 5.284997305062618e-05, + "loss": 0.0365, + "step": 28000 + }, + { + "epoch": 0.6563875361146762, + "grad_norm": 0.7840487957000732, + "learning_rate": 5.284347721501859e-05, + "loss": 0.1464, + "step": 28001 + }, + { + "epoch": 0.6564109776894812, + "grad_norm": 0.509428858757019, + "learning_rate": 5.283698163528361e-05, + "loss": 0.1052, + "step": 28002 + }, + { + "epoch": 0.6564344192642861, + "grad_norm": 0.11298901587724686, + "learning_rate": 5.283048631145645e-05, + "loss": 0.0256, + "step": 28003 + }, + { + "epoch": 0.6564578608390912, + "grad_norm": 0.45501279830932617, + "learning_rate": 5.2823991243572404e-05, + "loss": 0.0773, + "step": 28004 + }, + { + "epoch": 0.6564813024138961, + "grad_norm": 0.5112984776496887, + "learning_rate": 5.2817496431666736e-05, + "loss": 0.126, + "step": 28005 + }, + { + "epoch": 0.6565047439887012, + "grad_norm": 0.4558574855327606, + "learning_rate": 5.281100187577468e-05, + "loss": 0.6086, + "step": 28006 + }, + { + "epoch": 0.6565281855635061, + "grad_norm": 0.9751341938972473, + "learning_rate": 5.2804507575931424e-05, + "loss": 0.0296, + "step": 28007 + }, + { + "epoch": 0.6565516271383112, + "grad_norm": 0.8553405404090881, + "learning_rate": 5.279801353217226e-05, + "loss": 0.0844, + "step": 28008 + }, + { + "epoch": 0.6565750687131161, + "grad_norm": 0.26522865891456604, + "learning_rate": 5.27915197445324e-05, + "loss": 0.0513, + "step": 28009 + }, + { + "epoch": 0.6565985102879212, + "grad_norm": 0.11147301644086838, + "learning_rate": 5.278502621304704e-05, + "loss": 0.0211, + "step": 28010 + }, + { + "epoch": 0.6566219518627261, + "grad_norm": 0.5625455975532532, + "learning_rate": 5.27785329377515e-05, + "loss": 0.3233, + "step": 28011 + }, + { + "epoch": 0.6566453934375311, + "grad_norm": 0.5192399024963379, + "learning_rate": 5.277203991868096e-05, + "loss": 0.0552, + "step": 28012 + }, + { + "epoch": 0.6566688350123361, + "grad_norm": 0.916873574256897, + "learning_rate": 5.2765547155870654e-05, + "loss": 0.1773, + "step": 28013 + }, + { + "epoch": 0.6566922765871411, + "grad_norm": 0.5689818263053894, + "learning_rate": 5.2759054649355836e-05, + "loss": 0.0732, + "step": 28014 + }, + { + "epoch": 0.6567157181619461, + "grad_norm": 0.5082857012748718, + "learning_rate": 5.27525623991717e-05, + "loss": 0.1354, + "step": 28015 + }, + { + "epoch": 0.6567391597367511, + "grad_norm": 0.1090935543179512, + "learning_rate": 5.2746070405353456e-05, + "loss": 0.0088, + "step": 28016 + }, + { + "epoch": 0.6567626013115561, + "grad_norm": 0.3469180166721344, + "learning_rate": 5.2739578667936354e-05, + "loss": 0.0391, + "step": 28017 + }, + { + "epoch": 0.6567860428863611, + "grad_norm": 0.2864844799041748, + "learning_rate": 5.273308718695568e-05, + "loss": 0.0564, + "step": 28018 + }, + { + "epoch": 0.656809484461166, + "grad_norm": 0.16720111668109894, + "learning_rate": 5.27265959624466e-05, + "loss": 0.029, + "step": 28019 + }, + { + "epoch": 0.6568329260359711, + "grad_norm": 0.2857401967048645, + "learning_rate": 5.272010499444433e-05, + "loss": 0.0626, + "step": 28020 + }, + { + "epoch": 0.656856367610776, + "grad_norm": 0.8744239211082458, + "learning_rate": 5.27136142829841e-05, + "loss": 0.1063, + "step": 28021 + }, + { + "epoch": 0.6568798091855811, + "grad_norm": 0.6848298907279968, + "learning_rate": 5.27071238281011e-05, + "loss": 0.6589, + "step": 28022 + }, + { + "epoch": 0.6569032507603861, + "grad_norm": 0.5066357254981995, + "learning_rate": 5.270063362983061e-05, + "loss": 0.0479, + "step": 28023 + }, + { + "epoch": 0.6569266923351911, + "grad_norm": 0.7719139456748962, + "learning_rate": 5.269414368820782e-05, + "loss": 0.1273, + "step": 28024 + }, + { + "epoch": 0.6569501339099961, + "grad_norm": 0.1931040734052658, + "learning_rate": 5.2687654003267915e-05, + "loss": 0.0593, + "step": 28025 + }, + { + "epoch": 0.6569735754848011, + "grad_norm": 0.3684162497520447, + "learning_rate": 5.268116457504615e-05, + "loss": 0.0908, + "step": 28026 + }, + { + "epoch": 0.6569970170596061, + "grad_norm": 0.7366384267807007, + "learning_rate": 5.267467540357771e-05, + "loss": 0.0942, + "step": 28027 + }, + { + "epoch": 0.657020458634411, + "grad_norm": 0.44405749440193176, + "learning_rate": 5.2668186488897775e-05, + "loss": 0.048, + "step": 28028 + }, + { + "epoch": 0.6570439002092161, + "grad_norm": 0.251829594373703, + "learning_rate": 5.266169783104158e-05, + "loss": 0.0298, + "step": 28029 + }, + { + "epoch": 0.657067341784021, + "grad_norm": 0.19228218495845795, + "learning_rate": 5.2655209430044396e-05, + "loss": 0.0173, + "step": 28030 + }, + { + "epoch": 0.6570907833588261, + "grad_norm": 0.26008135080337524, + "learning_rate": 5.264872128594137e-05, + "loss": 0.0395, + "step": 28031 + }, + { + "epoch": 0.657114224933631, + "grad_norm": 0.09598146378993988, + "learning_rate": 5.264223339876773e-05, + "loss": 0.0189, + "step": 28032 + }, + { + "epoch": 0.6571376665084361, + "grad_norm": 0.49495914578437805, + "learning_rate": 5.263574576855863e-05, + "loss": 0.0706, + "step": 28033 + }, + { + "epoch": 0.657161108083241, + "grad_norm": 0.3100007176399231, + "learning_rate": 5.2629258395349334e-05, + "loss": 0.0479, + "step": 28034 + }, + { + "epoch": 0.6571845496580461, + "grad_norm": 0.3743472695350647, + "learning_rate": 5.2622771279174966e-05, + "loss": 0.0894, + "step": 28035 + }, + { + "epoch": 0.657207991232851, + "grad_norm": 0.41109150648117065, + "learning_rate": 5.261628442007081e-05, + "loss": 0.0617, + "step": 28036 + }, + { + "epoch": 0.6572314328076561, + "grad_norm": 0.17088782787322998, + "learning_rate": 5.260979781807203e-05, + "loss": 0.0227, + "step": 28037 + }, + { + "epoch": 0.657254874382461, + "grad_norm": 0.4037429392337799, + "learning_rate": 5.260331147321381e-05, + "loss": 0.0571, + "step": 28038 + }, + { + "epoch": 0.657278315957266, + "grad_norm": 0.31147852540016174, + "learning_rate": 5.259682538553137e-05, + "loss": 0.0309, + "step": 28039 + }, + { + "epoch": 0.657301757532071, + "grad_norm": 0.3005429804325104, + "learning_rate": 5.2590339555059833e-05, + "loss": 0.2191, + "step": 28040 + }, + { + "epoch": 0.657325199106876, + "grad_norm": 0.09863939881324768, + "learning_rate": 5.2583853981834495e-05, + "loss": 0.0137, + "step": 28041 + }, + { + "epoch": 0.657348640681681, + "grad_norm": 0.5075675845146179, + "learning_rate": 5.257736866589047e-05, + "loss": 0.695, + "step": 28042 + }, + { + "epoch": 0.657372082256486, + "grad_norm": 0.27452751994132996, + "learning_rate": 5.2570883607263e-05, + "loss": 0.0733, + "step": 28043 + }, + { + "epoch": 0.657395523831291, + "grad_norm": 0.6015178561210632, + "learning_rate": 5.2564398805987244e-05, + "loss": 0.6363, + "step": 28044 + }, + { + "epoch": 0.657418965406096, + "grad_norm": 0.3628285229206085, + "learning_rate": 5.255791426209842e-05, + "loss": 0.1024, + "step": 28045 + }, + { + "epoch": 0.657442406980901, + "grad_norm": 0.3399757742881775, + "learning_rate": 5.2551429975631663e-05, + "loss": 0.0584, + "step": 28046 + }, + { + "epoch": 0.657465848555706, + "grad_norm": 0.1388905644416809, + "learning_rate": 5.2544945946622184e-05, + "loss": 0.0254, + "step": 28047 + }, + { + "epoch": 0.6574892901305109, + "grad_norm": 0.17841076850891113, + "learning_rate": 5.253846217510513e-05, + "loss": 0.041, + "step": 28048 + }, + { + "epoch": 0.657512731705316, + "grad_norm": 0.344171404838562, + "learning_rate": 5.2531978661115745e-05, + "loss": 0.0763, + "step": 28049 + }, + { + "epoch": 0.6575361732801209, + "grad_norm": 0.2723619341850281, + "learning_rate": 5.2525495404689184e-05, + "loss": 0.0518, + "step": 28050 + }, + { + "epoch": 0.657559614854926, + "grad_norm": 0.3265049159526825, + "learning_rate": 5.251901240586064e-05, + "loss": 0.0535, + "step": 28051 + }, + { + "epoch": 0.6575830564297309, + "grad_norm": 0.3407440185546875, + "learning_rate": 5.25125296646652e-05, + "loss": 0.0629, + "step": 28052 + }, + { + "epoch": 0.657606498004536, + "grad_norm": 0.15714667737483978, + "learning_rate": 5.250604718113816e-05, + "loss": 0.0264, + "step": 28053 + }, + { + "epoch": 0.6576299395793409, + "grad_norm": 0.1607196182012558, + "learning_rate": 5.249956495531464e-05, + "loss": 0.0254, + "step": 28054 + }, + { + "epoch": 0.657653381154146, + "grad_norm": 0.33240053057670593, + "learning_rate": 5.249308298722978e-05, + "loss": 0.0634, + "step": 28055 + }, + { + "epoch": 0.6576768227289509, + "grad_norm": 0.17541174590587616, + "learning_rate": 5.2486601276918824e-05, + "loss": 0.0187, + "step": 28056 + }, + { + "epoch": 0.6577002643037559, + "grad_norm": 0.14522308111190796, + "learning_rate": 5.2480119824416915e-05, + "loss": 0.0247, + "step": 28057 + }, + { + "epoch": 0.6577237058785609, + "grad_norm": 0.28042498230934143, + "learning_rate": 5.24736386297592e-05, + "loss": 0.0645, + "step": 28058 + }, + { + "epoch": 0.6577471474533659, + "grad_norm": 0.3685391843318939, + "learning_rate": 5.246715769298087e-05, + "loss": 0.0488, + "step": 28059 + }, + { + "epoch": 0.6577705890281709, + "grad_norm": 0.12051037698984146, + "learning_rate": 5.246067701411703e-05, + "loss": 0.0238, + "step": 28060 + }, + { + "epoch": 0.6577940306029759, + "grad_norm": 0.828528881072998, + "learning_rate": 5.245419659320294e-05, + "loss": 0.1236, + "step": 28061 + }, + { + "epoch": 0.6578174721777809, + "grad_norm": 0.3682091236114502, + "learning_rate": 5.244771643027371e-05, + "loss": 0.0629, + "step": 28062 + }, + { + "epoch": 0.6578409137525859, + "grad_norm": 0.21491947770118713, + "learning_rate": 5.24412365253645e-05, + "loss": 0.0241, + "step": 28063 + }, + { + "epoch": 0.6578643553273908, + "grad_norm": 0.3184368312358856, + "learning_rate": 5.243475687851045e-05, + "loss": 0.0558, + "step": 28064 + }, + { + "epoch": 0.6578877969021959, + "grad_norm": 0.5136849880218506, + "learning_rate": 5.242827748974679e-05, + "loss": 0.0807, + "step": 28065 + }, + { + "epoch": 0.6579112384770008, + "grad_norm": 0.30715540051460266, + "learning_rate": 5.242179835910863e-05, + "loss": 0.0387, + "step": 28066 + }, + { + "epoch": 0.6579346800518059, + "grad_norm": 0.2021949589252472, + "learning_rate": 5.241531948663108e-05, + "loss": 0.0451, + "step": 28067 + }, + { + "epoch": 0.6579581216266108, + "grad_norm": 0.1583838015794754, + "learning_rate": 5.2408840872349395e-05, + "loss": 0.0267, + "step": 28068 + }, + { + "epoch": 0.6579815632014159, + "grad_norm": 0.16808651387691498, + "learning_rate": 5.2402362516298675e-05, + "loss": 0.0249, + "step": 28069 + }, + { + "epoch": 0.6580050047762208, + "grad_norm": 0.17039431631565094, + "learning_rate": 5.2395884418514074e-05, + "loss": 0.0181, + "step": 28070 + }, + { + "epoch": 0.6580284463510259, + "grad_norm": 0.5078511834144592, + "learning_rate": 5.238940657903074e-05, + "loss": 0.0744, + "step": 28071 + }, + { + "epoch": 0.6580518879258308, + "grad_norm": 0.49338915944099426, + "learning_rate": 5.238292899788383e-05, + "loss": 0.1073, + "step": 28072 + }, + { + "epoch": 0.6580753295006359, + "grad_norm": 0.4575555622577667, + "learning_rate": 5.237645167510845e-05, + "loss": 0.059, + "step": 28073 + }, + { + "epoch": 0.6580987710754409, + "grad_norm": 0.48093563318252563, + "learning_rate": 5.23699746107398e-05, + "loss": 0.0555, + "step": 28074 + }, + { + "epoch": 0.6581222126502458, + "grad_norm": 0.3699169158935547, + "learning_rate": 5.236349780481303e-05, + "loss": 0.0463, + "step": 28075 + }, + { + "epoch": 0.6581456542250509, + "grad_norm": 0.1760348379611969, + "learning_rate": 5.2357021257363213e-05, + "loss": 0.0155, + "step": 28076 + }, + { + "epoch": 0.6581690957998558, + "grad_norm": 0.2753232717514038, + "learning_rate": 5.2350544968425576e-05, + "loss": 0.0321, + "step": 28077 + }, + { + "epoch": 0.6581925373746609, + "grad_norm": 0.6445305347442627, + "learning_rate": 5.234406893803523e-05, + "loss": 0.6868, + "step": 28078 + }, + { + "epoch": 0.6582159789494658, + "grad_norm": 0.4176555275917053, + "learning_rate": 5.233759316622731e-05, + "loss": 0.0475, + "step": 28079 + }, + { + "epoch": 0.6582394205242709, + "grad_norm": 0.40057167410850525, + "learning_rate": 5.233111765303689e-05, + "loss": 0.026, + "step": 28080 + }, + { + "epoch": 0.6582628620990758, + "grad_norm": 0.708361804485321, + "learning_rate": 5.232464239849921e-05, + "loss": 0.1746, + "step": 28081 + }, + { + "epoch": 0.6582863036738809, + "grad_norm": 0.15692144632339478, + "learning_rate": 5.231816740264936e-05, + "loss": 0.0243, + "step": 28082 + }, + { + "epoch": 0.6583097452486858, + "grad_norm": 0.5600082278251648, + "learning_rate": 5.2311692665522495e-05, + "loss": 0.0943, + "step": 28083 + }, + { + "epoch": 0.6583331868234908, + "grad_norm": 0.41355881094932556, + "learning_rate": 5.23052181871537e-05, + "loss": 0.0625, + "step": 28084 + }, + { + "epoch": 0.6583566283982958, + "grad_norm": 0.4580429494380951, + "learning_rate": 5.229874396757815e-05, + "loss": 0.682, + "step": 28085 + }, + { + "epoch": 0.6583800699731008, + "grad_norm": 0.6989641189575195, + "learning_rate": 5.229227000683091e-05, + "loss": 0.1386, + "step": 28086 + }, + { + "epoch": 0.6584035115479058, + "grad_norm": 0.5257444381713867, + "learning_rate": 5.22857963049472e-05, + "loss": 0.0486, + "step": 28087 + }, + { + "epoch": 0.6584269531227108, + "grad_norm": 0.4186246693134308, + "learning_rate": 5.2279322861962085e-05, + "loss": 0.0943, + "step": 28088 + }, + { + "epoch": 0.6584503946975158, + "grad_norm": 0.08421328663825989, + "learning_rate": 5.227284967791068e-05, + "loss": 0.0126, + "step": 28089 + }, + { + "epoch": 0.6584738362723208, + "grad_norm": 0.21793833374977112, + "learning_rate": 5.226637675282817e-05, + "loss": 0.0362, + "step": 28090 + }, + { + "epoch": 0.6584972778471258, + "grad_norm": 0.4919894337654114, + "learning_rate": 5.2259904086749654e-05, + "loss": 0.0791, + "step": 28091 + }, + { + "epoch": 0.6585207194219308, + "grad_norm": 0.5892893075942993, + "learning_rate": 5.225343167971023e-05, + "loss": 0.1009, + "step": 28092 + }, + { + "epoch": 0.6585441609967357, + "grad_norm": 0.3309031128883362, + "learning_rate": 5.224695953174499e-05, + "loss": 0.0585, + "step": 28093 + }, + { + "epoch": 0.6585676025715408, + "grad_norm": 0.5960186719894409, + "learning_rate": 5.224048764288914e-05, + "loss": 0.1282, + "step": 28094 + }, + { + "epoch": 0.6585910441463457, + "grad_norm": 0.5890004634857178, + "learning_rate": 5.2234016013177724e-05, + "loss": 0.1138, + "step": 28095 + }, + { + "epoch": 0.6586144857211508, + "grad_norm": 0.48815563321113586, + "learning_rate": 5.22275446426459e-05, + "loss": 0.4315, + "step": 28096 + }, + { + "epoch": 0.6586379272959557, + "grad_norm": 0.2854706645011902, + "learning_rate": 5.2221073531328745e-05, + "loss": 0.0222, + "step": 28097 + }, + { + "epoch": 0.6586613688707608, + "grad_norm": 0.1043362244963646, + "learning_rate": 5.221460267926136e-05, + "loss": 0.0215, + "step": 28098 + }, + { + "epoch": 0.6586848104455657, + "grad_norm": 0.3968145251274109, + "learning_rate": 5.220813208647892e-05, + "loss": 0.5281, + "step": 28099 + }, + { + "epoch": 0.6587082520203708, + "grad_norm": 0.4760240316390991, + "learning_rate": 5.2201661753016504e-05, + "loss": 0.5704, + "step": 28100 + }, + { + "epoch": 0.6587316935951757, + "grad_norm": 0.3547535836696625, + "learning_rate": 5.219519167890916e-05, + "loss": 0.0735, + "step": 28101 + }, + { + "epoch": 0.6587551351699807, + "grad_norm": 0.4269246459007263, + "learning_rate": 5.2188721864192115e-05, + "loss": 0.2023, + "step": 28102 + }, + { + "epoch": 0.6587785767447857, + "grad_norm": 0.5354636907577515, + "learning_rate": 5.218225230890039e-05, + "loss": 0.0969, + "step": 28103 + }, + { + "epoch": 0.6588020183195907, + "grad_norm": 0.2794226408004761, + "learning_rate": 5.217578301306911e-05, + "loss": 0.028, + "step": 28104 + }, + { + "epoch": 0.6588254598943957, + "grad_norm": 0.22913707792758942, + "learning_rate": 5.216931397673335e-05, + "loss": 0.0223, + "step": 28105 + }, + { + "epoch": 0.6588489014692007, + "grad_norm": 0.2006826102733612, + "learning_rate": 5.216284519992827e-05, + "loss": 0.0209, + "step": 28106 + }, + { + "epoch": 0.6588723430440057, + "grad_norm": 0.5838109850883484, + "learning_rate": 5.215637668268893e-05, + "loss": 0.1645, + "step": 28107 + }, + { + "epoch": 0.6588957846188107, + "grad_norm": 0.19137927889823914, + "learning_rate": 5.214990842505044e-05, + "loss": 0.018, + "step": 28108 + }, + { + "epoch": 0.6589192261936156, + "grad_norm": 0.5868409872055054, + "learning_rate": 5.21434404270479e-05, + "loss": 0.0894, + "step": 28109 + }, + { + "epoch": 0.6589426677684207, + "grad_norm": 0.5246611833572388, + "learning_rate": 5.213697268871639e-05, + "loss": 0.1456, + "step": 28110 + }, + { + "epoch": 0.6589661093432256, + "grad_norm": 0.3629617393016815, + "learning_rate": 5.2130505210090977e-05, + "loss": 0.0713, + "step": 28111 + }, + { + "epoch": 0.6589895509180307, + "grad_norm": 0.18576428294181824, + "learning_rate": 5.212403799120682e-05, + "loss": 0.029, + "step": 28112 + }, + { + "epoch": 0.6590129924928356, + "grad_norm": 0.3147106468677521, + "learning_rate": 5.211757103209894e-05, + "loss": 0.4485, + "step": 28113 + }, + { + "epoch": 0.6590364340676407, + "grad_norm": 0.2105167657136917, + "learning_rate": 5.2111104332802505e-05, + "loss": 0.046, + "step": 28114 + }, + { + "epoch": 0.6590598756424456, + "grad_norm": 0.14183415472507477, + "learning_rate": 5.210463789335258e-05, + "loss": 0.0176, + "step": 28115 + }, + { + "epoch": 0.6590833172172507, + "grad_norm": 0.3884023129940033, + "learning_rate": 5.209817171378422e-05, + "loss": 0.0567, + "step": 28116 + }, + { + "epoch": 0.6591067587920556, + "grad_norm": 0.3106667399406433, + "learning_rate": 5.2091705794132536e-05, + "loss": 0.0628, + "step": 28117 + }, + { + "epoch": 0.6591302003668607, + "grad_norm": 0.42212215065956116, + "learning_rate": 5.208524013443257e-05, + "loss": 0.1126, + "step": 28118 + }, + { + "epoch": 0.6591536419416656, + "grad_norm": 0.5118401050567627, + "learning_rate": 5.207877473471946e-05, + "loss": 0.1063, + "step": 28119 + }, + { + "epoch": 0.6591770835164706, + "grad_norm": 0.5355120301246643, + "learning_rate": 5.207230959502827e-05, + "loss": 0.1077, + "step": 28120 + }, + { + "epoch": 0.6592005250912756, + "grad_norm": 0.15555696189403534, + "learning_rate": 5.2065844715394086e-05, + "loss": 0.0311, + "step": 28121 + }, + { + "epoch": 0.6592239666660806, + "grad_norm": 0.1227099746465683, + "learning_rate": 5.205938009585196e-05, + "loss": 0.0187, + "step": 28122 + }, + { + "epoch": 0.6592474082408856, + "grad_norm": 0.05286221578717232, + "learning_rate": 5.2052915736437e-05, + "loss": 0.0064, + "step": 28123 + }, + { + "epoch": 0.6592708498156906, + "grad_norm": 0.12100909650325775, + "learning_rate": 5.2046451637184224e-05, + "loss": 0.0136, + "step": 28124 + }, + { + "epoch": 0.6592942913904957, + "grad_norm": 0.2451726496219635, + "learning_rate": 5.2039987798128756e-05, + "loss": 0.0619, + "step": 28125 + }, + { + "epoch": 0.6593177329653006, + "grad_norm": 0.5499212741851807, + "learning_rate": 5.2033524219305695e-05, + "loss": 0.6395, + "step": 28126 + }, + { + "epoch": 0.6593411745401057, + "grad_norm": 0.6251830458641052, + "learning_rate": 5.202706090075008e-05, + "loss": 0.4998, + "step": 28127 + }, + { + "epoch": 0.6593646161149106, + "grad_norm": 0.36446064710617065, + "learning_rate": 5.202059784249698e-05, + "loss": 0.0325, + "step": 28128 + }, + { + "epoch": 0.6593880576897156, + "grad_norm": 0.3688691258430481, + "learning_rate": 5.201413504458147e-05, + "loss": 0.0674, + "step": 28129 + }, + { + "epoch": 0.6594114992645206, + "grad_norm": 0.6966937780380249, + "learning_rate": 5.200767250703861e-05, + "loss": 0.1275, + "step": 28130 + }, + { + "epoch": 0.6594349408393256, + "grad_norm": 0.18344008922576904, + "learning_rate": 5.200121022990343e-05, + "loss": 0.0482, + "step": 28131 + }, + { + "epoch": 0.6594583824141306, + "grad_norm": 0.44399869441986084, + "learning_rate": 5.1994748213211064e-05, + "loss": 0.1139, + "step": 28132 + }, + { + "epoch": 0.6594818239889356, + "grad_norm": 0.08366521447896957, + "learning_rate": 5.198828645699656e-05, + "loss": 0.0046, + "step": 28133 + }, + { + "epoch": 0.6595052655637406, + "grad_norm": 0.7431221604347229, + "learning_rate": 5.198182496129496e-05, + "loss": 0.1776, + "step": 28134 + }, + { + "epoch": 0.6595287071385456, + "grad_norm": 0.47486546635627747, + "learning_rate": 5.197536372614131e-05, + "loss": 0.5021, + "step": 28135 + }, + { + "epoch": 0.6595521487133506, + "grad_norm": 0.3093183934688568, + "learning_rate": 5.196890275157067e-05, + "loss": 0.0582, + "step": 28136 + }, + { + "epoch": 0.6595755902881556, + "grad_norm": 0.43905124068260193, + "learning_rate": 5.196244203761814e-05, + "loss": 0.0968, + "step": 28137 + }, + { + "epoch": 0.6595990318629605, + "grad_norm": 0.11855196207761765, + "learning_rate": 5.1955981584318716e-05, + "loss": 0.0201, + "step": 28138 + }, + { + "epoch": 0.6596224734377656, + "grad_norm": 0.23546330630779266, + "learning_rate": 5.1949521391707526e-05, + "loss": 0.0289, + "step": 28139 + }, + { + "epoch": 0.6596459150125705, + "grad_norm": 0.44894397258758545, + "learning_rate": 5.1943061459819575e-05, + "loss": 0.0209, + "step": 28140 + }, + { + "epoch": 0.6596693565873756, + "grad_norm": 0.17080116271972656, + "learning_rate": 5.193660178868993e-05, + "loss": 0.0335, + "step": 28141 + }, + { + "epoch": 0.6596927981621805, + "grad_norm": 0.876031756401062, + "learning_rate": 5.193014237835364e-05, + "loss": 0.1353, + "step": 28142 + }, + { + "epoch": 0.6597162397369856, + "grad_norm": 0.3360932469367981, + "learning_rate": 5.192368322884571e-05, + "loss": 0.0974, + "step": 28143 + }, + { + "epoch": 0.6597396813117905, + "grad_norm": 0.41032055020332336, + "learning_rate": 5.191722434020125e-05, + "loss": 0.0605, + "step": 28144 + }, + { + "epoch": 0.6597631228865956, + "grad_norm": 0.40003055334091187, + "learning_rate": 5.19107657124553e-05, + "loss": 0.0912, + "step": 28145 + }, + { + "epoch": 0.6597865644614005, + "grad_norm": 0.38091495633125305, + "learning_rate": 5.190430734564289e-05, + "loss": 0.0373, + "step": 28146 + }, + { + "epoch": 0.6598100060362055, + "grad_norm": 0.4442804455757141, + "learning_rate": 5.1897849239799056e-05, + "loss": 0.0608, + "step": 28147 + }, + { + "epoch": 0.6598334476110105, + "grad_norm": 0.1323106735944748, + "learning_rate": 5.1891391394958844e-05, + "loss": 0.0406, + "step": 28148 + }, + { + "epoch": 0.6598568891858155, + "grad_norm": 0.5375056862831116, + "learning_rate": 5.1884933811157245e-05, + "loss": 0.6221, + "step": 28149 + }, + { + "epoch": 0.6598803307606205, + "grad_norm": 0.3868650794029236, + "learning_rate": 5.187847648842936e-05, + "loss": 0.1043, + "step": 28150 + }, + { + "epoch": 0.6599037723354255, + "grad_norm": 0.37457624077796936, + "learning_rate": 5.1872019426810245e-05, + "loss": 0.099, + "step": 28151 + }, + { + "epoch": 0.6599272139102305, + "grad_norm": 0.4285475015640259, + "learning_rate": 5.186556262633492e-05, + "loss": 0.0948, + "step": 28152 + }, + { + "epoch": 0.6599506554850355, + "grad_norm": 0.4721386134624481, + "learning_rate": 5.185910608703839e-05, + "loss": 0.0808, + "step": 28153 + }, + { + "epoch": 0.6599740970598404, + "grad_norm": 0.4898693561553955, + "learning_rate": 5.185264980895571e-05, + "loss": 0.0828, + "step": 28154 + }, + { + "epoch": 0.6599975386346455, + "grad_norm": 0.15321913361549377, + "learning_rate": 5.184619379212191e-05, + "loss": 0.0191, + "step": 28155 + }, + { + "epoch": 0.6600209802094504, + "grad_norm": 0.06461233645677567, + "learning_rate": 5.183973803657197e-05, + "loss": 0.0115, + "step": 28156 + }, + { + "epoch": 0.6600444217842555, + "grad_norm": 0.5085901618003845, + "learning_rate": 5.183328254234101e-05, + "loss": 0.065, + "step": 28157 + }, + { + "epoch": 0.6600678633590604, + "grad_norm": 0.49196377396583557, + "learning_rate": 5.182682730946402e-05, + "loss": 0.0952, + "step": 28158 + }, + { + "epoch": 0.6600913049338655, + "grad_norm": 0.34492209553718567, + "learning_rate": 5.182037233797601e-05, + "loss": 0.0343, + "step": 28159 + }, + { + "epoch": 0.6601147465086704, + "grad_norm": 0.27213093638420105, + "learning_rate": 5.181391762791201e-05, + "loss": 0.0537, + "step": 28160 + }, + { + "epoch": 0.6601381880834755, + "grad_norm": 0.16500073671340942, + "learning_rate": 5.180746317930705e-05, + "loss": 0.0321, + "step": 28161 + }, + { + "epoch": 0.6601616296582804, + "grad_norm": 0.3279186189174652, + "learning_rate": 5.180100899219612e-05, + "loss": 0.0485, + "step": 28162 + }, + { + "epoch": 0.6601850712330855, + "grad_norm": 0.43769070506095886, + "learning_rate": 5.179455506661426e-05, + "loss": 0.0755, + "step": 28163 + }, + { + "epoch": 0.6602085128078904, + "grad_norm": 0.37669843435287476, + "learning_rate": 5.1788101402596544e-05, + "loss": 0.0326, + "step": 28164 + }, + { + "epoch": 0.6602319543826954, + "grad_norm": 0.38810044527053833, + "learning_rate": 5.178164800017795e-05, + "loss": 0.0739, + "step": 28165 + }, + { + "epoch": 0.6602553959575004, + "grad_norm": 0.48645928502082825, + "learning_rate": 5.1775194859393484e-05, + "loss": 0.0631, + "step": 28166 + }, + { + "epoch": 0.6602788375323054, + "grad_norm": 0.659227728843689, + "learning_rate": 5.1768741980278165e-05, + "loss": 0.0853, + "step": 28167 + }, + { + "epoch": 0.6603022791071104, + "grad_norm": 0.22334788739681244, + "learning_rate": 5.176228936286701e-05, + "loss": 0.0441, + "step": 28168 + }, + { + "epoch": 0.6603257206819154, + "grad_norm": 0.1041496992111206, + "learning_rate": 5.1755837007194994e-05, + "loss": 0.0127, + "step": 28169 + }, + { + "epoch": 0.6603491622567204, + "grad_norm": 0.11070197820663452, + "learning_rate": 5.1749384913297194e-05, + "loss": 0.0148, + "step": 28170 + }, + { + "epoch": 0.6603726038315254, + "grad_norm": 0.29126349091529846, + "learning_rate": 5.174293308120859e-05, + "loss": 0.0387, + "step": 28171 + }, + { + "epoch": 0.6603960454063303, + "grad_norm": 0.7481489181518555, + "learning_rate": 5.17364815109642e-05, + "loss": 0.128, + "step": 28172 + }, + { + "epoch": 0.6604194869811354, + "grad_norm": 0.9282432794570923, + "learning_rate": 5.173003020259901e-05, + "loss": 0.1047, + "step": 28173 + }, + { + "epoch": 0.6604429285559403, + "grad_norm": 0.4155847430229187, + "learning_rate": 5.1723579156147986e-05, + "loss": 0.1161, + "step": 28174 + }, + { + "epoch": 0.6604663701307454, + "grad_norm": 0.2180367112159729, + "learning_rate": 5.1717128371646226e-05, + "loss": 0.0643, + "step": 28175 + }, + { + "epoch": 0.6604898117055504, + "grad_norm": 0.4922817349433899, + "learning_rate": 5.171067784912864e-05, + "loss": 0.0767, + "step": 28176 + }, + { + "epoch": 0.6605132532803554, + "grad_norm": 0.17154309153556824, + "learning_rate": 5.1704227588630326e-05, + "loss": 0.0097, + "step": 28177 + }, + { + "epoch": 0.6605366948551604, + "grad_norm": 0.1290959268808365, + "learning_rate": 5.1697777590186214e-05, + "loss": 0.0144, + "step": 28178 + }, + { + "epoch": 0.6605601364299654, + "grad_norm": 0.26802143454551697, + "learning_rate": 5.169132785383133e-05, + "loss": 0.0156, + "step": 28179 + }, + { + "epoch": 0.6605835780047704, + "grad_norm": 0.678303062915802, + "learning_rate": 5.1684878379600665e-05, + "loss": 0.0856, + "step": 28180 + }, + { + "epoch": 0.6606070195795753, + "grad_norm": 0.5337589383125305, + "learning_rate": 5.167842916752916e-05, + "loss": 0.0359, + "step": 28181 + }, + { + "epoch": 0.6606304611543804, + "grad_norm": 0.25837430357933044, + "learning_rate": 5.167198021765189e-05, + "loss": 0.0464, + "step": 28182 + }, + { + "epoch": 0.6606539027291853, + "grad_norm": 0.1620989739894867, + "learning_rate": 5.166553153000383e-05, + "loss": 0.0273, + "step": 28183 + }, + { + "epoch": 0.6606773443039904, + "grad_norm": 0.7911409735679626, + "learning_rate": 5.1659083104619955e-05, + "loss": 0.1173, + "step": 28184 + }, + { + "epoch": 0.6607007858787953, + "grad_norm": 0.27627333998680115, + "learning_rate": 5.165263494153524e-05, + "loss": 0.0467, + "step": 28185 + }, + { + "epoch": 0.6607242274536004, + "grad_norm": 0.503750741481781, + "learning_rate": 5.1646187040784655e-05, + "loss": 0.0711, + "step": 28186 + }, + { + "epoch": 0.6607476690284053, + "grad_norm": 0.6003005504608154, + "learning_rate": 5.1639739402403255e-05, + "loss": 0.1035, + "step": 28187 + }, + { + "epoch": 0.6607711106032104, + "grad_norm": 0.5088851451873779, + "learning_rate": 5.163329202642595e-05, + "loss": 0.1012, + "step": 28188 + }, + { + "epoch": 0.6607945521780153, + "grad_norm": 0.2469402700662613, + "learning_rate": 5.1626844912887806e-05, + "loss": 0.0446, + "step": 28189 + }, + { + "epoch": 0.6608179937528204, + "grad_norm": 0.3808235824108124, + "learning_rate": 5.162039806182376e-05, + "loss": 0.0893, + "step": 28190 + }, + { + "epoch": 0.6608414353276253, + "grad_norm": 0.339932382106781, + "learning_rate": 5.1613951473268794e-05, + "loss": 0.3108, + "step": 28191 + }, + { + "epoch": 0.6608648769024303, + "grad_norm": 0.15789803862571716, + "learning_rate": 5.160750514725788e-05, + "loss": 0.0336, + "step": 28192 + }, + { + "epoch": 0.6608883184772353, + "grad_norm": 0.39362719655036926, + "learning_rate": 5.160105908382602e-05, + "loss": 0.0927, + "step": 28193 + }, + { + "epoch": 0.6609117600520403, + "grad_norm": 0.26237982511520386, + "learning_rate": 5.159461328300812e-05, + "loss": 0.0464, + "step": 28194 + }, + { + "epoch": 0.6609352016268453, + "grad_norm": 0.4740389585494995, + "learning_rate": 5.1588167744839254e-05, + "loss": 0.0896, + "step": 28195 + }, + { + "epoch": 0.6609586432016503, + "grad_norm": 0.2642329931259155, + "learning_rate": 5.1581722469354354e-05, + "loss": 0.0417, + "step": 28196 + }, + { + "epoch": 0.6609820847764553, + "grad_norm": 0.1660643219947815, + "learning_rate": 5.157527745658839e-05, + "loss": 0.0242, + "step": 28197 + }, + { + "epoch": 0.6610055263512603, + "grad_norm": 0.692538857460022, + "learning_rate": 5.1568832706576285e-05, + "loss": 0.0953, + "step": 28198 + }, + { + "epoch": 0.6610289679260652, + "grad_norm": 0.3844474256038666, + "learning_rate": 5.156238821935311e-05, + "loss": 0.0436, + "step": 28199 + }, + { + "epoch": 0.6610524095008703, + "grad_norm": 0.17852722108364105, + "learning_rate": 5.1555943994953756e-05, + "loss": 0.0263, + "step": 28200 + }, + { + "epoch": 0.6610758510756752, + "grad_norm": 0.19535090029239655, + "learning_rate": 5.154950003341319e-05, + "loss": 0.0268, + "step": 28201 + }, + { + "epoch": 0.6610992926504803, + "grad_norm": 0.43360430002212524, + "learning_rate": 5.154305633476644e-05, + "loss": 0.0866, + "step": 28202 + }, + { + "epoch": 0.6611227342252852, + "grad_norm": 0.36504706740379333, + "learning_rate": 5.153661289904842e-05, + "loss": 0.0799, + "step": 28203 + }, + { + "epoch": 0.6611461758000903, + "grad_norm": 0.23190142214298248, + "learning_rate": 5.15301697262941e-05, + "loss": 0.0337, + "step": 28204 + }, + { + "epoch": 0.6611696173748952, + "grad_norm": 0.6773149371147156, + "learning_rate": 5.1523726816538454e-05, + "loss": 0.1306, + "step": 28205 + }, + { + "epoch": 0.6611930589497003, + "grad_norm": 0.3256118595600128, + "learning_rate": 5.15172841698164e-05, + "loss": 0.0642, + "step": 28206 + }, + { + "epoch": 0.6612165005245052, + "grad_norm": 0.4709304869174957, + "learning_rate": 5.1510841786162924e-05, + "loss": 0.0926, + "step": 28207 + }, + { + "epoch": 0.6612399420993103, + "grad_norm": 0.36446481943130493, + "learning_rate": 5.1504399665613e-05, + "loss": 0.0482, + "step": 28208 + }, + { + "epoch": 0.6612633836741152, + "grad_norm": 0.23163726925849915, + "learning_rate": 5.149795780820157e-05, + "loss": 0.0435, + "step": 28209 + }, + { + "epoch": 0.6612868252489202, + "grad_norm": 0.7705327868461609, + "learning_rate": 5.1491516213963555e-05, + "loss": 0.1406, + "step": 28210 + }, + { + "epoch": 0.6613102668237252, + "grad_norm": 0.49254554510116577, + "learning_rate": 5.148507488293397e-05, + "loss": 0.0453, + "step": 28211 + }, + { + "epoch": 0.6613337083985302, + "grad_norm": 0.2841167449951172, + "learning_rate": 5.147863381514772e-05, + "loss": 0.0352, + "step": 28212 + }, + { + "epoch": 0.6613571499733352, + "grad_norm": 0.5613177418708801, + "learning_rate": 5.147219301063979e-05, + "loss": 0.1485, + "step": 28213 + }, + { + "epoch": 0.6613805915481402, + "grad_norm": 0.5653193593025208, + "learning_rate": 5.146575246944505e-05, + "loss": 0.0843, + "step": 28214 + }, + { + "epoch": 0.6614040331229452, + "grad_norm": 0.32428231835365295, + "learning_rate": 5.1459312191598544e-05, + "loss": 0.0278, + "step": 28215 + }, + { + "epoch": 0.6614274746977502, + "grad_norm": 0.31350889801979065, + "learning_rate": 5.145287217713518e-05, + "loss": 0.0518, + "step": 28216 + }, + { + "epoch": 0.6614509162725551, + "grad_norm": 0.517706573009491, + "learning_rate": 5.14464324260899e-05, + "loss": 0.1174, + "step": 28217 + }, + { + "epoch": 0.6614743578473602, + "grad_norm": 0.5248987674713135, + "learning_rate": 5.143999293849764e-05, + "loss": 0.0197, + "step": 28218 + }, + { + "epoch": 0.6614977994221651, + "grad_norm": 0.1547640711069107, + "learning_rate": 5.143355371439329e-05, + "loss": 0.0108, + "step": 28219 + }, + { + "epoch": 0.6615212409969702, + "grad_norm": 0.3622191846370697, + "learning_rate": 5.14271147538119e-05, + "loss": 0.0839, + "step": 28220 + }, + { + "epoch": 0.6615446825717751, + "grad_norm": 0.2446066439151764, + "learning_rate": 5.142067605678834e-05, + "loss": 0.0276, + "step": 28221 + }, + { + "epoch": 0.6615681241465802, + "grad_norm": 0.7390241026878357, + "learning_rate": 5.1414237623357534e-05, + "loss": 0.0852, + "step": 28222 + }, + { + "epoch": 0.6615915657213851, + "grad_norm": 0.6333001255989075, + "learning_rate": 5.140779945355446e-05, + "loss": 0.1563, + "step": 28223 + }, + { + "epoch": 0.6616150072961902, + "grad_norm": 0.6206126809120178, + "learning_rate": 5.140136154741404e-05, + "loss": 0.0639, + "step": 28224 + }, + { + "epoch": 0.6616384488709951, + "grad_norm": 0.4641294479370117, + "learning_rate": 5.13949239049712e-05, + "loss": 0.0971, + "step": 28225 + }, + { + "epoch": 0.6616618904458001, + "grad_norm": 0.3474828004837036, + "learning_rate": 5.1388486526260824e-05, + "loss": 0.0466, + "step": 28226 + }, + { + "epoch": 0.6616853320206052, + "grad_norm": 0.3471466898918152, + "learning_rate": 5.138204941131794e-05, + "loss": 0.0506, + "step": 28227 + }, + { + "epoch": 0.6617087735954101, + "grad_norm": 0.39513716101646423, + "learning_rate": 5.137561256017742e-05, + "loss": 0.0439, + "step": 28228 + }, + { + "epoch": 0.6617322151702152, + "grad_norm": 0.36138343811035156, + "learning_rate": 5.136917597287419e-05, + "loss": 0.0631, + "step": 28229 + }, + { + "epoch": 0.6617556567450201, + "grad_norm": 0.5033351182937622, + "learning_rate": 5.136273964944317e-05, + "loss": 0.4462, + "step": 28230 + }, + { + "epoch": 0.6617790983198252, + "grad_norm": 0.3690777122974396, + "learning_rate": 5.13563035899193e-05, + "loss": 0.0613, + "step": 28231 + }, + { + "epoch": 0.6618025398946301, + "grad_norm": 0.39262014627456665, + "learning_rate": 5.134986779433746e-05, + "loss": 0.1356, + "step": 28232 + }, + { + "epoch": 0.6618259814694352, + "grad_norm": 0.09901127219200134, + "learning_rate": 5.134343226273264e-05, + "loss": 0.0116, + "step": 28233 + }, + { + "epoch": 0.6618494230442401, + "grad_norm": 0.3259001076221466, + "learning_rate": 5.1336996995139676e-05, + "loss": 0.0565, + "step": 28234 + }, + { + "epoch": 0.6618728646190452, + "grad_norm": 0.13436587154865265, + "learning_rate": 5.1330561991593586e-05, + "loss": 0.0316, + "step": 28235 + }, + { + "epoch": 0.6618963061938501, + "grad_norm": 0.35831135511398315, + "learning_rate": 5.132412725212923e-05, + "loss": 0.0775, + "step": 28236 + }, + { + "epoch": 0.6619197477686551, + "grad_norm": 0.39546215534210205, + "learning_rate": 5.1317692776781535e-05, + "loss": 0.042, + "step": 28237 + }, + { + "epoch": 0.6619431893434601, + "grad_norm": 0.4194331467151642, + "learning_rate": 5.131125856558541e-05, + "loss": 0.0858, + "step": 28238 + }, + { + "epoch": 0.6619666309182651, + "grad_norm": 0.3765047788619995, + "learning_rate": 5.130482461857571e-05, + "loss": 0.0799, + "step": 28239 + }, + { + "epoch": 0.6619900724930701, + "grad_norm": 0.33359307050704956, + "learning_rate": 5.1298390935787454e-05, + "loss": 0.0843, + "step": 28240 + }, + { + "epoch": 0.6620135140678751, + "grad_norm": 0.537660539150238, + "learning_rate": 5.129195751725549e-05, + "loss": 0.655, + "step": 28241 + }, + { + "epoch": 0.6620369556426801, + "grad_norm": 0.3027273714542389, + "learning_rate": 5.1285524363014734e-05, + "loss": 0.0252, + "step": 28242 + }, + { + "epoch": 0.6620603972174851, + "grad_norm": 0.3504241704940796, + "learning_rate": 5.127909147310009e-05, + "loss": 0.0767, + "step": 28243 + }, + { + "epoch": 0.66208383879229, + "grad_norm": 0.396022230386734, + "learning_rate": 5.127265884754647e-05, + "loss": 0.0664, + "step": 28244 + }, + { + "epoch": 0.6621072803670951, + "grad_norm": 0.609973669052124, + "learning_rate": 5.1266226486388746e-05, + "loss": 0.1526, + "step": 28245 + }, + { + "epoch": 0.6621307219419, + "grad_norm": 0.5603159666061401, + "learning_rate": 5.1259794389661865e-05, + "loss": 0.0937, + "step": 28246 + }, + { + "epoch": 0.6621541635167051, + "grad_norm": 0.25875890254974365, + "learning_rate": 5.1253362557400686e-05, + "loss": 0.0558, + "step": 28247 + }, + { + "epoch": 0.66217760509151, + "grad_norm": 0.4430117607116699, + "learning_rate": 5.1246930989640174e-05, + "loss": 0.0907, + "step": 28248 + }, + { + "epoch": 0.6622010466663151, + "grad_norm": 0.4844094216823578, + "learning_rate": 5.124049968641518e-05, + "loss": 0.0523, + "step": 28249 + }, + { + "epoch": 0.66222448824112, + "grad_norm": 0.6430327296257019, + "learning_rate": 5.123406864776059e-05, + "loss": 0.1211, + "step": 28250 + }, + { + "epoch": 0.6622479298159251, + "grad_norm": 0.48949190974235535, + "learning_rate": 5.122763787371134e-05, + "loss": 0.0127, + "step": 28251 + }, + { + "epoch": 0.66227137139073, + "grad_norm": 0.5415177941322327, + "learning_rate": 5.1221207364302246e-05, + "loss": 0.0665, + "step": 28252 + }, + { + "epoch": 0.662294812965535, + "grad_norm": 0.49408435821533203, + "learning_rate": 5.121477711956829e-05, + "loss": 0.0773, + "step": 28253 + }, + { + "epoch": 0.66231825454034, + "grad_norm": 0.5191634297370911, + "learning_rate": 5.120834713954432e-05, + "loss": 0.0729, + "step": 28254 + }, + { + "epoch": 0.662341696115145, + "grad_norm": 0.24474288523197174, + "learning_rate": 5.1201917424265246e-05, + "loss": 0.0208, + "step": 28255 + }, + { + "epoch": 0.66236513768995, + "grad_norm": 0.4665239155292511, + "learning_rate": 5.1195487973765935e-05, + "loss": 0.0716, + "step": 28256 + }, + { + "epoch": 0.662388579264755, + "grad_norm": 0.6212839484214783, + "learning_rate": 5.1189058788081236e-05, + "loss": 0.0793, + "step": 28257 + }, + { + "epoch": 0.66241202083956, + "grad_norm": 0.14588148891925812, + "learning_rate": 5.1182629867246114e-05, + "loss": 0.0343, + "step": 28258 + }, + { + "epoch": 0.662435462414365, + "grad_norm": 0.10159442573785782, + "learning_rate": 5.117620121129538e-05, + "loss": 0.0111, + "step": 28259 + }, + { + "epoch": 0.66245890398917, + "grad_norm": 0.30252203345298767, + "learning_rate": 5.1169772820263985e-05, + "loss": 0.0375, + "step": 28260 + }, + { + "epoch": 0.662482345563975, + "grad_norm": 0.3430294990539551, + "learning_rate": 5.116334469418678e-05, + "loss": 0.4466, + "step": 28261 + }, + { + "epoch": 0.66250578713878, + "grad_norm": 0.450303316116333, + "learning_rate": 5.115691683309865e-05, + "loss": 0.0553, + "step": 28262 + }, + { + "epoch": 0.662529228713585, + "grad_norm": 0.5031708478927612, + "learning_rate": 5.1150489237034446e-05, + "loss": 0.0978, + "step": 28263 + }, + { + "epoch": 0.6625526702883899, + "grad_norm": 0.3026564121246338, + "learning_rate": 5.1144061906029026e-05, + "loss": 0.0387, + "step": 28264 + }, + { + "epoch": 0.662576111863195, + "grad_norm": 0.12467920780181885, + "learning_rate": 5.1137634840117346e-05, + "loss": 0.0263, + "step": 28265 + }, + { + "epoch": 0.6625995534379999, + "grad_norm": 0.14296954870224, + "learning_rate": 5.113120803933422e-05, + "loss": 0.0301, + "step": 28266 + }, + { + "epoch": 0.662622995012805, + "grad_norm": 0.6260349154472351, + "learning_rate": 5.1124781503714546e-05, + "loss": 0.129, + "step": 28267 + }, + { + "epoch": 0.6626464365876099, + "grad_norm": 0.5456358194351196, + "learning_rate": 5.111835523329318e-05, + "loss": 0.1084, + "step": 28268 + }, + { + "epoch": 0.662669878162415, + "grad_norm": 0.9221796989440918, + "learning_rate": 5.111192922810499e-05, + "loss": 0.1236, + "step": 28269 + }, + { + "epoch": 0.6626933197372199, + "grad_norm": 0.5061889886856079, + "learning_rate": 5.1105503488184805e-05, + "loss": 0.0729, + "step": 28270 + }, + { + "epoch": 0.662716761312025, + "grad_norm": 0.8500256538391113, + "learning_rate": 5.109907801356752e-05, + "loss": 0.1289, + "step": 28271 + }, + { + "epoch": 0.6627402028868299, + "grad_norm": 0.15280428528785706, + "learning_rate": 5.1092652804288075e-05, + "loss": 0.0271, + "step": 28272 + }, + { + "epoch": 0.6627636444616349, + "grad_norm": 0.5514042377471924, + "learning_rate": 5.1086227860381264e-05, + "loss": 0.1053, + "step": 28273 + }, + { + "epoch": 0.6627870860364399, + "grad_norm": 0.13309051096439362, + "learning_rate": 5.1079803181881955e-05, + "loss": 0.0299, + "step": 28274 + }, + { + "epoch": 0.6628105276112449, + "grad_norm": 0.13358066976070404, + "learning_rate": 5.107337876882501e-05, + "loss": 0.0109, + "step": 28275 + }, + { + "epoch": 0.6628339691860499, + "grad_norm": 0.18662790954113007, + "learning_rate": 5.106695462124529e-05, + "loss": 0.0392, + "step": 28276 + }, + { + "epoch": 0.6628574107608549, + "grad_norm": 0.6489805579185486, + "learning_rate": 5.1060530739177605e-05, + "loss": 0.1573, + "step": 28277 + }, + { + "epoch": 0.66288085233566, + "grad_norm": 0.4173782467842102, + "learning_rate": 5.1054107122656905e-05, + "loss": 0.0692, + "step": 28278 + }, + { + "epoch": 0.6629042939104649, + "grad_norm": 0.36239174008369446, + "learning_rate": 5.104768377171798e-05, + "loss": 0.0676, + "step": 28279 + }, + { + "epoch": 0.66292773548527, + "grad_norm": 0.37923184037208557, + "learning_rate": 5.104126068639572e-05, + "loss": 0.0836, + "step": 28280 + }, + { + "epoch": 0.6629511770600749, + "grad_norm": 0.15540383756160736, + "learning_rate": 5.103483786672494e-05, + "loss": 0.0335, + "step": 28281 + }, + { + "epoch": 0.6629746186348799, + "grad_norm": 0.4637517035007477, + "learning_rate": 5.102841531274052e-05, + "loss": 0.0557, + "step": 28282 + }, + { + "epoch": 0.6629980602096849, + "grad_norm": 0.4817960858345032, + "learning_rate": 5.1021993024477234e-05, + "loss": 0.0762, + "step": 28283 + }, + { + "epoch": 0.6630215017844899, + "grad_norm": 0.4052039682865143, + "learning_rate": 5.101557100197001e-05, + "loss": 0.0784, + "step": 28284 + }, + { + "epoch": 0.6630449433592949, + "grad_norm": 0.6320088505744934, + "learning_rate": 5.100914924525371e-05, + "loss": 0.1073, + "step": 28285 + }, + { + "epoch": 0.6630683849340999, + "grad_norm": 0.4513605237007141, + "learning_rate": 5.1002727754363144e-05, + "loss": 0.0721, + "step": 28286 + }, + { + "epoch": 0.6630918265089049, + "grad_norm": 0.4589761793613434, + "learning_rate": 5.099630652933315e-05, + "loss": 0.6351, + "step": 28287 + }, + { + "epoch": 0.6631152680837099, + "grad_norm": 0.39820778369903564, + "learning_rate": 5.098988557019857e-05, + "loss": 0.0646, + "step": 28288 + }, + { + "epoch": 0.6631387096585148, + "grad_norm": 0.34110227227211, + "learning_rate": 5.0983464876994256e-05, + "loss": 0.0529, + "step": 28289 + }, + { + "epoch": 0.6631621512333199, + "grad_norm": 0.4682006537914276, + "learning_rate": 5.097704444975501e-05, + "loss": 0.0799, + "step": 28290 + }, + { + "epoch": 0.6631855928081248, + "grad_norm": 0.5965479016304016, + "learning_rate": 5.0970624288515734e-05, + "loss": 0.1506, + "step": 28291 + }, + { + "epoch": 0.6632090343829299, + "grad_norm": 0.27034491300582886, + "learning_rate": 5.0964204393311225e-05, + "loss": 0.0589, + "step": 28292 + }, + { + "epoch": 0.6632324759577348, + "grad_norm": 0.6419584155082703, + "learning_rate": 5.095778476417632e-05, + "loss": 0.594, + "step": 28293 + }, + { + "epoch": 0.6632559175325399, + "grad_norm": 0.3898250162601471, + "learning_rate": 5.0951365401145845e-05, + "loss": 0.0846, + "step": 28294 + }, + { + "epoch": 0.6632793591073448, + "grad_norm": 0.18110205233097076, + "learning_rate": 5.0944946304254614e-05, + "loss": 0.0462, + "step": 28295 + }, + { + "epoch": 0.6633028006821499, + "grad_norm": 0.15120165050029755, + "learning_rate": 5.093852747353749e-05, + "loss": 0.029, + "step": 28296 + }, + { + "epoch": 0.6633262422569548, + "grad_norm": 0.7410808205604553, + "learning_rate": 5.093210890902933e-05, + "loss": 0.1088, + "step": 28297 + }, + { + "epoch": 0.6633496838317599, + "grad_norm": 0.4457915127277374, + "learning_rate": 5.092569061076491e-05, + "loss": 0.0927, + "step": 28298 + }, + { + "epoch": 0.6633731254065648, + "grad_norm": 0.17005732655525208, + "learning_rate": 5.0919272578779086e-05, + "loss": 0.0268, + "step": 28299 + }, + { + "epoch": 0.6633965669813698, + "grad_norm": 0.5445696115493774, + "learning_rate": 5.091285481310668e-05, + "loss": 0.7277, + "step": 28300 + }, + { + "epoch": 0.6634200085561748, + "grad_norm": 0.551390528678894, + "learning_rate": 5.090643731378249e-05, + "loss": 0.1342, + "step": 28301 + }, + { + "epoch": 0.6634434501309798, + "grad_norm": 0.2122267782688141, + "learning_rate": 5.090002008084132e-05, + "loss": 0.0178, + "step": 28302 + }, + { + "epoch": 0.6634668917057848, + "grad_norm": 0.18352840840816498, + "learning_rate": 5.089360311431806e-05, + "loss": 0.0275, + "step": 28303 + }, + { + "epoch": 0.6634903332805898, + "grad_norm": 0.8045634627342224, + "learning_rate": 5.088718641424748e-05, + "loss": 0.1154, + "step": 28304 + }, + { + "epoch": 0.6635137748553948, + "grad_norm": 0.5970107913017273, + "learning_rate": 5.08807699806644e-05, + "loss": 0.0904, + "step": 28305 + }, + { + "epoch": 0.6635372164301998, + "grad_norm": 0.2564712166786194, + "learning_rate": 5.087435381360366e-05, + "loss": 0.0239, + "step": 28306 + }, + { + "epoch": 0.6635606580050047, + "grad_norm": 0.3728437125682831, + "learning_rate": 5.0867937913100015e-05, + "loss": 0.3532, + "step": 28307 + }, + { + "epoch": 0.6635840995798098, + "grad_norm": 0.5636919736862183, + "learning_rate": 5.086152227918837e-05, + "loss": 0.1003, + "step": 28308 + }, + { + "epoch": 0.6636075411546147, + "grad_norm": 0.22151316702365875, + "learning_rate": 5.085510691190343e-05, + "loss": 0.0142, + "step": 28309 + }, + { + "epoch": 0.6636309827294198, + "grad_norm": 0.3352661430835724, + "learning_rate": 5.0848691811280115e-05, + "loss": 0.0796, + "step": 28310 + }, + { + "epoch": 0.6636544243042247, + "grad_norm": 0.4566314220428467, + "learning_rate": 5.084227697735317e-05, + "loss": 0.0497, + "step": 28311 + }, + { + "epoch": 0.6636778658790298, + "grad_norm": 0.7098851203918457, + "learning_rate": 5.083586241015741e-05, + "loss": 0.1827, + "step": 28312 + }, + { + "epoch": 0.6637013074538347, + "grad_norm": 0.30352410674095154, + "learning_rate": 5.082944810972765e-05, + "loss": 0.0356, + "step": 28313 + }, + { + "epoch": 0.6637247490286398, + "grad_norm": 0.23257429897785187, + "learning_rate": 5.0823034076098684e-05, + "loss": 0.0305, + "step": 28314 + }, + { + "epoch": 0.6637481906034447, + "grad_norm": 0.5495561361312866, + "learning_rate": 5.0816620309305274e-05, + "loss": 0.0573, + "step": 28315 + }, + { + "epoch": 0.6637716321782497, + "grad_norm": 0.34696707129478455, + "learning_rate": 5.081020680938231e-05, + "loss": 0.0539, + "step": 28316 + }, + { + "epoch": 0.6637950737530547, + "grad_norm": 0.15004102885723114, + "learning_rate": 5.0803793576364555e-05, + "loss": 0.0319, + "step": 28317 + }, + { + "epoch": 0.6638185153278597, + "grad_norm": 0.3833390772342682, + "learning_rate": 5.079738061028679e-05, + "loss": 0.0847, + "step": 28318 + }, + { + "epoch": 0.6638419569026647, + "grad_norm": 0.30419182777404785, + "learning_rate": 5.0790967911183826e-05, + "loss": 0.0728, + "step": 28319 + }, + { + "epoch": 0.6638653984774697, + "grad_norm": 0.4787718951702118, + "learning_rate": 5.078455547909041e-05, + "loss": 0.073, + "step": 28320 + }, + { + "epoch": 0.6638888400522747, + "grad_norm": 0.2216876894235611, + "learning_rate": 5.077814331404143e-05, + "loss": 0.0368, + "step": 28321 + }, + { + "epoch": 0.6639122816270797, + "grad_norm": 0.44054582715034485, + "learning_rate": 5.077173141607157e-05, + "loss": 0.0754, + "step": 28322 + }, + { + "epoch": 0.6639357232018847, + "grad_norm": 0.340481162071228, + "learning_rate": 5.076531978521573e-05, + "loss": 0.0794, + "step": 28323 + }, + { + "epoch": 0.6639591647766897, + "grad_norm": 0.39303654432296753, + "learning_rate": 5.0758908421508655e-05, + "loss": 0.0955, + "step": 28324 + }, + { + "epoch": 0.6639826063514946, + "grad_norm": 0.18006575107574463, + "learning_rate": 5.0752497324985106e-05, + "loss": 0.0167, + "step": 28325 + }, + { + "epoch": 0.6640060479262997, + "grad_norm": 0.391385018825531, + "learning_rate": 5.074608649567991e-05, + "loss": 0.0423, + "step": 28326 + }, + { + "epoch": 0.6640294895011046, + "grad_norm": 0.19216331839561462, + "learning_rate": 5.0739675933627785e-05, + "loss": 0.0285, + "step": 28327 + }, + { + "epoch": 0.6640529310759097, + "grad_norm": 0.24283349514007568, + "learning_rate": 5.073326563886361e-05, + "loss": 0.0286, + "step": 28328 + }, + { + "epoch": 0.6640763726507147, + "grad_norm": 0.44570714235305786, + "learning_rate": 5.072685561142212e-05, + "loss": 0.0649, + "step": 28329 + }, + { + "epoch": 0.6640998142255197, + "grad_norm": 0.3616807460784912, + "learning_rate": 5.0720445851338104e-05, + "loss": 0.0386, + "step": 28330 + }, + { + "epoch": 0.6641232558003247, + "grad_norm": 0.2200206220149994, + "learning_rate": 5.071403635864631e-05, + "loss": 0.0326, + "step": 28331 + }, + { + "epoch": 0.6641466973751297, + "grad_norm": 0.40961021184921265, + "learning_rate": 5.070762713338153e-05, + "loss": 0.0341, + "step": 28332 + }, + { + "epoch": 0.6641701389499347, + "grad_norm": 0.1763095259666443, + "learning_rate": 5.070121817557858e-05, + "loss": 0.0322, + "step": 28333 + }, + { + "epoch": 0.6641935805247396, + "grad_norm": 0.4611895978450775, + "learning_rate": 5.069480948527217e-05, + "loss": 0.0649, + "step": 28334 + }, + { + "epoch": 0.6642170220995447, + "grad_norm": 0.6971600651741028, + "learning_rate": 5.068840106249712e-05, + "loss": 0.061, + "step": 28335 + }, + { + "epoch": 0.6642404636743496, + "grad_norm": 0.2583531141281128, + "learning_rate": 5.0681992907288234e-05, + "loss": 0.0234, + "step": 28336 + }, + { + "epoch": 0.6642639052491547, + "grad_norm": 0.5048166513442993, + "learning_rate": 5.067558501968021e-05, + "loss": 0.0922, + "step": 28337 + }, + { + "epoch": 0.6642873468239596, + "grad_norm": 0.4350679814815521, + "learning_rate": 5.0669177399707865e-05, + "loss": 0.0197, + "step": 28338 + }, + { + "epoch": 0.6643107883987647, + "grad_norm": 0.2853572368621826, + "learning_rate": 5.066277004740595e-05, + "loss": 0.0528, + "step": 28339 + }, + { + "epoch": 0.6643342299735696, + "grad_norm": 0.5370116829872131, + "learning_rate": 5.065636296280919e-05, + "loss": 0.0908, + "step": 28340 + }, + { + "epoch": 0.6643576715483747, + "grad_norm": 0.3715203106403351, + "learning_rate": 5.064995614595243e-05, + "loss": 0.0337, + "step": 28341 + }, + { + "epoch": 0.6643811131231796, + "grad_norm": 0.8908746242523193, + "learning_rate": 5.06435495968704e-05, + "loss": 0.0783, + "step": 28342 + }, + { + "epoch": 0.6644045546979847, + "grad_norm": 0.18746715784072876, + "learning_rate": 5.063714331559786e-05, + "loss": 0.0252, + "step": 28343 + }, + { + "epoch": 0.6644279962727896, + "grad_norm": 0.6307457685470581, + "learning_rate": 5.0630737302169515e-05, + "loss": 0.0883, + "step": 28344 + }, + { + "epoch": 0.6644514378475946, + "grad_norm": 0.125565305352211, + "learning_rate": 5.062433155662023e-05, + "loss": 0.0182, + "step": 28345 + }, + { + "epoch": 0.6644748794223996, + "grad_norm": 0.4884243905544281, + "learning_rate": 5.0617926078984704e-05, + "loss": 0.1007, + "step": 28346 + }, + { + "epoch": 0.6644983209972046, + "grad_norm": 0.40764498710632324, + "learning_rate": 5.061152086929766e-05, + "loss": 0.0451, + "step": 28347 + }, + { + "epoch": 0.6645217625720096, + "grad_norm": 0.5557478070259094, + "learning_rate": 5.060511592759394e-05, + "loss": 0.0644, + "step": 28348 + }, + { + "epoch": 0.6645452041468146, + "grad_norm": 0.5547367334365845, + "learning_rate": 5.059871125390824e-05, + "loss": 0.1379, + "step": 28349 + }, + { + "epoch": 0.6645686457216196, + "grad_norm": 0.15423361957073212, + "learning_rate": 5.059230684827532e-05, + "loss": 0.0134, + "step": 28350 + }, + { + "epoch": 0.6645920872964246, + "grad_norm": 0.4561254680156708, + "learning_rate": 5.0585902710729935e-05, + "loss": 0.117, + "step": 28351 + }, + { + "epoch": 0.6646155288712295, + "grad_norm": 0.4103081524372101, + "learning_rate": 5.0579498841306835e-05, + "loss": 0.0949, + "step": 28352 + }, + { + "epoch": 0.6646389704460346, + "grad_norm": 0.19961203634738922, + "learning_rate": 5.0573095240040726e-05, + "loss": 0.0446, + "step": 28353 + }, + { + "epoch": 0.6646624120208395, + "grad_norm": 0.6402352452278137, + "learning_rate": 5.056669190696644e-05, + "loss": 0.1618, + "step": 28354 + }, + { + "epoch": 0.6646858535956446, + "grad_norm": 0.30236244201660156, + "learning_rate": 5.056028884211865e-05, + "loss": 0.0989, + "step": 28355 + }, + { + "epoch": 0.6647092951704495, + "grad_norm": 0.3948476016521454, + "learning_rate": 5.0553886045532106e-05, + "loss": 0.0559, + "step": 28356 + }, + { + "epoch": 0.6647327367452546, + "grad_norm": 0.30829918384552, + "learning_rate": 5.054748351724159e-05, + "loss": 0.0559, + "step": 28357 + }, + { + "epoch": 0.6647561783200595, + "grad_norm": 0.5427943468093872, + "learning_rate": 5.054108125728183e-05, + "loss": 0.0761, + "step": 28358 + }, + { + "epoch": 0.6647796198948646, + "grad_norm": 0.5031787753105164, + "learning_rate": 5.0534679265687535e-05, + "loss": 0.0728, + "step": 28359 + }, + { + "epoch": 0.6648030614696695, + "grad_norm": 0.6817097067832947, + "learning_rate": 5.052827754249344e-05, + "loss": 0.1101, + "step": 28360 + }, + { + "epoch": 0.6648265030444745, + "grad_norm": 0.5486842393875122, + "learning_rate": 5.0521876087734334e-05, + "loss": 0.0688, + "step": 28361 + }, + { + "epoch": 0.6648499446192795, + "grad_norm": 0.22226490080356598, + "learning_rate": 5.0515474901444924e-05, + "loss": 0.0365, + "step": 28362 + }, + { + "epoch": 0.6648733861940845, + "grad_norm": 0.5219497680664062, + "learning_rate": 5.050907398365994e-05, + "loss": 0.0489, + "step": 28363 + }, + { + "epoch": 0.6648968277688895, + "grad_norm": 0.0883244201540947, + "learning_rate": 5.050267333441411e-05, + "loss": 0.008, + "step": 28364 + }, + { + "epoch": 0.6649202693436945, + "grad_norm": 0.1894165575504303, + "learning_rate": 5.049627295374213e-05, + "loss": 0.0319, + "step": 28365 + }, + { + "epoch": 0.6649437109184995, + "grad_norm": 0.5996112823486328, + "learning_rate": 5.048987284167879e-05, + "loss": 0.7769, + "step": 28366 + }, + { + "epoch": 0.6649671524933045, + "grad_norm": 0.17191216349601746, + "learning_rate": 5.048347299825881e-05, + "loss": 0.0273, + "step": 28367 + }, + { + "epoch": 0.6649905940681095, + "grad_norm": 0.3644166886806488, + "learning_rate": 5.0477073423516844e-05, + "loss": 0.0668, + "step": 28368 + }, + { + "epoch": 0.6650140356429145, + "grad_norm": 0.3619696795940399, + "learning_rate": 5.047067411748772e-05, + "loss": 0.0454, + "step": 28369 + }, + { + "epoch": 0.6650374772177194, + "grad_norm": 0.36858031153678894, + "learning_rate": 5.04642750802061e-05, + "loss": 0.0598, + "step": 28370 + }, + { + "epoch": 0.6650609187925245, + "grad_norm": 0.244451105594635, + "learning_rate": 5.045787631170673e-05, + "loss": 0.0348, + "step": 28371 + }, + { + "epoch": 0.6650843603673294, + "grad_norm": 0.5458234548568726, + "learning_rate": 5.045147781202426e-05, + "loss": 0.116, + "step": 28372 + }, + { + "epoch": 0.6651078019421345, + "grad_norm": 0.5140501856803894, + "learning_rate": 5.044507958119352e-05, + "loss": 0.1095, + "step": 28373 + }, + { + "epoch": 0.6651312435169394, + "grad_norm": 0.4116092920303345, + "learning_rate": 5.043868161924916e-05, + "loss": 0.0373, + "step": 28374 + }, + { + "epoch": 0.6651546850917445, + "grad_norm": 0.08951465040445328, + "learning_rate": 5.043228392622592e-05, + "loss": 0.0161, + "step": 28375 + }, + { + "epoch": 0.6651781266665494, + "grad_norm": 0.6456086039543152, + "learning_rate": 5.042588650215849e-05, + "loss": 0.1515, + "step": 28376 + }, + { + "epoch": 0.6652015682413545, + "grad_norm": 0.34359049797058105, + "learning_rate": 5.041948934708161e-05, + "loss": 0.0288, + "step": 28377 + }, + { + "epoch": 0.6652250098161594, + "grad_norm": 0.5179802775382996, + "learning_rate": 5.041309246102992e-05, + "loss": 0.0921, + "step": 28378 + }, + { + "epoch": 0.6652484513909644, + "grad_norm": 0.22812797129154205, + "learning_rate": 5.040669584403822e-05, + "loss": 0.0361, + "step": 28379 + }, + { + "epoch": 0.6652718929657695, + "grad_norm": 0.14676228165626526, + "learning_rate": 5.0400299496141156e-05, + "loss": 0.0266, + "step": 28380 + }, + { + "epoch": 0.6652953345405744, + "grad_norm": 0.4596650004386902, + "learning_rate": 5.0393903417373486e-05, + "loss": 0.0883, + "step": 28381 + }, + { + "epoch": 0.6653187761153795, + "grad_norm": 0.3936333954334259, + "learning_rate": 5.0387507607769915e-05, + "loss": 0.0843, + "step": 28382 + }, + { + "epoch": 0.6653422176901844, + "grad_norm": 0.5543660521507263, + "learning_rate": 5.03811120673651e-05, + "loss": 0.0478, + "step": 28383 + }, + { + "epoch": 0.6653656592649895, + "grad_norm": 0.42411646246910095, + "learning_rate": 5.037471679619377e-05, + "loss": 0.1418, + "step": 28384 + }, + { + "epoch": 0.6653891008397944, + "grad_norm": 0.2840111553668976, + "learning_rate": 5.0368321794290585e-05, + "loss": 0.0242, + "step": 28385 + }, + { + "epoch": 0.6654125424145995, + "grad_norm": 0.5672928094863892, + "learning_rate": 5.0361927061690314e-05, + "loss": 0.0736, + "step": 28386 + }, + { + "epoch": 0.6654359839894044, + "grad_norm": 0.5599126219749451, + "learning_rate": 5.035553259842765e-05, + "loss": 0.125, + "step": 28387 + }, + { + "epoch": 0.6654594255642095, + "grad_norm": 0.4209873080253601, + "learning_rate": 5.034913840453723e-05, + "loss": 0.0519, + "step": 28388 + }, + { + "epoch": 0.6654828671390144, + "grad_norm": 0.5207265019416809, + "learning_rate": 5.0342744480053804e-05, + "loss": 0.1087, + "step": 28389 + }, + { + "epoch": 0.6655063087138194, + "grad_norm": 0.671403169631958, + "learning_rate": 5.0336350825012027e-05, + "loss": 0.0498, + "step": 28390 + }, + { + "epoch": 0.6655297502886244, + "grad_norm": 0.11767537891864777, + "learning_rate": 5.032995743944657e-05, + "loss": 0.0112, + "step": 28391 + }, + { + "epoch": 0.6655531918634294, + "grad_norm": 0.40457212924957275, + "learning_rate": 5.03235643233922e-05, + "loss": 0.0718, + "step": 28392 + }, + { + "epoch": 0.6655766334382344, + "grad_norm": 0.34363579750061035, + "learning_rate": 5.031717147688353e-05, + "loss": 0.0329, + "step": 28393 + }, + { + "epoch": 0.6656000750130394, + "grad_norm": 0.646125316619873, + "learning_rate": 5.031077889995532e-05, + "loss": 0.6442, + "step": 28394 + }, + { + "epoch": 0.6656235165878444, + "grad_norm": 0.35331442952156067, + "learning_rate": 5.030438659264223e-05, + "loss": 0.0719, + "step": 28395 + }, + { + "epoch": 0.6656469581626494, + "grad_norm": 0.4076174795627594, + "learning_rate": 5.029799455497893e-05, + "loss": 0.1127, + "step": 28396 + }, + { + "epoch": 0.6656703997374543, + "grad_norm": 0.38418200612068176, + "learning_rate": 5.0291602787000114e-05, + "loss": 0.0586, + "step": 28397 + }, + { + "epoch": 0.6656938413122594, + "grad_norm": 0.4517453610897064, + "learning_rate": 5.028521128874042e-05, + "loss": 0.0584, + "step": 28398 + }, + { + "epoch": 0.6657172828870643, + "grad_norm": 0.15787379443645477, + "learning_rate": 5.02788200602346e-05, + "loss": 0.0046, + "step": 28399 + }, + { + "epoch": 0.6657407244618694, + "grad_norm": 0.051389943808317184, + "learning_rate": 5.02724291015173e-05, + "loss": 0.0108, + "step": 28400 + }, + { + "epoch": 0.6657641660366743, + "grad_norm": 0.1385135054588318, + "learning_rate": 5.02660384126232e-05, + "loss": 0.0179, + "step": 28401 + }, + { + "epoch": 0.6657876076114794, + "grad_norm": 0.5167507529258728, + "learning_rate": 5.025964799358698e-05, + "loss": 0.104, + "step": 28402 + }, + { + "epoch": 0.6658110491862843, + "grad_norm": 0.3321988582611084, + "learning_rate": 5.025325784444326e-05, + "loss": 0.0538, + "step": 28403 + }, + { + "epoch": 0.6658344907610894, + "grad_norm": 0.15616488456726074, + "learning_rate": 5.02468679652268e-05, + "loss": 0.0416, + "step": 28404 + }, + { + "epoch": 0.6658579323358943, + "grad_norm": 0.4687100350856781, + "learning_rate": 5.024047835597221e-05, + "loss": 0.0386, + "step": 28405 + }, + { + "epoch": 0.6658813739106993, + "grad_norm": 0.34573912620544434, + "learning_rate": 5.02340890167142e-05, + "loss": 0.1109, + "step": 28406 + }, + { + "epoch": 0.6659048154855043, + "grad_norm": 0.33909836411476135, + "learning_rate": 5.022769994748744e-05, + "loss": 0.0892, + "step": 28407 + }, + { + "epoch": 0.6659282570603093, + "grad_norm": 0.23911236226558685, + "learning_rate": 5.022131114832658e-05, + "loss": 0.0339, + "step": 28408 + }, + { + "epoch": 0.6659516986351143, + "grad_norm": 0.6869605183601379, + "learning_rate": 5.021492261926627e-05, + "loss": 0.1466, + "step": 28409 + }, + { + "epoch": 0.6659751402099193, + "grad_norm": 0.5318489670753479, + "learning_rate": 5.0208534360341163e-05, + "loss": 0.1133, + "step": 28410 + }, + { + "epoch": 0.6659985817847243, + "grad_norm": 0.5865685343742371, + "learning_rate": 5.020214637158598e-05, + "loss": 0.1348, + "step": 28411 + }, + { + "epoch": 0.6660220233595293, + "grad_norm": 0.10426994413137436, + "learning_rate": 5.0195758653035355e-05, + "loss": 0.0257, + "step": 28412 + }, + { + "epoch": 0.6660454649343343, + "grad_norm": 0.4452437460422516, + "learning_rate": 5.018937120472393e-05, + "loss": 0.0912, + "step": 28413 + }, + { + "epoch": 0.6660689065091393, + "grad_norm": 0.45147427916526794, + "learning_rate": 5.0182984026686396e-05, + "loss": 0.06, + "step": 28414 + }, + { + "epoch": 0.6660923480839442, + "grad_norm": 0.1539076268672943, + "learning_rate": 5.017659711895739e-05, + "loss": 0.013, + "step": 28415 + }, + { + "epoch": 0.6661157896587493, + "grad_norm": 0.31824323534965515, + "learning_rate": 5.017021048157153e-05, + "loss": 0.048, + "step": 28416 + }, + { + "epoch": 0.6661392312335542, + "grad_norm": 0.45028263330459595, + "learning_rate": 5.0163824114563505e-05, + "loss": 0.0781, + "step": 28417 + }, + { + "epoch": 0.6661626728083593, + "grad_norm": 0.6176401972770691, + "learning_rate": 5.0157438017968014e-05, + "loss": 0.085, + "step": 28418 + }, + { + "epoch": 0.6661861143831642, + "grad_norm": 0.4870513677597046, + "learning_rate": 5.015105219181966e-05, + "loss": 0.1031, + "step": 28419 + }, + { + "epoch": 0.6662095559579693, + "grad_norm": 0.5210325121879578, + "learning_rate": 5.0144666636153105e-05, + "loss": 0.1314, + "step": 28420 + }, + { + "epoch": 0.6662329975327742, + "grad_norm": 0.4387798011302948, + "learning_rate": 5.013828135100298e-05, + "loss": 0.094, + "step": 28421 + }, + { + "epoch": 0.6662564391075793, + "grad_norm": 0.2023480087518692, + "learning_rate": 5.013189633640396e-05, + "loss": 0.0173, + "step": 28422 + }, + { + "epoch": 0.6662798806823842, + "grad_norm": 0.6933439373970032, + "learning_rate": 5.0125511592390615e-05, + "loss": 0.1148, + "step": 28423 + }, + { + "epoch": 0.6663033222571892, + "grad_norm": 0.31098800897598267, + "learning_rate": 5.011912711899769e-05, + "loss": 0.0255, + "step": 28424 + }, + { + "epoch": 0.6663267638319942, + "grad_norm": 0.2878481149673462, + "learning_rate": 5.011274291625979e-05, + "loss": 0.021, + "step": 28425 + }, + { + "epoch": 0.6663502054067992, + "grad_norm": 0.9652678370475769, + "learning_rate": 5.010635898421154e-05, + "loss": 0.1069, + "step": 28426 + }, + { + "epoch": 0.6663736469816042, + "grad_norm": 0.5004287362098694, + "learning_rate": 5.009997532288761e-05, + "loss": 0.0908, + "step": 28427 + }, + { + "epoch": 0.6663970885564092, + "grad_norm": 0.5986183285713196, + "learning_rate": 5.009359193232259e-05, + "loss": 0.0732, + "step": 28428 + }, + { + "epoch": 0.6664205301312142, + "grad_norm": 0.4295244514942169, + "learning_rate": 5.008720881255112e-05, + "loss": 0.0732, + "step": 28429 + }, + { + "epoch": 0.6664439717060192, + "grad_norm": 0.09262458235025406, + "learning_rate": 5.008082596360786e-05, + "loss": 0.0193, + "step": 28430 + }, + { + "epoch": 0.6664674132808243, + "grad_norm": 0.3994033634662628, + "learning_rate": 5.0074443385527467e-05, + "loss": 0.0801, + "step": 28431 + }, + { + "epoch": 0.6664908548556292, + "grad_norm": 0.37412452697753906, + "learning_rate": 5.006806107834457e-05, + "loss": 0.0838, + "step": 28432 + }, + { + "epoch": 0.6665142964304342, + "grad_norm": 0.2847629189491272, + "learning_rate": 5.0061679042093754e-05, + "loss": 0.055, + "step": 28433 + }, + { + "epoch": 0.6665377380052392, + "grad_norm": 0.5239623785018921, + "learning_rate": 5.005529727680969e-05, + "loss": 0.0718, + "step": 28434 + }, + { + "epoch": 0.6665611795800442, + "grad_norm": 0.4078652858734131, + "learning_rate": 5.0048915782526986e-05, + "loss": 0.0745, + "step": 28435 + }, + { + "epoch": 0.6665846211548492, + "grad_norm": 0.460440993309021, + "learning_rate": 5.0042534559280216e-05, + "loss": 0.1157, + "step": 28436 + }, + { + "epoch": 0.6666080627296542, + "grad_norm": 0.15916071832180023, + "learning_rate": 5.003615360710411e-05, + "loss": 0.0216, + "step": 28437 + }, + { + "epoch": 0.6666315043044592, + "grad_norm": 0.41518133878707886, + "learning_rate": 5.0029772926033234e-05, + "loss": 0.0896, + "step": 28438 + }, + { + "epoch": 0.6666549458792642, + "grad_norm": 0.07155003398656845, + "learning_rate": 5.002339251610222e-05, + "loss": 0.0076, + "step": 28439 + }, + { + "epoch": 0.6666783874540692, + "grad_norm": 0.39306047558784485, + "learning_rate": 5.001701237734568e-05, + "loss": 0.0459, + "step": 28440 + }, + { + "epoch": 0.6667018290288742, + "grad_norm": 0.5742803812026978, + "learning_rate": 5.00106325097982e-05, + "loss": 0.6858, + "step": 28441 + }, + { + "epoch": 0.6667252706036791, + "grad_norm": 0.3200405538082123, + "learning_rate": 5.0004252913494466e-05, + "loss": 0.054, + "step": 28442 + }, + { + "epoch": 0.6667487121784842, + "grad_norm": 0.787062406539917, + "learning_rate": 4.999787358846902e-05, + "loss": 0.0958, + "step": 28443 + }, + { + "epoch": 0.6667721537532891, + "grad_norm": 0.39872685074806213, + "learning_rate": 4.9991494534756566e-05, + "loss": 0.0716, + "step": 28444 + }, + { + "epoch": 0.6667955953280942, + "grad_norm": 0.15063020586967468, + "learning_rate": 4.9985115752391664e-05, + "loss": 0.0261, + "step": 28445 + }, + { + "epoch": 0.6668190369028991, + "grad_norm": 0.4146052896976471, + "learning_rate": 4.9978737241408934e-05, + "loss": 0.0765, + "step": 28446 + }, + { + "epoch": 0.6668424784777042, + "grad_norm": 0.4035850167274475, + "learning_rate": 4.9972359001842985e-05, + "loss": 0.0546, + "step": 28447 + }, + { + "epoch": 0.6668659200525091, + "grad_norm": 0.15160802006721497, + "learning_rate": 4.996598103372837e-05, + "loss": 0.0242, + "step": 28448 + }, + { + "epoch": 0.6668893616273142, + "grad_norm": 0.24172331392765045, + "learning_rate": 4.9959603337099804e-05, + "loss": 0.0345, + "step": 28449 + }, + { + "epoch": 0.6669128032021191, + "grad_norm": 0.2788741886615753, + "learning_rate": 4.995322591199183e-05, + "loss": 0.0599, + "step": 28450 + }, + { + "epoch": 0.6669362447769241, + "grad_norm": 0.52170330286026, + "learning_rate": 4.994684875843906e-05, + "loss": 0.0792, + "step": 28451 + }, + { + "epoch": 0.6669596863517291, + "grad_norm": 0.4237617552280426, + "learning_rate": 4.994047187647609e-05, + "loss": 0.0783, + "step": 28452 + }, + { + "epoch": 0.6669831279265341, + "grad_norm": 0.15363295376300812, + "learning_rate": 4.9934095266137494e-05, + "loss": 0.021, + "step": 28453 + }, + { + "epoch": 0.6670065695013391, + "grad_norm": 0.11445576697587967, + "learning_rate": 4.992771892745794e-05, + "loss": 0.0209, + "step": 28454 + }, + { + "epoch": 0.6670300110761441, + "grad_norm": 0.40082040429115295, + "learning_rate": 4.992134286047195e-05, + "loss": 0.5548, + "step": 28455 + }, + { + "epoch": 0.6670534526509491, + "grad_norm": 0.24456095695495605, + "learning_rate": 4.9914967065214214e-05, + "loss": 0.0551, + "step": 28456 + }, + { + "epoch": 0.6670768942257541, + "grad_norm": 0.39494678378105164, + "learning_rate": 4.9908591541719265e-05, + "loss": 0.111, + "step": 28457 + }, + { + "epoch": 0.667100335800559, + "grad_norm": 0.3234943449497223, + "learning_rate": 4.990221629002171e-05, + "loss": 0.07, + "step": 28458 + }, + { + "epoch": 0.6671237773753641, + "grad_norm": 0.35248252749443054, + "learning_rate": 4.989584131015613e-05, + "loss": 0.0355, + "step": 28459 + }, + { + "epoch": 0.667147218950169, + "grad_norm": 0.48000478744506836, + "learning_rate": 4.988946660215713e-05, + "loss": 0.0626, + "step": 28460 + }, + { + "epoch": 0.6671706605249741, + "grad_norm": 0.3983025550842285, + "learning_rate": 4.9883092166059266e-05, + "loss": 0.0721, + "step": 28461 + }, + { + "epoch": 0.667194102099779, + "grad_norm": 0.40291309356689453, + "learning_rate": 4.987671800189717e-05, + "loss": 0.0418, + "step": 28462 + }, + { + "epoch": 0.6672175436745841, + "grad_norm": 0.4665994644165039, + "learning_rate": 4.987034410970543e-05, + "loss": 0.1031, + "step": 28463 + }, + { + "epoch": 0.667240985249389, + "grad_norm": 0.6645774841308594, + "learning_rate": 4.9863970489518596e-05, + "loss": 0.1598, + "step": 28464 + }, + { + "epoch": 0.6672644268241941, + "grad_norm": 0.5431316494941711, + "learning_rate": 4.985759714137125e-05, + "loss": 0.0964, + "step": 28465 + }, + { + "epoch": 0.667287868398999, + "grad_norm": 0.5551738142967224, + "learning_rate": 4.985122406529801e-05, + "loss": 0.116, + "step": 28466 + }, + { + "epoch": 0.6673113099738041, + "grad_norm": 0.6114863753318787, + "learning_rate": 4.984485126133345e-05, + "loss": 0.1248, + "step": 28467 + }, + { + "epoch": 0.667334751548609, + "grad_norm": 0.9984340667724609, + "learning_rate": 4.983847872951209e-05, + "loss": 0.0579, + "step": 28468 + }, + { + "epoch": 0.667358193123414, + "grad_norm": 0.20049987733364105, + "learning_rate": 4.98321064698686e-05, + "loss": 0.0216, + "step": 28469 + }, + { + "epoch": 0.667381634698219, + "grad_norm": 0.1829061061143875, + "learning_rate": 4.9825734482437515e-05, + "loss": 0.029, + "step": 28470 + }, + { + "epoch": 0.667405076273024, + "grad_norm": 0.44706597924232483, + "learning_rate": 4.9819362767253396e-05, + "loss": 0.0737, + "step": 28471 + }, + { + "epoch": 0.667428517847829, + "grad_norm": 0.4665707051753998, + "learning_rate": 4.981299132435083e-05, + "loss": 0.2962, + "step": 28472 + }, + { + "epoch": 0.667451959422634, + "grad_norm": 0.345113068819046, + "learning_rate": 4.9806620153764384e-05, + "loss": 0.0779, + "step": 28473 + }, + { + "epoch": 0.667475400997439, + "grad_norm": 0.34951379895210266, + "learning_rate": 4.980024925552859e-05, + "loss": 0.2791, + "step": 28474 + }, + { + "epoch": 0.667498842572244, + "grad_norm": 0.5408532023429871, + "learning_rate": 4.979387862967809e-05, + "loss": 0.0812, + "step": 28475 + }, + { + "epoch": 0.667522284147049, + "grad_norm": 0.24665947258472443, + "learning_rate": 4.978750827624742e-05, + "loss": 0.0517, + "step": 28476 + }, + { + "epoch": 0.667545725721854, + "grad_norm": 0.39689335227012634, + "learning_rate": 4.9781138195271134e-05, + "loss": 0.0692, + "step": 28477 + }, + { + "epoch": 0.6675691672966589, + "grad_norm": 0.26061636209487915, + "learning_rate": 4.9774768386783774e-05, + "loss": 0.0482, + "step": 28478 + }, + { + "epoch": 0.667592608871464, + "grad_norm": 0.2915738523006439, + "learning_rate": 4.9768398850819975e-05, + "loss": 0.0427, + "step": 28479 + }, + { + "epoch": 0.6676160504462689, + "grad_norm": 0.5326611399650574, + "learning_rate": 4.976202958741425e-05, + "loss": 0.0756, + "step": 28480 + }, + { + "epoch": 0.667639492021074, + "grad_norm": 0.57705157995224, + "learning_rate": 4.975566059660113e-05, + "loss": 0.1428, + "step": 28481 + }, + { + "epoch": 0.6676629335958789, + "grad_norm": 0.08755559474229813, + "learning_rate": 4.974929187841525e-05, + "loss": 0.0151, + "step": 28482 + }, + { + "epoch": 0.667686375170684, + "grad_norm": 0.31070080399513245, + "learning_rate": 4.974292343289112e-05, + "loss": 0.0774, + "step": 28483 + }, + { + "epoch": 0.667709816745489, + "grad_norm": 0.4351177215576172, + "learning_rate": 4.97365552600633e-05, + "loss": 0.1258, + "step": 28484 + }, + { + "epoch": 0.667733258320294, + "grad_norm": 0.7874510288238525, + "learning_rate": 4.973018735996634e-05, + "loss": 0.1203, + "step": 28485 + }, + { + "epoch": 0.667756699895099, + "grad_norm": 0.3087928891181946, + "learning_rate": 4.972381973263477e-05, + "loss": 0.0615, + "step": 28486 + }, + { + "epoch": 0.6677801414699039, + "grad_norm": 0.6192195415496826, + "learning_rate": 4.971745237810319e-05, + "loss": 0.0797, + "step": 28487 + }, + { + "epoch": 0.667803583044709, + "grad_norm": 0.1747884601354599, + "learning_rate": 4.971108529640615e-05, + "loss": 0.0253, + "step": 28488 + }, + { + "epoch": 0.6678270246195139, + "grad_norm": 0.4648594260215759, + "learning_rate": 4.970471848757815e-05, + "loss": 0.0737, + "step": 28489 + }, + { + "epoch": 0.667850466194319, + "grad_norm": 0.37857934832572937, + "learning_rate": 4.9698351951653745e-05, + "loss": 0.0947, + "step": 28490 + }, + { + "epoch": 0.6678739077691239, + "grad_norm": 0.33534321188926697, + "learning_rate": 4.969198568866752e-05, + "loss": 0.0611, + "step": 28491 + }, + { + "epoch": 0.667897349343929, + "grad_norm": 0.6950438022613525, + "learning_rate": 4.9685619698653994e-05, + "loss": 0.0817, + "step": 28492 + }, + { + "epoch": 0.6679207909187339, + "grad_norm": 0.6523897051811218, + "learning_rate": 4.967925398164769e-05, + "loss": 0.1408, + "step": 28493 + }, + { + "epoch": 0.667944232493539, + "grad_norm": 0.5514576435089111, + "learning_rate": 4.967288853768319e-05, + "loss": 0.1064, + "step": 28494 + }, + { + "epoch": 0.6679676740683439, + "grad_norm": 0.22835730016231537, + "learning_rate": 4.966652336679501e-05, + "loss": 0.0557, + "step": 28495 + }, + { + "epoch": 0.667991115643149, + "grad_norm": 0.1366700977087021, + "learning_rate": 4.96601584690177e-05, + "loss": 0.0145, + "step": 28496 + }, + { + "epoch": 0.6680145572179539, + "grad_norm": 0.7347385883331299, + "learning_rate": 4.965379384438579e-05, + "loss": 0.2123, + "step": 28497 + }, + { + "epoch": 0.6680379987927589, + "grad_norm": 0.33210721611976624, + "learning_rate": 4.964742949293379e-05, + "loss": 0.0468, + "step": 28498 + }, + { + "epoch": 0.6680614403675639, + "grad_norm": 0.5187890529632568, + "learning_rate": 4.9641065414696244e-05, + "loss": 0.0278, + "step": 28499 + }, + { + "epoch": 0.6680848819423689, + "grad_norm": 0.7838027477264404, + "learning_rate": 4.963470160970771e-05, + "loss": 0.1572, + "step": 28500 + }, + { + "epoch": 0.6681083235171739, + "grad_norm": 0.5112844109535217, + "learning_rate": 4.96283380780027e-05, + "loss": 0.0927, + "step": 28501 + }, + { + "epoch": 0.6681317650919789, + "grad_norm": 0.411774605512619, + "learning_rate": 4.962197481961571e-05, + "loss": 0.0457, + "step": 28502 + }, + { + "epoch": 0.6681552066667839, + "grad_norm": 0.24685858190059662, + "learning_rate": 4.961561183458134e-05, + "loss": 0.0428, + "step": 28503 + }, + { + "epoch": 0.6681786482415889, + "grad_norm": 0.5281753540039062, + "learning_rate": 4.9609249122934086e-05, + "loss": 0.1279, + "step": 28504 + }, + { + "epoch": 0.6682020898163938, + "grad_norm": 0.33884337544441223, + "learning_rate": 4.960288668470846e-05, + "loss": 0.0523, + "step": 28505 + }, + { + "epoch": 0.6682255313911989, + "grad_norm": 0.44953489303588867, + "learning_rate": 4.9596524519938934e-05, + "loss": 0.09, + "step": 28506 + }, + { + "epoch": 0.6682489729660038, + "grad_norm": 0.3141661286354065, + "learning_rate": 4.959016262866014e-05, + "loss": 0.0475, + "step": 28507 + }, + { + "epoch": 0.6682724145408089, + "grad_norm": 0.6197552680969238, + "learning_rate": 4.9583801010906526e-05, + "loss": 0.6294, + "step": 28508 + }, + { + "epoch": 0.6682958561156138, + "grad_norm": 0.3226650655269623, + "learning_rate": 4.957743966671265e-05, + "loss": 0.0369, + "step": 28509 + }, + { + "epoch": 0.6683192976904189, + "grad_norm": 0.23475374281406403, + "learning_rate": 4.957107859611297e-05, + "loss": 0.0169, + "step": 28510 + }, + { + "epoch": 0.6683427392652238, + "grad_norm": 0.14648224413394928, + "learning_rate": 4.956471779914206e-05, + "loss": 0.046, + "step": 28511 + }, + { + "epoch": 0.6683661808400289, + "grad_norm": 0.5571216344833374, + "learning_rate": 4.955835727583437e-05, + "loss": 0.4917, + "step": 28512 + }, + { + "epoch": 0.6683896224148338, + "grad_norm": 0.552465558052063, + "learning_rate": 4.955199702622447e-05, + "loss": 0.1392, + "step": 28513 + }, + { + "epoch": 0.6684130639896388, + "grad_norm": 0.6392872333526611, + "learning_rate": 4.9545637050346835e-05, + "loss": 0.4195, + "step": 28514 + }, + { + "epoch": 0.6684365055644438, + "grad_norm": 0.6530625820159912, + "learning_rate": 4.9539277348236026e-05, + "loss": 0.1308, + "step": 28515 + }, + { + "epoch": 0.6684599471392488, + "grad_norm": 0.35539481043815613, + "learning_rate": 4.953291791992652e-05, + "loss": 0.039, + "step": 28516 + }, + { + "epoch": 0.6684833887140538, + "grad_norm": 0.1350201815366745, + "learning_rate": 4.9526558765452805e-05, + "loss": 0.0226, + "step": 28517 + }, + { + "epoch": 0.6685068302888588, + "grad_norm": 0.44032442569732666, + "learning_rate": 4.952019988484941e-05, + "loss": 0.0747, + "step": 28518 + }, + { + "epoch": 0.6685302718636638, + "grad_norm": 0.37160956859588623, + "learning_rate": 4.9513841278150786e-05, + "loss": 0.4429, + "step": 28519 + }, + { + "epoch": 0.6685537134384688, + "grad_norm": 0.47026047110557556, + "learning_rate": 4.950748294539153e-05, + "loss": 0.1012, + "step": 28520 + }, + { + "epoch": 0.6685771550132737, + "grad_norm": 0.508867084980011, + "learning_rate": 4.950112488660609e-05, + "loss": 0.1003, + "step": 28521 + }, + { + "epoch": 0.6686005965880788, + "grad_norm": 0.5524628758430481, + "learning_rate": 4.9494767101828955e-05, + "loss": 0.0845, + "step": 28522 + }, + { + "epoch": 0.6686240381628837, + "grad_norm": 0.2611493170261383, + "learning_rate": 4.948840959109464e-05, + "loss": 0.0365, + "step": 28523 + }, + { + "epoch": 0.6686474797376888, + "grad_norm": 0.2532016634941101, + "learning_rate": 4.948205235443759e-05, + "loss": 0.0165, + "step": 28524 + }, + { + "epoch": 0.6686709213124937, + "grad_norm": 0.7877594232559204, + "learning_rate": 4.9475695391892394e-05, + "loss": 0.1388, + "step": 28525 + }, + { + "epoch": 0.6686943628872988, + "grad_norm": 0.18348807096481323, + "learning_rate": 4.946933870349345e-05, + "loss": 0.0289, + "step": 28526 + }, + { + "epoch": 0.6687178044621037, + "grad_norm": 0.31675729155540466, + "learning_rate": 4.946298228927534e-05, + "loss": 0.0704, + "step": 28527 + }, + { + "epoch": 0.6687412460369088, + "grad_norm": 0.5079202055931091, + "learning_rate": 4.94566261492725e-05, + "loss": 0.0988, + "step": 28528 + }, + { + "epoch": 0.6687646876117137, + "grad_norm": 0.4770209789276123, + "learning_rate": 4.945027028351943e-05, + "loss": 0.0506, + "step": 28529 + }, + { + "epoch": 0.6687881291865188, + "grad_norm": 0.7341052293777466, + "learning_rate": 4.944391469205062e-05, + "loss": 0.1035, + "step": 28530 + }, + { + "epoch": 0.6688115707613237, + "grad_norm": 0.11178471893072128, + "learning_rate": 4.943755937490051e-05, + "loss": 0.0213, + "step": 28531 + }, + { + "epoch": 0.6688350123361287, + "grad_norm": 0.4176097512245178, + "learning_rate": 4.9431204332103654e-05, + "loss": 0.0734, + "step": 28532 + }, + { + "epoch": 0.6688584539109337, + "grad_norm": 0.5088422298431396, + "learning_rate": 4.942484956369452e-05, + "loss": 0.1059, + "step": 28533 + }, + { + "epoch": 0.6688818954857387, + "grad_norm": 0.7435379028320312, + "learning_rate": 4.941849506970756e-05, + "loss": 0.1569, + "step": 28534 + }, + { + "epoch": 0.6689053370605438, + "grad_norm": 0.688328206539154, + "learning_rate": 4.941214085017728e-05, + "loss": 0.0781, + "step": 28535 + }, + { + "epoch": 0.6689287786353487, + "grad_norm": 0.4606848359107971, + "learning_rate": 4.940578690513814e-05, + "loss": 0.051, + "step": 28536 + }, + { + "epoch": 0.6689522202101538, + "grad_norm": 0.15873540937900543, + "learning_rate": 4.939943323462458e-05, + "loss": 0.0373, + "step": 28537 + }, + { + "epoch": 0.6689756617849587, + "grad_norm": 0.36875227093696594, + "learning_rate": 4.939307983867113e-05, + "loss": 0.0538, + "step": 28538 + }, + { + "epoch": 0.6689991033597638, + "grad_norm": 0.08016548305749893, + "learning_rate": 4.938672671731228e-05, + "loss": 0.0162, + "step": 28539 + }, + { + "epoch": 0.6690225449345687, + "grad_norm": 0.14648312330245972, + "learning_rate": 4.9380373870582474e-05, + "loss": 0.0297, + "step": 28540 + }, + { + "epoch": 0.6690459865093737, + "grad_norm": 0.23885221779346466, + "learning_rate": 4.937402129851619e-05, + "loss": 0.0136, + "step": 28541 + }, + { + "epoch": 0.6690694280841787, + "grad_norm": 0.4757619798183441, + "learning_rate": 4.9367669001147886e-05, + "loss": 0.2868, + "step": 28542 + }, + { + "epoch": 0.6690928696589837, + "grad_norm": 0.5115289092063904, + "learning_rate": 4.936131697851204e-05, + "loss": 0.1326, + "step": 28543 + }, + { + "epoch": 0.6691163112337887, + "grad_norm": 0.5445044636726379, + "learning_rate": 4.935496523064307e-05, + "loss": 0.0798, + "step": 28544 + }, + { + "epoch": 0.6691397528085937, + "grad_norm": 0.18207882344722748, + "learning_rate": 4.9348613757575513e-05, + "loss": 0.0444, + "step": 28545 + }, + { + "epoch": 0.6691631943833987, + "grad_norm": 0.07324429601430893, + "learning_rate": 4.934226255934381e-05, + "loss": 0.0154, + "step": 28546 + }, + { + "epoch": 0.6691866359582037, + "grad_norm": 0.4133371412754059, + "learning_rate": 4.933591163598241e-05, + "loss": 0.0697, + "step": 28547 + }, + { + "epoch": 0.6692100775330087, + "grad_norm": 0.40201228857040405, + "learning_rate": 4.932956098752578e-05, + "loss": 0.0604, + "step": 28548 + }, + { + "epoch": 0.6692335191078137, + "grad_norm": 0.8265790343284607, + "learning_rate": 4.932321061400837e-05, + "loss": 0.1627, + "step": 28549 + }, + { + "epoch": 0.6692569606826186, + "grad_norm": 0.7814748287200928, + "learning_rate": 4.9316860515464623e-05, + "loss": 0.1649, + "step": 28550 + }, + { + "epoch": 0.6692804022574237, + "grad_norm": 0.556917130947113, + "learning_rate": 4.9310510691929e-05, + "loss": 0.1323, + "step": 28551 + }, + { + "epoch": 0.6693038438322286, + "grad_norm": 0.318017840385437, + "learning_rate": 4.9304161143436025e-05, + "loss": 0.0451, + "step": 28552 + }, + { + "epoch": 0.6693272854070337, + "grad_norm": 0.3931909203529358, + "learning_rate": 4.9297811870020084e-05, + "loss": 0.0275, + "step": 28553 + }, + { + "epoch": 0.6693507269818386, + "grad_norm": 0.503760039806366, + "learning_rate": 4.9291462871715647e-05, + "loss": 0.0677, + "step": 28554 + }, + { + "epoch": 0.6693741685566437, + "grad_norm": 0.3328442871570587, + "learning_rate": 4.9285114148557156e-05, + "loss": 0.0309, + "step": 28555 + }, + { + "epoch": 0.6693976101314486, + "grad_norm": 0.5780100226402283, + "learning_rate": 4.927876570057906e-05, + "loss": 0.1146, + "step": 28556 + }, + { + "epoch": 0.6694210517062537, + "grad_norm": 0.2992306351661682, + "learning_rate": 4.927241752781576e-05, + "loss": 0.0653, + "step": 28557 + }, + { + "epoch": 0.6694444932810586, + "grad_norm": 1.0859545469284058, + "learning_rate": 4.92660696303018e-05, + "loss": 0.298, + "step": 28558 + }, + { + "epoch": 0.6694679348558636, + "grad_norm": 0.38992980122566223, + "learning_rate": 4.925972200807157e-05, + "loss": 0.0398, + "step": 28559 + }, + { + "epoch": 0.6694913764306686, + "grad_norm": 0.6146215796470642, + "learning_rate": 4.925337466115952e-05, + "loss": 0.1284, + "step": 28560 + }, + { + "epoch": 0.6695148180054736, + "grad_norm": 0.15960894525051117, + "learning_rate": 4.9247027589600084e-05, + "loss": 0.0332, + "step": 28561 + }, + { + "epoch": 0.6695382595802786, + "grad_norm": 0.1398245394229889, + "learning_rate": 4.9240680793427664e-05, + "loss": 0.0223, + "step": 28562 + }, + { + "epoch": 0.6695617011550836, + "grad_norm": 0.29157519340515137, + "learning_rate": 4.923433427267673e-05, + "loss": 0.0599, + "step": 28563 + }, + { + "epoch": 0.6695851427298886, + "grad_norm": 0.2932252883911133, + "learning_rate": 4.9227988027381774e-05, + "loss": 0.0462, + "step": 28564 + }, + { + "epoch": 0.6696085843046936, + "grad_norm": 0.26237165927886963, + "learning_rate": 4.922164205757718e-05, + "loss": 0.0487, + "step": 28565 + }, + { + "epoch": 0.6696320258794985, + "grad_norm": 0.5873597264289856, + "learning_rate": 4.921529636329738e-05, + "loss": 0.0674, + "step": 28566 + }, + { + "epoch": 0.6696554674543036, + "grad_norm": 0.6830834746360779, + "learning_rate": 4.9208950944576813e-05, + "loss": 0.5569, + "step": 28567 + }, + { + "epoch": 0.6696789090291085, + "grad_norm": 0.37291419506073, + "learning_rate": 4.9202605801449906e-05, + "loss": 0.0565, + "step": 28568 + }, + { + "epoch": 0.6697023506039136, + "grad_norm": 0.7263683080673218, + "learning_rate": 4.919626093395104e-05, + "loss": 0.1902, + "step": 28569 + }, + { + "epoch": 0.6697257921787185, + "grad_norm": 0.49818670749664307, + "learning_rate": 4.918991634211475e-05, + "loss": 0.0878, + "step": 28570 + }, + { + "epoch": 0.6697492337535236, + "grad_norm": 0.4881179630756378, + "learning_rate": 4.918357202597539e-05, + "loss": 0.0604, + "step": 28571 + }, + { + "epoch": 0.6697726753283285, + "grad_norm": 0.46255069971084595, + "learning_rate": 4.9177227985567396e-05, + "loss": 0.0987, + "step": 28572 + }, + { + "epoch": 0.6697961169031336, + "grad_norm": 0.7907927632331848, + "learning_rate": 4.91708842209252e-05, + "loss": 0.183, + "step": 28573 + }, + { + "epoch": 0.6698195584779385, + "grad_norm": 0.38085946440696716, + "learning_rate": 4.91645407320832e-05, + "loss": 0.0737, + "step": 28574 + }, + { + "epoch": 0.6698430000527436, + "grad_norm": 0.5346066355705261, + "learning_rate": 4.91581975190758e-05, + "loss": 0.6204, + "step": 28575 + }, + { + "epoch": 0.6698664416275485, + "grad_norm": 0.7349157333374023, + "learning_rate": 4.915185458193746e-05, + "loss": 0.0963, + "step": 28576 + }, + { + "epoch": 0.6698898832023535, + "grad_norm": 0.32280269265174866, + "learning_rate": 4.914551192070262e-05, + "loss": 0.0594, + "step": 28577 + }, + { + "epoch": 0.6699133247771585, + "grad_norm": 0.13274629414081573, + "learning_rate": 4.913916953540565e-05, + "loss": 0.0255, + "step": 28578 + }, + { + "epoch": 0.6699367663519635, + "grad_norm": 0.09654557704925537, + "learning_rate": 4.913282742608098e-05, + "loss": 0.0193, + "step": 28579 + }, + { + "epoch": 0.6699602079267685, + "grad_norm": 0.7238373756408691, + "learning_rate": 4.912648559276304e-05, + "loss": 0.1424, + "step": 28580 + }, + { + "epoch": 0.6699836495015735, + "grad_norm": 0.3736684024333954, + "learning_rate": 4.912014403548619e-05, + "loss": 0.0962, + "step": 28581 + }, + { + "epoch": 0.6700070910763785, + "grad_norm": 0.052229706197977066, + "learning_rate": 4.9113802754284855e-05, + "loss": 0.0081, + "step": 28582 + }, + { + "epoch": 0.6700305326511835, + "grad_norm": 0.1393006592988968, + "learning_rate": 4.910746174919348e-05, + "loss": 0.0268, + "step": 28583 + }, + { + "epoch": 0.6700539742259884, + "grad_norm": 0.40604233741760254, + "learning_rate": 4.910112102024645e-05, + "loss": 0.0705, + "step": 28584 + }, + { + "epoch": 0.6700774158007935, + "grad_norm": 0.6732155680656433, + "learning_rate": 4.909478056747818e-05, + "loss": 0.1562, + "step": 28585 + }, + { + "epoch": 0.6701008573755985, + "grad_norm": 0.44659870862960815, + "learning_rate": 4.908844039092305e-05, + "loss": 0.0394, + "step": 28586 + }, + { + "epoch": 0.6701242989504035, + "grad_norm": 0.1504524052143097, + "learning_rate": 4.9082100490615444e-05, + "loss": 0.0285, + "step": 28587 + }, + { + "epoch": 0.6701477405252085, + "grad_norm": 0.5462537407875061, + "learning_rate": 4.9075760866589826e-05, + "loss": 0.062, + "step": 28588 + }, + { + "epoch": 0.6701711821000135, + "grad_norm": 0.3045935034751892, + "learning_rate": 4.9069421518880524e-05, + "loss": 0.0351, + "step": 28589 + }, + { + "epoch": 0.6701946236748185, + "grad_norm": 0.5108904838562012, + "learning_rate": 4.9063082447522016e-05, + "loss": 0.0527, + "step": 28590 + }, + { + "epoch": 0.6702180652496235, + "grad_norm": 0.707029402256012, + "learning_rate": 4.9056743652548645e-05, + "loss": 0.5559, + "step": 28591 + }, + { + "epoch": 0.6702415068244285, + "grad_norm": 0.4656253457069397, + "learning_rate": 4.905040513399481e-05, + "loss": 0.0366, + "step": 28592 + }, + { + "epoch": 0.6702649483992335, + "grad_norm": 0.4106595516204834, + "learning_rate": 4.904406689189492e-05, + "loss": 0.0749, + "step": 28593 + }, + { + "epoch": 0.6702883899740385, + "grad_norm": 0.3697616457939148, + "learning_rate": 4.9037728926283365e-05, + "loss": 0.0569, + "step": 28594 + }, + { + "epoch": 0.6703118315488434, + "grad_norm": 0.4717157185077667, + "learning_rate": 4.903139123719448e-05, + "loss": 0.1102, + "step": 28595 + }, + { + "epoch": 0.6703352731236485, + "grad_norm": 0.17247827351093292, + "learning_rate": 4.902505382466272e-05, + "loss": 0.0225, + "step": 28596 + }, + { + "epoch": 0.6703587146984534, + "grad_norm": 0.34271034598350525, + "learning_rate": 4.901871668872248e-05, + "loss": 0.0788, + "step": 28597 + }, + { + "epoch": 0.6703821562732585, + "grad_norm": 0.4733172357082367, + "learning_rate": 4.90123798294081e-05, + "loss": 0.0845, + "step": 28598 + }, + { + "epoch": 0.6704055978480634, + "grad_norm": 0.26458287239074707, + "learning_rate": 4.900604324675394e-05, + "loss": 0.0569, + "step": 28599 + }, + { + "epoch": 0.6704290394228685, + "grad_norm": 0.39497387409210205, + "learning_rate": 4.8999706940794474e-05, + "loss": 0.0305, + "step": 28600 + }, + { + "epoch": 0.6704524809976734, + "grad_norm": 0.38051837682724, + "learning_rate": 4.8993370911563985e-05, + "loss": 0.0357, + "step": 28601 + }, + { + "epoch": 0.6704759225724785, + "grad_norm": 0.6207098364830017, + "learning_rate": 4.898703515909694e-05, + "loss": 0.6361, + "step": 28602 + }, + { + "epoch": 0.6704993641472834, + "grad_norm": 0.2689967453479767, + "learning_rate": 4.898069968342768e-05, + "loss": 0.0608, + "step": 28603 + }, + { + "epoch": 0.6705228057220884, + "grad_norm": 0.44459789991378784, + "learning_rate": 4.8974364484590576e-05, + "loss": 0.0701, + "step": 28604 + }, + { + "epoch": 0.6705462472968934, + "grad_norm": 0.6355718970298767, + "learning_rate": 4.896802956262e-05, + "loss": 0.1712, + "step": 28605 + }, + { + "epoch": 0.6705696888716984, + "grad_norm": 0.3247922956943512, + "learning_rate": 4.896169491755033e-05, + "loss": 0.0551, + "step": 28606 + }, + { + "epoch": 0.6705931304465034, + "grad_norm": 0.3406846523284912, + "learning_rate": 4.8955360549415906e-05, + "loss": 0.0704, + "step": 28607 + }, + { + "epoch": 0.6706165720213084, + "grad_norm": 0.10046560317277908, + "learning_rate": 4.8949026458251167e-05, + "loss": 0.0172, + "step": 28608 + }, + { + "epoch": 0.6706400135961134, + "grad_norm": 0.6817405819892883, + "learning_rate": 4.8942692644090446e-05, + "loss": 0.0859, + "step": 28609 + }, + { + "epoch": 0.6706634551709184, + "grad_norm": 0.30149635672569275, + "learning_rate": 4.893635910696811e-05, + "loss": 0.0932, + "step": 28610 + }, + { + "epoch": 0.6706868967457233, + "grad_norm": 0.506761372089386, + "learning_rate": 4.893002584691848e-05, + "loss": 0.1113, + "step": 28611 + }, + { + "epoch": 0.6707103383205284, + "grad_norm": 0.3689015209674835, + "learning_rate": 4.892369286397601e-05, + "loss": 0.0677, + "step": 28612 + }, + { + "epoch": 0.6707337798953333, + "grad_norm": 0.4567647874355316, + "learning_rate": 4.891736015817503e-05, + "loss": 0.1085, + "step": 28613 + }, + { + "epoch": 0.6707572214701384, + "grad_norm": 0.45756328105926514, + "learning_rate": 4.8911027729549844e-05, + "loss": 0.0546, + "step": 28614 + }, + { + "epoch": 0.6707806630449433, + "grad_norm": 0.17761750519275665, + "learning_rate": 4.8904695578134883e-05, + "loss": 0.0353, + "step": 28615 + }, + { + "epoch": 0.6708041046197484, + "grad_norm": 0.4996294379234314, + "learning_rate": 4.88983637039645e-05, + "loss": 0.1011, + "step": 28616 + }, + { + "epoch": 0.6708275461945533, + "grad_norm": 0.2756860852241516, + "learning_rate": 4.8892032107073014e-05, + "loss": 0.0414, + "step": 28617 + }, + { + "epoch": 0.6708509877693584, + "grad_norm": 0.16417518258094788, + "learning_rate": 4.888570078749482e-05, + "loss": 0.0274, + "step": 28618 + }, + { + "epoch": 0.6708744293441633, + "grad_norm": 0.35372450947761536, + "learning_rate": 4.8879369745264236e-05, + "loss": 0.0647, + "step": 28619 + }, + { + "epoch": 0.6708978709189684, + "grad_norm": 0.3227350115776062, + "learning_rate": 4.88730389804156e-05, + "loss": 0.0509, + "step": 28620 + }, + { + "epoch": 0.6709213124937733, + "grad_norm": 0.13949130475521088, + "learning_rate": 4.886670849298331e-05, + "loss": 0.024, + "step": 28621 + }, + { + "epoch": 0.6709447540685783, + "grad_norm": 0.42811471223831177, + "learning_rate": 4.8860378283001714e-05, + "loss": 0.0727, + "step": 28622 + }, + { + "epoch": 0.6709681956433833, + "grad_norm": 0.2888771891593933, + "learning_rate": 4.885404835050514e-05, + "loss": 0.0981, + "step": 28623 + }, + { + "epoch": 0.6709916372181883, + "grad_norm": 0.21670059859752655, + "learning_rate": 4.884771869552791e-05, + "loss": 0.0324, + "step": 28624 + }, + { + "epoch": 0.6710150787929933, + "grad_norm": 0.3001973628997803, + "learning_rate": 4.884138931810442e-05, + "loss": 0.0285, + "step": 28625 + }, + { + "epoch": 0.6710385203677983, + "grad_norm": 0.3975062370300293, + "learning_rate": 4.8835060218269e-05, + "loss": 0.0746, + "step": 28626 + }, + { + "epoch": 0.6710619619426033, + "grad_norm": 0.6016857028007507, + "learning_rate": 4.882873139605594e-05, + "loss": 0.1046, + "step": 28627 + }, + { + "epoch": 0.6710854035174083, + "grad_norm": 0.343728631734848, + "learning_rate": 4.882240285149966e-05, + "loss": 0.0465, + "step": 28628 + }, + { + "epoch": 0.6711088450922132, + "grad_norm": 0.349038302898407, + "learning_rate": 4.8816074584634463e-05, + "loss": 0.0978, + "step": 28629 + }, + { + "epoch": 0.6711322866670183, + "grad_norm": 0.3560457229614258, + "learning_rate": 4.8809746595494676e-05, + "loss": 0.0615, + "step": 28630 + }, + { + "epoch": 0.6711557282418232, + "grad_norm": 0.4302406311035156, + "learning_rate": 4.880341888411466e-05, + "loss": 0.0988, + "step": 28631 + }, + { + "epoch": 0.6711791698166283, + "grad_norm": 0.0954551249742508, + "learning_rate": 4.8797091450528676e-05, + "loss": 0.0127, + "step": 28632 + }, + { + "epoch": 0.6712026113914332, + "grad_norm": 0.6851704120635986, + "learning_rate": 4.879076429477116e-05, + "loss": 0.5819, + "step": 28633 + }, + { + "epoch": 0.6712260529662383, + "grad_norm": 0.4968302249908447, + "learning_rate": 4.878443741687641e-05, + "loss": 0.1188, + "step": 28634 + }, + { + "epoch": 0.6712494945410432, + "grad_norm": 0.7522026896476746, + "learning_rate": 4.877811081687872e-05, + "loss": 0.069, + "step": 28635 + }, + { + "epoch": 0.6712729361158483, + "grad_norm": 0.2395457774400711, + "learning_rate": 4.877178449481241e-05, + "loss": 0.0351, + "step": 28636 + }, + { + "epoch": 0.6712963776906533, + "grad_norm": 0.8403981924057007, + "learning_rate": 4.8765458450711874e-05, + "loss": 0.5132, + "step": 28637 + }, + { + "epoch": 0.6713198192654583, + "grad_norm": 0.3189418315887451, + "learning_rate": 4.875913268461141e-05, + "loss": 0.0587, + "step": 28638 + }, + { + "epoch": 0.6713432608402633, + "grad_norm": 0.25012731552124023, + "learning_rate": 4.875280719654529e-05, + "loss": 0.0227, + "step": 28639 + }, + { + "epoch": 0.6713667024150682, + "grad_norm": 0.2401464432477951, + "learning_rate": 4.874648198654792e-05, + "loss": 0.0343, + "step": 28640 + }, + { + "epoch": 0.6713901439898733, + "grad_norm": 0.5912972688674927, + "learning_rate": 4.874015705465358e-05, + "loss": 0.1186, + "step": 28641 + }, + { + "epoch": 0.6714135855646782, + "grad_norm": 0.28419262170791626, + "learning_rate": 4.8733832400896586e-05, + "loss": 0.0281, + "step": 28642 + }, + { + "epoch": 0.6714370271394833, + "grad_norm": 0.10041435062885284, + "learning_rate": 4.8727508025311255e-05, + "loss": 0.0059, + "step": 28643 + }, + { + "epoch": 0.6714604687142882, + "grad_norm": 0.1718893200159073, + "learning_rate": 4.8721183927931916e-05, + "loss": 0.0206, + "step": 28644 + }, + { + "epoch": 0.6714839102890933, + "grad_norm": 0.4597967565059662, + "learning_rate": 4.8714860108792824e-05, + "loss": 0.0785, + "step": 28645 + }, + { + "epoch": 0.6715073518638982, + "grad_norm": 0.4254356920719147, + "learning_rate": 4.870853656792839e-05, + "loss": 0.0775, + "step": 28646 + }, + { + "epoch": 0.6715307934387033, + "grad_norm": 0.2510681748390198, + "learning_rate": 4.870221330537288e-05, + "loss": 0.0724, + "step": 28647 + }, + { + "epoch": 0.6715542350135082, + "grad_norm": 0.24532906711101532, + "learning_rate": 4.869589032116057e-05, + "loss": 0.0371, + "step": 28648 + }, + { + "epoch": 0.6715776765883132, + "grad_norm": 0.39619696140289307, + "learning_rate": 4.868956761532583e-05, + "loss": 0.3413, + "step": 28649 + }, + { + "epoch": 0.6716011181631182, + "grad_norm": 0.7089830040931702, + "learning_rate": 4.8683245187902934e-05, + "loss": 0.0726, + "step": 28650 + }, + { + "epoch": 0.6716245597379232, + "grad_norm": 0.6850772500038147, + "learning_rate": 4.86769230389262e-05, + "loss": 0.0582, + "step": 28651 + }, + { + "epoch": 0.6716480013127282, + "grad_norm": 0.7100622653961182, + "learning_rate": 4.867060116842988e-05, + "loss": 0.6453, + "step": 28652 + }, + { + "epoch": 0.6716714428875332, + "grad_norm": 0.17461487650871277, + "learning_rate": 4.866427957644838e-05, + "loss": 0.0393, + "step": 28653 + }, + { + "epoch": 0.6716948844623382, + "grad_norm": 0.7197893261909485, + "learning_rate": 4.8657958263015926e-05, + "loss": 0.1217, + "step": 28654 + }, + { + "epoch": 0.6717183260371432, + "grad_norm": 0.5221151113510132, + "learning_rate": 4.865163722816684e-05, + "loss": 0.0752, + "step": 28655 + }, + { + "epoch": 0.6717417676119481, + "grad_norm": 0.16651716828346252, + "learning_rate": 4.86453164719354e-05, + "loss": 0.0427, + "step": 28656 + }, + { + "epoch": 0.6717652091867532, + "grad_norm": 0.6494638323783875, + "learning_rate": 4.863899599435593e-05, + "loss": 0.1285, + "step": 28657 + }, + { + "epoch": 0.6717886507615581, + "grad_norm": 0.16498765349388123, + "learning_rate": 4.8632675795462665e-05, + "loss": 0.0372, + "step": 28658 + }, + { + "epoch": 0.6718120923363632, + "grad_norm": 0.3072676956653595, + "learning_rate": 4.8626355875289996e-05, + "loss": 0.0335, + "step": 28659 + }, + { + "epoch": 0.6718355339111681, + "grad_norm": 0.22614626586437225, + "learning_rate": 4.862003623387212e-05, + "loss": 0.0275, + "step": 28660 + }, + { + "epoch": 0.6718589754859732, + "grad_norm": 0.477251797914505, + "learning_rate": 4.861371687124341e-05, + "loss": 0.0632, + "step": 28661 + }, + { + "epoch": 0.6718824170607781, + "grad_norm": 0.33273833990097046, + "learning_rate": 4.860739778743811e-05, + "loss": 0.4698, + "step": 28662 + }, + { + "epoch": 0.6719058586355832, + "grad_norm": 0.5429359674453735, + "learning_rate": 4.8601078982490535e-05, + "loss": 0.0707, + "step": 28663 + }, + { + "epoch": 0.6719293002103881, + "grad_norm": 0.11180003732442856, + "learning_rate": 4.859476045643493e-05, + "loss": 0.0183, + "step": 28664 + }, + { + "epoch": 0.6719527417851932, + "grad_norm": 0.2707723081111908, + "learning_rate": 4.8588442209305574e-05, + "loss": 0.0552, + "step": 28665 + }, + { + "epoch": 0.6719761833599981, + "grad_norm": 0.42408430576324463, + "learning_rate": 4.8582124241136805e-05, + "loss": 0.0772, + "step": 28666 + }, + { + "epoch": 0.6719996249348031, + "grad_norm": 0.5259318947792053, + "learning_rate": 4.857580655196289e-05, + "loss": 0.0649, + "step": 28667 + }, + { + "epoch": 0.6720230665096081, + "grad_norm": 0.6201512813568115, + "learning_rate": 4.856948914181808e-05, + "loss": 0.1194, + "step": 28668 + }, + { + "epoch": 0.6720465080844131, + "grad_norm": 0.11957045644521713, + "learning_rate": 4.856317201073667e-05, + "loss": 0.0238, + "step": 28669 + }, + { + "epoch": 0.6720699496592181, + "grad_norm": 0.19964849948883057, + "learning_rate": 4.85568551587529e-05, + "loss": 0.0265, + "step": 28670 + }, + { + "epoch": 0.6720933912340231, + "grad_norm": 0.4230777621269226, + "learning_rate": 4.855053858590112e-05, + "loss": 0.1127, + "step": 28671 + }, + { + "epoch": 0.6721168328088281, + "grad_norm": 0.3156665861606598, + "learning_rate": 4.854422229221553e-05, + "loss": 0.0329, + "step": 28672 + }, + { + "epoch": 0.6721402743836331, + "grad_norm": 0.25410333275794983, + "learning_rate": 4.853790627773046e-05, + "loss": 0.0211, + "step": 28673 + }, + { + "epoch": 0.672163715958438, + "grad_norm": 0.43414634466171265, + "learning_rate": 4.8531590542480165e-05, + "loss": 0.5289, + "step": 28674 + }, + { + "epoch": 0.6721871575332431, + "grad_norm": 0.5048166513442993, + "learning_rate": 4.852527508649891e-05, + "loss": 0.0764, + "step": 28675 + }, + { + "epoch": 0.672210599108048, + "grad_norm": 0.11771021038293839, + "learning_rate": 4.851895990982096e-05, + "loss": 0.0192, + "step": 28676 + }, + { + "epoch": 0.6722340406828531, + "grad_norm": 0.4750308692455292, + "learning_rate": 4.851264501248054e-05, + "loss": 0.0974, + "step": 28677 + }, + { + "epoch": 0.672257482257658, + "grad_norm": 0.5204490423202515, + "learning_rate": 4.8506330394512e-05, + "loss": 0.1013, + "step": 28678 + }, + { + "epoch": 0.6722809238324631, + "grad_norm": 0.4841596782207489, + "learning_rate": 4.850001605594955e-05, + "loss": 0.0939, + "step": 28679 + }, + { + "epoch": 0.672304365407268, + "grad_norm": 0.6685733199119568, + "learning_rate": 4.849370199682747e-05, + "loss": 0.5676, + "step": 28680 + }, + { + "epoch": 0.6723278069820731, + "grad_norm": 0.5342115759849548, + "learning_rate": 4.8487388217180006e-05, + "loss": 0.1391, + "step": 28681 + }, + { + "epoch": 0.672351248556878, + "grad_norm": 0.13262677192687988, + "learning_rate": 4.848107471704142e-05, + "loss": 0.0181, + "step": 28682 + }, + { + "epoch": 0.672374690131683, + "grad_norm": 0.21347437798976898, + "learning_rate": 4.847476149644594e-05, + "loss": 0.0391, + "step": 28683 + }, + { + "epoch": 0.672398131706488, + "grad_norm": 0.5754510760307312, + "learning_rate": 4.8468448555427845e-05, + "loss": 0.5505, + "step": 28684 + }, + { + "epoch": 0.672421573281293, + "grad_norm": 0.6282299160957336, + "learning_rate": 4.846213589402144e-05, + "loss": 0.6752, + "step": 28685 + }, + { + "epoch": 0.672445014856098, + "grad_norm": 0.9596038460731506, + "learning_rate": 4.8455823512260936e-05, + "loss": 0.1527, + "step": 28686 + }, + { + "epoch": 0.672468456430903, + "grad_norm": 0.1087862104177475, + "learning_rate": 4.8449511410180585e-05, + "loss": 0.0178, + "step": 28687 + }, + { + "epoch": 0.6724918980057081, + "grad_norm": 0.349870502948761, + "learning_rate": 4.844319958781462e-05, + "loss": 0.0584, + "step": 28688 + }, + { + "epoch": 0.672515339580513, + "grad_norm": 0.21727435290813446, + "learning_rate": 4.8436888045197316e-05, + "loss": 0.0212, + "step": 28689 + }, + { + "epoch": 0.6725387811553181, + "grad_norm": 0.5477045774459839, + "learning_rate": 4.843057678236287e-05, + "loss": 0.0953, + "step": 28690 + }, + { + "epoch": 0.672562222730123, + "grad_norm": 0.24176515638828278, + "learning_rate": 4.842426579934559e-05, + "loss": 0.0379, + "step": 28691 + }, + { + "epoch": 0.672585664304928, + "grad_norm": 0.5733356475830078, + "learning_rate": 4.841795509617971e-05, + "loss": 0.1446, + "step": 28692 + }, + { + "epoch": 0.672609105879733, + "grad_norm": 0.19679822027683258, + "learning_rate": 4.841164467289945e-05, + "loss": 0.0347, + "step": 28693 + }, + { + "epoch": 0.672632547454538, + "grad_norm": 0.35993659496307373, + "learning_rate": 4.840533452953905e-05, + "loss": 0.0436, + "step": 28694 + }, + { + "epoch": 0.672655989029343, + "grad_norm": 0.3208659887313843, + "learning_rate": 4.8399024666132765e-05, + "loss": 0.0475, + "step": 28695 + }, + { + "epoch": 0.672679430604148, + "grad_norm": 0.6977130770683289, + "learning_rate": 4.8392715082714776e-05, + "loss": 0.6762, + "step": 28696 + }, + { + "epoch": 0.672702872178953, + "grad_norm": 0.401128888130188, + "learning_rate": 4.838640577931938e-05, + "loss": 0.7186, + "step": 28697 + }, + { + "epoch": 0.672726313753758, + "grad_norm": 0.08730019629001617, + "learning_rate": 4.838009675598082e-05, + "loss": 0.0129, + "step": 28698 + }, + { + "epoch": 0.672749755328563, + "grad_norm": 0.36802706122398376, + "learning_rate": 4.837378801273331e-05, + "loss": 0.0608, + "step": 28699 + }, + { + "epoch": 0.672773196903368, + "grad_norm": 1.2595267295837402, + "learning_rate": 4.836747954961108e-05, + "loss": 0.2735, + "step": 28700 + }, + { + "epoch": 0.672796638478173, + "grad_norm": 0.44357016682624817, + "learning_rate": 4.836117136664836e-05, + "loss": 0.0972, + "step": 28701 + }, + { + "epoch": 0.672820080052978, + "grad_norm": 0.12308670580387115, + "learning_rate": 4.835486346387939e-05, + "loss": 0.0112, + "step": 28702 + }, + { + "epoch": 0.6728435216277829, + "grad_norm": 0.34871751070022583, + "learning_rate": 4.8348555841338314e-05, + "loss": 0.0466, + "step": 28703 + }, + { + "epoch": 0.672866963202588, + "grad_norm": 0.19327066838741302, + "learning_rate": 4.8342248499059484e-05, + "loss": 0.0242, + "step": 28704 + }, + { + "epoch": 0.6728904047773929, + "grad_norm": 0.5107447504997253, + "learning_rate": 4.833594143707707e-05, + "loss": 0.1191, + "step": 28705 + }, + { + "epoch": 0.672913846352198, + "grad_norm": 0.3114562928676605, + "learning_rate": 4.8329634655425284e-05, + "loss": 0.0461, + "step": 28706 + }, + { + "epoch": 0.6729372879270029, + "grad_norm": 0.35452744364738464, + "learning_rate": 4.832332815413835e-05, + "loss": 0.0759, + "step": 28707 + }, + { + "epoch": 0.672960729501808, + "grad_norm": 0.26191920042037964, + "learning_rate": 4.831702193325045e-05, + "loss": 0.0605, + "step": 28708 + }, + { + "epoch": 0.6729841710766129, + "grad_norm": 0.41006818413734436, + "learning_rate": 4.8310715992795885e-05, + "loss": 0.1115, + "step": 28709 + }, + { + "epoch": 0.673007612651418, + "grad_norm": 0.5588099360466003, + "learning_rate": 4.830441033280879e-05, + "loss": 0.1092, + "step": 28710 + }, + { + "epoch": 0.6730310542262229, + "grad_norm": 0.23124350607395172, + "learning_rate": 4.8298104953323455e-05, + "loss": 0.0279, + "step": 28711 + }, + { + "epoch": 0.6730544958010279, + "grad_norm": 0.4754113256931305, + "learning_rate": 4.829179985437407e-05, + "loss": 0.0513, + "step": 28712 + }, + { + "epoch": 0.6730779373758329, + "grad_norm": 0.3641429841518402, + "learning_rate": 4.828549503599481e-05, + "loss": 0.0192, + "step": 28713 + }, + { + "epoch": 0.6731013789506379, + "grad_norm": 0.1707412600517273, + "learning_rate": 4.827919049821992e-05, + "loss": 0.0227, + "step": 28714 + }, + { + "epoch": 0.6731248205254429, + "grad_norm": 0.1146281510591507, + "learning_rate": 4.827288624108356e-05, + "loss": 0.0164, + "step": 28715 + }, + { + "epoch": 0.6731482621002479, + "grad_norm": 0.21657784283161163, + "learning_rate": 4.8266582264620005e-05, + "loss": 0.0457, + "step": 28716 + }, + { + "epoch": 0.6731717036750529, + "grad_norm": 0.33283594250679016, + "learning_rate": 4.826027856886344e-05, + "loss": 0.0537, + "step": 28717 + }, + { + "epoch": 0.6731951452498579, + "grad_norm": 0.1334344446659088, + "learning_rate": 4.825397515384804e-05, + "loss": 0.0133, + "step": 28718 + }, + { + "epoch": 0.6732185868246628, + "grad_norm": 0.5029234290122986, + "learning_rate": 4.824767201960804e-05, + "loss": 0.1019, + "step": 28719 + }, + { + "epoch": 0.6732420283994679, + "grad_norm": 0.41502198576927185, + "learning_rate": 4.8241369166177616e-05, + "loss": 0.0599, + "step": 28720 + }, + { + "epoch": 0.6732654699742728, + "grad_norm": 0.23555149137973785, + "learning_rate": 4.823506659359096e-05, + "loss": 0.0438, + "step": 28721 + }, + { + "epoch": 0.6732889115490779, + "grad_norm": 0.07944568991661072, + "learning_rate": 4.822876430188227e-05, + "loss": 0.0104, + "step": 28722 + }, + { + "epoch": 0.6733123531238828, + "grad_norm": 0.3904203772544861, + "learning_rate": 4.8222462291085804e-05, + "loss": 0.042, + "step": 28723 + }, + { + "epoch": 0.6733357946986879, + "grad_norm": 0.07662945240736008, + "learning_rate": 4.8216160561235715e-05, + "loss": 0.0147, + "step": 28724 + }, + { + "epoch": 0.6733592362734928, + "grad_norm": 0.5019428133964539, + "learning_rate": 4.820985911236619e-05, + "loss": 0.0204, + "step": 28725 + }, + { + "epoch": 0.6733826778482979, + "grad_norm": 0.39714905619621277, + "learning_rate": 4.820355794451143e-05, + "loss": 0.0539, + "step": 28726 + }, + { + "epoch": 0.6734061194231028, + "grad_norm": 0.7986775636672974, + "learning_rate": 4.8197257057705626e-05, + "loss": 0.0648, + "step": 28727 + }, + { + "epoch": 0.6734295609979079, + "grad_norm": 0.6980736255645752, + "learning_rate": 4.819095645198291e-05, + "loss": 0.1092, + "step": 28728 + }, + { + "epoch": 0.6734530025727128, + "grad_norm": 0.6373505592346191, + "learning_rate": 4.818465612737758e-05, + "loss": 0.1264, + "step": 28729 + }, + { + "epoch": 0.6734764441475178, + "grad_norm": 0.4962691068649292, + "learning_rate": 4.8178356083923746e-05, + "loss": 0.0997, + "step": 28730 + }, + { + "epoch": 0.6734998857223228, + "grad_norm": 0.5397667288780212, + "learning_rate": 4.817205632165561e-05, + "loss": 0.7101, + "step": 28731 + }, + { + "epoch": 0.6735233272971278, + "grad_norm": 0.5825006365776062, + "learning_rate": 4.8165756840607354e-05, + "loss": 0.6614, + "step": 28732 + }, + { + "epoch": 0.6735467688719328, + "grad_norm": 0.28223589062690735, + "learning_rate": 4.8159457640813135e-05, + "loss": 0.0667, + "step": 28733 + }, + { + "epoch": 0.6735702104467378, + "grad_norm": 0.44626176357269287, + "learning_rate": 4.815315872230718e-05, + "loss": 0.0774, + "step": 28734 + }, + { + "epoch": 0.6735936520215428, + "grad_norm": 0.2169940322637558, + "learning_rate": 4.81468600851236e-05, + "loss": 0.0512, + "step": 28735 + }, + { + "epoch": 0.6736170935963478, + "grad_norm": 0.17291657626628876, + "learning_rate": 4.814056172929667e-05, + "loss": 0.0188, + "step": 28736 + }, + { + "epoch": 0.6736405351711527, + "grad_norm": 0.45248833298683167, + "learning_rate": 4.81342636548605e-05, + "loss": 0.4416, + "step": 28737 + }, + { + "epoch": 0.6736639767459578, + "grad_norm": 0.2840931713581085, + "learning_rate": 4.8127965861849276e-05, + "loss": 0.0412, + "step": 28738 + }, + { + "epoch": 0.6736874183207628, + "grad_norm": 0.13756783306598663, + "learning_rate": 4.8121668350297166e-05, + "loss": 0.0179, + "step": 28739 + }, + { + "epoch": 0.6737108598955678, + "grad_norm": 0.05304621905088425, + "learning_rate": 4.811537112023835e-05, + "loss": 0.0087, + "step": 28740 + }, + { + "epoch": 0.6737343014703728, + "grad_norm": 0.583335816860199, + "learning_rate": 4.810907417170693e-05, + "loss": 0.738, + "step": 28741 + }, + { + "epoch": 0.6737577430451778, + "grad_norm": 0.45139482617378235, + "learning_rate": 4.8102777504737196e-05, + "loss": 0.3811, + "step": 28742 + }, + { + "epoch": 0.6737811846199828, + "grad_norm": 0.13657565414905548, + "learning_rate": 4.809648111936324e-05, + "loss": 0.0125, + "step": 28743 + }, + { + "epoch": 0.6738046261947878, + "grad_norm": 0.43804478645324707, + "learning_rate": 4.8090185015619235e-05, + "loss": 0.0863, + "step": 28744 + }, + { + "epoch": 0.6738280677695928, + "grad_norm": 0.3399201035499573, + "learning_rate": 4.808388919353931e-05, + "loss": 0.4661, + "step": 28745 + }, + { + "epoch": 0.6738515093443977, + "grad_norm": 0.6255207061767578, + "learning_rate": 4.8077593653157707e-05, + "loss": 0.1333, + "step": 28746 + }, + { + "epoch": 0.6738749509192028, + "grad_norm": 0.4710705280303955, + "learning_rate": 4.8071298394508525e-05, + "loss": 0.1105, + "step": 28747 + }, + { + "epoch": 0.6738983924940077, + "grad_norm": 0.37048575282096863, + "learning_rate": 4.806500341762591e-05, + "loss": 0.101, + "step": 28748 + }, + { + "epoch": 0.6739218340688128, + "grad_norm": 0.3506886959075928, + "learning_rate": 4.805870872254409e-05, + "loss": 0.0737, + "step": 28749 + }, + { + "epoch": 0.6739452756436177, + "grad_norm": 0.4861688017845154, + "learning_rate": 4.805241430929717e-05, + "loss": 0.0819, + "step": 28750 + }, + { + "epoch": 0.6739687172184228, + "grad_norm": 0.5107505917549133, + "learning_rate": 4.8046120177919307e-05, + "loss": 0.5745, + "step": 28751 + }, + { + "epoch": 0.6739921587932277, + "grad_norm": 0.10883653163909912, + "learning_rate": 4.803982632844466e-05, + "loss": 0.009, + "step": 28752 + }, + { + "epoch": 0.6740156003680328, + "grad_norm": 0.6195923089981079, + "learning_rate": 4.803353276090734e-05, + "loss": 0.7829, + "step": 28753 + }, + { + "epoch": 0.6740390419428377, + "grad_norm": 0.3789694607257843, + "learning_rate": 4.802723947534157e-05, + "loss": 0.0495, + "step": 28754 + }, + { + "epoch": 0.6740624835176428, + "grad_norm": 0.9705666899681091, + "learning_rate": 4.802094647178146e-05, + "loss": 0.2644, + "step": 28755 + }, + { + "epoch": 0.6740859250924477, + "grad_norm": 0.4875200688838959, + "learning_rate": 4.801465375026115e-05, + "loss": 0.0868, + "step": 28756 + }, + { + "epoch": 0.6741093666672527, + "grad_norm": 0.5698903203010559, + "learning_rate": 4.8008361310814755e-05, + "loss": 0.5974, + "step": 28757 + }, + { + "epoch": 0.6741328082420577, + "grad_norm": 0.3776206970214844, + "learning_rate": 4.800206915347649e-05, + "loss": 0.0349, + "step": 28758 + }, + { + "epoch": 0.6741562498168627, + "grad_norm": 0.19353333115577698, + "learning_rate": 4.799577727828047e-05, + "loss": 0.0228, + "step": 28759 + }, + { + "epoch": 0.6741796913916677, + "grad_norm": 0.3427525460720062, + "learning_rate": 4.7989485685260785e-05, + "loss": 0.0533, + "step": 28760 + }, + { + "epoch": 0.6742031329664727, + "grad_norm": 0.28866124153137207, + "learning_rate": 4.7983194374451645e-05, + "loss": 0.0566, + "step": 28761 + }, + { + "epoch": 0.6742265745412777, + "grad_norm": 0.4243490695953369, + "learning_rate": 4.797690334588716e-05, + "loss": 0.0677, + "step": 28762 + }, + { + "epoch": 0.6742500161160827, + "grad_norm": 0.3718085289001465, + "learning_rate": 4.797061259960145e-05, + "loss": 0.0501, + "step": 28763 + }, + { + "epoch": 0.6742734576908876, + "grad_norm": 0.510513961315155, + "learning_rate": 4.796432213562867e-05, + "loss": 0.13, + "step": 28764 + }, + { + "epoch": 0.6742968992656927, + "grad_norm": 0.8996477723121643, + "learning_rate": 4.795803195400294e-05, + "loss": 0.1729, + "step": 28765 + }, + { + "epoch": 0.6743203408404976, + "grad_norm": 0.4742629826068878, + "learning_rate": 4.795174205475834e-05, + "loss": 0.0704, + "step": 28766 + }, + { + "epoch": 0.6743437824153027, + "grad_norm": 0.3412352204322815, + "learning_rate": 4.7945452437929104e-05, + "loss": 0.0659, + "step": 28767 + }, + { + "epoch": 0.6743672239901076, + "grad_norm": 0.39049363136291504, + "learning_rate": 4.79391631035493e-05, + "loss": 0.0567, + "step": 28768 + }, + { + "epoch": 0.6743906655649127, + "grad_norm": 0.2301558405160904, + "learning_rate": 4.793287405165303e-05, + "loss": 0.0448, + "step": 28769 + }, + { + "epoch": 0.6744141071397176, + "grad_norm": 0.23548150062561035, + "learning_rate": 4.7926585282274484e-05, + "loss": 0.0337, + "step": 28770 + }, + { + "epoch": 0.6744375487145227, + "grad_norm": 0.1721247434616089, + "learning_rate": 4.792029679544775e-05, + "loss": 0.011, + "step": 28771 + }, + { + "epoch": 0.6744609902893276, + "grad_norm": 0.3454623520374298, + "learning_rate": 4.791400859120695e-05, + "loss": 0.0379, + "step": 28772 + }, + { + "epoch": 0.6744844318641326, + "grad_norm": 0.5071001648902893, + "learning_rate": 4.790772066958616e-05, + "loss": 0.0854, + "step": 28773 + }, + { + "epoch": 0.6745078734389376, + "grad_norm": 0.5937814712524414, + "learning_rate": 4.790143303061958e-05, + "loss": 0.5312, + "step": 28774 + }, + { + "epoch": 0.6745313150137426, + "grad_norm": 0.4943859279155731, + "learning_rate": 4.7895145674341305e-05, + "loss": 0.077, + "step": 28775 + }, + { + "epoch": 0.6745547565885476, + "grad_norm": 0.1863768845796585, + "learning_rate": 4.788885860078541e-05, + "loss": 0.0369, + "step": 28776 + }, + { + "epoch": 0.6745781981633526, + "grad_norm": 0.28207385540008545, + "learning_rate": 4.788257180998605e-05, + "loss": 0.0507, + "step": 28777 + }, + { + "epoch": 0.6746016397381576, + "grad_norm": 0.3584989607334137, + "learning_rate": 4.787628530197732e-05, + "loss": 0.0828, + "step": 28778 + }, + { + "epoch": 0.6746250813129626, + "grad_norm": 0.2457929402589798, + "learning_rate": 4.786999907679329e-05, + "loss": 0.0616, + "step": 28779 + }, + { + "epoch": 0.6746485228877676, + "grad_norm": 0.47728151082992554, + "learning_rate": 4.786371313446816e-05, + "loss": 0.2632, + "step": 28780 + }, + { + "epoch": 0.6746719644625726, + "grad_norm": 0.4955573081970215, + "learning_rate": 4.785742747503598e-05, + "loss": 0.0433, + "step": 28781 + }, + { + "epoch": 0.6746954060373775, + "grad_norm": 0.14338429272174835, + "learning_rate": 4.7851142098530825e-05, + "loss": 0.0139, + "step": 28782 + }, + { + "epoch": 0.6747188476121826, + "grad_norm": 0.4809635281562805, + "learning_rate": 4.784485700498688e-05, + "loss": 0.0699, + "step": 28783 + }, + { + "epoch": 0.6747422891869875, + "grad_norm": 0.4453854560852051, + "learning_rate": 4.7838572194438205e-05, + "loss": 0.0661, + "step": 28784 + }, + { + "epoch": 0.6747657307617926, + "grad_norm": 0.3314657211303711, + "learning_rate": 4.7832287666918896e-05, + "loss": 0.0289, + "step": 28785 + }, + { + "epoch": 0.6747891723365975, + "grad_norm": 0.3548530340194702, + "learning_rate": 4.7826003422463036e-05, + "loss": 0.0648, + "step": 28786 + }, + { + "epoch": 0.6748126139114026, + "grad_norm": 0.5485713481903076, + "learning_rate": 4.7819719461104785e-05, + "loss": 0.1345, + "step": 28787 + }, + { + "epoch": 0.6748360554862075, + "grad_norm": 0.17701201140880585, + "learning_rate": 4.7813435782878205e-05, + "loss": 0.0358, + "step": 28788 + }, + { + "epoch": 0.6748594970610126, + "grad_norm": 0.22712868452072144, + "learning_rate": 4.7807152387817376e-05, + "loss": 0.0352, + "step": 28789 + }, + { + "epoch": 0.6748829386358176, + "grad_norm": 0.50874263048172, + "learning_rate": 4.7800869275956416e-05, + "loss": 0.1441, + "step": 28790 + }, + { + "epoch": 0.6749063802106225, + "grad_norm": 0.5276727080345154, + "learning_rate": 4.779458644732937e-05, + "loss": 0.0615, + "step": 28791 + }, + { + "epoch": 0.6749298217854276, + "grad_norm": 0.37880828976631165, + "learning_rate": 4.77883039019704e-05, + "loss": 0.3464, + "step": 28792 + }, + { + "epoch": 0.6749532633602325, + "grad_norm": 0.33912569284439087, + "learning_rate": 4.7782021639913566e-05, + "loss": 0.0899, + "step": 28793 + }, + { + "epoch": 0.6749767049350376, + "grad_norm": 0.3434385359287262, + "learning_rate": 4.7775739661192906e-05, + "loss": 0.0573, + "step": 28794 + }, + { + "epoch": 0.6750001465098425, + "grad_norm": 0.49414804577827454, + "learning_rate": 4.776945796584259e-05, + "loss": 0.1237, + "step": 28795 + }, + { + "epoch": 0.6750235880846476, + "grad_norm": 0.3071519732475281, + "learning_rate": 4.776317655389667e-05, + "loss": 0.0571, + "step": 28796 + }, + { + "epoch": 0.6750470296594525, + "grad_norm": 1.4946107864379883, + "learning_rate": 4.775689542538921e-05, + "loss": 0.1337, + "step": 28797 + }, + { + "epoch": 0.6750704712342576, + "grad_norm": 0.3670696020126343, + "learning_rate": 4.775061458035428e-05, + "loss": 0.0453, + "step": 28798 + }, + { + "epoch": 0.6750939128090625, + "grad_norm": 0.20612551271915436, + "learning_rate": 4.774433401882602e-05, + "loss": 0.0467, + "step": 28799 + }, + { + "epoch": 0.6751173543838676, + "grad_norm": 0.2242586612701416, + "learning_rate": 4.7738053740838476e-05, + "loss": 0.0338, + "step": 28800 + }, + { + "epoch": 0.6751407959586725, + "grad_norm": 0.5851647853851318, + "learning_rate": 4.773177374642571e-05, + "loss": 0.0705, + "step": 28801 + }, + { + "epoch": 0.6751642375334775, + "grad_norm": 0.5727745294570923, + "learning_rate": 4.772549403562182e-05, + "loss": 0.1265, + "step": 28802 + }, + { + "epoch": 0.6751876791082825, + "grad_norm": 0.9377263784408569, + "learning_rate": 4.771921460846086e-05, + "loss": 0.12, + "step": 28803 + }, + { + "epoch": 0.6752111206830875, + "grad_norm": 0.589060366153717, + "learning_rate": 4.7712935464976884e-05, + "loss": 0.1027, + "step": 28804 + }, + { + "epoch": 0.6752345622578925, + "grad_norm": 0.6335610747337341, + "learning_rate": 4.770665660520401e-05, + "loss": 0.5421, + "step": 28805 + }, + { + "epoch": 0.6752580038326975, + "grad_norm": 0.16655214130878448, + "learning_rate": 4.7700378029176275e-05, + "loss": 0.0234, + "step": 28806 + }, + { + "epoch": 0.6752814454075025, + "grad_norm": 0.15232129395008087, + "learning_rate": 4.769409973692778e-05, + "loss": 0.0202, + "step": 28807 + }, + { + "epoch": 0.6753048869823075, + "grad_norm": 0.7939947247505188, + "learning_rate": 4.7687821728492575e-05, + "loss": 0.0681, + "step": 28808 + }, + { + "epoch": 0.6753283285571124, + "grad_norm": 0.5043224096298218, + "learning_rate": 4.768154400390472e-05, + "loss": 0.0867, + "step": 28809 + }, + { + "epoch": 0.6753517701319175, + "grad_norm": 0.3413265347480774, + "learning_rate": 4.7675266563198286e-05, + "loss": 0.0587, + "step": 28810 + }, + { + "epoch": 0.6753752117067224, + "grad_norm": 0.13291941583156586, + "learning_rate": 4.7668989406407286e-05, + "loss": 0.0155, + "step": 28811 + }, + { + "epoch": 0.6753986532815275, + "grad_norm": 0.45097044110298157, + "learning_rate": 4.766271253356585e-05, + "loss": 0.0816, + "step": 28812 + }, + { + "epoch": 0.6754220948563324, + "grad_norm": 0.4220118224620819, + "learning_rate": 4.765643594470801e-05, + "loss": 0.1007, + "step": 28813 + }, + { + "epoch": 0.6754455364311375, + "grad_norm": 0.5886841416358948, + "learning_rate": 4.765015963986782e-05, + "loss": 0.1693, + "step": 28814 + }, + { + "epoch": 0.6754689780059424, + "grad_norm": 0.44422149658203125, + "learning_rate": 4.7643883619079345e-05, + "loss": 0.056, + "step": 28815 + }, + { + "epoch": 0.6754924195807475, + "grad_norm": 0.39588218927383423, + "learning_rate": 4.763760788237661e-05, + "loss": 0.0527, + "step": 28816 + }, + { + "epoch": 0.6755158611555524, + "grad_norm": 0.32862961292266846, + "learning_rate": 4.763133242979367e-05, + "loss": 0.0412, + "step": 28817 + }, + { + "epoch": 0.6755393027303574, + "grad_norm": 0.17427927255630493, + "learning_rate": 4.762505726136458e-05, + "loss": 0.0233, + "step": 28818 + }, + { + "epoch": 0.6755627443051624, + "grad_norm": 0.20634892582893372, + "learning_rate": 4.761878237712345e-05, + "loss": 0.0254, + "step": 28819 + }, + { + "epoch": 0.6755861858799674, + "grad_norm": 0.6014614701271057, + "learning_rate": 4.7612507777104277e-05, + "loss": 0.0845, + "step": 28820 + }, + { + "epoch": 0.6756096274547724, + "grad_norm": 0.499713271856308, + "learning_rate": 4.76062334613411e-05, + "loss": 0.5702, + "step": 28821 + }, + { + "epoch": 0.6756330690295774, + "grad_norm": 0.3458818197250366, + "learning_rate": 4.759995942986798e-05, + "loss": 0.0888, + "step": 28822 + }, + { + "epoch": 0.6756565106043824, + "grad_norm": 0.6105064749717712, + "learning_rate": 4.759368568271896e-05, + "loss": 0.1032, + "step": 28823 + }, + { + "epoch": 0.6756799521791874, + "grad_norm": 0.5200347304344177, + "learning_rate": 4.758741221992803e-05, + "loss": 0.1256, + "step": 28824 + }, + { + "epoch": 0.6757033937539924, + "grad_norm": 0.2821417450904846, + "learning_rate": 4.758113904152931e-05, + "loss": 0.0613, + "step": 28825 + }, + { + "epoch": 0.6757268353287974, + "grad_norm": 0.11341479420661926, + "learning_rate": 4.757486614755681e-05, + "loss": 0.0155, + "step": 28826 + }, + { + "epoch": 0.6757502769036023, + "grad_norm": 0.3491863012313843, + "learning_rate": 4.756859353804456e-05, + "loss": 0.0372, + "step": 28827 + }, + { + "epoch": 0.6757737184784074, + "grad_norm": 0.2697318494319916, + "learning_rate": 4.7562321213026585e-05, + "loss": 0.0493, + "step": 28828 + }, + { + "epoch": 0.6757971600532123, + "grad_norm": 0.223122239112854, + "learning_rate": 4.7556049172536895e-05, + "loss": 0.0342, + "step": 28829 + }, + { + "epoch": 0.6758206016280174, + "grad_norm": 0.24879090487957, + "learning_rate": 4.754977741660961e-05, + "loss": 0.0434, + "step": 28830 + }, + { + "epoch": 0.6758440432028223, + "grad_norm": 0.5031630992889404, + "learning_rate": 4.7543505945278656e-05, + "loss": 0.0804, + "step": 28831 + }, + { + "epoch": 0.6758674847776274, + "grad_norm": 0.5514788031578064, + "learning_rate": 4.753723475857815e-05, + "loss": 0.1228, + "step": 28832 + }, + { + "epoch": 0.6758909263524323, + "grad_norm": 0.5206978917121887, + "learning_rate": 4.7530963856542086e-05, + "loss": 0.0596, + "step": 28833 + }, + { + "epoch": 0.6759143679272374, + "grad_norm": 0.35306739807128906, + "learning_rate": 4.75246932392045e-05, + "loss": 0.0461, + "step": 28834 + }, + { + "epoch": 0.6759378095020423, + "grad_norm": 0.4897594153881073, + "learning_rate": 4.751842290659939e-05, + "loss": 0.091, + "step": 28835 + }, + { + "epoch": 0.6759612510768473, + "grad_norm": 0.3550621569156647, + "learning_rate": 4.751215285876076e-05, + "loss": 0.0576, + "step": 28836 + }, + { + "epoch": 0.6759846926516523, + "grad_norm": 0.5076864361763, + "learning_rate": 4.75058830957227e-05, + "loss": 0.0975, + "step": 28837 + }, + { + "epoch": 0.6760081342264573, + "grad_norm": 0.27077576518058777, + "learning_rate": 4.749961361751919e-05, + "loss": 0.0441, + "step": 28838 + }, + { + "epoch": 0.6760315758012623, + "grad_norm": 0.4424635171890259, + "learning_rate": 4.749334442418426e-05, + "loss": 0.0752, + "step": 28839 + }, + { + "epoch": 0.6760550173760673, + "grad_norm": 0.2177126407623291, + "learning_rate": 4.7487075515751924e-05, + "loss": 0.0366, + "step": 28840 + }, + { + "epoch": 0.6760784589508724, + "grad_norm": 0.20986883342266083, + "learning_rate": 4.748080689225618e-05, + "loss": 0.0168, + "step": 28841 + }, + { + "epoch": 0.6761019005256773, + "grad_norm": 0.11785594373941422, + "learning_rate": 4.7474538553731016e-05, + "loss": 0.0131, + "step": 28842 + }, + { + "epoch": 0.6761253421004824, + "grad_norm": 0.16106893122196198, + "learning_rate": 4.74682705002105e-05, + "loss": 0.021, + "step": 28843 + }, + { + "epoch": 0.6761487836752873, + "grad_norm": 0.21594128012657166, + "learning_rate": 4.746200273172865e-05, + "loss": 0.0306, + "step": 28844 + }, + { + "epoch": 0.6761722252500924, + "grad_norm": 0.17310115694999695, + "learning_rate": 4.7455735248319455e-05, + "loss": 0.0137, + "step": 28845 + }, + { + "epoch": 0.6761956668248973, + "grad_norm": 0.31394270062446594, + "learning_rate": 4.744946805001691e-05, + "loss": 0.0574, + "step": 28846 + }, + { + "epoch": 0.6762191083997023, + "grad_norm": 0.21866168081760406, + "learning_rate": 4.744320113685504e-05, + "loss": 0.0362, + "step": 28847 + }, + { + "epoch": 0.6762425499745073, + "grad_norm": 0.43523740768432617, + "learning_rate": 4.7436934508867836e-05, + "loss": 0.0294, + "step": 28848 + }, + { + "epoch": 0.6762659915493123, + "grad_norm": 0.1689285933971405, + "learning_rate": 4.7430668166089255e-05, + "loss": 0.0141, + "step": 28849 + }, + { + "epoch": 0.6762894331241173, + "grad_norm": 0.4311099648475647, + "learning_rate": 4.74244021085534e-05, + "loss": 0.1024, + "step": 28850 + }, + { + "epoch": 0.6763128746989223, + "grad_norm": 0.23915442824363708, + "learning_rate": 4.74181363362942e-05, + "loss": 0.0211, + "step": 28851 + }, + { + "epoch": 0.6763363162737273, + "grad_norm": 0.5586560368537903, + "learning_rate": 4.741187084934568e-05, + "loss": 0.0617, + "step": 28852 + }, + { + "epoch": 0.6763597578485323, + "grad_norm": 0.2943570613861084, + "learning_rate": 4.7405605647741835e-05, + "loss": 0.0577, + "step": 28853 + }, + { + "epoch": 0.6763831994233372, + "grad_norm": 0.19018150866031647, + "learning_rate": 4.739934073151664e-05, + "loss": 0.0485, + "step": 28854 + }, + { + "epoch": 0.6764066409981423, + "grad_norm": 0.5150271058082581, + "learning_rate": 4.739307610070407e-05, + "loss": 0.0631, + "step": 28855 + }, + { + "epoch": 0.6764300825729472, + "grad_norm": 0.34551045298576355, + "learning_rate": 4.738681175533814e-05, + "loss": 0.0801, + "step": 28856 + }, + { + "epoch": 0.6764535241477523, + "grad_norm": 0.6677862405776978, + "learning_rate": 4.7380547695452905e-05, + "loss": 0.0798, + "step": 28857 + }, + { + "epoch": 0.6764769657225572, + "grad_norm": 0.3759819269180298, + "learning_rate": 4.737428392108229e-05, + "loss": 0.1176, + "step": 28858 + }, + { + "epoch": 0.6765004072973623, + "grad_norm": 0.40817880630493164, + "learning_rate": 4.7368020432260286e-05, + "loss": 0.0974, + "step": 28859 + }, + { + "epoch": 0.6765238488721672, + "grad_norm": 0.5985540747642517, + "learning_rate": 4.736175722902088e-05, + "loss": 0.0935, + "step": 28860 + }, + { + "epoch": 0.6765472904469723, + "grad_norm": 0.43284696340560913, + "learning_rate": 4.7355494311398064e-05, + "loss": 0.1009, + "step": 28861 + }, + { + "epoch": 0.6765707320217772, + "grad_norm": 0.5649815201759338, + "learning_rate": 4.7349231679425776e-05, + "loss": 0.1973, + "step": 28862 + }, + { + "epoch": 0.6765941735965822, + "grad_norm": 0.8019556403160095, + "learning_rate": 4.734296933313808e-05, + "loss": 0.1077, + "step": 28863 + }, + { + "epoch": 0.6766176151713872, + "grad_norm": 0.40859559178352356, + "learning_rate": 4.73367072725689e-05, + "loss": 0.074, + "step": 28864 + }, + { + "epoch": 0.6766410567461922, + "grad_norm": 0.1267879158258438, + "learning_rate": 4.7330445497752243e-05, + "loss": 0.0282, + "step": 28865 + }, + { + "epoch": 0.6766644983209972, + "grad_norm": 0.2631258964538574, + "learning_rate": 4.732418400872206e-05, + "loss": 0.0476, + "step": 28866 + }, + { + "epoch": 0.6766879398958022, + "grad_norm": 0.6184464693069458, + "learning_rate": 4.7317922805512304e-05, + "loss": 0.0343, + "step": 28867 + }, + { + "epoch": 0.6767113814706072, + "grad_norm": 0.9073468446731567, + "learning_rate": 4.731166188815701e-05, + "loss": 0.1547, + "step": 28868 + }, + { + "epoch": 0.6767348230454122, + "grad_norm": 0.3316599130630493, + "learning_rate": 4.730540125669009e-05, + "loss": 0.0631, + "step": 28869 + }, + { + "epoch": 0.6767582646202172, + "grad_norm": 0.4267377555370331, + "learning_rate": 4.729914091114558e-05, + "loss": 0.1019, + "step": 28870 + }, + { + "epoch": 0.6767817061950222, + "grad_norm": 0.4264541268348694, + "learning_rate": 4.7292880851557405e-05, + "loss": 0.0333, + "step": 28871 + }, + { + "epoch": 0.6768051477698271, + "grad_norm": 0.07644341886043549, + "learning_rate": 4.728662107795956e-05, + "loss": 0.0115, + "step": 28872 + }, + { + "epoch": 0.6768285893446322, + "grad_norm": 0.6808376908302307, + "learning_rate": 4.728036159038597e-05, + "loss": 0.1249, + "step": 28873 + }, + { + "epoch": 0.6768520309194371, + "grad_norm": 0.3239646553993225, + "learning_rate": 4.72741023888706e-05, + "loss": 0.0519, + "step": 28874 + }, + { + "epoch": 0.6768754724942422, + "grad_norm": 0.18034082651138306, + "learning_rate": 4.726784347344746e-05, + "loss": 0.0436, + "step": 28875 + }, + { + "epoch": 0.6768989140690471, + "grad_norm": 0.41966307163238525, + "learning_rate": 4.7261584844150485e-05, + "loss": 0.0948, + "step": 28876 + }, + { + "epoch": 0.6769223556438522, + "grad_norm": 0.2739535868167877, + "learning_rate": 4.725532650101363e-05, + "loss": 0.0517, + "step": 28877 + }, + { + "epoch": 0.6769457972186571, + "grad_norm": 0.24158553779125214, + "learning_rate": 4.724906844407087e-05, + "loss": 0.0344, + "step": 28878 + }, + { + "epoch": 0.6769692387934622, + "grad_norm": 0.42316934466362, + "learning_rate": 4.7242810673356094e-05, + "loss": 0.0567, + "step": 28879 + }, + { + "epoch": 0.6769926803682671, + "grad_norm": 0.21672023832798004, + "learning_rate": 4.723655318890336e-05, + "loss": 0.0213, + "step": 28880 + }, + { + "epoch": 0.6770161219430721, + "grad_norm": 0.40482768416404724, + "learning_rate": 4.7230295990746534e-05, + "loss": 0.053, + "step": 28881 + }, + { + "epoch": 0.6770395635178771, + "grad_norm": 0.5008749961853027, + "learning_rate": 4.722403907891963e-05, + "loss": 0.1586, + "step": 28882 + }, + { + "epoch": 0.6770630050926821, + "grad_norm": 0.4004700481891632, + "learning_rate": 4.721778245345659e-05, + "loss": 0.0363, + "step": 28883 + }, + { + "epoch": 0.6770864466674871, + "grad_norm": 0.3397768437862396, + "learning_rate": 4.721152611439134e-05, + "loss": 0.0464, + "step": 28884 + }, + { + "epoch": 0.6771098882422921, + "grad_norm": 0.49673253297805786, + "learning_rate": 4.7205270061757834e-05, + "loss": 0.062, + "step": 28885 + }, + { + "epoch": 0.6771333298170971, + "grad_norm": 0.42893490195274353, + "learning_rate": 4.7199014295590014e-05, + "loss": 0.1232, + "step": 28886 + }, + { + "epoch": 0.6771567713919021, + "grad_norm": 0.37271788716316223, + "learning_rate": 4.71927588159218e-05, + "loss": 0.1128, + "step": 28887 + }, + { + "epoch": 0.677180212966707, + "grad_norm": 0.21587371826171875, + "learning_rate": 4.718650362278718e-05, + "loss": 0.0434, + "step": 28888 + }, + { + "epoch": 0.6772036545415121, + "grad_norm": 0.23750440776348114, + "learning_rate": 4.71802487162201e-05, + "loss": 0.0295, + "step": 28889 + }, + { + "epoch": 0.677227096116317, + "grad_norm": 0.7422781586647034, + "learning_rate": 4.717399409625446e-05, + "loss": 0.0716, + "step": 28890 + }, + { + "epoch": 0.6772505376911221, + "grad_norm": 0.5861826539039612, + "learning_rate": 4.716773976292418e-05, + "loss": 0.1255, + "step": 28891 + }, + { + "epoch": 0.6772739792659271, + "grad_norm": 0.6079548001289368, + "learning_rate": 4.716148571626326e-05, + "loss": 0.1179, + "step": 28892 + }, + { + "epoch": 0.6772974208407321, + "grad_norm": 0.16871897876262665, + "learning_rate": 4.7155231956305615e-05, + "loss": 0.0315, + "step": 28893 + }, + { + "epoch": 0.6773208624155371, + "grad_norm": 0.2445821613073349, + "learning_rate": 4.714897848308513e-05, + "loss": 0.0231, + "step": 28894 + }, + { + "epoch": 0.6773443039903421, + "grad_norm": 0.6428461074829102, + "learning_rate": 4.7142725296635804e-05, + "loss": 0.5631, + "step": 28895 + }, + { + "epoch": 0.6773677455651471, + "grad_norm": 0.6656506657600403, + "learning_rate": 4.713647239699154e-05, + "loss": 0.0822, + "step": 28896 + }, + { + "epoch": 0.6773911871399521, + "grad_norm": 0.33831140398979187, + "learning_rate": 4.713021978418626e-05, + "loss": 0.0704, + "step": 28897 + }, + { + "epoch": 0.6774146287147571, + "grad_norm": 0.2508928179740906, + "learning_rate": 4.712396745825389e-05, + "loss": 0.0569, + "step": 28898 + }, + { + "epoch": 0.677438070289562, + "grad_norm": 0.4480001628398895, + "learning_rate": 4.711771541922837e-05, + "loss": 0.0946, + "step": 28899 + }, + { + "epoch": 0.6774615118643671, + "grad_norm": 0.3667832016944885, + "learning_rate": 4.7111463667143576e-05, + "loss": 0.0656, + "step": 28900 + }, + { + "epoch": 0.677484953439172, + "grad_norm": 0.4920231103897095, + "learning_rate": 4.71052122020335e-05, + "loss": 0.1271, + "step": 28901 + }, + { + "epoch": 0.6775083950139771, + "grad_norm": 0.6956365704536438, + "learning_rate": 4.709896102393202e-05, + "loss": 0.1197, + "step": 28902 + }, + { + "epoch": 0.677531836588782, + "grad_norm": 0.5174343585968018, + "learning_rate": 4.709271013287304e-05, + "loss": 0.6267, + "step": 28903 + }, + { + "epoch": 0.6775552781635871, + "grad_norm": 0.1568698137998581, + "learning_rate": 4.7086459528890535e-05, + "loss": 0.0132, + "step": 28904 + }, + { + "epoch": 0.677578719738392, + "grad_norm": 0.44564342498779297, + "learning_rate": 4.70802092120184e-05, + "loss": 0.0641, + "step": 28905 + }, + { + "epoch": 0.6776021613131971, + "grad_norm": 0.1904197782278061, + "learning_rate": 4.707395918229048e-05, + "loss": 0.0373, + "step": 28906 + }, + { + "epoch": 0.677625602888002, + "grad_norm": 0.5628216862678528, + "learning_rate": 4.70677094397408e-05, + "loss": 0.0891, + "step": 28907 + }, + { + "epoch": 0.677649044462807, + "grad_norm": 0.11617902666330338, + "learning_rate": 4.706145998440321e-05, + "loss": 0.0188, + "step": 28908 + }, + { + "epoch": 0.677672486037612, + "grad_norm": 0.5818219184875488, + "learning_rate": 4.705521081631162e-05, + "loss": 0.1703, + "step": 28909 + }, + { + "epoch": 0.677695927612417, + "grad_norm": 0.39983147382736206, + "learning_rate": 4.704896193549996e-05, + "loss": 0.088, + "step": 28910 + }, + { + "epoch": 0.677719369187222, + "grad_norm": 0.36410224437713623, + "learning_rate": 4.704271334200211e-05, + "loss": 0.0476, + "step": 28911 + }, + { + "epoch": 0.677742810762027, + "grad_norm": 0.44517621397972107, + "learning_rate": 4.7036465035851964e-05, + "loss": 0.0605, + "step": 28912 + }, + { + "epoch": 0.677766252336832, + "grad_norm": 0.1089971661567688, + "learning_rate": 4.7030217017083464e-05, + "loss": 0.0144, + "step": 28913 + }, + { + "epoch": 0.677789693911637, + "grad_norm": 0.10159698873758316, + "learning_rate": 4.702396928573051e-05, + "loss": 0.0208, + "step": 28914 + }, + { + "epoch": 0.677813135486442, + "grad_norm": 0.37788426876068115, + "learning_rate": 4.701772184182695e-05, + "loss": 0.0513, + "step": 28915 + }, + { + "epoch": 0.677836577061247, + "grad_norm": 0.2238127440214157, + "learning_rate": 4.701147468540675e-05, + "loss": 0.0322, + "step": 28916 + }, + { + "epoch": 0.6778600186360519, + "grad_norm": 1.0285099744796753, + "learning_rate": 4.700522781650378e-05, + "loss": 0.0374, + "step": 28917 + }, + { + "epoch": 0.677883460210857, + "grad_norm": 0.2875700891017914, + "learning_rate": 4.699898123515194e-05, + "loss": 0.0465, + "step": 28918 + }, + { + "epoch": 0.6779069017856619, + "grad_norm": 0.3446476459503174, + "learning_rate": 4.699273494138507e-05, + "loss": 0.0726, + "step": 28919 + }, + { + "epoch": 0.677930343360467, + "grad_norm": 0.5618042945861816, + "learning_rate": 4.698648893523716e-05, + "loss": 0.1772, + "step": 28920 + }, + { + "epoch": 0.6779537849352719, + "grad_norm": 0.10344043374061584, + "learning_rate": 4.698024321674206e-05, + "loss": 0.0062, + "step": 28921 + }, + { + "epoch": 0.677977226510077, + "grad_norm": 0.25585082173347473, + "learning_rate": 4.6973997785933635e-05, + "loss": 0.0543, + "step": 28922 + }, + { + "epoch": 0.6780006680848819, + "grad_norm": 0.5044899582862854, + "learning_rate": 4.6967752642845786e-05, + "loss": 0.1318, + "step": 28923 + }, + { + "epoch": 0.678024109659687, + "grad_norm": 0.3892919719219208, + "learning_rate": 4.6961507787512416e-05, + "loss": 0.0316, + "step": 28924 + }, + { + "epoch": 0.6780475512344919, + "grad_norm": 0.19533705711364746, + "learning_rate": 4.695526321996735e-05, + "loss": 0.0375, + "step": 28925 + }, + { + "epoch": 0.678070992809297, + "grad_norm": 0.21820536255836487, + "learning_rate": 4.694901894024456e-05, + "loss": 0.0355, + "step": 28926 + }, + { + "epoch": 0.6780944343841019, + "grad_norm": 0.19891013205051422, + "learning_rate": 4.694277494837784e-05, + "loss": 0.0358, + "step": 28927 + }, + { + "epoch": 0.6781178759589069, + "grad_norm": 0.49692219495773315, + "learning_rate": 4.6936531244401164e-05, + "loss": 0.105, + "step": 28928 + }, + { + "epoch": 0.6781413175337119, + "grad_norm": 0.5290038585662842, + "learning_rate": 4.693028782834834e-05, + "loss": 0.0916, + "step": 28929 + }, + { + "epoch": 0.6781647591085169, + "grad_norm": 0.37793388962745667, + "learning_rate": 4.692404470025329e-05, + "loss": 0.0706, + "step": 28930 + }, + { + "epoch": 0.6781882006833219, + "grad_norm": 0.17896215617656708, + "learning_rate": 4.691780186014986e-05, + "loss": 0.0119, + "step": 28931 + }, + { + "epoch": 0.6782116422581269, + "grad_norm": 0.4221512973308563, + "learning_rate": 4.691155930807188e-05, + "loss": 0.1224, + "step": 28932 + }, + { + "epoch": 0.6782350838329319, + "grad_norm": 0.5319796800613403, + "learning_rate": 4.690531704405332e-05, + "loss": 0.5442, + "step": 28933 + }, + { + "epoch": 0.6782585254077369, + "grad_norm": 0.3554883599281311, + "learning_rate": 4.6899075068127996e-05, + "loss": 0.077, + "step": 28934 + }, + { + "epoch": 0.6782819669825418, + "grad_norm": 0.8088017702102661, + "learning_rate": 4.689283338032979e-05, + "loss": 0.1259, + "step": 28935 + }, + { + "epoch": 0.6783054085573469, + "grad_norm": 0.47904688119888306, + "learning_rate": 4.6886591980692566e-05, + "loss": 0.0517, + "step": 28936 + }, + { + "epoch": 0.6783288501321518, + "grad_norm": 0.47910258173942566, + "learning_rate": 4.688035086925018e-05, + "loss": 0.09, + "step": 28937 + }, + { + "epoch": 0.6783522917069569, + "grad_norm": 0.5410827994346619, + "learning_rate": 4.687411004603647e-05, + "loss": 0.1217, + "step": 28938 + }, + { + "epoch": 0.6783757332817618, + "grad_norm": 0.33506500720977783, + "learning_rate": 4.686786951108536e-05, + "loss": 0.0562, + "step": 28939 + }, + { + "epoch": 0.6783991748565669, + "grad_norm": 0.13353300094604492, + "learning_rate": 4.686162926443066e-05, + "loss": 0.0325, + "step": 28940 + }, + { + "epoch": 0.6784226164313718, + "grad_norm": 0.5159915685653687, + "learning_rate": 4.685538930610629e-05, + "loss": 0.053, + "step": 28941 + }, + { + "epoch": 0.6784460580061769, + "grad_norm": 1.093704104423523, + "learning_rate": 4.684914963614606e-05, + "loss": 0.0822, + "step": 28942 + }, + { + "epoch": 0.6784694995809819, + "grad_norm": 0.48517581820487976, + "learning_rate": 4.684291025458385e-05, + "loss": 0.0995, + "step": 28943 + }, + { + "epoch": 0.6784929411557868, + "grad_norm": 0.31129077076911926, + "learning_rate": 4.683667116145347e-05, + "loss": 0.0589, + "step": 28944 + }, + { + "epoch": 0.6785163827305919, + "grad_norm": 0.39022207260131836, + "learning_rate": 4.6830432356788837e-05, + "loss": 0.0588, + "step": 28945 + }, + { + "epoch": 0.6785398243053968, + "grad_norm": 0.16049234569072723, + "learning_rate": 4.682419384062377e-05, + "loss": 0.0356, + "step": 28946 + }, + { + "epoch": 0.6785632658802019, + "grad_norm": 0.4806535840034485, + "learning_rate": 4.681795561299213e-05, + "loss": 0.0572, + "step": 28947 + }, + { + "epoch": 0.6785867074550068, + "grad_norm": 0.34376588463783264, + "learning_rate": 4.6811717673927755e-05, + "loss": 0.0535, + "step": 28948 + }, + { + "epoch": 0.6786101490298119, + "grad_norm": 0.11394540220499039, + "learning_rate": 4.680548002346449e-05, + "loss": 0.0157, + "step": 28949 + }, + { + "epoch": 0.6786335906046168, + "grad_norm": 1.0253524780273438, + "learning_rate": 4.679924266163616e-05, + "loss": 0.1509, + "step": 28950 + }, + { + "epoch": 0.6786570321794219, + "grad_norm": 0.4502949118614197, + "learning_rate": 4.679300558847666e-05, + "loss": 0.0569, + "step": 28951 + }, + { + "epoch": 0.6786804737542268, + "grad_norm": 0.561015784740448, + "learning_rate": 4.678676880401978e-05, + "loss": 0.0547, + "step": 28952 + }, + { + "epoch": 0.6787039153290318, + "grad_norm": 0.4197440445423126, + "learning_rate": 4.678053230829942e-05, + "loss": 0.078, + "step": 28953 + }, + { + "epoch": 0.6787273569038368, + "grad_norm": 0.5779187083244324, + "learning_rate": 4.677429610134939e-05, + "loss": 0.5664, + "step": 28954 + }, + { + "epoch": 0.6787507984786418, + "grad_norm": 0.33271321654319763, + "learning_rate": 4.676806018320353e-05, + "loss": 0.0463, + "step": 28955 + }, + { + "epoch": 0.6787742400534468, + "grad_norm": 0.6061186790466309, + "learning_rate": 4.676182455389567e-05, + "loss": 0.1836, + "step": 28956 + }, + { + "epoch": 0.6787976816282518, + "grad_norm": 0.2279859483242035, + "learning_rate": 4.675558921345961e-05, + "loss": 0.0413, + "step": 28957 + }, + { + "epoch": 0.6788211232030568, + "grad_norm": 0.4915178716182709, + "learning_rate": 4.674935416192925e-05, + "loss": 0.1103, + "step": 28958 + }, + { + "epoch": 0.6788445647778618, + "grad_norm": 0.35343724489212036, + "learning_rate": 4.6743119399338394e-05, + "loss": 0.0754, + "step": 28959 + }, + { + "epoch": 0.6788680063526668, + "grad_norm": 0.24360854923725128, + "learning_rate": 4.673688492572088e-05, + "loss": 0.0572, + "step": 28960 + }, + { + "epoch": 0.6788914479274718, + "grad_norm": 0.4132600724697113, + "learning_rate": 4.6730650741110515e-05, + "loss": 0.0509, + "step": 28961 + }, + { + "epoch": 0.6789148895022767, + "grad_norm": 0.16021397709846497, + "learning_rate": 4.672441684554114e-05, + "loss": 0.0369, + "step": 28962 + }, + { + "epoch": 0.6789383310770818, + "grad_norm": 0.1703559160232544, + "learning_rate": 4.6718183239046545e-05, + "loss": 0.0503, + "step": 28963 + }, + { + "epoch": 0.6789617726518867, + "grad_norm": 0.580049455165863, + "learning_rate": 4.671194992166058e-05, + "loss": 0.1206, + "step": 28964 + }, + { + "epoch": 0.6789852142266918, + "grad_norm": 0.2281406968832016, + "learning_rate": 4.670571689341711e-05, + "loss": 0.0288, + "step": 28965 + }, + { + "epoch": 0.6790086558014967, + "grad_norm": 0.40209171175956726, + "learning_rate": 4.669948415434993e-05, + "loss": 0.0418, + "step": 28966 + }, + { + "epoch": 0.6790320973763018, + "grad_norm": 0.13518428802490234, + "learning_rate": 4.6693251704492835e-05, + "loss": 0.0356, + "step": 28967 + }, + { + "epoch": 0.6790555389511067, + "grad_norm": 0.5402597188949585, + "learning_rate": 4.668701954387965e-05, + "loss": 0.0739, + "step": 28968 + }, + { + "epoch": 0.6790789805259118, + "grad_norm": 0.5993406772613525, + "learning_rate": 4.6680787672544214e-05, + "loss": 0.1197, + "step": 28969 + }, + { + "epoch": 0.6791024221007167, + "grad_norm": 0.4811520576477051, + "learning_rate": 4.667455609052027e-05, + "loss": 0.0988, + "step": 28970 + }, + { + "epoch": 0.6791258636755217, + "grad_norm": 0.21089060604572296, + "learning_rate": 4.666832479784172e-05, + "loss": 0.0273, + "step": 28971 + }, + { + "epoch": 0.6791493052503267, + "grad_norm": 0.24498271942138672, + "learning_rate": 4.666209379454235e-05, + "loss": 0.0755, + "step": 28972 + }, + { + "epoch": 0.6791727468251317, + "grad_norm": 0.44015592336654663, + "learning_rate": 4.665586308065596e-05, + "loss": 0.0867, + "step": 28973 + }, + { + "epoch": 0.6791961883999367, + "grad_norm": 0.4310590326786041, + "learning_rate": 4.6649632656216356e-05, + "loss": 0.089, + "step": 28974 + }, + { + "epoch": 0.6792196299747417, + "grad_norm": 0.47206416726112366, + "learning_rate": 4.6643402521257306e-05, + "loss": 0.0708, + "step": 28975 + }, + { + "epoch": 0.6792430715495467, + "grad_norm": 0.18444786965847015, + "learning_rate": 4.663717267581269e-05, + "loss": 0.0306, + "step": 28976 + }, + { + "epoch": 0.6792665131243517, + "grad_norm": 0.8654069900512695, + "learning_rate": 4.6630943119916235e-05, + "loss": 0.068, + "step": 28977 + }, + { + "epoch": 0.6792899546991567, + "grad_norm": 0.46220824122428894, + "learning_rate": 4.662471385360182e-05, + "loss": 0.1342, + "step": 28978 + }, + { + "epoch": 0.6793133962739617, + "grad_norm": 0.4498431384563446, + "learning_rate": 4.6618484876903216e-05, + "loss": 0.0695, + "step": 28979 + }, + { + "epoch": 0.6793368378487666, + "grad_norm": 0.6272438764572144, + "learning_rate": 4.661225618985421e-05, + "loss": 0.6107, + "step": 28980 + }, + { + "epoch": 0.6793602794235717, + "grad_norm": 0.27908751368522644, + "learning_rate": 4.660602779248861e-05, + "loss": 0.0438, + "step": 28981 + }, + { + "epoch": 0.6793837209983766, + "grad_norm": 0.5465636849403381, + "learning_rate": 4.659979968484015e-05, + "loss": 0.1102, + "step": 28982 + }, + { + "epoch": 0.6794071625731817, + "grad_norm": 0.34153077006340027, + "learning_rate": 4.659357186694273e-05, + "loss": 0.0557, + "step": 28983 + }, + { + "epoch": 0.6794306041479866, + "grad_norm": 0.5350881814956665, + "learning_rate": 4.6587344338830086e-05, + "loss": 0.1286, + "step": 28984 + }, + { + "epoch": 0.6794540457227917, + "grad_norm": 0.5818782448768616, + "learning_rate": 4.658111710053601e-05, + "loss": 0.0795, + "step": 28985 + }, + { + "epoch": 0.6794774872975966, + "grad_norm": 0.48578888177871704, + "learning_rate": 4.65748901520943e-05, + "loss": 0.1164, + "step": 28986 + }, + { + "epoch": 0.6795009288724017, + "grad_norm": 0.1254393458366394, + "learning_rate": 4.6568663493538735e-05, + "loss": 0.0152, + "step": 28987 + }, + { + "epoch": 0.6795243704472066, + "grad_norm": 0.2567026913166046, + "learning_rate": 4.6562437124903066e-05, + "loss": 0.0573, + "step": 28988 + }, + { + "epoch": 0.6795478120220116, + "grad_norm": 0.7008722424507141, + "learning_rate": 4.655621104622112e-05, + "loss": 0.1284, + "step": 28989 + }, + { + "epoch": 0.6795712535968166, + "grad_norm": 0.36648792028427124, + "learning_rate": 4.654998525752672e-05, + "loss": 0.0531, + "step": 28990 + }, + { + "epoch": 0.6795946951716216, + "grad_norm": 0.31331679224967957, + "learning_rate": 4.6543759758853586e-05, + "loss": 0.0424, + "step": 28991 + }, + { + "epoch": 0.6796181367464266, + "grad_norm": 0.3579375147819519, + "learning_rate": 4.653753455023553e-05, + "loss": 0.0325, + "step": 28992 + }, + { + "epoch": 0.6796415783212316, + "grad_norm": 0.311242938041687, + "learning_rate": 4.653130963170631e-05, + "loss": 0.2246, + "step": 28993 + }, + { + "epoch": 0.6796650198960367, + "grad_norm": 0.23519295454025269, + "learning_rate": 4.652508500329969e-05, + "loss": 0.0604, + "step": 28994 + }, + { + "epoch": 0.6796884614708416, + "grad_norm": 0.5421041250228882, + "learning_rate": 4.6518860665049447e-05, + "loss": 0.51, + "step": 28995 + }, + { + "epoch": 0.6797119030456467, + "grad_norm": 0.47780102491378784, + "learning_rate": 4.65126366169894e-05, + "loss": 0.4339, + "step": 28996 + }, + { + "epoch": 0.6797353446204516, + "grad_norm": 0.2026064693927765, + "learning_rate": 4.650641285915329e-05, + "loss": 0.034, + "step": 28997 + }, + { + "epoch": 0.6797587861952566, + "grad_norm": 0.6055313348770142, + "learning_rate": 4.650018939157489e-05, + "loss": 0.1359, + "step": 28998 + }, + { + "epoch": 0.6797822277700616, + "grad_norm": 0.12868636846542358, + "learning_rate": 4.6493966214287964e-05, + "loss": 0.023, + "step": 28999 + }, + { + "epoch": 0.6798056693448666, + "grad_norm": 0.587916910648346, + "learning_rate": 4.648774332732625e-05, + "loss": 0.103, + "step": 29000 + }, + { + "epoch": 0.6798291109196716, + "grad_norm": 0.03531899303197861, + "learning_rate": 4.648152073072358e-05, + "loss": 0.0055, + "step": 29001 + }, + { + "epoch": 0.6798525524944766, + "grad_norm": 0.21440277993679047, + "learning_rate": 4.647529842451365e-05, + "loss": 0.0586, + "step": 29002 + }, + { + "epoch": 0.6798759940692816, + "grad_norm": 0.1580432802438736, + "learning_rate": 4.646907640873028e-05, + "loss": 0.0226, + "step": 29003 + }, + { + "epoch": 0.6798994356440866, + "grad_norm": 0.1346701830625534, + "learning_rate": 4.646285468340722e-05, + "loss": 0.0278, + "step": 29004 + }, + { + "epoch": 0.6799228772188916, + "grad_norm": 0.5345361828804016, + "learning_rate": 4.6456633248578205e-05, + "loss": 0.0609, + "step": 29005 + }, + { + "epoch": 0.6799463187936966, + "grad_norm": 0.2657026946544647, + "learning_rate": 4.645041210427701e-05, + "loss": 0.0641, + "step": 29006 + }, + { + "epoch": 0.6799697603685015, + "grad_norm": 0.27560892701148987, + "learning_rate": 4.6444191250537384e-05, + "loss": 0.0558, + "step": 29007 + }, + { + "epoch": 0.6799932019433066, + "grad_norm": 0.6455404162406921, + "learning_rate": 4.643797068739304e-05, + "loss": 0.6629, + "step": 29008 + }, + { + "epoch": 0.6800166435181115, + "grad_norm": 0.33062484860420227, + "learning_rate": 4.6431750414877816e-05, + "loss": 0.0632, + "step": 29009 + }, + { + "epoch": 0.6800400850929166, + "grad_norm": 0.43707993626594543, + "learning_rate": 4.642553043302542e-05, + "loss": 0.0753, + "step": 29010 + }, + { + "epoch": 0.6800635266677215, + "grad_norm": 0.3388044834136963, + "learning_rate": 4.6419310741869604e-05, + "loss": 0.0489, + "step": 29011 + }, + { + "epoch": 0.6800869682425266, + "grad_norm": 0.13353252410888672, + "learning_rate": 4.6413091341444104e-05, + "loss": 0.0242, + "step": 29012 + }, + { + "epoch": 0.6801104098173315, + "grad_norm": 0.33629268407821655, + "learning_rate": 4.640687223178265e-05, + "loss": 0.0536, + "step": 29013 + }, + { + "epoch": 0.6801338513921366, + "grad_norm": 0.13613572716712952, + "learning_rate": 4.640065341291905e-05, + "loss": 0.0358, + "step": 29014 + }, + { + "epoch": 0.6801572929669415, + "grad_norm": 0.6178110837936401, + "learning_rate": 4.6394434884886976e-05, + "loss": 0.1116, + "step": 29015 + }, + { + "epoch": 0.6801807345417465, + "grad_norm": 0.364755243062973, + "learning_rate": 4.638821664772024e-05, + "loss": 0.049, + "step": 29016 + }, + { + "epoch": 0.6802041761165515, + "grad_norm": 0.1494835466146469, + "learning_rate": 4.6381998701452546e-05, + "loss": 0.0325, + "step": 29017 + }, + { + "epoch": 0.6802276176913565, + "grad_norm": 0.20282121002674103, + "learning_rate": 4.6375781046117636e-05, + "loss": 0.0208, + "step": 29018 + }, + { + "epoch": 0.6802510592661615, + "grad_norm": 0.366531640291214, + "learning_rate": 4.636956368174925e-05, + "loss": 0.3717, + "step": 29019 + }, + { + "epoch": 0.6802745008409665, + "grad_norm": 0.08288682997226715, + "learning_rate": 4.636334660838108e-05, + "loss": 0.0117, + "step": 29020 + }, + { + "epoch": 0.6802979424157715, + "grad_norm": 0.7029181718826294, + "learning_rate": 4.635712982604693e-05, + "loss": 0.0688, + "step": 29021 + }, + { + "epoch": 0.6803213839905765, + "grad_norm": 0.2176266461610794, + "learning_rate": 4.635091333478051e-05, + "loss": 0.044, + "step": 29022 + }, + { + "epoch": 0.6803448255653815, + "grad_norm": 0.3540840446949005, + "learning_rate": 4.634469713461553e-05, + "loss": 0.0757, + "step": 29023 + }, + { + "epoch": 0.6803682671401865, + "grad_norm": 0.27260005474090576, + "learning_rate": 4.633848122558575e-05, + "loss": 0.0329, + "step": 29024 + }, + { + "epoch": 0.6803917087149914, + "grad_norm": 0.28658026456832886, + "learning_rate": 4.6332265607724836e-05, + "loss": 0.0543, + "step": 29025 + }, + { + "epoch": 0.6804151502897965, + "grad_norm": 0.27164116501808167, + "learning_rate": 4.632605028106659e-05, + "loss": 0.0802, + "step": 29026 + }, + { + "epoch": 0.6804385918646014, + "grad_norm": 0.6261324882507324, + "learning_rate": 4.631983524564467e-05, + "loss": 0.1109, + "step": 29027 + }, + { + "epoch": 0.6804620334394065, + "grad_norm": 0.12812155485153198, + "learning_rate": 4.6313620501492875e-05, + "loss": 0.0261, + "step": 29028 + }, + { + "epoch": 0.6804854750142114, + "grad_norm": 0.5807161927223206, + "learning_rate": 4.630740604864487e-05, + "loss": 0.0843, + "step": 29029 + }, + { + "epoch": 0.6805089165890165, + "grad_norm": 0.20881783962249756, + "learning_rate": 4.630119188713441e-05, + "loss": 0.0321, + "step": 29030 + }, + { + "epoch": 0.6805323581638214, + "grad_norm": 0.43011119961738586, + "learning_rate": 4.6294978016995184e-05, + "loss": 0.0673, + "step": 29031 + }, + { + "epoch": 0.6805557997386265, + "grad_norm": 0.3108176589012146, + "learning_rate": 4.628876443826091e-05, + "loss": 0.0236, + "step": 29032 + }, + { + "epoch": 0.6805792413134314, + "grad_norm": 0.6532148718833923, + "learning_rate": 4.628255115096528e-05, + "loss": 0.0922, + "step": 29033 + }, + { + "epoch": 0.6806026828882364, + "grad_norm": 0.1320759803056717, + "learning_rate": 4.6276338155142074e-05, + "loss": 0.0275, + "step": 29034 + }, + { + "epoch": 0.6806261244630414, + "grad_norm": 0.29814383387565613, + "learning_rate": 4.6270125450824976e-05, + "loss": 0.08, + "step": 29035 + }, + { + "epoch": 0.6806495660378464, + "grad_norm": 0.43209993839263916, + "learning_rate": 4.626391303804768e-05, + "loss": 0.1118, + "step": 29036 + }, + { + "epoch": 0.6806730076126514, + "grad_norm": 0.13222001492977142, + "learning_rate": 4.625770091684386e-05, + "loss": 0.0175, + "step": 29037 + }, + { + "epoch": 0.6806964491874564, + "grad_norm": 0.4921208322048187, + "learning_rate": 4.625148908724731e-05, + "loss": 0.0417, + "step": 29038 + }, + { + "epoch": 0.6807198907622614, + "grad_norm": 0.6278828382492065, + "learning_rate": 4.624527754929169e-05, + "loss": 0.0827, + "step": 29039 + }, + { + "epoch": 0.6807433323370664, + "grad_norm": 0.52955162525177, + "learning_rate": 4.623906630301067e-05, + "loss": 0.1172, + "step": 29040 + }, + { + "epoch": 0.6807667739118713, + "grad_norm": 0.46353471279144287, + "learning_rate": 4.623285534843802e-05, + "loss": 0.0797, + "step": 29041 + }, + { + "epoch": 0.6807902154866764, + "grad_norm": 0.5043457746505737, + "learning_rate": 4.622664468560741e-05, + "loss": 0.0809, + "step": 29042 + }, + { + "epoch": 0.6808136570614813, + "grad_norm": 0.40744462609291077, + "learning_rate": 4.622043431455254e-05, + "loss": 0.0566, + "step": 29043 + }, + { + "epoch": 0.6808370986362864, + "grad_norm": 0.3248712420463562, + "learning_rate": 4.6214224235307116e-05, + "loss": 0.0926, + "step": 29044 + }, + { + "epoch": 0.6808605402110914, + "grad_norm": 0.1622607260942459, + "learning_rate": 4.6208014447904816e-05, + "loss": 0.0325, + "step": 29045 + }, + { + "epoch": 0.6808839817858964, + "grad_norm": 0.5250590443611145, + "learning_rate": 4.620180495237929e-05, + "loss": 0.0807, + "step": 29046 + }, + { + "epoch": 0.6809074233607014, + "grad_norm": 0.4392034411430359, + "learning_rate": 4.619559574876434e-05, + "loss": 0.0966, + "step": 29047 + }, + { + "epoch": 0.6809308649355064, + "grad_norm": 0.4551655054092407, + "learning_rate": 4.6189386837093595e-05, + "loss": 0.0566, + "step": 29048 + }, + { + "epoch": 0.6809543065103114, + "grad_norm": 0.22107669711112976, + "learning_rate": 4.6183178217400705e-05, + "loss": 0.015, + "step": 29049 + }, + { + "epoch": 0.6809777480851164, + "grad_norm": 0.5699146389961243, + "learning_rate": 4.617696988971945e-05, + "loss": 0.7179, + "step": 29050 + }, + { + "epoch": 0.6810011896599214, + "grad_norm": 0.5340283513069153, + "learning_rate": 4.617076185408347e-05, + "loss": 0.6097, + "step": 29051 + }, + { + "epoch": 0.6810246312347263, + "grad_norm": 0.48845309019088745, + "learning_rate": 4.616455411052644e-05, + "loss": 0.102, + "step": 29052 + }, + { + "epoch": 0.6810480728095314, + "grad_norm": 0.1184341311454773, + "learning_rate": 4.615834665908202e-05, + "loss": 0.016, + "step": 29053 + }, + { + "epoch": 0.6810715143843363, + "grad_norm": 0.6008573174476624, + "learning_rate": 4.6152139499783965e-05, + "loss": 0.1291, + "step": 29054 + }, + { + "epoch": 0.6810949559591414, + "grad_norm": 0.44569796323776245, + "learning_rate": 4.614593263266591e-05, + "loss": 0.0916, + "step": 29055 + }, + { + "epoch": 0.6811183975339463, + "grad_norm": 0.40513014793395996, + "learning_rate": 4.613972605776155e-05, + "loss": 0.0483, + "step": 29056 + }, + { + "epoch": 0.6811418391087514, + "grad_norm": 0.6774947047233582, + "learning_rate": 4.6133519775104536e-05, + "loss": 0.1062, + "step": 29057 + }, + { + "epoch": 0.6811652806835563, + "grad_norm": 0.3513610064983368, + "learning_rate": 4.6127313784728534e-05, + "loss": 0.0574, + "step": 29058 + }, + { + "epoch": 0.6811887222583614, + "grad_norm": 0.7059875130653381, + "learning_rate": 4.6121108086667276e-05, + "loss": 0.1411, + "step": 29059 + }, + { + "epoch": 0.6812121638331663, + "grad_norm": 0.4487185776233673, + "learning_rate": 4.61149026809544e-05, + "loss": 0.0695, + "step": 29060 + }, + { + "epoch": 0.6812356054079713, + "grad_norm": 0.36481091380119324, + "learning_rate": 4.610869756762354e-05, + "loss": 0.0496, + "step": 29061 + }, + { + "epoch": 0.6812590469827763, + "grad_norm": 0.3542248606681824, + "learning_rate": 4.610249274670844e-05, + "loss": 0.0769, + "step": 29062 + }, + { + "epoch": 0.6812824885575813, + "grad_norm": 0.40929293632507324, + "learning_rate": 4.609628821824272e-05, + "loss": 0.072, + "step": 29063 + }, + { + "epoch": 0.6813059301323863, + "grad_norm": 0.4174495339393616, + "learning_rate": 4.609008398226006e-05, + "loss": 0.0803, + "step": 29064 + }, + { + "epoch": 0.6813293717071913, + "grad_norm": 0.34862685203552246, + "learning_rate": 4.608388003879409e-05, + "loss": 0.0514, + "step": 29065 + }, + { + "epoch": 0.6813528132819963, + "grad_norm": 0.5288670063018799, + "learning_rate": 4.6077676387878533e-05, + "loss": 0.1411, + "step": 29066 + }, + { + "epoch": 0.6813762548568013, + "grad_norm": 0.4973105192184448, + "learning_rate": 4.607147302954703e-05, + "loss": 0.0569, + "step": 29067 + }, + { + "epoch": 0.6813996964316062, + "grad_norm": 0.13071869313716888, + "learning_rate": 4.6065269963833223e-05, + "loss": 0.0089, + "step": 29068 + }, + { + "epoch": 0.6814231380064113, + "grad_norm": 0.4635296165943146, + "learning_rate": 4.605906719077078e-05, + "loss": 0.076, + "step": 29069 + }, + { + "epoch": 0.6814465795812162, + "grad_norm": 0.47170624136924744, + "learning_rate": 4.605286471039335e-05, + "loss": 0.1072, + "step": 29070 + }, + { + "epoch": 0.6814700211560213, + "grad_norm": 0.34420615434646606, + "learning_rate": 4.604666252273456e-05, + "loss": 0.0783, + "step": 29071 + }, + { + "epoch": 0.6814934627308262, + "grad_norm": 0.40669628977775574, + "learning_rate": 4.604046062782813e-05, + "loss": 0.0711, + "step": 29072 + }, + { + "epoch": 0.6815169043056313, + "grad_norm": 0.3114185035228729, + "learning_rate": 4.603425902570765e-05, + "loss": 0.0355, + "step": 29073 + }, + { + "epoch": 0.6815403458804362, + "grad_norm": 0.29404497146606445, + "learning_rate": 4.602805771640682e-05, + "loss": 0.0544, + "step": 29074 + }, + { + "epoch": 0.6815637874552413, + "grad_norm": 0.14023543894290924, + "learning_rate": 4.6021856699959276e-05, + "loss": 0.0154, + "step": 29075 + }, + { + "epoch": 0.6815872290300462, + "grad_norm": 0.28400933742523193, + "learning_rate": 4.601565597639867e-05, + "loss": 0.0247, + "step": 29076 + }, + { + "epoch": 0.6816106706048513, + "grad_norm": 0.2551324665546417, + "learning_rate": 4.6009455545758606e-05, + "loss": 0.0395, + "step": 29077 + }, + { + "epoch": 0.6816341121796562, + "grad_norm": 0.19831474125385284, + "learning_rate": 4.600325540807273e-05, + "loss": 0.0308, + "step": 29078 + }, + { + "epoch": 0.6816575537544612, + "grad_norm": 0.4104183316230774, + "learning_rate": 4.599705556337475e-05, + "loss": 0.072, + "step": 29079 + }, + { + "epoch": 0.6816809953292662, + "grad_norm": 0.387094646692276, + "learning_rate": 4.599085601169825e-05, + "loss": 0.0975, + "step": 29080 + }, + { + "epoch": 0.6817044369040712, + "grad_norm": 0.2183384746313095, + "learning_rate": 4.5984656753076895e-05, + "loss": 0.0196, + "step": 29081 + }, + { + "epoch": 0.6817278784788762, + "grad_norm": 0.11888542771339417, + "learning_rate": 4.59784577875443e-05, + "loss": 0.0244, + "step": 29082 + }, + { + "epoch": 0.6817513200536812, + "grad_norm": 0.5243421196937561, + "learning_rate": 4.5972259115134134e-05, + "loss": 0.0902, + "step": 29083 + }, + { + "epoch": 0.6817747616284862, + "grad_norm": 0.16957862675189972, + "learning_rate": 4.5966060735879955e-05, + "loss": 0.029, + "step": 29084 + }, + { + "epoch": 0.6817982032032912, + "grad_norm": 0.48342931270599365, + "learning_rate": 4.595986264981546e-05, + "loss": 0.0629, + "step": 29085 + }, + { + "epoch": 0.6818216447780961, + "grad_norm": 1.1033434867858887, + "learning_rate": 4.595366485697429e-05, + "loss": 0.1582, + "step": 29086 + }, + { + "epoch": 0.6818450863529012, + "grad_norm": 0.2823685109615326, + "learning_rate": 4.5947467357390065e-05, + "loss": 0.0385, + "step": 29087 + }, + { + "epoch": 0.6818685279277061, + "grad_norm": 0.12469752877950668, + "learning_rate": 4.5941270151096405e-05, + "loss": 0.0327, + "step": 29088 + }, + { + "epoch": 0.6818919695025112, + "grad_norm": 0.49589964747428894, + "learning_rate": 4.593507323812693e-05, + "loss": 0.0396, + "step": 29089 + }, + { + "epoch": 0.6819154110773161, + "grad_norm": 0.40378835797309875, + "learning_rate": 4.592887661851527e-05, + "loss": 0.0582, + "step": 29090 + }, + { + "epoch": 0.6819388526521212, + "grad_norm": 0.10948783159255981, + "learning_rate": 4.592268029229501e-05, + "loss": 0.0141, + "step": 29091 + }, + { + "epoch": 0.6819622942269261, + "grad_norm": 0.28478148579597473, + "learning_rate": 4.5916484259499835e-05, + "loss": 0.0607, + "step": 29092 + }, + { + "epoch": 0.6819857358017312, + "grad_norm": 0.3615381121635437, + "learning_rate": 4.591028852016333e-05, + "loss": 0.0511, + "step": 29093 + }, + { + "epoch": 0.6820091773765361, + "grad_norm": 0.2829545736312866, + "learning_rate": 4.5904093074319146e-05, + "loss": 0.0551, + "step": 29094 + }, + { + "epoch": 0.6820326189513412, + "grad_norm": 0.3494263291358948, + "learning_rate": 4.589789792200086e-05, + "loss": 0.0928, + "step": 29095 + }, + { + "epoch": 0.6820560605261462, + "grad_norm": 0.555284321308136, + "learning_rate": 4.589170306324206e-05, + "loss": 0.1221, + "step": 29096 + }, + { + "epoch": 0.6820795021009511, + "grad_norm": 0.3902960419654846, + "learning_rate": 4.588550849807645e-05, + "loss": 0.0697, + "step": 29097 + }, + { + "epoch": 0.6821029436757562, + "grad_norm": 0.5751007795333862, + "learning_rate": 4.587931422653754e-05, + "loss": 0.0824, + "step": 29098 + }, + { + "epoch": 0.6821263852505611, + "grad_norm": 0.8761656880378723, + "learning_rate": 4.587312024865904e-05, + "loss": 0.1378, + "step": 29099 + }, + { + "epoch": 0.6821498268253662, + "grad_norm": 0.2624456286430359, + "learning_rate": 4.586692656447451e-05, + "loss": 0.0571, + "step": 29100 + }, + { + "epoch": 0.6821732684001711, + "grad_norm": 0.48764175176620483, + "learning_rate": 4.5860733174017555e-05, + "loss": 0.0659, + "step": 29101 + }, + { + "epoch": 0.6821967099749762, + "grad_norm": 0.37218204140663147, + "learning_rate": 4.58545400773218e-05, + "loss": 0.0586, + "step": 29102 + }, + { + "epoch": 0.6822201515497811, + "grad_norm": 0.6778411269187927, + "learning_rate": 4.584834727442078e-05, + "loss": 0.1665, + "step": 29103 + }, + { + "epoch": 0.6822435931245862, + "grad_norm": 0.6716911792755127, + "learning_rate": 4.584215476534819e-05, + "loss": 0.145, + "step": 29104 + }, + { + "epoch": 0.6822670346993911, + "grad_norm": 0.24453586339950562, + "learning_rate": 4.583596255013759e-05, + "loss": 0.0292, + "step": 29105 + }, + { + "epoch": 0.6822904762741961, + "grad_norm": 0.4874790906906128, + "learning_rate": 4.582977062882258e-05, + "loss": 0.0977, + "step": 29106 + }, + { + "epoch": 0.6823139178490011, + "grad_norm": 0.6585705876350403, + "learning_rate": 4.582357900143675e-05, + "loss": 0.1644, + "step": 29107 + }, + { + "epoch": 0.6823373594238061, + "grad_norm": 0.5429721474647522, + "learning_rate": 4.5817387668013724e-05, + "loss": 0.0789, + "step": 29108 + }, + { + "epoch": 0.6823608009986111, + "grad_norm": 0.17011503875255585, + "learning_rate": 4.581119662858703e-05, + "loss": 0.0364, + "step": 29109 + }, + { + "epoch": 0.6823842425734161, + "grad_norm": 0.3220880627632141, + "learning_rate": 4.5805005883190297e-05, + "loss": 0.0495, + "step": 29110 + }, + { + "epoch": 0.6824076841482211, + "grad_norm": 0.09843912720680237, + "learning_rate": 4.579881543185717e-05, + "loss": 0.0289, + "step": 29111 + }, + { + "epoch": 0.6824311257230261, + "grad_norm": 0.07767719775438309, + "learning_rate": 4.5792625274621195e-05, + "loss": 0.0045, + "step": 29112 + }, + { + "epoch": 0.682454567297831, + "grad_norm": 0.9068291187286377, + "learning_rate": 4.578643541151596e-05, + "loss": 0.1276, + "step": 29113 + }, + { + "epoch": 0.6824780088726361, + "grad_norm": 0.19129444658756256, + "learning_rate": 4.5780245842575056e-05, + "loss": 0.0431, + "step": 29114 + }, + { + "epoch": 0.682501450447441, + "grad_norm": 0.4799512028694153, + "learning_rate": 4.577405656783205e-05, + "loss": 0.1059, + "step": 29115 + }, + { + "epoch": 0.6825248920222461, + "grad_norm": 0.6245864629745483, + "learning_rate": 4.576786758732051e-05, + "loss": 0.1275, + "step": 29116 + }, + { + "epoch": 0.682548333597051, + "grad_norm": 0.17141595482826233, + "learning_rate": 4.576167890107408e-05, + "loss": 0.0338, + "step": 29117 + }, + { + "epoch": 0.6825717751718561, + "grad_norm": 0.41049718856811523, + "learning_rate": 4.575549050912631e-05, + "loss": 0.0625, + "step": 29118 + }, + { + "epoch": 0.682595216746661, + "grad_norm": 0.1576731652021408, + "learning_rate": 4.574930241151076e-05, + "loss": 0.0221, + "step": 29119 + }, + { + "epoch": 0.6826186583214661, + "grad_norm": 0.16860376298427582, + "learning_rate": 4.574311460826103e-05, + "loss": 0.0299, + "step": 29120 + }, + { + "epoch": 0.682642099896271, + "grad_norm": 0.45095306634902954, + "learning_rate": 4.5736927099410687e-05, + "loss": 0.0988, + "step": 29121 + }, + { + "epoch": 0.682665541471076, + "grad_norm": 0.14751183986663818, + "learning_rate": 4.573073988499327e-05, + "loss": 0.0195, + "step": 29122 + }, + { + "epoch": 0.682688983045881, + "grad_norm": 0.5606047511100769, + "learning_rate": 4.572455296504238e-05, + "loss": 0.0773, + "step": 29123 + }, + { + "epoch": 0.682712424620686, + "grad_norm": 0.5620273947715759, + "learning_rate": 4.571836633959162e-05, + "loss": 0.0781, + "step": 29124 + }, + { + "epoch": 0.682735866195491, + "grad_norm": 0.4506196081638336, + "learning_rate": 4.571218000867453e-05, + "loss": 0.0955, + "step": 29125 + }, + { + "epoch": 0.682759307770296, + "grad_norm": 0.5011918544769287, + "learning_rate": 4.570599397232469e-05, + "loss": 0.0738, + "step": 29126 + }, + { + "epoch": 0.682782749345101, + "grad_norm": 0.43005385994911194, + "learning_rate": 4.5699808230575644e-05, + "loss": 0.0937, + "step": 29127 + }, + { + "epoch": 0.682806190919906, + "grad_norm": 0.6142768263816833, + "learning_rate": 4.5693622783460976e-05, + "loss": 0.1097, + "step": 29128 + }, + { + "epoch": 0.682829632494711, + "grad_norm": 0.31333446502685547, + "learning_rate": 4.5687437631014196e-05, + "loss": 0.0599, + "step": 29129 + }, + { + "epoch": 0.682853074069516, + "grad_norm": 0.47678935527801514, + "learning_rate": 4.568125277326895e-05, + "loss": 0.083, + "step": 29130 + }, + { + "epoch": 0.682876515644321, + "grad_norm": 0.3316918909549713, + "learning_rate": 4.567506821025874e-05, + "loss": 0.058, + "step": 29131 + }, + { + "epoch": 0.682899957219126, + "grad_norm": 0.36626124382019043, + "learning_rate": 4.566888394201715e-05, + "loss": 0.0413, + "step": 29132 + }, + { + "epoch": 0.6829233987939309, + "grad_norm": 0.48469430208206177, + "learning_rate": 4.566269996857771e-05, + "loss": 0.0636, + "step": 29133 + }, + { + "epoch": 0.682946840368736, + "grad_norm": 0.360711008310318, + "learning_rate": 4.565651628997396e-05, + "loss": 0.0317, + "step": 29134 + }, + { + "epoch": 0.6829702819435409, + "grad_norm": 0.562394380569458, + "learning_rate": 4.565033290623951e-05, + "loss": 0.1296, + "step": 29135 + }, + { + "epoch": 0.682993723518346, + "grad_norm": 0.17354555428028107, + "learning_rate": 4.5644149817407854e-05, + "loss": 0.0341, + "step": 29136 + }, + { + "epoch": 0.6830171650931509, + "grad_norm": 0.35247063636779785, + "learning_rate": 4.563796702351261e-05, + "loss": 0.0475, + "step": 29137 + }, + { + "epoch": 0.683040606667956, + "grad_norm": 0.3441675007343292, + "learning_rate": 4.563178452458727e-05, + "loss": 0.0316, + "step": 29138 + }, + { + "epoch": 0.6830640482427609, + "grad_norm": 0.10461518913507462, + "learning_rate": 4.562560232066541e-05, + "loss": 0.0257, + "step": 29139 + }, + { + "epoch": 0.683087489817566, + "grad_norm": 0.18513987958431244, + "learning_rate": 4.5619420411780555e-05, + "loss": 0.0393, + "step": 29140 + }, + { + "epoch": 0.6831109313923709, + "grad_norm": 0.5110247135162354, + "learning_rate": 4.561323879796621e-05, + "loss": 0.6031, + "step": 29141 + }, + { + "epoch": 0.6831343729671759, + "grad_norm": 0.4281167685985565, + "learning_rate": 4.560705747925601e-05, + "loss": 0.0912, + "step": 29142 + }, + { + "epoch": 0.6831578145419809, + "grad_norm": 0.6182174682617188, + "learning_rate": 4.560087645568344e-05, + "loss": 0.1462, + "step": 29143 + }, + { + "epoch": 0.6831812561167859, + "grad_norm": 0.4982452690601349, + "learning_rate": 4.5594695727282043e-05, + "loss": 0.0916, + "step": 29144 + }, + { + "epoch": 0.6832046976915909, + "grad_norm": 0.30287986993789673, + "learning_rate": 4.5588515294085353e-05, + "loss": 0.0472, + "step": 29145 + }, + { + "epoch": 0.6832281392663959, + "grad_norm": 0.3830949664115906, + "learning_rate": 4.5582335156126876e-05, + "loss": 0.0668, + "step": 29146 + }, + { + "epoch": 0.683251580841201, + "grad_norm": 0.21298442780971527, + "learning_rate": 4.5576155313440215e-05, + "loss": 0.0375, + "step": 29147 + }, + { + "epoch": 0.6832750224160059, + "grad_norm": 0.28823885321617126, + "learning_rate": 4.5569975766058824e-05, + "loss": 0.0501, + "step": 29148 + }, + { + "epoch": 0.683298463990811, + "grad_norm": 0.8202692866325378, + "learning_rate": 4.556379651401632e-05, + "loss": 0.0785, + "step": 29149 + }, + { + "epoch": 0.6833219055656159, + "grad_norm": 0.4361020624637604, + "learning_rate": 4.555761755734619e-05, + "loss": 0.098, + "step": 29150 + }, + { + "epoch": 0.683345347140421, + "grad_norm": 0.19310812652111053, + "learning_rate": 4.555143889608195e-05, + "loss": 0.0277, + "step": 29151 + }, + { + "epoch": 0.6833687887152259, + "grad_norm": 0.09135040640830994, + "learning_rate": 4.5545260530257136e-05, + "loss": 0.0144, + "step": 29152 + }, + { + "epoch": 0.6833922302900309, + "grad_norm": 0.6568027138710022, + "learning_rate": 4.553908245990527e-05, + "loss": 0.5637, + "step": 29153 + }, + { + "epoch": 0.6834156718648359, + "grad_norm": 0.44817832112312317, + "learning_rate": 4.553290468505983e-05, + "loss": 0.0926, + "step": 29154 + }, + { + "epoch": 0.6834391134396409, + "grad_norm": 0.7181469798088074, + "learning_rate": 4.5526727205754426e-05, + "loss": 0.0881, + "step": 29155 + }, + { + "epoch": 0.6834625550144459, + "grad_norm": 0.4164294898509979, + "learning_rate": 4.5520550022022534e-05, + "loss": 0.6088, + "step": 29156 + }, + { + "epoch": 0.6834859965892509, + "grad_norm": 0.41186395287513733, + "learning_rate": 4.551437313389766e-05, + "loss": 0.0907, + "step": 29157 + }, + { + "epoch": 0.6835094381640558, + "grad_norm": 0.369920551776886, + "learning_rate": 4.55081965414133e-05, + "loss": 0.0623, + "step": 29158 + }, + { + "epoch": 0.6835328797388609, + "grad_norm": 0.31138649582862854, + "learning_rate": 4.550202024460303e-05, + "loss": 0.0228, + "step": 29159 + }, + { + "epoch": 0.6835563213136658, + "grad_norm": 0.16405680775642395, + "learning_rate": 4.549584424350033e-05, + "loss": 0.0266, + "step": 29160 + }, + { + "epoch": 0.6835797628884709, + "grad_norm": 0.49508339166641235, + "learning_rate": 4.548966853813867e-05, + "loss": 0.1106, + "step": 29161 + }, + { + "epoch": 0.6836032044632758, + "grad_norm": 0.3997266888618469, + "learning_rate": 4.5483493128551656e-05, + "loss": 0.0456, + "step": 29162 + }, + { + "epoch": 0.6836266460380809, + "grad_norm": 0.5037578344345093, + "learning_rate": 4.547731801477273e-05, + "loss": 0.0975, + "step": 29163 + }, + { + "epoch": 0.6836500876128858, + "grad_norm": 0.44831401109695435, + "learning_rate": 4.547114319683541e-05, + "loss": 0.0912, + "step": 29164 + }, + { + "epoch": 0.6836735291876909, + "grad_norm": 0.1772632747888565, + "learning_rate": 4.5464968674773204e-05, + "loss": 0.0267, + "step": 29165 + }, + { + "epoch": 0.6836969707624958, + "grad_norm": 0.20691218972206116, + "learning_rate": 4.54587944486196e-05, + "loss": 0.0405, + "step": 29166 + }, + { + "epoch": 0.6837204123373009, + "grad_norm": 0.38522934913635254, + "learning_rate": 4.5452620518408084e-05, + "loss": 0.0646, + "step": 29167 + }, + { + "epoch": 0.6837438539121058, + "grad_norm": 0.4106585681438446, + "learning_rate": 4.544644688417221e-05, + "loss": 0.0993, + "step": 29168 + }, + { + "epoch": 0.6837672954869108, + "grad_norm": 0.5580462217330933, + "learning_rate": 4.544027354594547e-05, + "loss": 0.1381, + "step": 29169 + }, + { + "epoch": 0.6837907370617158, + "grad_norm": 0.41466957330703735, + "learning_rate": 4.5434100503761325e-05, + "loss": 0.0456, + "step": 29170 + }, + { + "epoch": 0.6838141786365208, + "grad_norm": 0.34591907262802124, + "learning_rate": 4.5427927757653244e-05, + "loss": 0.0694, + "step": 29171 + }, + { + "epoch": 0.6838376202113258, + "grad_norm": 0.78836590051651, + "learning_rate": 4.5421755307654804e-05, + "loss": 0.4738, + "step": 29172 + }, + { + "epoch": 0.6838610617861308, + "grad_norm": 0.23362977802753448, + "learning_rate": 4.5415583153799444e-05, + "loss": 0.0236, + "step": 29173 + }, + { + "epoch": 0.6838845033609358, + "grad_norm": 0.3306083381175995, + "learning_rate": 4.540941129612064e-05, + "loss": 0.0831, + "step": 29174 + }, + { + "epoch": 0.6839079449357408, + "grad_norm": 0.48035794496536255, + "learning_rate": 4.540323973465195e-05, + "loss": 0.0409, + "step": 29175 + }, + { + "epoch": 0.6839313865105457, + "grad_norm": 0.582676112651825, + "learning_rate": 4.5397068469426796e-05, + "loss": 0.6109, + "step": 29176 + }, + { + "epoch": 0.6839548280853508, + "grad_norm": 0.5170612335205078, + "learning_rate": 4.5390897500478705e-05, + "loss": 0.6434, + "step": 29177 + }, + { + "epoch": 0.6839782696601557, + "grad_norm": 0.4435722827911377, + "learning_rate": 4.5384726827841116e-05, + "loss": 0.0957, + "step": 29178 + }, + { + "epoch": 0.6840017112349608, + "grad_norm": 0.7218664884567261, + "learning_rate": 4.537855645154752e-05, + "loss": 0.1137, + "step": 29179 + }, + { + "epoch": 0.6840251528097657, + "grad_norm": 1.373054027557373, + "learning_rate": 4.537238637163144e-05, + "loss": 0.1477, + "step": 29180 + }, + { + "epoch": 0.6840485943845708, + "grad_norm": 0.10940413177013397, + "learning_rate": 4.536621658812632e-05, + "loss": 0.0084, + "step": 29181 + }, + { + "epoch": 0.6840720359593757, + "grad_norm": 0.3649160861968994, + "learning_rate": 4.536004710106565e-05, + "loss": 0.0689, + "step": 29182 + }, + { + "epoch": 0.6840954775341808, + "grad_norm": 0.3681482672691345, + "learning_rate": 4.535387791048287e-05, + "loss": 0.0638, + "step": 29183 + }, + { + "epoch": 0.6841189191089857, + "grad_norm": 0.37270867824554443, + "learning_rate": 4.534770901641152e-05, + "loss": 0.0785, + "step": 29184 + }, + { + "epoch": 0.6841423606837908, + "grad_norm": 0.29285943508148193, + "learning_rate": 4.534154041888503e-05, + "loss": 0.0401, + "step": 29185 + }, + { + "epoch": 0.6841658022585957, + "grad_norm": 0.3598034679889679, + "learning_rate": 4.5335372117936846e-05, + "loss": 0.0413, + "step": 29186 + }, + { + "epoch": 0.6841892438334007, + "grad_norm": 0.1936706006526947, + "learning_rate": 4.532920411360051e-05, + "loss": 0.0396, + "step": 29187 + }, + { + "epoch": 0.6842126854082057, + "grad_norm": 0.5144121050834656, + "learning_rate": 4.532303640590945e-05, + "loss": 0.0837, + "step": 29188 + }, + { + "epoch": 0.6842361269830107, + "grad_norm": 0.4930986166000366, + "learning_rate": 4.531686899489713e-05, + "loss": 0.0466, + "step": 29189 + }, + { + "epoch": 0.6842595685578157, + "grad_norm": 0.5526109337806702, + "learning_rate": 4.5310701880597017e-05, + "loss": 0.7038, + "step": 29190 + }, + { + "epoch": 0.6842830101326207, + "grad_norm": 0.3350347876548767, + "learning_rate": 4.530453506304258e-05, + "loss": 0.0563, + "step": 29191 + }, + { + "epoch": 0.6843064517074257, + "grad_norm": 0.30522581934928894, + "learning_rate": 4.529836854226722e-05, + "loss": 0.0531, + "step": 29192 + }, + { + "epoch": 0.6843298932822307, + "grad_norm": 0.2081710696220398, + "learning_rate": 4.52922023183045e-05, + "loss": 0.0453, + "step": 29193 + }, + { + "epoch": 0.6843533348570356, + "grad_norm": 0.26584964990615845, + "learning_rate": 4.528603639118781e-05, + "loss": 0.025, + "step": 29194 + }, + { + "epoch": 0.6843767764318407, + "grad_norm": 0.4257507920265198, + "learning_rate": 4.527987076095061e-05, + "loss": 0.0511, + "step": 29195 + }, + { + "epoch": 0.6844002180066456, + "grad_norm": 0.414703369140625, + "learning_rate": 4.52737054276264e-05, + "loss": 0.0549, + "step": 29196 + }, + { + "epoch": 0.6844236595814507, + "grad_norm": 0.14041554927825928, + "learning_rate": 4.526754039124861e-05, + "loss": 0.025, + "step": 29197 + }, + { + "epoch": 0.6844471011562556, + "grad_norm": 0.3185497224330902, + "learning_rate": 4.526137565185067e-05, + "loss": 0.0609, + "step": 29198 + }, + { + "epoch": 0.6844705427310607, + "grad_norm": 0.23196056485176086, + "learning_rate": 4.525521120946602e-05, + "loss": 0.0352, + "step": 29199 + }, + { + "epoch": 0.6844939843058657, + "grad_norm": 0.2686256766319275, + "learning_rate": 4.524904706412816e-05, + "loss": 0.0162, + "step": 29200 + }, + { + "epoch": 0.6845174258806707, + "grad_norm": 0.4828563332557678, + "learning_rate": 4.5242883215870514e-05, + "loss": 0.1278, + "step": 29201 + }, + { + "epoch": 0.6845408674554757, + "grad_norm": 0.6989254355430603, + "learning_rate": 4.5236719664726526e-05, + "loss": 0.6242, + "step": 29202 + }, + { + "epoch": 0.6845643090302806, + "grad_norm": 0.5164340734481812, + "learning_rate": 4.5230556410729644e-05, + "loss": 0.0614, + "step": 29203 + }, + { + "epoch": 0.6845877506050857, + "grad_norm": 0.229746013879776, + "learning_rate": 4.522439345391328e-05, + "loss": 0.0345, + "step": 29204 + }, + { + "epoch": 0.6846111921798906, + "grad_norm": 0.5521003603935242, + "learning_rate": 4.521823079431088e-05, + "loss": 0.6246, + "step": 29205 + }, + { + "epoch": 0.6846346337546957, + "grad_norm": 0.2505420446395874, + "learning_rate": 4.521206843195594e-05, + "loss": 0.0315, + "step": 29206 + }, + { + "epoch": 0.6846580753295006, + "grad_norm": 0.580730676651001, + "learning_rate": 4.5205906366881815e-05, + "loss": 0.1448, + "step": 29207 + }, + { + "epoch": 0.6846815169043057, + "grad_norm": 0.28776460886001587, + "learning_rate": 4.519974459912201e-05, + "loss": 0.0897, + "step": 29208 + }, + { + "epoch": 0.6847049584791106, + "grad_norm": 0.4027746319770813, + "learning_rate": 4.519358312870994e-05, + "loss": 0.0792, + "step": 29209 + }, + { + "epoch": 0.6847284000539157, + "grad_norm": 0.39248913526535034, + "learning_rate": 4.518742195567902e-05, + "loss": 0.0883, + "step": 29210 + }, + { + "epoch": 0.6847518416287206, + "grad_norm": 0.17686127126216888, + "learning_rate": 4.5181261080062666e-05, + "loss": 0.0308, + "step": 29211 + }, + { + "epoch": 0.6847752832035257, + "grad_norm": 0.16379784047603607, + "learning_rate": 4.5175100501894364e-05, + "loss": 0.0247, + "step": 29212 + }, + { + "epoch": 0.6847987247783306, + "grad_norm": 0.3200197219848633, + "learning_rate": 4.51689402212075e-05, + "loss": 0.0543, + "step": 29213 + }, + { + "epoch": 0.6848221663531356, + "grad_norm": 0.3867306113243103, + "learning_rate": 4.5162780238035516e-05, + "loss": 0.0822, + "step": 29214 + }, + { + "epoch": 0.6848456079279406, + "grad_norm": 0.575684666633606, + "learning_rate": 4.515662055241183e-05, + "loss": 0.0883, + "step": 29215 + }, + { + "epoch": 0.6848690495027456, + "grad_norm": 0.3243056535720825, + "learning_rate": 4.515046116436986e-05, + "loss": 0.0685, + "step": 29216 + }, + { + "epoch": 0.6848924910775506, + "grad_norm": 0.5972105860710144, + "learning_rate": 4.514430207394299e-05, + "loss": 0.0403, + "step": 29217 + }, + { + "epoch": 0.6849159326523556, + "grad_norm": 0.5310975313186646, + "learning_rate": 4.513814328116471e-05, + "loss": 0.0401, + "step": 29218 + }, + { + "epoch": 0.6849393742271606, + "grad_norm": 0.2345496267080307, + "learning_rate": 4.5131984786068385e-05, + "loss": 0.0263, + "step": 29219 + }, + { + "epoch": 0.6849628158019656, + "grad_norm": 0.3526085317134857, + "learning_rate": 4.5125826588687484e-05, + "loss": 0.0539, + "step": 29220 + }, + { + "epoch": 0.6849862573767705, + "grad_norm": 0.5664432644844055, + "learning_rate": 4.5119668689055394e-05, + "loss": 0.1022, + "step": 29221 + }, + { + "epoch": 0.6850096989515756, + "grad_norm": 0.4877135455608368, + "learning_rate": 4.5113511087205516e-05, + "loss": 0.5609, + "step": 29222 + }, + { + "epoch": 0.6850331405263805, + "grad_norm": 0.47220322489738464, + "learning_rate": 4.510735378317128e-05, + "loss": 0.0791, + "step": 29223 + }, + { + "epoch": 0.6850565821011856, + "grad_norm": 0.5424910187721252, + "learning_rate": 4.510119677698604e-05, + "loss": 0.0856, + "step": 29224 + }, + { + "epoch": 0.6850800236759905, + "grad_norm": 0.5618336200714111, + "learning_rate": 4.509504006868329e-05, + "loss": 0.0865, + "step": 29225 + }, + { + "epoch": 0.6851034652507956, + "grad_norm": 0.2830384373664856, + "learning_rate": 4.508888365829641e-05, + "loss": 0.0308, + "step": 29226 + }, + { + "epoch": 0.6851269068256005, + "grad_norm": 0.47575902938842773, + "learning_rate": 4.508272754585877e-05, + "loss": 0.0556, + "step": 29227 + }, + { + "epoch": 0.6851503484004056, + "grad_norm": 0.9180028438568115, + "learning_rate": 4.5076571731403804e-05, + "loss": 0.146, + "step": 29228 + }, + { + "epoch": 0.6851737899752105, + "grad_norm": 0.20985481142997742, + "learning_rate": 4.5070416214964905e-05, + "loss": 0.0457, + "step": 29229 + }, + { + "epoch": 0.6851972315500156, + "grad_norm": 0.2552390396595001, + "learning_rate": 4.5064260996575434e-05, + "loss": 0.0273, + "step": 29230 + }, + { + "epoch": 0.6852206731248205, + "grad_norm": 0.5313352942466736, + "learning_rate": 4.505810607626884e-05, + "loss": 0.1431, + "step": 29231 + }, + { + "epoch": 0.6852441146996255, + "grad_norm": 0.28444117307662964, + "learning_rate": 4.505195145407852e-05, + "loss": 0.0506, + "step": 29232 + }, + { + "epoch": 0.6852675562744305, + "grad_norm": 0.3579305112361908, + "learning_rate": 4.504579713003787e-05, + "loss": 0.0881, + "step": 29233 + }, + { + "epoch": 0.6852909978492355, + "grad_norm": 0.42777204513549805, + "learning_rate": 4.5039643104180274e-05, + "loss": 0.0588, + "step": 29234 + }, + { + "epoch": 0.6853144394240405, + "grad_norm": 0.47971054911613464, + "learning_rate": 4.50334893765391e-05, + "loss": 0.0501, + "step": 29235 + }, + { + "epoch": 0.6853378809988455, + "grad_norm": 0.5934578776359558, + "learning_rate": 4.5027335947147773e-05, + "loss": 0.0736, + "step": 29236 + }, + { + "epoch": 0.6853613225736505, + "grad_norm": 0.4060242772102356, + "learning_rate": 4.502118281603963e-05, + "loss": 0.0899, + "step": 29237 + }, + { + "epoch": 0.6853847641484555, + "grad_norm": 0.3774917721748352, + "learning_rate": 4.501502998324814e-05, + "loss": 0.0535, + "step": 29238 + }, + { + "epoch": 0.6854082057232604, + "grad_norm": 0.33791685104370117, + "learning_rate": 4.500887744880663e-05, + "loss": 0.0641, + "step": 29239 + }, + { + "epoch": 0.6854316472980655, + "grad_norm": 0.1477973461151123, + "learning_rate": 4.5002725212748497e-05, + "loss": 0.0135, + "step": 29240 + }, + { + "epoch": 0.6854550888728704, + "grad_norm": 0.049793362617492676, + "learning_rate": 4.4996573275107135e-05, + "loss": 0.0061, + "step": 29241 + }, + { + "epoch": 0.6854785304476755, + "grad_norm": 0.5972795486450195, + "learning_rate": 4.49904216359159e-05, + "loss": 0.1268, + "step": 29242 + }, + { + "epoch": 0.6855019720224804, + "grad_norm": 0.27322444319725037, + "learning_rate": 4.498427029520815e-05, + "loss": 0.0395, + "step": 29243 + }, + { + "epoch": 0.6855254135972855, + "grad_norm": 0.4308645725250244, + "learning_rate": 4.49781192530173e-05, + "loss": 0.0443, + "step": 29244 + }, + { + "epoch": 0.6855488551720904, + "grad_norm": 0.14280541241168976, + "learning_rate": 4.497196850937676e-05, + "loss": 0.0239, + "step": 29245 + }, + { + "epoch": 0.6855722967468955, + "grad_norm": 0.48439937829971313, + "learning_rate": 4.4965818064319865e-05, + "loss": 0.6954, + "step": 29246 + }, + { + "epoch": 0.6855957383217004, + "grad_norm": 0.5200634598731995, + "learning_rate": 4.495966791787999e-05, + "loss": 0.0484, + "step": 29247 + }, + { + "epoch": 0.6856191798965054, + "grad_norm": 0.43969449400901794, + "learning_rate": 4.4953518070090504e-05, + "loss": 0.0831, + "step": 29248 + }, + { + "epoch": 0.6856426214713104, + "grad_norm": 0.7140870094299316, + "learning_rate": 4.494736852098473e-05, + "loss": 0.1084, + "step": 29249 + }, + { + "epoch": 0.6856660630461154, + "grad_norm": 0.17013543844223022, + "learning_rate": 4.4941219270596135e-05, + "loss": 0.016, + "step": 29250 + }, + { + "epoch": 0.6856895046209205, + "grad_norm": 0.22055095434188843, + "learning_rate": 4.493507031895803e-05, + "loss": 0.032, + "step": 29251 + }, + { + "epoch": 0.6857129461957254, + "grad_norm": 0.4551583528518677, + "learning_rate": 4.4928921666103776e-05, + "loss": 0.0554, + "step": 29252 + }, + { + "epoch": 0.6857363877705305, + "grad_norm": 0.19443614780902863, + "learning_rate": 4.4922773312066734e-05, + "loss": 0.0308, + "step": 29253 + }, + { + "epoch": 0.6857598293453354, + "grad_norm": 0.49208909273147583, + "learning_rate": 4.491662525688028e-05, + "loss": 0.0851, + "step": 29254 + }, + { + "epoch": 0.6857832709201405, + "grad_norm": 0.5243831276893616, + "learning_rate": 4.491047750057773e-05, + "loss": 0.4942, + "step": 29255 + }, + { + "epoch": 0.6858067124949454, + "grad_norm": 0.0804833397269249, + "learning_rate": 4.490433004319248e-05, + "loss": 0.0052, + "step": 29256 + }, + { + "epoch": 0.6858301540697505, + "grad_norm": 0.6872670650482178, + "learning_rate": 4.4898182884757925e-05, + "loss": 0.0961, + "step": 29257 + }, + { + "epoch": 0.6858535956445554, + "grad_norm": 0.5979205965995789, + "learning_rate": 4.4892036025307374e-05, + "loss": 0.095, + "step": 29258 + }, + { + "epoch": 0.6858770372193604, + "grad_norm": 0.7526393532752991, + "learning_rate": 4.488588946487418e-05, + "loss": 0.0277, + "step": 29259 + }, + { + "epoch": 0.6859004787941654, + "grad_norm": 0.3538334369659424, + "learning_rate": 4.48797432034917e-05, + "loss": 0.0362, + "step": 29260 + }, + { + "epoch": 0.6859239203689704, + "grad_norm": 0.08931272476911545, + "learning_rate": 4.487359724119329e-05, + "loss": 0.0141, + "step": 29261 + }, + { + "epoch": 0.6859473619437754, + "grad_norm": 0.24243859946727753, + "learning_rate": 4.4867451578012244e-05, + "loss": 0.0669, + "step": 29262 + }, + { + "epoch": 0.6859708035185804, + "grad_norm": 1.4319649934768677, + "learning_rate": 4.4861306213982e-05, + "loss": 0.0836, + "step": 29263 + }, + { + "epoch": 0.6859942450933854, + "grad_norm": 0.37006041407585144, + "learning_rate": 4.485516114913586e-05, + "loss": 0.0565, + "step": 29264 + }, + { + "epoch": 0.6860176866681904, + "grad_norm": 0.31019511818885803, + "learning_rate": 4.484901638350717e-05, + "loss": 0.0718, + "step": 29265 + }, + { + "epoch": 0.6860411282429953, + "grad_norm": 0.44857001304626465, + "learning_rate": 4.484287191712926e-05, + "loss": 0.0734, + "step": 29266 + }, + { + "epoch": 0.6860645698178004, + "grad_norm": 0.5921787023544312, + "learning_rate": 4.483672775003548e-05, + "loss": 0.1221, + "step": 29267 + }, + { + "epoch": 0.6860880113926053, + "grad_norm": 0.2734074592590332, + "learning_rate": 4.483058388225913e-05, + "loss": 0.0347, + "step": 29268 + }, + { + "epoch": 0.6861114529674104, + "grad_norm": 0.42395374178886414, + "learning_rate": 4.482444031383358e-05, + "loss": 0.073, + "step": 29269 + }, + { + "epoch": 0.6861348945422153, + "grad_norm": 0.6032238602638245, + "learning_rate": 4.481829704479221e-05, + "loss": 0.5856, + "step": 29270 + }, + { + "epoch": 0.6861583361170204, + "grad_norm": 0.13623443245887756, + "learning_rate": 4.481215407516831e-05, + "loss": 0.0314, + "step": 29271 + }, + { + "epoch": 0.6861817776918253, + "grad_norm": 0.4111020565032959, + "learning_rate": 4.480601140499522e-05, + "loss": 0.0811, + "step": 29272 + }, + { + "epoch": 0.6862052192666304, + "grad_norm": 0.17926737666130066, + "learning_rate": 4.479986903430625e-05, + "loss": 0.0361, + "step": 29273 + }, + { + "epoch": 0.6862286608414353, + "grad_norm": 0.5506975054740906, + "learning_rate": 4.479372696313475e-05, + "loss": 0.0966, + "step": 29274 + }, + { + "epoch": 0.6862521024162404, + "grad_norm": 0.5360279083251953, + "learning_rate": 4.4787585191514004e-05, + "loss": 0.0708, + "step": 29275 + }, + { + "epoch": 0.6862755439910453, + "grad_norm": 0.44588273763656616, + "learning_rate": 4.478144371947741e-05, + "loss": 0.0841, + "step": 29276 + }, + { + "epoch": 0.6862989855658503, + "grad_norm": 0.30954182147979736, + "learning_rate": 4.4775302547058265e-05, + "loss": 0.0513, + "step": 29277 + }, + { + "epoch": 0.6863224271406553, + "grad_norm": 0.4519312381744385, + "learning_rate": 4.476916167428986e-05, + "loss": 0.0804, + "step": 29278 + }, + { + "epoch": 0.6863458687154603, + "grad_norm": 0.29472073912620544, + "learning_rate": 4.476302110120555e-05, + "loss": 0.0423, + "step": 29279 + }, + { + "epoch": 0.6863693102902653, + "grad_norm": 0.5852122902870178, + "learning_rate": 4.47568808278386e-05, + "loss": 0.1012, + "step": 29280 + }, + { + "epoch": 0.6863927518650703, + "grad_norm": 0.32495054602622986, + "learning_rate": 4.475074085422241e-05, + "loss": 0.0718, + "step": 29281 + }, + { + "epoch": 0.6864161934398753, + "grad_norm": 0.18736755847930908, + "learning_rate": 4.474460118039021e-05, + "loss": 0.0311, + "step": 29282 + }, + { + "epoch": 0.6864396350146803, + "grad_norm": 0.701408326625824, + "learning_rate": 4.47384618063754e-05, + "loss": 0.1223, + "step": 29283 + }, + { + "epoch": 0.6864630765894852, + "grad_norm": 0.28081685304641724, + "learning_rate": 4.473232273221124e-05, + "loss": 0.0493, + "step": 29284 + }, + { + "epoch": 0.6864865181642903, + "grad_norm": 0.6082607507705688, + "learning_rate": 4.472618395793107e-05, + "loss": 0.1384, + "step": 29285 + }, + { + "epoch": 0.6865099597390952, + "grad_norm": 0.11642293632030487, + "learning_rate": 4.4720045483568164e-05, + "loss": 0.027, + "step": 29286 + }, + { + "epoch": 0.6865334013139003, + "grad_norm": 0.2716987431049347, + "learning_rate": 4.471390730915581e-05, + "loss": 0.0366, + "step": 29287 + }, + { + "epoch": 0.6865568428887052, + "grad_norm": 0.5711411833763123, + "learning_rate": 4.470776943472739e-05, + "loss": 0.0986, + "step": 29288 + }, + { + "epoch": 0.6865802844635103, + "grad_norm": 0.4183114469051361, + "learning_rate": 4.470163186031616e-05, + "loss": 0.0897, + "step": 29289 + }, + { + "epoch": 0.6866037260383152, + "grad_norm": 0.35605525970458984, + "learning_rate": 4.469549458595545e-05, + "loss": 0.0489, + "step": 29290 + }, + { + "epoch": 0.6866271676131203, + "grad_norm": 0.025626540184020996, + "learning_rate": 4.4689357611678516e-05, + "loss": 0.002, + "step": 29291 + }, + { + "epoch": 0.6866506091879252, + "grad_norm": 0.11590144783258438, + "learning_rate": 4.4683220937518666e-05, + "loss": 0.0168, + "step": 29292 + }, + { + "epoch": 0.6866740507627302, + "grad_norm": 0.5835863351821899, + "learning_rate": 4.467708456350924e-05, + "loss": 0.1066, + "step": 29293 + }, + { + "epoch": 0.6866974923375352, + "grad_norm": 0.15672865509986877, + "learning_rate": 4.467094848968347e-05, + "loss": 0.0184, + "step": 29294 + }, + { + "epoch": 0.6867209339123402, + "grad_norm": 0.17852796614170074, + "learning_rate": 4.4664812716074734e-05, + "loss": 0.0373, + "step": 29295 + }, + { + "epoch": 0.6867443754871452, + "grad_norm": 0.4606982469558716, + "learning_rate": 4.4658677242716285e-05, + "loss": 0.0638, + "step": 29296 + }, + { + "epoch": 0.6867678170619502, + "grad_norm": 0.4016520380973816, + "learning_rate": 4.4652542069641404e-05, + "loss": 0.0688, + "step": 29297 + }, + { + "epoch": 0.6867912586367552, + "grad_norm": 0.32678112387657166, + "learning_rate": 4.4646407196883375e-05, + "loss": 0.043, + "step": 29298 + }, + { + "epoch": 0.6868147002115602, + "grad_norm": 0.5063546895980835, + "learning_rate": 4.464027262447551e-05, + "loss": 0.0848, + "step": 29299 + }, + { + "epoch": 0.6868381417863652, + "grad_norm": 0.1909504383802414, + "learning_rate": 4.463413835245104e-05, + "loss": 0.0352, + "step": 29300 + }, + { + "epoch": 0.6868615833611702, + "grad_norm": 0.3907061815261841, + "learning_rate": 4.462800438084334e-05, + "loss": 0.0552, + "step": 29301 + }, + { + "epoch": 0.6868850249359753, + "grad_norm": 0.23320356011390686, + "learning_rate": 4.462187070968563e-05, + "loss": 0.0258, + "step": 29302 + }, + { + "epoch": 0.6869084665107802, + "grad_norm": 0.13269157707691193, + "learning_rate": 4.461573733901121e-05, + "loss": 0.0241, + "step": 29303 + }, + { + "epoch": 0.6869319080855852, + "grad_norm": 0.38109180331230164, + "learning_rate": 4.4609604268853324e-05, + "loss": 0.0853, + "step": 29304 + }, + { + "epoch": 0.6869553496603902, + "grad_norm": 0.17463219165802002, + "learning_rate": 4.460347149924532e-05, + "loss": 0.0438, + "step": 29305 + }, + { + "epoch": 0.6869787912351952, + "grad_norm": 0.7676807045936584, + "learning_rate": 4.459733903022043e-05, + "loss": 0.1796, + "step": 29306 + }, + { + "epoch": 0.6870022328100002, + "grad_norm": 0.1762419044971466, + "learning_rate": 4.459120686181191e-05, + "loss": 0.0336, + "step": 29307 + }, + { + "epoch": 0.6870256743848052, + "grad_norm": 0.4967581331729889, + "learning_rate": 4.458507499405309e-05, + "loss": 0.0779, + "step": 29308 + }, + { + "epoch": 0.6870491159596102, + "grad_norm": 0.26839715242385864, + "learning_rate": 4.4578943426977205e-05, + "loss": 0.0621, + "step": 29309 + }, + { + "epoch": 0.6870725575344152, + "grad_norm": 0.26922571659088135, + "learning_rate": 4.457281216061754e-05, + "loss": 0.0474, + "step": 29310 + }, + { + "epoch": 0.6870959991092201, + "grad_norm": 0.5123644471168518, + "learning_rate": 4.456668119500736e-05, + "loss": 0.081, + "step": 29311 + }, + { + "epoch": 0.6871194406840252, + "grad_norm": 0.49943307042121887, + "learning_rate": 4.456055053017991e-05, + "loss": 0.0373, + "step": 29312 + }, + { + "epoch": 0.6871428822588301, + "grad_norm": 0.515767514705658, + "learning_rate": 4.455442016616844e-05, + "loss": 0.0813, + "step": 29313 + }, + { + "epoch": 0.6871663238336352, + "grad_norm": 0.34426724910736084, + "learning_rate": 4.4548290103006295e-05, + "loss": 0.0315, + "step": 29314 + }, + { + "epoch": 0.6871897654084401, + "grad_norm": 0.1645120233297348, + "learning_rate": 4.454216034072668e-05, + "loss": 0.0399, + "step": 29315 + }, + { + "epoch": 0.6872132069832452, + "grad_norm": 0.49393168091773987, + "learning_rate": 4.453603087936282e-05, + "loss": 0.0732, + "step": 29316 + }, + { + "epoch": 0.6872366485580501, + "grad_norm": 0.21911005675792694, + "learning_rate": 4.452990171894805e-05, + "loss": 0.0265, + "step": 29317 + }, + { + "epoch": 0.6872600901328552, + "grad_norm": 0.428117573261261, + "learning_rate": 4.45237728595156e-05, + "loss": 0.0912, + "step": 29318 + }, + { + "epoch": 0.6872835317076601, + "grad_norm": 0.5816209316253662, + "learning_rate": 4.451764430109873e-05, + "loss": 0.4899, + "step": 29319 + }, + { + "epoch": 0.6873069732824652, + "grad_norm": 0.3434212803840637, + "learning_rate": 4.451151604373063e-05, + "loss": 0.047, + "step": 29320 + }, + { + "epoch": 0.6873304148572701, + "grad_norm": 0.31703242659568787, + "learning_rate": 4.4505388087444635e-05, + "loss": 0.0516, + "step": 29321 + }, + { + "epoch": 0.6873538564320751, + "grad_norm": 0.3653956651687622, + "learning_rate": 4.449926043227398e-05, + "loss": 0.0502, + "step": 29322 + }, + { + "epoch": 0.6873772980068801, + "grad_norm": 0.4116315543651581, + "learning_rate": 4.449313307825189e-05, + "loss": 0.0484, + "step": 29323 + }, + { + "epoch": 0.6874007395816851, + "grad_norm": 0.0741182416677475, + "learning_rate": 4.4487006025411625e-05, + "loss": 0.0136, + "step": 29324 + }, + { + "epoch": 0.6874241811564901, + "grad_norm": 0.20149768888950348, + "learning_rate": 4.448087927378638e-05, + "loss": 0.027, + "step": 29325 + }, + { + "epoch": 0.6874476227312951, + "grad_norm": 0.4535369575023651, + "learning_rate": 4.4474752823409495e-05, + "loss": 0.4781, + "step": 29326 + }, + { + "epoch": 0.6874710643061, + "grad_norm": 0.17014092206954956, + "learning_rate": 4.446862667431415e-05, + "loss": 0.0261, + "step": 29327 + }, + { + "epoch": 0.6874945058809051, + "grad_norm": 0.6046413779258728, + "learning_rate": 4.44625008265336e-05, + "loss": 0.0883, + "step": 29328 + }, + { + "epoch": 0.68751794745571, + "grad_norm": 0.08537764102220535, + "learning_rate": 4.445637528010106e-05, + "loss": 0.0106, + "step": 29329 + }, + { + "epoch": 0.6875413890305151, + "grad_norm": 0.4139406383037567, + "learning_rate": 4.445025003504981e-05, + "loss": 0.1309, + "step": 29330 + }, + { + "epoch": 0.68756483060532, + "grad_norm": 0.1959051638841629, + "learning_rate": 4.444412509141307e-05, + "loss": 0.0403, + "step": 29331 + }, + { + "epoch": 0.6875882721801251, + "grad_norm": 0.4131796360015869, + "learning_rate": 4.4438000449224036e-05, + "loss": 0.0558, + "step": 29332 + }, + { + "epoch": 0.68761171375493, + "grad_norm": 0.6394479870796204, + "learning_rate": 4.443187610851601e-05, + "loss": 0.1022, + "step": 29333 + }, + { + "epoch": 0.6876351553297351, + "grad_norm": 0.5731245875358582, + "learning_rate": 4.442575206932218e-05, + "loss": 0.0803, + "step": 29334 + }, + { + "epoch": 0.68765859690454, + "grad_norm": 0.48130515217781067, + "learning_rate": 4.441962833167579e-05, + "loss": 0.0604, + "step": 29335 + }, + { + "epoch": 0.6876820384793451, + "grad_norm": 0.24378913640975952, + "learning_rate": 4.441350489561006e-05, + "loss": 0.0542, + "step": 29336 + }, + { + "epoch": 0.68770548005415, + "grad_norm": 0.5733461976051331, + "learning_rate": 4.4407381761158205e-05, + "loss": 0.0993, + "step": 29337 + }, + { + "epoch": 0.687728921628955, + "grad_norm": 0.06943874061107635, + "learning_rate": 4.4401258928353426e-05, + "loss": 0.0155, + "step": 29338 + }, + { + "epoch": 0.68775236320376, + "grad_norm": 0.22283372282981873, + "learning_rate": 4.439513639722902e-05, + "loss": 0.0175, + "step": 29339 + }, + { + "epoch": 0.687775804778565, + "grad_norm": 0.19175788760185242, + "learning_rate": 4.4389014167818175e-05, + "loss": 0.0455, + "step": 29340 + }, + { + "epoch": 0.68779924635337, + "grad_norm": 0.30002260208129883, + "learning_rate": 4.438289224015405e-05, + "loss": 0.0638, + "step": 29341 + }, + { + "epoch": 0.687822687928175, + "grad_norm": 0.689887523651123, + "learning_rate": 4.437677061426997e-05, + "loss": 0.1614, + "step": 29342 + }, + { + "epoch": 0.68784612950298, + "grad_norm": 0.19346953928470612, + "learning_rate": 4.4370649290199084e-05, + "loss": 0.0186, + "step": 29343 + }, + { + "epoch": 0.687869571077785, + "grad_norm": 0.3758496344089508, + "learning_rate": 4.436452826797462e-05, + "loss": 0.0765, + "step": 29344 + }, + { + "epoch": 0.68789301265259, + "grad_norm": 0.4393666982650757, + "learning_rate": 4.4358407547629756e-05, + "loss": 0.1229, + "step": 29345 + }, + { + "epoch": 0.687916454227395, + "grad_norm": 0.09471824765205383, + "learning_rate": 4.4352287129197776e-05, + "loss": 0.0239, + "step": 29346 + }, + { + "epoch": 0.6879398958021999, + "grad_norm": 0.3840487599372864, + "learning_rate": 4.4346167012711846e-05, + "loss": 0.0895, + "step": 29347 + }, + { + "epoch": 0.687963337377005, + "grad_norm": 0.7972932457923889, + "learning_rate": 4.434004719820518e-05, + "loss": 0.1092, + "step": 29348 + }, + { + "epoch": 0.6879867789518099, + "grad_norm": 0.49068039655685425, + "learning_rate": 4.433392768571098e-05, + "loss": 0.0798, + "step": 29349 + }, + { + "epoch": 0.688010220526615, + "grad_norm": 0.08136012405157089, + "learning_rate": 4.4327808475262456e-05, + "loss": 0.0129, + "step": 29350 + }, + { + "epoch": 0.6880336621014199, + "grad_norm": 0.16841472685337067, + "learning_rate": 4.4321689566892774e-05, + "loss": 0.0303, + "step": 29351 + }, + { + "epoch": 0.688057103676225, + "grad_norm": 0.242139995098114, + "learning_rate": 4.431557096063521e-05, + "loss": 0.0574, + "step": 29352 + }, + { + "epoch": 0.68808054525103, + "grad_norm": 0.6273643374443054, + "learning_rate": 4.430945265652288e-05, + "loss": 0.8073, + "step": 29353 + }, + { + "epoch": 0.688103986825835, + "grad_norm": 0.34240737557411194, + "learning_rate": 4.430333465458907e-05, + "loss": 0.0486, + "step": 29354 + }, + { + "epoch": 0.68812742840064, + "grad_norm": 0.09309925138950348, + "learning_rate": 4.429721695486693e-05, + "loss": 0.0165, + "step": 29355 + }, + { + "epoch": 0.688150869975445, + "grad_norm": 0.2855572998523712, + "learning_rate": 4.429109955738965e-05, + "loss": 0.0399, + "step": 29356 + }, + { + "epoch": 0.68817431155025, + "grad_norm": 0.3167896568775177, + "learning_rate": 4.428498246219043e-05, + "loss": 0.0609, + "step": 29357 + }, + { + "epoch": 0.6881977531250549, + "grad_norm": 0.5010282397270203, + "learning_rate": 4.427886566930243e-05, + "loss": 0.0943, + "step": 29358 + }, + { + "epoch": 0.68822119469986, + "grad_norm": 0.8245481848716736, + "learning_rate": 4.427274917875891e-05, + "loss": 0.1469, + "step": 29359 + }, + { + "epoch": 0.6882446362746649, + "grad_norm": 0.9075501561164856, + "learning_rate": 4.426663299059301e-05, + "loss": 0.1511, + "step": 29360 + }, + { + "epoch": 0.68826807784947, + "grad_norm": 0.06344813108444214, + "learning_rate": 4.426051710483794e-05, + "loss": 0.009, + "step": 29361 + }, + { + "epoch": 0.6882915194242749, + "grad_norm": 0.31860217452049255, + "learning_rate": 4.425440152152687e-05, + "loss": 0.0537, + "step": 29362 + }, + { + "epoch": 0.68831496099908, + "grad_norm": 0.3397076427936554, + "learning_rate": 4.424828624069294e-05, + "loss": 0.0768, + "step": 29363 + }, + { + "epoch": 0.6883384025738849, + "grad_norm": 0.6412818431854248, + "learning_rate": 4.424217126236941e-05, + "loss": 0.0788, + "step": 29364 + }, + { + "epoch": 0.68836184414869, + "grad_norm": 0.16865742206573486, + "learning_rate": 4.4236056586589406e-05, + "loss": 0.0311, + "step": 29365 + }, + { + "epoch": 0.6883852857234949, + "grad_norm": 0.5380172729492188, + "learning_rate": 4.4229942213386156e-05, + "loss": 0.5892, + "step": 29366 + }, + { + "epoch": 0.6884087272982999, + "grad_norm": 0.13724805414676666, + "learning_rate": 4.42238281427928e-05, + "loss": 0.0188, + "step": 29367 + }, + { + "epoch": 0.6884321688731049, + "grad_norm": 0.4884427785873413, + "learning_rate": 4.4217714374842525e-05, + "loss": 0.4363, + "step": 29368 + }, + { + "epoch": 0.6884556104479099, + "grad_norm": 0.26765909790992737, + "learning_rate": 4.4211600909568504e-05, + "loss": 0.0627, + "step": 29369 + }, + { + "epoch": 0.6884790520227149, + "grad_norm": 0.5027700066566467, + "learning_rate": 4.420548774700387e-05, + "loss": 0.7099, + "step": 29370 + }, + { + "epoch": 0.6885024935975199, + "grad_norm": 0.38804158568382263, + "learning_rate": 4.419937488718185e-05, + "loss": 0.066, + "step": 29371 + }, + { + "epoch": 0.6885259351723249, + "grad_norm": 0.45764559507369995, + "learning_rate": 4.419326233013561e-05, + "loss": 0.1021, + "step": 29372 + }, + { + "epoch": 0.6885493767471299, + "grad_norm": 0.810255765914917, + "learning_rate": 4.4187150075898285e-05, + "loss": 0.0623, + "step": 29373 + }, + { + "epoch": 0.6885728183219348, + "grad_norm": 0.5943421125411987, + "learning_rate": 4.418103812450306e-05, + "loss": 0.5363, + "step": 29374 + }, + { + "epoch": 0.6885962598967399, + "grad_norm": 0.30794644355773926, + "learning_rate": 4.417492647598309e-05, + "loss": 0.0642, + "step": 29375 + }, + { + "epoch": 0.6886197014715448, + "grad_norm": 0.48728466033935547, + "learning_rate": 4.41688151303715e-05, + "loss": 0.1392, + "step": 29376 + }, + { + "epoch": 0.6886431430463499, + "grad_norm": 0.15206818282604218, + "learning_rate": 4.416270408770149e-05, + "loss": 0.0152, + "step": 29377 + }, + { + "epoch": 0.6886665846211548, + "grad_norm": 0.5466905832290649, + "learning_rate": 4.415659334800626e-05, + "loss": 0.1042, + "step": 29378 + }, + { + "epoch": 0.6886900261959599, + "grad_norm": 0.5694205164909363, + "learning_rate": 4.4150482911318914e-05, + "loss": 0.0615, + "step": 29379 + }, + { + "epoch": 0.6887134677707648, + "grad_norm": 0.12272796779870987, + "learning_rate": 4.4144372777672625e-05, + "loss": 0.0273, + "step": 29380 + }, + { + "epoch": 0.6887369093455699, + "grad_norm": 0.2839137017726898, + "learning_rate": 4.413826294710054e-05, + "loss": 0.0189, + "step": 29381 + }, + { + "epoch": 0.6887603509203748, + "grad_norm": 0.4409278929233551, + "learning_rate": 4.413215341963581e-05, + "loss": 0.111, + "step": 29382 + }, + { + "epoch": 0.6887837924951798, + "grad_norm": 0.5704292058944702, + "learning_rate": 4.4126044195311545e-05, + "loss": 0.632, + "step": 29383 + }, + { + "epoch": 0.6888072340699848, + "grad_norm": 0.37953561544418335, + "learning_rate": 4.411993527416097e-05, + "loss": 0.0814, + "step": 29384 + }, + { + "epoch": 0.6888306756447898, + "grad_norm": 0.5270426869392395, + "learning_rate": 4.41138266562172e-05, + "loss": 0.1066, + "step": 29385 + }, + { + "epoch": 0.6888541172195948, + "grad_norm": 0.5527058243751526, + "learning_rate": 4.410771834151339e-05, + "loss": 0.7241, + "step": 29386 + }, + { + "epoch": 0.6888775587943998, + "grad_norm": 0.6244117021560669, + "learning_rate": 4.410161033008267e-05, + "loss": 0.1165, + "step": 29387 + }, + { + "epoch": 0.6889010003692048, + "grad_norm": 0.3613879382610321, + "learning_rate": 4.409550262195817e-05, + "loss": 0.0419, + "step": 29388 + }, + { + "epoch": 0.6889244419440098, + "grad_norm": 1.0875507593154907, + "learning_rate": 4.408939521717301e-05, + "loss": 0.0495, + "step": 29389 + }, + { + "epoch": 0.6889478835188148, + "grad_norm": 0.40764209628105164, + "learning_rate": 4.408328811576038e-05, + "loss": 0.0802, + "step": 29390 + }, + { + "epoch": 0.6889713250936198, + "grad_norm": 0.4939621090888977, + "learning_rate": 4.4077181317753426e-05, + "loss": 0.5765, + "step": 29391 + }, + { + "epoch": 0.6889947666684247, + "grad_norm": 0.4482174217700958, + "learning_rate": 4.407107482318526e-05, + "loss": 0.0613, + "step": 29392 + }, + { + "epoch": 0.6890182082432298, + "grad_norm": 0.5578584671020508, + "learning_rate": 4.406496863208901e-05, + "loss": 0.0745, + "step": 29393 + }, + { + "epoch": 0.6890416498180347, + "grad_norm": 0.062209077179431915, + "learning_rate": 4.405886274449782e-05, + "loss": 0.011, + "step": 29394 + }, + { + "epoch": 0.6890650913928398, + "grad_norm": 0.5116752982139587, + "learning_rate": 4.40527571604448e-05, + "loss": 0.0714, + "step": 29395 + }, + { + "epoch": 0.6890885329676447, + "grad_norm": 0.27214452624320984, + "learning_rate": 4.404665187996307e-05, + "loss": 0.0393, + "step": 29396 + }, + { + "epoch": 0.6891119745424498, + "grad_norm": 0.5762783288955688, + "learning_rate": 4.404054690308581e-05, + "loss": 0.0921, + "step": 29397 + }, + { + "epoch": 0.6891354161172547, + "grad_norm": 0.4212060868740082, + "learning_rate": 4.403444222984612e-05, + "loss": 0.0559, + "step": 29398 + }, + { + "epoch": 0.6891588576920598, + "grad_norm": 0.24516893923282623, + "learning_rate": 4.402833786027712e-05, + "loss": 0.0462, + "step": 29399 + }, + { + "epoch": 0.6891822992668647, + "grad_norm": 0.4127066433429718, + "learning_rate": 4.402223379441194e-05, + "loss": 0.042, + "step": 29400 + }, + { + "epoch": 0.6892057408416697, + "grad_norm": 0.12185612320899963, + "learning_rate": 4.401613003228364e-05, + "loss": 0.0219, + "step": 29401 + }, + { + "epoch": 0.6892291824164747, + "grad_norm": 0.14754092693328857, + "learning_rate": 4.4010026573925434e-05, + "loss": 0.0334, + "step": 29402 + }, + { + "epoch": 0.6892526239912797, + "grad_norm": 0.7530162930488586, + "learning_rate": 4.400392341937037e-05, + "loss": 0.1309, + "step": 29403 + }, + { + "epoch": 0.6892760655660848, + "grad_norm": 0.5047410130500793, + "learning_rate": 4.399782056865163e-05, + "loss": 0.1256, + "step": 29404 + }, + { + "epoch": 0.6892995071408897, + "grad_norm": 0.4207640290260315, + "learning_rate": 4.399171802180227e-05, + "loss": 0.0703, + "step": 29405 + }, + { + "epoch": 0.6893229487156948, + "grad_norm": 0.5817299485206604, + "learning_rate": 4.398561577885545e-05, + "loss": 0.1451, + "step": 29406 + }, + { + "epoch": 0.6893463902904997, + "grad_norm": 0.4160415530204773, + "learning_rate": 4.397951383984423e-05, + "loss": 0.0634, + "step": 29407 + }, + { + "epoch": 0.6893698318653048, + "grad_norm": 0.26025575399398804, + "learning_rate": 4.3973412204801726e-05, + "loss": 0.0487, + "step": 29408 + }, + { + "epoch": 0.6893932734401097, + "grad_norm": 0.44163358211517334, + "learning_rate": 4.396731087376108e-05, + "loss": 0.0735, + "step": 29409 + }, + { + "epoch": 0.6894167150149147, + "grad_norm": 0.07259613275527954, + "learning_rate": 4.396120984675539e-05, + "loss": 0.0073, + "step": 29410 + }, + { + "epoch": 0.6894401565897197, + "grad_norm": 0.43626388907432556, + "learning_rate": 4.395510912381776e-05, + "loss": 0.0936, + "step": 29411 + }, + { + "epoch": 0.6894635981645247, + "grad_norm": 0.5113814473152161, + "learning_rate": 4.394900870498128e-05, + "loss": 0.0969, + "step": 29412 + }, + { + "epoch": 0.6894870397393297, + "grad_norm": 0.2643420100212097, + "learning_rate": 4.3942908590279044e-05, + "loss": 0.0633, + "step": 29413 + }, + { + "epoch": 0.6895104813141347, + "grad_norm": 0.8291943669319153, + "learning_rate": 4.393680877974413e-05, + "loss": 0.4906, + "step": 29414 + }, + { + "epoch": 0.6895339228889397, + "grad_norm": 0.09909083694219589, + "learning_rate": 4.393070927340967e-05, + "loss": 0.0095, + "step": 29415 + }, + { + "epoch": 0.6895573644637447, + "grad_norm": 0.2801491618156433, + "learning_rate": 4.3924610071308794e-05, + "loss": 0.0512, + "step": 29416 + }, + { + "epoch": 0.6895808060385497, + "grad_norm": 0.5645958185195923, + "learning_rate": 4.391851117347456e-05, + "loss": 0.1212, + "step": 29417 + }, + { + "epoch": 0.6896042476133547, + "grad_norm": 0.48498520255088806, + "learning_rate": 4.391241257994006e-05, + "loss": 0.0913, + "step": 29418 + }, + { + "epoch": 0.6896276891881596, + "grad_norm": 0.1907031387090683, + "learning_rate": 4.3906314290738396e-05, + "loss": 0.0456, + "step": 29419 + }, + { + "epoch": 0.6896511307629647, + "grad_norm": 0.5870035290718079, + "learning_rate": 4.390021630590264e-05, + "loss": 0.5202, + "step": 29420 + }, + { + "epoch": 0.6896745723377696, + "grad_norm": 0.4218825697898865, + "learning_rate": 4.389411862546585e-05, + "loss": 0.2764, + "step": 29421 + }, + { + "epoch": 0.6896980139125747, + "grad_norm": 0.3844227194786072, + "learning_rate": 4.3888021249461186e-05, + "loss": 0.0738, + "step": 29422 + }, + { + "epoch": 0.6897214554873796, + "grad_norm": 0.38710543513298035, + "learning_rate": 4.388192417792169e-05, + "loss": 0.085, + "step": 29423 + }, + { + "epoch": 0.6897448970621847, + "grad_norm": 0.07937503606081009, + "learning_rate": 4.387582741088047e-05, + "loss": 0.0146, + "step": 29424 + }, + { + "epoch": 0.6897683386369896, + "grad_norm": 0.45490172505378723, + "learning_rate": 4.386973094837058e-05, + "loss": 0.1052, + "step": 29425 + }, + { + "epoch": 0.6897917802117947, + "grad_norm": 0.16921202838420868, + "learning_rate": 4.386363479042507e-05, + "loss": 0.0402, + "step": 29426 + }, + { + "epoch": 0.6898152217865996, + "grad_norm": 0.4123769998550415, + "learning_rate": 4.385753893707709e-05, + "loss": 0.0761, + "step": 29427 + }, + { + "epoch": 0.6898386633614046, + "grad_norm": 0.08198793232440948, + "learning_rate": 4.385144338835965e-05, + "loss": 0.0093, + "step": 29428 + }, + { + "epoch": 0.6898621049362096, + "grad_norm": 0.5173786878585815, + "learning_rate": 4.384534814430589e-05, + "loss": 0.6077, + "step": 29429 + }, + { + "epoch": 0.6898855465110146, + "grad_norm": 0.23183290660381317, + "learning_rate": 4.383925320494885e-05, + "loss": 0.0205, + "step": 29430 + }, + { + "epoch": 0.6899089880858196, + "grad_norm": 0.46614810824394226, + "learning_rate": 4.3833158570321606e-05, + "loss": 0.1106, + "step": 29431 + }, + { + "epoch": 0.6899324296606246, + "grad_norm": 0.4918380379676819, + "learning_rate": 4.382706424045722e-05, + "loss": 0.5291, + "step": 29432 + }, + { + "epoch": 0.6899558712354296, + "grad_norm": 0.45954185724258423, + "learning_rate": 4.3820970215388766e-05, + "loss": 0.0618, + "step": 29433 + }, + { + "epoch": 0.6899793128102346, + "grad_norm": 0.4992761015892029, + "learning_rate": 4.381487649514926e-05, + "loss": 0.0735, + "step": 29434 + }, + { + "epoch": 0.6900027543850396, + "grad_norm": 0.13697290420532227, + "learning_rate": 4.380878307977186e-05, + "loss": 0.0185, + "step": 29435 + }, + { + "epoch": 0.6900261959598446, + "grad_norm": 0.14509746432304382, + "learning_rate": 4.380268996928958e-05, + "loss": 0.0315, + "step": 29436 + }, + { + "epoch": 0.6900496375346495, + "grad_norm": 0.5030688047409058, + "learning_rate": 4.3796597163735486e-05, + "loss": 0.0537, + "step": 29437 + }, + { + "epoch": 0.6900730791094546, + "grad_norm": 0.20267513394355774, + "learning_rate": 4.3790504663142594e-05, + "loss": 0.0425, + "step": 29438 + }, + { + "epoch": 0.6900965206842595, + "grad_norm": 0.39608901739120483, + "learning_rate": 4.3784412467544054e-05, + "loss": 0.0547, + "step": 29439 + }, + { + "epoch": 0.6901199622590646, + "grad_norm": 0.13212621212005615, + "learning_rate": 4.3778320576972865e-05, + "loss": 0.0243, + "step": 29440 + }, + { + "epoch": 0.6901434038338695, + "grad_norm": 0.17821460962295532, + "learning_rate": 4.377222899146205e-05, + "loss": 0.0414, + "step": 29441 + }, + { + "epoch": 0.6901668454086746, + "grad_norm": 0.5516034364700317, + "learning_rate": 4.3766137711044744e-05, + "loss": 0.1292, + "step": 29442 + }, + { + "epoch": 0.6901902869834795, + "grad_norm": 0.1283344328403473, + "learning_rate": 4.376004673575395e-05, + "loss": 0.0177, + "step": 29443 + }, + { + "epoch": 0.6902137285582846, + "grad_norm": 0.29334181547164917, + "learning_rate": 4.375395606562273e-05, + "loss": 0.0614, + "step": 29444 + }, + { + "epoch": 0.6902371701330895, + "grad_norm": 0.23135411739349365, + "learning_rate": 4.374786570068411e-05, + "loss": 0.0479, + "step": 29445 + }, + { + "epoch": 0.6902606117078945, + "grad_norm": 0.45750004053115845, + "learning_rate": 4.374177564097113e-05, + "loss": 0.0527, + "step": 29446 + }, + { + "epoch": 0.6902840532826995, + "grad_norm": 0.356755793094635, + "learning_rate": 4.373568588651689e-05, + "loss": 0.0995, + "step": 29447 + }, + { + "epoch": 0.6903074948575045, + "grad_norm": 0.3740568161010742, + "learning_rate": 4.37295964373544e-05, + "loss": 0.0635, + "step": 29448 + }, + { + "epoch": 0.6903309364323095, + "grad_norm": 0.3282945156097412, + "learning_rate": 4.3723507293516704e-05, + "loss": 0.0466, + "step": 29449 + }, + { + "epoch": 0.6903543780071145, + "grad_norm": 0.4936343729496002, + "learning_rate": 4.3717418455036804e-05, + "loss": 0.0504, + "step": 29450 + }, + { + "epoch": 0.6903778195819195, + "grad_norm": 0.557339608669281, + "learning_rate": 4.371132992194781e-05, + "loss": 0.6662, + "step": 29451 + }, + { + "epoch": 0.6904012611567245, + "grad_norm": 0.2222144603729248, + "learning_rate": 4.3705241694282714e-05, + "loss": 0.0439, + "step": 29452 + }, + { + "epoch": 0.6904247027315294, + "grad_norm": 0.16993963718414307, + "learning_rate": 4.3699153772074544e-05, + "loss": 0.0294, + "step": 29453 + }, + { + "epoch": 0.6904481443063345, + "grad_norm": 0.16285733878612518, + "learning_rate": 4.369306615535636e-05, + "loss": 0.0295, + "step": 29454 + }, + { + "epoch": 0.6904715858811395, + "grad_norm": 0.7439338564872742, + "learning_rate": 4.36869788441612e-05, + "loss": 0.172, + "step": 29455 + }, + { + "epoch": 0.6904950274559445, + "grad_norm": 0.4850082993507385, + "learning_rate": 4.368089183852207e-05, + "loss": 0.0805, + "step": 29456 + }, + { + "epoch": 0.6905184690307495, + "grad_norm": 0.7631945013999939, + "learning_rate": 4.367480513847201e-05, + "loss": 0.1322, + "step": 29457 + }, + { + "epoch": 0.6905419106055545, + "grad_norm": 0.8411234617233276, + "learning_rate": 4.366871874404402e-05, + "loss": 0.1138, + "step": 29458 + }, + { + "epoch": 0.6905653521803595, + "grad_norm": 0.1792808175086975, + "learning_rate": 4.366263265527113e-05, + "loss": 0.0467, + "step": 29459 + }, + { + "epoch": 0.6905887937551645, + "grad_norm": 0.5906083583831787, + "learning_rate": 4.365654687218641e-05, + "loss": 0.1259, + "step": 29460 + }, + { + "epoch": 0.6906122353299695, + "grad_norm": 0.6074609160423279, + "learning_rate": 4.365046139482286e-05, + "loss": 0.0801, + "step": 29461 + }, + { + "epoch": 0.6906356769047745, + "grad_norm": 0.6192612051963806, + "learning_rate": 4.364437622321345e-05, + "loss": 0.4816, + "step": 29462 + }, + { + "epoch": 0.6906591184795795, + "grad_norm": 0.4070624113082886, + "learning_rate": 4.363829135739126e-05, + "loss": 0.0758, + "step": 29463 + }, + { + "epoch": 0.6906825600543844, + "grad_norm": 0.5663468837738037, + "learning_rate": 4.36322067973893e-05, + "loss": 0.1223, + "step": 29464 + }, + { + "epoch": 0.6907060016291895, + "grad_norm": 0.3942295014858246, + "learning_rate": 4.362612254324057e-05, + "loss": 0.0727, + "step": 29465 + }, + { + "epoch": 0.6907294432039944, + "grad_norm": 0.22647762298583984, + "learning_rate": 4.362003859497805e-05, + "loss": 0.0679, + "step": 29466 + }, + { + "epoch": 0.6907528847787995, + "grad_norm": 0.2920174300670624, + "learning_rate": 4.3613954952634815e-05, + "loss": 0.0313, + "step": 29467 + }, + { + "epoch": 0.6907763263536044, + "grad_norm": 0.59133380651474, + "learning_rate": 4.3607871616243846e-05, + "loss": 0.0707, + "step": 29468 + }, + { + "epoch": 0.6907997679284095, + "grad_norm": 0.19913053512573242, + "learning_rate": 4.3601788585838155e-05, + "loss": 0.0636, + "step": 29469 + }, + { + "epoch": 0.6908232095032144, + "grad_norm": 0.12629009783267975, + "learning_rate": 4.3595705861450756e-05, + "loss": 0.0193, + "step": 29470 + }, + { + "epoch": 0.6908466510780195, + "grad_norm": 0.2693828344345093, + "learning_rate": 4.3589623443114624e-05, + "loss": 0.0321, + "step": 29471 + }, + { + "epoch": 0.6908700926528244, + "grad_norm": 0.050266701728105545, + "learning_rate": 4.358354133086275e-05, + "loss": 0.007, + "step": 29472 + }, + { + "epoch": 0.6908935342276294, + "grad_norm": 0.32999783754348755, + "learning_rate": 4.3577459524728205e-05, + "loss": 0.0684, + "step": 29473 + }, + { + "epoch": 0.6909169758024344, + "grad_norm": 0.44447898864746094, + "learning_rate": 4.357137802474396e-05, + "loss": 0.0795, + "step": 29474 + }, + { + "epoch": 0.6909404173772394, + "grad_norm": 0.04922901466488838, + "learning_rate": 4.356529683094296e-05, + "loss": 0.0065, + "step": 29475 + }, + { + "epoch": 0.6909638589520444, + "grad_norm": 0.4378131031990051, + "learning_rate": 4.3559215943358275e-05, + "loss": 0.0668, + "step": 29476 + }, + { + "epoch": 0.6909873005268494, + "grad_norm": 0.17876890301704407, + "learning_rate": 4.355313536202288e-05, + "loss": 0.035, + "step": 29477 + }, + { + "epoch": 0.6910107421016544, + "grad_norm": 0.44262129068374634, + "learning_rate": 4.354705508696976e-05, + "loss": 0.0721, + "step": 29478 + }, + { + "epoch": 0.6910341836764594, + "grad_norm": 0.23996751010417938, + "learning_rate": 4.354097511823186e-05, + "loss": 0.0413, + "step": 29479 + }, + { + "epoch": 0.6910576252512644, + "grad_norm": 1.0305393934249878, + "learning_rate": 4.3534895455842275e-05, + "loss": 0.1496, + "step": 29480 + }, + { + "epoch": 0.6910810668260694, + "grad_norm": 0.11450081318616867, + "learning_rate": 4.352881609983391e-05, + "loss": 0.0244, + "step": 29481 + }, + { + "epoch": 0.6911045084008743, + "grad_norm": 0.19413137435913086, + "learning_rate": 4.352273705023979e-05, + "loss": 0.0177, + "step": 29482 + }, + { + "epoch": 0.6911279499756794, + "grad_norm": 0.3574766218662262, + "learning_rate": 4.351665830709288e-05, + "loss": 0.0531, + "step": 29483 + }, + { + "epoch": 0.6911513915504843, + "grad_norm": 0.3929481506347656, + "learning_rate": 4.3510579870426146e-05, + "loss": 0.0782, + "step": 29484 + }, + { + "epoch": 0.6911748331252894, + "grad_norm": 0.5509026646614075, + "learning_rate": 4.3504501740272616e-05, + "loss": 0.0586, + "step": 29485 + }, + { + "epoch": 0.6911982747000943, + "grad_norm": 0.34061023592948914, + "learning_rate": 4.349842391666526e-05, + "loss": 0.0873, + "step": 29486 + }, + { + "epoch": 0.6912217162748994, + "grad_norm": 0.5580277442932129, + "learning_rate": 4.3492346399637005e-05, + "loss": 0.0448, + "step": 29487 + }, + { + "epoch": 0.6912451578497043, + "grad_norm": 0.2319241613149643, + "learning_rate": 4.3486269189220905e-05, + "loss": 0.0408, + "step": 29488 + }, + { + "epoch": 0.6912685994245094, + "grad_norm": 0.20586714148521423, + "learning_rate": 4.3480192285449896e-05, + "loss": 0.0394, + "step": 29489 + }, + { + "epoch": 0.6912920409993143, + "grad_norm": 0.5823667049407959, + "learning_rate": 4.347411568835694e-05, + "loss": 0.5382, + "step": 29490 + }, + { + "epoch": 0.6913154825741193, + "grad_norm": 0.5936117768287659, + "learning_rate": 4.3468039397975e-05, + "loss": 0.7627, + "step": 29491 + }, + { + "epoch": 0.6913389241489243, + "grad_norm": 0.8388369083404541, + "learning_rate": 4.346196341433711e-05, + "loss": 0.1425, + "step": 29492 + }, + { + "epoch": 0.6913623657237293, + "grad_norm": 0.13101421296596527, + "learning_rate": 4.345588773747618e-05, + "loss": 0.0284, + "step": 29493 + }, + { + "epoch": 0.6913858072985343, + "grad_norm": 0.34151479601860046, + "learning_rate": 4.3449812367425205e-05, + "loss": 0.0335, + "step": 29494 + }, + { + "epoch": 0.6914092488733393, + "grad_norm": 0.20186424255371094, + "learning_rate": 4.344373730421713e-05, + "loss": 0.0358, + "step": 29495 + }, + { + "epoch": 0.6914326904481443, + "grad_norm": 0.3452325761318207, + "learning_rate": 4.343766254788493e-05, + "loss": 0.0801, + "step": 29496 + }, + { + "epoch": 0.6914561320229493, + "grad_norm": 0.3640103340148926, + "learning_rate": 4.343158809846153e-05, + "loss": 0.0884, + "step": 29497 + }, + { + "epoch": 0.6914795735977542, + "grad_norm": 0.3743104934692383, + "learning_rate": 4.3425513955979946e-05, + "loss": 0.077, + "step": 29498 + }, + { + "epoch": 0.6915030151725593, + "grad_norm": 0.3241625130176544, + "learning_rate": 4.3419440120473085e-05, + "loss": 0.0829, + "step": 29499 + }, + { + "epoch": 0.6915264567473642, + "grad_norm": 0.2725018262863159, + "learning_rate": 4.341336659197398e-05, + "loss": 0.03, + "step": 29500 + }, + { + "epoch": 0.6915498983221693, + "grad_norm": 0.13593535125255585, + "learning_rate": 4.3407293370515536e-05, + "loss": 0.027, + "step": 29501 + }, + { + "epoch": 0.6915733398969742, + "grad_norm": 0.09563044458627701, + "learning_rate": 4.34012204561307e-05, + "loss": 0.0096, + "step": 29502 + }, + { + "epoch": 0.6915967814717793, + "grad_norm": 0.7476012110710144, + "learning_rate": 4.339514784885244e-05, + "loss": 0.1059, + "step": 29503 + }, + { + "epoch": 0.6916202230465842, + "grad_norm": 0.213187113404274, + "learning_rate": 4.3389075548713655e-05, + "loss": 0.0368, + "step": 29504 + }, + { + "epoch": 0.6916436646213893, + "grad_norm": 0.6453302502632141, + "learning_rate": 4.338300355574738e-05, + "loss": 0.0807, + "step": 29505 + }, + { + "epoch": 0.6916671061961943, + "grad_norm": 0.34531667828559875, + "learning_rate": 4.337693186998652e-05, + "loss": 0.0498, + "step": 29506 + }, + { + "epoch": 0.6916905477709993, + "grad_norm": 0.33820268511772156, + "learning_rate": 4.337086049146402e-05, + "loss": 0.0689, + "step": 29507 + }, + { + "epoch": 0.6917139893458043, + "grad_norm": 0.30315810441970825, + "learning_rate": 4.336478942021283e-05, + "loss": 0.0386, + "step": 29508 + }, + { + "epoch": 0.6917374309206092, + "grad_norm": 0.2624950706958771, + "learning_rate": 4.335871865626587e-05, + "loss": 0.0404, + "step": 29509 + }, + { + "epoch": 0.6917608724954143, + "grad_norm": 0.48034754395484924, + "learning_rate": 4.335264819965607e-05, + "loss": 0.0604, + "step": 29510 + }, + { + "epoch": 0.6917843140702192, + "grad_norm": 0.3594238758087158, + "learning_rate": 4.3346578050416386e-05, + "loss": 0.0855, + "step": 29511 + }, + { + "epoch": 0.6918077556450243, + "grad_norm": 0.5084994435310364, + "learning_rate": 4.33405082085798e-05, + "loss": 0.1082, + "step": 29512 + }, + { + "epoch": 0.6918311972198292, + "grad_norm": 0.3128865361213684, + "learning_rate": 4.3334438674179225e-05, + "loss": 0.029, + "step": 29513 + }, + { + "epoch": 0.6918546387946343, + "grad_norm": 0.12515875697135925, + "learning_rate": 4.3328369447247564e-05, + "loss": 0.0203, + "step": 29514 + }, + { + "epoch": 0.6918780803694392, + "grad_norm": 0.39579322934150696, + "learning_rate": 4.332230052781776e-05, + "loss": 0.0775, + "step": 29515 + }, + { + "epoch": 0.6919015219442443, + "grad_norm": 0.24897366762161255, + "learning_rate": 4.3316231915922755e-05, + "loss": 0.0258, + "step": 29516 + }, + { + "epoch": 0.6919249635190492, + "grad_norm": 0.21725408732891083, + "learning_rate": 4.331016361159543e-05, + "loss": 0.0326, + "step": 29517 + }, + { + "epoch": 0.6919484050938542, + "grad_norm": 0.25680604577064514, + "learning_rate": 4.330409561486879e-05, + "loss": 0.0238, + "step": 29518 + }, + { + "epoch": 0.6919718466686592, + "grad_norm": 0.1823868602514267, + "learning_rate": 4.329802792577572e-05, + "loss": 0.0205, + "step": 29519 + }, + { + "epoch": 0.6919952882434642, + "grad_norm": 0.3352489173412323, + "learning_rate": 4.329196054434914e-05, + "loss": 0.0698, + "step": 29520 + }, + { + "epoch": 0.6920187298182692, + "grad_norm": 0.5681436657905579, + "learning_rate": 4.328589347062199e-05, + "loss": 0.1049, + "step": 29521 + }, + { + "epoch": 0.6920421713930742, + "grad_norm": 0.22551506757736206, + "learning_rate": 4.3279826704627135e-05, + "loss": 0.0414, + "step": 29522 + }, + { + "epoch": 0.6920656129678792, + "grad_norm": 0.08065759390592575, + "learning_rate": 4.327376024639753e-05, + "loss": 0.0155, + "step": 29523 + }, + { + "epoch": 0.6920890545426842, + "grad_norm": 0.24749444425106049, + "learning_rate": 4.326769409596614e-05, + "loss": 0.0417, + "step": 29524 + }, + { + "epoch": 0.6921124961174892, + "grad_norm": 0.7097871899604797, + "learning_rate": 4.3261628253365826e-05, + "loss": 0.1146, + "step": 29525 + }, + { + "epoch": 0.6921359376922942, + "grad_norm": 0.4082363545894623, + "learning_rate": 4.325556271862952e-05, + "loss": 0.0756, + "step": 29526 + }, + { + "epoch": 0.6921593792670991, + "grad_norm": 0.559176504611969, + "learning_rate": 4.3249497491790126e-05, + "loss": 0.0987, + "step": 29527 + }, + { + "epoch": 0.6921828208419042, + "grad_norm": 0.2205466330051422, + "learning_rate": 4.324343257288055e-05, + "loss": 0.025, + "step": 29528 + }, + { + "epoch": 0.6922062624167091, + "grad_norm": 0.23743340373039246, + "learning_rate": 4.3237367961933675e-05, + "loss": 0.0514, + "step": 29529 + }, + { + "epoch": 0.6922297039915142, + "grad_norm": 0.6556021571159363, + "learning_rate": 4.323130365898247e-05, + "loss": 0.1302, + "step": 29530 + }, + { + "epoch": 0.6922531455663191, + "grad_norm": 0.3222949504852295, + "learning_rate": 4.3225239664059805e-05, + "loss": 0.0637, + "step": 29531 + }, + { + "epoch": 0.6922765871411242, + "grad_norm": 0.3518887460231781, + "learning_rate": 4.321917597719859e-05, + "loss": 0.0239, + "step": 29532 + }, + { + "epoch": 0.6923000287159291, + "grad_norm": 0.3354063034057617, + "learning_rate": 4.3213112598431726e-05, + "loss": 0.0671, + "step": 29533 + }, + { + "epoch": 0.6923234702907342, + "grad_norm": 0.2991091012954712, + "learning_rate": 4.32070495277921e-05, + "loss": 0.0582, + "step": 29534 + }, + { + "epoch": 0.6923469118655391, + "grad_norm": 0.12384452670812607, + "learning_rate": 4.320098676531259e-05, + "loss": 0.0224, + "step": 29535 + }, + { + "epoch": 0.6923703534403441, + "grad_norm": 0.11922107636928558, + "learning_rate": 4.319492431102612e-05, + "loss": 0.0173, + "step": 29536 + }, + { + "epoch": 0.6923937950151491, + "grad_norm": 0.4509769082069397, + "learning_rate": 4.318886216496563e-05, + "loss": 0.0976, + "step": 29537 + }, + { + "epoch": 0.6924172365899541, + "grad_norm": 0.15595200657844543, + "learning_rate": 4.3182800327163965e-05, + "loss": 0.0204, + "step": 29538 + }, + { + "epoch": 0.6924406781647591, + "grad_norm": 0.2687362730503082, + "learning_rate": 4.317673879765402e-05, + "loss": 0.0302, + "step": 29539 + }, + { + "epoch": 0.6924641197395641, + "grad_norm": 0.5197204947471619, + "learning_rate": 4.3170677576468696e-05, + "loss": 0.0803, + "step": 29540 + }, + { + "epoch": 0.6924875613143691, + "grad_norm": 0.31872737407684326, + "learning_rate": 4.3164616663640875e-05, + "loss": 0.0467, + "step": 29541 + }, + { + "epoch": 0.6925110028891741, + "grad_norm": 0.476905882358551, + "learning_rate": 4.3158556059203405e-05, + "loss": 0.064, + "step": 29542 + }, + { + "epoch": 0.692534444463979, + "grad_norm": 0.558320164680481, + "learning_rate": 4.315249576318924e-05, + "loss": 0.1209, + "step": 29543 + }, + { + "epoch": 0.6925578860387841, + "grad_norm": 0.13661141693592072, + "learning_rate": 4.3146435775631234e-05, + "loss": 0.0271, + "step": 29544 + }, + { + "epoch": 0.692581327613589, + "grad_norm": 0.7747587561607361, + "learning_rate": 4.314037609656226e-05, + "loss": 0.5926, + "step": 29545 + }, + { + "epoch": 0.6926047691883941, + "grad_norm": 0.26423805952072144, + "learning_rate": 4.313431672601521e-05, + "loss": 0.0437, + "step": 29546 + }, + { + "epoch": 0.692628210763199, + "grad_norm": 0.11085048317909241, + "learning_rate": 4.312825766402292e-05, + "loss": 0.0175, + "step": 29547 + }, + { + "epoch": 0.6926516523380041, + "grad_norm": 0.43472743034362793, + "learning_rate": 4.312219891061834e-05, + "loss": 0.0753, + "step": 29548 + }, + { + "epoch": 0.692675093912809, + "grad_norm": 0.531295120716095, + "learning_rate": 4.311614046583427e-05, + "loss": 0.062, + "step": 29549 + }, + { + "epoch": 0.6926985354876141, + "grad_norm": 0.2870560586452484, + "learning_rate": 4.311008232970366e-05, + "loss": 0.0492, + "step": 29550 + }, + { + "epoch": 0.692721977062419, + "grad_norm": 0.11721479147672653, + "learning_rate": 4.310402450225933e-05, + "loss": 0.0278, + "step": 29551 + }, + { + "epoch": 0.692745418637224, + "grad_norm": 0.5166517496109009, + "learning_rate": 4.309796698353418e-05, + "loss": 0.099, + "step": 29552 + }, + { + "epoch": 0.692768860212029, + "grad_norm": 0.5523580312728882, + "learning_rate": 4.3091909773561056e-05, + "loss": 0.0899, + "step": 29553 + }, + { + "epoch": 0.692792301786834, + "grad_norm": 0.03457072377204895, + "learning_rate": 4.308585287237279e-05, + "loss": 0.0048, + "step": 29554 + }, + { + "epoch": 0.692815743361639, + "grad_norm": 0.5541488528251648, + "learning_rate": 4.307979628000233e-05, + "loss": 0.0748, + "step": 29555 + }, + { + "epoch": 0.692839184936444, + "grad_norm": 0.5098506212234497, + "learning_rate": 4.3073739996482486e-05, + "loss": 0.4235, + "step": 29556 + }, + { + "epoch": 0.6928626265112491, + "grad_norm": 0.07566513121128082, + "learning_rate": 4.3067684021846136e-05, + "loss": 0.0078, + "step": 29557 + }, + { + "epoch": 0.692886068086054, + "grad_norm": 0.47637516260147095, + "learning_rate": 4.3061628356126124e-05, + "loss": 0.0618, + "step": 29558 + }, + { + "epoch": 0.6929095096608591, + "grad_norm": 0.31243371963500977, + "learning_rate": 4.3055572999355334e-05, + "loss": 0.0257, + "step": 29559 + }, + { + "epoch": 0.692932951235664, + "grad_norm": 0.2892855703830719, + "learning_rate": 4.304951795156655e-05, + "loss": 0.0659, + "step": 29560 + }, + { + "epoch": 0.6929563928104691, + "grad_norm": 0.1186307817697525, + "learning_rate": 4.30434632127927e-05, + "loss": 0.0177, + "step": 29561 + }, + { + "epoch": 0.692979834385274, + "grad_norm": 0.41001757979393005, + "learning_rate": 4.303740878306664e-05, + "loss": 0.0443, + "step": 29562 + }, + { + "epoch": 0.693003275960079, + "grad_norm": 0.16428756713867188, + "learning_rate": 4.303135466242122e-05, + "loss": 0.0228, + "step": 29563 + }, + { + "epoch": 0.693026717534884, + "grad_norm": 0.37133392691612244, + "learning_rate": 4.302530085088926e-05, + "loss": 0.0751, + "step": 29564 + }, + { + "epoch": 0.693050159109689, + "grad_norm": 0.30699023604393005, + "learning_rate": 4.301924734850362e-05, + "loss": 0.0499, + "step": 29565 + }, + { + "epoch": 0.693073600684494, + "grad_norm": 0.13535068929195404, + "learning_rate": 4.3013194155297155e-05, + "loss": 0.0263, + "step": 29566 + }, + { + "epoch": 0.693097042259299, + "grad_norm": 0.4322977066040039, + "learning_rate": 4.3007141271302656e-05, + "loss": 0.0821, + "step": 29567 + }, + { + "epoch": 0.693120483834104, + "grad_norm": 0.12367673218250275, + "learning_rate": 4.300108869655305e-05, + "loss": 0.0245, + "step": 29568 + }, + { + "epoch": 0.693143925408909, + "grad_norm": 0.6639864444732666, + "learning_rate": 4.299503643108114e-05, + "loss": 0.5397, + "step": 29569 + }, + { + "epoch": 0.693167366983714, + "grad_norm": 0.5101279020309448, + "learning_rate": 4.298898447491976e-05, + "loss": 0.0763, + "step": 29570 + }, + { + "epoch": 0.693190808558519, + "grad_norm": 0.18352718651294708, + "learning_rate": 4.298293282810175e-05, + "loss": 0.022, + "step": 29571 + }, + { + "epoch": 0.6932142501333239, + "grad_norm": 0.5767455697059631, + "learning_rate": 4.297688149065993e-05, + "loss": 0.116, + "step": 29572 + }, + { + "epoch": 0.693237691708129, + "grad_norm": 0.6999170184135437, + "learning_rate": 4.297083046262718e-05, + "loss": 0.0621, + "step": 29573 + }, + { + "epoch": 0.6932611332829339, + "grad_norm": 0.23367463052272797, + "learning_rate": 4.296477974403626e-05, + "loss": 0.056, + "step": 29574 + }, + { + "epoch": 0.693284574857739, + "grad_norm": 0.3671087324619293, + "learning_rate": 4.29587293349201e-05, + "loss": 0.0755, + "step": 29575 + }, + { + "epoch": 0.6933080164325439, + "grad_norm": 0.6347596049308777, + "learning_rate": 4.2952679235311474e-05, + "loss": 0.4689, + "step": 29576 + }, + { + "epoch": 0.693331458007349, + "grad_norm": 0.18668046593666077, + "learning_rate": 4.294662944524323e-05, + "loss": 0.0233, + "step": 29577 + }, + { + "epoch": 0.6933548995821539, + "grad_norm": 0.21851606667041779, + "learning_rate": 4.294057996474815e-05, + "loss": 0.0251, + "step": 29578 + }, + { + "epoch": 0.693378341156959, + "grad_norm": 0.20488771796226501, + "learning_rate": 4.293453079385911e-05, + "loss": 0.0407, + "step": 29579 + }, + { + "epoch": 0.6934017827317639, + "grad_norm": 0.8390741348266602, + "learning_rate": 4.2928481932608864e-05, + "loss": 0.1331, + "step": 29580 + }, + { + "epoch": 0.6934252243065689, + "grad_norm": 0.1877753734588623, + "learning_rate": 4.2922433381030315e-05, + "loss": 0.0383, + "step": 29581 + }, + { + "epoch": 0.6934486658813739, + "grad_norm": 0.42557790875434875, + "learning_rate": 4.291638513915626e-05, + "loss": 0.0516, + "step": 29582 + }, + { + "epoch": 0.6934721074561789, + "grad_norm": 0.5994849801063538, + "learning_rate": 4.291033720701949e-05, + "loss": 0.4243, + "step": 29583 + }, + { + "epoch": 0.6934955490309839, + "grad_norm": 0.6420972943305969, + "learning_rate": 4.29042895846528e-05, + "loss": 0.141, + "step": 29584 + }, + { + "epoch": 0.6935189906057889, + "grad_norm": 0.5814490914344788, + "learning_rate": 4.2898242272089084e-05, + "loss": 0.561, + "step": 29585 + }, + { + "epoch": 0.6935424321805939, + "grad_norm": 0.43006420135498047, + "learning_rate": 4.2892195269361114e-05, + "loss": 0.0702, + "step": 29586 + }, + { + "epoch": 0.6935658737553989, + "grad_norm": 0.47439485788345337, + "learning_rate": 4.288614857650165e-05, + "loss": 0.0596, + "step": 29587 + }, + { + "epoch": 0.6935893153302038, + "grad_norm": 0.4978570342063904, + "learning_rate": 4.2880102193543595e-05, + "loss": 0.0772, + "step": 29588 + }, + { + "epoch": 0.6936127569050089, + "grad_norm": 0.7198064923286438, + "learning_rate": 4.28740561205197e-05, + "loss": 0.1425, + "step": 29589 + }, + { + "epoch": 0.6936361984798138, + "grad_norm": 0.3597222864627838, + "learning_rate": 4.2868010357462795e-05, + "loss": 0.0516, + "step": 29590 + }, + { + "epoch": 0.6936596400546189, + "grad_norm": 0.4128827750682831, + "learning_rate": 4.286196490440567e-05, + "loss": 0.0642, + "step": 29591 + }, + { + "epoch": 0.6936830816294238, + "grad_norm": 0.3199004828929901, + "learning_rate": 4.2855919761381094e-05, + "loss": 0.262, + "step": 29592 + }, + { + "epoch": 0.6937065232042289, + "grad_norm": 0.38151609897613525, + "learning_rate": 4.284987492842194e-05, + "loss": 0.0618, + "step": 29593 + }, + { + "epoch": 0.6937299647790338, + "grad_norm": 0.13065136969089508, + "learning_rate": 4.284383040556097e-05, + "loss": 0.0248, + "step": 29594 + }, + { + "epoch": 0.6937534063538389, + "grad_norm": 0.3796585202217102, + "learning_rate": 4.2837786192830974e-05, + "loss": 0.0717, + "step": 29595 + }, + { + "epoch": 0.6937768479286438, + "grad_norm": 0.4220837354660034, + "learning_rate": 4.2831742290264733e-05, + "loss": 0.0737, + "step": 29596 + }, + { + "epoch": 0.6938002895034489, + "grad_norm": 0.5113556981086731, + "learning_rate": 4.28256986978951e-05, + "loss": 0.1279, + "step": 29597 + }, + { + "epoch": 0.6938237310782538, + "grad_norm": 0.6046357154846191, + "learning_rate": 4.2819655415754836e-05, + "loss": 0.0712, + "step": 29598 + }, + { + "epoch": 0.6938471726530588, + "grad_norm": 0.2931435704231262, + "learning_rate": 4.2813612443876685e-05, + "loss": 0.046, + "step": 29599 + }, + { + "epoch": 0.6938706142278638, + "grad_norm": 0.6992090344429016, + "learning_rate": 4.280756978229352e-05, + "loss": 0.3696, + "step": 29600 + }, + { + "epoch": 0.6938940558026688, + "grad_norm": 0.6238741874694824, + "learning_rate": 4.280152743103809e-05, + "loss": 0.0673, + "step": 29601 + }, + { + "epoch": 0.6939174973774738, + "grad_norm": 0.4084000587463379, + "learning_rate": 4.2795485390143185e-05, + "loss": 0.062, + "step": 29602 + }, + { + "epoch": 0.6939409389522788, + "grad_norm": 0.5416755676269531, + "learning_rate": 4.278944365964158e-05, + "loss": 0.0612, + "step": 29603 + }, + { + "epoch": 0.6939643805270838, + "grad_norm": 0.19665341079235077, + "learning_rate": 4.278340223956607e-05, + "loss": 0.0226, + "step": 29604 + }, + { + "epoch": 0.6939878221018888, + "grad_norm": 0.47008588910102844, + "learning_rate": 4.277736112994939e-05, + "loss": 0.0604, + "step": 29605 + }, + { + "epoch": 0.6940112636766937, + "grad_norm": 0.33550789952278137, + "learning_rate": 4.277132033082438e-05, + "loss": 0.0676, + "step": 29606 + }, + { + "epoch": 0.6940347052514988, + "grad_norm": 0.3632495701313019, + "learning_rate": 4.2765279842223804e-05, + "loss": 0.0684, + "step": 29607 + }, + { + "epoch": 0.6940581468263038, + "grad_norm": 0.5335947871208191, + "learning_rate": 4.2759239664180396e-05, + "loss": 0.1708, + "step": 29608 + }, + { + "epoch": 0.6940815884011088, + "grad_norm": 0.7573287487030029, + "learning_rate": 4.2753199796726996e-05, + "loss": 0.148, + "step": 29609 + }, + { + "epoch": 0.6941050299759138, + "grad_norm": 0.8288047909736633, + "learning_rate": 4.2747160239896346e-05, + "loss": 0.1619, + "step": 29610 + }, + { + "epoch": 0.6941284715507188, + "grad_norm": 0.17361651360988617, + "learning_rate": 4.2741120993721216e-05, + "loss": 0.0298, + "step": 29611 + }, + { + "epoch": 0.6941519131255238, + "grad_norm": 1.0357943773269653, + "learning_rate": 4.273508205823433e-05, + "loss": 0.1554, + "step": 29612 + }, + { + "epoch": 0.6941753547003288, + "grad_norm": 0.38446861505508423, + "learning_rate": 4.272904343346854e-05, + "loss": 0.0939, + "step": 29613 + }, + { + "epoch": 0.6941987962751338, + "grad_norm": 0.5227115154266357, + "learning_rate": 4.272300511945657e-05, + "loss": 0.0829, + "step": 29614 + }, + { + "epoch": 0.6942222378499388, + "grad_norm": 0.40604168176651, + "learning_rate": 4.271696711623119e-05, + "loss": 0.0814, + "step": 29615 + }, + { + "epoch": 0.6942456794247438, + "grad_norm": 0.4155290424823761, + "learning_rate": 4.271092942382514e-05, + "loss": 0.1408, + "step": 29616 + }, + { + "epoch": 0.6942691209995487, + "grad_norm": 0.46107980608940125, + "learning_rate": 4.2704892042271216e-05, + "loss": 0.0898, + "step": 29617 + }, + { + "epoch": 0.6942925625743538, + "grad_norm": 0.658237099647522, + "learning_rate": 4.269885497160211e-05, + "loss": 0.1767, + "step": 29618 + }, + { + "epoch": 0.6943160041491587, + "grad_norm": 0.5096779465675354, + "learning_rate": 4.269281821185067e-05, + "loss": 0.1141, + "step": 29619 + }, + { + "epoch": 0.6943394457239638, + "grad_norm": 0.10505365580320358, + "learning_rate": 4.268678176304958e-05, + "loss": 0.0134, + "step": 29620 + }, + { + "epoch": 0.6943628872987687, + "grad_norm": 0.3920670747756958, + "learning_rate": 4.2680745625231653e-05, + "loss": 0.082, + "step": 29621 + }, + { + "epoch": 0.6943863288735738, + "grad_norm": 0.16572318971157074, + "learning_rate": 4.267470979842961e-05, + "loss": 0.0268, + "step": 29622 + }, + { + "epoch": 0.6944097704483787, + "grad_norm": 0.31222647428512573, + "learning_rate": 4.26686742826762e-05, + "loss": 0.0722, + "step": 29623 + }, + { + "epoch": 0.6944332120231838, + "grad_norm": 0.22883769869804382, + "learning_rate": 4.2662639078004186e-05, + "loss": 0.0263, + "step": 29624 + }, + { + "epoch": 0.6944566535979887, + "grad_norm": 0.7641086578369141, + "learning_rate": 4.265660418444627e-05, + "loss": 0.1481, + "step": 29625 + }, + { + "epoch": 0.6944800951727937, + "grad_norm": 0.15807944536209106, + "learning_rate": 4.2650569602035264e-05, + "loss": 0.0223, + "step": 29626 + }, + { + "epoch": 0.6945035367475987, + "grad_norm": 0.5019787549972534, + "learning_rate": 4.264453533080387e-05, + "loss": 0.0876, + "step": 29627 + }, + { + "epoch": 0.6945269783224037, + "grad_norm": 0.1460198163986206, + "learning_rate": 4.263850137078484e-05, + "loss": 0.0278, + "step": 29628 + }, + { + "epoch": 0.6945504198972087, + "grad_norm": 0.6165609955787659, + "learning_rate": 4.263246772201093e-05, + "loss": 0.1279, + "step": 29629 + }, + { + "epoch": 0.6945738614720137, + "grad_norm": 0.4307672679424286, + "learning_rate": 4.262643438451482e-05, + "loss": 0.1112, + "step": 29630 + }, + { + "epoch": 0.6945973030468187, + "grad_norm": 0.10638082027435303, + "learning_rate": 4.262040135832932e-05, + "loss": 0.0244, + "step": 29631 + }, + { + "epoch": 0.6946207446216237, + "grad_norm": 0.240327388048172, + "learning_rate": 4.2614368643487144e-05, + "loss": 0.0351, + "step": 29632 + }, + { + "epoch": 0.6946441861964286, + "grad_norm": 0.6267570853233337, + "learning_rate": 4.260833624002098e-05, + "loss": 0.1955, + "step": 29633 + }, + { + "epoch": 0.6946676277712337, + "grad_norm": 0.16144903004169464, + "learning_rate": 4.260230414796363e-05, + "loss": 0.0218, + "step": 29634 + }, + { + "epoch": 0.6946910693460386, + "grad_norm": 0.995100736618042, + "learning_rate": 4.259627236734779e-05, + "loss": 0.1446, + "step": 29635 + }, + { + "epoch": 0.6947145109208437, + "grad_norm": 0.19785456359386444, + "learning_rate": 4.25902408982062e-05, + "loss": 0.0424, + "step": 29636 + }, + { + "epoch": 0.6947379524956486, + "grad_norm": 0.1877002865076065, + "learning_rate": 4.2584209740571534e-05, + "loss": 0.0171, + "step": 29637 + }, + { + "epoch": 0.6947613940704537, + "grad_norm": 0.4666259288787842, + "learning_rate": 4.2578178894476596e-05, + "loss": 0.0507, + "step": 29638 + }, + { + "epoch": 0.6947848356452586, + "grad_norm": 0.5464783906936646, + "learning_rate": 4.257214835995407e-05, + "loss": 0.5857, + "step": 29639 + }, + { + "epoch": 0.6948082772200637, + "grad_norm": 0.2225683480501175, + "learning_rate": 4.256611813703669e-05, + "loss": 0.0244, + "step": 29640 + }, + { + "epoch": 0.6948317187948686, + "grad_norm": 0.33068329095840454, + "learning_rate": 4.2560088225757164e-05, + "loss": 0.0342, + "step": 29641 + }, + { + "epoch": 0.6948551603696737, + "grad_norm": 0.43914172053337097, + "learning_rate": 4.255405862614822e-05, + "loss": 0.0673, + "step": 29642 + }, + { + "epoch": 0.6948786019444786, + "grad_norm": 0.45127540826797485, + "learning_rate": 4.254802933824252e-05, + "loss": 0.1018, + "step": 29643 + }, + { + "epoch": 0.6949020435192836, + "grad_norm": 0.11250902712345123, + "learning_rate": 4.254200036207287e-05, + "loss": 0.0154, + "step": 29644 + }, + { + "epoch": 0.6949254850940886, + "grad_norm": 0.3629227876663208, + "learning_rate": 4.253597169767191e-05, + "loss": 0.0673, + "step": 29645 + }, + { + "epoch": 0.6949489266688936, + "grad_norm": 0.4341474771499634, + "learning_rate": 4.252994334507242e-05, + "loss": 0.054, + "step": 29646 + }, + { + "epoch": 0.6949723682436986, + "grad_norm": 0.36599451303482056, + "learning_rate": 4.252391530430707e-05, + "loss": 0.0669, + "step": 29647 + }, + { + "epoch": 0.6949958098185036, + "grad_norm": 0.12291303277015686, + "learning_rate": 4.251788757540857e-05, + "loss": 0.0257, + "step": 29648 + }, + { + "epoch": 0.6950192513933086, + "grad_norm": 0.3235187232494354, + "learning_rate": 4.2511860158409634e-05, + "loss": 0.054, + "step": 29649 + }, + { + "epoch": 0.6950426929681136, + "grad_norm": 0.39366641640663147, + "learning_rate": 4.2505833053342915e-05, + "loss": 0.0611, + "step": 29650 + }, + { + "epoch": 0.6950661345429185, + "grad_norm": 0.42573851346969604, + "learning_rate": 4.24998062602412e-05, + "loss": 0.0848, + "step": 29651 + }, + { + "epoch": 0.6950895761177236, + "grad_norm": 0.19979223608970642, + "learning_rate": 4.2493779779137164e-05, + "loss": 0.0317, + "step": 29652 + }, + { + "epoch": 0.6951130176925285, + "grad_norm": 0.1685018390417099, + "learning_rate": 4.248775361006349e-05, + "loss": 0.0276, + "step": 29653 + }, + { + "epoch": 0.6951364592673336, + "grad_norm": 0.1400948315858841, + "learning_rate": 4.248172775305288e-05, + "loss": 0.0149, + "step": 29654 + }, + { + "epoch": 0.6951599008421385, + "grad_norm": 0.30494558811187744, + "learning_rate": 4.2475702208138034e-05, + "loss": 0.0545, + "step": 29655 + }, + { + "epoch": 0.6951833424169436, + "grad_norm": 0.17164120078086853, + "learning_rate": 4.2469676975351614e-05, + "loss": 0.0375, + "step": 29656 + }, + { + "epoch": 0.6952067839917485, + "grad_norm": 0.49770936369895935, + "learning_rate": 4.246365205472633e-05, + "loss": 0.5648, + "step": 29657 + }, + { + "epoch": 0.6952302255665536, + "grad_norm": 0.3223225772380829, + "learning_rate": 4.2457627446294937e-05, + "loss": 0.0428, + "step": 29658 + }, + { + "epoch": 0.6952536671413586, + "grad_norm": 0.4168921113014221, + "learning_rate": 4.245160315009007e-05, + "loss": 0.0502, + "step": 29659 + }, + { + "epoch": 0.6952771087161635, + "grad_norm": 0.6848971843719482, + "learning_rate": 4.244557916614444e-05, + "loss": 0.1368, + "step": 29660 + }, + { + "epoch": 0.6953005502909686, + "grad_norm": 0.360662579536438, + "learning_rate": 4.243955549449069e-05, + "loss": 0.1042, + "step": 29661 + }, + { + "epoch": 0.6953239918657735, + "grad_norm": 0.23445698618888855, + "learning_rate": 4.2433532135161536e-05, + "loss": 0.0605, + "step": 29662 + }, + { + "epoch": 0.6953474334405786, + "grad_norm": 0.33699384331703186, + "learning_rate": 4.242750908818962e-05, + "loss": 0.0617, + "step": 29663 + }, + { + "epoch": 0.6953708750153835, + "grad_norm": 0.488007515668869, + "learning_rate": 4.242148635360771e-05, + "loss": 0.119, + "step": 29664 + }, + { + "epoch": 0.6953943165901886, + "grad_norm": 0.6955742239952087, + "learning_rate": 4.241546393144842e-05, + "loss": 0.1296, + "step": 29665 + }, + { + "epoch": 0.6954177581649935, + "grad_norm": 0.44676676392555237, + "learning_rate": 4.2409441821744446e-05, + "loss": 0.0835, + "step": 29666 + }, + { + "epoch": 0.6954411997397986, + "grad_norm": 0.29095327854156494, + "learning_rate": 4.240342002452845e-05, + "loss": 0.0642, + "step": 29667 + }, + { + "epoch": 0.6954646413146035, + "grad_norm": 0.4773673117160797, + "learning_rate": 4.239739853983309e-05, + "loss": 0.0826, + "step": 29668 + }, + { + "epoch": 0.6954880828894086, + "grad_norm": 0.3014891445636749, + "learning_rate": 4.23913773676911e-05, + "loss": 0.0362, + "step": 29669 + }, + { + "epoch": 0.6955115244642135, + "grad_norm": 0.3742421865463257, + "learning_rate": 4.2385356508135074e-05, + "loss": 0.0872, + "step": 29670 + }, + { + "epoch": 0.6955349660390185, + "grad_norm": 0.578829288482666, + "learning_rate": 4.2379335961197774e-05, + "loss": 0.0967, + "step": 29671 + }, + { + "epoch": 0.6955584076138235, + "grad_norm": 0.43317919969558716, + "learning_rate": 4.23733157269118e-05, + "loss": 0.1114, + "step": 29672 + }, + { + "epoch": 0.6955818491886285, + "grad_norm": 0.5883894562721252, + "learning_rate": 4.236729580530985e-05, + "loss": 0.0609, + "step": 29673 + }, + { + "epoch": 0.6956052907634335, + "grad_norm": 0.5228391289710999, + "learning_rate": 4.2361276196424574e-05, + "loss": 0.0705, + "step": 29674 + }, + { + "epoch": 0.6956287323382385, + "grad_norm": 0.5003002285957336, + "learning_rate": 4.235525690028859e-05, + "loss": 0.0944, + "step": 29675 + }, + { + "epoch": 0.6956521739130435, + "grad_norm": 0.22929027676582336, + "learning_rate": 4.234923791693464e-05, + "loss": 0.0374, + "step": 29676 + }, + { + "epoch": 0.6956756154878485, + "grad_norm": 0.5065937638282776, + "learning_rate": 4.2343219246395335e-05, + "loss": 0.1006, + "step": 29677 + }, + { + "epoch": 0.6956990570626534, + "grad_norm": 0.33023446798324585, + "learning_rate": 4.2337200888703355e-05, + "loss": 0.3139, + "step": 29678 + }, + { + "epoch": 0.6957224986374585, + "grad_norm": 0.39825671911239624, + "learning_rate": 4.2331182843891346e-05, + "loss": 0.0581, + "step": 29679 + }, + { + "epoch": 0.6957459402122634, + "grad_norm": 0.5128811597824097, + "learning_rate": 4.232516511199196e-05, + "loss": 0.0671, + "step": 29680 + }, + { + "epoch": 0.6957693817870685, + "grad_norm": 0.31522536277770996, + "learning_rate": 4.2319147693037805e-05, + "loss": 0.0383, + "step": 29681 + }, + { + "epoch": 0.6957928233618734, + "grad_norm": 0.4483901560306549, + "learning_rate": 4.231313058706159e-05, + "loss": 0.0662, + "step": 29682 + }, + { + "epoch": 0.6958162649366785, + "grad_norm": 0.4046872854232788, + "learning_rate": 4.230711379409598e-05, + "loss": 0.0945, + "step": 29683 + }, + { + "epoch": 0.6958397065114834, + "grad_norm": 0.13877926766872406, + "learning_rate": 4.2301097314173586e-05, + "loss": 0.0363, + "step": 29684 + }, + { + "epoch": 0.6958631480862885, + "grad_norm": 0.1500226855278015, + "learning_rate": 4.229508114732708e-05, + "loss": 0.0299, + "step": 29685 + }, + { + "epoch": 0.6958865896610934, + "grad_norm": 0.3193095028400421, + "learning_rate": 4.228906529358907e-05, + "loss": 0.0544, + "step": 29686 + }, + { + "epoch": 0.6959100312358985, + "grad_norm": 0.6651887893676758, + "learning_rate": 4.228304975299221e-05, + "loss": 0.1032, + "step": 29687 + }, + { + "epoch": 0.6959334728107034, + "grad_norm": 0.5870097279548645, + "learning_rate": 4.2277034525569126e-05, + "loss": 0.0815, + "step": 29688 + }, + { + "epoch": 0.6959569143855084, + "grad_norm": 0.333508163690567, + "learning_rate": 4.22710196113525e-05, + "loss": 0.0597, + "step": 29689 + }, + { + "epoch": 0.6959803559603134, + "grad_norm": 0.22323912382125854, + "learning_rate": 4.226500501037495e-05, + "loss": 0.0381, + "step": 29690 + }, + { + "epoch": 0.6960037975351184, + "grad_norm": 0.45425423979759216, + "learning_rate": 4.22589907226691e-05, + "loss": 0.0824, + "step": 29691 + }, + { + "epoch": 0.6960272391099234, + "grad_norm": 0.5791987180709839, + "learning_rate": 4.225297674826759e-05, + "loss": 0.0586, + "step": 29692 + }, + { + "epoch": 0.6960506806847284, + "grad_norm": 0.48169219493865967, + "learning_rate": 4.224696308720302e-05, + "loss": 0.0876, + "step": 29693 + }, + { + "epoch": 0.6960741222595334, + "grad_norm": 1.0789062976837158, + "learning_rate": 4.224094973950808e-05, + "loss": 0.2287, + "step": 29694 + }, + { + "epoch": 0.6960975638343384, + "grad_norm": 0.4349098205566406, + "learning_rate": 4.223493670521534e-05, + "loss": 0.1013, + "step": 29695 + }, + { + "epoch": 0.6961210054091433, + "grad_norm": 0.0967424213886261, + "learning_rate": 4.222892398435751e-05, + "loss": 0.0202, + "step": 29696 + }, + { + "epoch": 0.6961444469839484, + "grad_norm": 0.2580551505088806, + "learning_rate": 4.222291157696714e-05, + "loss": 0.0453, + "step": 29697 + }, + { + "epoch": 0.6961678885587533, + "grad_norm": 0.45834675431251526, + "learning_rate": 4.2216899483076886e-05, + "loss": 0.0761, + "step": 29698 + }, + { + "epoch": 0.6961913301335584, + "grad_norm": 0.18666115403175354, + "learning_rate": 4.221088770271936e-05, + "loss": 0.0365, + "step": 29699 + }, + { + "epoch": 0.6962147717083633, + "grad_norm": 0.7062011361122131, + "learning_rate": 4.2204876235927184e-05, + "loss": 0.1397, + "step": 29700 + }, + { + "epoch": 0.6962382132831684, + "grad_norm": 0.24549388885498047, + "learning_rate": 4.219886508273293e-05, + "loss": 0.0591, + "step": 29701 + }, + { + "epoch": 0.6962616548579733, + "grad_norm": 0.5821103453636169, + "learning_rate": 4.21928542431693e-05, + "loss": 0.109, + "step": 29702 + }, + { + "epoch": 0.6962850964327784, + "grad_norm": 0.1677217036485672, + "learning_rate": 4.2186843717268874e-05, + "loss": 0.0269, + "step": 29703 + }, + { + "epoch": 0.6963085380075833, + "grad_norm": 0.4568209648132324, + "learning_rate": 4.218083350506426e-05, + "loss": 0.0614, + "step": 29704 + }, + { + "epoch": 0.6963319795823883, + "grad_norm": 0.4260123372077942, + "learning_rate": 4.2174823606588054e-05, + "loss": 0.0626, + "step": 29705 + }, + { + "epoch": 0.6963554211571933, + "grad_norm": 0.42126110196113586, + "learning_rate": 4.216881402187286e-05, + "loss": 0.1065, + "step": 29706 + }, + { + "epoch": 0.6963788627319983, + "grad_norm": 0.3251963257789612, + "learning_rate": 4.216280475095134e-05, + "loss": 0.0443, + "step": 29707 + }, + { + "epoch": 0.6964023043068033, + "grad_norm": 0.3816803991794586, + "learning_rate": 4.215679579385603e-05, + "loss": 0.0648, + "step": 29708 + }, + { + "epoch": 0.6964257458816083, + "grad_norm": 0.3168644607067108, + "learning_rate": 4.215078715061961e-05, + "loss": 0.0323, + "step": 29709 + }, + { + "epoch": 0.6964491874564134, + "grad_norm": 0.6995836496353149, + "learning_rate": 4.214477882127463e-05, + "loss": 0.1606, + "step": 29710 + }, + { + "epoch": 0.6964726290312183, + "grad_norm": 0.532528281211853, + "learning_rate": 4.213877080585373e-05, + "loss": 0.0838, + "step": 29711 + }, + { + "epoch": 0.6964960706060234, + "grad_norm": 0.5233678817749023, + "learning_rate": 4.2132763104389475e-05, + "loss": 0.0957, + "step": 29712 + }, + { + "epoch": 0.6965195121808283, + "grad_norm": 0.07871775329113007, + "learning_rate": 4.2126755716914444e-05, + "loss": 0.0213, + "step": 29713 + }, + { + "epoch": 0.6965429537556334, + "grad_norm": 0.3771030306816101, + "learning_rate": 4.2120748643461293e-05, + "loss": 0.0598, + "step": 29714 + }, + { + "epoch": 0.6965663953304383, + "grad_norm": 0.7212544083595276, + "learning_rate": 4.211474188406259e-05, + "loss": 0.1106, + "step": 29715 + }, + { + "epoch": 0.6965898369052433, + "grad_norm": 0.7096558809280396, + "learning_rate": 4.2108735438750925e-05, + "loss": 0.4407, + "step": 29716 + }, + { + "epoch": 0.6966132784800483, + "grad_norm": 0.33317041397094727, + "learning_rate": 4.2102729307558895e-05, + "loss": 0.0544, + "step": 29717 + }, + { + "epoch": 0.6966367200548533, + "grad_norm": 0.39386168122291565, + "learning_rate": 4.209672349051904e-05, + "loss": 0.0769, + "step": 29718 + }, + { + "epoch": 0.6966601616296583, + "grad_norm": 0.4414738416671753, + "learning_rate": 4.209071798766403e-05, + "loss": 0.0768, + "step": 29719 + }, + { + "epoch": 0.6966836032044633, + "grad_norm": 0.30144941806793213, + "learning_rate": 4.208471279902637e-05, + "loss": 0.0263, + "step": 29720 + }, + { + "epoch": 0.6967070447792683, + "grad_norm": 0.46374252438545227, + "learning_rate": 4.2078707924638736e-05, + "loss": 0.0843, + "step": 29721 + }, + { + "epoch": 0.6967304863540733, + "grad_norm": 0.12271233648061752, + "learning_rate": 4.207270336453365e-05, + "loss": 0.0255, + "step": 29722 + }, + { + "epoch": 0.6967539279288782, + "grad_norm": 0.7636536359786987, + "learning_rate": 4.206669911874371e-05, + "loss": 0.1322, + "step": 29723 + }, + { + "epoch": 0.6967773695036833, + "grad_norm": 0.6229961514472961, + "learning_rate": 4.206069518730149e-05, + "loss": 0.0585, + "step": 29724 + }, + { + "epoch": 0.6968008110784882, + "grad_norm": 0.4379698634147644, + "learning_rate": 4.2054691570239566e-05, + "loss": 0.0588, + "step": 29725 + }, + { + "epoch": 0.6968242526532933, + "grad_norm": 0.2841872274875641, + "learning_rate": 4.204868826759048e-05, + "loss": 0.0267, + "step": 29726 + }, + { + "epoch": 0.6968476942280982, + "grad_norm": 0.5727654695510864, + "learning_rate": 4.2042685279386874e-05, + "loss": 0.0839, + "step": 29727 + }, + { + "epoch": 0.6968711358029033, + "grad_norm": 0.26328879594802856, + "learning_rate": 4.2036682605661294e-05, + "loss": 0.027, + "step": 29728 + }, + { + "epoch": 0.6968945773777082, + "grad_norm": 0.3638010025024414, + "learning_rate": 4.20306802464463e-05, + "loss": 0.0526, + "step": 29729 + }, + { + "epoch": 0.6969180189525133, + "grad_norm": 0.6054860949516296, + "learning_rate": 4.2024678201774426e-05, + "loss": 0.0848, + "step": 29730 + }, + { + "epoch": 0.6969414605273182, + "grad_norm": 0.154421865940094, + "learning_rate": 4.201867647167831e-05, + "loss": 0.0308, + "step": 29731 + }, + { + "epoch": 0.6969649021021233, + "grad_norm": 0.07868499308824539, + "learning_rate": 4.201267505619049e-05, + "loss": 0.0165, + "step": 29732 + }, + { + "epoch": 0.6969883436769282, + "grad_norm": 0.09496976435184479, + "learning_rate": 4.200667395534349e-05, + "loss": 0.0133, + "step": 29733 + }, + { + "epoch": 0.6970117852517332, + "grad_norm": 0.19059108197689056, + "learning_rate": 4.200067316916995e-05, + "loss": 0.0318, + "step": 29734 + }, + { + "epoch": 0.6970352268265382, + "grad_norm": 0.14630325138568878, + "learning_rate": 4.1994672697702383e-05, + "loss": 0.0294, + "step": 29735 + }, + { + "epoch": 0.6970586684013432, + "grad_norm": 0.5564596056938171, + "learning_rate": 4.198867254097335e-05, + "loss": 0.541, + "step": 29736 + }, + { + "epoch": 0.6970821099761482, + "grad_norm": 0.10822103917598724, + "learning_rate": 4.1982672699015414e-05, + "loss": 0.0101, + "step": 29737 + }, + { + "epoch": 0.6971055515509532, + "grad_norm": 0.41916653513908386, + "learning_rate": 4.197667317186112e-05, + "loss": 0.0973, + "step": 29738 + }, + { + "epoch": 0.6971289931257582, + "grad_norm": 0.5009477734565735, + "learning_rate": 4.1970673959543006e-05, + "loss": 0.0964, + "step": 29739 + }, + { + "epoch": 0.6971524347005632, + "grad_norm": 0.5091927647590637, + "learning_rate": 4.196467506209368e-05, + "loss": 0.5386, + "step": 29740 + }, + { + "epoch": 0.6971758762753681, + "grad_norm": 0.2447061985731125, + "learning_rate": 4.195867647954565e-05, + "loss": 0.0509, + "step": 29741 + }, + { + "epoch": 0.6971993178501732, + "grad_norm": 0.527931809425354, + "learning_rate": 4.1952678211931454e-05, + "loss": 0.1375, + "step": 29742 + }, + { + "epoch": 0.6972227594249781, + "grad_norm": 0.3353457748889923, + "learning_rate": 4.194668025928368e-05, + "loss": 0.055, + "step": 29743 + }, + { + "epoch": 0.6972462009997832, + "grad_norm": 0.2359834909439087, + "learning_rate": 4.194068262163485e-05, + "loss": 0.0648, + "step": 29744 + }, + { + "epoch": 0.6972696425745881, + "grad_norm": 0.3235529661178589, + "learning_rate": 4.193468529901752e-05, + "loss": 0.0718, + "step": 29745 + }, + { + "epoch": 0.6972930841493932, + "grad_norm": 0.07866575568914413, + "learning_rate": 4.192868829146417e-05, + "loss": 0.0134, + "step": 29746 + }, + { + "epoch": 0.6973165257241981, + "grad_norm": 0.4001418948173523, + "learning_rate": 4.1922691599007436e-05, + "loss": 0.0617, + "step": 29747 + }, + { + "epoch": 0.6973399672990032, + "grad_norm": 0.5336717367172241, + "learning_rate": 4.19166952216798e-05, + "loss": 0.1329, + "step": 29748 + }, + { + "epoch": 0.6973634088738081, + "grad_norm": 0.5027238726615906, + "learning_rate": 4.191069915951382e-05, + "loss": 0.088, + "step": 29749 + }, + { + "epoch": 0.6973868504486131, + "grad_norm": 0.6419147253036499, + "learning_rate": 4.1904703412542024e-05, + "loss": 0.1081, + "step": 29750 + }, + { + "epoch": 0.6974102920234181, + "grad_norm": 0.34730252623558044, + "learning_rate": 4.189870798079689e-05, + "loss": 0.0883, + "step": 29751 + }, + { + "epoch": 0.6974337335982231, + "grad_norm": 0.035761792212724686, + "learning_rate": 4.189271286431104e-05, + "loss": 0.0046, + "step": 29752 + }, + { + "epoch": 0.6974571751730281, + "grad_norm": 0.31176698207855225, + "learning_rate": 4.188671806311697e-05, + "loss": 0.0635, + "step": 29753 + }, + { + "epoch": 0.6974806167478331, + "grad_norm": 0.6338899731636047, + "learning_rate": 4.188072357724716e-05, + "loss": 0.0951, + "step": 29754 + }, + { + "epoch": 0.6975040583226381, + "grad_norm": 0.5244521498680115, + "learning_rate": 4.1874729406734223e-05, + "loss": 0.4539, + "step": 29755 + }, + { + "epoch": 0.6975274998974431, + "grad_norm": 0.4786930978298187, + "learning_rate": 4.186873555161063e-05, + "loss": 0.1029, + "step": 29756 + }, + { + "epoch": 0.697550941472248, + "grad_norm": 0.2677917182445526, + "learning_rate": 4.186274201190892e-05, + "loss": 0.0461, + "step": 29757 + }, + { + "epoch": 0.6975743830470531, + "grad_norm": 0.3698938488960266, + "learning_rate": 4.185674878766156e-05, + "loss": 0.0489, + "step": 29758 + }, + { + "epoch": 0.697597824621858, + "grad_norm": 0.40615883469581604, + "learning_rate": 4.185075587890116e-05, + "loss": 0.0849, + "step": 29759 + }, + { + "epoch": 0.6976212661966631, + "grad_norm": 0.44950562715530396, + "learning_rate": 4.1844763285660196e-05, + "loss": 0.0819, + "step": 29760 + }, + { + "epoch": 0.6976447077714681, + "grad_norm": 0.5191647410392761, + "learning_rate": 4.183877100797117e-05, + "loss": 0.0744, + "step": 29761 + }, + { + "epoch": 0.6976681493462731, + "grad_norm": 0.24024637043476105, + "learning_rate": 4.183277904586661e-05, + "loss": 0.0425, + "step": 29762 + }, + { + "epoch": 0.6976915909210781, + "grad_norm": 0.6595410704612732, + "learning_rate": 4.182678739937903e-05, + "loss": 0.1442, + "step": 29763 + }, + { + "epoch": 0.6977150324958831, + "grad_norm": 0.7225103378295898, + "learning_rate": 4.1820796068540903e-05, + "loss": 0.6471, + "step": 29764 + }, + { + "epoch": 0.6977384740706881, + "grad_norm": 0.11959873884916306, + "learning_rate": 4.1814805053384806e-05, + "loss": 0.0258, + "step": 29765 + }, + { + "epoch": 0.6977619156454931, + "grad_norm": 0.19043242931365967, + "learning_rate": 4.180881435394318e-05, + "loss": 0.0278, + "step": 29766 + }, + { + "epoch": 0.6977853572202981, + "grad_norm": 0.3772942125797272, + "learning_rate": 4.18028239702486e-05, + "loss": 0.0288, + "step": 29767 + }, + { + "epoch": 0.697808798795103, + "grad_norm": 0.3976403772830963, + "learning_rate": 4.179683390233353e-05, + "loss": 0.0568, + "step": 29768 + }, + { + "epoch": 0.6978322403699081, + "grad_norm": 0.31101611256599426, + "learning_rate": 4.179084415023048e-05, + "loss": 0.0348, + "step": 29769 + }, + { + "epoch": 0.697855681944713, + "grad_norm": 0.45979180932044983, + "learning_rate": 4.1784854713971946e-05, + "loss": 0.0983, + "step": 29770 + }, + { + "epoch": 0.6978791235195181, + "grad_norm": 0.4651201665401459, + "learning_rate": 4.177886559359039e-05, + "loss": 0.1179, + "step": 29771 + }, + { + "epoch": 0.697902565094323, + "grad_norm": 0.23865017294883728, + "learning_rate": 4.177287678911837e-05, + "loss": 0.0139, + "step": 29772 + }, + { + "epoch": 0.6979260066691281, + "grad_norm": 0.15713191032409668, + "learning_rate": 4.1766888300588366e-05, + "loss": 0.0179, + "step": 29773 + }, + { + "epoch": 0.697949448243933, + "grad_norm": 0.4360118508338928, + "learning_rate": 4.1760900128032874e-05, + "loss": 0.0593, + "step": 29774 + }, + { + "epoch": 0.6979728898187381, + "grad_norm": 0.44468846917152405, + "learning_rate": 4.175491227148437e-05, + "loss": 0.0791, + "step": 29775 + }, + { + "epoch": 0.697996331393543, + "grad_norm": 0.45483896136283875, + "learning_rate": 4.174892473097535e-05, + "loss": 0.0704, + "step": 29776 + }, + { + "epoch": 0.698019772968348, + "grad_norm": 0.2313118875026703, + "learning_rate": 4.174293750653826e-05, + "loss": 0.0306, + "step": 29777 + }, + { + "epoch": 0.698043214543153, + "grad_norm": 0.4662638306617737, + "learning_rate": 4.173695059820563e-05, + "loss": 0.049, + "step": 29778 + }, + { + "epoch": 0.698066656117958, + "grad_norm": 0.4166519343852997, + "learning_rate": 4.173096400600999e-05, + "loss": 0.0694, + "step": 29779 + }, + { + "epoch": 0.698090097692763, + "grad_norm": 0.348588764667511, + "learning_rate": 4.172497772998376e-05, + "loss": 0.0499, + "step": 29780 + }, + { + "epoch": 0.698113539267568, + "grad_norm": 0.4038833975791931, + "learning_rate": 4.1718991770159456e-05, + "loss": 0.4233, + "step": 29781 + }, + { + "epoch": 0.698136980842373, + "grad_norm": 0.42522841691970825, + "learning_rate": 4.171300612656953e-05, + "loss": 0.0831, + "step": 29782 + }, + { + "epoch": 0.698160422417178, + "grad_norm": 0.09361671656370163, + "learning_rate": 4.170702079924648e-05, + "loss": 0.0208, + "step": 29783 + }, + { + "epoch": 0.698183863991983, + "grad_norm": 0.3212188482284546, + "learning_rate": 4.170103578822273e-05, + "loss": 0.2867, + "step": 29784 + }, + { + "epoch": 0.698207305566788, + "grad_norm": 0.15599416196346283, + "learning_rate": 4.169505109353085e-05, + "loss": 0.0228, + "step": 29785 + }, + { + "epoch": 0.698230747141593, + "grad_norm": 0.3299502730369568, + "learning_rate": 4.1689066715203254e-05, + "loss": 0.0448, + "step": 29786 + }, + { + "epoch": 0.698254188716398, + "grad_norm": 0.47431108355522156, + "learning_rate": 4.168308265327241e-05, + "loss": 0.1083, + "step": 29787 + }, + { + "epoch": 0.6982776302912029, + "grad_norm": 0.2540467381477356, + "learning_rate": 4.1677098907770816e-05, + "loss": 0.0362, + "step": 29788 + }, + { + "epoch": 0.698301071866008, + "grad_norm": 0.6630289554595947, + "learning_rate": 4.167111547873088e-05, + "loss": 0.1311, + "step": 29789 + }, + { + "epoch": 0.6983245134408129, + "grad_norm": 1.2821776866912842, + "learning_rate": 4.1665132366185146e-05, + "loss": 0.1952, + "step": 29790 + }, + { + "epoch": 0.698347955015618, + "grad_norm": 0.21721914410591125, + "learning_rate": 4.1659149570166015e-05, + "loss": 0.0282, + "step": 29791 + }, + { + "epoch": 0.6983713965904229, + "grad_norm": 0.3601081073284149, + "learning_rate": 4.1653167090706e-05, + "loss": 0.0789, + "step": 29792 + }, + { + "epoch": 0.698394838165228, + "grad_norm": 0.13946734368801117, + "learning_rate": 4.164718492783756e-05, + "loss": 0.0369, + "step": 29793 + }, + { + "epoch": 0.6984182797400329, + "grad_norm": 0.5269402861595154, + "learning_rate": 4.164120308159312e-05, + "loss": 0.0905, + "step": 29794 + }, + { + "epoch": 0.698441721314838, + "grad_norm": 0.4279910922050476, + "learning_rate": 4.163522155200515e-05, + "loss": 0.5647, + "step": 29795 + }, + { + "epoch": 0.6984651628896429, + "grad_norm": 0.44345328211784363, + "learning_rate": 4.162924033910609e-05, + "loss": 0.0554, + "step": 29796 + }, + { + "epoch": 0.6984886044644479, + "grad_norm": 0.5511431097984314, + "learning_rate": 4.162325944292844e-05, + "loss": 0.1091, + "step": 29797 + }, + { + "epoch": 0.6985120460392529, + "grad_norm": 0.16553908586502075, + "learning_rate": 4.1617278863504626e-05, + "loss": 0.0187, + "step": 29798 + }, + { + "epoch": 0.6985354876140579, + "grad_norm": 0.11821522563695908, + "learning_rate": 4.16112986008671e-05, + "loss": 0.0238, + "step": 29799 + }, + { + "epoch": 0.6985589291888629, + "grad_norm": 0.2063172310590744, + "learning_rate": 4.160531865504831e-05, + "loss": 0.0523, + "step": 29800 + }, + { + "epoch": 0.6985823707636679, + "grad_norm": 0.3123539388179779, + "learning_rate": 4.159933902608071e-05, + "loss": 0.0572, + "step": 29801 + }, + { + "epoch": 0.6986058123384729, + "grad_norm": 0.31961074471473694, + "learning_rate": 4.159335971399668e-05, + "loss": 0.0505, + "step": 29802 + }, + { + "epoch": 0.6986292539132779, + "grad_norm": 0.34012821316719055, + "learning_rate": 4.158738071882874e-05, + "loss": 0.0683, + "step": 29803 + }, + { + "epoch": 0.6986526954880828, + "grad_norm": 0.5986794233322144, + "learning_rate": 4.1581402040609355e-05, + "loss": 0.1143, + "step": 29804 + }, + { + "epoch": 0.6986761370628879, + "grad_norm": 0.3798918128013611, + "learning_rate": 4.157542367937092e-05, + "loss": 0.2903, + "step": 29805 + }, + { + "epoch": 0.6986995786376928, + "grad_norm": 0.6976273655891418, + "learning_rate": 4.1569445635145865e-05, + "loss": 0.1473, + "step": 29806 + }, + { + "epoch": 0.6987230202124979, + "grad_norm": 0.39086318016052246, + "learning_rate": 4.156346790796666e-05, + "loss": 0.0793, + "step": 29807 + }, + { + "epoch": 0.6987464617873028, + "grad_norm": 0.44348806142807007, + "learning_rate": 4.1557490497865715e-05, + "loss": 0.0756, + "step": 29808 + }, + { + "epoch": 0.6987699033621079, + "grad_norm": 0.33662745356559753, + "learning_rate": 4.155151340487543e-05, + "loss": 0.0536, + "step": 29809 + }, + { + "epoch": 0.6987933449369128, + "grad_norm": 0.3376355469226837, + "learning_rate": 4.154553662902831e-05, + "loss": 0.0824, + "step": 29810 + }, + { + "epoch": 0.6988167865117179, + "grad_norm": 0.5433335304260254, + "learning_rate": 4.153956017035676e-05, + "loss": 0.13, + "step": 29811 + }, + { + "epoch": 0.6988402280865229, + "grad_norm": 0.19044546782970428, + "learning_rate": 4.153358402889319e-05, + "loss": 0.0236, + "step": 29812 + }, + { + "epoch": 0.6988636696613278, + "grad_norm": 0.3378973603248596, + "learning_rate": 4.1527608204670056e-05, + "loss": 0.0699, + "step": 29813 + }, + { + "epoch": 0.6988871112361329, + "grad_norm": 0.39605456590652466, + "learning_rate": 4.152163269771975e-05, + "loss": 0.0483, + "step": 29814 + }, + { + "epoch": 0.6989105528109378, + "grad_norm": 0.08838135004043579, + "learning_rate": 4.151565750807467e-05, + "loss": 0.0199, + "step": 29815 + }, + { + "epoch": 0.6989339943857429, + "grad_norm": 0.4323650896549225, + "learning_rate": 4.1509682635767274e-05, + "loss": 0.0719, + "step": 29816 + }, + { + "epoch": 0.6989574359605478, + "grad_norm": 0.2813461124897003, + "learning_rate": 4.150370808083004e-05, + "loss": 0.0364, + "step": 29817 + }, + { + "epoch": 0.6989808775353529, + "grad_norm": 0.2112782746553421, + "learning_rate": 4.1497733843295314e-05, + "loss": 0.0272, + "step": 29818 + }, + { + "epoch": 0.6990043191101578, + "grad_norm": 0.16273954510688782, + "learning_rate": 4.149175992319554e-05, + "loss": 0.0246, + "step": 29819 + }, + { + "epoch": 0.6990277606849629, + "grad_norm": 0.547878623008728, + "learning_rate": 4.148578632056311e-05, + "loss": 0.0507, + "step": 29820 + }, + { + "epoch": 0.6990512022597678, + "grad_norm": 0.8167666792869568, + "learning_rate": 4.147981303543046e-05, + "loss": 0.1561, + "step": 29821 + }, + { + "epoch": 0.6990746438345729, + "grad_norm": 0.1371716558933258, + "learning_rate": 4.147384006782995e-05, + "loss": 0.0265, + "step": 29822 + }, + { + "epoch": 0.6990980854093778, + "grad_norm": 0.30635735392570496, + "learning_rate": 4.146786741779406e-05, + "loss": 0.0551, + "step": 29823 + }, + { + "epoch": 0.6991215269841828, + "grad_norm": 0.44520387053489685, + "learning_rate": 4.1461895085355175e-05, + "loss": 0.4754, + "step": 29824 + }, + { + "epoch": 0.6991449685589878, + "grad_norm": 0.10325460880994797, + "learning_rate": 4.145592307054569e-05, + "loss": 0.0205, + "step": 29825 + }, + { + "epoch": 0.6991684101337928, + "grad_norm": 0.24096760153770447, + "learning_rate": 4.144995137339801e-05, + "loss": 0.0424, + "step": 29826 + }, + { + "epoch": 0.6991918517085978, + "grad_norm": 0.3826867938041687, + "learning_rate": 4.1443979993944504e-05, + "loss": 0.0419, + "step": 29827 + }, + { + "epoch": 0.6992152932834028, + "grad_norm": 0.5070840120315552, + "learning_rate": 4.143800893221761e-05, + "loss": 0.0388, + "step": 29828 + }, + { + "epoch": 0.6992387348582078, + "grad_norm": 0.07986951619386673, + "learning_rate": 4.143203818824977e-05, + "loss": 0.0094, + "step": 29829 + }, + { + "epoch": 0.6992621764330128, + "grad_norm": 0.4055461883544922, + "learning_rate": 4.142606776207333e-05, + "loss": 0.0682, + "step": 29830 + }, + { + "epoch": 0.6992856180078177, + "grad_norm": 0.3115746080875397, + "learning_rate": 4.142009765372069e-05, + "loss": 0.0318, + "step": 29831 + }, + { + "epoch": 0.6993090595826228, + "grad_norm": 0.11460905522108078, + "learning_rate": 4.141412786322425e-05, + "loss": 0.0257, + "step": 29832 + }, + { + "epoch": 0.6993325011574277, + "grad_norm": 0.46692192554473877, + "learning_rate": 4.1408158390616406e-05, + "loss": 0.0869, + "step": 29833 + }, + { + "epoch": 0.6993559427322328, + "grad_norm": 0.3428802788257599, + "learning_rate": 4.14021892359295e-05, + "loss": 0.0579, + "step": 29834 + }, + { + "epoch": 0.6993793843070377, + "grad_norm": 0.44847649335861206, + "learning_rate": 4.139622039919599e-05, + "loss": 0.0793, + "step": 29835 + }, + { + "epoch": 0.6994028258818428, + "grad_norm": 0.17247003316879272, + "learning_rate": 4.139025188044825e-05, + "loss": 0.0222, + "step": 29836 + }, + { + "epoch": 0.6994262674566477, + "grad_norm": 0.39696407318115234, + "learning_rate": 4.138428367971865e-05, + "loss": 0.0845, + "step": 29837 + }, + { + "epoch": 0.6994497090314528, + "grad_norm": 0.6118237376213074, + "learning_rate": 4.137831579703957e-05, + "loss": 0.6182, + "step": 29838 + }, + { + "epoch": 0.6994731506062577, + "grad_norm": 0.26295340061187744, + "learning_rate": 4.137234823244337e-05, + "loss": 0.0455, + "step": 29839 + }, + { + "epoch": 0.6994965921810627, + "grad_norm": 0.5707114934921265, + "learning_rate": 4.136638098596248e-05, + "loss": 0.6813, + "step": 29840 + }, + { + "epoch": 0.6995200337558677, + "grad_norm": 0.5592187643051147, + "learning_rate": 4.1360414057629224e-05, + "loss": 0.0463, + "step": 29841 + }, + { + "epoch": 0.6995434753306727, + "grad_norm": 0.7555748224258423, + "learning_rate": 4.1354447447476054e-05, + "loss": 0.0944, + "step": 29842 + }, + { + "epoch": 0.6995669169054777, + "grad_norm": 0.13335193693637848, + "learning_rate": 4.1348481155535294e-05, + "loss": 0.0244, + "step": 29843 + }, + { + "epoch": 0.6995903584802827, + "grad_norm": 0.4213569164276123, + "learning_rate": 4.1342515181839326e-05, + "loss": 0.1039, + "step": 29844 + }, + { + "epoch": 0.6996138000550877, + "grad_norm": 0.13210757076740265, + "learning_rate": 4.133654952642051e-05, + "loss": 0.0247, + "step": 29845 + }, + { + "epoch": 0.6996372416298927, + "grad_norm": 0.21024568378925323, + "learning_rate": 4.133058418931124e-05, + "loss": 0.0269, + "step": 29846 + }, + { + "epoch": 0.6996606832046977, + "grad_norm": 0.5778928399085999, + "learning_rate": 4.132461917054382e-05, + "loss": 0.1799, + "step": 29847 + }, + { + "epoch": 0.6996841247795027, + "grad_norm": 0.20181021094322205, + "learning_rate": 4.131865447015072e-05, + "loss": 0.0324, + "step": 29848 + }, + { + "epoch": 0.6997075663543076, + "grad_norm": 0.5281000733375549, + "learning_rate": 4.131269008816424e-05, + "loss": 0.1117, + "step": 29849 + }, + { + "epoch": 0.6997310079291127, + "grad_norm": 0.4167478382587433, + "learning_rate": 4.130672602461675e-05, + "loss": 0.5127, + "step": 29850 + }, + { + "epoch": 0.6997544495039176, + "grad_norm": 0.3083464801311493, + "learning_rate": 4.1300762279540583e-05, + "loss": 0.0534, + "step": 29851 + }, + { + "epoch": 0.6997778910787227, + "grad_norm": 0.3751870095729828, + "learning_rate": 4.1294798852968166e-05, + "loss": 0.0507, + "step": 29852 + }, + { + "epoch": 0.6998013326535276, + "grad_norm": 0.5210330486297607, + "learning_rate": 4.128883574493183e-05, + "loss": 0.1222, + "step": 29853 + }, + { + "epoch": 0.6998247742283327, + "grad_norm": 0.37171223759651184, + "learning_rate": 4.128287295546387e-05, + "loss": 0.0419, + "step": 29854 + }, + { + "epoch": 0.6998482158031376, + "grad_norm": 0.5429425835609436, + "learning_rate": 4.127691048459674e-05, + "loss": 0.1047, + "step": 29855 + }, + { + "epoch": 0.6998716573779427, + "grad_norm": 0.37153175473213196, + "learning_rate": 4.127094833236275e-05, + "loss": 0.0628, + "step": 29856 + }, + { + "epoch": 0.6998950989527476, + "grad_norm": 0.5106928944587708, + "learning_rate": 4.126498649879424e-05, + "loss": 0.0827, + "step": 29857 + }, + { + "epoch": 0.6999185405275526, + "grad_norm": 0.48677632212638855, + "learning_rate": 4.125902498392355e-05, + "loss": 0.1194, + "step": 29858 + }, + { + "epoch": 0.6999419821023576, + "grad_norm": 0.15962514281272888, + "learning_rate": 4.125306378778302e-05, + "loss": 0.037, + "step": 29859 + }, + { + "epoch": 0.6999654236771626, + "grad_norm": 0.3082221746444702, + "learning_rate": 4.124710291040506e-05, + "loss": 0.0338, + "step": 29860 + }, + { + "epoch": 0.6999888652519676, + "grad_norm": 0.39983758330345154, + "learning_rate": 4.124114235182195e-05, + "loss": 0.4713, + "step": 29861 + }, + { + "epoch": 0.7000123068267726, + "grad_norm": 0.39243564009666443, + "learning_rate": 4.1235182112066076e-05, + "loss": 0.0319, + "step": 29862 + }, + { + "epoch": 0.7000357484015777, + "grad_norm": 0.670035183429718, + "learning_rate": 4.1229222191169734e-05, + "loss": 0.5801, + "step": 29863 + }, + { + "epoch": 0.7000591899763826, + "grad_norm": 0.4953172504901886, + "learning_rate": 4.122326258916527e-05, + "loss": 0.0847, + "step": 29864 + }, + { + "epoch": 0.7000826315511877, + "grad_norm": 0.11566688120365143, + "learning_rate": 4.121730330608506e-05, + "loss": 0.025, + "step": 29865 + }, + { + "epoch": 0.7001060731259926, + "grad_norm": 0.07955815643072128, + "learning_rate": 4.121134434196139e-05, + "loss": 0.0176, + "step": 29866 + }, + { + "epoch": 0.7001295147007977, + "grad_norm": 0.33263975381851196, + "learning_rate": 4.1205385696826646e-05, + "loss": 0.0486, + "step": 29867 + }, + { + "epoch": 0.7001529562756026, + "grad_norm": 0.4518606960773468, + "learning_rate": 4.119942737071313e-05, + "loss": 0.0432, + "step": 29868 + }, + { + "epoch": 0.7001763978504076, + "grad_norm": 0.3857332170009613, + "learning_rate": 4.119346936365318e-05, + "loss": 0.1068, + "step": 29869 + }, + { + "epoch": 0.7001998394252126, + "grad_norm": 0.36197638511657715, + "learning_rate": 4.118751167567911e-05, + "loss": 0.3748, + "step": 29870 + }, + { + "epoch": 0.7002232810000176, + "grad_norm": 0.5061367750167847, + "learning_rate": 4.1181554306823265e-05, + "loss": 0.0656, + "step": 29871 + }, + { + "epoch": 0.7002467225748226, + "grad_norm": 0.4679441750049591, + "learning_rate": 4.1175597257117926e-05, + "loss": 0.0885, + "step": 29872 + }, + { + "epoch": 0.7002701641496276, + "grad_norm": 0.103702113032341, + "learning_rate": 4.1169640526595487e-05, + "loss": 0.008, + "step": 29873 + }, + { + "epoch": 0.7002936057244326, + "grad_norm": 0.12676896154880524, + "learning_rate": 4.116368411528823e-05, + "loss": 0.0259, + "step": 29874 + }, + { + "epoch": 0.7003170472992376, + "grad_norm": 0.3929433822631836, + "learning_rate": 4.115772802322847e-05, + "loss": 0.0742, + "step": 29875 + }, + { + "epoch": 0.7003404888740425, + "grad_norm": 0.36922770738601685, + "learning_rate": 4.11517722504485e-05, + "loss": 0.0597, + "step": 29876 + }, + { + "epoch": 0.7003639304488476, + "grad_norm": 0.45289140939712524, + "learning_rate": 4.1145816796980705e-05, + "loss": 0.0458, + "step": 29877 + }, + { + "epoch": 0.7003873720236525, + "grad_norm": 0.38575151562690735, + "learning_rate": 4.113986166285736e-05, + "loss": 0.0841, + "step": 29878 + }, + { + "epoch": 0.7004108135984576, + "grad_norm": 0.14775486290454865, + "learning_rate": 4.1133906848110745e-05, + "loss": 0.0334, + "step": 29879 + }, + { + "epoch": 0.7004342551732625, + "grad_norm": 0.5899816155433655, + "learning_rate": 4.112795235277325e-05, + "loss": 0.1008, + "step": 29880 + }, + { + "epoch": 0.7004576967480676, + "grad_norm": 0.27866408228874207, + "learning_rate": 4.112199817687714e-05, + "loss": 0.0595, + "step": 29881 + }, + { + "epoch": 0.7004811383228725, + "grad_norm": 0.07481184601783752, + "learning_rate": 4.111604432045471e-05, + "loss": 0.0145, + "step": 29882 + }, + { + "epoch": 0.7005045798976776, + "grad_norm": 0.1320251226425171, + "learning_rate": 4.111009078353829e-05, + "loss": 0.0129, + "step": 29883 + }, + { + "epoch": 0.7005280214724825, + "grad_norm": 0.5503548979759216, + "learning_rate": 4.110413756616018e-05, + "loss": 0.1537, + "step": 29884 + }, + { + "epoch": 0.7005514630472875, + "grad_norm": 0.1323990821838379, + "learning_rate": 4.1098184668352636e-05, + "loss": 0.0285, + "step": 29885 + }, + { + "epoch": 0.7005749046220925, + "grad_norm": 0.47140565514564514, + "learning_rate": 4.109223209014802e-05, + "loss": 0.0519, + "step": 29886 + }, + { + "epoch": 0.7005983461968975, + "grad_norm": 0.5400969982147217, + "learning_rate": 4.1086279831578624e-05, + "loss": 0.0654, + "step": 29887 + }, + { + "epoch": 0.7006217877717025, + "grad_norm": 0.16181856393814087, + "learning_rate": 4.108032789267669e-05, + "loss": 0.0392, + "step": 29888 + }, + { + "epoch": 0.7006452293465075, + "grad_norm": 0.2625575661659241, + "learning_rate": 4.107437627347459e-05, + "loss": 0.0215, + "step": 29889 + }, + { + "epoch": 0.7006686709213125, + "grad_norm": 0.1284146010875702, + "learning_rate": 4.106842497400458e-05, + "loss": 0.0337, + "step": 29890 + }, + { + "epoch": 0.7006921124961175, + "grad_norm": 0.47867223620414734, + "learning_rate": 4.106247399429896e-05, + "loss": 0.5973, + "step": 29891 + }, + { + "epoch": 0.7007155540709225, + "grad_norm": 0.6026161909103394, + "learning_rate": 4.1056523334389965e-05, + "loss": 0.0813, + "step": 29892 + }, + { + "epoch": 0.7007389956457275, + "grad_norm": 0.43652960658073425, + "learning_rate": 4.105057299430997e-05, + "loss": 0.0907, + "step": 29893 + }, + { + "epoch": 0.7007624372205324, + "grad_norm": 0.24888773262500763, + "learning_rate": 4.104462297409123e-05, + "loss": 0.0386, + "step": 29894 + }, + { + "epoch": 0.7007858787953375, + "grad_norm": 0.15661761164665222, + "learning_rate": 4.103867327376601e-05, + "loss": 0.025, + "step": 29895 + }, + { + "epoch": 0.7008093203701424, + "grad_norm": 0.20380111038684845, + "learning_rate": 4.103272389336662e-05, + "loss": 0.0549, + "step": 29896 + }, + { + "epoch": 0.7008327619449475, + "grad_norm": 0.4900831878185272, + "learning_rate": 4.102677483292529e-05, + "loss": 0.116, + "step": 29897 + }, + { + "epoch": 0.7008562035197524, + "grad_norm": 0.38491445779800415, + "learning_rate": 4.102082609247437e-05, + "loss": 0.0363, + "step": 29898 + }, + { + "epoch": 0.7008796450945575, + "grad_norm": 0.3657548427581787, + "learning_rate": 4.1014877672046115e-05, + "loss": 0.0983, + "step": 29899 + }, + { + "epoch": 0.7009030866693624, + "grad_norm": 0.23513638973236084, + "learning_rate": 4.100892957167275e-05, + "loss": 0.0331, + "step": 29900 + }, + { + "epoch": 0.7009265282441675, + "grad_norm": 0.9562849998474121, + "learning_rate": 4.100298179138663e-05, + "loss": 0.0878, + "step": 29901 + }, + { + "epoch": 0.7009499698189724, + "grad_norm": 0.09070336818695068, + "learning_rate": 4.099703433121998e-05, + "loss": 0.0206, + "step": 29902 + }, + { + "epoch": 0.7009734113937774, + "grad_norm": 0.6717003583908081, + "learning_rate": 4.099108719120509e-05, + "loss": 0.0714, + "step": 29903 + }, + { + "epoch": 0.7009968529685824, + "grad_norm": 0.396664559841156, + "learning_rate": 4.098514037137419e-05, + "loss": 0.0864, + "step": 29904 + }, + { + "epoch": 0.7010202945433874, + "grad_norm": 0.24274390935897827, + "learning_rate": 4.0979193871759604e-05, + "loss": 0.0569, + "step": 29905 + }, + { + "epoch": 0.7010437361181924, + "grad_norm": 0.32334622740745544, + "learning_rate": 4.097324769239357e-05, + "loss": 0.0884, + "step": 29906 + }, + { + "epoch": 0.7010671776929974, + "grad_norm": 0.09613874554634094, + "learning_rate": 4.0967301833308357e-05, + "loss": 0.0168, + "step": 29907 + }, + { + "epoch": 0.7010906192678024, + "grad_norm": 0.37527334690093994, + "learning_rate": 4.096135629453622e-05, + "loss": 0.0619, + "step": 29908 + }, + { + "epoch": 0.7011140608426074, + "grad_norm": 0.4900341331958771, + "learning_rate": 4.095541107610943e-05, + "loss": 0.114, + "step": 29909 + }, + { + "epoch": 0.7011375024174124, + "grad_norm": 0.9731540679931641, + "learning_rate": 4.09494661780602e-05, + "loss": 0.1479, + "step": 29910 + }, + { + "epoch": 0.7011609439922174, + "grad_norm": 0.825448215007782, + "learning_rate": 4.0943521600420866e-05, + "loss": 0.188, + "step": 29911 + }, + { + "epoch": 0.7011843855670223, + "grad_norm": 0.388094037771225, + "learning_rate": 4.093757734322361e-05, + "loss": 0.0465, + "step": 29912 + }, + { + "epoch": 0.7012078271418274, + "grad_norm": 0.6928248405456543, + "learning_rate": 4.0931633406500744e-05, + "loss": 0.1492, + "step": 29913 + }, + { + "epoch": 0.7012312687166324, + "grad_norm": 0.32659095525741577, + "learning_rate": 4.0925689790284515e-05, + "loss": 0.0496, + "step": 29914 + }, + { + "epoch": 0.7012547102914374, + "grad_norm": 0.3755934238433838, + "learning_rate": 4.0919746494607135e-05, + "loss": 0.0892, + "step": 29915 + }, + { + "epoch": 0.7012781518662424, + "grad_norm": 0.6999070644378662, + "learning_rate": 4.091380351950088e-05, + "loss": 0.0899, + "step": 29916 + }, + { + "epoch": 0.7013015934410474, + "grad_norm": 0.5259009599685669, + "learning_rate": 4.090786086499796e-05, + "loss": 0.0918, + "step": 29917 + }, + { + "epoch": 0.7013250350158524, + "grad_norm": 0.348969966173172, + "learning_rate": 4.090191853113068e-05, + "loss": 0.3772, + "step": 29918 + }, + { + "epoch": 0.7013484765906574, + "grad_norm": 0.1532200127840042, + "learning_rate": 4.089597651793124e-05, + "loss": 0.0192, + "step": 29919 + }, + { + "epoch": 0.7013719181654624, + "grad_norm": 0.5449056625366211, + "learning_rate": 4.089003482543191e-05, + "loss": 0.1411, + "step": 29920 + }, + { + "epoch": 0.7013953597402673, + "grad_norm": 0.31934890151023865, + "learning_rate": 4.088409345366491e-05, + "loss": 0.0832, + "step": 29921 + }, + { + "epoch": 0.7014188013150724, + "grad_norm": 0.46444594860076904, + "learning_rate": 4.0878152402662475e-05, + "loss": 0.0908, + "step": 29922 + }, + { + "epoch": 0.7014422428898773, + "grad_norm": 0.1794469654560089, + "learning_rate": 4.087221167245681e-05, + "loss": 0.0358, + "step": 29923 + }, + { + "epoch": 0.7014656844646824, + "grad_norm": 0.5881567597389221, + "learning_rate": 4.086627126308019e-05, + "loss": 0.1065, + "step": 29924 + }, + { + "epoch": 0.7014891260394873, + "grad_norm": 0.34298175573349, + "learning_rate": 4.086033117456488e-05, + "loss": 0.0587, + "step": 29925 + }, + { + "epoch": 0.7015125676142924, + "grad_norm": 0.6051222085952759, + "learning_rate": 4.085439140694308e-05, + "loss": 0.0986, + "step": 29926 + }, + { + "epoch": 0.7015360091890973, + "grad_norm": 0.5157023072242737, + "learning_rate": 4.084845196024701e-05, + "loss": 0.6573, + "step": 29927 + }, + { + "epoch": 0.7015594507639024, + "grad_norm": 0.30403342843055725, + "learning_rate": 4.084251283450891e-05, + "loss": 0.0674, + "step": 29928 + }, + { + "epoch": 0.7015828923387073, + "grad_norm": 0.1345342993736267, + "learning_rate": 4.083657402976099e-05, + "loss": 0.0164, + "step": 29929 + }, + { + "epoch": 0.7016063339135123, + "grad_norm": 0.517936110496521, + "learning_rate": 4.083063554603546e-05, + "loss": 0.4574, + "step": 29930 + }, + { + "epoch": 0.7016297754883173, + "grad_norm": 0.7082779407501221, + "learning_rate": 4.0824697383364585e-05, + "loss": 0.1214, + "step": 29931 + }, + { + "epoch": 0.7016532170631223, + "grad_norm": 0.12558160722255707, + "learning_rate": 4.081875954178057e-05, + "loss": 0.0339, + "step": 29932 + }, + { + "epoch": 0.7016766586379273, + "grad_norm": 0.24458502233028412, + "learning_rate": 4.0812822021315635e-05, + "loss": 0.032, + "step": 29933 + }, + { + "epoch": 0.7017001002127323, + "grad_norm": 0.2101912945508957, + "learning_rate": 4.080688482200199e-05, + "loss": 0.035, + "step": 29934 + }, + { + "epoch": 0.7017235417875373, + "grad_norm": 0.2997920513153076, + "learning_rate": 4.080094794387181e-05, + "loss": 0.0753, + "step": 29935 + }, + { + "epoch": 0.7017469833623423, + "grad_norm": 0.46469882130622864, + "learning_rate": 4.0795011386957395e-05, + "loss": 0.4804, + "step": 29936 + }, + { + "epoch": 0.7017704249371473, + "grad_norm": 0.2803923785686493, + "learning_rate": 4.0789075151290876e-05, + "loss": 0.0327, + "step": 29937 + }, + { + "epoch": 0.7017938665119523, + "grad_norm": 0.40542852878570557, + "learning_rate": 4.078313923690453e-05, + "loss": 0.027, + "step": 29938 + }, + { + "epoch": 0.7018173080867572, + "grad_norm": 0.8277601003646851, + "learning_rate": 4.077720364383055e-05, + "loss": 0.0949, + "step": 29939 + }, + { + "epoch": 0.7018407496615623, + "grad_norm": 0.5421988368034363, + "learning_rate": 4.077126837210111e-05, + "loss": 0.1019, + "step": 29940 + }, + { + "epoch": 0.7018641912363672, + "grad_norm": 0.298258900642395, + "learning_rate": 4.0765333421748456e-05, + "loss": 0.0468, + "step": 29941 + }, + { + "epoch": 0.7018876328111723, + "grad_norm": 0.36979615688323975, + "learning_rate": 4.0759398792804715e-05, + "loss": 0.0417, + "step": 29942 + }, + { + "epoch": 0.7019110743859772, + "grad_norm": 0.14915338158607483, + "learning_rate": 4.075346448530217e-05, + "loss": 0.0357, + "step": 29943 + }, + { + "epoch": 0.7019345159607823, + "grad_norm": 0.09614070504903793, + "learning_rate": 4.074753049927301e-05, + "loss": 0.0183, + "step": 29944 + }, + { + "epoch": 0.7019579575355872, + "grad_norm": 0.3339653015136719, + "learning_rate": 4.074159683474941e-05, + "loss": 0.0605, + "step": 29945 + }, + { + "epoch": 0.7019813991103923, + "grad_norm": 0.3635723292827606, + "learning_rate": 4.073566349176357e-05, + "loss": 0.0663, + "step": 29946 + }, + { + "epoch": 0.7020048406851972, + "grad_norm": 0.4166530668735504, + "learning_rate": 4.0729730470347685e-05, + "loss": 0.102, + "step": 29947 + }, + { + "epoch": 0.7020282822600022, + "grad_norm": 0.4966287612915039, + "learning_rate": 4.072379777053392e-05, + "loss": 0.1051, + "step": 29948 + }, + { + "epoch": 0.7020517238348072, + "grad_norm": 0.10979878902435303, + "learning_rate": 4.0717865392354494e-05, + "loss": 0.0126, + "step": 29949 + }, + { + "epoch": 0.7020751654096122, + "grad_norm": 0.2065083384513855, + "learning_rate": 4.071193333584162e-05, + "loss": 0.0102, + "step": 29950 + }, + { + "epoch": 0.7020986069844172, + "grad_norm": 0.2500617504119873, + "learning_rate": 4.070600160102748e-05, + "loss": 0.0361, + "step": 29951 + }, + { + "epoch": 0.7021220485592222, + "grad_norm": 0.2213549017906189, + "learning_rate": 4.070007018794423e-05, + "loss": 0.0522, + "step": 29952 + }, + { + "epoch": 0.7021454901340272, + "grad_norm": 0.41236793994903564, + "learning_rate": 4.069413909662407e-05, + "loss": 0.0297, + "step": 29953 + }, + { + "epoch": 0.7021689317088322, + "grad_norm": 0.707714319229126, + "learning_rate": 4.0688208327099184e-05, + "loss": 0.7531, + "step": 29954 + }, + { + "epoch": 0.7021923732836372, + "grad_norm": 0.3396390676498413, + "learning_rate": 4.068227787940171e-05, + "loss": 0.0678, + "step": 29955 + }, + { + "epoch": 0.7022158148584422, + "grad_norm": 0.37957191467285156, + "learning_rate": 4.0676347753563896e-05, + "loss": 0.0874, + "step": 29956 + }, + { + "epoch": 0.7022392564332471, + "grad_norm": 0.27073681354522705, + "learning_rate": 4.067041794961789e-05, + "loss": 0.0584, + "step": 29957 + }, + { + "epoch": 0.7022626980080522, + "grad_norm": 0.2612968683242798, + "learning_rate": 4.066448846759586e-05, + "loss": 0.0429, + "step": 29958 + }, + { + "epoch": 0.7022861395828571, + "grad_norm": 0.2056065797805786, + "learning_rate": 4.065855930752999e-05, + "loss": 0.0212, + "step": 29959 + }, + { + "epoch": 0.7023095811576622, + "grad_norm": 0.6470343470573425, + "learning_rate": 4.065263046945243e-05, + "loss": 0.1369, + "step": 29960 + }, + { + "epoch": 0.7023330227324671, + "grad_norm": 0.7034005522727966, + "learning_rate": 4.064670195339535e-05, + "loss": 0.1176, + "step": 29961 + }, + { + "epoch": 0.7023564643072722, + "grad_norm": 0.5621327757835388, + "learning_rate": 4.064077375939093e-05, + "loss": 0.1011, + "step": 29962 + }, + { + "epoch": 0.7023799058820771, + "grad_norm": 0.22254379093647003, + "learning_rate": 4.063484588747137e-05, + "loss": 0.0461, + "step": 29963 + }, + { + "epoch": 0.7024033474568822, + "grad_norm": 0.25654053688049316, + "learning_rate": 4.0628918337668806e-05, + "loss": 0.0405, + "step": 29964 + }, + { + "epoch": 0.7024267890316871, + "grad_norm": 0.6519685983657837, + "learning_rate": 4.0622991110015407e-05, + "loss": 0.1115, + "step": 29965 + }, + { + "epoch": 0.7024502306064921, + "grad_norm": 0.4341566264629364, + "learning_rate": 4.061706420454332e-05, + "loss": 0.0885, + "step": 29966 + }, + { + "epoch": 0.7024736721812972, + "grad_norm": 0.2962876856327057, + "learning_rate": 4.0611137621284724e-05, + "loss": 0.045, + "step": 29967 + }, + { + "epoch": 0.7024971137561021, + "grad_norm": 0.06797405332326889, + "learning_rate": 4.0605211360271714e-05, + "loss": 0.0086, + "step": 29968 + }, + { + "epoch": 0.7025205553309072, + "grad_norm": 0.4569087028503418, + "learning_rate": 4.059928542153654e-05, + "loss": 0.1369, + "step": 29969 + }, + { + "epoch": 0.7025439969057121, + "grad_norm": 0.15104222297668457, + "learning_rate": 4.059335980511132e-05, + "loss": 0.0258, + "step": 29970 + }, + { + "epoch": 0.7025674384805172, + "grad_norm": 0.30508875846862793, + "learning_rate": 4.058743451102819e-05, + "loss": 0.0534, + "step": 29971 + }, + { + "epoch": 0.7025908800553221, + "grad_norm": 0.36288392543792725, + "learning_rate": 4.058150953931932e-05, + "loss": 0.0762, + "step": 29972 + }, + { + "epoch": 0.7026143216301272, + "grad_norm": 0.6083279252052307, + "learning_rate": 4.057558489001681e-05, + "loss": 0.1163, + "step": 29973 + }, + { + "epoch": 0.7026377632049321, + "grad_norm": 0.4451356828212738, + "learning_rate": 4.056966056315288e-05, + "loss": 0.0642, + "step": 29974 + }, + { + "epoch": 0.7026612047797371, + "grad_norm": 0.28771647810935974, + "learning_rate": 4.05637365587596e-05, + "loss": 0.0364, + "step": 29975 + }, + { + "epoch": 0.7026846463545421, + "grad_norm": 0.3111538290977478, + "learning_rate": 4.055781287686921e-05, + "loss": 0.0437, + "step": 29976 + }, + { + "epoch": 0.7027080879293471, + "grad_norm": 0.10314740240573883, + "learning_rate": 4.0551889517513784e-05, + "loss": 0.0115, + "step": 29977 + }, + { + "epoch": 0.7027315295041521, + "grad_norm": 0.328154981136322, + "learning_rate": 4.054596648072549e-05, + "loss": 0.0407, + "step": 29978 + }, + { + "epoch": 0.7027549710789571, + "grad_norm": 0.4435855448246002, + "learning_rate": 4.054004376653644e-05, + "loss": 0.0894, + "step": 29979 + }, + { + "epoch": 0.7027784126537621, + "grad_norm": 0.2863585352897644, + "learning_rate": 4.053412137497876e-05, + "loss": 0.0348, + "step": 29980 + }, + { + "epoch": 0.7028018542285671, + "grad_norm": 0.5012056231498718, + "learning_rate": 4.052819930608463e-05, + "loss": 0.1077, + "step": 29981 + }, + { + "epoch": 0.702825295803372, + "grad_norm": 0.45684540271759033, + "learning_rate": 4.0522277559886176e-05, + "loss": 0.0596, + "step": 29982 + }, + { + "epoch": 0.7028487373781771, + "grad_norm": 0.4954434335231781, + "learning_rate": 4.051635613641552e-05, + "loss": 0.6714, + "step": 29983 + }, + { + "epoch": 0.702872178952982, + "grad_norm": 0.43400657176971436, + "learning_rate": 4.0510435035704777e-05, + "loss": 0.0753, + "step": 29984 + }, + { + "epoch": 0.7028956205277871, + "grad_norm": 0.2455093413591385, + "learning_rate": 4.050451425778605e-05, + "loss": 0.0464, + "step": 29985 + }, + { + "epoch": 0.702919062102592, + "grad_norm": 0.20897993445396423, + "learning_rate": 4.049859380269155e-05, + "loss": 0.034, + "step": 29986 + }, + { + "epoch": 0.7029425036773971, + "grad_norm": 0.28240978717803955, + "learning_rate": 4.049267367045331e-05, + "loss": 0.0535, + "step": 29987 + }, + { + "epoch": 0.702965945252202, + "grad_norm": 0.3359493911266327, + "learning_rate": 4.048675386110353e-05, + "loss": 0.0635, + "step": 29988 + }, + { + "epoch": 0.7029893868270071, + "grad_norm": 0.3681545555591583, + "learning_rate": 4.0480834374674304e-05, + "loss": 0.0674, + "step": 29989 + }, + { + "epoch": 0.703012828401812, + "grad_norm": 0.45668095350265503, + "learning_rate": 4.047491521119774e-05, + "loss": 0.073, + "step": 29990 + }, + { + "epoch": 0.7030362699766171, + "grad_norm": 0.5994359850883484, + "learning_rate": 4.046899637070596e-05, + "loss": 0.1395, + "step": 29991 + }, + { + "epoch": 0.703059711551422, + "grad_norm": 0.38330087065696716, + "learning_rate": 4.0463077853231066e-05, + "loss": 0.378, + "step": 29992 + }, + { + "epoch": 0.703083153126227, + "grad_norm": 0.2841334640979767, + "learning_rate": 4.045715965880517e-05, + "loss": 0.0386, + "step": 29993 + }, + { + "epoch": 0.703106594701032, + "grad_norm": 0.29853710532188416, + "learning_rate": 4.0451241787460424e-05, + "loss": 0.0324, + "step": 29994 + }, + { + "epoch": 0.703130036275837, + "grad_norm": 0.1307695358991623, + "learning_rate": 4.044532423922892e-05, + "loss": 0.017, + "step": 29995 + }, + { + "epoch": 0.703153477850642, + "grad_norm": 0.4772081673145294, + "learning_rate": 4.0439407014142764e-05, + "loss": 0.0306, + "step": 29996 + }, + { + "epoch": 0.703176919425447, + "grad_norm": 0.6406065225601196, + "learning_rate": 4.0433490112234004e-05, + "loss": 0.5219, + "step": 29997 + }, + { + "epoch": 0.703200361000252, + "grad_norm": 0.5285840034484863, + "learning_rate": 4.042757353353486e-05, + "loss": 0.6706, + "step": 29998 + }, + { + "epoch": 0.703223802575057, + "grad_norm": 0.16662408411502838, + "learning_rate": 4.0421657278077375e-05, + "loss": 0.0201, + "step": 29999 + }, + { + "epoch": 0.703247244149862, + "grad_norm": 0.42298153042793274, + "learning_rate": 4.041574134589361e-05, + "loss": 0.0767, + "step": 30000 + }, + { + "epoch": 0.703270685724667, + "grad_norm": 0.1798776388168335, + "learning_rate": 4.0409825737015736e-05, + "loss": 0.0359, + "step": 30001 + }, + { + "epoch": 0.7032941272994719, + "grad_norm": 0.34202444553375244, + "learning_rate": 4.0403910451475834e-05, + "loss": 0.0561, + "step": 30002 + }, + { + "epoch": 0.703317568874277, + "grad_norm": 0.2536643147468567, + "learning_rate": 4.0397995489306e-05, + "loss": 0.0303, + "step": 30003 + }, + { + "epoch": 0.7033410104490819, + "grad_norm": 0.27811503410339355, + "learning_rate": 4.0392080850538296e-05, + "loss": 0.0432, + "step": 30004 + }, + { + "epoch": 0.703364452023887, + "grad_norm": 0.3949863612651825, + "learning_rate": 4.038616653520485e-05, + "loss": 0.0798, + "step": 30005 + }, + { + "epoch": 0.7033878935986919, + "grad_norm": 0.18758843839168549, + "learning_rate": 4.0380252543337706e-05, + "loss": 0.0305, + "step": 30006 + }, + { + "epoch": 0.703411335173497, + "grad_norm": 0.6156074404716492, + "learning_rate": 4.037433887496902e-05, + "loss": 0.5646, + "step": 30007 + }, + { + "epoch": 0.7034347767483019, + "grad_norm": 0.13287317752838135, + "learning_rate": 4.036842553013085e-05, + "loss": 0.0285, + "step": 30008 + }, + { + "epoch": 0.703458218323107, + "grad_norm": 0.523288369178772, + "learning_rate": 4.0362512508855244e-05, + "loss": 0.0824, + "step": 30009 + }, + { + "epoch": 0.7034816598979119, + "grad_norm": 0.17687265574932098, + "learning_rate": 4.035659981117435e-05, + "loss": 0.0406, + "step": 30010 + }, + { + "epoch": 0.7035051014727169, + "grad_norm": 0.3016526699066162, + "learning_rate": 4.035068743712023e-05, + "loss": 0.056, + "step": 30011 + }, + { + "epoch": 0.7035285430475219, + "grad_norm": 0.38737931847572327, + "learning_rate": 4.034477538672495e-05, + "loss": 0.0851, + "step": 30012 + }, + { + "epoch": 0.7035519846223269, + "grad_norm": 0.6169300675392151, + "learning_rate": 4.033886366002057e-05, + "loss": 0.0988, + "step": 30013 + }, + { + "epoch": 0.7035754261971319, + "grad_norm": 0.14690639078617096, + "learning_rate": 4.0332952257039226e-05, + "loss": 0.0326, + "step": 30014 + }, + { + "epoch": 0.7035988677719369, + "grad_norm": 0.322762131690979, + "learning_rate": 4.0327041177812965e-05, + "loss": 0.0356, + "step": 30015 + }, + { + "epoch": 0.7036223093467419, + "grad_norm": 0.29665789008140564, + "learning_rate": 4.0321130422373844e-05, + "loss": 0.0531, + "step": 30016 + }, + { + "epoch": 0.7036457509215469, + "grad_norm": 0.42864036560058594, + "learning_rate": 4.031521999075395e-05, + "loss": 0.0934, + "step": 30017 + }, + { + "epoch": 0.703669192496352, + "grad_norm": 0.20086383819580078, + "learning_rate": 4.0309309882985323e-05, + "loss": 0.0241, + "step": 30018 + }, + { + "epoch": 0.7036926340711569, + "grad_norm": 0.026468226686120033, + "learning_rate": 4.0303400099100095e-05, + "loss": 0.0024, + "step": 30019 + }, + { + "epoch": 0.703716075645962, + "grad_norm": 0.22039292752742767, + "learning_rate": 4.029749063913029e-05, + "loss": 0.0226, + "step": 30020 + }, + { + "epoch": 0.7037395172207669, + "grad_norm": 0.308984637260437, + "learning_rate": 4.029158150310798e-05, + "loss": 0.0467, + "step": 30021 + }, + { + "epoch": 0.7037629587955719, + "grad_norm": 0.42389389872550964, + "learning_rate": 4.028567269106519e-05, + "loss": 0.1031, + "step": 30022 + }, + { + "epoch": 0.7037864003703769, + "grad_norm": 0.1892707794904709, + "learning_rate": 4.0279764203034063e-05, + "loss": 0.0305, + "step": 30023 + }, + { + "epoch": 0.7038098419451819, + "grad_norm": 0.14454051852226257, + "learning_rate": 4.0273856039046607e-05, + "loss": 0.0155, + "step": 30024 + }, + { + "epoch": 0.7038332835199869, + "grad_norm": 0.5389055013656616, + "learning_rate": 4.026794819913486e-05, + "loss": 0.1107, + "step": 30025 + }, + { + "epoch": 0.7038567250947919, + "grad_norm": 0.20802633464336395, + "learning_rate": 4.026204068333094e-05, + "loss": 0.0347, + "step": 30026 + }, + { + "epoch": 0.7038801666695969, + "grad_norm": 0.09630019962787628, + "learning_rate": 4.025613349166686e-05, + "loss": 0.0076, + "step": 30027 + }, + { + "epoch": 0.7039036082444019, + "grad_norm": 0.44212082028388977, + "learning_rate": 4.025022662417468e-05, + "loss": 0.12, + "step": 30028 + }, + { + "epoch": 0.7039270498192068, + "grad_norm": 0.6195065975189209, + "learning_rate": 4.0244320080886454e-05, + "loss": 0.5228, + "step": 30029 + }, + { + "epoch": 0.7039504913940119, + "grad_norm": 0.2855461537837982, + "learning_rate": 4.0238413861834224e-05, + "loss": 0.0337, + "step": 30030 + }, + { + "epoch": 0.7039739329688168, + "grad_norm": 0.13097795844078064, + "learning_rate": 4.023250796705001e-05, + "loss": 0.0209, + "step": 30031 + }, + { + "epoch": 0.7039973745436219, + "grad_norm": 0.3074691891670227, + "learning_rate": 4.022660239656593e-05, + "loss": 0.0551, + "step": 30032 + }, + { + "epoch": 0.7040208161184268, + "grad_norm": 0.17598024010658264, + "learning_rate": 4.022069715041397e-05, + "loss": 0.0121, + "step": 30033 + }, + { + "epoch": 0.7040442576932319, + "grad_norm": 0.43889984488487244, + "learning_rate": 4.0214792228626164e-05, + "loss": 0.087, + "step": 30034 + }, + { + "epoch": 0.7040676992680368, + "grad_norm": 0.4290090501308441, + "learning_rate": 4.0208887631234606e-05, + "loss": 0.6142, + "step": 30035 + }, + { + "epoch": 0.7040911408428419, + "grad_norm": 0.47759050130844116, + "learning_rate": 4.020298335827131e-05, + "loss": 0.0651, + "step": 30036 + }, + { + "epoch": 0.7041145824176468, + "grad_norm": 0.1603042334318161, + "learning_rate": 4.01970794097683e-05, + "loss": 0.0272, + "step": 30037 + }, + { + "epoch": 0.7041380239924518, + "grad_norm": 0.45506638288497925, + "learning_rate": 4.019117578575758e-05, + "loss": 0.6511, + "step": 30038 + }, + { + "epoch": 0.7041614655672568, + "grad_norm": 0.411196768283844, + "learning_rate": 4.0185272486271274e-05, + "loss": 0.0925, + "step": 30039 + }, + { + "epoch": 0.7041849071420618, + "grad_norm": 0.5220876336097717, + "learning_rate": 4.017936951134135e-05, + "loss": 0.5631, + "step": 30040 + }, + { + "epoch": 0.7042083487168668, + "grad_norm": 0.26027217507362366, + "learning_rate": 4.0173466860999854e-05, + "loss": 0.025, + "step": 30041 + }, + { + "epoch": 0.7042317902916718, + "grad_norm": 0.48590075969696045, + "learning_rate": 4.01675645352788e-05, + "loss": 0.1049, + "step": 30042 + }, + { + "epoch": 0.7042552318664768, + "grad_norm": 0.35135215520858765, + "learning_rate": 4.016166253421023e-05, + "loss": 0.0408, + "step": 30043 + }, + { + "epoch": 0.7042786734412818, + "grad_norm": 0.4995422661304474, + "learning_rate": 4.015576085782613e-05, + "loss": 0.1026, + "step": 30044 + }, + { + "epoch": 0.7043021150160867, + "grad_norm": 0.3547271192073822, + "learning_rate": 4.014985950615858e-05, + "loss": 0.0426, + "step": 30045 + }, + { + "epoch": 0.7043255565908918, + "grad_norm": 0.1873210221529007, + "learning_rate": 4.014395847923955e-05, + "loss": 0.0359, + "step": 30046 + }, + { + "epoch": 0.7043489981656967, + "grad_norm": 0.22035957872867584, + "learning_rate": 4.013805777710111e-05, + "loss": 0.0283, + "step": 30047 + }, + { + "epoch": 0.7043724397405018, + "grad_norm": 0.4775814414024353, + "learning_rate": 4.013215739977526e-05, + "loss": 0.1184, + "step": 30048 + }, + { + "epoch": 0.7043958813153067, + "grad_norm": 0.08423707634210587, + "learning_rate": 4.0126257347294e-05, + "loss": 0.0067, + "step": 30049 + }, + { + "epoch": 0.7044193228901118, + "grad_norm": 0.23380115628242493, + "learning_rate": 4.0120357619689364e-05, + "loss": 0.0193, + "step": 30050 + }, + { + "epoch": 0.7044427644649167, + "grad_norm": 0.26419156789779663, + "learning_rate": 4.01144582169933e-05, + "loss": 0.0296, + "step": 30051 + }, + { + "epoch": 0.7044662060397218, + "grad_norm": 0.8640868067741394, + "learning_rate": 4.010855913923791e-05, + "loss": 0.1137, + "step": 30052 + }, + { + "epoch": 0.7044896476145267, + "grad_norm": 0.32826054096221924, + "learning_rate": 4.010266038645517e-05, + "loss": 0.0303, + "step": 30053 + }, + { + "epoch": 0.7045130891893318, + "grad_norm": 0.1804044544696808, + "learning_rate": 4.0096761958677065e-05, + "loss": 0.0129, + "step": 30054 + }, + { + "epoch": 0.7045365307641367, + "grad_norm": 0.3138252794742584, + "learning_rate": 4.009086385593562e-05, + "loss": 0.2538, + "step": 30055 + }, + { + "epoch": 0.7045599723389417, + "grad_norm": 0.36816996335983276, + "learning_rate": 4.008496607826281e-05, + "loss": 0.0401, + "step": 30056 + }, + { + "epoch": 0.7045834139137467, + "grad_norm": 0.19913743436336517, + "learning_rate": 4.0079068625690675e-05, + "loss": 0.0159, + "step": 30057 + }, + { + "epoch": 0.7046068554885517, + "grad_norm": 1.0448100566864014, + "learning_rate": 4.0073171498251175e-05, + "loss": 0.0633, + "step": 30058 + }, + { + "epoch": 0.7046302970633567, + "grad_norm": 0.1407671570777893, + "learning_rate": 4.006727469597636e-05, + "loss": 0.024, + "step": 30059 + }, + { + "epoch": 0.7046537386381617, + "grad_norm": 0.2017335593700409, + "learning_rate": 4.0061378218898205e-05, + "loss": 0.024, + "step": 30060 + }, + { + "epoch": 0.7046771802129667, + "grad_norm": 0.5310574173927307, + "learning_rate": 4.005548206704869e-05, + "loss": 0.091, + "step": 30061 + }, + { + "epoch": 0.7047006217877717, + "grad_norm": 0.21796618402004242, + "learning_rate": 4.004958624045981e-05, + "loss": 0.0357, + "step": 30062 + }, + { + "epoch": 0.7047240633625766, + "grad_norm": 0.4876863956451416, + "learning_rate": 4.004369073916354e-05, + "loss": 0.0684, + "step": 30063 + }, + { + "epoch": 0.7047475049373817, + "grad_norm": 0.17703519761562347, + "learning_rate": 4.003779556319191e-05, + "loss": 0.0277, + "step": 30064 + }, + { + "epoch": 0.7047709465121866, + "grad_norm": 0.23813877999782562, + "learning_rate": 4.0031900712576894e-05, + "loss": 0.0501, + "step": 30065 + }, + { + "epoch": 0.7047943880869917, + "grad_norm": 0.2087380588054657, + "learning_rate": 4.002600618735047e-05, + "loss": 0.0485, + "step": 30066 + }, + { + "epoch": 0.7048178296617966, + "grad_norm": 0.15208809077739716, + "learning_rate": 4.002011198754463e-05, + "loss": 0.0362, + "step": 30067 + }, + { + "epoch": 0.7048412712366017, + "grad_norm": 0.3701154589653015, + "learning_rate": 4.001421811319135e-05, + "loss": 0.0458, + "step": 30068 + }, + { + "epoch": 0.7048647128114067, + "grad_norm": 0.28558167815208435, + "learning_rate": 4.000832456432256e-05, + "loss": 0.0283, + "step": 30069 + }, + { + "epoch": 0.7048881543862117, + "grad_norm": 0.34367257356643677, + "learning_rate": 4.00024313409703e-05, + "loss": 0.0453, + "step": 30070 + }, + { + "epoch": 0.7049115959610167, + "grad_norm": 0.1703110933303833, + "learning_rate": 3.999653844316657e-05, + "loss": 0.0347, + "step": 30071 + }, + { + "epoch": 0.7049350375358217, + "grad_norm": 0.17145167291164398, + "learning_rate": 3.99906458709433e-05, + "loss": 0.0307, + "step": 30072 + }, + { + "epoch": 0.7049584791106267, + "grad_norm": 0.7200592160224915, + "learning_rate": 3.998475362433247e-05, + "loss": 0.0457, + "step": 30073 + }, + { + "epoch": 0.7049819206854316, + "grad_norm": 0.23388423025608063, + "learning_rate": 3.997886170336607e-05, + "loss": 0.0147, + "step": 30074 + }, + { + "epoch": 0.7050053622602367, + "grad_norm": 0.49116677045822144, + "learning_rate": 3.997297010807605e-05, + "loss": 0.0852, + "step": 30075 + }, + { + "epoch": 0.7050288038350416, + "grad_norm": 0.3600509464740753, + "learning_rate": 3.996707883849433e-05, + "loss": 0.0597, + "step": 30076 + }, + { + "epoch": 0.7050522454098467, + "grad_norm": 0.05584977939724922, + "learning_rate": 3.9961187894652974e-05, + "loss": 0.0118, + "step": 30077 + }, + { + "epoch": 0.7050756869846516, + "grad_norm": 0.4098352789878845, + "learning_rate": 3.99552972765839e-05, + "loss": 0.0554, + "step": 30078 + }, + { + "epoch": 0.7050991285594567, + "grad_norm": 0.7345855236053467, + "learning_rate": 3.994940698431908e-05, + "loss": 0.1274, + "step": 30079 + }, + { + "epoch": 0.7051225701342616, + "grad_norm": 0.5110424160957336, + "learning_rate": 3.994351701789046e-05, + "loss": 0.0721, + "step": 30080 + }, + { + "epoch": 0.7051460117090667, + "grad_norm": 0.17820143699645996, + "learning_rate": 3.993762737733001e-05, + "loss": 0.044, + "step": 30081 + }, + { + "epoch": 0.7051694532838716, + "grad_norm": 0.3291499614715576, + "learning_rate": 3.993173806266964e-05, + "loss": 0.0483, + "step": 30082 + }, + { + "epoch": 0.7051928948586766, + "grad_norm": 0.5215846300125122, + "learning_rate": 3.9925849073941345e-05, + "loss": 0.6043, + "step": 30083 + }, + { + "epoch": 0.7052163364334816, + "grad_norm": 0.3294014632701874, + "learning_rate": 3.991996041117713e-05, + "loss": 0.061, + "step": 30084 + }, + { + "epoch": 0.7052397780082866, + "grad_norm": 0.27176690101623535, + "learning_rate": 3.991407207440889e-05, + "loss": 0.0246, + "step": 30085 + }, + { + "epoch": 0.7052632195830916, + "grad_norm": 0.36189112067222595, + "learning_rate": 3.990818406366859e-05, + "loss": 0.0691, + "step": 30086 + }, + { + "epoch": 0.7052866611578966, + "grad_norm": 0.531258225440979, + "learning_rate": 3.990229637898817e-05, + "loss": 0.0844, + "step": 30087 + }, + { + "epoch": 0.7053101027327016, + "grad_norm": 0.44244953989982605, + "learning_rate": 3.989640902039957e-05, + "loss": 0.0926, + "step": 30088 + }, + { + "epoch": 0.7053335443075066, + "grad_norm": 0.49328726530075073, + "learning_rate": 3.989052198793473e-05, + "loss": 0.0689, + "step": 30089 + }, + { + "epoch": 0.7053569858823115, + "grad_norm": 0.20115667581558228, + "learning_rate": 3.9884635281625624e-05, + "loss": 0.04, + "step": 30090 + }, + { + "epoch": 0.7053804274571166, + "grad_norm": 0.25211793184280396, + "learning_rate": 3.987874890150418e-05, + "loss": 0.0533, + "step": 30091 + }, + { + "epoch": 0.7054038690319215, + "grad_norm": 0.4176565706729889, + "learning_rate": 3.987286284760234e-05, + "loss": 0.4317, + "step": 30092 + }, + { + "epoch": 0.7054273106067266, + "grad_norm": 0.17085789144039154, + "learning_rate": 3.986697711995204e-05, + "loss": 0.0574, + "step": 30093 + }, + { + "epoch": 0.7054507521815315, + "grad_norm": 0.1366024762392044, + "learning_rate": 3.986109171858518e-05, + "loss": 0.0183, + "step": 30094 + }, + { + "epoch": 0.7054741937563366, + "grad_norm": 0.6979473233222961, + "learning_rate": 3.985520664353375e-05, + "loss": 0.1291, + "step": 30095 + }, + { + "epoch": 0.7054976353311415, + "grad_norm": 0.452656626701355, + "learning_rate": 3.9849321894829636e-05, + "loss": 0.1096, + "step": 30096 + }, + { + "epoch": 0.7055210769059466, + "grad_norm": 0.5671362280845642, + "learning_rate": 3.984343747250482e-05, + "loss": 0.0493, + "step": 30097 + }, + { + "epoch": 0.7055445184807515, + "grad_norm": 0.3277110457420349, + "learning_rate": 3.9837553376591205e-05, + "loss": 0.0467, + "step": 30098 + }, + { + "epoch": 0.7055679600555566, + "grad_norm": 0.36539801955223083, + "learning_rate": 3.9831669607120726e-05, + "loss": 0.0403, + "step": 30099 + }, + { + "epoch": 0.7055914016303615, + "grad_norm": 0.1577654778957367, + "learning_rate": 3.982578616412529e-05, + "loss": 0.0247, + "step": 30100 + }, + { + "epoch": 0.7056148432051665, + "grad_norm": 1.2522883415222168, + "learning_rate": 3.98199030476368e-05, + "loss": 0.1998, + "step": 30101 + }, + { + "epoch": 0.7056382847799715, + "grad_norm": 0.4888036251068115, + "learning_rate": 3.981402025768723e-05, + "loss": 0.4235, + "step": 30102 + }, + { + "epoch": 0.7056617263547765, + "grad_norm": 0.15293416380882263, + "learning_rate": 3.98081377943085e-05, + "loss": 0.0245, + "step": 30103 + }, + { + "epoch": 0.7056851679295815, + "grad_norm": 0.4719367027282715, + "learning_rate": 3.9802255657532494e-05, + "loss": 0.131, + "step": 30104 + }, + { + "epoch": 0.7057086095043865, + "grad_norm": 0.20556746423244476, + "learning_rate": 3.9796373847391144e-05, + "loss": 0.0242, + "step": 30105 + }, + { + "epoch": 0.7057320510791915, + "grad_norm": 0.5311018228530884, + "learning_rate": 3.9790492363916365e-05, + "loss": 0.5511, + "step": 30106 + }, + { + "epoch": 0.7057554926539965, + "grad_norm": 0.15706130862236023, + "learning_rate": 3.978461120714002e-05, + "loss": 0.0233, + "step": 30107 + }, + { + "epoch": 0.7057789342288014, + "grad_norm": 0.4844011068344116, + "learning_rate": 3.9778730377094075e-05, + "loss": 0.0581, + "step": 30108 + }, + { + "epoch": 0.7058023758036065, + "grad_norm": 0.22533437609672546, + "learning_rate": 3.977284987381047e-05, + "loss": 0.0509, + "step": 30109 + }, + { + "epoch": 0.7058258173784114, + "grad_norm": 0.739559531211853, + "learning_rate": 3.9766969697321076e-05, + "loss": 0.1055, + "step": 30110 + }, + { + "epoch": 0.7058492589532165, + "grad_norm": 0.4041120707988739, + "learning_rate": 3.97610898476578e-05, + "loss": 0.0646, + "step": 30111 + }, + { + "epoch": 0.7058727005280214, + "grad_norm": 0.37070077657699585, + "learning_rate": 3.975521032485254e-05, + "loss": 0.0443, + "step": 30112 + }, + { + "epoch": 0.7058961421028265, + "grad_norm": 0.3129575252532959, + "learning_rate": 3.9749331128937206e-05, + "loss": 0.0363, + "step": 30113 + }, + { + "epoch": 0.7059195836776314, + "grad_norm": 0.5655437111854553, + "learning_rate": 3.9743452259943656e-05, + "loss": 0.0776, + "step": 30114 + }, + { + "epoch": 0.7059430252524365, + "grad_norm": 0.6671190857887268, + "learning_rate": 3.973757371790386e-05, + "loss": 0.1994, + "step": 30115 + }, + { + "epoch": 0.7059664668272414, + "grad_norm": 0.503105103969574, + "learning_rate": 3.9731695502849685e-05, + "loss": 0.114, + "step": 30116 + }, + { + "epoch": 0.7059899084020465, + "grad_norm": 0.5959482192993164, + "learning_rate": 3.9725817614813035e-05, + "loss": 0.1478, + "step": 30117 + }, + { + "epoch": 0.7060133499768514, + "grad_norm": 0.1581960767507553, + "learning_rate": 3.971994005382579e-05, + "loss": 0.0229, + "step": 30118 + }, + { + "epoch": 0.7060367915516564, + "grad_norm": 0.6448829174041748, + "learning_rate": 3.9714062819919805e-05, + "loss": 0.0834, + "step": 30119 + }, + { + "epoch": 0.7060602331264615, + "grad_norm": 1.3979604244232178, + "learning_rate": 3.9708185913127036e-05, + "loss": 0.1401, + "step": 30120 + }, + { + "epoch": 0.7060836747012664, + "grad_norm": 0.23840200901031494, + "learning_rate": 3.970230933347932e-05, + "loss": 0.0344, + "step": 30121 + }, + { + "epoch": 0.7061071162760715, + "grad_norm": 0.10680875927209854, + "learning_rate": 3.969643308100861e-05, + "loss": 0.0346, + "step": 30122 + }, + { + "epoch": 0.7061305578508764, + "grad_norm": 0.6068520545959473, + "learning_rate": 3.9690557155746745e-05, + "loss": 0.0939, + "step": 30123 + }, + { + "epoch": 0.7061539994256815, + "grad_norm": 0.3465782105922699, + "learning_rate": 3.968468155772561e-05, + "loss": 0.032, + "step": 30124 + }, + { + "epoch": 0.7061774410004864, + "grad_norm": 0.5325697660446167, + "learning_rate": 3.96788062869771e-05, + "loss": 0.1038, + "step": 30125 + }, + { + "epoch": 0.7062008825752915, + "grad_norm": 0.4256582260131836, + "learning_rate": 3.967293134353307e-05, + "loss": 0.0581, + "step": 30126 + }, + { + "epoch": 0.7062243241500964, + "grad_norm": 0.5530184507369995, + "learning_rate": 3.966705672742538e-05, + "loss": 0.083, + "step": 30127 + }, + { + "epoch": 0.7062477657249014, + "grad_norm": 0.4952695071697235, + "learning_rate": 3.9661182438685976e-05, + "loss": 0.0752, + "step": 30128 + }, + { + "epoch": 0.7062712072997064, + "grad_norm": 0.21114951372146606, + "learning_rate": 3.9655308477346686e-05, + "loss": 0.0185, + "step": 30129 + }, + { + "epoch": 0.7062946488745114, + "grad_norm": 0.384424090385437, + "learning_rate": 3.964943484343939e-05, + "loss": 0.43, + "step": 30130 + }, + { + "epoch": 0.7063180904493164, + "grad_norm": 0.5287460684776306, + "learning_rate": 3.964356153699593e-05, + "loss": 0.0924, + "step": 30131 + }, + { + "epoch": 0.7063415320241214, + "grad_norm": 0.548140823841095, + "learning_rate": 3.9637688558048225e-05, + "loss": 0.0858, + "step": 30132 + }, + { + "epoch": 0.7063649735989264, + "grad_norm": 0.23897632956504822, + "learning_rate": 3.963181590662809e-05, + "loss": 0.019, + "step": 30133 + }, + { + "epoch": 0.7063884151737314, + "grad_norm": 0.49405092000961304, + "learning_rate": 3.962594358276746e-05, + "loss": 0.1287, + "step": 30134 + }, + { + "epoch": 0.7064118567485363, + "grad_norm": 0.15985962748527527, + "learning_rate": 3.962007158649815e-05, + "loss": 0.013, + "step": 30135 + }, + { + "epoch": 0.7064352983233414, + "grad_norm": 0.4291081726551056, + "learning_rate": 3.961419991785203e-05, + "loss": 0.0609, + "step": 30136 + }, + { + "epoch": 0.7064587398981463, + "grad_norm": 0.06594140827655792, + "learning_rate": 3.960832857686096e-05, + "loss": 0.0098, + "step": 30137 + }, + { + "epoch": 0.7064821814729514, + "grad_norm": 0.2700475752353668, + "learning_rate": 3.9602457563556805e-05, + "loss": 0.0771, + "step": 30138 + }, + { + "epoch": 0.7065056230477563, + "grad_norm": 0.6237230896949768, + "learning_rate": 3.959658687797136e-05, + "loss": 0.033, + "step": 30139 + }, + { + "epoch": 0.7065290646225614, + "grad_norm": 0.5075575709342957, + "learning_rate": 3.9590716520136584e-05, + "loss": 0.1016, + "step": 30140 + }, + { + "epoch": 0.7065525061973663, + "grad_norm": 0.42451831698417664, + "learning_rate": 3.958484649008427e-05, + "loss": 0.0654, + "step": 30141 + }, + { + "epoch": 0.7065759477721714, + "grad_norm": 0.43138501048088074, + "learning_rate": 3.957897678784629e-05, + "loss": 0.067, + "step": 30142 + }, + { + "epoch": 0.7065993893469763, + "grad_norm": 0.367923378944397, + "learning_rate": 3.9573107413454435e-05, + "loss": 0.3297, + "step": 30143 + }, + { + "epoch": 0.7066228309217814, + "grad_norm": 0.5815855264663696, + "learning_rate": 3.956723836694064e-05, + "loss": 0.0838, + "step": 30144 + }, + { + "epoch": 0.7066462724965863, + "grad_norm": 0.3978692591190338, + "learning_rate": 3.9561369648336696e-05, + "loss": 0.0789, + "step": 30145 + }, + { + "epoch": 0.7066697140713913, + "grad_norm": 0.4981532394886017, + "learning_rate": 3.955550125767444e-05, + "loss": 0.1109, + "step": 30146 + }, + { + "epoch": 0.7066931556461963, + "grad_norm": 0.3543907105922699, + "learning_rate": 3.954963319498576e-05, + "loss": 0.0668, + "step": 30147 + }, + { + "epoch": 0.7067165972210013, + "grad_norm": 0.30007025599479675, + "learning_rate": 3.9543765460302474e-05, + "loss": 0.0476, + "step": 30148 + }, + { + "epoch": 0.7067400387958063, + "grad_norm": 0.3187255859375, + "learning_rate": 3.953789805365641e-05, + "loss": 0.0738, + "step": 30149 + }, + { + "epoch": 0.7067634803706113, + "grad_norm": 0.28382399678230286, + "learning_rate": 3.9532030975079415e-05, + "loss": 0.0566, + "step": 30150 + }, + { + "epoch": 0.7067869219454163, + "grad_norm": 0.23379746079444885, + "learning_rate": 3.9526164224603325e-05, + "loss": 0.0295, + "step": 30151 + }, + { + "epoch": 0.7068103635202213, + "grad_norm": 0.4027659595012665, + "learning_rate": 3.952029780225993e-05, + "loss": 0.0601, + "step": 30152 + }, + { + "epoch": 0.7068338050950262, + "grad_norm": 0.3524141311645508, + "learning_rate": 3.9514431708081137e-05, + "loss": 0.0479, + "step": 30153 + }, + { + "epoch": 0.7068572466698313, + "grad_norm": 0.4070574641227722, + "learning_rate": 3.950856594209873e-05, + "loss": 0.1348, + "step": 30154 + }, + { + "epoch": 0.7068806882446362, + "grad_norm": 0.7124492526054382, + "learning_rate": 3.9502700504344515e-05, + "loss": 0.1421, + "step": 30155 + }, + { + "epoch": 0.7069041298194413, + "grad_norm": 0.15050099790096283, + "learning_rate": 3.9496835394850385e-05, + "loss": 0.0246, + "step": 30156 + }, + { + "epoch": 0.7069275713942462, + "grad_norm": 0.5407605171203613, + "learning_rate": 3.9490970613648126e-05, + "loss": 0.6127, + "step": 30157 + }, + { + "epoch": 0.7069510129690513, + "grad_norm": 0.4456852674484253, + "learning_rate": 3.948510616076956e-05, + "loss": 0.0775, + "step": 30158 + }, + { + "epoch": 0.7069744545438562, + "grad_norm": 0.25555843114852905, + "learning_rate": 3.947924203624648e-05, + "loss": 0.0578, + "step": 30159 + }, + { + "epoch": 0.7069978961186613, + "grad_norm": 0.8453837633132935, + "learning_rate": 3.947337824011076e-05, + "loss": 0.1127, + "step": 30160 + }, + { + "epoch": 0.7070213376934662, + "grad_norm": 0.6503477096557617, + "learning_rate": 3.94675147723942e-05, + "loss": 0.1406, + "step": 30161 + }, + { + "epoch": 0.7070447792682713, + "grad_norm": 0.11164085566997528, + "learning_rate": 3.94616516331286e-05, + "loss": 0.0121, + "step": 30162 + }, + { + "epoch": 0.7070682208430762, + "grad_norm": 0.4403376281261444, + "learning_rate": 3.945578882234578e-05, + "loss": 0.0455, + "step": 30163 + }, + { + "epoch": 0.7070916624178812, + "grad_norm": 0.6332032084465027, + "learning_rate": 3.9449926340077556e-05, + "loss": 0.5326, + "step": 30164 + }, + { + "epoch": 0.7071151039926862, + "grad_norm": 0.625721275806427, + "learning_rate": 3.944406418635571e-05, + "loss": 0.0803, + "step": 30165 + }, + { + "epoch": 0.7071385455674912, + "grad_norm": 0.23409807682037354, + "learning_rate": 3.9438202361212095e-05, + "loss": 0.0419, + "step": 30166 + }, + { + "epoch": 0.7071619871422962, + "grad_norm": 0.15535393357276917, + "learning_rate": 3.9432340864678495e-05, + "loss": 0.0141, + "step": 30167 + }, + { + "epoch": 0.7071854287171012, + "grad_norm": 0.17686855792999268, + "learning_rate": 3.942647969678669e-05, + "loss": 0.0297, + "step": 30168 + }, + { + "epoch": 0.7072088702919062, + "grad_norm": 0.5921773910522461, + "learning_rate": 3.942061885756855e-05, + "loss": 0.4114, + "step": 30169 + }, + { + "epoch": 0.7072323118667112, + "grad_norm": 0.3315688967704773, + "learning_rate": 3.9414758347055825e-05, + "loss": 0.0588, + "step": 30170 + }, + { + "epoch": 0.7072557534415163, + "grad_norm": 0.33259549736976624, + "learning_rate": 3.9408898165280295e-05, + "loss": 0.0465, + "step": 30171 + }, + { + "epoch": 0.7072791950163212, + "grad_norm": 0.6548675894737244, + "learning_rate": 3.940303831227381e-05, + "loss": 0.0815, + "step": 30172 + }, + { + "epoch": 0.7073026365911262, + "grad_norm": 0.17237836122512817, + "learning_rate": 3.9397178788068154e-05, + "loss": 0.0109, + "step": 30173 + }, + { + "epoch": 0.7073260781659312, + "grad_norm": 0.45046719908714294, + "learning_rate": 3.939131959269511e-05, + "loss": 0.6005, + "step": 30174 + }, + { + "epoch": 0.7073495197407362, + "grad_norm": 0.22509098052978516, + "learning_rate": 3.938546072618646e-05, + "loss": 0.0091, + "step": 30175 + }, + { + "epoch": 0.7073729613155412, + "grad_norm": 0.08224055916070938, + "learning_rate": 3.9379602188574014e-05, + "loss": 0.0177, + "step": 30176 + }, + { + "epoch": 0.7073964028903462, + "grad_norm": 0.10825493931770325, + "learning_rate": 3.937374397988952e-05, + "loss": 0.0102, + "step": 30177 + }, + { + "epoch": 0.7074198444651512, + "grad_norm": 0.12143047899007797, + "learning_rate": 3.936788610016483e-05, + "loss": 0.0284, + "step": 30178 + }, + { + "epoch": 0.7074432860399562, + "grad_norm": 0.2521718144416809, + "learning_rate": 3.9362028549431686e-05, + "loss": 0.0508, + "step": 30179 + }, + { + "epoch": 0.7074667276147611, + "grad_norm": 0.23295281827449799, + "learning_rate": 3.935617132772185e-05, + "loss": 0.0463, + "step": 30180 + }, + { + "epoch": 0.7074901691895662, + "grad_norm": 0.4939238727092743, + "learning_rate": 3.935031443506717e-05, + "loss": 0.109, + "step": 30181 + }, + { + "epoch": 0.7075136107643711, + "grad_norm": 0.2206616848707199, + "learning_rate": 3.9344457871499384e-05, + "loss": 0.049, + "step": 30182 + }, + { + "epoch": 0.7075370523391762, + "grad_norm": 0.3742554783821106, + "learning_rate": 3.9338601637050274e-05, + "loss": 0.0406, + "step": 30183 + }, + { + "epoch": 0.7075604939139811, + "grad_norm": 0.26514923572540283, + "learning_rate": 3.933274573175159e-05, + "loss": 0.0326, + "step": 30184 + }, + { + "epoch": 0.7075839354887862, + "grad_norm": 0.450458824634552, + "learning_rate": 3.9326890155635166e-05, + "loss": 0.0875, + "step": 30185 + }, + { + "epoch": 0.7076073770635911, + "grad_norm": 0.6877342462539673, + "learning_rate": 3.9321034908732736e-05, + "loss": 0.0991, + "step": 30186 + }, + { + "epoch": 0.7076308186383962, + "grad_norm": 0.3237369656562805, + "learning_rate": 3.931517999107608e-05, + "loss": 0.0617, + "step": 30187 + }, + { + "epoch": 0.7076542602132011, + "grad_norm": 0.3093944787979126, + "learning_rate": 3.9309325402696975e-05, + "loss": 0.0444, + "step": 30188 + }, + { + "epoch": 0.7076777017880062, + "grad_norm": 0.7415910363197327, + "learning_rate": 3.930347114362716e-05, + "loss": 0.137, + "step": 30189 + }, + { + "epoch": 0.7077011433628111, + "grad_norm": 0.698699414730072, + "learning_rate": 3.9297617213898385e-05, + "loss": 0.5047, + "step": 30190 + }, + { + "epoch": 0.7077245849376161, + "grad_norm": 0.08369269967079163, + "learning_rate": 3.929176361354249e-05, + "loss": 0.0179, + "step": 30191 + }, + { + "epoch": 0.7077480265124211, + "grad_norm": 0.4424338638782501, + "learning_rate": 3.9285910342591146e-05, + "loss": 0.1041, + "step": 30192 + }, + { + "epoch": 0.7077714680872261, + "grad_norm": 1.1947582960128784, + "learning_rate": 3.928005740107621e-05, + "loss": 0.1244, + "step": 30193 + }, + { + "epoch": 0.7077949096620311, + "grad_norm": 0.4555054306983948, + "learning_rate": 3.927420478902938e-05, + "loss": 0.0882, + "step": 30194 + }, + { + "epoch": 0.7078183512368361, + "grad_norm": 0.33928588032722473, + "learning_rate": 3.926835250648242e-05, + "loss": 0.0552, + "step": 30195 + }, + { + "epoch": 0.7078417928116411, + "grad_norm": 0.21315084397792816, + "learning_rate": 3.926250055346709e-05, + "loss": 0.0591, + "step": 30196 + }, + { + "epoch": 0.7078652343864461, + "grad_norm": 0.37614917755126953, + "learning_rate": 3.92566489300151e-05, + "loss": 0.061, + "step": 30197 + }, + { + "epoch": 0.707888675961251, + "grad_norm": 0.36655503511428833, + "learning_rate": 3.9250797636158287e-05, + "loss": 0.0495, + "step": 30198 + }, + { + "epoch": 0.7079121175360561, + "grad_norm": 0.13189221918582916, + "learning_rate": 3.9244946671928354e-05, + "loss": 0.0229, + "step": 30199 + }, + { + "epoch": 0.707935559110861, + "grad_norm": 0.3498014211654663, + "learning_rate": 3.923909603735704e-05, + "loss": 0.0606, + "step": 30200 + }, + { + "epoch": 0.7079590006856661, + "grad_norm": 0.33027851581573486, + "learning_rate": 3.92332457324761e-05, + "loss": 0.0607, + "step": 30201 + }, + { + "epoch": 0.707982442260471, + "grad_norm": 0.20741163194179535, + "learning_rate": 3.922739575731724e-05, + "loss": 0.0267, + "step": 30202 + }, + { + "epoch": 0.7080058838352761, + "grad_norm": 0.3234676122665405, + "learning_rate": 3.922154611191228e-05, + "loss": 0.0509, + "step": 30203 + }, + { + "epoch": 0.708029325410081, + "grad_norm": 0.5699513554573059, + "learning_rate": 3.9215696796292875e-05, + "loss": 0.1117, + "step": 30204 + }, + { + "epoch": 0.7080527669848861, + "grad_norm": 0.2858223617076874, + "learning_rate": 3.920984781049086e-05, + "loss": 0.0425, + "step": 30205 + }, + { + "epoch": 0.708076208559691, + "grad_norm": 0.5590239763259888, + "learning_rate": 3.9203999154537905e-05, + "loss": 0.1466, + "step": 30206 + }, + { + "epoch": 0.708099650134496, + "grad_norm": 0.36792466044425964, + "learning_rate": 3.919815082846577e-05, + "loss": 0.0732, + "step": 30207 + }, + { + "epoch": 0.708123091709301, + "grad_norm": 0.15501804649829865, + "learning_rate": 3.919230283230618e-05, + "loss": 0.0363, + "step": 30208 + }, + { + "epoch": 0.708146533284106, + "grad_norm": 0.17074401676654816, + "learning_rate": 3.918645516609082e-05, + "loss": 0.0331, + "step": 30209 + }, + { + "epoch": 0.708169974858911, + "grad_norm": 0.09866643697023392, + "learning_rate": 3.91806078298515e-05, + "loss": 0.0101, + "step": 30210 + }, + { + "epoch": 0.708193416433716, + "grad_norm": 0.6376604437828064, + "learning_rate": 3.917476082361991e-05, + "loss": 0.1942, + "step": 30211 + }, + { + "epoch": 0.708216858008521, + "grad_norm": 0.7402248978614807, + "learning_rate": 3.9168914147427794e-05, + "loss": 0.1343, + "step": 30212 + }, + { + "epoch": 0.708240299583326, + "grad_norm": 0.43479031324386597, + "learning_rate": 3.916306780130685e-05, + "loss": 0.097, + "step": 30213 + }, + { + "epoch": 0.708263741158131, + "grad_norm": 0.5791536569595337, + "learning_rate": 3.915722178528881e-05, + "loss": 0.0675, + "step": 30214 + }, + { + "epoch": 0.708287182732936, + "grad_norm": 0.16621585190296173, + "learning_rate": 3.9151376099405354e-05, + "loss": 0.0251, + "step": 30215 + }, + { + "epoch": 0.7083106243077409, + "grad_norm": 0.47031205892562866, + "learning_rate": 3.9145530743688255e-05, + "loss": 0.1201, + "step": 30216 + }, + { + "epoch": 0.708334065882546, + "grad_norm": 0.5367496013641357, + "learning_rate": 3.913968571816925e-05, + "loss": 0.1122, + "step": 30217 + }, + { + "epoch": 0.7083575074573509, + "grad_norm": 0.34842178225517273, + "learning_rate": 3.9133841022880013e-05, + "loss": 0.077, + "step": 30218 + }, + { + "epoch": 0.708380949032156, + "grad_norm": 0.532221257686615, + "learning_rate": 3.912799665785227e-05, + "loss": 0.0757, + "step": 30219 + }, + { + "epoch": 0.7084043906069609, + "grad_norm": 0.6031200885772705, + "learning_rate": 3.912215262311773e-05, + "loss": 0.1147, + "step": 30220 + }, + { + "epoch": 0.708427832181766, + "grad_norm": 0.19588220119476318, + "learning_rate": 3.91163089187081e-05, + "loss": 0.0257, + "step": 30221 + }, + { + "epoch": 0.708451273756571, + "grad_norm": 0.4917605221271515, + "learning_rate": 3.911046554465506e-05, + "loss": 0.121, + "step": 30222 + }, + { + "epoch": 0.708474715331376, + "grad_norm": 0.8011378049850464, + "learning_rate": 3.9104622500990374e-05, + "loss": 0.1022, + "step": 30223 + }, + { + "epoch": 0.708498156906181, + "grad_norm": 0.2147931605577469, + "learning_rate": 3.909877978774572e-05, + "loss": 0.0295, + "step": 30224 + }, + { + "epoch": 0.708521598480986, + "grad_norm": 0.7922327518463135, + "learning_rate": 3.90929374049528e-05, + "loss": 0.1574, + "step": 30225 + }, + { + "epoch": 0.708545040055791, + "grad_norm": 0.3762997090816498, + "learning_rate": 3.908709535264332e-05, + "loss": 0.0683, + "step": 30226 + }, + { + "epoch": 0.7085684816305959, + "grad_norm": 0.4404061734676361, + "learning_rate": 3.908125363084897e-05, + "loss": 0.0496, + "step": 30227 + }, + { + "epoch": 0.708591923205401, + "grad_norm": 0.4074198305606842, + "learning_rate": 3.907541223960142e-05, + "loss": 0.0583, + "step": 30228 + }, + { + "epoch": 0.7086153647802059, + "grad_norm": 0.5031999349594116, + "learning_rate": 3.9069571178932386e-05, + "loss": 0.0873, + "step": 30229 + }, + { + "epoch": 0.708638806355011, + "grad_norm": 0.2870378792285919, + "learning_rate": 3.90637304488736e-05, + "loss": 0.0283, + "step": 30230 + }, + { + "epoch": 0.7086622479298159, + "grad_norm": 0.48462435603141785, + "learning_rate": 3.9057890049456746e-05, + "loss": 0.0971, + "step": 30231 + }, + { + "epoch": 0.708685689504621, + "grad_norm": 0.38590171933174133, + "learning_rate": 3.905204998071348e-05, + "loss": 0.0677, + "step": 30232 + }, + { + "epoch": 0.7087091310794259, + "grad_norm": 0.37877440452575684, + "learning_rate": 3.9046210242675515e-05, + "loss": 0.0719, + "step": 30233 + }, + { + "epoch": 0.708732572654231, + "grad_norm": 0.2805720865726471, + "learning_rate": 3.904037083537452e-05, + "loss": 0.0429, + "step": 30234 + }, + { + "epoch": 0.7087560142290359, + "grad_norm": 0.36072680354118347, + "learning_rate": 3.903453175884215e-05, + "loss": 0.041, + "step": 30235 + }, + { + "epoch": 0.7087794558038409, + "grad_norm": 0.4592061936855316, + "learning_rate": 3.902869301311016e-05, + "loss": 0.041, + "step": 30236 + }, + { + "epoch": 0.7088028973786459, + "grad_norm": 0.43456408381462097, + "learning_rate": 3.902285459821021e-05, + "loss": 0.0582, + "step": 30237 + }, + { + "epoch": 0.7088263389534509, + "grad_norm": 0.18361179530620575, + "learning_rate": 3.9017016514173956e-05, + "loss": 0.0225, + "step": 30238 + }, + { + "epoch": 0.7088497805282559, + "grad_norm": 0.47485312819480896, + "learning_rate": 3.9011178761033086e-05, + "loss": 0.0788, + "step": 30239 + }, + { + "epoch": 0.7088732221030609, + "grad_norm": 0.3632694482803345, + "learning_rate": 3.900534133881923e-05, + "loss": 0.0664, + "step": 30240 + }, + { + "epoch": 0.7088966636778659, + "grad_norm": 0.14184102416038513, + "learning_rate": 3.899950424756416e-05, + "loss": 0.0178, + "step": 30241 + }, + { + "epoch": 0.7089201052526709, + "grad_norm": 0.12193451076745987, + "learning_rate": 3.8993667487299446e-05, + "loss": 0.0268, + "step": 30242 + }, + { + "epoch": 0.7089435468274758, + "grad_norm": 0.8039519786834717, + "learning_rate": 3.898783105805684e-05, + "loss": 0.1634, + "step": 30243 + }, + { + "epoch": 0.7089669884022809, + "grad_norm": 1.0678143501281738, + "learning_rate": 3.8981994959867986e-05, + "loss": 0.1115, + "step": 30244 + }, + { + "epoch": 0.7089904299770858, + "grad_norm": 0.42082953453063965, + "learning_rate": 3.897615919276455e-05, + "loss": 0.0877, + "step": 30245 + }, + { + "epoch": 0.7090138715518909, + "grad_norm": 0.4969789683818817, + "learning_rate": 3.897032375677818e-05, + "loss": 0.3771, + "step": 30246 + }, + { + "epoch": 0.7090373131266958, + "grad_norm": 0.213052898645401, + "learning_rate": 3.8964488651940514e-05, + "loss": 0.027, + "step": 30247 + }, + { + "epoch": 0.7090607547015009, + "grad_norm": 0.13195110857486725, + "learning_rate": 3.895865387828328e-05, + "loss": 0.0178, + "step": 30248 + }, + { + "epoch": 0.7090841962763058, + "grad_norm": 0.45938587188720703, + "learning_rate": 3.8952819435838116e-05, + "loss": 0.5491, + "step": 30249 + }, + { + "epoch": 0.7091076378511109, + "grad_norm": 0.4338461458683014, + "learning_rate": 3.8946985324636664e-05, + "loss": 0.0726, + "step": 30250 + }, + { + "epoch": 0.7091310794259158, + "grad_norm": 0.4822236895561218, + "learning_rate": 3.89411515447106e-05, + "loss": 0.0905, + "step": 30251 + }, + { + "epoch": 0.7091545210007208, + "grad_norm": 0.6070044040679932, + "learning_rate": 3.893531809609155e-05, + "loss": 0.1183, + "step": 30252 + }, + { + "epoch": 0.7091779625755258, + "grad_norm": 0.40193265676498413, + "learning_rate": 3.8929484978811145e-05, + "loss": 0.0455, + "step": 30253 + }, + { + "epoch": 0.7092014041503308, + "grad_norm": 0.27303892374038696, + "learning_rate": 3.892365219290107e-05, + "loss": 0.0511, + "step": 30254 + }, + { + "epoch": 0.7092248457251358, + "grad_norm": 0.22907216846942902, + "learning_rate": 3.8917819738393024e-05, + "loss": 0.0305, + "step": 30255 + }, + { + "epoch": 0.7092482872999408, + "grad_norm": 0.5898997783660889, + "learning_rate": 3.89119876153186e-05, + "loss": 0.1392, + "step": 30256 + }, + { + "epoch": 0.7092717288747458, + "grad_norm": 0.48841097950935364, + "learning_rate": 3.890615582370944e-05, + "loss": 0.1252, + "step": 30257 + }, + { + "epoch": 0.7092951704495508, + "grad_norm": 0.37190836668014526, + "learning_rate": 3.890032436359721e-05, + "loss": 0.0821, + "step": 30258 + }, + { + "epoch": 0.7093186120243558, + "grad_norm": 0.537367582321167, + "learning_rate": 3.889449323501353e-05, + "loss": 0.6786, + "step": 30259 + }, + { + "epoch": 0.7093420535991608, + "grad_norm": 0.15965873003005981, + "learning_rate": 3.888866243799002e-05, + "loss": 0.0196, + "step": 30260 + }, + { + "epoch": 0.7093654951739657, + "grad_norm": 0.3794498145580292, + "learning_rate": 3.888283197255836e-05, + "loss": 0.0533, + "step": 30261 + }, + { + "epoch": 0.7093889367487708, + "grad_norm": 0.4303656220436096, + "learning_rate": 3.8877001838750195e-05, + "loss": 0.0712, + "step": 30262 + }, + { + "epoch": 0.7094123783235757, + "grad_norm": 0.4963603615760803, + "learning_rate": 3.8871172036597134e-05, + "loss": 0.0892, + "step": 30263 + }, + { + "epoch": 0.7094358198983808, + "grad_norm": 0.33427923917770386, + "learning_rate": 3.8865342566130805e-05, + "loss": 0.0709, + "step": 30264 + }, + { + "epoch": 0.7094592614731857, + "grad_norm": 0.24692244827747345, + "learning_rate": 3.885951342738281e-05, + "loss": 0.0475, + "step": 30265 + }, + { + "epoch": 0.7094827030479908, + "grad_norm": 0.4421684443950653, + "learning_rate": 3.8853684620384854e-05, + "loss": 0.0874, + "step": 30266 + }, + { + "epoch": 0.7095061446227957, + "grad_norm": 0.3905905485153198, + "learning_rate": 3.8847856145168495e-05, + "loss": 0.0571, + "step": 30267 + }, + { + "epoch": 0.7095295861976008, + "grad_norm": 0.19986899197101593, + "learning_rate": 3.884202800176541e-05, + "loss": 0.0214, + "step": 30268 + }, + { + "epoch": 0.7095530277724057, + "grad_norm": 0.3080548048019409, + "learning_rate": 3.8836200190207204e-05, + "loss": 0.0415, + "step": 30269 + }, + { + "epoch": 0.7095764693472107, + "grad_norm": 0.18163661658763885, + "learning_rate": 3.88303727105255e-05, + "loss": 0.0266, + "step": 30270 + }, + { + "epoch": 0.7095999109220157, + "grad_norm": 0.4124385714530945, + "learning_rate": 3.8824545562751916e-05, + "loss": 0.1372, + "step": 30271 + }, + { + "epoch": 0.7096233524968207, + "grad_norm": 0.20843800902366638, + "learning_rate": 3.881871874691806e-05, + "loss": 0.0116, + "step": 30272 + }, + { + "epoch": 0.7096467940716258, + "grad_norm": 0.47093063592910767, + "learning_rate": 3.881289226305552e-05, + "loss": 0.0743, + "step": 30273 + }, + { + "epoch": 0.7096702356464307, + "grad_norm": 0.3665149211883545, + "learning_rate": 3.8807066111196e-05, + "loss": 0.0612, + "step": 30274 + }, + { + "epoch": 0.7096936772212358, + "grad_norm": 0.5890291929244995, + "learning_rate": 3.8801240291371046e-05, + "loss": 0.1633, + "step": 30275 + }, + { + "epoch": 0.7097171187960407, + "grad_norm": 0.08377236872911453, + "learning_rate": 3.8795414803612286e-05, + "loss": 0.011, + "step": 30276 + }, + { + "epoch": 0.7097405603708458, + "grad_norm": 0.3207210600376129, + "learning_rate": 3.8789589647951286e-05, + "loss": 0.0463, + "step": 30277 + }, + { + "epoch": 0.7097640019456507, + "grad_norm": 0.24305987358093262, + "learning_rate": 3.878376482441973e-05, + "loss": 0.0596, + "step": 30278 + }, + { + "epoch": 0.7097874435204558, + "grad_norm": 0.4430772066116333, + "learning_rate": 3.877794033304919e-05, + "loss": 0.6932, + "step": 30279 + }, + { + "epoch": 0.7098108850952607, + "grad_norm": 0.5921272039413452, + "learning_rate": 3.877211617387123e-05, + "loss": 0.0978, + "step": 30280 + }, + { + "epoch": 0.7098343266700657, + "grad_norm": 0.25140804052352905, + "learning_rate": 3.876629234691752e-05, + "loss": 0.0255, + "step": 30281 + }, + { + "epoch": 0.7098577682448707, + "grad_norm": 0.14625757932662964, + "learning_rate": 3.8760468852219644e-05, + "loss": 0.038, + "step": 30282 + }, + { + "epoch": 0.7098812098196757, + "grad_norm": 0.4048948287963867, + "learning_rate": 3.875464568980918e-05, + "loss": 0.0536, + "step": 30283 + }, + { + "epoch": 0.7099046513944807, + "grad_norm": 0.5511255860328674, + "learning_rate": 3.8748822859717724e-05, + "loss": 0.0856, + "step": 30284 + }, + { + "epoch": 0.7099280929692857, + "grad_norm": 0.2801097631454468, + "learning_rate": 3.8743000361976845e-05, + "loss": 0.0432, + "step": 30285 + }, + { + "epoch": 0.7099515345440907, + "grad_norm": 0.23306819796562195, + "learning_rate": 3.8737178196618206e-05, + "loss": 0.017, + "step": 30286 + }, + { + "epoch": 0.7099749761188957, + "grad_norm": 0.20891185104846954, + "learning_rate": 3.873135636367335e-05, + "loss": 0.0246, + "step": 30287 + }, + { + "epoch": 0.7099984176937006, + "grad_norm": 0.5169095396995544, + "learning_rate": 3.87255348631739e-05, + "loss": 0.072, + "step": 30288 + }, + { + "epoch": 0.7100218592685057, + "grad_norm": 0.34849002957344055, + "learning_rate": 3.871971369515136e-05, + "loss": 0.0604, + "step": 30289 + }, + { + "epoch": 0.7100453008433106, + "grad_norm": 0.837479829788208, + "learning_rate": 3.871389285963742e-05, + "loss": 0.1101, + "step": 30290 + }, + { + "epoch": 0.7100687424181157, + "grad_norm": 0.6227378249168396, + "learning_rate": 3.8708072356663625e-05, + "loss": 0.1187, + "step": 30291 + }, + { + "epoch": 0.7100921839929206, + "grad_norm": 0.13864438235759735, + "learning_rate": 3.870225218626152e-05, + "loss": 0.0175, + "step": 30292 + }, + { + "epoch": 0.7101156255677257, + "grad_norm": 0.3921884596347809, + "learning_rate": 3.869643234846275e-05, + "loss": 0.0774, + "step": 30293 + }, + { + "epoch": 0.7101390671425306, + "grad_norm": 0.3592425286769867, + "learning_rate": 3.8690612843298865e-05, + "loss": 0.0414, + "step": 30294 + }, + { + "epoch": 0.7101625087173357, + "grad_norm": 0.3924258053302765, + "learning_rate": 3.8684793670801434e-05, + "loss": 0.0788, + "step": 30295 + }, + { + "epoch": 0.7101859502921406, + "grad_norm": 0.5016171336174011, + "learning_rate": 3.867897483100204e-05, + "loss": 0.0487, + "step": 30296 + }, + { + "epoch": 0.7102093918669456, + "grad_norm": 0.6014180183410645, + "learning_rate": 3.867315632393225e-05, + "loss": 0.0924, + "step": 30297 + }, + { + "epoch": 0.7102328334417506, + "grad_norm": 0.506176233291626, + "learning_rate": 3.86673381496236e-05, + "loss": 0.6303, + "step": 30298 + }, + { + "epoch": 0.7102562750165556, + "grad_norm": 0.34123024344444275, + "learning_rate": 3.8661520308107744e-05, + "loss": 0.0245, + "step": 30299 + }, + { + "epoch": 0.7102797165913606, + "grad_norm": 0.26379722356796265, + "learning_rate": 3.865570279941619e-05, + "loss": 0.0329, + "step": 30300 + }, + { + "epoch": 0.7103031581661656, + "grad_norm": 0.48700034618377686, + "learning_rate": 3.8649885623580495e-05, + "loss": 0.0684, + "step": 30301 + }, + { + "epoch": 0.7103265997409706, + "grad_norm": 0.5207278728485107, + "learning_rate": 3.864406878063227e-05, + "loss": 0.0619, + "step": 30302 + }, + { + "epoch": 0.7103500413157756, + "grad_norm": 0.17050786316394806, + "learning_rate": 3.863825227060306e-05, + "loss": 0.0278, + "step": 30303 + }, + { + "epoch": 0.7103734828905806, + "grad_norm": 0.3908897042274475, + "learning_rate": 3.8632436093524416e-05, + "loss": 0.085, + "step": 30304 + }, + { + "epoch": 0.7103969244653856, + "grad_norm": 0.48490044474601746, + "learning_rate": 3.862662024942786e-05, + "loss": 0.0772, + "step": 30305 + }, + { + "epoch": 0.7104203660401905, + "grad_norm": 0.417161226272583, + "learning_rate": 3.8620804738345026e-05, + "loss": 0.0489, + "step": 30306 + }, + { + "epoch": 0.7104438076149956, + "grad_norm": 0.15743117034435272, + "learning_rate": 3.8614989560307435e-05, + "loss": 0.0488, + "step": 30307 + }, + { + "epoch": 0.7104672491898005, + "grad_norm": 0.4772047698497772, + "learning_rate": 3.860917471534663e-05, + "loss": 0.0637, + "step": 30308 + }, + { + "epoch": 0.7104906907646056, + "grad_norm": 0.25783342123031616, + "learning_rate": 3.860336020349418e-05, + "loss": 0.04, + "step": 30309 + }, + { + "epoch": 0.7105141323394105, + "grad_norm": 0.49679723381996155, + "learning_rate": 3.859754602478162e-05, + "loss": 0.0516, + "step": 30310 + }, + { + "epoch": 0.7105375739142156, + "grad_norm": 0.2732165455818176, + "learning_rate": 3.859173217924046e-05, + "loss": 0.0445, + "step": 30311 + }, + { + "epoch": 0.7105610154890205, + "grad_norm": 0.340923547744751, + "learning_rate": 3.858591866690232e-05, + "loss": 0.0415, + "step": 30312 + }, + { + "epoch": 0.7105844570638256, + "grad_norm": 0.3155737519264221, + "learning_rate": 3.858010548779869e-05, + "loss": 0.0518, + "step": 30313 + }, + { + "epoch": 0.7106078986386305, + "grad_norm": 0.4390828311443329, + "learning_rate": 3.857429264196116e-05, + "loss": 0.06, + "step": 30314 + }, + { + "epoch": 0.7106313402134355, + "grad_norm": 0.1264733225107193, + "learning_rate": 3.856848012942125e-05, + "loss": 0.0234, + "step": 30315 + }, + { + "epoch": 0.7106547817882405, + "grad_norm": 0.5403963327407837, + "learning_rate": 3.85626679502105e-05, + "loss": 0.1038, + "step": 30316 + }, + { + "epoch": 0.7106782233630455, + "grad_norm": 0.40711960196495056, + "learning_rate": 3.8556856104360425e-05, + "loss": 0.0755, + "step": 30317 + }, + { + "epoch": 0.7107016649378505, + "grad_norm": 0.13561728596687317, + "learning_rate": 3.855104459190255e-05, + "loss": 0.0133, + "step": 30318 + }, + { + "epoch": 0.7107251065126555, + "grad_norm": 0.31185778975486755, + "learning_rate": 3.854523341286848e-05, + "loss": 0.0706, + "step": 30319 + }, + { + "epoch": 0.7107485480874605, + "grad_norm": 0.3522926867008209, + "learning_rate": 3.8539422567289684e-05, + "loss": 0.0611, + "step": 30320 + }, + { + "epoch": 0.7107719896622655, + "grad_norm": 0.31814226508140564, + "learning_rate": 3.853361205519772e-05, + "loss": 0.2407, + "step": 30321 + }, + { + "epoch": 0.7107954312370705, + "grad_norm": 0.3647376298904419, + "learning_rate": 3.85278018766241e-05, + "loss": 0.046, + "step": 30322 + }, + { + "epoch": 0.7108188728118755, + "grad_norm": 0.3774951100349426, + "learning_rate": 3.852199203160032e-05, + "loss": 0.049, + "step": 30323 + }, + { + "epoch": 0.7108423143866806, + "grad_norm": 0.1342839002609253, + "learning_rate": 3.8516182520157986e-05, + "loss": 0.0201, + "step": 30324 + }, + { + "epoch": 0.7108657559614855, + "grad_norm": 0.5109419226646423, + "learning_rate": 3.851037334232855e-05, + "loss": 0.0902, + "step": 30325 + }, + { + "epoch": 0.7108891975362905, + "grad_norm": 0.33516550064086914, + "learning_rate": 3.8504564498143545e-05, + "loss": 0.0477, + "step": 30326 + }, + { + "epoch": 0.7109126391110955, + "grad_norm": 0.42826682329177856, + "learning_rate": 3.849875598763453e-05, + "loss": 0.0394, + "step": 30327 + }, + { + "epoch": 0.7109360806859005, + "grad_norm": 0.44832974672317505, + "learning_rate": 3.849294781083299e-05, + "loss": 0.0472, + "step": 30328 + }, + { + "epoch": 0.7109595222607055, + "grad_norm": 0.5767010450363159, + "learning_rate": 3.848713996777046e-05, + "loss": 0.1283, + "step": 30329 + }, + { + "epoch": 0.7109829638355105, + "grad_norm": 0.3793241083621979, + "learning_rate": 3.848133245847839e-05, + "loss": 0.0625, + "step": 30330 + }, + { + "epoch": 0.7110064054103155, + "grad_norm": 0.2117769718170166, + "learning_rate": 3.847552528298838e-05, + "loss": 0.0291, + "step": 30331 + }, + { + "epoch": 0.7110298469851205, + "grad_norm": 0.3635149300098419, + "learning_rate": 3.846971844133189e-05, + "loss": 0.0712, + "step": 30332 + }, + { + "epoch": 0.7110532885599254, + "grad_norm": 0.37979820370674133, + "learning_rate": 3.846391193354044e-05, + "loss": 0.0379, + "step": 30333 + }, + { + "epoch": 0.7110767301347305, + "grad_norm": 0.18260663747787476, + "learning_rate": 3.8458105759645545e-05, + "loss": 0.0442, + "step": 30334 + }, + { + "epoch": 0.7111001717095354, + "grad_norm": 0.24733662605285645, + "learning_rate": 3.845229991967869e-05, + "loss": 0.0553, + "step": 30335 + }, + { + "epoch": 0.7111236132843405, + "grad_norm": 0.2817247807979584, + "learning_rate": 3.844649441367135e-05, + "loss": 0.0415, + "step": 30336 + }, + { + "epoch": 0.7111470548591454, + "grad_norm": 0.22704197466373444, + "learning_rate": 3.84406892416551e-05, + "loss": 0.05, + "step": 30337 + }, + { + "epoch": 0.7111704964339505, + "grad_norm": 0.5513148307800293, + "learning_rate": 3.8434884403661374e-05, + "loss": 0.061, + "step": 30338 + }, + { + "epoch": 0.7111939380087554, + "grad_norm": 0.32151126861572266, + "learning_rate": 3.842907989972172e-05, + "loss": 0.0432, + "step": 30339 + }, + { + "epoch": 0.7112173795835605, + "grad_norm": 0.08099156618118286, + "learning_rate": 3.8423275729867616e-05, + "loss": 0.0096, + "step": 30340 + }, + { + "epoch": 0.7112408211583654, + "grad_norm": 0.22288843989372253, + "learning_rate": 3.8417471894130544e-05, + "loss": 0.0293, + "step": 30341 + }, + { + "epoch": 0.7112642627331704, + "grad_norm": 0.5266022086143494, + "learning_rate": 3.841166839254201e-05, + "loss": 0.2555, + "step": 30342 + }, + { + "epoch": 0.7112877043079754, + "grad_norm": 0.2828497886657715, + "learning_rate": 3.840586522513345e-05, + "loss": 0.046, + "step": 30343 + }, + { + "epoch": 0.7113111458827804, + "grad_norm": 0.40410783886909485, + "learning_rate": 3.840006239193643e-05, + "loss": 0.0787, + "step": 30344 + }, + { + "epoch": 0.7113345874575854, + "grad_norm": 0.516676127910614, + "learning_rate": 3.839425989298242e-05, + "loss": 0.4827, + "step": 30345 + }, + { + "epoch": 0.7113580290323904, + "grad_norm": 0.3452019989490509, + "learning_rate": 3.838845772830287e-05, + "loss": 0.038, + "step": 30346 + }, + { + "epoch": 0.7113814706071954, + "grad_norm": 0.5414109230041504, + "learning_rate": 3.8382655897929285e-05, + "loss": 0.1189, + "step": 30347 + }, + { + "epoch": 0.7114049121820004, + "grad_norm": 0.11340803653001785, + "learning_rate": 3.837685440189315e-05, + "loss": 0.0188, + "step": 30348 + }, + { + "epoch": 0.7114283537568054, + "grad_norm": 0.5987166166305542, + "learning_rate": 3.83710532402259e-05, + "loss": 0.1135, + "step": 30349 + }, + { + "epoch": 0.7114517953316104, + "grad_norm": 0.24800308048725128, + "learning_rate": 3.836525241295904e-05, + "loss": 0.0404, + "step": 30350 + }, + { + "epoch": 0.7114752369064153, + "grad_norm": 0.3494308590888977, + "learning_rate": 3.835945192012409e-05, + "loss": 0.0364, + "step": 30351 + }, + { + "epoch": 0.7114986784812204, + "grad_norm": 0.556312620639801, + "learning_rate": 3.8353651761752495e-05, + "loss": 0.118, + "step": 30352 + }, + { + "epoch": 0.7115221200560253, + "grad_norm": 0.591620922088623, + "learning_rate": 3.834785193787571e-05, + "loss": 0.0942, + "step": 30353 + }, + { + "epoch": 0.7115455616308304, + "grad_norm": 0.3365205228328705, + "learning_rate": 3.834205244852521e-05, + "loss": 0.0613, + "step": 30354 + }, + { + "epoch": 0.7115690032056353, + "grad_norm": 0.49433067440986633, + "learning_rate": 3.8336253293732485e-05, + "loss": 0.072, + "step": 30355 + }, + { + "epoch": 0.7115924447804404, + "grad_norm": 0.4935184419155121, + "learning_rate": 3.8330454473528933e-05, + "loss": 0.0635, + "step": 30356 + }, + { + "epoch": 0.7116158863552453, + "grad_norm": 0.09675396978855133, + "learning_rate": 3.8324655987946113e-05, + "loss": 0.0302, + "step": 30357 + }, + { + "epoch": 0.7116393279300504, + "grad_norm": 0.4953470528125763, + "learning_rate": 3.831885783701543e-05, + "loss": 0.0775, + "step": 30358 + }, + { + "epoch": 0.7116627695048553, + "grad_norm": 0.162831649184227, + "learning_rate": 3.8313060020768375e-05, + "loss": 0.0297, + "step": 30359 + }, + { + "epoch": 0.7116862110796603, + "grad_norm": 0.35753166675567627, + "learning_rate": 3.8307262539236374e-05, + "loss": 0.024, + "step": 30360 + }, + { + "epoch": 0.7117096526544653, + "grad_norm": 0.5459455251693726, + "learning_rate": 3.8301465392450886e-05, + "loss": 0.101, + "step": 30361 + }, + { + "epoch": 0.7117330942292703, + "grad_norm": 0.6089855432510376, + "learning_rate": 3.82956685804434e-05, + "loss": 0.1228, + "step": 30362 + }, + { + "epoch": 0.7117565358040753, + "grad_norm": 0.18783922493457794, + "learning_rate": 3.8289872103245325e-05, + "loss": 0.0198, + "step": 30363 + }, + { + "epoch": 0.7117799773788803, + "grad_norm": 0.5044400691986084, + "learning_rate": 3.828407596088817e-05, + "loss": 0.0809, + "step": 30364 + }, + { + "epoch": 0.7118034189536853, + "grad_norm": 0.4558841586112976, + "learning_rate": 3.827828015340335e-05, + "loss": 0.0685, + "step": 30365 + }, + { + "epoch": 0.7118268605284903, + "grad_norm": 0.5616546869277954, + "learning_rate": 3.827248468082233e-05, + "loss": 0.0953, + "step": 30366 + }, + { + "epoch": 0.7118503021032953, + "grad_norm": 0.05958117917180061, + "learning_rate": 3.8266689543176534e-05, + "loss": 0.0039, + "step": 30367 + }, + { + "epoch": 0.7118737436781003, + "grad_norm": 0.30291321873664856, + "learning_rate": 3.826089474049739e-05, + "loss": 0.0266, + "step": 30368 + }, + { + "epoch": 0.7118971852529052, + "grad_norm": 0.4010552763938904, + "learning_rate": 3.825510027281639e-05, + "loss": 0.0596, + "step": 30369 + }, + { + "epoch": 0.7119206268277103, + "grad_norm": 0.24649666249752045, + "learning_rate": 3.824930614016496e-05, + "loss": 0.0623, + "step": 30370 + }, + { + "epoch": 0.7119440684025152, + "grad_norm": 0.10919367522001266, + "learning_rate": 3.824351234257453e-05, + "loss": 0.0155, + "step": 30371 + }, + { + "epoch": 0.7119675099773203, + "grad_norm": 0.07876071333885193, + "learning_rate": 3.823771888007653e-05, + "loss": 0.005, + "step": 30372 + }, + { + "epoch": 0.7119909515521252, + "grad_norm": 0.38860487937927246, + "learning_rate": 3.823192575270242e-05, + "loss": 0.0346, + "step": 30373 + }, + { + "epoch": 0.7120143931269303, + "grad_norm": 0.34186896681785583, + "learning_rate": 3.8226132960483575e-05, + "loss": 0.0507, + "step": 30374 + }, + { + "epoch": 0.7120378347017353, + "grad_norm": 0.31943419575691223, + "learning_rate": 3.822034050345147e-05, + "loss": 0.043, + "step": 30375 + }, + { + "epoch": 0.7120612762765403, + "grad_norm": 0.19543081521987915, + "learning_rate": 3.821454838163757e-05, + "loss": 0.0339, + "step": 30376 + }, + { + "epoch": 0.7120847178513453, + "grad_norm": 0.10598374158143997, + "learning_rate": 3.820875659507327e-05, + "loss": 0.0202, + "step": 30377 + }, + { + "epoch": 0.7121081594261502, + "grad_norm": 0.4987431466579437, + "learning_rate": 3.820296514378998e-05, + "loss": 0.0795, + "step": 30378 + }, + { + "epoch": 0.7121316010009553, + "grad_norm": 0.16337373852729797, + "learning_rate": 3.819717402781915e-05, + "loss": 0.0264, + "step": 30379 + }, + { + "epoch": 0.7121550425757602, + "grad_norm": 0.6714569926261902, + "learning_rate": 3.8191383247192195e-05, + "loss": 0.1401, + "step": 30380 + }, + { + "epoch": 0.7121784841505653, + "grad_norm": 0.30216431617736816, + "learning_rate": 3.81855928019405e-05, + "loss": 0.0331, + "step": 30381 + }, + { + "epoch": 0.7122019257253702, + "grad_norm": 0.37866273522377014, + "learning_rate": 3.817980269209553e-05, + "loss": 0.0488, + "step": 30382 + }, + { + "epoch": 0.7122253673001753, + "grad_norm": 0.3748489320278168, + "learning_rate": 3.8174012917688706e-05, + "loss": 0.3983, + "step": 30383 + }, + { + "epoch": 0.7122488088749802, + "grad_norm": 0.4603804647922516, + "learning_rate": 3.816822347875143e-05, + "loss": 0.0852, + "step": 30384 + }, + { + "epoch": 0.7122722504497853, + "grad_norm": 0.3333405554294586, + "learning_rate": 3.816243437531511e-05, + "loss": 0.0536, + "step": 30385 + }, + { + "epoch": 0.7122956920245902, + "grad_norm": 0.7136861681938171, + "learning_rate": 3.815664560741111e-05, + "loss": 0.8523, + "step": 30386 + }, + { + "epoch": 0.7123191335993952, + "grad_norm": 0.375093936920166, + "learning_rate": 3.815085717507094e-05, + "loss": 0.0385, + "step": 30387 + }, + { + "epoch": 0.7123425751742002, + "grad_norm": 0.34664902091026306, + "learning_rate": 3.8145069078325926e-05, + "loss": 0.3828, + "step": 30388 + }, + { + "epoch": 0.7123660167490052, + "grad_norm": 0.21709957718849182, + "learning_rate": 3.8139281317207534e-05, + "loss": 0.048, + "step": 30389 + }, + { + "epoch": 0.7123894583238102, + "grad_norm": 0.06869346648454666, + "learning_rate": 3.813349389174713e-05, + "loss": 0.005, + "step": 30390 + }, + { + "epoch": 0.7124128998986152, + "grad_norm": 0.5926958322525024, + "learning_rate": 3.8127706801976135e-05, + "loss": 0.4921, + "step": 30391 + }, + { + "epoch": 0.7124363414734202, + "grad_norm": 0.47352755069732666, + "learning_rate": 3.812192004792595e-05, + "loss": 0.1312, + "step": 30392 + }, + { + "epoch": 0.7124597830482252, + "grad_norm": 0.675929069519043, + "learning_rate": 3.811613362962796e-05, + "loss": 0.2053, + "step": 30393 + }, + { + "epoch": 0.7124832246230302, + "grad_norm": 0.08395028114318848, + "learning_rate": 3.8110347547113545e-05, + "loss": 0.0121, + "step": 30394 + }, + { + "epoch": 0.7125066661978352, + "grad_norm": 0.4487432837486267, + "learning_rate": 3.810456180041415e-05, + "loss": 0.0666, + "step": 30395 + }, + { + "epoch": 0.7125301077726401, + "grad_norm": 0.3031884729862213, + "learning_rate": 3.809877638956114e-05, + "loss": 0.3435, + "step": 30396 + }, + { + "epoch": 0.7125535493474452, + "grad_norm": 0.3210521340370178, + "learning_rate": 3.8092991314585914e-05, + "loss": 0.0534, + "step": 30397 + }, + { + "epoch": 0.7125769909222501, + "grad_norm": 0.5151185393333435, + "learning_rate": 3.808720657551986e-05, + "loss": 0.0943, + "step": 30398 + }, + { + "epoch": 0.7126004324970552, + "grad_norm": 0.7436886429786682, + "learning_rate": 3.808142217239432e-05, + "loss": 0.1152, + "step": 30399 + }, + { + "epoch": 0.7126238740718601, + "grad_norm": 0.28660404682159424, + "learning_rate": 3.8075638105240777e-05, + "loss": 0.0459, + "step": 30400 + }, + { + "epoch": 0.7126473156466652, + "grad_norm": 0.3656728267669678, + "learning_rate": 3.806985437409051e-05, + "loss": 0.0538, + "step": 30401 + }, + { + "epoch": 0.7126707572214701, + "grad_norm": 0.4372439980506897, + "learning_rate": 3.8064070978975e-05, + "loss": 0.0864, + "step": 30402 + }, + { + "epoch": 0.7126941987962752, + "grad_norm": 0.11034388840198517, + "learning_rate": 3.805828791992557e-05, + "loss": 0.0154, + "step": 30403 + }, + { + "epoch": 0.7127176403710801, + "grad_norm": 0.3597986698150635, + "learning_rate": 3.805250519697362e-05, + "loss": 0.0479, + "step": 30404 + }, + { + "epoch": 0.7127410819458851, + "grad_norm": 0.6707448363304138, + "learning_rate": 3.804672281015053e-05, + "loss": 0.0451, + "step": 30405 + }, + { + "epoch": 0.7127645235206901, + "grad_norm": 0.5697504281997681, + "learning_rate": 3.8040940759487606e-05, + "loss": 0.1012, + "step": 30406 + }, + { + "epoch": 0.7127879650954951, + "grad_norm": 0.44493812322616577, + "learning_rate": 3.803515904501633e-05, + "loss": 0.0903, + "step": 30407 + }, + { + "epoch": 0.7128114066703001, + "grad_norm": 0.3633723258972168, + "learning_rate": 3.8029377666768015e-05, + "loss": 0.035, + "step": 30408 + }, + { + "epoch": 0.7128348482451051, + "grad_norm": 0.9618049263954163, + "learning_rate": 3.802359662477405e-05, + "loss": 0.1585, + "step": 30409 + }, + { + "epoch": 0.7128582898199101, + "grad_norm": 0.2571277916431427, + "learning_rate": 3.801781591906578e-05, + "loss": 0.0285, + "step": 30410 + }, + { + "epoch": 0.7128817313947151, + "grad_norm": 0.6925053596496582, + "learning_rate": 3.801203554967456e-05, + "loss": 0.1053, + "step": 30411 + }, + { + "epoch": 0.71290517296952, + "grad_norm": 0.12285669147968292, + "learning_rate": 3.800625551663181e-05, + "loss": 0.0178, + "step": 30412 + }, + { + "epoch": 0.7129286145443251, + "grad_norm": 0.35010525584220886, + "learning_rate": 3.8000475819968816e-05, + "loss": 0.0455, + "step": 30413 + }, + { + "epoch": 0.71295205611913, + "grad_norm": 0.20027564465999603, + "learning_rate": 3.799469645971703e-05, + "loss": 0.0301, + "step": 30414 + }, + { + "epoch": 0.7129754976939351, + "grad_norm": 0.2411433458328247, + "learning_rate": 3.798891743590776e-05, + "loss": 0.0354, + "step": 30415 + }, + { + "epoch": 0.71299893926874, + "grad_norm": 0.3597937822341919, + "learning_rate": 3.798313874857237e-05, + "loss": 0.0647, + "step": 30416 + }, + { + "epoch": 0.7130223808435451, + "grad_norm": 0.4153986871242523, + "learning_rate": 3.7977360397742214e-05, + "loss": 0.1059, + "step": 30417 + }, + { + "epoch": 0.71304582241835, + "grad_norm": 0.43485766649246216, + "learning_rate": 3.7971582383448635e-05, + "loss": 0.083, + "step": 30418 + }, + { + "epoch": 0.7130692639931551, + "grad_norm": 0.20941762626171112, + "learning_rate": 3.796580470572296e-05, + "loss": 0.0407, + "step": 30419 + }, + { + "epoch": 0.71309270556796, + "grad_norm": 0.4104501008987427, + "learning_rate": 3.796002736459662e-05, + "loss": 0.0261, + "step": 30420 + }, + { + "epoch": 0.7131161471427651, + "grad_norm": 0.2370707243680954, + "learning_rate": 3.7954250360100906e-05, + "loss": 0.0522, + "step": 30421 + }, + { + "epoch": 0.71313958871757, + "grad_norm": 0.44040095806121826, + "learning_rate": 3.794847369226717e-05, + "loss": 0.4903, + "step": 30422 + }, + { + "epoch": 0.713163030292375, + "grad_norm": 0.4098712205886841, + "learning_rate": 3.794269736112675e-05, + "loss": 0.0666, + "step": 30423 + }, + { + "epoch": 0.71318647186718, + "grad_norm": 0.09281289577484131, + "learning_rate": 3.793692136671101e-05, + "loss": 0.0116, + "step": 30424 + }, + { + "epoch": 0.713209913441985, + "grad_norm": 0.4267388880252838, + "learning_rate": 3.793114570905128e-05, + "loss": 0.0446, + "step": 30425 + }, + { + "epoch": 0.7132333550167901, + "grad_norm": 0.5120425820350647, + "learning_rate": 3.792537038817887e-05, + "loss": 0.1021, + "step": 30426 + }, + { + "epoch": 0.713256796591595, + "grad_norm": 0.47327566146850586, + "learning_rate": 3.791959540412519e-05, + "loss": 0.0952, + "step": 30427 + }, + { + "epoch": 0.7132802381664001, + "grad_norm": 0.24980439245700836, + "learning_rate": 3.7913820756921514e-05, + "loss": 0.0711, + "step": 30428 + }, + { + "epoch": 0.713303679741205, + "grad_norm": 0.38122832775115967, + "learning_rate": 3.79080464465992e-05, + "loss": 0.0789, + "step": 30429 + }, + { + "epoch": 0.7133271213160101, + "grad_norm": 0.1824866533279419, + "learning_rate": 3.790227247318957e-05, + "loss": 0.0152, + "step": 30430 + }, + { + "epoch": 0.713350562890815, + "grad_norm": 0.44786664843559265, + "learning_rate": 3.789649883672396e-05, + "loss": 0.1181, + "step": 30431 + }, + { + "epoch": 0.71337400446562, + "grad_norm": 0.6226620078086853, + "learning_rate": 3.789072553723365e-05, + "loss": 0.1429, + "step": 30432 + }, + { + "epoch": 0.713397446040425, + "grad_norm": 0.35929074883461, + "learning_rate": 3.7884952574750055e-05, + "loss": 0.0443, + "step": 30433 + }, + { + "epoch": 0.71342088761523, + "grad_norm": 0.10294108837842941, + "learning_rate": 3.787917994930444e-05, + "loss": 0.0163, + "step": 30434 + }, + { + "epoch": 0.713444329190035, + "grad_norm": 0.0872291550040245, + "learning_rate": 3.787340766092813e-05, + "loss": 0.02, + "step": 30435 + }, + { + "epoch": 0.71346777076484, + "grad_norm": 0.37179800868034363, + "learning_rate": 3.786763570965247e-05, + "loss": 0.0362, + "step": 30436 + }, + { + "epoch": 0.713491212339645, + "grad_norm": 0.49383974075317383, + "learning_rate": 3.786186409550877e-05, + "loss": 0.054, + "step": 30437 + }, + { + "epoch": 0.71351465391445, + "grad_norm": 0.543498694896698, + "learning_rate": 3.785609281852831e-05, + "loss": 0.5568, + "step": 30438 + }, + { + "epoch": 0.713538095489255, + "grad_norm": 0.23778881132602692, + "learning_rate": 3.7850321878742476e-05, + "loss": 0.0276, + "step": 30439 + }, + { + "epoch": 0.71356153706406, + "grad_norm": 0.5887418985366821, + "learning_rate": 3.784455127618254e-05, + "loss": 0.0713, + "step": 30440 + }, + { + "epoch": 0.7135849786388649, + "grad_norm": 0.3376191556453705, + "learning_rate": 3.783878101087982e-05, + "loss": 0.0335, + "step": 30441 + }, + { + "epoch": 0.71360842021367, + "grad_norm": 0.06577702611684799, + "learning_rate": 3.7833011082865625e-05, + "loss": 0.0068, + "step": 30442 + }, + { + "epoch": 0.7136318617884749, + "grad_norm": 0.5362644791603088, + "learning_rate": 3.7827241492171253e-05, + "loss": 0.0904, + "step": 30443 + }, + { + "epoch": 0.71365530336328, + "grad_norm": 0.45401832461357117, + "learning_rate": 3.7821472238827974e-05, + "loss": 0.0824, + "step": 30444 + }, + { + "epoch": 0.7136787449380849, + "grad_norm": 0.2313794642686844, + "learning_rate": 3.781570332286718e-05, + "loss": 0.0412, + "step": 30445 + }, + { + "epoch": 0.71370218651289, + "grad_norm": 1.2069467306137085, + "learning_rate": 3.780993474432013e-05, + "loss": 0.0851, + "step": 30446 + }, + { + "epoch": 0.7137256280876949, + "grad_norm": 0.08653680980205536, + "learning_rate": 3.7804166503218086e-05, + "loss": 0.0133, + "step": 30447 + }, + { + "epoch": 0.7137490696625, + "grad_norm": 0.5266069769859314, + "learning_rate": 3.779839859959241e-05, + "loss": 0.5965, + "step": 30448 + }, + { + "epoch": 0.7137725112373049, + "grad_norm": 0.6574520468711853, + "learning_rate": 3.779263103347438e-05, + "loss": 0.1372, + "step": 30449 + }, + { + "epoch": 0.71379595281211, + "grad_norm": 0.5351167917251587, + "learning_rate": 3.7786863804895275e-05, + "loss": 0.0616, + "step": 30450 + }, + { + "epoch": 0.7138193943869149, + "grad_norm": 0.5101478695869446, + "learning_rate": 3.778109691388637e-05, + "loss": 0.0633, + "step": 30451 + }, + { + "epoch": 0.7138428359617199, + "grad_norm": 0.634668231010437, + "learning_rate": 3.777533036047902e-05, + "loss": 0.0961, + "step": 30452 + }, + { + "epoch": 0.7138662775365249, + "grad_norm": 0.22853723168373108, + "learning_rate": 3.776956414470447e-05, + "loss": 0.0453, + "step": 30453 + }, + { + "epoch": 0.7138897191113299, + "grad_norm": 0.7007309794425964, + "learning_rate": 3.776379826659402e-05, + "loss": 0.1359, + "step": 30454 + }, + { + "epoch": 0.7139131606861349, + "grad_norm": 0.3566116690635681, + "learning_rate": 3.775803272617895e-05, + "loss": 0.0728, + "step": 30455 + }, + { + "epoch": 0.7139366022609399, + "grad_norm": 0.3088283836841583, + "learning_rate": 3.7752267523490537e-05, + "loss": 0.348, + "step": 30456 + }, + { + "epoch": 0.7139600438357449, + "grad_norm": 0.797776460647583, + "learning_rate": 3.774650265856005e-05, + "loss": 0.1192, + "step": 30457 + }, + { + "epoch": 0.7139834854105499, + "grad_norm": 0.3141334056854248, + "learning_rate": 3.7740738131418815e-05, + "loss": 0.0543, + "step": 30458 + }, + { + "epoch": 0.7140069269853548, + "grad_norm": 0.4061136841773987, + "learning_rate": 3.7734973942098053e-05, + "loss": 0.086, + "step": 30459 + }, + { + "epoch": 0.7140303685601599, + "grad_norm": 0.08775409311056137, + "learning_rate": 3.772921009062911e-05, + "loss": 0.0095, + "step": 30460 + }, + { + "epoch": 0.7140538101349648, + "grad_norm": 0.135850191116333, + "learning_rate": 3.7723446577043206e-05, + "loss": 0.02, + "step": 30461 + }, + { + "epoch": 0.7140772517097699, + "grad_norm": 0.4594215452671051, + "learning_rate": 3.771768340137165e-05, + "loss": 0.0774, + "step": 30462 + }, + { + "epoch": 0.7141006932845748, + "grad_norm": 0.7105091214179993, + "learning_rate": 3.7711920563645696e-05, + "loss": 0.1183, + "step": 30463 + }, + { + "epoch": 0.7141241348593799, + "grad_norm": 0.13633985817432404, + "learning_rate": 3.770615806389657e-05, + "loss": 0.0184, + "step": 30464 + }, + { + "epoch": 0.7141475764341848, + "grad_norm": 0.6181209087371826, + "learning_rate": 3.7700395902155624e-05, + "loss": 0.1413, + "step": 30465 + }, + { + "epoch": 0.7141710180089899, + "grad_norm": 0.31425201892852783, + "learning_rate": 3.769463407845407e-05, + "loss": 0.0421, + "step": 30466 + }, + { + "epoch": 0.7141944595837948, + "grad_norm": 0.4581905007362366, + "learning_rate": 3.76888725928232e-05, + "loss": 0.0974, + "step": 30467 + }, + { + "epoch": 0.7142179011585998, + "grad_norm": 0.4365830421447754, + "learning_rate": 3.768311144529425e-05, + "loss": 0.0892, + "step": 30468 + }, + { + "epoch": 0.7142413427334048, + "grad_norm": 0.4202549457550049, + "learning_rate": 3.767735063589849e-05, + "loss": 0.0782, + "step": 30469 + }, + { + "epoch": 0.7142647843082098, + "grad_norm": 0.09584157913923264, + "learning_rate": 3.7671590164667135e-05, + "loss": 0.0182, + "step": 30470 + }, + { + "epoch": 0.7142882258830148, + "grad_norm": 0.2489253133535385, + "learning_rate": 3.76658300316315e-05, + "loss": 0.1835, + "step": 30471 + }, + { + "epoch": 0.7143116674578198, + "grad_norm": 0.3328339457511902, + "learning_rate": 3.766007023682285e-05, + "loss": 0.0723, + "step": 30472 + }, + { + "epoch": 0.7143351090326248, + "grad_norm": 0.45998039841651917, + "learning_rate": 3.765431078027241e-05, + "loss": 0.1305, + "step": 30473 + }, + { + "epoch": 0.7143585506074298, + "grad_norm": 0.3877652585506439, + "learning_rate": 3.764855166201143e-05, + "loss": 0.1102, + "step": 30474 + }, + { + "epoch": 0.7143819921822347, + "grad_norm": 0.5621222257614136, + "learning_rate": 3.764279288207117e-05, + "loss": 0.6836, + "step": 30475 + }, + { + "epoch": 0.7144054337570398, + "grad_norm": 0.4905662536621094, + "learning_rate": 3.7637034440482834e-05, + "loss": 0.0866, + "step": 30476 + }, + { + "epoch": 0.7144288753318448, + "grad_norm": 0.3207966387271881, + "learning_rate": 3.763127633727773e-05, + "loss": 0.0455, + "step": 30477 + }, + { + "epoch": 0.7144523169066498, + "grad_norm": 1.172392725944519, + "learning_rate": 3.762551857248708e-05, + "loss": 0.3136, + "step": 30478 + }, + { + "epoch": 0.7144757584814548, + "grad_norm": 0.25328144431114197, + "learning_rate": 3.761976114614213e-05, + "loss": 0.0251, + "step": 30479 + }, + { + "epoch": 0.7144992000562598, + "grad_norm": 0.31552132964134216, + "learning_rate": 3.7614004058274085e-05, + "loss": 0.029, + "step": 30480 + }, + { + "epoch": 0.7145226416310648, + "grad_norm": 0.19967100024223328, + "learning_rate": 3.7608247308914236e-05, + "loss": 0.0356, + "step": 30481 + }, + { + "epoch": 0.7145460832058698, + "grad_norm": 0.3837345838546753, + "learning_rate": 3.760249089809374e-05, + "loss": 0.0555, + "step": 30482 + }, + { + "epoch": 0.7145695247806748, + "grad_norm": 0.6848743557929993, + "learning_rate": 3.7596734825843915e-05, + "loss": 0.1486, + "step": 30483 + }, + { + "epoch": 0.7145929663554798, + "grad_norm": 0.4677309989929199, + "learning_rate": 3.7590979092195934e-05, + "loss": 0.5378, + "step": 30484 + }, + { + "epoch": 0.7146164079302848, + "grad_norm": 0.4673340618610382, + "learning_rate": 3.758522369718109e-05, + "loss": 0.0926, + "step": 30485 + }, + { + "epoch": 0.7146398495050897, + "grad_norm": 0.3820054531097412, + "learning_rate": 3.757946864083057e-05, + "loss": 0.053, + "step": 30486 + }, + { + "epoch": 0.7146632910798948, + "grad_norm": 0.394228458404541, + "learning_rate": 3.7573713923175624e-05, + "loss": 0.039, + "step": 30487 + }, + { + "epoch": 0.7146867326546997, + "grad_norm": 0.3356671929359436, + "learning_rate": 3.7567959544247445e-05, + "loss": 0.0373, + "step": 30488 + }, + { + "epoch": 0.7147101742295048, + "grad_norm": 0.5992522835731506, + "learning_rate": 3.756220550407724e-05, + "loss": 0.6511, + "step": 30489 + }, + { + "epoch": 0.7147336158043097, + "grad_norm": 0.4737433195114136, + "learning_rate": 3.755645180269631e-05, + "loss": 0.0807, + "step": 30490 + }, + { + "epoch": 0.7147570573791148, + "grad_norm": 0.2838090658187866, + "learning_rate": 3.755069844013582e-05, + "loss": 0.0314, + "step": 30491 + }, + { + "epoch": 0.7147804989539197, + "grad_norm": 0.24065616726875305, + "learning_rate": 3.7544945416426994e-05, + "loss": 0.0631, + "step": 30492 + }, + { + "epoch": 0.7148039405287248, + "grad_norm": 0.5584911704063416, + "learning_rate": 3.753919273160106e-05, + "loss": 0.6488, + "step": 30493 + }, + { + "epoch": 0.7148273821035297, + "grad_norm": 0.6332605481147766, + "learning_rate": 3.7533440385689214e-05, + "loss": 0.4553, + "step": 30494 + }, + { + "epoch": 0.7148508236783347, + "grad_norm": 0.07337916642427444, + "learning_rate": 3.752768837872264e-05, + "loss": 0.0114, + "step": 30495 + }, + { + "epoch": 0.7148742652531397, + "grad_norm": 0.6097245812416077, + "learning_rate": 3.752193671073259e-05, + "loss": 0.6459, + "step": 30496 + }, + { + "epoch": 0.7148977068279447, + "grad_norm": 0.05118069797754288, + "learning_rate": 3.75161853817503e-05, + "loss": 0.0047, + "step": 30497 + }, + { + "epoch": 0.7149211484027497, + "grad_norm": 0.1538894772529602, + "learning_rate": 3.751043439180695e-05, + "loss": 0.0146, + "step": 30498 + }, + { + "epoch": 0.7149445899775547, + "grad_norm": 0.4889289438724518, + "learning_rate": 3.750468374093373e-05, + "loss": 0.0609, + "step": 30499 + }, + { + "epoch": 0.7149680315523597, + "grad_norm": 0.3441523015499115, + "learning_rate": 3.7498933429161865e-05, + "loss": 0.0704, + "step": 30500 + }, + { + "epoch": 0.7149914731271647, + "grad_norm": 0.4135861098766327, + "learning_rate": 3.749318345652254e-05, + "loss": 0.0688, + "step": 30501 + }, + { + "epoch": 0.7150149147019697, + "grad_norm": 0.32189279794692993, + "learning_rate": 3.7487433823046926e-05, + "loss": 0.0507, + "step": 30502 + }, + { + "epoch": 0.7150383562767747, + "grad_norm": 0.15794524550437927, + "learning_rate": 3.7481684528766294e-05, + "loss": 0.0136, + "step": 30503 + }, + { + "epoch": 0.7150617978515796, + "grad_norm": 0.7730216383934021, + "learning_rate": 3.747593557371179e-05, + "loss": 0.0481, + "step": 30504 + }, + { + "epoch": 0.7150852394263847, + "grad_norm": 0.46378299593925476, + "learning_rate": 3.747018695791462e-05, + "loss": 0.0833, + "step": 30505 + }, + { + "epoch": 0.7151086810011896, + "grad_norm": 0.391120970249176, + "learning_rate": 3.7464438681405965e-05, + "loss": 0.0724, + "step": 30506 + }, + { + "epoch": 0.7151321225759947, + "grad_norm": 0.4893011748790741, + "learning_rate": 3.7458690744217005e-05, + "loss": 0.0637, + "step": 30507 + }, + { + "epoch": 0.7151555641507996, + "grad_norm": 0.49269944429397583, + "learning_rate": 3.745294314637897e-05, + "loss": 0.1349, + "step": 30508 + }, + { + "epoch": 0.7151790057256047, + "grad_norm": 0.675369381904602, + "learning_rate": 3.7447195887923005e-05, + "loss": 0.0484, + "step": 30509 + }, + { + "epoch": 0.7152024473004096, + "grad_norm": 0.6200929880142212, + "learning_rate": 3.744144896888033e-05, + "loss": 0.5052, + "step": 30510 + }, + { + "epoch": 0.7152258888752147, + "grad_norm": 0.3436442017555237, + "learning_rate": 3.743570238928212e-05, + "loss": 0.057, + "step": 30511 + }, + { + "epoch": 0.7152493304500196, + "grad_norm": 0.28344297409057617, + "learning_rate": 3.742995614915955e-05, + "loss": 0.0516, + "step": 30512 + }, + { + "epoch": 0.7152727720248246, + "grad_norm": 0.38440215587615967, + "learning_rate": 3.74242102485438e-05, + "loss": 0.0558, + "step": 30513 + }, + { + "epoch": 0.7152962135996296, + "grad_norm": 0.26306626200675964, + "learning_rate": 3.7418464687466e-05, + "loss": 0.0401, + "step": 30514 + }, + { + "epoch": 0.7153196551744346, + "grad_norm": 0.09504205733537674, + "learning_rate": 3.741271946595741e-05, + "loss": 0.0119, + "step": 30515 + }, + { + "epoch": 0.7153430967492396, + "grad_norm": 0.07457628101110458, + "learning_rate": 3.740697458404917e-05, + "loss": 0.0126, + "step": 30516 + }, + { + "epoch": 0.7153665383240446, + "grad_norm": 0.31742027401924133, + "learning_rate": 3.740123004177244e-05, + "loss": 0.0458, + "step": 30517 + }, + { + "epoch": 0.7153899798988496, + "grad_norm": 0.4349190592765808, + "learning_rate": 3.73954858391584e-05, + "loss": 0.0685, + "step": 30518 + }, + { + "epoch": 0.7154134214736546, + "grad_norm": 0.43455609679222107, + "learning_rate": 3.738974197623821e-05, + "loss": 0.121, + "step": 30519 + }, + { + "epoch": 0.7154368630484595, + "grad_norm": 0.13069090247154236, + "learning_rate": 3.7383998453043e-05, + "loss": 0.0132, + "step": 30520 + }, + { + "epoch": 0.7154603046232646, + "grad_norm": 0.2751975655555725, + "learning_rate": 3.737825526960399e-05, + "loss": 0.0398, + "step": 30521 + }, + { + "epoch": 0.7154837461980695, + "grad_norm": 0.121795155107975, + "learning_rate": 3.7372512425952345e-05, + "loss": 0.0464, + "step": 30522 + }, + { + "epoch": 0.7155071877728746, + "grad_norm": 0.21698565781116486, + "learning_rate": 3.7366769922119224e-05, + "loss": 0.0145, + "step": 30523 + }, + { + "epoch": 0.7155306293476795, + "grad_norm": 0.621476948261261, + "learning_rate": 3.736102775813576e-05, + "loss": 0.0902, + "step": 30524 + }, + { + "epoch": 0.7155540709224846, + "grad_norm": 0.4459172487258911, + "learning_rate": 3.735528593403312e-05, + "loss": 0.1049, + "step": 30525 + }, + { + "epoch": 0.7155775124972895, + "grad_norm": 0.4127280116081238, + "learning_rate": 3.734954444984247e-05, + "loss": 0.0631, + "step": 30526 + }, + { + "epoch": 0.7156009540720946, + "grad_norm": 0.20615807175636292, + "learning_rate": 3.734380330559491e-05, + "loss": 0.0224, + "step": 30527 + }, + { + "epoch": 0.7156243956468996, + "grad_norm": 0.3202511966228485, + "learning_rate": 3.733806250132168e-05, + "loss": 0.055, + "step": 30528 + }, + { + "epoch": 0.7156478372217046, + "grad_norm": 0.25570908188819885, + "learning_rate": 3.733232203705387e-05, + "loss": 0.0457, + "step": 30529 + }, + { + "epoch": 0.7156712787965096, + "grad_norm": 0.4533878564834595, + "learning_rate": 3.732658191282266e-05, + "loss": 0.0953, + "step": 30530 + }, + { + "epoch": 0.7156947203713145, + "grad_norm": 0.3715413212776184, + "learning_rate": 3.732084212865916e-05, + "loss": 0.0417, + "step": 30531 + }, + { + "epoch": 0.7157181619461196, + "grad_norm": 0.3535982668399811, + "learning_rate": 3.731510268459451e-05, + "loss": 0.0514, + "step": 30532 + }, + { + "epoch": 0.7157416035209245, + "grad_norm": 0.4603332579135895, + "learning_rate": 3.730936358065991e-05, + "loss": 0.0796, + "step": 30533 + }, + { + "epoch": 0.7157650450957296, + "grad_norm": 0.692254900932312, + "learning_rate": 3.730362481688643e-05, + "loss": 0.6653, + "step": 30534 + }, + { + "epoch": 0.7157884866705345, + "grad_norm": 0.19211138784885406, + "learning_rate": 3.729788639330527e-05, + "loss": 0.0184, + "step": 30535 + }, + { + "epoch": 0.7158119282453396, + "grad_norm": 0.4904460906982422, + "learning_rate": 3.729214830994755e-05, + "loss": 0.676, + "step": 30536 + }, + { + "epoch": 0.7158353698201445, + "grad_norm": 0.27495500445365906, + "learning_rate": 3.72864105668444e-05, + "loss": 0.0451, + "step": 30537 + }, + { + "epoch": 0.7158588113949496, + "grad_norm": 0.23596560955047607, + "learning_rate": 3.7280673164026944e-05, + "loss": 0.0582, + "step": 30538 + }, + { + "epoch": 0.7158822529697545, + "grad_norm": 0.5155832171440125, + "learning_rate": 3.7274936101526314e-05, + "loss": 0.1124, + "step": 30539 + }, + { + "epoch": 0.7159056945445595, + "grad_norm": 0.42324888706207275, + "learning_rate": 3.726919937937362e-05, + "loss": 0.0562, + "step": 30540 + }, + { + "epoch": 0.7159291361193645, + "grad_norm": 0.3920568525791168, + "learning_rate": 3.726346299760005e-05, + "loss": 0.0578, + "step": 30541 + }, + { + "epoch": 0.7159525776941695, + "grad_norm": 0.30388280749320984, + "learning_rate": 3.725772695623668e-05, + "loss": 0.055, + "step": 30542 + }, + { + "epoch": 0.7159760192689745, + "grad_norm": 0.6241393089294434, + "learning_rate": 3.7251991255314654e-05, + "loss": 0.0868, + "step": 30543 + }, + { + "epoch": 0.7159994608437795, + "grad_norm": 0.1524875909090042, + "learning_rate": 3.724625589486505e-05, + "loss": 0.0226, + "step": 30544 + }, + { + "epoch": 0.7160229024185845, + "grad_norm": 0.4632708430290222, + "learning_rate": 3.724052087491908e-05, + "loss": 0.0508, + "step": 30545 + }, + { + "epoch": 0.7160463439933895, + "grad_norm": 0.47848814725875854, + "learning_rate": 3.723478619550779e-05, + "loss": 0.0444, + "step": 30546 + }, + { + "epoch": 0.7160697855681945, + "grad_norm": 0.39996305108070374, + "learning_rate": 3.722905185666229e-05, + "loss": 0.0585, + "step": 30547 + }, + { + "epoch": 0.7160932271429995, + "grad_norm": 0.5493667721748352, + "learning_rate": 3.722331785841374e-05, + "loss": 0.073, + "step": 30548 + }, + { + "epoch": 0.7161166687178044, + "grad_norm": 0.14744697511196136, + "learning_rate": 3.721758420079324e-05, + "loss": 0.0232, + "step": 30549 + }, + { + "epoch": 0.7161401102926095, + "grad_norm": 0.7372650504112244, + "learning_rate": 3.72118508838319e-05, + "loss": 0.044, + "step": 30550 + }, + { + "epoch": 0.7161635518674144, + "grad_norm": 0.6120419502258301, + "learning_rate": 3.720611790756082e-05, + "loss": 0.1085, + "step": 30551 + }, + { + "epoch": 0.7161869934422195, + "grad_norm": 0.19575728476047516, + "learning_rate": 3.7200385272011075e-05, + "loss": 0.054, + "step": 30552 + }, + { + "epoch": 0.7162104350170244, + "grad_norm": 0.1052921712398529, + "learning_rate": 3.719465297721383e-05, + "loss": 0.0113, + "step": 30553 + }, + { + "epoch": 0.7162338765918295, + "grad_norm": 0.1919008046388626, + "learning_rate": 3.7188921023200175e-05, + "loss": 0.0302, + "step": 30554 + }, + { + "epoch": 0.7162573181666344, + "grad_norm": 0.485818475484848, + "learning_rate": 3.71831894100012e-05, + "loss": 0.0645, + "step": 30555 + }, + { + "epoch": 0.7162807597414395, + "grad_norm": 0.12443750351667404, + "learning_rate": 3.7177458137648e-05, + "loss": 0.0142, + "step": 30556 + }, + { + "epoch": 0.7163042013162444, + "grad_norm": 0.14635083079338074, + "learning_rate": 3.717172720617165e-05, + "loss": 0.0273, + "step": 30557 + }, + { + "epoch": 0.7163276428910494, + "grad_norm": 0.11408203840255737, + "learning_rate": 3.7165996615603304e-05, + "loss": 0.0147, + "step": 30558 + }, + { + "epoch": 0.7163510844658544, + "grad_norm": 0.1838829070329666, + "learning_rate": 3.716026636597399e-05, + "loss": 0.0356, + "step": 30559 + }, + { + "epoch": 0.7163745260406594, + "grad_norm": 0.33123835921287537, + "learning_rate": 3.715453645731488e-05, + "loss": 0.0203, + "step": 30560 + }, + { + "epoch": 0.7163979676154644, + "grad_norm": 0.1711944192647934, + "learning_rate": 3.7148806889657016e-05, + "loss": 0.0374, + "step": 30561 + }, + { + "epoch": 0.7164214091902694, + "grad_norm": 0.083324134349823, + "learning_rate": 3.714307766303149e-05, + "loss": 0.0089, + "step": 30562 + }, + { + "epoch": 0.7164448507650744, + "grad_norm": 0.37749019265174866, + "learning_rate": 3.71373487774694e-05, + "loss": 0.0643, + "step": 30563 + }, + { + "epoch": 0.7164682923398794, + "grad_norm": 0.35820862650871277, + "learning_rate": 3.713162023300182e-05, + "loss": 0.0793, + "step": 30564 + }, + { + "epoch": 0.7164917339146843, + "grad_norm": 0.3086417317390442, + "learning_rate": 3.71258920296598e-05, + "loss": 0.0479, + "step": 30565 + }, + { + "epoch": 0.7165151754894894, + "grad_norm": 0.431240439414978, + "learning_rate": 3.71201641674745e-05, + "loss": 0.0652, + "step": 30566 + }, + { + "epoch": 0.7165386170642943, + "grad_norm": 0.36357641220092773, + "learning_rate": 3.711443664647695e-05, + "loss": 0.5642, + "step": 30567 + }, + { + "epoch": 0.7165620586390994, + "grad_norm": 0.260201096534729, + "learning_rate": 3.710870946669823e-05, + "loss": 0.0572, + "step": 30568 + }, + { + "epoch": 0.7165855002139043, + "grad_norm": 0.6943192481994629, + "learning_rate": 3.71029826281694e-05, + "loss": 0.0822, + "step": 30569 + }, + { + "epoch": 0.7166089417887094, + "grad_norm": 0.5266618132591248, + "learning_rate": 3.709725613092159e-05, + "loss": 0.0661, + "step": 30570 + }, + { + "epoch": 0.7166323833635143, + "grad_norm": 0.2132066786289215, + "learning_rate": 3.7091529974985814e-05, + "loss": 0.0594, + "step": 30571 + }, + { + "epoch": 0.7166558249383194, + "grad_norm": 0.40779080986976624, + "learning_rate": 3.708580416039316e-05, + "loss": 0.0807, + "step": 30572 + }, + { + "epoch": 0.7166792665131243, + "grad_norm": 0.19186188280582428, + "learning_rate": 3.7080078687174714e-05, + "loss": 0.0332, + "step": 30573 + }, + { + "epoch": 0.7167027080879294, + "grad_norm": 0.5597171187400818, + "learning_rate": 3.707435355536154e-05, + "loss": 0.1292, + "step": 30574 + }, + { + "epoch": 0.7167261496627343, + "grad_norm": 0.1445489376783371, + "learning_rate": 3.706862876498469e-05, + "loss": 0.0369, + "step": 30575 + }, + { + "epoch": 0.7167495912375393, + "grad_norm": 0.44013988971710205, + "learning_rate": 3.706290431607523e-05, + "loss": 0.041, + "step": 30576 + }, + { + "epoch": 0.7167730328123443, + "grad_norm": 0.2012087106704712, + "learning_rate": 3.705718020866422e-05, + "loss": 0.0541, + "step": 30577 + }, + { + "epoch": 0.7167964743871493, + "grad_norm": 0.5525338649749756, + "learning_rate": 3.7051456442782694e-05, + "loss": 0.1167, + "step": 30578 + }, + { + "epoch": 0.7168199159619544, + "grad_norm": 0.4792790114879608, + "learning_rate": 3.704573301846176e-05, + "loss": 0.0526, + "step": 30579 + }, + { + "epoch": 0.7168433575367593, + "grad_norm": 0.22283786535263062, + "learning_rate": 3.704000993573246e-05, + "loss": 0.0294, + "step": 30580 + }, + { + "epoch": 0.7168667991115644, + "grad_norm": 0.36026161909103394, + "learning_rate": 3.7034287194625786e-05, + "loss": 0.0357, + "step": 30581 + }, + { + "epoch": 0.7168902406863693, + "grad_norm": 0.43687373399734497, + "learning_rate": 3.702856479517288e-05, + "loss": 0.0706, + "step": 30582 + }, + { + "epoch": 0.7169136822611744, + "grad_norm": 0.08246130496263504, + "learning_rate": 3.702284273740476e-05, + "loss": 0.0152, + "step": 30583 + }, + { + "epoch": 0.7169371238359793, + "grad_norm": 0.0909360870718956, + "learning_rate": 3.701712102135245e-05, + "loss": 0.0171, + "step": 30584 + }, + { + "epoch": 0.7169605654107843, + "grad_norm": 0.6666103005409241, + "learning_rate": 3.7011399647047e-05, + "loss": 0.0935, + "step": 30585 + }, + { + "epoch": 0.7169840069855893, + "grad_norm": 0.3663959205150604, + "learning_rate": 3.700567861451948e-05, + "loss": 0.0349, + "step": 30586 + }, + { + "epoch": 0.7170074485603943, + "grad_norm": 0.6339995265007019, + "learning_rate": 3.699995792380093e-05, + "loss": 0.118, + "step": 30587 + }, + { + "epoch": 0.7170308901351993, + "grad_norm": 0.27144020795822144, + "learning_rate": 3.699423757492239e-05, + "loss": 0.0414, + "step": 30588 + }, + { + "epoch": 0.7170543317100043, + "grad_norm": 0.34738439321517944, + "learning_rate": 3.698851756791487e-05, + "loss": 0.0457, + "step": 30589 + }, + { + "epoch": 0.7170777732848093, + "grad_norm": 0.29889512062072754, + "learning_rate": 3.6982797902809395e-05, + "loss": 0.0519, + "step": 30590 + }, + { + "epoch": 0.7171012148596143, + "grad_norm": 0.6534624695777893, + "learning_rate": 3.697707857963707e-05, + "loss": 0.6309, + "step": 30591 + }, + { + "epoch": 0.7171246564344192, + "grad_norm": 0.6937369704246521, + "learning_rate": 3.6971359598428876e-05, + "loss": 0.1316, + "step": 30592 + }, + { + "epoch": 0.7171480980092243, + "grad_norm": 0.46248361468315125, + "learning_rate": 3.696564095921584e-05, + "loss": 0.0745, + "step": 30593 + }, + { + "epoch": 0.7171715395840292, + "grad_norm": 0.30517691373825073, + "learning_rate": 3.695992266202903e-05, + "loss": 0.0352, + "step": 30594 + }, + { + "epoch": 0.7171949811588343, + "grad_norm": 0.346303790807724, + "learning_rate": 3.695420470689947e-05, + "loss": 0.0306, + "step": 30595 + }, + { + "epoch": 0.7172184227336392, + "grad_norm": 0.3012593388557434, + "learning_rate": 3.6948487093858154e-05, + "loss": 0.0561, + "step": 30596 + }, + { + "epoch": 0.7172418643084443, + "grad_norm": 0.1941872090101242, + "learning_rate": 3.694276982293608e-05, + "loss": 0.0468, + "step": 30597 + }, + { + "epoch": 0.7172653058832492, + "grad_norm": 0.48406538367271423, + "learning_rate": 3.693705289416436e-05, + "loss": 0.0679, + "step": 30598 + }, + { + "epoch": 0.7172887474580543, + "grad_norm": 0.46041303873062134, + "learning_rate": 3.693133630757395e-05, + "loss": 0.1056, + "step": 30599 + }, + { + "epoch": 0.7173121890328592, + "grad_norm": 0.3214108645915985, + "learning_rate": 3.692562006319589e-05, + "loss": 0.0489, + "step": 30600 + }, + { + "epoch": 0.7173356306076643, + "grad_norm": 0.27382728457450867, + "learning_rate": 3.6919904161061183e-05, + "loss": 0.0325, + "step": 30601 + }, + { + "epoch": 0.7173590721824692, + "grad_norm": 0.42099887132644653, + "learning_rate": 3.691418860120086e-05, + "loss": 0.0613, + "step": 30602 + }, + { + "epoch": 0.7173825137572742, + "grad_norm": 0.05207018554210663, + "learning_rate": 3.690847338364588e-05, + "loss": 0.005, + "step": 30603 + }, + { + "epoch": 0.7174059553320792, + "grad_norm": 0.5290929079055786, + "learning_rate": 3.690275850842734e-05, + "loss": 0.0372, + "step": 30604 + }, + { + "epoch": 0.7174293969068842, + "grad_norm": 0.5103658437728882, + "learning_rate": 3.689704397557616e-05, + "loss": 0.0449, + "step": 30605 + }, + { + "epoch": 0.7174528384816892, + "grad_norm": 0.5137535333633423, + "learning_rate": 3.689132978512344e-05, + "loss": 0.1419, + "step": 30606 + }, + { + "epoch": 0.7174762800564942, + "grad_norm": 0.48186224699020386, + "learning_rate": 3.688561593710013e-05, + "loss": 0.0585, + "step": 30607 + }, + { + "epoch": 0.7174997216312992, + "grad_norm": 0.5285707712173462, + "learning_rate": 3.687990243153725e-05, + "loss": 0.141, + "step": 30608 + }, + { + "epoch": 0.7175231632061042, + "grad_norm": 0.5142881870269775, + "learning_rate": 3.687418926846579e-05, + "loss": 0.1024, + "step": 30609 + }, + { + "epoch": 0.7175466047809091, + "grad_norm": 0.5512416958808899, + "learning_rate": 3.6868476447916724e-05, + "loss": 0.7521, + "step": 30610 + }, + { + "epoch": 0.7175700463557142, + "grad_norm": 0.32706212997436523, + "learning_rate": 3.68627639699211e-05, + "loss": 0.0666, + "step": 30611 + }, + { + "epoch": 0.7175934879305191, + "grad_norm": 0.5796104669570923, + "learning_rate": 3.6857051834509906e-05, + "loss": 0.1182, + "step": 30612 + }, + { + "epoch": 0.7176169295053242, + "grad_norm": 0.6272682547569275, + "learning_rate": 3.685134004171412e-05, + "loss": 0.1516, + "step": 30613 + }, + { + "epoch": 0.7176403710801291, + "grad_norm": 0.45344945788383484, + "learning_rate": 3.684562859156475e-05, + "loss": 0.6002, + "step": 30614 + }, + { + "epoch": 0.7176638126549342, + "grad_norm": 0.29339879751205444, + "learning_rate": 3.6839917484092766e-05, + "loss": 0.041, + "step": 30615 + }, + { + "epoch": 0.7176872542297391, + "grad_norm": 0.27878984808921814, + "learning_rate": 3.683420671932912e-05, + "loss": 0.0253, + "step": 30616 + }, + { + "epoch": 0.7177106958045442, + "grad_norm": 0.46854525804519653, + "learning_rate": 3.682849629730486e-05, + "loss": 0.1013, + "step": 30617 + }, + { + "epoch": 0.7177341373793491, + "grad_norm": 0.2119196057319641, + "learning_rate": 3.6822786218050984e-05, + "loss": 0.0127, + "step": 30618 + }, + { + "epoch": 0.7177575789541542, + "grad_norm": 0.33707889914512634, + "learning_rate": 3.6817076481598456e-05, + "loss": 0.0819, + "step": 30619 + }, + { + "epoch": 0.7177810205289591, + "grad_norm": 0.4794977903366089, + "learning_rate": 3.681136708797823e-05, + "loss": 0.0474, + "step": 30620 + }, + { + "epoch": 0.7178044621037641, + "grad_norm": 0.31105345487594604, + "learning_rate": 3.680565803722132e-05, + "loss": 0.0385, + "step": 30621 + }, + { + "epoch": 0.7178279036785691, + "grad_norm": 0.5941876769065857, + "learning_rate": 3.679994932935867e-05, + "loss": 0.0752, + "step": 30622 + }, + { + "epoch": 0.7178513452533741, + "grad_norm": 0.4307471513748169, + "learning_rate": 3.6794240964421255e-05, + "loss": 0.0484, + "step": 30623 + }, + { + "epoch": 0.7178747868281791, + "grad_norm": 0.37748876214027405, + "learning_rate": 3.678853294244009e-05, + "loss": 0.0469, + "step": 30624 + }, + { + "epoch": 0.7178982284029841, + "grad_norm": 0.15033337473869324, + "learning_rate": 3.678282526344613e-05, + "loss": 0.0372, + "step": 30625 + }, + { + "epoch": 0.7179216699777891, + "grad_norm": 0.4665899872779846, + "learning_rate": 3.677711792747033e-05, + "loss": 0.0884, + "step": 30626 + }, + { + "epoch": 0.7179451115525941, + "grad_norm": 0.34337562322616577, + "learning_rate": 3.677141093454366e-05, + "loss": 0.0775, + "step": 30627 + }, + { + "epoch": 0.717968553127399, + "grad_norm": 0.2725405693054199, + "learning_rate": 3.676570428469707e-05, + "loss": 0.036, + "step": 30628 + }, + { + "epoch": 0.7179919947022041, + "grad_norm": 0.41645142436027527, + "learning_rate": 3.675999797796158e-05, + "loss": 0.0484, + "step": 30629 + }, + { + "epoch": 0.7180154362770091, + "grad_norm": 1.0988627672195435, + "learning_rate": 3.675429201436808e-05, + "loss": 0.1536, + "step": 30630 + }, + { + "epoch": 0.7180388778518141, + "grad_norm": 0.2637357711791992, + "learning_rate": 3.67485863939476e-05, + "loss": 0.0202, + "step": 30631 + }, + { + "epoch": 0.7180623194266191, + "grad_norm": 0.6580572724342346, + "learning_rate": 3.674288111673108e-05, + "loss": 0.4488, + "step": 30632 + }, + { + "epoch": 0.7180857610014241, + "grad_norm": 0.4658445417881012, + "learning_rate": 3.673717618274946e-05, + "loss": 0.0681, + "step": 30633 + }, + { + "epoch": 0.7181092025762291, + "grad_norm": 0.13360804319381714, + "learning_rate": 3.673147159203369e-05, + "loss": 0.0248, + "step": 30634 + }, + { + "epoch": 0.7181326441510341, + "grad_norm": 0.26935485005378723, + "learning_rate": 3.6725767344614714e-05, + "loss": 0.0582, + "step": 30635 + }, + { + "epoch": 0.7181560857258391, + "grad_norm": 0.40632525086402893, + "learning_rate": 3.672006344052353e-05, + "loss": 0.0672, + "step": 30636 + }, + { + "epoch": 0.718179527300644, + "grad_norm": 0.6462616324424744, + "learning_rate": 3.671435987979107e-05, + "loss": 0.1472, + "step": 30637 + }, + { + "epoch": 0.7182029688754491, + "grad_norm": 0.4952114224433899, + "learning_rate": 3.670865666244827e-05, + "loss": 0.0959, + "step": 30638 + }, + { + "epoch": 0.718226410450254, + "grad_norm": 0.3973379135131836, + "learning_rate": 3.670295378852607e-05, + "loss": 0.0373, + "step": 30639 + }, + { + "epoch": 0.7182498520250591, + "grad_norm": 0.19370467960834503, + "learning_rate": 3.669725125805543e-05, + "loss": 0.055, + "step": 30640 + }, + { + "epoch": 0.718273293599864, + "grad_norm": 0.20069003105163574, + "learning_rate": 3.669154907106726e-05, + "loss": 0.0195, + "step": 30641 + }, + { + "epoch": 0.7182967351746691, + "grad_norm": 0.570690393447876, + "learning_rate": 3.6685847227592504e-05, + "loss": 0.085, + "step": 30642 + }, + { + "epoch": 0.718320176749474, + "grad_norm": 0.4527473747730255, + "learning_rate": 3.6680145727662164e-05, + "loss": 0.0594, + "step": 30643 + }, + { + "epoch": 0.7183436183242791, + "grad_norm": 0.40878963470458984, + "learning_rate": 3.6674444571307144e-05, + "loss": 0.0739, + "step": 30644 + }, + { + "epoch": 0.718367059899084, + "grad_norm": 0.1640022099018097, + "learning_rate": 3.666874375855836e-05, + "loss": 0.0176, + "step": 30645 + }, + { + "epoch": 0.718390501473889, + "grad_norm": 0.3873932659626007, + "learning_rate": 3.666304328944674e-05, + "loss": 0.0775, + "step": 30646 + }, + { + "epoch": 0.718413943048694, + "grad_norm": 0.30333542823791504, + "learning_rate": 3.665734316400324e-05, + "loss": 0.0315, + "step": 30647 + }, + { + "epoch": 0.718437384623499, + "grad_norm": 0.461148202419281, + "learning_rate": 3.665164338225875e-05, + "loss": 0.0842, + "step": 30648 + }, + { + "epoch": 0.718460826198304, + "grad_norm": 0.22587496042251587, + "learning_rate": 3.6645943944244254e-05, + "loss": 0.0243, + "step": 30649 + }, + { + "epoch": 0.718484267773109, + "grad_norm": 0.14398838579654694, + "learning_rate": 3.6640244849990646e-05, + "loss": 0.0266, + "step": 30650 + }, + { + "epoch": 0.718507709347914, + "grad_norm": 0.25618308782577515, + "learning_rate": 3.663454609952885e-05, + "loss": 0.2508, + "step": 30651 + }, + { + "epoch": 0.718531150922719, + "grad_norm": 0.14337265491485596, + "learning_rate": 3.662884769288979e-05, + "loss": 0.0193, + "step": 30652 + }, + { + "epoch": 0.718554592497524, + "grad_norm": 0.47790515422821045, + "learning_rate": 3.662314963010438e-05, + "loss": 0.0737, + "step": 30653 + }, + { + "epoch": 0.718578034072329, + "grad_norm": 0.4289834499359131, + "learning_rate": 3.6617451911203505e-05, + "loss": 0.054, + "step": 30654 + }, + { + "epoch": 0.718601475647134, + "grad_norm": 0.35298845171928406, + "learning_rate": 3.6611754536218126e-05, + "loss": 0.0806, + "step": 30655 + }, + { + "epoch": 0.718624917221939, + "grad_norm": 0.21424852311611176, + "learning_rate": 3.660605750517918e-05, + "loss": 0.0368, + "step": 30656 + }, + { + "epoch": 0.7186483587967439, + "grad_norm": 0.1962999701499939, + "learning_rate": 3.660036081811755e-05, + "loss": 0.0244, + "step": 30657 + }, + { + "epoch": 0.718671800371549, + "grad_norm": 0.7421748638153076, + "learning_rate": 3.6594664475064146e-05, + "loss": 0.1146, + "step": 30658 + }, + { + "epoch": 0.7186952419463539, + "grad_norm": 0.43420639634132385, + "learning_rate": 3.658896847604988e-05, + "loss": 0.0585, + "step": 30659 + }, + { + "epoch": 0.718718683521159, + "grad_norm": 0.3261392414569855, + "learning_rate": 3.6583272821105644e-05, + "loss": 0.2833, + "step": 30660 + }, + { + "epoch": 0.7187421250959639, + "grad_norm": 0.3913629949092865, + "learning_rate": 3.6577577510262314e-05, + "loss": 0.0661, + "step": 30661 + }, + { + "epoch": 0.718765566670769, + "grad_norm": 1.0256495475769043, + "learning_rate": 3.657188254355089e-05, + "loss": 0.1807, + "step": 30662 + }, + { + "epoch": 0.7187890082455739, + "grad_norm": 0.14137689769268036, + "learning_rate": 3.656618792100219e-05, + "loss": 0.027, + "step": 30663 + }, + { + "epoch": 0.718812449820379, + "grad_norm": 0.5097703337669373, + "learning_rate": 3.6560493642647154e-05, + "loss": 0.0745, + "step": 30664 + }, + { + "epoch": 0.7188358913951839, + "grad_norm": 0.38791829347610474, + "learning_rate": 3.6554799708516655e-05, + "loss": 0.0455, + "step": 30665 + }, + { + "epoch": 0.7188593329699889, + "grad_norm": 0.29760366678237915, + "learning_rate": 3.654910611864156e-05, + "loss": 0.0878, + "step": 30666 + }, + { + "epoch": 0.7188827745447939, + "grad_norm": 0.4529300928115845, + "learning_rate": 3.654341287305284e-05, + "loss": 0.077, + "step": 30667 + }, + { + "epoch": 0.7189062161195989, + "grad_norm": 0.34693604707717896, + "learning_rate": 3.65377199717813e-05, + "loss": 0.0405, + "step": 30668 + }, + { + "epoch": 0.7189296576944039, + "grad_norm": 0.24960607290267944, + "learning_rate": 3.653202741485793e-05, + "loss": 0.0444, + "step": 30669 + }, + { + "epoch": 0.7189530992692089, + "grad_norm": 0.4905325174331665, + "learning_rate": 3.652633520231355e-05, + "loss": 0.6455, + "step": 30670 + }, + { + "epoch": 0.7189765408440139, + "grad_norm": 0.14363215863704681, + "learning_rate": 3.652064333417906e-05, + "loss": 0.0349, + "step": 30671 + }, + { + "epoch": 0.7189999824188189, + "grad_norm": 0.39730724692344666, + "learning_rate": 3.651495181048536e-05, + "loss": 0.0567, + "step": 30672 + }, + { + "epoch": 0.7190234239936238, + "grad_norm": 0.29533207416534424, + "learning_rate": 3.6509260631263254e-05, + "loss": 0.0341, + "step": 30673 + }, + { + "epoch": 0.7190468655684289, + "grad_norm": 0.5291714072227478, + "learning_rate": 3.6503569796543735e-05, + "loss": 0.5566, + "step": 30674 + }, + { + "epoch": 0.7190703071432338, + "grad_norm": 0.3681759834289551, + "learning_rate": 3.649787930635764e-05, + "loss": 0.0781, + "step": 30675 + }, + { + "epoch": 0.7190937487180389, + "grad_norm": 0.3816951811313629, + "learning_rate": 3.6492189160735824e-05, + "loss": 0.0256, + "step": 30676 + }, + { + "epoch": 0.7191171902928438, + "grad_norm": 0.6988081336021423, + "learning_rate": 3.648649935970918e-05, + "loss": 0.6162, + "step": 30677 + }, + { + "epoch": 0.7191406318676489, + "grad_norm": 0.25130778551101685, + "learning_rate": 3.6480809903308545e-05, + "loss": 0.0365, + "step": 30678 + }, + { + "epoch": 0.7191640734424538, + "grad_norm": 0.3987826704978943, + "learning_rate": 3.647512079156485e-05, + "loss": 0.1061, + "step": 30679 + }, + { + "epoch": 0.7191875150172589, + "grad_norm": 0.31516948342323303, + "learning_rate": 3.6469432024508906e-05, + "loss": 0.3528, + "step": 30680 + }, + { + "epoch": 0.7192109565920639, + "grad_norm": 0.5398398637771606, + "learning_rate": 3.646374360217164e-05, + "loss": 0.4576, + "step": 30681 + }, + { + "epoch": 0.7192343981668688, + "grad_norm": 0.13971783220767975, + "learning_rate": 3.645805552458389e-05, + "loss": 0.027, + "step": 30682 + }, + { + "epoch": 0.7192578397416739, + "grad_norm": 0.32260212302207947, + "learning_rate": 3.645236779177652e-05, + "loss": 0.0381, + "step": 30683 + }, + { + "epoch": 0.7192812813164788, + "grad_norm": 0.4158742129802704, + "learning_rate": 3.6446680403780386e-05, + "loss": 0.0726, + "step": 30684 + }, + { + "epoch": 0.7193047228912839, + "grad_norm": 0.6160982847213745, + "learning_rate": 3.644099336062635e-05, + "loss": 0.0394, + "step": 30685 + }, + { + "epoch": 0.7193281644660888, + "grad_norm": 0.179814413189888, + "learning_rate": 3.6435306662345236e-05, + "loss": 0.0349, + "step": 30686 + }, + { + "epoch": 0.7193516060408939, + "grad_norm": 0.3573399484157562, + "learning_rate": 3.642962030896798e-05, + "loss": 0.3326, + "step": 30687 + }, + { + "epoch": 0.7193750476156988, + "grad_norm": 0.5742230415344238, + "learning_rate": 3.642393430052539e-05, + "loss": 0.068, + "step": 30688 + }, + { + "epoch": 0.7193984891905039, + "grad_norm": 0.3667587339878082, + "learning_rate": 3.6418248637048304e-05, + "loss": 0.0331, + "step": 30689 + }, + { + "epoch": 0.7194219307653088, + "grad_norm": 0.33944767713546753, + "learning_rate": 3.641256331856757e-05, + "loss": 0.0557, + "step": 30690 + }, + { + "epoch": 0.7194453723401139, + "grad_norm": 0.321911484003067, + "learning_rate": 3.640687834511408e-05, + "loss": 0.0613, + "step": 30691 + }, + { + "epoch": 0.7194688139149188, + "grad_norm": 0.39116793870925903, + "learning_rate": 3.6401193716718664e-05, + "loss": 0.0647, + "step": 30692 + }, + { + "epoch": 0.7194922554897238, + "grad_norm": 0.8140527606010437, + "learning_rate": 3.639550943341212e-05, + "loss": 0.1958, + "step": 30693 + }, + { + "epoch": 0.7195156970645288, + "grad_norm": 0.7041392922401428, + "learning_rate": 3.638982549522536e-05, + "loss": 0.4889, + "step": 30694 + }, + { + "epoch": 0.7195391386393338, + "grad_norm": 0.45480209589004517, + "learning_rate": 3.638414190218921e-05, + "loss": 0.0614, + "step": 30695 + }, + { + "epoch": 0.7195625802141388, + "grad_norm": 0.11545708030462265, + "learning_rate": 3.637845865433448e-05, + "loss": 0.0281, + "step": 30696 + }, + { + "epoch": 0.7195860217889438, + "grad_norm": 0.3787640631198883, + "learning_rate": 3.6372775751692033e-05, + "loss": 0.0611, + "step": 30697 + }, + { + "epoch": 0.7196094633637488, + "grad_norm": 0.4013719856739044, + "learning_rate": 3.636709319429269e-05, + "loss": 0.4632, + "step": 30698 + }, + { + "epoch": 0.7196329049385538, + "grad_norm": 0.09815280139446259, + "learning_rate": 3.636141098216726e-05, + "loss": 0.0175, + "step": 30699 + }, + { + "epoch": 0.7196563465133587, + "grad_norm": 0.5532071590423584, + "learning_rate": 3.635572911534662e-05, + "loss": 0.1018, + "step": 30700 + }, + { + "epoch": 0.7196797880881638, + "grad_norm": 0.391679584980011, + "learning_rate": 3.6350047593861594e-05, + "loss": 0.0611, + "step": 30701 + }, + { + "epoch": 0.7197032296629687, + "grad_norm": 0.3297937214374542, + "learning_rate": 3.634436641774297e-05, + "loss": 0.0453, + "step": 30702 + }, + { + "epoch": 0.7197266712377738, + "grad_norm": 0.6207306385040283, + "learning_rate": 3.6338685587021626e-05, + "loss": 0.1112, + "step": 30703 + }, + { + "epoch": 0.7197501128125787, + "grad_norm": 0.6355524063110352, + "learning_rate": 3.633300510172838e-05, + "loss": 0.4801, + "step": 30704 + }, + { + "epoch": 0.7197735543873838, + "grad_norm": 0.3548431396484375, + "learning_rate": 3.632732496189403e-05, + "loss": 0.3257, + "step": 30705 + }, + { + "epoch": 0.7197969959621887, + "grad_norm": 0.15396277606487274, + "learning_rate": 3.632164516754937e-05, + "loss": 0.0362, + "step": 30706 + }, + { + "epoch": 0.7198204375369938, + "grad_norm": 0.28218260407447815, + "learning_rate": 3.6315965718725284e-05, + "loss": 0.222, + "step": 30707 + }, + { + "epoch": 0.7198438791117987, + "grad_norm": 0.4868735373020172, + "learning_rate": 3.631028661545256e-05, + "loss": 0.0858, + "step": 30708 + }, + { + "epoch": 0.7198673206866038, + "grad_norm": 0.3863790035247803, + "learning_rate": 3.630460785776202e-05, + "loss": 0.4599, + "step": 30709 + }, + { + "epoch": 0.7198907622614087, + "grad_norm": 0.17592161893844604, + "learning_rate": 3.629892944568446e-05, + "loss": 0.0297, + "step": 30710 + }, + { + "epoch": 0.7199142038362137, + "grad_norm": 0.2827954590320587, + "learning_rate": 3.629325137925067e-05, + "loss": 0.0298, + "step": 30711 + }, + { + "epoch": 0.7199376454110187, + "grad_norm": 0.13474340736865997, + "learning_rate": 3.6287573658491525e-05, + "loss": 0.0272, + "step": 30712 + }, + { + "epoch": 0.7199610869858237, + "grad_norm": 0.3592105209827423, + "learning_rate": 3.628189628343779e-05, + "loss": 0.0895, + "step": 30713 + }, + { + "epoch": 0.7199845285606287, + "grad_norm": 0.395834356546402, + "learning_rate": 3.627621925412028e-05, + "loss": 0.0234, + "step": 30714 + }, + { + "epoch": 0.7200079701354337, + "grad_norm": 0.7324946522712708, + "learning_rate": 3.627054257056977e-05, + "loss": 0.0799, + "step": 30715 + }, + { + "epoch": 0.7200314117102387, + "grad_norm": 0.4889356195926666, + "learning_rate": 3.626486623281712e-05, + "loss": 0.1054, + "step": 30716 + }, + { + "epoch": 0.7200548532850437, + "grad_norm": 0.34552499651908875, + "learning_rate": 3.625919024089309e-05, + "loss": 0.0507, + "step": 30717 + }, + { + "epoch": 0.7200782948598486, + "grad_norm": 0.30331623554229736, + "learning_rate": 3.625351459482845e-05, + "loss": 0.039, + "step": 30718 + }, + { + "epoch": 0.7201017364346537, + "grad_norm": 0.28296229243278503, + "learning_rate": 3.6247839294654074e-05, + "loss": 0.0464, + "step": 30719 + }, + { + "epoch": 0.7201251780094586, + "grad_norm": 0.10432786494493484, + "learning_rate": 3.624216434040072e-05, + "loss": 0.0188, + "step": 30720 + }, + { + "epoch": 0.7201486195842637, + "grad_norm": 0.2820586860179901, + "learning_rate": 3.623648973209916e-05, + "loss": 0.039, + "step": 30721 + }, + { + "epoch": 0.7201720611590686, + "grad_norm": 0.5113007426261902, + "learning_rate": 3.6230815469780214e-05, + "loss": 0.0801, + "step": 30722 + }, + { + "epoch": 0.7201955027338737, + "grad_norm": 0.15947148203849792, + "learning_rate": 3.622514155347464e-05, + "loss": 0.0339, + "step": 30723 + }, + { + "epoch": 0.7202189443086786, + "grad_norm": 0.4482424855232239, + "learning_rate": 3.621946798321322e-05, + "loss": 0.0901, + "step": 30724 + }, + { + "epoch": 0.7202423858834837, + "grad_norm": 0.5159710049629211, + "learning_rate": 3.6213794759026785e-05, + "loss": 0.4419, + "step": 30725 + }, + { + "epoch": 0.7202658274582886, + "grad_norm": 0.5035274028778076, + "learning_rate": 3.6208121880946114e-05, + "loss": 0.0849, + "step": 30726 + }, + { + "epoch": 0.7202892690330936, + "grad_norm": 0.38982057571411133, + "learning_rate": 3.620244934900191e-05, + "loss": 0.0772, + "step": 30727 + }, + { + "epoch": 0.7203127106078986, + "grad_norm": 0.11739516258239746, + "learning_rate": 3.619677716322505e-05, + "loss": 0.016, + "step": 30728 + }, + { + "epoch": 0.7203361521827036, + "grad_norm": 0.33855873346328735, + "learning_rate": 3.619110532364628e-05, + "loss": 0.0576, + "step": 30729 + }, + { + "epoch": 0.7203595937575086, + "grad_norm": 0.5922478437423706, + "learning_rate": 3.618543383029635e-05, + "loss": 0.1447, + "step": 30730 + }, + { + "epoch": 0.7203830353323136, + "grad_norm": 0.6251793503761292, + "learning_rate": 3.617976268320603e-05, + "loss": 0.1065, + "step": 30731 + }, + { + "epoch": 0.7204064769071186, + "grad_norm": 0.6387127637863159, + "learning_rate": 3.6174091882406146e-05, + "loss": 0.0846, + "step": 30732 + }, + { + "epoch": 0.7204299184819236, + "grad_norm": 0.48764628171920776, + "learning_rate": 3.616842142792742e-05, + "loss": 0.4137, + "step": 30733 + }, + { + "epoch": 0.7204533600567287, + "grad_norm": 0.3508889675140381, + "learning_rate": 3.616275131980065e-05, + "loss": 0.0658, + "step": 30734 + }, + { + "epoch": 0.7204768016315336, + "grad_norm": 0.4151337146759033, + "learning_rate": 3.615708155805658e-05, + "loss": 0.049, + "step": 30735 + }, + { + "epoch": 0.7205002432063387, + "grad_norm": 0.12705287337303162, + "learning_rate": 3.615141214272598e-05, + "loss": 0.0138, + "step": 30736 + }, + { + "epoch": 0.7205236847811436, + "grad_norm": 0.3149087131023407, + "learning_rate": 3.6145743073839575e-05, + "loss": 0.0306, + "step": 30737 + }, + { + "epoch": 0.7205471263559486, + "grad_norm": 0.9704195261001587, + "learning_rate": 3.614007435142821e-05, + "loss": 0.1279, + "step": 30738 + }, + { + "epoch": 0.7205705679307536, + "grad_norm": 0.29188570380210876, + "learning_rate": 3.613440597552256e-05, + "loss": 0.0509, + "step": 30739 + }, + { + "epoch": 0.7205940095055586, + "grad_norm": 0.1741762012243271, + "learning_rate": 3.612873794615345e-05, + "loss": 0.0311, + "step": 30740 + }, + { + "epoch": 0.7206174510803636, + "grad_norm": 0.5470182299613953, + "learning_rate": 3.61230702633516e-05, + "loss": 0.1135, + "step": 30741 + }, + { + "epoch": 0.7206408926551686, + "grad_norm": 0.08834680914878845, + "learning_rate": 3.611740292714777e-05, + "loss": 0.0106, + "step": 30742 + }, + { + "epoch": 0.7206643342299736, + "grad_norm": 0.4614849388599396, + "learning_rate": 3.6111735937572696e-05, + "loss": 0.1177, + "step": 30743 + }, + { + "epoch": 0.7206877758047786, + "grad_norm": 0.5396619439125061, + "learning_rate": 3.610606929465712e-05, + "loss": 0.0795, + "step": 30744 + }, + { + "epoch": 0.7207112173795835, + "grad_norm": 0.38301852345466614, + "learning_rate": 3.6100402998431836e-05, + "loss": 0.068, + "step": 30745 + }, + { + "epoch": 0.7207346589543886, + "grad_norm": 0.2373145967721939, + "learning_rate": 3.609473704892757e-05, + "loss": 0.0364, + "step": 30746 + }, + { + "epoch": 0.7207581005291935, + "grad_norm": 0.18937760591506958, + "learning_rate": 3.6089071446175046e-05, + "loss": 0.0213, + "step": 30747 + }, + { + "epoch": 0.7207815421039986, + "grad_norm": 0.6899057030677795, + "learning_rate": 3.608340619020503e-05, + "loss": 0.1019, + "step": 30748 + }, + { + "epoch": 0.7208049836788035, + "grad_norm": 0.35510364174842834, + "learning_rate": 3.6077741281048216e-05, + "loss": 0.0707, + "step": 30749 + }, + { + "epoch": 0.7208284252536086, + "grad_norm": 0.41066721081733704, + "learning_rate": 3.607207671873539e-05, + "loss": 0.071, + "step": 30750 + }, + { + "epoch": 0.7208518668284135, + "grad_norm": 0.17033126950263977, + "learning_rate": 3.606641250329725e-05, + "loss": 0.0142, + "step": 30751 + }, + { + "epoch": 0.7208753084032186, + "grad_norm": 0.1140613853931427, + "learning_rate": 3.6060748634764597e-05, + "loss": 0.0256, + "step": 30752 + }, + { + "epoch": 0.7208987499780235, + "grad_norm": 0.5355569124221802, + "learning_rate": 3.605508511316811e-05, + "loss": 0.0977, + "step": 30753 + }, + { + "epoch": 0.7209221915528286, + "grad_norm": 0.18089424073696136, + "learning_rate": 3.604942193853853e-05, + "loss": 0.0279, + "step": 30754 + }, + { + "epoch": 0.7209456331276335, + "grad_norm": 0.3880893588066101, + "learning_rate": 3.6043759110906594e-05, + "loss": 0.3359, + "step": 30755 + }, + { + "epoch": 0.7209690747024385, + "grad_norm": 0.24110457301139832, + "learning_rate": 3.603809663030298e-05, + "loss": 0.0344, + "step": 30756 + }, + { + "epoch": 0.7209925162772435, + "grad_norm": 0.33953630924224854, + "learning_rate": 3.6032434496758475e-05, + "loss": 0.3401, + "step": 30757 + }, + { + "epoch": 0.7210159578520485, + "grad_norm": 0.49369409680366516, + "learning_rate": 3.602677271030379e-05, + "loss": 0.0831, + "step": 30758 + }, + { + "epoch": 0.7210393994268535, + "grad_norm": 0.8319414854049683, + "learning_rate": 3.602111127096963e-05, + "loss": 0.1249, + "step": 30759 + }, + { + "epoch": 0.7210628410016585, + "grad_norm": 0.396770179271698, + "learning_rate": 3.6015450178786725e-05, + "loss": 0.0614, + "step": 30760 + }, + { + "epoch": 0.7210862825764635, + "grad_norm": 0.25719472765922546, + "learning_rate": 3.600978943378578e-05, + "loss": 0.0483, + "step": 30761 + }, + { + "epoch": 0.7211097241512685, + "grad_norm": 0.3817456364631653, + "learning_rate": 3.600412903599748e-05, + "loss": 0.0496, + "step": 30762 + }, + { + "epoch": 0.7211331657260734, + "grad_norm": 0.3310115933418274, + "learning_rate": 3.599846898545258e-05, + "loss": 0.0546, + "step": 30763 + }, + { + "epoch": 0.7211566073008785, + "grad_norm": 0.45305708050727844, + "learning_rate": 3.5992809282181827e-05, + "loss": 0.059, + "step": 30764 + }, + { + "epoch": 0.7211800488756834, + "grad_norm": 0.6923176646232605, + "learning_rate": 3.598714992621588e-05, + "loss": 0.1053, + "step": 30765 + }, + { + "epoch": 0.7212034904504885, + "grad_norm": 0.37517550587654114, + "learning_rate": 3.598149091758547e-05, + "loss": 0.058, + "step": 30766 + }, + { + "epoch": 0.7212269320252934, + "grad_norm": 0.3437325656414032, + "learning_rate": 3.597583225632127e-05, + "loss": 0.0685, + "step": 30767 + }, + { + "epoch": 0.7212503736000985, + "grad_norm": 0.47056931257247925, + "learning_rate": 3.5970173942454024e-05, + "loss": 0.0783, + "step": 30768 + }, + { + "epoch": 0.7212738151749034, + "grad_norm": 0.45816338062286377, + "learning_rate": 3.596451597601438e-05, + "loss": 0.0944, + "step": 30769 + }, + { + "epoch": 0.7212972567497085, + "grad_norm": 0.37411361932754517, + "learning_rate": 3.59588583570331e-05, + "loss": 0.0914, + "step": 30770 + }, + { + "epoch": 0.7213206983245134, + "grad_norm": 0.3755223751068115, + "learning_rate": 3.595320108554086e-05, + "loss": 0.0667, + "step": 30771 + }, + { + "epoch": 0.7213441398993184, + "grad_norm": 0.5809519290924072, + "learning_rate": 3.594754416156834e-05, + "loss": 0.0764, + "step": 30772 + }, + { + "epoch": 0.7213675814741234, + "grad_norm": 0.6136699318885803, + "learning_rate": 3.594188758514625e-05, + "loss": 0.145, + "step": 30773 + }, + { + "epoch": 0.7213910230489284, + "grad_norm": 0.3957189917564392, + "learning_rate": 3.5936231356305285e-05, + "loss": 0.0713, + "step": 30774 + }, + { + "epoch": 0.7214144646237334, + "grad_norm": 0.3467644453048706, + "learning_rate": 3.593057547507609e-05, + "loss": 0.0763, + "step": 30775 + }, + { + "epoch": 0.7214379061985384, + "grad_norm": 0.5253431797027588, + "learning_rate": 3.59249199414894e-05, + "loss": 0.0656, + "step": 30776 + }, + { + "epoch": 0.7214613477733434, + "grad_norm": 0.35718321800231934, + "learning_rate": 3.5919264755575935e-05, + "loss": 0.0618, + "step": 30777 + }, + { + "epoch": 0.7214847893481484, + "grad_norm": 0.7409364581108093, + "learning_rate": 3.5913609917366334e-05, + "loss": 0.1617, + "step": 30778 + }, + { + "epoch": 0.7215082309229534, + "grad_norm": 0.5367249846458435, + "learning_rate": 3.590795542689128e-05, + "loss": 0.0922, + "step": 30779 + }, + { + "epoch": 0.7215316724977584, + "grad_norm": 0.5101888179779053, + "learning_rate": 3.590230128418147e-05, + "loss": 0.0666, + "step": 30780 + }, + { + "epoch": 0.7215551140725633, + "grad_norm": 1.3767759799957275, + "learning_rate": 3.589664748926754e-05, + "loss": 0.0809, + "step": 30781 + }, + { + "epoch": 0.7215785556473684, + "grad_norm": 0.1005244106054306, + "learning_rate": 3.589099404218024e-05, + "loss": 0.0182, + "step": 30782 + }, + { + "epoch": 0.7216019972221733, + "grad_norm": 0.509421706199646, + "learning_rate": 3.588534094295021e-05, + "loss": 0.0609, + "step": 30783 + }, + { + "epoch": 0.7216254387969784, + "grad_norm": 1.095951795578003, + "learning_rate": 3.587968819160812e-05, + "loss": 0.0726, + "step": 30784 + }, + { + "epoch": 0.7216488803717834, + "grad_norm": 0.15189889073371887, + "learning_rate": 3.587403578818465e-05, + "loss": 0.0274, + "step": 30785 + }, + { + "epoch": 0.7216723219465884, + "grad_norm": 0.5935319066047668, + "learning_rate": 3.586838373271045e-05, + "loss": 0.132, + "step": 30786 + }, + { + "epoch": 0.7216957635213934, + "grad_norm": 0.4405374228954315, + "learning_rate": 3.586273202521618e-05, + "loss": 0.0901, + "step": 30787 + }, + { + "epoch": 0.7217192050961984, + "grad_norm": 0.7526108622550964, + "learning_rate": 3.5857080665732536e-05, + "loss": 0.0956, + "step": 30788 + }, + { + "epoch": 0.7217426466710034, + "grad_norm": 0.32283177971839905, + "learning_rate": 3.585142965429019e-05, + "loss": 0.0666, + "step": 30789 + }, + { + "epoch": 0.7217660882458083, + "grad_norm": 0.214606374502182, + "learning_rate": 3.584577899091981e-05, + "loss": 0.0363, + "step": 30790 + }, + { + "epoch": 0.7217895298206134, + "grad_norm": 0.45873498916625977, + "learning_rate": 3.584012867565203e-05, + "loss": 0.5962, + "step": 30791 + }, + { + "epoch": 0.7218129713954183, + "grad_norm": 0.5377060770988464, + "learning_rate": 3.583447870851752e-05, + "loss": 0.0902, + "step": 30792 + }, + { + "epoch": 0.7218364129702234, + "grad_norm": 0.3350474536418915, + "learning_rate": 3.5828829089546925e-05, + "loss": 0.0329, + "step": 30793 + }, + { + "epoch": 0.7218598545450283, + "grad_norm": 0.12926654517650604, + "learning_rate": 3.582317981877087e-05, + "loss": 0.0214, + "step": 30794 + }, + { + "epoch": 0.7218832961198334, + "grad_norm": 0.10914143174886703, + "learning_rate": 3.5817530896220085e-05, + "loss": 0.0117, + "step": 30795 + }, + { + "epoch": 0.7219067376946383, + "grad_norm": 0.7108084559440613, + "learning_rate": 3.581188232192519e-05, + "loss": 0.1081, + "step": 30796 + }, + { + "epoch": 0.7219301792694434, + "grad_norm": 1.1119705438613892, + "learning_rate": 3.580623409591682e-05, + "loss": 0.0982, + "step": 30797 + }, + { + "epoch": 0.7219536208442483, + "grad_norm": 1.1817641258239746, + "learning_rate": 3.580058621822563e-05, + "loss": 0.527, + "step": 30798 + }, + { + "epoch": 0.7219770624190534, + "grad_norm": 0.4216926693916321, + "learning_rate": 3.579493868888226e-05, + "loss": 0.0667, + "step": 30799 + }, + { + "epoch": 0.7220005039938583, + "grad_norm": 0.4790501892566681, + "learning_rate": 3.5789291507917325e-05, + "loss": 0.0977, + "step": 30800 + }, + { + "epoch": 0.7220239455686633, + "grad_norm": 0.13901202380657196, + "learning_rate": 3.578364467536151e-05, + "loss": 0.0378, + "step": 30801 + }, + { + "epoch": 0.7220473871434683, + "grad_norm": 0.23645108938217163, + "learning_rate": 3.577799819124547e-05, + "loss": 0.0411, + "step": 30802 + }, + { + "epoch": 0.7220708287182733, + "grad_norm": 0.05493483319878578, + "learning_rate": 3.577235205559981e-05, + "loss": 0.0077, + "step": 30803 + }, + { + "epoch": 0.7220942702930783, + "grad_norm": 0.3490878641605377, + "learning_rate": 3.576670626845519e-05, + "loss": 0.0579, + "step": 30804 + }, + { + "epoch": 0.7221177118678833, + "grad_norm": 0.4680781960487366, + "learning_rate": 3.5761060829842206e-05, + "loss": 0.0693, + "step": 30805 + }, + { + "epoch": 0.7221411534426883, + "grad_norm": 0.42134517431259155, + "learning_rate": 3.575541573979153e-05, + "loss": 0.0645, + "step": 30806 + }, + { + "epoch": 0.7221645950174933, + "grad_norm": 0.5536753535270691, + "learning_rate": 3.5749770998333746e-05, + "loss": 0.0934, + "step": 30807 + }, + { + "epoch": 0.7221880365922982, + "grad_norm": 0.3796239197254181, + "learning_rate": 3.574412660549953e-05, + "loss": 0.0882, + "step": 30808 + }, + { + "epoch": 0.7222114781671033, + "grad_norm": 0.40447625517845154, + "learning_rate": 3.57384825613195e-05, + "loss": 0.1047, + "step": 30809 + }, + { + "epoch": 0.7222349197419082, + "grad_norm": 0.10592947900295258, + "learning_rate": 3.573283886582426e-05, + "loss": 0.0175, + "step": 30810 + }, + { + "epoch": 0.7222583613167133, + "grad_norm": 0.5410692095756531, + "learning_rate": 3.572719551904444e-05, + "loss": 0.4682, + "step": 30811 + }, + { + "epoch": 0.7222818028915182, + "grad_norm": 0.2311287820339203, + "learning_rate": 3.5721552521010646e-05, + "loss": 0.0309, + "step": 30812 + }, + { + "epoch": 0.7223052444663233, + "grad_norm": 0.29670417308807373, + "learning_rate": 3.571590987175354e-05, + "loss": 0.064, + "step": 30813 + }, + { + "epoch": 0.7223286860411282, + "grad_norm": 0.26503685116767883, + "learning_rate": 3.571026757130368e-05, + "loss": 0.0263, + "step": 30814 + }, + { + "epoch": 0.7223521276159333, + "grad_norm": 0.3294391632080078, + "learning_rate": 3.5704625619691756e-05, + "loss": 0.0481, + "step": 30815 + }, + { + "epoch": 0.7223755691907382, + "grad_norm": 0.08834054321050644, + "learning_rate": 3.5698984016948344e-05, + "loss": 0.0085, + "step": 30816 + }, + { + "epoch": 0.7223990107655432, + "grad_norm": 0.08826707303524017, + "learning_rate": 3.5693342763104045e-05, + "loss": 0.0106, + "step": 30817 + }, + { + "epoch": 0.7224224523403482, + "grad_norm": 0.2258082628250122, + "learning_rate": 3.568770185818947e-05, + "loss": 0.0638, + "step": 30818 + }, + { + "epoch": 0.7224458939151532, + "grad_norm": 0.28137773275375366, + "learning_rate": 3.5682061302235215e-05, + "loss": 0.0547, + "step": 30819 + }, + { + "epoch": 0.7224693354899582, + "grad_norm": 0.2563057839870453, + "learning_rate": 3.567642109527193e-05, + "loss": 0.0304, + "step": 30820 + }, + { + "epoch": 0.7224927770647632, + "grad_norm": 0.46381521224975586, + "learning_rate": 3.567078123733019e-05, + "loss": 0.0982, + "step": 30821 + }, + { + "epoch": 0.7225162186395682, + "grad_norm": 0.32474273443222046, + "learning_rate": 3.56651417284406e-05, + "loss": 0.0618, + "step": 30822 + }, + { + "epoch": 0.7225396602143732, + "grad_norm": 0.3739745616912842, + "learning_rate": 3.565950256863376e-05, + "loss": 0.0967, + "step": 30823 + }, + { + "epoch": 0.7225631017891782, + "grad_norm": 0.6728315353393555, + "learning_rate": 3.565386375794024e-05, + "loss": 0.133, + "step": 30824 + }, + { + "epoch": 0.7225865433639832, + "grad_norm": 0.12813030183315277, + "learning_rate": 3.564822529639069e-05, + "loss": 0.0202, + "step": 30825 + }, + { + "epoch": 0.7226099849387881, + "grad_norm": 0.27729156613349915, + "learning_rate": 3.564258718401565e-05, + "loss": 0.0391, + "step": 30826 + }, + { + "epoch": 0.7226334265135932, + "grad_norm": 0.3214815855026245, + "learning_rate": 3.563694942084578e-05, + "loss": 0.0773, + "step": 30827 + }, + { + "epoch": 0.7226568680883981, + "grad_norm": 0.4123552441596985, + "learning_rate": 3.563131200691163e-05, + "loss": 0.0663, + "step": 30828 + }, + { + "epoch": 0.7226803096632032, + "grad_norm": 0.40912553668022156, + "learning_rate": 3.562567494224379e-05, + "loss": 0.0962, + "step": 30829 + }, + { + "epoch": 0.7227037512380081, + "grad_norm": 0.1527855098247528, + "learning_rate": 3.562003822687284e-05, + "loss": 0.0245, + "step": 30830 + }, + { + "epoch": 0.7227271928128132, + "grad_norm": 0.3103848695755005, + "learning_rate": 3.5614401860829384e-05, + "loss": 0.014, + "step": 30831 + }, + { + "epoch": 0.7227506343876181, + "grad_norm": 0.5943885445594788, + "learning_rate": 3.5608765844143956e-05, + "loss": 0.1139, + "step": 30832 + }, + { + "epoch": 0.7227740759624232, + "grad_norm": 0.4753250777721405, + "learning_rate": 3.560313017684721e-05, + "loss": 0.053, + "step": 30833 + }, + { + "epoch": 0.7227975175372281, + "grad_norm": 0.18788737058639526, + "learning_rate": 3.5597494858969684e-05, + "loss": 0.0155, + "step": 30834 + }, + { + "epoch": 0.7228209591120331, + "grad_norm": 0.1649436056613922, + "learning_rate": 3.559185989054197e-05, + "loss": 0.0231, + "step": 30835 + }, + { + "epoch": 0.7228444006868382, + "grad_norm": 0.49157389998435974, + "learning_rate": 3.55862252715946e-05, + "loss": 0.0776, + "step": 30836 + }, + { + "epoch": 0.7228678422616431, + "grad_norm": 0.5000529885292053, + "learning_rate": 3.558059100215822e-05, + "loss": 0.0582, + "step": 30837 + }, + { + "epoch": 0.7228912838364482, + "grad_norm": 0.24530355632305145, + "learning_rate": 3.557495708226335e-05, + "loss": 0.0536, + "step": 30838 + }, + { + "epoch": 0.7229147254112531, + "grad_norm": 0.243129700422287, + "learning_rate": 3.556932351194054e-05, + "loss": 0.0208, + "step": 30839 + }, + { + "epoch": 0.7229381669860582, + "grad_norm": 0.6113969087600708, + "learning_rate": 3.556369029122044e-05, + "loss": 0.0694, + "step": 30840 + }, + { + "epoch": 0.7229616085608631, + "grad_norm": 0.1988603174686432, + "learning_rate": 3.555805742013356e-05, + "loss": 0.0265, + "step": 30841 + }, + { + "epoch": 0.7229850501356682, + "grad_norm": 0.44877487421035767, + "learning_rate": 3.5552424898710476e-05, + "loss": 0.0924, + "step": 30842 + }, + { + "epoch": 0.7230084917104731, + "grad_norm": 0.376099169254303, + "learning_rate": 3.554679272698175e-05, + "loss": 0.238, + "step": 30843 + }, + { + "epoch": 0.7230319332852781, + "grad_norm": 0.45566269755363464, + "learning_rate": 3.5541160904977944e-05, + "loss": 0.0778, + "step": 30844 + }, + { + "epoch": 0.7230553748600831, + "grad_norm": 0.4349779486656189, + "learning_rate": 3.553552943272956e-05, + "loss": 0.0536, + "step": 30845 + }, + { + "epoch": 0.7230788164348881, + "grad_norm": 0.3391857445240021, + "learning_rate": 3.552989831026725e-05, + "loss": 0.0443, + "step": 30846 + }, + { + "epoch": 0.7231022580096931, + "grad_norm": 0.20917963981628418, + "learning_rate": 3.5524267537621527e-05, + "loss": 0.0775, + "step": 30847 + }, + { + "epoch": 0.7231256995844981, + "grad_norm": 0.3850719928741455, + "learning_rate": 3.551863711482291e-05, + "loss": 0.0862, + "step": 30848 + }, + { + "epoch": 0.7231491411593031, + "grad_norm": 0.17262527346611023, + "learning_rate": 3.5513007041902004e-05, + "loss": 0.0243, + "step": 30849 + }, + { + "epoch": 0.7231725827341081, + "grad_norm": 0.5711292028427124, + "learning_rate": 3.550737731888933e-05, + "loss": 0.1165, + "step": 30850 + }, + { + "epoch": 0.723196024308913, + "grad_norm": 0.3968559801578522, + "learning_rate": 3.5501747945815455e-05, + "loss": 0.0707, + "step": 30851 + }, + { + "epoch": 0.7232194658837181, + "grad_norm": 0.5394582152366638, + "learning_rate": 3.549611892271087e-05, + "loss": 0.4118, + "step": 30852 + }, + { + "epoch": 0.723242907458523, + "grad_norm": 0.3783855438232422, + "learning_rate": 3.549049024960619e-05, + "loss": 0.0615, + "step": 30853 + }, + { + "epoch": 0.7232663490333281, + "grad_norm": 0.16611577570438385, + "learning_rate": 3.548486192653193e-05, + "loss": 0.0276, + "step": 30854 + }, + { + "epoch": 0.723289790608133, + "grad_norm": 0.44763660430908203, + "learning_rate": 3.5479233953518624e-05, + "loss": 0.0792, + "step": 30855 + }, + { + "epoch": 0.7233132321829381, + "grad_norm": 0.4236777424812317, + "learning_rate": 3.54736063305968e-05, + "loss": 0.0397, + "step": 30856 + }, + { + "epoch": 0.723336673757743, + "grad_norm": 0.41089868545532227, + "learning_rate": 3.5467979057796964e-05, + "loss": 0.3846, + "step": 30857 + }, + { + "epoch": 0.7233601153325481, + "grad_norm": 0.46993058919906616, + "learning_rate": 3.5462352135149734e-05, + "loss": 0.0552, + "step": 30858 + }, + { + "epoch": 0.723383556907353, + "grad_norm": 0.5040888786315918, + "learning_rate": 3.545672556268559e-05, + "loss": 0.0687, + "step": 30859 + }, + { + "epoch": 0.7234069984821581, + "grad_norm": 0.44826340675354004, + "learning_rate": 3.5451099340435065e-05, + "loss": 0.0668, + "step": 30860 + }, + { + "epoch": 0.723430440056963, + "grad_norm": 0.4699088931083679, + "learning_rate": 3.544547346842866e-05, + "loss": 0.1161, + "step": 30861 + }, + { + "epoch": 0.723453881631768, + "grad_norm": 0.7577853798866272, + "learning_rate": 3.5439847946696966e-05, + "loss": 0.0694, + "step": 30862 + }, + { + "epoch": 0.723477323206573, + "grad_norm": 0.49332481622695923, + "learning_rate": 3.543422277527047e-05, + "loss": 0.0875, + "step": 30863 + }, + { + "epoch": 0.723500764781378, + "grad_norm": 0.3581642508506775, + "learning_rate": 3.542859795417967e-05, + "loss": 0.3598, + "step": 30864 + }, + { + "epoch": 0.723524206356183, + "grad_norm": 0.4405156970024109, + "learning_rate": 3.5422973483455136e-05, + "loss": 0.5205, + "step": 30865 + }, + { + "epoch": 0.723547647930988, + "grad_norm": 0.38673537969589233, + "learning_rate": 3.541734936312737e-05, + "loss": 0.0275, + "step": 30866 + }, + { + "epoch": 0.723571089505793, + "grad_norm": 0.5743582248687744, + "learning_rate": 3.5411725593226884e-05, + "loss": 0.0676, + "step": 30867 + }, + { + "epoch": 0.723594531080598, + "grad_norm": 0.2581048905849457, + "learning_rate": 3.540610217378418e-05, + "loss": 0.2119, + "step": 30868 + }, + { + "epoch": 0.723617972655403, + "grad_norm": 0.12640026211738586, + "learning_rate": 3.54004791048298e-05, + "loss": 0.0324, + "step": 30869 + }, + { + "epoch": 0.723641414230208, + "grad_norm": 0.41091859340667725, + "learning_rate": 3.539485638639419e-05, + "loss": 0.0495, + "step": 30870 + }, + { + "epoch": 0.7236648558050129, + "grad_norm": 1.2746714353561401, + "learning_rate": 3.5389234018507944e-05, + "loss": 0.1084, + "step": 30871 + }, + { + "epoch": 0.723688297379818, + "grad_norm": 0.25000786781311035, + "learning_rate": 3.538361200120153e-05, + "loss": 0.0531, + "step": 30872 + }, + { + "epoch": 0.7237117389546229, + "grad_norm": 0.6535929441452026, + "learning_rate": 3.537799033450542e-05, + "loss": 0.3886, + "step": 30873 + }, + { + "epoch": 0.723735180529428, + "grad_norm": 0.14675748348236084, + "learning_rate": 3.537236901845018e-05, + "loss": 0.0141, + "step": 30874 + }, + { + "epoch": 0.7237586221042329, + "grad_norm": 0.1339161992073059, + "learning_rate": 3.536674805306629e-05, + "loss": 0.0306, + "step": 30875 + }, + { + "epoch": 0.723782063679038, + "grad_norm": 0.23222662508487701, + "learning_rate": 3.536112743838424e-05, + "loss": 0.0535, + "step": 30876 + }, + { + "epoch": 0.7238055052538429, + "grad_norm": 0.4370507001876831, + "learning_rate": 3.53555071744345e-05, + "loss": 0.0377, + "step": 30877 + }, + { + "epoch": 0.723828946828648, + "grad_norm": 0.7403219938278198, + "learning_rate": 3.534988726124763e-05, + "loss": 0.0783, + "step": 30878 + }, + { + "epoch": 0.7238523884034529, + "grad_norm": 0.41441911458969116, + "learning_rate": 3.5344267698854084e-05, + "loss": 0.0606, + "step": 30879 + }, + { + "epoch": 0.723875829978258, + "grad_norm": 0.5496088266372681, + "learning_rate": 3.533864848728435e-05, + "loss": 0.3982, + "step": 30880 + }, + { + "epoch": 0.7238992715530629, + "grad_norm": 0.3497782349586487, + "learning_rate": 3.533302962656895e-05, + "loss": 0.0465, + "step": 30881 + }, + { + "epoch": 0.7239227131278679, + "grad_norm": 0.6500545740127563, + "learning_rate": 3.532741111673833e-05, + "loss": 0.0804, + "step": 30882 + }, + { + "epoch": 0.7239461547026729, + "grad_norm": 0.6774952411651611, + "learning_rate": 3.5321792957822976e-05, + "loss": 0.1305, + "step": 30883 + }, + { + "epoch": 0.7239695962774779, + "grad_norm": 0.8787769079208374, + "learning_rate": 3.5316175149853414e-05, + "loss": 0.1392, + "step": 30884 + }, + { + "epoch": 0.7239930378522829, + "grad_norm": 0.2357856035232544, + "learning_rate": 3.531055769286008e-05, + "loss": 0.0329, + "step": 30885 + }, + { + "epoch": 0.7240164794270879, + "grad_norm": 0.5009635090827942, + "learning_rate": 3.5304940586873504e-05, + "loss": 0.0873, + "step": 30886 + }, + { + "epoch": 0.724039921001893, + "grad_norm": 0.4463862478733063, + "learning_rate": 3.529932383192415e-05, + "loss": 0.0917, + "step": 30887 + }, + { + "epoch": 0.7240633625766979, + "grad_norm": 0.10702614486217499, + "learning_rate": 3.529370742804248e-05, + "loss": 0.0115, + "step": 30888 + }, + { + "epoch": 0.724086804151503, + "grad_norm": 0.36879050731658936, + "learning_rate": 3.528809137525897e-05, + "loss": 0.1197, + "step": 30889 + }, + { + "epoch": 0.7241102457263079, + "grad_norm": 0.5421666502952576, + "learning_rate": 3.528247567360407e-05, + "loss": 0.0924, + "step": 30890 + }, + { + "epoch": 0.7241336873011129, + "grad_norm": 0.1443052589893341, + "learning_rate": 3.52768603231083e-05, + "loss": 0.0291, + "step": 30891 + }, + { + "epoch": 0.7241571288759179, + "grad_norm": 0.5702922344207764, + "learning_rate": 3.527124532380211e-05, + "loss": 0.0803, + "step": 30892 + }, + { + "epoch": 0.7241805704507229, + "grad_norm": 0.23035059869289398, + "learning_rate": 3.526563067571595e-05, + "loss": 0.0289, + "step": 30893 + }, + { + "epoch": 0.7242040120255279, + "grad_norm": 0.25484272837638855, + "learning_rate": 3.526001637888031e-05, + "loss": 0.0359, + "step": 30894 + }, + { + "epoch": 0.7242274536003329, + "grad_norm": 0.5231600403785706, + "learning_rate": 3.5254402433325614e-05, + "loss": 0.589, + "step": 30895 + }, + { + "epoch": 0.7242508951751379, + "grad_norm": 0.42453473806381226, + "learning_rate": 3.524878883908238e-05, + "loss": 0.0677, + "step": 30896 + }, + { + "epoch": 0.7242743367499429, + "grad_norm": 0.1869434267282486, + "learning_rate": 3.5243175596181e-05, + "loss": 0.0194, + "step": 30897 + }, + { + "epoch": 0.7242977783247478, + "grad_norm": 0.5732495784759521, + "learning_rate": 3.5237562704652014e-05, + "loss": 0.1131, + "step": 30898 + }, + { + "epoch": 0.7243212198995529, + "grad_norm": 0.08399143069982529, + "learning_rate": 3.523195016452583e-05, + "loss": 0.0147, + "step": 30899 + }, + { + "epoch": 0.7243446614743578, + "grad_norm": 0.48166996240615845, + "learning_rate": 3.52263379758329e-05, + "loss": 0.066, + "step": 30900 + }, + { + "epoch": 0.7243681030491629, + "grad_norm": 0.6257413625717163, + "learning_rate": 3.522072613860368e-05, + "loss": 0.1102, + "step": 30901 + }, + { + "epoch": 0.7243915446239678, + "grad_norm": 0.41237202286720276, + "learning_rate": 3.52151146528686e-05, + "loss": 0.0592, + "step": 30902 + }, + { + "epoch": 0.7244149861987729, + "grad_norm": 0.21509617567062378, + "learning_rate": 3.520950351865815e-05, + "loss": 0.0345, + "step": 30903 + }, + { + "epoch": 0.7244384277735778, + "grad_norm": 0.9894478917121887, + "learning_rate": 3.520389273600275e-05, + "loss": 0.1108, + "step": 30904 + }, + { + "epoch": 0.7244618693483829, + "grad_norm": 0.15360327064990997, + "learning_rate": 3.519828230493287e-05, + "loss": 0.0339, + "step": 30905 + }, + { + "epoch": 0.7244853109231878, + "grad_norm": 0.453159898519516, + "learning_rate": 3.519267222547892e-05, + "loss": 0.0666, + "step": 30906 + }, + { + "epoch": 0.7245087524979928, + "grad_norm": 0.20890571177005768, + "learning_rate": 3.518706249767135e-05, + "loss": 0.0283, + "step": 30907 + }, + { + "epoch": 0.7245321940727978, + "grad_norm": 0.18853458762168884, + "learning_rate": 3.5181453121540575e-05, + "loss": 0.1273, + "step": 30908 + }, + { + "epoch": 0.7245556356476028, + "grad_norm": 0.4908795654773712, + "learning_rate": 3.517584409711706e-05, + "loss": 0.0769, + "step": 30909 + }, + { + "epoch": 0.7245790772224078, + "grad_norm": 0.3166494369506836, + "learning_rate": 3.517023542443126e-05, + "loss": 0.0331, + "step": 30910 + }, + { + "epoch": 0.7246025187972128, + "grad_norm": 0.3826662003993988, + "learning_rate": 3.5164627103513605e-05, + "loss": 0.0284, + "step": 30911 + }, + { + "epoch": 0.7246259603720178, + "grad_norm": 0.2005612850189209, + "learning_rate": 3.515901913439449e-05, + "loss": 0.0216, + "step": 30912 + }, + { + "epoch": 0.7246494019468228, + "grad_norm": 0.6549564003944397, + "learning_rate": 3.515341151710436e-05, + "loss": 0.1208, + "step": 30913 + }, + { + "epoch": 0.7246728435216278, + "grad_norm": 1.1110013723373413, + "learning_rate": 3.514780425167364e-05, + "loss": 0.103, + "step": 30914 + }, + { + "epoch": 0.7246962850964328, + "grad_norm": 0.3585177958011627, + "learning_rate": 3.514219733813274e-05, + "loss": 0.0598, + "step": 30915 + }, + { + "epoch": 0.7247197266712377, + "grad_norm": 0.3263819217681885, + "learning_rate": 3.5136590776512126e-05, + "loss": 0.0319, + "step": 30916 + }, + { + "epoch": 0.7247431682460428, + "grad_norm": 0.82209312915802, + "learning_rate": 3.5130984566842196e-05, + "loss": 0.1614, + "step": 30917 + }, + { + "epoch": 0.7247666098208477, + "grad_norm": 0.48381608724594116, + "learning_rate": 3.512537870915336e-05, + "loss": 0.0926, + "step": 30918 + }, + { + "epoch": 0.7247900513956528, + "grad_norm": 0.46604084968566895, + "learning_rate": 3.5119773203476045e-05, + "loss": 0.0985, + "step": 30919 + }, + { + "epoch": 0.7248134929704577, + "grad_norm": 0.46957141160964966, + "learning_rate": 3.511416804984067e-05, + "loss": 0.1022, + "step": 30920 + }, + { + "epoch": 0.7248369345452628, + "grad_norm": 0.08073394745588303, + "learning_rate": 3.510856324827761e-05, + "loss": 0.0169, + "step": 30921 + }, + { + "epoch": 0.7248603761200677, + "grad_norm": 0.5151752233505249, + "learning_rate": 3.510295879881731e-05, + "loss": 0.0703, + "step": 30922 + }, + { + "epoch": 0.7248838176948728, + "grad_norm": 0.5882045030593872, + "learning_rate": 3.509735470149022e-05, + "loss": 0.1022, + "step": 30923 + }, + { + "epoch": 0.7249072592696777, + "grad_norm": 0.5530440807342529, + "learning_rate": 3.50917509563267e-05, + "loss": 0.1385, + "step": 30924 + }, + { + "epoch": 0.7249307008444827, + "grad_norm": 0.35822704434394836, + "learning_rate": 3.508614756335715e-05, + "loss": 0.0301, + "step": 30925 + }, + { + "epoch": 0.7249541424192877, + "grad_norm": 0.4206465184688568, + "learning_rate": 3.5080544522612014e-05, + "loss": 0.0586, + "step": 30926 + }, + { + "epoch": 0.7249775839940927, + "grad_norm": 0.1723010390996933, + "learning_rate": 3.5074941834121655e-05, + "loss": 0.0251, + "step": 30927 + }, + { + "epoch": 0.7250010255688977, + "grad_norm": 0.2017393857240677, + "learning_rate": 3.506933949791645e-05, + "loss": 0.0384, + "step": 30928 + }, + { + "epoch": 0.7250244671437027, + "grad_norm": 0.146988183259964, + "learning_rate": 3.5063737514026886e-05, + "loss": 0.0311, + "step": 30929 + }, + { + "epoch": 0.7250479087185077, + "grad_norm": 0.6131684184074402, + "learning_rate": 3.505813588248329e-05, + "loss": 0.1163, + "step": 30930 + }, + { + "epoch": 0.7250713502933127, + "grad_norm": 0.3093116879463196, + "learning_rate": 3.505253460331609e-05, + "loss": 0.0419, + "step": 30931 + }, + { + "epoch": 0.7250947918681176, + "grad_norm": 0.33908611536026, + "learning_rate": 3.504693367655565e-05, + "loss": 0.0502, + "step": 30932 + }, + { + "epoch": 0.7251182334429227, + "grad_norm": 0.14975646138191223, + "learning_rate": 3.5041333102232346e-05, + "loss": 0.0188, + "step": 30933 + }, + { + "epoch": 0.7251416750177276, + "grad_norm": 0.3179084062576294, + "learning_rate": 3.5035732880376624e-05, + "loss": 0.0553, + "step": 30934 + }, + { + "epoch": 0.7251651165925327, + "grad_norm": 0.5476715564727783, + "learning_rate": 3.503013301101882e-05, + "loss": 0.4544, + "step": 30935 + }, + { + "epoch": 0.7251885581673376, + "grad_norm": 0.07079246640205383, + "learning_rate": 3.502453349418937e-05, + "loss": 0.0081, + "step": 30936 + }, + { + "epoch": 0.7252119997421427, + "grad_norm": 0.4191160500049591, + "learning_rate": 3.501893432991862e-05, + "loss": 0.0566, + "step": 30937 + }, + { + "epoch": 0.7252354413169477, + "grad_norm": 0.3818896412849426, + "learning_rate": 3.5013335518236964e-05, + "loss": 0.0936, + "step": 30938 + }, + { + "epoch": 0.7252588828917527, + "grad_norm": 0.5900419354438782, + "learning_rate": 3.500773705917478e-05, + "loss": 0.0583, + "step": 30939 + }, + { + "epoch": 0.7252823244665577, + "grad_norm": 0.6707642078399658, + "learning_rate": 3.500213895276241e-05, + "loss": 0.1088, + "step": 30940 + }, + { + "epoch": 0.7253057660413627, + "grad_norm": 0.4402214586734772, + "learning_rate": 3.499654119903029e-05, + "loss": 0.0931, + "step": 30941 + }, + { + "epoch": 0.7253292076161677, + "grad_norm": 0.16705258190631866, + "learning_rate": 3.499094379800877e-05, + "loss": 0.032, + "step": 30942 + }, + { + "epoch": 0.7253526491909726, + "grad_norm": 0.39052343368530273, + "learning_rate": 3.498534674972821e-05, + "loss": 0.0577, + "step": 30943 + }, + { + "epoch": 0.7253760907657777, + "grad_norm": 0.23622453212738037, + "learning_rate": 3.4979750054218996e-05, + "loss": 0.0166, + "step": 30944 + }, + { + "epoch": 0.7253995323405826, + "grad_norm": 0.6245509386062622, + "learning_rate": 3.497415371151148e-05, + "loss": 0.0732, + "step": 30945 + }, + { + "epoch": 0.7254229739153877, + "grad_norm": 0.8075331449508667, + "learning_rate": 3.4968557721636e-05, + "loss": 0.0967, + "step": 30946 + }, + { + "epoch": 0.7254464154901926, + "grad_norm": 0.11214989423751831, + "learning_rate": 3.4962962084622955e-05, + "loss": 0.0162, + "step": 30947 + }, + { + "epoch": 0.7254698570649977, + "grad_norm": 0.29019811749458313, + "learning_rate": 3.495736680050274e-05, + "loss": 0.0381, + "step": 30948 + }, + { + "epoch": 0.7254932986398026, + "grad_norm": 0.13617627322673798, + "learning_rate": 3.4951771869305686e-05, + "loss": 0.0156, + "step": 30949 + }, + { + "epoch": 0.7255167402146077, + "grad_norm": 0.20891818404197693, + "learning_rate": 3.494617729106214e-05, + "loss": 0.0375, + "step": 30950 + }, + { + "epoch": 0.7255401817894126, + "grad_norm": 0.2421106994152069, + "learning_rate": 3.494058306580247e-05, + "loss": 0.0157, + "step": 30951 + }, + { + "epoch": 0.7255636233642176, + "grad_norm": 0.20184744894504547, + "learning_rate": 3.4934989193557023e-05, + "loss": 0.0345, + "step": 30952 + }, + { + "epoch": 0.7255870649390226, + "grad_norm": 0.45607152581214905, + "learning_rate": 3.4929395674356124e-05, + "loss": 0.0325, + "step": 30953 + }, + { + "epoch": 0.7256105065138276, + "grad_norm": 0.2791590392589569, + "learning_rate": 3.492380250823019e-05, + "loss": 0.0399, + "step": 30954 + }, + { + "epoch": 0.7256339480886326, + "grad_norm": 0.061481673270463943, + "learning_rate": 3.491820969520953e-05, + "loss": 0.0096, + "step": 30955 + }, + { + "epoch": 0.7256573896634376, + "grad_norm": 0.8471044301986694, + "learning_rate": 3.4912617235324484e-05, + "loss": 0.2368, + "step": 30956 + }, + { + "epoch": 0.7256808312382426, + "grad_norm": 0.5776920318603516, + "learning_rate": 3.490702512860542e-05, + "loss": 0.4687, + "step": 30957 + }, + { + "epoch": 0.7257042728130476, + "grad_norm": 0.60518878698349, + "learning_rate": 3.490143337508263e-05, + "loss": 0.5489, + "step": 30958 + }, + { + "epoch": 0.7257277143878526, + "grad_norm": 0.8786543607711792, + "learning_rate": 3.4895841974786527e-05, + "loss": 0.2118, + "step": 30959 + }, + { + "epoch": 0.7257511559626576, + "grad_norm": 0.5485391020774841, + "learning_rate": 3.489025092774738e-05, + "loss": 0.1269, + "step": 30960 + }, + { + "epoch": 0.7257745975374625, + "grad_norm": 0.24107959866523743, + "learning_rate": 3.4884660233995595e-05, + "loss": 0.0234, + "step": 30961 + }, + { + "epoch": 0.7257980391122676, + "grad_norm": 0.6442404985427856, + "learning_rate": 3.4879069893561475e-05, + "loss": 0.1061, + "step": 30962 + }, + { + "epoch": 0.7258214806870725, + "grad_norm": 0.6497030258178711, + "learning_rate": 3.487347990647536e-05, + "loss": 0.0886, + "step": 30963 + }, + { + "epoch": 0.7258449222618776, + "grad_norm": 0.4111140966415405, + "learning_rate": 3.4867890272767565e-05, + "loss": 0.0364, + "step": 30964 + }, + { + "epoch": 0.7258683638366825, + "grad_norm": 0.5844666957855225, + "learning_rate": 3.486230099246843e-05, + "loss": 0.0857, + "step": 30965 + }, + { + "epoch": 0.7258918054114876, + "grad_norm": 0.3160073161125183, + "learning_rate": 3.485671206560825e-05, + "loss": 0.0378, + "step": 30966 + }, + { + "epoch": 0.7259152469862925, + "grad_norm": 0.7850914001464844, + "learning_rate": 3.485112349221742e-05, + "loss": 0.5178, + "step": 30967 + }, + { + "epoch": 0.7259386885610976, + "grad_norm": 0.6221737861633301, + "learning_rate": 3.484553527232622e-05, + "loss": 0.0928, + "step": 30968 + }, + { + "epoch": 0.7259621301359025, + "grad_norm": 1.4437651634216309, + "learning_rate": 3.483994740596498e-05, + "loss": 0.1693, + "step": 30969 + }, + { + "epoch": 0.7259855717107075, + "grad_norm": 0.12524525821208954, + "learning_rate": 3.483435989316398e-05, + "loss": 0.0154, + "step": 30970 + }, + { + "epoch": 0.7260090132855125, + "grad_norm": 0.08889163285493851, + "learning_rate": 3.482877273395362e-05, + "loss": 0.0109, + "step": 30971 + }, + { + "epoch": 0.7260324548603175, + "grad_norm": 0.563848614692688, + "learning_rate": 3.482318592836416e-05, + "loss": 0.0838, + "step": 30972 + }, + { + "epoch": 0.7260558964351225, + "grad_norm": 0.5369212031364441, + "learning_rate": 3.48175994764259e-05, + "loss": 0.0297, + "step": 30973 + }, + { + "epoch": 0.7260793380099275, + "grad_norm": 0.4980204105377197, + "learning_rate": 3.48120133781692e-05, + "loss": 0.0523, + "step": 30974 + }, + { + "epoch": 0.7261027795847325, + "grad_norm": 0.4996804893016815, + "learning_rate": 3.480642763362436e-05, + "loss": 0.4063, + "step": 30975 + }, + { + "epoch": 0.7261262211595375, + "grad_norm": 0.15796637535095215, + "learning_rate": 3.4800842242821675e-05, + "loss": 0.0241, + "step": 30976 + }, + { + "epoch": 0.7261496627343424, + "grad_norm": 0.448308527469635, + "learning_rate": 3.479525720579146e-05, + "loss": 0.1211, + "step": 30977 + }, + { + "epoch": 0.7261731043091475, + "grad_norm": 0.5732307434082031, + "learning_rate": 3.478967252256395e-05, + "loss": 0.0921, + "step": 30978 + }, + { + "epoch": 0.7261965458839524, + "grad_norm": 0.3089749813079834, + "learning_rate": 3.4784088193169575e-05, + "loss": 0.0701, + "step": 30979 + }, + { + "epoch": 0.7262199874587575, + "grad_norm": 0.44627901911735535, + "learning_rate": 3.477850421763857e-05, + "loss": 0.4859, + "step": 30980 + }, + { + "epoch": 0.7262434290335624, + "grad_norm": 0.5936005711555481, + "learning_rate": 3.477292059600121e-05, + "loss": 0.0805, + "step": 30981 + }, + { + "epoch": 0.7262668706083675, + "grad_norm": 0.25889304280281067, + "learning_rate": 3.476733732828781e-05, + "loss": 0.0469, + "step": 30982 + }, + { + "epoch": 0.7262903121831724, + "grad_norm": 0.40779855847358704, + "learning_rate": 3.476175441452869e-05, + "loss": 0.0691, + "step": 30983 + }, + { + "epoch": 0.7263137537579775, + "grad_norm": 0.5274901986122131, + "learning_rate": 3.4756171854754125e-05, + "loss": 0.6371, + "step": 30984 + }, + { + "epoch": 0.7263371953327824, + "grad_norm": 0.28314563632011414, + "learning_rate": 3.475058964899438e-05, + "loss": 0.0318, + "step": 30985 + }, + { + "epoch": 0.7263606369075875, + "grad_norm": 0.24143756926059723, + "learning_rate": 3.4745007797279797e-05, + "loss": 0.0352, + "step": 30986 + }, + { + "epoch": 0.7263840784823924, + "grad_norm": 0.16340424120426178, + "learning_rate": 3.473942629964064e-05, + "loss": 0.0121, + "step": 30987 + }, + { + "epoch": 0.7264075200571974, + "grad_norm": 0.11806721240282059, + "learning_rate": 3.473384515610719e-05, + "loss": 0.0231, + "step": 30988 + }, + { + "epoch": 0.7264309616320025, + "grad_norm": 0.5044209361076355, + "learning_rate": 3.472826436670973e-05, + "loss": 0.0673, + "step": 30989 + }, + { + "epoch": 0.7264544032068074, + "grad_norm": 0.18813367187976837, + "learning_rate": 3.472268393147855e-05, + "loss": 0.0328, + "step": 30990 + }, + { + "epoch": 0.7264778447816125, + "grad_norm": 0.38967373967170715, + "learning_rate": 3.4717103850443876e-05, + "loss": 0.0586, + "step": 30991 + }, + { + "epoch": 0.7265012863564174, + "grad_norm": 0.17992126941680908, + "learning_rate": 3.4711524123636073e-05, + "loss": 0.0256, + "step": 30992 + }, + { + "epoch": 0.7265247279312225, + "grad_norm": 0.16157196462154388, + "learning_rate": 3.470594475108538e-05, + "loss": 0.0216, + "step": 30993 + }, + { + "epoch": 0.7265481695060274, + "grad_norm": 0.24573186039924622, + "learning_rate": 3.470036573282203e-05, + "loss": 0.0371, + "step": 30994 + }, + { + "epoch": 0.7265716110808325, + "grad_norm": 0.5332486629486084, + "learning_rate": 3.4694787068876365e-05, + "loss": 0.0863, + "step": 30995 + }, + { + "epoch": 0.7265950526556374, + "grad_norm": 0.18721486628055573, + "learning_rate": 3.468920875927862e-05, + "loss": 0.0198, + "step": 30996 + }, + { + "epoch": 0.7266184942304424, + "grad_norm": 0.37185904383659363, + "learning_rate": 3.4683630804059066e-05, + "loss": 0.0788, + "step": 30997 + }, + { + "epoch": 0.7266419358052474, + "grad_norm": 0.3554271459579468, + "learning_rate": 3.467805320324794e-05, + "loss": 0.0497, + "step": 30998 + }, + { + "epoch": 0.7266653773800524, + "grad_norm": 0.6196630001068115, + "learning_rate": 3.467247595687556e-05, + "loss": 0.5807, + "step": 30999 + }, + { + "epoch": 0.7266888189548574, + "grad_norm": 0.5583145022392273, + "learning_rate": 3.466689906497216e-05, + "loss": 0.0886, + "step": 31000 + }, + { + "epoch": 0.7267122605296624, + "grad_norm": 0.2180614471435547, + "learning_rate": 3.4661322527568006e-05, + "loss": 0.0349, + "step": 31001 + }, + { + "epoch": 0.7267357021044674, + "grad_norm": 0.5628218650817871, + "learning_rate": 3.465574634469335e-05, + "loss": 0.0742, + "step": 31002 + }, + { + "epoch": 0.7267591436792724, + "grad_norm": 0.3960147500038147, + "learning_rate": 3.465017051637846e-05, + "loss": 0.0555, + "step": 31003 + }, + { + "epoch": 0.7267825852540774, + "grad_norm": 0.15421344339847565, + "learning_rate": 3.464459504265355e-05, + "loss": 0.0417, + "step": 31004 + }, + { + "epoch": 0.7268060268288824, + "grad_norm": 0.6479560136795044, + "learning_rate": 3.4639019923548934e-05, + "loss": 0.7149, + "step": 31005 + }, + { + "epoch": 0.7268294684036873, + "grad_norm": 0.29272952675819397, + "learning_rate": 3.463344515909479e-05, + "loss": 0.0331, + "step": 31006 + }, + { + "epoch": 0.7268529099784924, + "grad_norm": 0.18522299826145172, + "learning_rate": 3.4627870749321455e-05, + "loss": 0.0207, + "step": 31007 + }, + { + "epoch": 0.7268763515532973, + "grad_norm": 0.6321225166320801, + "learning_rate": 3.462229669425913e-05, + "loss": 0.0275, + "step": 31008 + }, + { + "epoch": 0.7268997931281024, + "grad_norm": 0.0591479055583477, + "learning_rate": 3.4616722993938056e-05, + "loss": 0.0049, + "step": 31009 + }, + { + "epoch": 0.7269232347029073, + "grad_norm": 0.17470493912696838, + "learning_rate": 3.461114964838849e-05, + "loss": 0.0315, + "step": 31010 + }, + { + "epoch": 0.7269466762777124, + "grad_norm": 0.16918057203292847, + "learning_rate": 3.460557665764062e-05, + "loss": 0.0229, + "step": 31011 + }, + { + "epoch": 0.7269701178525173, + "grad_norm": 0.4516272246837616, + "learning_rate": 3.4600004021724774e-05, + "loss": 0.0663, + "step": 31012 + }, + { + "epoch": 0.7269935594273224, + "grad_norm": 0.23943902552127838, + "learning_rate": 3.459443174067114e-05, + "loss": 0.0466, + "step": 31013 + }, + { + "epoch": 0.7270170010021273, + "grad_norm": 0.4593064785003662, + "learning_rate": 3.458885981450996e-05, + "loss": 0.1029, + "step": 31014 + }, + { + "epoch": 0.7270404425769323, + "grad_norm": 0.16767477989196777, + "learning_rate": 3.458328824327147e-05, + "loss": 0.0292, + "step": 31015 + }, + { + "epoch": 0.7270638841517373, + "grad_norm": 0.22723771631717682, + "learning_rate": 3.457771702698586e-05, + "loss": 0.0311, + "step": 31016 + }, + { + "epoch": 0.7270873257265423, + "grad_norm": 0.25892072916030884, + "learning_rate": 3.4572146165683425e-05, + "loss": 0.0606, + "step": 31017 + }, + { + "epoch": 0.7271107673013473, + "grad_norm": 0.4428500235080719, + "learning_rate": 3.4566575659394374e-05, + "loss": 0.0389, + "step": 31018 + }, + { + "epoch": 0.7271342088761523, + "grad_norm": 0.14735494554042816, + "learning_rate": 3.456100550814888e-05, + "loss": 0.0248, + "step": 31019 + }, + { + "epoch": 0.7271576504509573, + "grad_norm": 0.21329566836357117, + "learning_rate": 3.4555435711977244e-05, + "loss": 0.029, + "step": 31020 + }, + { + "epoch": 0.7271810920257623, + "grad_norm": 0.3304567039012909, + "learning_rate": 3.454986627090967e-05, + "loss": 0.0446, + "step": 31021 + }, + { + "epoch": 0.7272045336005672, + "grad_norm": 1.4517221450805664, + "learning_rate": 3.454429718497635e-05, + "loss": 0.0793, + "step": 31022 + }, + { + "epoch": 0.7272279751753723, + "grad_norm": 0.3041183054447174, + "learning_rate": 3.453872845420748e-05, + "loss": 0.0219, + "step": 31023 + }, + { + "epoch": 0.7272514167501772, + "grad_norm": 0.5965265035629272, + "learning_rate": 3.453316007863334e-05, + "loss": 0.0606, + "step": 31024 + }, + { + "epoch": 0.7272748583249823, + "grad_norm": 0.5072116255760193, + "learning_rate": 3.4527592058284116e-05, + "loss": 0.0789, + "step": 31025 + }, + { + "epoch": 0.7272982998997872, + "grad_norm": 0.6054565906524658, + "learning_rate": 3.452202439319002e-05, + "loss": 0.1424, + "step": 31026 + }, + { + "epoch": 0.7273217414745923, + "grad_norm": 0.29022419452667236, + "learning_rate": 3.451645708338127e-05, + "loss": 0.0498, + "step": 31027 + }, + { + "epoch": 0.7273451830493972, + "grad_norm": 0.5481155514717102, + "learning_rate": 3.451089012888805e-05, + "loss": 0.0727, + "step": 31028 + }, + { + "epoch": 0.7273686246242023, + "grad_norm": 0.21783016622066498, + "learning_rate": 3.4505323529740564e-05, + "loss": 0.027, + "step": 31029 + }, + { + "epoch": 0.7273920661990072, + "grad_norm": 0.14865757524967194, + "learning_rate": 3.449975728596906e-05, + "loss": 0.0336, + "step": 31030 + }, + { + "epoch": 0.7274155077738123, + "grad_norm": 0.3612464666366577, + "learning_rate": 3.449419139760367e-05, + "loss": 0.0782, + "step": 31031 + }, + { + "epoch": 0.7274389493486172, + "grad_norm": 0.448869526386261, + "learning_rate": 3.448862586467469e-05, + "loss": 0.0613, + "step": 31032 + }, + { + "epoch": 0.7274623909234222, + "grad_norm": 0.34082141518592834, + "learning_rate": 3.4483060687212256e-05, + "loss": 0.0424, + "step": 31033 + }, + { + "epoch": 0.7274858324982272, + "grad_norm": 0.6005061268806458, + "learning_rate": 3.447749586524659e-05, + "loss": 0.0997, + "step": 31034 + }, + { + "epoch": 0.7275092740730322, + "grad_norm": 0.26126617193222046, + "learning_rate": 3.4471931398807854e-05, + "loss": 0.0293, + "step": 31035 + }, + { + "epoch": 0.7275327156478372, + "grad_norm": 0.6753287315368652, + "learning_rate": 3.446636728792625e-05, + "loss": 0.1608, + "step": 31036 + }, + { + "epoch": 0.7275561572226422, + "grad_norm": 0.1490069031715393, + "learning_rate": 3.446080353263199e-05, + "loss": 0.0265, + "step": 31037 + }, + { + "epoch": 0.7275795987974472, + "grad_norm": 0.5220131278038025, + "learning_rate": 3.445524013295527e-05, + "loss": 0.0763, + "step": 31038 + }, + { + "epoch": 0.7276030403722522, + "grad_norm": 0.13476872444152832, + "learning_rate": 3.4449677088926256e-05, + "loss": 0.0223, + "step": 31039 + }, + { + "epoch": 0.7276264819470573, + "grad_norm": 0.2657475471496582, + "learning_rate": 3.444411440057513e-05, + "loss": 0.0309, + "step": 31040 + }, + { + "epoch": 0.7276499235218622, + "grad_norm": 0.5273421406745911, + "learning_rate": 3.44385520679321e-05, + "loss": 0.0502, + "step": 31041 + }, + { + "epoch": 0.7276733650966672, + "grad_norm": 0.8137086629867554, + "learning_rate": 3.443299009102728e-05, + "loss": 0.1661, + "step": 31042 + }, + { + "epoch": 0.7276968066714722, + "grad_norm": 0.4835010766983032, + "learning_rate": 3.4427428469890896e-05, + "loss": 0.0813, + "step": 31043 + }, + { + "epoch": 0.7277202482462772, + "grad_norm": 0.20599505305290222, + "learning_rate": 3.4421867204553184e-05, + "loss": 0.0208, + "step": 31044 + }, + { + "epoch": 0.7277436898210822, + "grad_norm": 0.12621010839939117, + "learning_rate": 3.441630629504424e-05, + "loss": 0.0145, + "step": 31045 + }, + { + "epoch": 0.7277671313958872, + "grad_norm": 0.29634109139442444, + "learning_rate": 3.441074574139428e-05, + "loss": 0.0485, + "step": 31046 + }, + { + "epoch": 0.7277905729706922, + "grad_norm": 0.40462103486061096, + "learning_rate": 3.440518554363344e-05, + "loss": 0.0648, + "step": 31047 + }, + { + "epoch": 0.7278140145454972, + "grad_norm": 0.7309973835945129, + "learning_rate": 3.439962570179193e-05, + "loss": 0.104, + "step": 31048 + }, + { + "epoch": 0.7278374561203022, + "grad_norm": 0.5352310538291931, + "learning_rate": 3.4394066215899845e-05, + "loss": 0.0999, + "step": 31049 + }, + { + "epoch": 0.7278608976951072, + "grad_norm": 0.4263315796852112, + "learning_rate": 3.438850708598744e-05, + "loss": 0.0452, + "step": 31050 + }, + { + "epoch": 0.7278843392699121, + "grad_norm": 0.4534221291542053, + "learning_rate": 3.438294831208484e-05, + "loss": 0.0368, + "step": 31051 + }, + { + "epoch": 0.7279077808447172, + "grad_norm": 0.21940645575523376, + "learning_rate": 3.4377389894222214e-05, + "loss": 0.0422, + "step": 31052 + }, + { + "epoch": 0.7279312224195221, + "grad_norm": 0.28362852334976196, + "learning_rate": 3.437183183242971e-05, + "loss": 0.0447, + "step": 31053 + }, + { + "epoch": 0.7279546639943272, + "grad_norm": 0.5543888807296753, + "learning_rate": 3.436627412673745e-05, + "loss": 0.1152, + "step": 31054 + }, + { + "epoch": 0.7279781055691321, + "grad_norm": 0.13290196657180786, + "learning_rate": 3.436071677717565e-05, + "loss": 0.0219, + "step": 31055 + }, + { + "epoch": 0.7280015471439372, + "grad_norm": 0.2785733640193939, + "learning_rate": 3.435515978377446e-05, + "loss": 0.0329, + "step": 31056 + }, + { + "epoch": 0.7280249887187421, + "grad_norm": 0.282770037651062, + "learning_rate": 3.434960314656404e-05, + "loss": 0.0443, + "step": 31057 + }, + { + "epoch": 0.7280484302935472, + "grad_norm": 0.3531181216239929, + "learning_rate": 3.43440468655745e-05, + "loss": 0.0659, + "step": 31058 + }, + { + "epoch": 0.7280718718683521, + "grad_norm": 0.4343528747558594, + "learning_rate": 3.433849094083602e-05, + "loss": 0.0543, + "step": 31059 + }, + { + "epoch": 0.7280953134431571, + "grad_norm": 0.5402979850769043, + "learning_rate": 3.433293537237873e-05, + "loss": 0.1246, + "step": 31060 + }, + { + "epoch": 0.7281187550179621, + "grad_norm": 0.3565995395183563, + "learning_rate": 3.4327380160232746e-05, + "loss": 0.0547, + "step": 31061 + }, + { + "epoch": 0.7281421965927671, + "grad_norm": 0.25712257623672485, + "learning_rate": 3.432182530442828e-05, + "loss": 0.0603, + "step": 31062 + }, + { + "epoch": 0.7281656381675721, + "grad_norm": 0.41663604974746704, + "learning_rate": 3.4316270804995434e-05, + "loss": 0.0853, + "step": 31063 + }, + { + "epoch": 0.7281890797423771, + "grad_norm": 0.17093342542648315, + "learning_rate": 3.4310716661964346e-05, + "loss": 0.0271, + "step": 31064 + }, + { + "epoch": 0.7282125213171821, + "grad_norm": 0.5044246912002563, + "learning_rate": 3.430516287536516e-05, + "loss": 0.0934, + "step": 31065 + }, + { + "epoch": 0.7282359628919871, + "grad_norm": 0.6986443996429443, + "learning_rate": 3.429960944522801e-05, + "loss": 0.1213, + "step": 31066 + }, + { + "epoch": 0.728259404466792, + "grad_norm": 0.6563909649848938, + "learning_rate": 3.429405637158298e-05, + "loss": 0.0874, + "step": 31067 + }, + { + "epoch": 0.7282828460415971, + "grad_norm": 0.15807829797267914, + "learning_rate": 3.428850365446026e-05, + "loss": 0.0303, + "step": 31068 + }, + { + "epoch": 0.728306287616402, + "grad_norm": 0.2593332827091217, + "learning_rate": 3.428295129389e-05, + "loss": 0.0397, + "step": 31069 + }, + { + "epoch": 0.7283297291912071, + "grad_norm": 0.42547741532325745, + "learning_rate": 3.427739928990229e-05, + "loss": 0.0507, + "step": 31070 + }, + { + "epoch": 0.728353170766012, + "grad_norm": 0.4985032379627228, + "learning_rate": 3.4271847642527265e-05, + "loss": 0.1817, + "step": 31071 + }, + { + "epoch": 0.7283766123408171, + "grad_norm": 0.5113300085067749, + "learning_rate": 3.426629635179503e-05, + "loss": 0.1284, + "step": 31072 + }, + { + "epoch": 0.728400053915622, + "grad_norm": 0.2012944370508194, + "learning_rate": 3.426074541773574e-05, + "loss": 0.0213, + "step": 31073 + }, + { + "epoch": 0.7284234954904271, + "grad_norm": 0.10916491597890854, + "learning_rate": 3.425519484037944e-05, + "loss": 0.0223, + "step": 31074 + }, + { + "epoch": 0.728446937065232, + "grad_norm": 0.5505293011665344, + "learning_rate": 3.424964461975635e-05, + "loss": 0.4973, + "step": 31075 + }, + { + "epoch": 0.728470378640037, + "grad_norm": 0.43804967403411865, + "learning_rate": 3.424409475589653e-05, + "loss": 0.0728, + "step": 31076 + }, + { + "epoch": 0.728493820214842, + "grad_norm": 0.43884459137916565, + "learning_rate": 3.42385452488301e-05, + "loss": 0.0654, + "step": 31077 + }, + { + "epoch": 0.728517261789647, + "grad_norm": 0.20851294696331024, + "learning_rate": 3.423299609858718e-05, + "loss": 0.0241, + "step": 31078 + }, + { + "epoch": 0.728540703364452, + "grad_norm": 0.08811792731285095, + "learning_rate": 3.422744730519782e-05, + "loss": 0.0124, + "step": 31079 + }, + { + "epoch": 0.728564144939257, + "grad_norm": 0.39866355061531067, + "learning_rate": 3.422189886869224e-05, + "loss": 0.0439, + "step": 31080 + }, + { + "epoch": 0.728587586514062, + "grad_norm": 0.19302107393741608, + "learning_rate": 3.421635078910044e-05, + "loss": 0.0113, + "step": 31081 + }, + { + "epoch": 0.728611028088867, + "grad_norm": 0.4227679371833801, + "learning_rate": 3.421080306645259e-05, + "loss": 0.0879, + "step": 31082 + }, + { + "epoch": 0.728634469663672, + "grad_norm": 0.3802861273288727, + "learning_rate": 3.420525570077879e-05, + "loss": 0.0426, + "step": 31083 + }, + { + "epoch": 0.728657911238477, + "grad_norm": 0.4560907185077667, + "learning_rate": 3.419970869210912e-05, + "loss": 0.0736, + "step": 31084 + }, + { + "epoch": 0.728681352813282, + "grad_norm": 0.206712543964386, + "learning_rate": 3.4194162040473675e-05, + "loss": 0.0423, + "step": 31085 + }, + { + "epoch": 0.728704794388087, + "grad_norm": 0.16635997593402863, + "learning_rate": 3.418861574590253e-05, + "loss": 0.0155, + "step": 31086 + }, + { + "epoch": 0.7287282359628919, + "grad_norm": 0.5826452970504761, + "learning_rate": 3.418306980842584e-05, + "loss": 0.0662, + "step": 31087 + }, + { + "epoch": 0.728751677537697, + "grad_norm": 0.24879583716392517, + "learning_rate": 3.417752422807365e-05, + "loss": 0.0341, + "step": 31088 + }, + { + "epoch": 0.7287751191125019, + "grad_norm": 0.4123935401439667, + "learning_rate": 3.4171979004876084e-05, + "loss": 0.06, + "step": 31089 + }, + { + "epoch": 0.728798560687307, + "grad_norm": 0.6238150596618652, + "learning_rate": 3.41664341388632e-05, + "loss": 0.0599, + "step": 31090 + }, + { + "epoch": 0.728822002262112, + "grad_norm": 0.18363620340824127, + "learning_rate": 3.4160889630065095e-05, + "loss": 0.0278, + "step": 31091 + }, + { + "epoch": 0.728845443836917, + "grad_norm": 0.3857955038547516, + "learning_rate": 3.415534547851183e-05, + "loss": 0.0574, + "step": 31092 + }, + { + "epoch": 0.728868885411722, + "grad_norm": 0.4019825756549835, + "learning_rate": 3.4149801684233505e-05, + "loss": 0.0787, + "step": 31093 + }, + { + "epoch": 0.728892326986527, + "grad_norm": 0.1764519214630127, + "learning_rate": 3.414425824726024e-05, + "loss": 0.0253, + "step": 31094 + }, + { + "epoch": 0.728915768561332, + "grad_norm": 0.11134830862283707, + "learning_rate": 3.413871516762208e-05, + "loss": 0.0251, + "step": 31095 + }, + { + "epoch": 0.7289392101361369, + "grad_norm": 0.510033905506134, + "learning_rate": 3.4133172445349104e-05, + "loss": 0.7013, + "step": 31096 + }, + { + "epoch": 0.728962651710942, + "grad_norm": 0.06214233115315437, + "learning_rate": 3.412763008047139e-05, + "loss": 0.0092, + "step": 31097 + }, + { + "epoch": 0.7289860932857469, + "grad_norm": 0.5519568920135498, + "learning_rate": 3.412208807301901e-05, + "loss": 0.0386, + "step": 31098 + }, + { + "epoch": 0.729009534860552, + "grad_norm": 0.5364765524864197, + "learning_rate": 3.4116546423021986e-05, + "loss": 0.0985, + "step": 31099 + }, + { + "epoch": 0.7290329764353569, + "grad_norm": 0.8433031439781189, + "learning_rate": 3.411100513051049e-05, + "loss": 0.0705, + "step": 31100 + }, + { + "epoch": 0.729056418010162, + "grad_norm": 0.33908167481422424, + "learning_rate": 3.410546419551451e-05, + "loss": 0.0841, + "step": 31101 + }, + { + "epoch": 0.7290798595849669, + "grad_norm": 0.4249778389930725, + "learning_rate": 3.4099923618064144e-05, + "loss": 0.0596, + "step": 31102 + }, + { + "epoch": 0.729103301159772, + "grad_norm": 0.33820709586143494, + "learning_rate": 3.409438339818945e-05, + "loss": 0.0509, + "step": 31103 + }, + { + "epoch": 0.7291267427345769, + "grad_norm": 0.6441258788108826, + "learning_rate": 3.408884353592043e-05, + "loss": 0.1478, + "step": 31104 + }, + { + "epoch": 0.7291501843093819, + "grad_norm": 0.09610345214605331, + "learning_rate": 3.408330403128725e-05, + "loss": 0.01, + "step": 31105 + }, + { + "epoch": 0.7291736258841869, + "grad_norm": 0.8387285470962524, + "learning_rate": 3.407776488431988e-05, + "loss": 0.1123, + "step": 31106 + }, + { + "epoch": 0.7291970674589919, + "grad_norm": 0.15785731375217438, + "learning_rate": 3.4072226095048454e-05, + "loss": 0.0352, + "step": 31107 + }, + { + "epoch": 0.7292205090337969, + "grad_norm": 0.18791308999061584, + "learning_rate": 3.406668766350297e-05, + "loss": 0.0314, + "step": 31108 + }, + { + "epoch": 0.7292439506086019, + "grad_norm": 0.39760056138038635, + "learning_rate": 3.406114958971349e-05, + "loss": 0.0782, + "step": 31109 + }, + { + "epoch": 0.7292673921834069, + "grad_norm": 0.08360408991575241, + "learning_rate": 3.4055611873710084e-05, + "loss": 0.0163, + "step": 31110 + }, + { + "epoch": 0.7292908337582119, + "grad_norm": 0.35574132204055786, + "learning_rate": 3.405007451552277e-05, + "loss": 0.0235, + "step": 31111 + }, + { + "epoch": 0.7293142753330168, + "grad_norm": 0.15960362553596497, + "learning_rate": 3.404453751518156e-05, + "loss": 0.0346, + "step": 31112 + }, + { + "epoch": 0.7293377169078219, + "grad_norm": 0.1927308589220047, + "learning_rate": 3.4039000872716596e-05, + "loss": 0.0355, + "step": 31113 + }, + { + "epoch": 0.7293611584826268, + "grad_norm": 0.4936002790927887, + "learning_rate": 3.403346458815786e-05, + "loss": 0.0593, + "step": 31114 + }, + { + "epoch": 0.7293846000574319, + "grad_norm": 0.6161425113677979, + "learning_rate": 3.4027928661535405e-05, + "loss": 0.0729, + "step": 31115 + }, + { + "epoch": 0.7294080416322368, + "grad_norm": 0.5348708629608154, + "learning_rate": 3.4022393092879225e-05, + "loss": 0.1131, + "step": 31116 + }, + { + "epoch": 0.7294314832070419, + "grad_norm": 0.40548884868621826, + "learning_rate": 3.401685788221943e-05, + "loss": 0.0509, + "step": 31117 + }, + { + "epoch": 0.7294549247818468, + "grad_norm": 0.2689186632633209, + "learning_rate": 3.4011323029586014e-05, + "loss": 0.0412, + "step": 31118 + }, + { + "epoch": 0.7294783663566519, + "grad_norm": 0.6012427806854248, + "learning_rate": 3.4005788535009e-05, + "loss": 0.761, + "step": 31119 + }, + { + "epoch": 0.7295018079314568, + "grad_norm": 0.19195450842380524, + "learning_rate": 3.400025439851845e-05, + "loss": 0.0235, + "step": 31120 + }, + { + "epoch": 0.7295252495062619, + "grad_norm": 0.45919227600097656, + "learning_rate": 3.399472062014438e-05, + "loss": 0.0718, + "step": 31121 + }, + { + "epoch": 0.7295486910810668, + "grad_norm": 0.4427320063114166, + "learning_rate": 3.398918719991682e-05, + "loss": 0.0571, + "step": 31122 + }, + { + "epoch": 0.7295721326558718, + "grad_norm": 0.46420639753341675, + "learning_rate": 3.398365413786577e-05, + "loss": 0.5374, + "step": 31123 + }, + { + "epoch": 0.7295955742306768, + "grad_norm": 0.5415639281272888, + "learning_rate": 3.397812143402125e-05, + "loss": 0.4574, + "step": 31124 + }, + { + "epoch": 0.7296190158054818, + "grad_norm": 0.26686447858810425, + "learning_rate": 3.397258908841332e-05, + "loss": 0.0277, + "step": 31125 + }, + { + "epoch": 0.7296424573802868, + "grad_norm": 0.13508492708206177, + "learning_rate": 3.396705710107199e-05, + "loss": 0.0298, + "step": 31126 + }, + { + "epoch": 0.7296658989550918, + "grad_norm": 0.3854724168777466, + "learning_rate": 3.3961525472027255e-05, + "loss": 0.0475, + "step": 31127 + }, + { + "epoch": 0.7296893405298968, + "grad_norm": 0.33212074637413025, + "learning_rate": 3.3955994201309105e-05, + "loss": 0.0529, + "step": 31128 + }, + { + "epoch": 0.7297127821047018, + "grad_norm": 0.5927364826202393, + "learning_rate": 3.395046328894763e-05, + "loss": 0.0951, + "step": 31129 + }, + { + "epoch": 0.7297362236795067, + "grad_norm": 0.6060454249382019, + "learning_rate": 3.394493273497279e-05, + "loss": 0.6658, + "step": 31130 + }, + { + "epoch": 0.7297596652543118, + "grad_norm": 0.23714397847652435, + "learning_rate": 3.393940253941458e-05, + "loss": 0.0166, + "step": 31131 + }, + { + "epoch": 0.7297831068291167, + "grad_norm": 0.12214761227369308, + "learning_rate": 3.393387270230305e-05, + "loss": 0.0224, + "step": 31132 + }, + { + "epoch": 0.7298065484039218, + "grad_norm": 0.256168007850647, + "learning_rate": 3.3928343223668194e-05, + "loss": 0.0284, + "step": 31133 + }, + { + "epoch": 0.7298299899787267, + "grad_norm": 0.8391091823577881, + "learning_rate": 3.3922814103540005e-05, + "loss": 0.0937, + "step": 31134 + }, + { + "epoch": 0.7298534315535318, + "grad_norm": 0.7279921174049377, + "learning_rate": 3.391728534194848e-05, + "loss": 0.4667, + "step": 31135 + }, + { + "epoch": 0.7298768731283367, + "grad_norm": 0.3044934868812561, + "learning_rate": 3.3911756938923624e-05, + "loss": 0.0444, + "step": 31136 + }, + { + "epoch": 0.7299003147031418, + "grad_norm": 0.6535782814025879, + "learning_rate": 3.3906228894495396e-05, + "loss": 0.1389, + "step": 31137 + }, + { + "epoch": 0.7299237562779467, + "grad_norm": 0.23420125246047974, + "learning_rate": 3.390070120869386e-05, + "loss": 0.0691, + "step": 31138 + }, + { + "epoch": 0.7299471978527518, + "grad_norm": 0.6560924649238586, + "learning_rate": 3.389517388154899e-05, + "loss": 0.1061, + "step": 31139 + }, + { + "epoch": 0.7299706394275567, + "grad_norm": 0.3162025213241577, + "learning_rate": 3.388964691309074e-05, + "loss": 0.0619, + "step": 31140 + }, + { + "epoch": 0.7299940810023617, + "grad_norm": 0.34886932373046875, + "learning_rate": 3.3884120303349134e-05, + "loss": 0.0554, + "step": 31141 + }, + { + "epoch": 0.7300175225771668, + "grad_norm": 0.4427899420261383, + "learning_rate": 3.387859405235416e-05, + "loss": 0.0802, + "step": 31142 + }, + { + "epoch": 0.7300409641519717, + "grad_norm": 0.41000229120254517, + "learning_rate": 3.387306816013579e-05, + "loss": 0.0662, + "step": 31143 + }, + { + "epoch": 0.7300644057267768, + "grad_norm": 0.35559022426605225, + "learning_rate": 3.386754262672398e-05, + "loss": 0.0511, + "step": 31144 + }, + { + "epoch": 0.7300878473015817, + "grad_norm": 0.3890027403831482, + "learning_rate": 3.3862017452148784e-05, + "loss": 0.0609, + "step": 31145 + }, + { + "epoch": 0.7301112888763868, + "grad_norm": 0.5617232918739319, + "learning_rate": 3.385649263644013e-05, + "loss": 0.0569, + "step": 31146 + }, + { + "epoch": 0.7301347304511917, + "grad_norm": 0.7142744064331055, + "learning_rate": 3.385096817962802e-05, + "loss": 0.1167, + "step": 31147 + }, + { + "epoch": 0.7301581720259968, + "grad_norm": 0.6377702951431274, + "learning_rate": 3.3845444081742406e-05, + "loss": 0.1294, + "step": 31148 + }, + { + "epoch": 0.7301816136008017, + "grad_norm": 0.4022050201892853, + "learning_rate": 3.383992034281327e-05, + "loss": 0.0527, + "step": 31149 + }, + { + "epoch": 0.7302050551756067, + "grad_norm": 0.17506466805934906, + "learning_rate": 3.383439696287056e-05, + "loss": 0.0153, + "step": 31150 + }, + { + "epoch": 0.7302284967504117, + "grad_norm": 0.6025483012199402, + "learning_rate": 3.382887394194431e-05, + "loss": 0.647, + "step": 31151 + }, + { + "epoch": 0.7302519383252167, + "grad_norm": 0.48435088992118835, + "learning_rate": 3.382335128006441e-05, + "loss": 0.0654, + "step": 31152 + }, + { + "epoch": 0.7302753799000217, + "grad_norm": 0.3191887140274048, + "learning_rate": 3.38178289772609e-05, + "loss": 0.0703, + "step": 31153 + }, + { + "epoch": 0.7302988214748267, + "grad_norm": 0.43841519951820374, + "learning_rate": 3.381230703356371e-05, + "loss": 0.0527, + "step": 31154 + }, + { + "epoch": 0.7303222630496317, + "grad_norm": 0.3693575859069824, + "learning_rate": 3.380678544900281e-05, + "loss": 0.0461, + "step": 31155 + }, + { + "epoch": 0.7303457046244367, + "grad_norm": 0.07430879771709442, + "learning_rate": 3.380126422360814e-05, + "loss": 0.0086, + "step": 31156 + }, + { + "epoch": 0.7303691461992416, + "grad_norm": 0.1391899287700653, + "learning_rate": 3.379574335740965e-05, + "loss": 0.0153, + "step": 31157 + }, + { + "epoch": 0.7303925877740467, + "grad_norm": 0.2072363644838333, + "learning_rate": 3.379022285043735e-05, + "loss": 0.0204, + "step": 31158 + }, + { + "epoch": 0.7304160293488516, + "grad_norm": 0.276533305644989, + "learning_rate": 3.378470270272116e-05, + "loss": 0.0555, + "step": 31159 + }, + { + "epoch": 0.7304394709236567, + "grad_norm": 0.32837170362472534, + "learning_rate": 3.377918291429103e-05, + "loss": 0.0583, + "step": 31160 + }, + { + "epoch": 0.7304629124984616, + "grad_norm": 0.07357314974069595, + "learning_rate": 3.377366348517692e-05, + "loss": 0.0113, + "step": 31161 + }, + { + "epoch": 0.7304863540732667, + "grad_norm": 0.4551827907562256, + "learning_rate": 3.3768144415408755e-05, + "loss": 0.0616, + "step": 31162 + }, + { + "epoch": 0.7305097956480716, + "grad_norm": 0.2005990445613861, + "learning_rate": 3.376262570501651e-05, + "loss": 0.0188, + "step": 31163 + }, + { + "epoch": 0.7305332372228767, + "grad_norm": 0.3500541150569916, + "learning_rate": 3.375710735403011e-05, + "loss": 0.0445, + "step": 31164 + }, + { + "epoch": 0.7305566787976816, + "grad_norm": 0.34999239444732666, + "learning_rate": 3.375158936247952e-05, + "loss": 0.0551, + "step": 31165 + }, + { + "epoch": 0.7305801203724867, + "grad_norm": 0.6258870363235474, + "learning_rate": 3.374607173039468e-05, + "loss": 0.1167, + "step": 31166 + }, + { + "epoch": 0.7306035619472916, + "grad_norm": 0.18768607079982758, + "learning_rate": 3.374055445780552e-05, + "loss": 0.0284, + "step": 31167 + }, + { + "epoch": 0.7306270035220966, + "grad_norm": 0.17189978063106537, + "learning_rate": 3.373503754474197e-05, + "loss": 0.0166, + "step": 31168 + }, + { + "epoch": 0.7306504450969016, + "grad_norm": 0.12922565639019012, + "learning_rate": 3.3729520991233934e-05, + "loss": 0.0366, + "step": 31169 + }, + { + "epoch": 0.7306738866717066, + "grad_norm": 0.48775896430015564, + "learning_rate": 3.372400479731142e-05, + "loss": 0.097, + "step": 31170 + }, + { + "epoch": 0.7306973282465116, + "grad_norm": 0.34304365515708923, + "learning_rate": 3.3718488963004314e-05, + "loss": 0.0458, + "step": 31171 + }, + { + "epoch": 0.7307207698213166, + "grad_norm": 0.20822517573833466, + "learning_rate": 3.371297348834255e-05, + "loss": 0.0483, + "step": 31172 + }, + { + "epoch": 0.7307442113961216, + "grad_norm": 0.3170999586582184, + "learning_rate": 3.370745837335606e-05, + "loss": 0.0295, + "step": 31173 + }, + { + "epoch": 0.7307676529709266, + "grad_norm": 0.595348596572876, + "learning_rate": 3.370194361807476e-05, + "loss": 0.1199, + "step": 31174 + }, + { + "epoch": 0.7307910945457315, + "grad_norm": 0.13719238340854645, + "learning_rate": 3.369642922252855e-05, + "loss": 0.0168, + "step": 31175 + }, + { + "epoch": 0.7308145361205366, + "grad_norm": 0.43736860156059265, + "learning_rate": 3.369091518674741e-05, + "loss": 0.069, + "step": 31176 + }, + { + "epoch": 0.7308379776953415, + "grad_norm": 0.8433274626731873, + "learning_rate": 3.3685401510761196e-05, + "loss": 0.0756, + "step": 31177 + }, + { + "epoch": 0.7308614192701466, + "grad_norm": 0.12005329132080078, + "learning_rate": 3.367988819459989e-05, + "loss": 0.0187, + "step": 31178 + }, + { + "epoch": 0.7308848608449515, + "grad_norm": 0.12051411718130112, + "learning_rate": 3.3674375238293376e-05, + "loss": 0.0209, + "step": 31179 + }, + { + "epoch": 0.7309083024197566, + "grad_norm": 0.4223064184188843, + "learning_rate": 3.366886264187157e-05, + "loss": 0.0583, + "step": 31180 + }, + { + "epoch": 0.7309317439945615, + "grad_norm": 0.6432526111602783, + "learning_rate": 3.366335040536438e-05, + "loss": 0.135, + "step": 31181 + }, + { + "epoch": 0.7309551855693666, + "grad_norm": 0.41705122590065, + "learning_rate": 3.365783852880168e-05, + "loss": 0.0371, + "step": 31182 + }, + { + "epoch": 0.7309786271441715, + "grad_norm": 0.1368257850408554, + "learning_rate": 3.3652327012213446e-05, + "loss": 0.0209, + "step": 31183 + }, + { + "epoch": 0.7310020687189765, + "grad_norm": 0.25590434670448303, + "learning_rate": 3.3646815855629544e-05, + "loss": 0.0423, + "step": 31184 + }, + { + "epoch": 0.7310255102937815, + "grad_norm": 0.1881840080022812, + "learning_rate": 3.36413050590799e-05, + "loss": 0.0167, + "step": 31185 + }, + { + "epoch": 0.7310489518685865, + "grad_norm": 0.32120752334594727, + "learning_rate": 3.3635794622594394e-05, + "loss": 0.07, + "step": 31186 + }, + { + "epoch": 0.7310723934433915, + "grad_norm": 0.2037954330444336, + "learning_rate": 3.3630284546202926e-05, + "loss": 0.0402, + "step": 31187 + }, + { + "epoch": 0.7310958350181965, + "grad_norm": 0.5309741497039795, + "learning_rate": 3.3624774829935376e-05, + "loss": 0.5167, + "step": 31188 + }, + { + "epoch": 0.7311192765930015, + "grad_norm": 0.6462573409080505, + "learning_rate": 3.361926547382166e-05, + "loss": 0.6223, + "step": 31189 + }, + { + "epoch": 0.7311427181678065, + "grad_norm": 0.22861945629119873, + "learning_rate": 3.361375647789171e-05, + "loss": 0.031, + "step": 31190 + }, + { + "epoch": 0.7311661597426115, + "grad_norm": 0.60658860206604, + "learning_rate": 3.3608247842175386e-05, + "loss": 0.0762, + "step": 31191 + }, + { + "epoch": 0.7311896013174165, + "grad_norm": 0.29203420877456665, + "learning_rate": 3.360273956670258e-05, + "loss": 0.0305, + "step": 31192 + }, + { + "epoch": 0.7312130428922216, + "grad_norm": 0.6382582783699036, + "learning_rate": 3.3597231651503166e-05, + "loss": 0.0992, + "step": 31193 + }, + { + "epoch": 0.7312364844670265, + "grad_norm": 0.10049322992563248, + "learning_rate": 3.3591724096607046e-05, + "loss": 0.0186, + "step": 31194 + }, + { + "epoch": 0.7312599260418315, + "grad_norm": 0.7357425689697266, + "learning_rate": 3.358621690204407e-05, + "loss": 0.5172, + "step": 31195 + }, + { + "epoch": 0.7312833676166365, + "grad_norm": 0.15487931668758392, + "learning_rate": 3.358071006784417e-05, + "loss": 0.0342, + "step": 31196 + }, + { + "epoch": 0.7313068091914415, + "grad_norm": 0.14010822772979736, + "learning_rate": 3.357520359403721e-05, + "loss": 0.0281, + "step": 31197 + }, + { + "epoch": 0.7313302507662465, + "grad_norm": 0.1919609010219574, + "learning_rate": 3.3569697480653064e-05, + "loss": 0.0343, + "step": 31198 + }, + { + "epoch": 0.7313536923410515, + "grad_norm": 0.19386586546897888, + "learning_rate": 3.35641917277216e-05, + "loss": 0.0478, + "step": 31199 + }, + { + "epoch": 0.7313771339158565, + "grad_norm": 0.4376804828643799, + "learning_rate": 3.355868633527267e-05, + "loss": 0.0556, + "step": 31200 + }, + { + "epoch": 0.7314005754906615, + "grad_norm": 0.5287317037582397, + "learning_rate": 3.355318130333621e-05, + "loss": 0.6453, + "step": 31201 + }, + { + "epoch": 0.7314240170654664, + "grad_norm": 0.3334319591522217, + "learning_rate": 3.354767663194203e-05, + "loss": 0.0652, + "step": 31202 + }, + { + "epoch": 0.7314474586402715, + "grad_norm": 0.30728456377983093, + "learning_rate": 3.354217232112006e-05, + "loss": 0.0972, + "step": 31203 + }, + { + "epoch": 0.7314709002150764, + "grad_norm": 0.68560791015625, + "learning_rate": 3.353666837090013e-05, + "loss": 0.1119, + "step": 31204 + }, + { + "epoch": 0.7314943417898815, + "grad_norm": 0.2985338866710663, + "learning_rate": 3.35311647813121e-05, + "loss": 0.0512, + "step": 31205 + }, + { + "epoch": 0.7315177833646864, + "grad_norm": 0.37165483832359314, + "learning_rate": 3.352566155238583e-05, + "loss": 0.0658, + "step": 31206 + }, + { + "epoch": 0.7315412249394915, + "grad_norm": 0.44966164231300354, + "learning_rate": 3.352015868415117e-05, + "loss": 0.098, + "step": 31207 + }, + { + "epoch": 0.7315646665142964, + "grad_norm": 0.3981713652610779, + "learning_rate": 3.351465617663803e-05, + "loss": 0.0586, + "step": 31208 + }, + { + "epoch": 0.7315881080891015, + "grad_norm": 0.14176172018051147, + "learning_rate": 3.3509154029876236e-05, + "loss": 0.0275, + "step": 31209 + }, + { + "epoch": 0.7316115496639064, + "grad_norm": 0.5701539516448975, + "learning_rate": 3.3503652243895634e-05, + "loss": 0.0919, + "step": 31210 + }, + { + "epoch": 0.7316349912387115, + "grad_norm": 0.40670254826545715, + "learning_rate": 3.3498150818726096e-05, + "loss": 0.073, + "step": 31211 + }, + { + "epoch": 0.7316584328135164, + "grad_norm": 0.09643696993589401, + "learning_rate": 3.3492649754397446e-05, + "loss": 0.0151, + "step": 31212 + }, + { + "epoch": 0.7316818743883214, + "grad_norm": 0.12660419940948486, + "learning_rate": 3.3487149050939525e-05, + "loss": 0.0068, + "step": 31213 + }, + { + "epoch": 0.7317053159631264, + "grad_norm": 0.9214216470718384, + "learning_rate": 3.3481648708382205e-05, + "loss": 0.67, + "step": 31214 + }, + { + "epoch": 0.7317287575379314, + "grad_norm": 0.5644482970237732, + "learning_rate": 3.347614872675536e-05, + "loss": 0.0622, + "step": 31215 + }, + { + "epoch": 0.7317521991127364, + "grad_norm": 0.7338663339614868, + "learning_rate": 3.34706491060888e-05, + "loss": 0.1194, + "step": 31216 + }, + { + "epoch": 0.7317756406875414, + "grad_norm": 0.27184200286865234, + "learning_rate": 3.346514984641237e-05, + "loss": 0.0338, + "step": 31217 + }, + { + "epoch": 0.7317990822623464, + "grad_norm": 0.20247118175029755, + "learning_rate": 3.3459650947755916e-05, + "loss": 0.0312, + "step": 31218 + }, + { + "epoch": 0.7318225238371514, + "grad_norm": 0.4786282777786255, + "learning_rate": 3.3454152410149255e-05, + "loss": 0.0911, + "step": 31219 + }, + { + "epoch": 0.7318459654119563, + "grad_norm": 0.3858387768268585, + "learning_rate": 3.3448654233622215e-05, + "loss": 0.0384, + "step": 31220 + }, + { + "epoch": 0.7318694069867614, + "grad_norm": 0.5050376057624817, + "learning_rate": 3.344315641820467e-05, + "loss": 0.1163, + "step": 31221 + }, + { + "epoch": 0.7318928485615663, + "grad_norm": 0.22945687174797058, + "learning_rate": 3.343765896392643e-05, + "loss": 0.0505, + "step": 31222 + }, + { + "epoch": 0.7319162901363714, + "grad_norm": 0.670066773891449, + "learning_rate": 3.343216187081732e-05, + "loss": 0.097, + "step": 31223 + }, + { + "epoch": 0.7319397317111763, + "grad_norm": 0.3578874170780182, + "learning_rate": 3.342666513890719e-05, + "loss": 0.0398, + "step": 31224 + }, + { + "epoch": 0.7319631732859814, + "grad_norm": 0.9367839097976685, + "learning_rate": 3.34211687682258e-05, + "loss": 0.1642, + "step": 31225 + }, + { + "epoch": 0.7319866148607863, + "grad_norm": 0.28007546067237854, + "learning_rate": 3.341567275880305e-05, + "loss": 0.0404, + "step": 31226 + }, + { + "epoch": 0.7320100564355914, + "grad_norm": 0.19649724662303925, + "learning_rate": 3.341017711066871e-05, + "loss": 0.0342, + "step": 31227 + }, + { + "epoch": 0.7320334980103963, + "grad_norm": 0.48240625858306885, + "learning_rate": 3.3404681823852634e-05, + "loss": 0.0605, + "step": 31228 + }, + { + "epoch": 0.7320569395852013, + "grad_norm": 0.5965736508369446, + "learning_rate": 3.339918689838465e-05, + "loss": 0.1466, + "step": 31229 + }, + { + "epoch": 0.7320803811600063, + "grad_norm": 0.3370371460914612, + "learning_rate": 3.339369233429453e-05, + "loss": 0.3682, + "step": 31230 + }, + { + "epoch": 0.7321038227348113, + "grad_norm": 0.1700957715511322, + "learning_rate": 3.338819813161212e-05, + "loss": 0.0407, + "step": 31231 + }, + { + "epoch": 0.7321272643096163, + "grad_norm": 0.2331358641386032, + "learning_rate": 3.338270429036723e-05, + "loss": 0.0269, + "step": 31232 + }, + { + "epoch": 0.7321507058844213, + "grad_norm": 0.4911852180957794, + "learning_rate": 3.337721081058961e-05, + "loss": 0.1006, + "step": 31233 + }, + { + "epoch": 0.7321741474592263, + "grad_norm": 0.5325526595115662, + "learning_rate": 3.3371717692309156e-05, + "loss": 0.0879, + "step": 31234 + }, + { + "epoch": 0.7321975890340313, + "grad_norm": 0.5696461200714111, + "learning_rate": 3.336622493555565e-05, + "loss": 0.6755, + "step": 31235 + }, + { + "epoch": 0.7322210306088363, + "grad_norm": 0.4057728052139282, + "learning_rate": 3.336073254035885e-05, + "loss": 0.074, + "step": 31236 + }, + { + "epoch": 0.7322444721836413, + "grad_norm": 0.4244988262653351, + "learning_rate": 3.335524050674857e-05, + "loss": 0.0679, + "step": 31237 + }, + { + "epoch": 0.7322679137584462, + "grad_norm": 0.45093071460723877, + "learning_rate": 3.334974883475468e-05, + "loss": 0.1085, + "step": 31238 + }, + { + "epoch": 0.7322913553332513, + "grad_norm": 0.6030367016792297, + "learning_rate": 3.334425752440691e-05, + "loss": 0.162, + "step": 31239 + }, + { + "epoch": 0.7323147969080562, + "grad_norm": 0.13884791731834412, + "learning_rate": 3.333876657573504e-05, + "loss": 0.0222, + "step": 31240 + }, + { + "epoch": 0.7323382384828613, + "grad_norm": 0.38133081793785095, + "learning_rate": 3.333327598876893e-05, + "loss": 0.0416, + "step": 31241 + }, + { + "epoch": 0.7323616800576662, + "grad_norm": 0.11276727169752121, + "learning_rate": 3.3327785763538343e-05, + "loss": 0.0163, + "step": 31242 + }, + { + "epoch": 0.7323851216324713, + "grad_norm": 0.7605501413345337, + "learning_rate": 3.3322295900073064e-05, + "loss": 0.1521, + "step": 31243 + }, + { + "epoch": 0.7324085632072763, + "grad_norm": 0.10304580628871918, + "learning_rate": 3.331680639840288e-05, + "loss": 0.0225, + "step": 31244 + }, + { + "epoch": 0.7324320047820813, + "grad_norm": 0.4587010145187378, + "learning_rate": 3.331131725855754e-05, + "loss": 0.0852, + "step": 31245 + }, + { + "epoch": 0.7324554463568863, + "grad_norm": 0.48745524883270264, + "learning_rate": 3.330582848056691e-05, + "loss": 0.5098, + "step": 31246 + }, + { + "epoch": 0.7324788879316912, + "grad_norm": 0.4061453938484192, + "learning_rate": 3.330034006446073e-05, + "loss": 0.0933, + "step": 31247 + }, + { + "epoch": 0.7325023295064963, + "grad_norm": 0.16692714393138885, + "learning_rate": 3.329485201026877e-05, + "loss": 0.028, + "step": 31248 + }, + { + "epoch": 0.7325257710813012, + "grad_norm": 0.31504955887794495, + "learning_rate": 3.328936431802083e-05, + "loss": 0.0421, + "step": 31249 + }, + { + "epoch": 0.7325492126561063, + "grad_norm": 0.3812701404094696, + "learning_rate": 3.3283876987746644e-05, + "loss": 0.0741, + "step": 31250 + }, + { + "epoch": 0.7325726542309112, + "grad_norm": 0.1253756880760193, + "learning_rate": 3.327839001947605e-05, + "loss": 0.0134, + "step": 31251 + }, + { + "epoch": 0.7325960958057163, + "grad_norm": 0.21537357568740845, + "learning_rate": 3.3272903413238745e-05, + "loss": 0.1585, + "step": 31252 + }, + { + "epoch": 0.7326195373805212, + "grad_norm": 0.21989120543003082, + "learning_rate": 3.326741716906458e-05, + "loss": 0.0235, + "step": 31253 + }, + { + "epoch": 0.7326429789553263, + "grad_norm": 0.22254645824432373, + "learning_rate": 3.3261931286983286e-05, + "loss": 0.0333, + "step": 31254 + }, + { + "epoch": 0.7326664205301312, + "grad_norm": 0.41468292474746704, + "learning_rate": 3.3256445767024626e-05, + "loss": 0.1001, + "step": 31255 + }, + { + "epoch": 0.7326898621049363, + "grad_norm": 0.2872529625892639, + "learning_rate": 3.325096060921837e-05, + "loss": 0.0342, + "step": 31256 + }, + { + "epoch": 0.7327133036797412, + "grad_norm": 0.2214190661907196, + "learning_rate": 3.3245475813594286e-05, + "loss": 0.032, + "step": 31257 + }, + { + "epoch": 0.7327367452545462, + "grad_norm": 0.5285380482673645, + "learning_rate": 3.323999138018208e-05, + "loss": 0.1059, + "step": 31258 + }, + { + "epoch": 0.7327601868293512, + "grad_norm": 0.3022468686103821, + "learning_rate": 3.32345073090116e-05, + "loss": 0.0277, + "step": 31259 + }, + { + "epoch": 0.7327836284041562, + "grad_norm": 0.3511439263820648, + "learning_rate": 3.322902360011256e-05, + "loss": 0.0434, + "step": 31260 + }, + { + "epoch": 0.7328070699789612, + "grad_norm": 0.14822949469089508, + "learning_rate": 3.322354025351472e-05, + "loss": 0.0204, + "step": 31261 + }, + { + "epoch": 0.7328305115537662, + "grad_norm": 0.5900216102600098, + "learning_rate": 3.321805726924778e-05, + "loss": 0.1256, + "step": 31262 + }, + { + "epoch": 0.7328539531285712, + "grad_norm": 0.5335663557052612, + "learning_rate": 3.3212574647341586e-05, + "loss": 0.7361, + "step": 31263 + }, + { + "epoch": 0.7328773947033762, + "grad_norm": 0.24653741717338562, + "learning_rate": 3.320709238782583e-05, + "loss": 0.0222, + "step": 31264 + }, + { + "epoch": 0.7329008362781811, + "grad_norm": 0.526483952999115, + "learning_rate": 3.320161049073025e-05, + "loss": 0.115, + "step": 31265 + }, + { + "epoch": 0.7329242778529862, + "grad_norm": 0.10771024972200394, + "learning_rate": 3.3196128956084626e-05, + "loss": 0.0205, + "step": 31266 + }, + { + "epoch": 0.7329477194277911, + "grad_norm": 0.3567804992198944, + "learning_rate": 3.3190647783918695e-05, + "loss": 0.019, + "step": 31267 + }, + { + "epoch": 0.7329711610025962, + "grad_norm": 0.24533316493034363, + "learning_rate": 3.318516697426219e-05, + "loss": 0.0251, + "step": 31268 + }, + { + "epoch": 0.7329946025774011, + "grad_norm": 0.5565934181213379, + "learning_rate": 3.3179686527144834e-05, + "loss": 0.0918, + "step": 31269 + }, + { + "epoch": 0.7330180441522062, + "grad_norm": 0.10806337743997574, + "learning_rate": 3.31742064425964e-05, + "loss": 0.018, + "step": 31270 + }, + { + "epoch": 0.7330414857270111, + "grad_norm": 1.0213738679885864, + "learning_rate": 3.316872672064655e-05, + "loss": 0.2059, + "step": 31271 + }, + { + "epoch": 0.7330649273018162, + "grad_norm": 0.048198334872722626, + "learning_rate": 3.31632473613251e-05, + "loss": 0.0056, + "step": 31272 + }, + { + "epoch": 0.7330883688766211, + "grad_norm": 0.12152162194252014, + "learning_rate": 3.315776836466175e-05, + "loss": 0.0196, + "step": 31273 + }, + { + "epoch": 0.7331118104514261, + "grad_norm": 0.6553642749786377, + "learning_rate": 3.315228973068622e-05, + "loss": 0.0768, + "step": 31274 + }, + { + "epoch": 0.7331352520262311, + "grad_norm": 0.1283283829689026, + "learning_rate": 3.3146811459428254e-05, + "loss": 0.0175, + "step": 31275 + }, + { + "epoch": 0.7331586936010361, + "grad_norm": 1.797663688659668, + "learning_rate": 3.3141333550917565e-05, + "loss": 0.0894, + "step": 31276 + }, + { + "epoch": 0.7331821351758411, + "grad_norm": 0.493161678314209, + "learning_rate": 3.31358560051839e-05, + "loss": 0.0727, + "step": 31277 + }, + { + "epoch": 0.7332055767506461, + "grad_norm": 0.5315675139427185, + "learning_rate": 3.3130378822256905e-05, + "loss": 0.0992, + "step": 31278 + }, + { + "epoch": 0.7332290183254511, + "grad_norm": 0.3428191542625427, + "learning_rate": 3.31249020021664e-05, + "loss": 0.077, + "step": 31279 + }, + { + "epoch": 0.7332524599002561, + "grad_norm": 0.4078989028930664, + "learning_rate": 3.311942554494205e-05, + "loss": 0.0879, + "step": 31280 + }, + { + "epoch": 0.733275901475061, + "grad_norm": 0.4749375879764557, + "learning_rate": 3.311394945061359e-05, + "loss": 0.0893, + "step": 31281 + }, + { + "epoch": 0.7332993430498661, + "grad_norm": 0.8266035318374634, + "learning_rate": 3.310847371921071e-05, + "loss": 0.1773, + "step": 31282 + }, + { + "epoch": 0.733322784624671, + "grad_norm": 0.3475908637046814, + "learning_rate": 3.3102998350763105e-05, + "loss": 0.0659, + "step": 31283 + }, + { + "epoch": 0.7333462261994761, + "grad_norm": 0.447782039642334, + "learning_rate": 3.309752334530054e-05, + "loss": 0.0429, + "step": 31284 + }, + { + "epoch": 0.733369667774281, + "grad_norm": 0.22862137854099274, + "learning_rate": 3.3092048702852705e-05, + "loss": 0.0354, + "step": 31285 + }, + { + "epoch": 0.7333931093490861, + "grad_norm": 0.1553971916437149, + "learning_rate": 3.308657442344926e-05, + "loss": 0.0263, + "step": 31286 + }, + { + "epoch": 0.733416550923891, + "grad_norm": 0.41954824328422546, + "learning_rate": 3.3081100507119965e-05, + "loss": 0.0506, + "step": 31287 + }, + { + "epoch": 0.7334399924986961, + "grad_norm": 0.43432143330574036, + "learning_rate": 3.30756269538945e-05, + "loss": 0.5557, + "step": 31288 + }, + { + "epoch": 0.733463434073501, + "grad_norm": 0.14207379519939423, + "learning_rate": 3.307015376380258e-05, + "loss": 0.0308, + "step": 31289 + }, + { + "epoch": 0.7334868756483061, + "grad_norm": 0.2754247486591339, + "learning_rate": 3.306468093687385e-05, + "loss": 0.0229, + "step": 31290 + }, + { + "epoch": 0.733510317223111, + "grad_norm": 0.4180386960506439, + "learning_rate": 3.305920847313807e-05, + "loss": 0.0583, + "step": 31291 + }, + { + "epoch": 0.733533758797916, + "grad_norm": 0.9258527159690857, + "learning_rate": 3.305373637262491e-05, + "loss": 0.2092, + "step": 31292 + }, + { + "epoch": 0.733557200372721, + "grad_norm": 0.5689691305160522, + "learning_rate": 3.304826463536406e-05, + "loss": 0.0856, + "step": 31293 + }, + { + "epoch": 0.733580641947526, + "grad_norm": 0.27271342277526855, + "learning_rate": 3.3042793261385195e-05, + "loss": 0.0419, + "step": 31294 + }, + { + "epoch": 0.7336040835223311, + "grad_norm": 0.5146679878234863, + "learning_rate": 3.303732225071804e-05, + "loss": 0.0816, + "step": 31295 + }, + { + "epoch": 0.733627525097136, + "grad_norm": 0.3698579967021942, + "learning_rate": 3.3031851603392205e-05, + "loss": 0.0512, + "step": 31296 + }, + { + "epoch": 0.7336509666719411, + "grad_norm": 0.15940190851688385, + "learning_rate": 3.302638131943747e-05, + "loss": 0.027, + "step": 31297 + }, + { + "epoch": 0.733674408246746, + "grad_norm": 0.32244518399238586, + "learning_rate": 3.302091139888344e-05, + "loss": 0.0493, + "step": 31298 + }, + { + "epoch": 0.7336978498215511, + "grad_norm": 0.3039165139198303, + "learning_rate": 3.301544184175985e-05, + "loss": 0.047, + "step": 31299 + }, + { + "epoch": 0.733721291396356, + "grad_norm": 0.0270344540476799, + "learning_rate": 3.3009972648096364e-05, + "loss": 0.0016, + "step": 31300 + }, + { + "epoch": 0.733744732971161, + "grad_norm": 0.17524327337741852, + "learning_rate": 3.300450381792265e-05, + "loss": 0.0323, + "step": 31301 + }, + { + "epoch": 0.733768174545966, + "grad_norm": 0.2150505781173706, + "learning_rate": 3.299903535126838e-05, + "loss": 0.0582, + "step": 31302 + }, + { + "epoch": 0.733791616120771, + "grad_norm": 0.4271916151046753, + "learning_rate": 3.29935672481632e-05, + "loss": 0.0508, + "step": 31303 + }, + { + "epoch": 0.733815057695576, + "grad_norm": 0.14581939578056335, + "learning_rate": 3.298809950863684e-05, + "loss": 0.03, + "step": 31304 + }, + { + "epoch": 0.733838499270381, + "grad_norm": 0.6059701442718506, + "learning_rate": 3.2982632132718935e-05, + "loss": 0.0591, + "step": 31305 + }, + { + "epoch": 0.733861940845186, + "grad_norm": 0.3366732597351074, + "learning_rate": 3.2977165120439145e-05, + "loss": 0.0688, + "step": 31306 + }, + { + "epoch": 0.733885382419991, + "grad_norm": 0.39717897772789, + "learning_rate": 3.297169847182715e-05, + "loss": 0.0886, + "step": 31307 + }, + { + "epoch": 0.733908823994796, + "grad_norm": 0.5508422255516052, + "learning_rate": 3.29662321869126e-05, + "loss": 0.5159, + "step": 31308 + }, + { + "epoch": 0.733932265569601, + "grad_norm": 0.659581184387207, + "learning_rate": 3.2960766265725144e-05, + "loss": 0.1151, + "step": 31309 + }, + { + "epoch": 0.7339557071444059, + "grad_norm": 0.39442178606987, + "learning_rate": 3.295530070829443e-05, + "loss": 0.1007, + "step": 31310 + }, + { + "epoch": 0.733979148719211, + "grad_norm": 0.15672218799591064, + "learning_rate": 3.294983551465018e-05, + "loss": 0.0212, + "step": 31311 + }, + { + "epoch": 0.7340025902940159, + "grad_norm": 0.4804988503456116, + "learning_rate": 3.294437068482201e-05, + "loss": 0.5235, + "step": 31312 + }, + { + "epoch": 0.734026031868821, + "grad_norm": 0.3654344379901886, + "learning_rate": 3.293890621883957e-05, + "loss": 0.0425, + "step": 31313 + }, + { + "epoch": 0.7340494734436259, + "grad_norm": 0.5216490626335144, + "learning_rate": 3.293344211673252e-05, + "loss": 0.1135, + "step": 31314 + }, + { + "epoch": 0.734072915018431, + "grad_norm": 0.494693398475647, + "learning_rate": 3.292797837853048e-05, + "loss": 0.0919, + "step": 31315 + }, + { + "epoch": 0.7340963565932359, + "grad_norm": 0.5782077312469482, + "learning_rate": 3.2922515004263086e-05, + "loss": 0.1344, + "step": 31316 + }, + { + "epoch": 0.734119798168041, + "grad_norm": 0.2307511866092682, + "learning_rate": 3.291705199396006e-05, + "loss": 0.0439, + "step": 31317 + }, + { + "epoch": 0.7341432397428459, + "grad_norm": 0.34513458609580994, + "learning_rate": 3.291158934765099e-05, + "loss": 0.055, + "step": 31318 + }, + { + "epoch": 0.734166681317651, + "grad_norm": 0.4150850474834442, + "learning_rate": 3.290612706536551e-05, + "loss": 0.083, + "step": 31319 + }, + { + "epoch": 0.7341901228924559, + "grad_norm": 0.6710712313652039, + "learning_rate": 3.290066514713328e-05, + "loss": 0.0963, + "step": 31320 + }, + { + "epoch": 0.7342135644672609, + "grad_norm": 0.18266446888446808, + "learning_rate": 3.289520359298389e-05, + "loss": 0.0226, + "step": 31321 + }, + { + "epoch": 0.7342370060420659, + "grad_norm": 0.43020668625831604, + "learning_rate": 3.288974240294706e-05, + "loss": 0.09, + "step": 31322 + }, + { + "epoch": 0.7342604476168709, + "grad_norm": 0.4558519124984741, + "learning_rate": 3.288428157705232e-05, + "loss": 0.0875, + "step": 31323 + }, + { + "epoch": 0.7342838891916759, + "grad_norm": 0.362234503030777, + "learning_rate": 3.28788211153294e-05, + "loss": 0.0593, + "step": 31324 + }, + { + "epoch": 0.7343073307664809, + "grad_norm": 0.5053828358650208, + "learning_rate": 3.287336101780789e-05, + "loss": 0.0789, + "step": 31325 + }, + { + "epoch": 0.7343307723412859, + "grad_norm": 0.8809502720832825, + "learning_rate": 3.28679012845174e-05, + "loss": 0.1219, + "step": 31326 + }, + { + "epoch": 0.7343542139160909, + "grad_norm": 0.2595783472061157, + "learning_rate": 3.286244191548757e-05, + "loss": 0.0311, + "step": 31327 + }, + { + "epoch": 0.7343776554908958, + "grad_norm": 0.4022914469242096, + "learning_rate": 3.2856982910747977e-05, + "loss": 0.0377, + "step": 31328 + }, + { + "epoch": 0.7344010970657009, + "grad_norm": 0.2447831928730011, + "learning_rate": 3.285152427032832e-05, + "loss": 0.0296, + "step": 31329 + }, + { + "epoch": 0.7344245386405058, + "grad_norm": 0.3793604373931885, + "learning_rate": 3.284606599425819e-05, + "loss": 0.0793, + "step": 31330 + }, + { + "epoch": 0.7344479802153109, + "grad_norm": 0.8508111238479614, + "learning_rate": 3.284060808256719e-05, + "loss": 0.1056, + "step": 31331 + }, + { + "epoch": 0.7344714217901158, + "grad_norm": 0.4552057087421417, + "learning_rate": 3.283515053528493e-05, + "loss": 0.0709, + "step": 31332 + }, + { + "epoch": 0.7344948633649209, + "grad_norm": 0.09743682295084, + "learning_rate": 3.282969335244104e-05, + "loss": 0.0173, + "step": 31333 + }, + { + "epoch": 0.7345183049397258, + "grad_norm": 0.5989415645599365, + "learning_rate": 3.282423653406509e-05, + "loss": 0.0926, + "step": 31334 + }, + { + "epoch": 0.7345417465145309, + "grad_norm": 0.44269636273384094, + "learning_rate": 3.2818780080186715e-05, + "loss": 0.0782, + "step": 31335 + }, + { + "epoch": 0.7345651880893358, + "grad_norm": 0.5360259413719177, + "learning_rate": 3.281332399083556e-05, + "loss": 0.5897, + "step": 31336 + }, + { + "epoch": 0.7345886296641408, + "grad_norm": 0.3851271867752075, + "learning_rate": 3.280786826604121e-05, + "loss": 0.0522, + "step": 31337 + }, + { + "epoch": 0.7346120712389458, + "grad_norm": 0.35648050904273987, + "learning_rate": 3.2802412905833245e-05, + "loss": 0.0497, + "step": 31338 + }, + { + "epoch": 0.7346355128137508, + "grad_norm": 0.26929983496665955, + "learning_rate": 3.279695791024129e-05, + "loss": 0.0463, + "step": 31339 + }, + { + "epoch": 0.7346589543885558, + "grad_norm": 0.529978334903717, + "learning_rate": 3.279150327929492e-05, + "loss": 0.0681, + "step": 31340 + }, + { + "epoch": 0.7346823959633608, + "grad_norm": 0.9988787770271301, + "learning_rate": 3.278604901302371e-05, + "loss": 0.0473, + "step": 31341 + }, + { + "epoch": 0.7347058375381658, + "grad_norm": 0.4037635624408722, + "learning_rate": 3.278059511145733e-05, + "loss": 0.0534, + "step": 31342 + }, + { + "epoch": 0.7347292791129708, + "grad_norm": 0.3613613247871399, + "learning_rate": 3.277514157462531e-05, + "loss": 0.0571, + "step": 31343 + }, + { + "epoch": 0.7347527206877758, + "grad_norm": 1.6444365978240967, + "learning_rate": 3.2769688402557287e-05, + "loss": 0.0843, + "step": 31344 + }, + { + "epoch": 0.7347761622625808, + "grad_norm": 0.2999073266983032, + "learning_rate": 3.276423559528281e-05, + "loss": 0.0291, + "step": 31345 + }, + { + "epoch": 0.7347996038373859, + "grad_norm": 0.4839411973953247, + "learning_rate": 3.2758783152831485e-05, + "loss": 0.4881, + "step": 31346 + }, + { + "epoch": 0.7348230454121908, + "grad_norm": 0.3462049663066864, + "learning_rate": 3.2753331075232865e-05, + "loss": 0.0501, + "step": 31347 + }, + { + "epoch": 0.7348464869869958, + "grad_norm": 0.5455712676048279, + "learning_rate": 3.274787936251656e-05, + "loss": 0.1316, + "step": 31348 + }, + { + "epoch": 0.7348699285618008, + "grad_norm": 0.6524198055267334, + "learning_rate": 3.274242801471218e-05, + "loss": 0.0377, + "step": 31349 + }, + { + "epoch": 0.7348933701366058, + "grad_norm": 0.256996214389801, + "learning_rate": 3.273697703184927e-05, + "loss": 0.0291, + "step": 31350 + }, + { + "epoch": 0.7349168117114108, + "grad_norm": 0.6320885419845581, + "learning_rate": 3.273152641395742e-05, + "loss": 0.0818, + "step": 31351 + }, + { + "epoch": 0.7349402532862158, + "grad_norm": 0.6179653406143188, + "learning_rate": 3.272607616106621e-05, + "loss": 0.1067, + "step": 31352 + }, + { + "epoch": 0.7349636948610208, + "grad_norm": 0.3481098413467407, + "learning_rate": 3.272062627320519e-05, + "loss": 0.0639, + "step": 31353 + }, + { + "epoch": 0.7349871364358258, + "grad_norm": 0.45104748010635376, + "learning_rate": 3.271517675040392e-05, + "loss": 0.3715, + "step": 31354 + }, + { + "epoch": 0.7350105780106307, + "grad_norm": 0.2563926875591278, + "learning_rate": 3.2709727592692016e-05, + "loss": 0.0692, + "step": 31355 + }, + { + "epoch": 0.7350340195854358, + "grad_norm": 0.28796595335006714, + "learning_rate": 3.270427880009903e-05, + "loss": 0.0594, + "step": 31356 + }, + { + "epoch": 0.7350574611602407, + "grad_norm": 0.4408982992172241, + "learning_rate": 3.269883037265451e-05, + "loss": 0.0504, + "step": 31357 + }, + { + "epoch": 0.7350809027350458, + "grad_norm": 0.3722529709339142, + "learning_rate": 3.2693382310388046e-05, + "loss": 0.0571, + "step": 31358 + }, + { + "epoch": 0.7351043443098507, + "grad_norm": 0.5951363444328308, + "learning_rate": 3.268793461332913e-05, + "loss": 0.5971, + "step": 31359 + }, + { + "epoch": 0.7351277858846558, + "grad_norm": 0.5863786935806274, + "learning_rate": 3.268248728150738e-05, + "loss": 0.7549, + "step": 31360 + }, + { + "epoch": 0.7351512274594607, + "grad_norm": 0.23356059193611145, + "learning_rate": 3.267704031495239e-05, + "loss": 0.26, + "step": 31361 + }, + { + "epoch": 0.7351746690342658, + "grad_norm": 0.5821707248687744, + "learning_rate": 3.267159371369366e-05, + "loss": 0.0655, + "step": 31362 + }, + { + "epoch": 0.7351981106090707, + "grad_norm": 0.0862622931599617, + "learning_rate": 3.266614747776075e-05, + "loss": 0.0121, + "step": 31363 + }, + { + "epoch": 0.7352215521838757, + "grad_norm": 0.6324855089187622, + "learning_rate": 3.266070160718323e-05, + "loss": 0.0846, + "step": 31364 + }, + { + "epoch": 0.7352449937586807, + "grad_norm": 0.5742659568786621, + "learning_rate": 3.265525610199064e-05, + "loss": 0.5508, + "step": 31365 + }, + { + "epoch": 0.7352684353334857, + "grad_norm": 0.6979444622993469, + "learning_rate": 3.264981096221248e-05, + "loss": 0.1053, + "step": 31366 + }, + { + "epoch": 0.7352918769082907, + "grad_norm": 0.377060204744339, + "learning_rate": 3.2644366187878375e-05, + "loss": 0.0632, + "step": 31367 + }, + { + "epoch": 0.7353153184830957, + "grad_norm": 0.2694321870803833, + "learning_rate": 3.263892177901784e-05, + "loss": 0.0316, + "step": 31368 + }, + { + "epoch": 0.7353387600579007, + "grad_norm": 0.37928661704063416, + "learning_rate": 3.26334777356604e-05, + "loss": 0.0704, + "step": 31369 + }, + { + "epoch": 0.7353622016327057, + "grad_norm": 0.5159097909927368, + "learning_rate": 3.262803405783561e-05, + "loss": 0.4977, + "step": 31370 + }, + { + "epoch": 0.7353856432075107, + "grad_norm": 1.1265941858291626, + "learning_rate": 3.2622590745572957e-05, + "loss": 0.2203, + "step": 31371 + }, + { + "epoch": 0.7354090847823157, + "grad_norm": 0.3681134283542633, + "learning_rate": 3.261714779890207e-05, + "loss": 0.058, + "step": 31372 + }, + { + "epoch": 0.7354325263571206, + "grad_norm": 0.41896340250968933, + "learning_rate": 3.2611705217852393e-05, + "loss": 0.1166, + "step": 31373 + }, + { + "epoch": 0.7354559679319257, + "grad_norm": 0.44997894763946533, + "learning_rate": 3.260626300245353e-05, + "loss": 0.0999, + "step": 31374 + }, + { + "epoch": 0.7354794095067306, + "grad_norm": 0.3082268536090851, + "learning_rate": 3.260082115273498e-05, + "loss": 0.0317, + "step": 31375 + }, + { + "epoch": 0.7355028510815357, + "grad_norm": 0.3171860873699188, + "learning_rate": 3.259537966872627e-05, + "loss": 0.0534, + "step": 31376 + }, + { + "epoch": 0.7355262926563406, + "grad_norm": 0.08353181183338165, + "learning_rate": 3.258993855045692e-05, + "loss": 0.0138, + "step": 31377 + }, + { + "epoch": 0.7355497342311457, + "grad_norm": 0.39598584175109863, + "learning_rate": 3.258449779795647e-05, + "loss": 0.0343, + "step": 31378 + }, + { + "epoch": 0.7355731758059506, + "grad_norm": 0.5644581913948059, + "learning_rate": 3.257905741125439e-05, + "loss": 0.7497, + "step": 31379 + }, + { + "epoch": 0.7355966173807557, + "grad_norm": 0.45603615045547485, + "learning_rate": 3.2573617390380276e-05, + "loss": 0.0783, + "step": 31380 + }, + { + "epoch": 0.7356200589555606, + "grad_norm": 0.33469635248184204, + "learning_rate": 3.2568177735363615e-05, + "loss": 0.0476, + "step": 31381 + }, + { + "epoch": 0.7356435005303656, + "grad_norm": 0.2911401689052582, + "learning_rate": 3.256273844623391e-05, + "loss": 0.0233, + "step": 31382 + }, + { + "epoch": 0.7356669421051706, + "grad_norm": 0.23629243671894073, + "learning_rate": 3.2557299523020644e-05, + "loss": 0.0394, + "step": 31383 + }, + { + "epoch": 0.7356903836799756, + "grad_norm": 0.08420107513666153, + "learning_rate": 3.255186096575341e-05, + "loss": 0.0194, + "step": 31384 + }, + { + "epoch": 0.7357138252547806, + "grad_norm": 0.47661444544792175, + "learning_rate": 3.254642277446167e-05, + "loss": 0.0891, + "step": 31385 + }, + { + "epoch": 0.7357372668295856, + "grad_norm": 0.39781221747398376, + "learning_rate": 3.254098494917491e-05, + "loss": 0.0401, + "step": 31386 + }, + { + "epoch": 0.7357607084043906, + "grad_norm": 0.1755579710006714, + "learning_rate": 3.2535547489922694e-05, + "loss": 0.0128, + "step": 31387 + }, + { + "epoch": 0.7357841499791956, + "grad_norm": 0.6744460463523865, + "learning_rate": 3.2530110396734494e-05, + "loss": 0.0678, + "step": 31388 + }, + { + "epoch": 0.7358075915540006, + "grad_norm": 0.12771263718605042, + "learning_rate": 3.25246736696398e-05, + "loss": 0.0199, + "step": 31389 + }, + { + "epoch": 0.7358310331288056, + "grad_norm": 0.4670644998550415, + "learning_rate": 3.251923730866814e-05, + "loss": 0.0733, + "step": 31390 + }, + { + "epoch": 0.7358544747036105, + "grad_norm": 0.530553936958313, + "learning_rate": 3.251380131384899e-05, + "loss": 0.1319, + "step": 31391 + }, + { + "epoch": 0.7358779162784156, + "grad_norm": 0.23538430035114288, + "learning_rate": 3.2508365685211815e-05, + "loss": 0.0304, + "step": 31392 + }, + { + "epoch": 0.7359013578532205, + "grad_norm": 0.19300374388694763, + "learning_rate": 3.250293042278618e-05, + "loss": 0.0343, + "step": 31393 + }, + { + "epoch": 0.7359247994280256, + "grad_norm": 0.13766463100910187, + "learning_rate": 3.249749552660154e-05, + "loss": 0.0398, + "step": 31394 + }, + { + "epoch": 0.7359482410028305, + "grad_norm": 0.5233825445175171, + "learning_rate": 3.2492060996687356e-05, + "loss": 0.1203, + "step": 31395 + }, + { + "epoch": 0.7359716825776356, + "grad_norm": 0.5087752938270569, + "learning_rate": 3.2486626833073186e-05, + "loss": 0.0683, + "step": 31396 + }, + { + "epoch": 0.7359951241524406, + "grad_norm": 0.2307388037443161, + "learning_rate": 3.248119303578847e-05, + "loss": 0.0393, + "step": 31397 + }, + { + "epoch": 0.7360185657272456, + "grad_norm": 0.5663823485374451, + "learning_rate": 3.2475759604862676e-05, + "loss": 0.105, + "step": 31398 + }, + { + "epoch": 0.7360420073020506, + "grad_norm": 0.33329057693481445, + "learning_rate": 3.247032654032534e-05, + "loss": 0.0374, + "step": 31399 + }, + { + "epoch": 0.7360654488768555, + "grad_norm": 0.1692197322845459, + "learning_rate": 3.24648938422059e-05, + "loss": 0.0361, + "step": 31400 + }, + { + "epoch": 0.7360888904516606, + "grad_norm": 0.7317336201667786, + "learning_rate": 3.245946151053385e-05, + "loss": 0.0713, + "step": 31401 + }, + { + "epoch": 0.7361123320264655, + "grad_norm": 0.39799103140830994, + "learning_rate": 3.245402954533866e-05, + "loss": 0.0666, + "step": 31402 + }, + { + "epoch": 0.7361357736012706, + "grad_norm": 0.14212796092033386, + "learning_rate": 3.244859794664981e-05, + "loss": 0.0148, + "step": 31403 + }, + { + "epoch": 0.7361592151760755, + "grad_norm": 0.5270010232925415, + "learning_rate": 3.244316671449673e-05, + "loss": 0.1159, + "step": 31404 + }, + { + "epoch": 0.7361826567508806, + "grad_norm": 0.480446457862854, + "learning_rate": 3.2437735848908954e-05, + "loss": 0.0636, + "step": 31405 + }, + { + "epoch": 0.7362060983256855, + "grad_norm": 0.16621455550193787, + "learning_rate": 3.2432305349915936e-05, + "loss": 0.0375, + "step": 31406 + }, + { + "epoch": 0.7362295399004906, + "grad_norm": 0.4862705171108246, + "learning_rate": 3.242687521754713e-05, + "loss": 0.1149, + "step": 31407 + }, + { + "epoch": 0.7362529814752955, + "grad_norm": 0.33318573236465454, + "learning_rate": 3.2421445451831956e-05, + "loss": 0.1268, + "step": 31408 + }, + { + "epoch": 0.7362764230501005, + "grad_norm": 0.23936057090759277, + "learning_rate": 3.241601605279995e-05, + "loss": 0.0458, + "step": 31409 + }, + { + "epoch": 0.7362998646249055, + "grad_norm": 0.6197687983512878, + "learning_rate": 3.241058702048054e-05, + "loss": 0.1349, + "step": 31410 + }, + { + "epoch": 0.7363233061997105, + "grad_norm": 0.5197668671607971, + "learning_rate": 3.2405158354903164e-05, + "loss": 0.6552, + "step": 31411 + }, + { + "epoch": 0.7363467477745155, + "grad_norm": 0.31676918268203735, + "learning_rate": 3.2399730056097335e-05, + "loss": 0.0448, + "step": 31412 + }, + { + "epoch": 0.7363701893493205, + "grad_norm": 0.531955361366272, + "learning_rate": 3.2394302124092456e-05, + "loss": 0.5051, + "step": 31413 + }, + { + "epoch": 0.7363936309241255, + "grad_norm": 0.4075627624988556, + "learning_rate": 3.2388874558918004e-05, + "loss": 0.0794, + "step": 31414 + }, + { + "epoch": 0.7364170724989305, + "grad_norm": 0.3194440007209778, + "learning_rate": 3.2383447360603414e-05, + "loss": 0.0363, + "step": 31415 + }, + { + "epoch": 0.7364405140737355, + "grad_norm": 0.35859695076942444, + "learning_rate": 3.2378020529178144e-05, + "loss": 0.0418, + "step": 31416 + }, + { + "epoch": 0.7364639556485405, + "grad_norm": 0.3203330636024475, + "learning_rate": 3.23725940646716e-05, + "loss": 0.0554, + "step": 31417 + }, + { + "epoch": 0.7364873972233454, + "grad_norm": 0.642353355884552, + "learning_rate": 3.23671679671133e-05, + "loss": 0.1095, + "step": 31418 + }, + { + "epoch": 0.7365108387981505, + "grad_norm": 0.3758309483528137, + "learning_rate": 3.236174223653264e-05, + "loss": 0.0587, + "step": 31419 + }, + { + "epoch": 0.7365342803729554, + "grad_norm": 0.12499048560857773, + "learning_rate": 3.2356316872959044e-05, + "loss": 0.0279, + "step": 31420 + }, + { + "epoch": 0.7365577219477605, + "grad_norm": 0.326261967420578, + "learning_rate": 3.2350891876422007e-05, + "loss": 0.0562, + "step": 31421 + }, + { + "epoch": 0.7365811635225654, + "grad_norm": 0.06700674444437027, + "learning_rate": 3.234546724695093e-05, + "loss": 0.0069, + "step": 31422 + }, + { + "epoch": 0.7366046050973705, + "grad_norm": 0.561851441860199, + "learning_rate": 3.234004298457525e-05, + "loss": 0.6278, + "step": 31423 + }, + { + "epoch": 0.7366280466721754, + "grad_norm": 0.29044830799102783, + "learning_rate": 3.233461908932437e-05, + "loss": 0.04, + "step": 31424 + }, + { + "epoch": 0.7366514882469805, + "grad_norm": 0.25580060482025146, + "learning_rate": 3.2329195561227776e-05, + "loss": 0.0575, + "step": 31425 + }, + { + "epoch": 0.7366749298217854, + "grad_norm": 0.3363209068775177, + "learning_rate": 3.232377240031488e-05, + "loss": 0.0372, + "step": 31426 + }, + { + "epoch": 0.7366983713965904, + "grad_norm": 0.5428639054298401, + "learning_rate": 3.2318349606615086e-05, + "loss": 0.1177, + "step": 31427 + }, + { + "epoch": 0.7367218129713954, + "grad_norm": 0.4024301767349243, + "learning_rate": 3.231292718015784e-05, + "loss": 0.0604, + "step": 31428 + }, + { + "epoch": 0.7367452545462004, + "grad_norm": 0.5607210397720337, + "learning_rate": 3.2307505120972516e-05, + "loss": 0.0708, + "step": 31429 + }, + { + "epoch": 0.7367686961210054, + "grad_norm": 0.4504684507846832, + "learning_rate": 3.2302083429088606e-05, + "loss": 0.0801, + "step": 31430 + }, + { + "epoch": 0.7367921376958104, + "grad_norm": 0.13982953131198883, + "learning_rate": 3.229666210453549e-05, + "loss": 0.0223, + "step": 31431 + }, + { + "epoch": 0.7368155792706154, + "grad_norm": 0.34812700748443604, + "learning_rate": 3.229124114734255e-05, + "loss": 0.0616, + "step": 31432 + }, + { + "epoch": 0.7368390208454204, + "grad_norm": 0.4061059355735779, + "learning_rate": 3.2285820557539295e-05, + "loss": 0.0659, + "step": 31433 + }, + { + "epoch": 0.7368624624202254, + "grad_norm": 0.3154778480529785, + "learning_rate": 3.228040033515507e-05, + "loss": 0.0453, + "step": 31434 + }, + { + "epoch": 0.7368859039950304, + "grad_norm": 0.548225462436676, + "learning_rate": 3.2274980480219296e-05, + "loss": 0.0979, + "step": 31435 + }, + { + "epoch": 0.7369093455698353, + "grad_norm": 0.1759803742170334, + "learning_rate": 3.226956099276135e-05, + "loss": 0.0413, + "step": 31436 + }, + { + "epoch": 0.7369327871446404, + "grad_norm": 0.5614923238754272, + "learning_rate": 3.22641418728107e-05, + "loss": 0.0845, + "step": 31437 + }, + { + "epoch": 0.7369562287194453, + "grad_norm": 0.570938229560852, + "learning_rate": 3.225872312039673e-05, + "loss": 0.0536, + "step": 31438 + }, + { + "epoch": 0.7369796702942504, + "grad_norm": 0.29462969303131104, + "learning_rate": 3.225330473554882e-05, + "loss": 0.0424, + "step": 31439 + }, + { + "epoch": 0.7370031118690553, + "grad_norm": 0.5625381469726562, + "learning_rate": 3.2247886718296396e-05, + "loss": 0.5353, + "step": 31440 + }, + { + "epoch": 0.7370265534438604, + "grad_norm": 0.668481707572937, + "learning_rate": 3.2242469068668836e-05, + "loss": 0.6044, + "step": 31441 + }, + { + "epoch": 0.7370499950186653, + "grad_norm": 0.46768200397491455, + "learning_rate": 3.223705178669551e-05, + "loss": 0.1242, + "step": 31442 + }, + { + "epoch": 0.7370734365934704, + "grad_norm": 0.34067970514297485, + "learning_rate": 3.223163487240588e-05, + "loss": 0.0812, + "step": 31443 + }, + { + "epoch": 0.7370968781682753, + "grad_norm": 0.34162020683288574, + "learning_rate": 3.222621832582927e-05, + "loss": 0.0628, + "step": 31444 + }, + { + "epoch": 0.7371203197430803, + "grad_norm": 0.2786158621311188, + "learning_rate": 3.222080214699514e-05, + "loss": 0.0613, + "step": 31445 + }, + { + "epoch": 0.7371437613178853, + "grad_norm": 0.34471428394317627, + "learning_rate": 3.221538633593285e-05, + "loss": 0.0666, + "step": 31446 + }, + { + "epoch": 0.7371672028926903, + "grad_norm": 0.21951454877853394, + "learning_rate": 3.220997089267178e-05, + "loss": 0.0213, + "step": 31447 + }, + { + "epoch": 0.7371906444674954, + "grad_norm": 0.5863929390907288, + "learning_rate": 3.22045558172413e-05, + "loss": 0.0705, + "step": 31448 + }, + { + "epoch": 0.7372140860423003, + "grad_norm": 0.18768927454948425, + "learning_rate": 3.2199141109670784e-05, + "loss": 0.0173, + "step": 31449 + }, + { + "epoch": 0.7372375276171054, + "grad_norm": 0.3370802104473114, + "learning_rate": 3.219372676998967e-05, + "loss": 0.0505, + "step": 31450 + }, + { + "epoch": 0.7372609691919103, + "grad_norm": 0.1500324159860611, + "learning_rate": 3.218831279822729e-05, + "loss": 0.0109, + "step": 31451 + }, + { + "epoch": 0.7372844107667154, + "grad_norm": 0.23025834560394287, + "learning_rate": 3.218289919441304e-05, + "loss": 0.1095, + "step": 31452 + }, + { + "epoch": 0.7373078523415203, + "grad_norm": 0.4304470717906952, + "learning_rate": 3.217748595857629e-05, + "loss": 0.0635, + "step": 31453 + }, + { + "epoch": 0.7373312939163253, + "grad_norm": 0.31831517815589905, + "learning_rate": 3.21720730907464e-05, + "loss": 0.0389, + "step": 31454 + }, + { + "epoch": 0.7373547354911303, + "grad_norm": 0.683035671710968, + "learning_rate": 3.2166660590952724e-05, + "loss": 0.1017, + "step": 31455 + }, + { + "epoch": 0.7373781770659353, + "grad_norm": 0.21405957639217377, + "learning_rate": 3.216124845922465e-05, + "loss": 0.0254, + "step": 31456 + }, + { + "epoch": 0.7374016186407403, + "grad_norm": 0.5394545197486877, + "learning_rate": 3.215583669559158e-05, + "loss": 0.3951, + "step": 31457 + }, + { + "epoch": 0.7374250602155453, + "grad_norm": 0.4192970395088196, + "learning_rate": 3.215042530008284e-05, + "loss": 0.1043, + "step": 31458 + }, + { + "epoch": 0.7374485017903503, + "grad_norm": 0.1550295650959015, + "learning_rate": 3.2145014272727814e-05, + "loss": 0.0289, + "step": 31459 + }, + { + "epoch": 0.7374719433651553, + "grad_norm": 0.3451479375362396, + "learning_rate": 3.2139603613555845e-05, + "loss": 0.0673, + "step": 31460 + }, + { + "epoch": 0.7374953849399603, + "grad_norm": 0.29475656151771545, + "learning_rate": 3.2134193322596295e-05, + "loss": 0.0447, + "step": 31461 + }, + { + "epoch": 0.7375188265147653, + "grad_norm": 0.15460456907749176, + "learning_rate": 3.2128783399878484e-05, + "loss": 0.018, + "step": 31462 + }, + { + "epoch": 0.7375422680895702, + "grad_norm": 0.1496286392211914, + "learning_rate": 3.2123373845431836e-05, + "loss": 0.032, + "step": 31463 + }, + { + "epoch": 0.7375657096643753, + "grad_norm": 0.45244747400283813, + "learning_rate": 3.211796465928567e-05, + "loss": 0.0939, + "step": 31464 + }, + { + "epoch": 0.7375891512391802, + "grad_norm": 0.45380985736846924, + "learning_rate": 3.2112555841469335e-05, + "loss": 0.0487, + "step": 31465 + }, + { + "epoch": 0.7376125928139853, + "grad_norm": 0.3614214062690735, + "learning_rate": 3.210714739201218e-05, + "loss": 0.5337, + "step": 31466 + }, + { + "epoch": 0.7376360343887902, + "grad_norm": 0.3672406077384949, + "learning_rate": 3.2101739310943536e-05, + "loss": 0.0406, + "step": 31467 + }, + { + "epoch": 0.7376594759635953, + "grad_norm": 0.29971548914909363, + "learning_rate": 3.209633159829278e-05, + "loss": 0.0311, + "step": 31468 + }, + { + "epoch": 0.7376829175384002, + "grad_norm": 0.09602921456098557, + "learning_rate": 3.20909242540892e-05, + "loss": 0.0228, + "step": 31469 + }, + { + "epoch": 0.7377063591132053, + "grad_norm": 0.3025699257850647, + "learning_rate": 3.208551727836222e-05, + "loss": 0.0785, + "step": 31470 + }, + { + "epoch": 0.7377298006880102, + "grad_norm": 0.1206260621547699, + "learning_rate": 3.208011067114113e-05, + "loss": 0.0174, + "step": 31471 + }, + { + "epoch": 0.7377532422628152, + "grad_norm": 0.15226857364177704, + "learning_rate": 3.207470443245527e-05, + "loss": 0.0321, + "step": 31472 + }, + { + "epoch": 0.7377766838376202, + "grad_norm": 0.40362632274627686, + "learning_rate": 3.206929856233397e-05, + "loss": 0.0683, + "step": 31473 + }, + { + "epoch": 0.7378001254124252, + "grad_norm": 0.15244561433792114, + "learning_rate": 3.206389306080654e-05, + "loss": 0.0309, + "step": 31474 + }, + { + "epoch": 0.7378235669872302, + "grad_norm": 0.07603360712528229, + "learning_rate": 3.2058487927902357e-05, + "loss": 0.014, + "step": 31475 + }, + { + "epoch": 0.7378470085620352, + "grad_norm": 0.2950538694858551, + "learning_rate": 3.205308316365074e-05, + "loss": 0.0326, + "step": 31476 + }, + { + "epoch": 0.7378704501368402, + "grad_norm": 0.5017889142036438, + "learning_rate": 3.2047678768081e-05, + "loss": 0.1203, + "step": 31477 + }, + { + "epoch": 0.7378938917116452, + "grad_norm": 0.5525312423706055, + "learning_rate": 3.2042274741222464e-05, + "loss": 0.0991, + "step": 31478 + }, + { + "epoch": 0.7379173332864501, + "grad_norm": 0.2767516076564789, + "learning_rate": 3.203687108310446e-05, + "loss": 0.022, + "step": 31479 + }, + { + "epoch": 0.7379407748612552, + "grad_norm": 0.6724290251731873, + "learning_rate": 3.2031467793756265e-05, + "loss": 0.5009, + "step": 31480 + }, + { + "epoch": 0.7379642164360601, + "grad_norm": 0.368570476770401, + "learning_rate": 3.202606487320724e-05, + "loss": 0.0409, + "step": 31481 + }, + { + "epoch": 0.7379876580108652, + "grad_norm": 0.21163693070411682, + "learning_rate": 3.202066232148674e-05, + "loss": 0.049, + "step": 31482 + }, + { + "epoch": 0.7380110995856701, + "grad_norm": 0.1397382915019989, + "learning_rate": 3.2015260138624025e-05, + "loss": 0.0235, + "step": 31483 + }, + { + "epoch": 0.7380345411604752, + "grad_norm": 0.4030397832393646, + "learning_rate": 3.200985832464842e-05, + "loss": 0.0362, + "step": 31484 + }, + { + "epoch": 0.7380579827352801, + "grad_norm": 0.38708606362342834, + "learning_rate": 3.200445687958925e-05, + "loss": 0.0868, + "step": 31485 + }, + { + "epoch": 0.7380814243100852, + "grad_norm": 0.44898512959480286, + "learning_rate": 3.1999055803475796e-05, + "loss": 0.0813, + "step": 31486 + }, + { + "epoch": 0.7381048658848901, + "grad_norm": 0.22460541129112244, + "learning_rate": 3.199365509633735e-05, + "loss": 0.0485, + "step": 31487 + }, + { + "epoch": 0.7381283074596952, + "grad_norm": 0.5290973782539368, + "learning_rate": 3.1988254758203275e-05, + "loss": 0.4346, + "step": 31488 + }, + { + "epoch": 0.7381517490345001, + "grad_norm": 0.2302536815404892, + "learning_rate": 3.198285478910284e-05, + "loss": 0.0141, + "step": 31489 + }, + { + "epoch": 0.7381751906093051, + "grad_norm": 0.49903562664985657, + "learning_rate": 3.197745518906535e-05, + "loss": 0.0395, + "step": 31490 + }, + { + "epoch": 0.7381986321841101, + "grad_norm": 0.47948020696640015, + "learning_rate": 3.1972055958120094e-05, + "loss": 0.0862, + "step": 31491 + }, + { + "epoch": 0.7382220737589151, + "grad_norm": 0.29049232602119446, + "learning_rate": 3.196665709629638e-05, + "loss": 0.064, + "step": 31492 + }, + { + "epoch": 0.7382455153337201, + "grad_norm": 0.45810016989707947, + "learning_rate": 3.196125860362347e-05, + "loss": 0.1222, + "step": 31493 + }, + { + "epoch": 0.7382689569085251, + "grad_norm": 0.7314538359642029, + "learning_rate": 3.1955860480130686e-05, + "loss": 0.0937, + "step": 31494 + }, + { + "epoch": 0.7382923984833301, + "grad_norm": 0.21332256495952606, + "learning_rate": 3.195046272584734e-05, + "loss": 0.0318, + "step": 31495 + }, + { + "epoch": 0.7383158400581351, + "grad_norm": 0.6167590618133545, + "learning_rate": 3.1945065340802694e-05, + "loss": 0.1104, + "step": 31496 + }, + { + "epoch": 0.73833928163294, + "grad_norm": 0.15990427136421204, + "learning_rate": 3.193966832502604e-05, + "loss": 0.0352, + "step": 31497 + }, + { + "epoch": 0.7383627232077451, + "grad_norm": 0.31712910532951355, + "learning_rate": 3.193427167854667e-05, + "loss": 0.0473, + "step": 31498 + }, + { + "epoch": 0.73838616478255, + "grad_norm": 0.19987350702285767, + "learning_rate": 3.192887540139384e-05, + "loss": 0.0311, + "step": 31499 + }, + { + "epoch": 0.7384096063573551, + "grad_norm": 0.5766708850860596, + "learning_rate": 3.192347949359682e-05, + "loss": 0.7336, + "step": 31500 + }, + { + "epoch": 0.7384330479321601, + "grad_norm": 0.13142059743404388, + "learning_rate": 3.1918083955184955e-05, + "loss": 0.0266, + "step": 31501 + }, + { + "epoch": 0.7384564895069651, + "grad_norm": 0.5711579322814941, + "learning_rate": 3.191268878618747e-05, + "loss": 0.546, + "step": 31502 + }, + { + "epoch": 0.7384799310817701, + "grad_norm": 0.2976781129837036, + "learning_rate": 3.1907293986633646e-05, + "loss": 0.0519, + "step": 31503 + }, + { + "epoch": 0.7385033726565751, + "grad_norm": 0.3302682936191559, + "learning_rate": 3.1901899556552774e-05, + "loss": 0.0599, + "step": 31504 + }, + { + "epoch": 0.7385268142313801, + "grad_norm": 0.1970827430486679, + "learning_rate": 3.189650549597407e-05, + "loss": 0.03, + "step": 31505 + }, + { + "epoch": 0.738550255806185, + "grad_norm": 0.34979212284088135, + "learning_rate": 3.189111180492688e-05, + "loss": 0.0838, + "step": 31506 + }, + { + "epoch": 0.7385736973809901, + "grad_norm": 0.42373549938201904, + "learning_rate": 3.188571848344038e-05, + "loss": 0.0669, + "step": 31507 + }, + { + "epoch": 0.738597138955795, + "grad_norm": 0.5421416759490967, + "learning_rate": 3.188032553154394e-05, + "loss": 0.1133, + "step": 31508 + }, + { + "epoch": 0.7386205805306001, + "grad_norm": 0.49645769596099854, + "learning_rate": 3.1874932949266764e-05, + "loss": 0.0653, + "step": 31509 + }, + { + "epoch": 0.738644022105405, + "grad_norm": 0.21957054734230042, + "learning_rate": 3.186954073663811e-05, + "loss": 0.0475, + "step": 31510 + }, + { + "epoch": 0.7386674636802101, + "grad_norm": 0.2374170869588852, + "learning_rate": 3.1864148893687254e-05, + "loss": 0.0587, + "step": 31511 + }, + { + "epoch": 0.738690905255015, + "grad_norm": 0.5661266446113586, + "learning_rate": 3.185875742044341e-05, + "loss": 0.1218, + "step": 31512 + }, + { + "epoch": 0.7387143468298201, + "grad_norm": 0.31578725576400757, + "learning_rate": 3.1853366316935886e-05, + "loss": 0.075, + "step": 31513 + }, + { + "epoch": 0.738737788404625, + "grad_norm": 0.17707152664661407, + "learning_rate": 3.184797558319392e-05, + "loss": 0.0303, + "step": 31514 + }, + { + "epoch": 0.7387612299794301, + "grad_norm": 0.16620922088623047, + "learning_rate": 3.184258521924676e-05, + "loss": 0.0287, + "step": 31515 + }, + { + "epoch": 0.738784671554235, + "grad_norm": 0.5953701138496399, + "learning_rate": 3.183719522512364e-05, + "loss": 0.0658, + "step": 31516 + }, + { + "epoch": 0.73880811312904, + "grad_norm": 0.5871471166610718, + "learning_rate": 3.183180560085378e-05, + "loss": 0.1016, + "step": 31517 + }, + { + "epoch": 0.738831554703845, + "grad_norm": 0.6457770466804504, + "learning_rate": 3.1826416346466495e-05, + "loss": 0.1253, + "step": 31518 + }, + { + "epoch": 0.73885499627865, + "grad_norm": 0.08052252233028412, + "learning_rate": 3.182102746199096e-05, + "loss": 0.0115, + "step": 31519 + }, + { + "epoch": 0.738878437853455, + "grad_norm": 0.7060831785202026, + "learning_rate": 3.181563894745648e-05, + "loss": 0.0733, + "step": 31520 + }, + { + "epoch": 0.73890187942826, + "grad_norm": 0.413522869348526, + "learning_rate": 3.1810250802892264e-05, + "loss": 0.0664, + "step": 31521 + }, + { + "epoch": 0.738925321003065, + "grad_norm": 0.42833077907562256, + "learning_rate": 3.1804863028327536e-05, + "loss": 0.068, + "step": 31522 + }, + { + "epoch": 0.73894876257787, + "grad_norm": 0.14042900502681732, + "learning_rate": 3.179947562379155e-05, + "loss": 0.0091, + "step": 31523 + }, + { + "epoch": 0.738972204152675, + "grad_norm": 0.27654168009757996, + "learning_rate": 3.179408858931352e-05, + "loss": 0.0556, + "step": 31524 + }, + { + "epoch": 0.73899564572748, + "grad_norm": 0.3704703450202942, + "learning_rate": 3.178870192492265e-05, + "loss": 0.0728, + "step": 31525 + }, + { + "epoch": 0.7390190873022849, + "grad_norm": 0.4014981985092163, + "learning_rate": 3.1783315630648236e-05, + "loss": 0.0866, + "step": 31526 + }, + { + "epoch": 0.73904252887709, + "grad_norm": 0.1497081071138382, + "learning_rate": 3.177792970651947e-05, + "loss": 0.0269, + "step": 31527 + }, + { + "epoch": 0.7390659704518949, + "grad_norm": 0.4072642922401428, + "learning_rate": 3.177254415256558e-05, + "loss": 0.0491, + "step": 31528 + }, + { + "epoch": 0.7390894120267, + "grad_norm": 0.3290241062641144, + "learning_rate": 3.1767158968815734e-05, + "loss": 0.0309, + "step": 31529 + }, + { + "epoch": 0.7391128536015049, + "grad_norm": 0.40019071102142334, + "learning_rate": 3.176177415529925e-05, + "loss": 0.046, + "step": 31530 + }, + { + "epoch": 0.73913629517631, + "grad_norm": 0.28804072737693787, + "learning_rate": 3.17563897120453e-05, + "loss": 0.2709, + "step": 31531 + }, + { + "epoch": 0.7391597367511149, + "grad_norm": 0.22082191705703735, + "learning_rate": 3.1751005639083054e-05, + "loss": 0.0332, + "step": 31532 + }, + { + "epoch": 0.73918317832592, + "grad_norm": 0.6320930123329163, + "learning_rate": 3.17456219364418e-05, + "loss": 0.7719, + "step": 31533 + }, + { + "epoch": 0.7392066199007249, + "grad_norm": 0.31686604022979736, + "learning_rate": 3.174023860415074e-05, + "loss": 0.0325, + "step": 31534 + }, + { + "epoch": 0.7392300614755299, + "grad_norm": 0.3863254487514496, + "learning_rate": 3.173485564223906e-05, + "loss": 0.1181, + "step": 31535 + }, + { + "epoch": 0.7392535030503349, + "grad_norm": 0.4805198311805725, + "learning_rate": 3.172947305073596e-05, + "loss": 0.6442, + "step": 31536 + }, + { + "epoch": 0.7392769446251399, + "grad_norm": 0.8449408411979675, + "learning_rate": 3.172409082967066e-05, + "loss": 0.1312, + "step": 31537 + }, + { + "epoch": 0.7393003861999449, + "grad_norm": 0.31474196910858154, + "learning_rate": 3.171870897907234e-05, + "loss": 0.0512, + "step": 31538 + }, + { + "epoch": 0.7393238277747499, + "grad_norm": 0.8503243923187256, + "learning_rate": 3.171332749897026e-05, + "loss": 0.2005, + "step": 31539 + }, + { + "epoch": 0.7393472693495549, + "grad_norm": 0.3982442319393158, + "learning_rate": 3.170794638939357e-05, + "loss": 0.0886, + "step": 31540 + }, + { + "epoch": 0.7393707109243599, + "grad_norm": 1.510622262954712, + "learning_rate": 3.170256565037146e-05, + "loss": 0.105, + "step": 31541 + }, + { + "epoch": 0.7393941524991648, + "grad_norm": 0.702934980392456, + "learning_rate": 3.1697185281933186e-05, + "loss": 0.2195, + "step": 31542 + }, + { + "epoch": 0.7394175940739699, + "grad_norm": 0.13048343360424042, + "learning_rate": 3.16918052841079e-05, + "loss": 0.0138, + "step": 31543 + }, + { + "epoch": 0.7394410356487748, + "grad_norm": 0.3032730519771576, + "learning_rate": 3.16864256569248e-05, + "loss": 0.0291, + "step": 31544 + }, + { + "epoch": 0.7394644772235799, + "grad_norm": 0.46090003848075867, + "learning_rate": 3.1681046400413036e-05, + "loss": 0.0802, + "step": 31545 + }, + { + "epoch": 0.7394879187983848, + "grad_norm": 0.10883702337741852, + "learning_rate": 3.167566751460188e-05, + "loss": 0.0175, + "step": 31546 + }, + { + "epoch": 0.7395113603731899, + "grad_norm": 0.1580943614244461, + "learning_rate": 3.167028899952046e-05, + "loss": 0.0338, + "step": 31547 + }, + { + "epoch": 0.7395348019479948, + "grad_norm": 0.48586079478263855, + "learning_rate": 3.166491085519798e-05, + "loss": 0.0949, + "step": 31548 + }, + { + "epoch": 0.7395582435227999, + "grad_norm": 0.22089001536369324, + "learning_rate": 3.165953308166362e-05, + "loss": 0.0412, + "step": 31549 + }, + { + "epoch": 0.7395816850976048, + "grad_norm": 0.26527002453804016, + "learning_rate": 3.165415567894652e-05, + "loss": 0.0598, + "step": 31550 + }, + { + "epoch": 0.7396051266724099, + "grad_norm": 0.2554432451725006, + "learning_rate": 3.164877864707591e-05, + "loss": 0.0495, + "step": 31551 + }, + { + "epoch": 0.7396285682472149, + "grad_norm": 0.161605104804039, + "learning_rate": 3.1643401986080967e-05, + "loss": 0.0331, + "step": 31552 + }, + { + "epoch": 0.7396520098220198, + "grad_norm": 0.1797737032175064, + "learning_rate": 3.163802569599081e-05, + "loss": 0.0384, + "step": 31553 + }, + { + "epoch": 0.7396754513968249, + "grad_norm": 0.37356922030448914, + "learning_rate": 3.163264977683468e-05, + "loss": 0.0586, + "step": 31554 + }, + { + "epoch": 0.7396988929716298, + "grad_norm": 0.5972776412963867, + "learning_rate": 3.1627274228641714e-05, + "loss": 0.0815, + "step": 31555 + }, + { + "epoch": 0.7397223345464349, + "grad_norm": 0.51659095287323, + "learning_rate": 3.1621899051441084e-05, + "loss": 0.1035, + "step": 31556 + }, + { + "epoch": 0.7397457761212398, + "grad_norm": 0.32028329372406006, + "learning_rate": 3.161652424526191e-05, + "loss": 0.0465, + "step": 31557 + }, + { + "epoch": 0.7397692176960449, + "grad_norm": 0.34195390343666077, + "learning_rate": 3.161114981013345e-05, + "loss": 0.0662, + "step": 31558 + }, + { + "epoch": 0.7397926592708498, + "grad_norm": 0.4850555956363678, + "learning_rate": 3.16057757460848e-05, + "loss": 0.0826, + "step": 31559 + }, + { + "epoch": 0.7398161008456549, + "grad_norm": 0.0961974710226059, + "learning_rate": 3.1600402053145126e-05, + "loss": 0.0127, + "step": 31560 + }, + { + "epoch": 0.7398395424204598, + "grad_norm": 0.35371774435043335, + "learning_rate": 3.159502873134361e-05, + "loss": 0.3185, + "step": 31561 + }, + { + "epoch": 0.7398629839952648, + "grad_norm": 0.4905270040035248, + "learning_rate": 3.1589655780709384e-05, + "loss": 0.1512, + "step": 31562 + }, + { + "epoch": 0.7398864255700698, + "grad_norm": 0.4198099970817566, + "learning_rate": 3.158428320127157e-05, + "loss": 0.0547, + "step": 31563 + }, + { + "epoch": 0.7399098671448748, + "grad_norm": 0.3951341211795807, + "learning_rate": 3.1578910993059406e-05, + "loss": 0.0881, + "step": 31564 + }, + { + "epoch": 0.7399333087196798, + "grad_norm": 0.2740994393825531, + "learning_rate": 3.1573539156101984e-05, + "loss": 0.0566, + "step": 31565 + }, + { + "epoch": 0.7399567502944848, + "grad_norm": 0.5424454808235168, + "learning_rate": 3.156816769042843e-05, + "loss": 0.0893, + "step": 31566 + }, + { + "epoch": 0.7399801918692898, + "grad_norm": 0.20899556577205658, + "learning_rate": 3.1562796596067965e-05, + "loss": 0.0269, + "step": 31567 + }, + { + "epoch": 0.7400036334440948, + "grad_norm": 0.14559337496757507, + "learning_rate": 3.1557425873049685e-05, + "loss": 0.029, + "step": 31568 + }, + { + "epoch": 0.7400270750188997, + "grad_norm": 0.5129328966140747, + "learning_rate": 3.1552055521402734e-05, + "loss": 0.1423, + "step": 31569 + }, + { + "epoch": 0.7400505165937048, + "grad_norm": 0.5605493187904358, + "learning_rate": 3.154668554115622e-05, + "loss": 0.1061, + "step": 31570 + }, + { + "epoch": 0.7400739581685097, + "grad_norm": 0.6577597856521606, + "learning_rate": 3.1541315932339346e-05, + "loss": 0.1025, + "step": 31571 + }, + { + "epoch": 0.7400973997433148, + "grad_norm": 0.0933222770690918, + "learning_rate": 3.153594669498122e-05, + "loss": 0.0176, + "step": 31572 + }, + { + "epoch": 0.7401208413181197, + "grad_norm": 0.30750465393066406, + "learning_rate": 3.153057782911097e-05, + "loss": 0.0333, + "step": 31573 + }, + { + "epoch": 0.7401442828929248, + "grad_norm": 0.38344553112983704, + "learning_rate": 3.152520933475775e-05, + "loss": 0.1023, + "step": 31574 + }, + { + "epoch": 0.7401677244677297, + "grad_norm": 0.3163962960243225, + "learning_rate": 3.1519841211950644e-05, + "loss": 0.0492, + "step": 31575 + }, + { + "epoch": 0.7401911660425348, + "grad_norm": 0.3659043610095978, + "learning_rate": 3.1514473460718775e-05, + "loss": 0.0359, + "step": 31576 + }, + { + "epoch": 0.7402146076173397, + "grad_norm": 0.2703528106212616, + "learning_rate": 3.1509106081091346e-05, + "loss": 0.0523, + "step": 31577 + }, + { + "epoch": 0.7402380491921448, + "grad_norm": 0.9766110181808472, + "learning_rate": 3.150373907309738e-05, + "loss": 0.1386, + "step": 31578 + }, + { + "epoch": 0.7402614907669497, + "grad_norm": 0.12230727821588516, + "learning_rate": 3.14983724367661e-05, + "loss": 0.0217, + "step": 31579 + }, + { + "epoch": 0.7402849323417547, + "grad_norm": 0.4481053948402405, + "learning_rate": 3.149300617212657e-05, + "loss": 0.1022, + "step": 31580 + }, + { + "epoch": 0.7403083739165597, + "grad_norm": 0.5561386346817017, + "learning_rate": 3.148764027920792e-05, + "loss": 0.089, + "step": 31581 + }, + { + "epoch": 0.7403318154913647, + "grad_norm": 0.287210077047348, + "learning_rate": 3.1482274758039265e-05, + "loss": 0.0356, + "step": 31582 + }, + { + "epoch": 0.7403552570661697, + "grad_norm": 0.3494323790073395, + "learning_rate": 3.1476909608649675e-05, + "loss": 0.0519, + "step": 31583 + }, + { + "epoch": 0.7403786986409747, + "grad_norm": 0.4277951121330261, + "learning_rate": 3.147154483106833e-05, + "loss": 0.0573, + "step": 31584 + }, + { + "epoch": 0.7404021402157797, + "grad_norm": 0.16624104976654053, + "learning_rate": 3.146618042532431e-05, + "loss": 0.0313, + "step": 31585 + }, + { + "epoch": 0.7404255817905847, + "grad_norm": 0.49877816438674927, + "learning_rate": 3.146081639144672e-05, + "loss": 0.0777, + "step": 31586 + }, + { + "epoch": 0.7404490233653896, + "grad_norm": 0.5489854216575623, + "learning_rate": 3.145545272946467e-05, + "loss": 0.1116, + "step": 31587 + }, + { + "epoch": 0.7404724649401947, + "grad_norm": 0.4318607747554779, + "learning_rate": 3.145008943940724e-05, + "loss": 0.0866, + "step": 31588 + }, + { + "epoch": 0.7404959065149996, + "grad_norm": 0.21823827922344208, + "learning_rate": 3.1444726521303566e-05, + "loss": 0.0181, + "step": 31589 + }, + { + "epoch": 0.7405193480898047, + "grad_norm": 0.32853055000305176, + "learning_rate": 3.143936397518271e-05, + "loss": 0.0889, + "step": 31590 + }, + { + "epoch": 0.7405427896646096, + "grad_norm": 0.4500082731246948, + "learning_rate": 3.1434001801073834e-05, + "loss": 0.0698, + "step": 31591 + }, + { + "epoch": 0.7405662312394147, + "grad_norm": 0.7387334704399109, + "learning_rate": 3.1428639999005985e-05, + "loss": 0.0418, + "step": 31592 + }, + { + "epoch": 0.7405896728142196, + "grad_norm": 0.4913634657859802, + "learning_rate": 3.1423278569008254e-05, + "loss": 0.1217, + "step": 31593 + }, + { + "epoch": 0.7406131143890247, + "grad_norm": 0.14175944030284882, + "learning_rate": 3.141791751110975e-05, + "loss": 0.019, + "step": 31594 + }, + { + "epoch": 0.7406365559638296, + "grad_norm": 0.5436981320381165, + "learning_rate": 3.141255682533952e-05, + "loss": 0.1323, + "step": 31595 + }, + { + "epoch": 0.7406599975386347, + "grad_norm": 0.4490354657173157, + "learning_rate": 3.140719651172671e-05, + "loss": 0.0606, + "step": 31596 + }, + { + "epoch": 0.7406834391134396, + "grad_norm": 0.23193691670894623, + "learning_rate": 3.140183657030039e-05, + "loss": 0.0438, + "step": 31597 + }, + { + "epoch": 0.7407068806882446, + "grad_norm": 0.36366015672683716, + "learning_rate": 3.139647700108963e-05, + "loss": 0.0594, + "step": 31598 + }, + { + "epoch": 0.7407303222630496, + "grad_norm": 0.44158774614334106, + "learning_rate": 3.1391117804123523e-05, + "loss": 0.0639, + "step": 31599 + }, + { + "epoch": 0.7407537638378546, + "grad_norm": 0.6593990921974182, + "learning_rate": 3.1385758979431134e-05, + "loss": 0.4707, + "step": 31600 + }, + { + "epoch": 0.7407772054126596, + "grad_norm": 0.6262936592102051, + "learning_rate": 3.1380400527041506e-05, + "loss": 0.124, + "step": 31601 + }, + { + "epoch": 0.7408006469874646, + "grad_norm": 0.3285525143146515, + "learning_rate": 3.137504244698377e-05, + "loss": 0.0634, + "step": 31602 + }, + { + "epoch": 0.7408240885622697, + "grad_norm": 0.5400421023368835, + "learning_rate": 3.136968473928701e-05, + "loss": 0.0965, + "step": 31603 + }, + { + "epoch": 0.7408475301370746, + "grad_norm": 0.17947521805763245, + "learning_rate": 3.136432740398026e-05, + "loss": 0.0111, + "step": 31604 + }, + { + "epoch": 0.7408709717118797, + "grad_norm": 0.9830317497253418, + "learning_rate": 3.1358970441092604e-05, + "loss": 0.1368, + "step": 31605 + }, + { + "epoch": 0.7408944132866846, + "grad_norm": 0.3867342472076416, + "learning_rate": 3.1353613850653105e-05, + "loss": 0.069, + "step": 31606 + }, + { + "epoch": 0.7409178548614896, + "grad_norm": 0.42602303624153137, + "learning_rate": 3.1348257632690835e-05, + "loss": 0.1037, + "step": 31607 + }, + { + "epoch": 0.7409412964362946, + "grad_norm": 0.30505868792533875, + "learning_rate": 3.134290178723481e-05, + "loss": 0.0532, + "step": 31608 + }, + { + "epoch": 0.7409647380110996, + "grad_norm": 0.46261852979660034, + "learning_rate": 3.133754631431417e-05, + "loss": 0.0947, + "step": 31609 + }, + { + "epoch": 0.7409881795859046, + "grad_norm": 0.35404667258262634, + "learning_rate": 3.133219121395793e-05, + "loss": 0.0635, + "step": 31610 + }, + { + "epoch": 0.7410116211607096, + "grad_norm": 0.3186536729335785, + "learning_rate": 3.132683648619514e-05, + "loss": 0.0258, + "step": 31611 + }, + { + "epoch": 0.7410350627355146, + "grad_norm": 0.6797031164169312, + "learning_rate": 3.132148213105488e-05, + "loss": 0.1394, + "step": 31612 + }, + { + "epoch": 0.7410585043103196, + "grad_norm": 0.19859153032302856, + "learning_rate": 3.1316128148566184e-05, + "loss": 0.0205, + "step": 31613 + }, + { + "epoch": 0.7410819458851245, + "grad_norm": 0.3944317102432251, + "learning_rate": 3.131077453875808e-05, + "loss": 0.0633, + "step": 31614 + }, + { + "epoch": 0.7411053874599296, + "grad_norm": 0.4167160093784332, + "learning_rate": 3.130542130165964e-05, + "loss": 0.1138, + "step": 31615 + }, + { + "epoch": 0.7411288290347345, + "grad_norm": 0.14262613654136658, + "learning_rate": 3.1300068437299944e-05, + "loss": 0.0225, + "step": 31616 + }, + { + "epoch": 0.7411522706095396, + "grad_norm": 0.25955823063850403, + "learning_rate": 3.129471594570802e-05, + "loss": 0.1971, + "step": 31617 + }, + { + "epoch": 0.7411757121843445, + "grad_norm": 0.21270671486854553, + "learning_rate": 3.128936382691289e-05, + "loss": 0.0458, + "step": 31618 + }, + { + "epoch": 0.7411991537591496, + "grad_norm": 0.6518259048461914, + "learning_rate": 3.12840120809436e-05, + "loss": 0.553, + "step": 31619 + }, + { + "epoch": 0.7412225953339545, + "grad_norm": 0.0871332511305809, + "learning_rate": 3.12786607078292e-05, + "loss": 0.0126, + "step": 31620 + }, + { + "epoch": 0.7412460369087596, + "grad_norm": 0.36468395590782166, + "learning_rate": 3.127330970759869e-05, + "loss": 0.0788, + "step": 31621 + }, + { + "epoch": 0.7412694784835645, + "grad_norm": 0.09993957728147507, + "learning_rate": 3.126795908028116e-05, + "loss": 0.0175, + "step": 31622 + }, + { + "epoch": 0.7412929200583696, + "grad_norm": 0.5175031423568726, + "learning_rate": 3.1262608825905624e-05, + "loss": 0.0727, + "step": 31623 + }, + { + "epoch": 0.7413163616331745, + "grad_norm": 0.4268779754638672, + "learning_rate": 3.12572589445011e-05, + "loss": 0.0351, + "step": 31624 + }, + { + "epoch": 0.7413398032079795, + "grad_norm": 0.38241302967071533, + "learning_rate": 3.1251909436096616e-05, + "loss": 0.0733, + "step": 31625 + }, + { + "epoch": 0.7413632447827845, + "grad_norm": 0.4346758723258972, + "learning_rate": 3.124656030072118e-05, + "loss": 0.0542, + "step": 31626 + }, + { + "epoch": 0.7413866863575895, + "grad_norm": 0.23550574481487274, + "learning_rate": 3.124121153840387e-05, + "loss": 0.0203, + "step": 31627 + }, + { + "epoch": 0.7414101279323945, + "grad_norm": 0.40944501757621765, + "learning_rate": 3.123586314917365e-05, + "loss": 0.5803, + "step": 31628 + }, + { + "epoch": 0.7414335695071995, + "grad_norm": 0.40622958540916443, + "learning_rate": 3.12305151330596e-05, + "loss": 0.0491, + "step": 31629 + }, + { + "epoch": 0.7414570110820045, + "grad_norm": 0.258924275636673, + "learning_rate": 3.122516749009071e-05, + "loss": 0.0245, + "step": 31630 + }, + { + "epoch": 0.7414804526568095, + "grad_norm": 0.14943629503250122, + "learning_rate": 3.1219820220295994e-05, + "loss": 0.0145, + "step": 31631 + }, + { + "epoch": 0.7415038942316144, + "grad_norm": 0.35146573185920715, + "learning_rate": 3.121447332370447e-05, + "loss": 0.064, + "step": 31632 + }, + { + "epoch": 0.7415273358064195, + "grad_norm": 0.1959364414215088, + "learning_rate": 3.120912680034511e-05, + "loss": 0.0137, + "step": 31633 + }, + { + "epoch": 0.7415507773812244, + "grad_norm": 0.45841678977012634, + "learning_rate": 3.120378065024699e-05, + "loss": 0.1037, + "step": 31634 + }, + { + "epoch": 0.7415742189560295, + "grad_norm": 0.23571567237377167, + "learning_rate": 3.119843487343911e-05, + "loss": 0.058, + "step": 31635 + }, + { + "epoch": 0.7415976605308344, + "grad_norm": 0.16964730620384216, + "learning_rate": 3.119308946995044e-05, + "loss": 0.0357, + "step": 31636 + }, + { + "epoch": 0.7416211021056395, + "grad_norm": 0.4367101490497589, + "learning_rate": 3.1187744439810006e-05, + "loss": 0.0563, + "step": 31637 + }, + { + "epoch": 0.7416445436804444, + "grad_norm": 0.23772305250167847, + "learning_rate": 3.118239978304681e-05, + "loss": 0.0289, + "step": 31638 + }, + { + "epoch": 0.7416679852552495, + "grad_norm": 0.5471802949905396, + "learning_rate": 3.117705549968981e-05, + "loss": 0.1176, + "step": 31639 + }, + { + "epoch": 0.7416914268300544, + "grad_norm": 0.5271496772766113, + "learning_rate": 3.117171158976805e-05, + "loss": 0.6253, + "step": 31640 + }, + { + "epoch": 0.7417148684048595, + "grad_norm": 0.18783541023731232, + "learning_rate": 3.116636805331053e-05, + "loss": 0.0241, + "step": 31641 + }, + { + "epoch": 0.7417383099796644, + "grad_norm": 0.28301817178726196, + "learning_rate": 3.1161024890346256e-05, + "loss": 0.044, + "step": 31642 + }, + { + "epoch": 0.7417617515544694, + "grad_norm": 0.34524911642074585, + "learning_rate": 3.1155682100904184e-05, + "loss": 0.0387, + "step": 31643 + }, + { + "epoch": 0.7417851931292744, + "grad_norm": 0.23790322244167328, + "learning_rate": 3.115033968501332e-05, + "loss": 0.0327, + "step": 31644 + }, + { + "epoch": 0.7418086347040794, + "grad_norm": 0.24464009702205658, + "learning_rate": 3.1144997642702647e-05, + "loss": 0.0402, + "step": 31645 + }, + { + "epoch": 0.7418320762788844, + "grad_norm": 0.3078160583972931, + "learning_rate": 3.1139655974001126e-05, + "loss": 0.043, + "step": 31646 + }, + { + "epoch": 0.7418555178536894, + "grad_norm": 0.47190535068511963, + "learning_rate": 3.113431467893779e-05, + "loss": 0.0686, + "step": 31647 + }, + { + "epoch": 0.7418789594284944, + "grad_norm": 0.18498101830482483, + "learning_rate": 3.112897375754161e-05, + "loss": 0.031, + "step": 31648 + }, + { + "epoch": 0.7419024010032994, + "grad_norm": 0.5080516934394836, + "learning_rate": 3.112363320984155e-05, + "loss": 0.1186, + "step": 31649 + }, + { + "epoch": 0.7419258425781043, + "grad_norm": 0.3188779354095459, + "learning_rate": 3.1118293035866584e-05, + "loss": 0.0567, + "step": 31650 + }, + { + "epoch": 0.7419492841529094, + "grad_norm": 0.5055118203163147, + "learning_rate": 3.111295323564568e-05, + "loss": 0.0955, + "step": 31651 + }, + { + "epoch": 0.7419727257277143, + "grad_norm": 0.32143694162368774, + "learning_rate": 3.1107613809207846e-05, + "loss": 0.0412, + "step": 31652 + }, + { + "epoch": 0.7419961673025194, + "grad_norm": 0.3898167610168457, + "learning_rate": 3.1102274756582017e-05, + "loss": 0.0886, + "step": 31653 + }, + { + "epoch": 0.7420196088773244, + "grad_norm": 0.08010522276163101, + "learning_rate": 3.1096936077797214e-05, + "loss": 0.012, + "step": 31654 + }, + { + "epoch": 0.7420430504521294, + "grad_norm": 0.5057243704795837, + "learning_rate": 3.1091597772882366e-05, + "loss": 0.081, + "step": 31655 + }, + { + "epoch": 0.7420664920269344, + "grad_norm": 0.7703374624252319, + "learning_rate": 3.108625984186645e-05, + "loss": 0.0954, + "step": 31656 + }, + { + "epoch": 0.7420899336017394, + "grad_norm": 0.6245713829994202, + "learning_rate": 3.108092228477842e-05, + "loss": 0.1275, + "step": 31657 + }, + { + "epoch": 0.7421133751765444, + "grad_norm": 0.23004165291786194, + "learning_rate": 3.107558510164725e-05, + "loss": 0.0267, + "step": 31658 + }, + { + "epoch": 0.7421368167513493, + "grad_norm": 0.4820396602153778, + "learning_rate": 3.107024829250186e-05, + "loss": 0.1134, + "step": 31659 + }, + { + "epoch": 0.7421602583261544, + "grad_norm": 0.24840673804283142, + "learning_rate": 3.1064911857371274e-05, + "loss": 0.0473, + "step": 31660 + }, + { + "epoch": 0.7421836999009593, + "grad_norm": 0.5039926767349243, + "learning_rate": 3.105957579628441e-05, + "loss": 0.0994, + "step": 31661 + }, + { + "epoch": 0.7422071414757644, + "grad_norm": 0.17794449627399445, + "learning_rate": 3.1054240109270226e-05, + "loss": 0.025, + "step": 31662 + }, + { + "epoch": 0.7422305830505693, + "grad_norm": 0.3235822916030884, + "learning_rate": 3.104890479635765e-05, + "loss": 0.032, + "step": 31663 + }, + { + "epoch": 0.7422540246253744, + "grad_norm": 0.3508094251155853, + "learning_rate": 3.1043569857575686e-05, + "loss": 0.0521, + "step": 31664 + }, + { + "epoch": 0.7422774662001793, + "grad_norm": 0.06425318121910095, + "learning_rate": 3.103823529295321e-05, + "loss": 0.0088, + "step": 31665 + }, + { + "epoch": 0.7423009077749844, + "grad_norm": 0.27767980098724365, + "learning_rate": 3.103290110251925e-05, + "loss": 0.0382, + "step": 31666 + }, + { + "epoch": 0.7423243493497893, + "grad_norm": 0.4996749460697174, + "learning_rate": 3.1027567286302703e-05, + "loss": 0.0595, + "step": 31667 + }, + { + "epoch": 0.7423477909245944, + "grad_norm": 0.21494410932064056, + "learning_rate": 3.102223384433252e-05, + "loss": 0.0196, + "step": 31668 + }, + { + "epoch": 0.7423712324993993, + "grad_norm": 0.27496767044067383, + "learning_rate": 3.101690077663762e-05, + "loss": 0.0144, + "step": 31669 + }, + { + "epoch": 0.7423946740742043, + "grad_norm": 0.38837751746177673, + "learning_rate": 3.1011568083246966e-05, + "loss": 0.0822, + "step": 31670 + }, + { + "epoch": 0.7424181156490093, + "grad_norm": 0.40474584698677063, + "learning_rate": 3.100623576418946e-05, + "loss": 0.0528, + "step": 31671 + }, + { + "epoch": 0.7424415572238143, + "grad_norm": 0.31229203939437866, + "learning_rate": 3.1000903819494075e-05, + "loss": 0.0682, + "step": 31672 + }, + { + "epoch": 0.7424649987986193, + "grad_norm": 0.6563658118247986, + "learning_rate": 3.099557224918974e-05, + "loss": 0.0582, + "step": 31673 + }, + { + "epoch": 0.7424884403734243, + "grad_norm": 0.1075059100985527, + "learning_rate": 3.099024105330537e-05, + "loss": 0.0174, + "step": 31674 + }, + { + "epoch": 0.7425118819482293, + "grad_norm": 0.530161440372467, + "learning_rate": 3.0984910231869844e-05, + "loss": 0.0832, + "step": 31675 + }, + { + "epoch": 0.7425353235230343, + "grad_norm": 0.6652054786682129, + "learning_rate": 3.097957978491218e-05, + "loss": 0.1393, + "step": 31676 + }, + { + "epoch": 0.7425587650978392, + "grad_norm": 0.40751534700393677, + "learning_rate": 3.097424971246126e-05, + "loss": 0.0415, + "step": 31677 + }, + { + "epoch": 0.7425822066726443, + "grad_norm": 0.14053277671337128, + "learning_rate": 3.0968920014545964e-05, + "loss": 0.0225, + "step": 31678 + }, + { + "epoch": 0.7426056482474492, + "grad_norm": 0.47676199674606323, + "learning_rate": 3.096359069119528e-05, + "loss": 0.0577, + "step": 31679 + }, + { + "epoch": 0.7426290898222543, + "grad_norm": 0.11392191797494888, + "learning_rate": 3.09582617424381e-05, + "loss": 0.0126, + "step": 31680 + }, + { + "epoch": 0.7426525313970592, + "grad_norm": 0.2829308807849884, + "learning_rate": 3.095293316830332e-05, + "loss": 0.0405, + "step": 31681 + }, + { + "epoch": 0.7426759729718643, + "grad_norm": 0.11954543739557266, + "learning_rate": 3.094760496881988e-05, + "loss": 0.0222, + "step": 31682 + }, + { + "epoch": 0.7426994145466692, + "grad_norm": 0.06083543226122856, + "learning_rate": 3.0942277144016676e-05, + "loss": 0.0086, + "step": 31683 + }, + { + "epoch": 0.7427228561214743, + "grad_norm": 0.4120975434780121, + "learning_rate": 3.093694969392258e-05, + "loss": 0.0593, + "step": 31684 + }, + { + "epoch": 0.7427462976962792, + "grad_norm": 0.49207258224487305, + "learning_rate": 3.0931622618566566e-05, + "loss": 0.0849, + "step": 31685 + }, + { + "epoch": 0.7427697392710843, + "grad_norm": 0.06801309436559677, + "learning_rate": 3.092629591797752e-05, + "loss": 0.012, + "step": 31686 + }, + { + "epoch": 0.7427931808458892, + "grad_norm": 0.24560698866844177, + "learning_rate": 3.09209695921843e-05, + "loss": 0.0231, + "step": 31687 + }, + { + "epoch": 0.7428166224206942, + "grad_norm": 0.2395155131816864, + "learning_rate": 3.0915643641215864e-05, + "loss": 0.0272, + "step": 31688 + }, + { + "epoch": 0.7428400639954992, + "grad_norm": 0.27009356021881104, + "learning_rate": 3.091031806510109e-05, + "loss": 0.0347, + "step": 31689 + }, + { + "epoch": 0.7428635055703042, + "grad_norm": 0.40836432576179504, + "learning_rate": 3.090499286386889e-05, + "loss": 0.05, + "step": 31690 + }, + { + "epoch": 0.7428869471451092, + "grad_norm": 0.5821489691734314, + "learning_rate": 3.0899668037548094e-05, + "loss": 0.08, + "step": 31691 + }, + { + "epoch": 0.7429103887199142, + "grad_norm": 0.8789841532707214, + "learning_rate": 3.0894343586167675e-05, + "loss": 0.1468, + "step": 31692 + }, + { + "epoch": 0.7429338302947192, + "grad_norm": 0.4334692358970642, + "learning_rate": 3.088901950975649e-05, + "loss": 0.0835, + "step": 31693 + }, + { + "epoch": 0.7429572718695242, + "grad_norm": 0.44386544823646545, + "learning_rate": 3.088369580834344e-05, + "loss": 0.0493, + "step": 31694 + }, + { + "epoch": 0.7429807134443291, + "grad_norm": 0.40917548537254333, + "learning_rate": 3.0878372481957386e-05, + "loss": 0.0304, + "step": 31695 + }, + { + "epoch": 0.7430041550191342, + "grad_norm": 0.48676374554634094, + "learning_rate": 3.087304953062723e-05, + "loss": 0.4331, + "step": 31696 + }, + { + "epoch": 0.7430275965939391, + "grad_norm": 0.4197919964790344, + "learning_rate": 3.0867726954381824e-05, + "loss": 0.0705, + "step": 31697 + }, + { + "epoch": 0.7430510381687442, + "grad_norm": 0.6256449222564697, + "learning_rate": 3.08624047532501e-05, + "loss": 0.065, + "step": 31698 + }, + { + "epoch": 0.7430744797435491, + "grad_norm": 0.32289019227027893, + "learning_rate": 3.085708292726088e-05, + "loss": 0.0525, + "step": 31699 + }, + { + "epoch": 0.7430979213183542, + "grad_norm": 0.6150311827659607, + "learning_rate": 3.0851761476443105e-05, + "loss": 0.1169, + "step": 31700 + }, + { + "epoch": 0.7431213628931591, + "grad_norm": 0.0890028178691864, + "learning_rate": 3.0846440400825626e-05, + "loss": 0.0135, + "step": 31701 + }, + { + "epoch": 0.7431448044679642, + "grad_norm": 0.47511038184165955, + "learning_rate": 3.0841119700437305e-05, + "loss": 0.083, + "step": 31702 + }, + { + "epoch": 0.7431682460427691, + "grad_norm": 0.3245977759361267, + "learning_rate": 3.0835799375306974e-05, + "loss": 0.0421, + "step": 31703 + }, + { + "epoch": 0.7431916876175741, + "grad_norm": 0.6160818338394165, + "learning_rate": 3.0830479425463563e-05, + "loss": 0.1241, + "step": 31704 + }, + { + "epoch": 0.7432151291923792, + "grad_norm": 0.5583149790763855, + "learning_rate": 3.082515985093594e-05, + "loss": 0.1112, + "step": 31705 + }, + { + "epoch": 0.7432385707671841, + "grad_norm": 0.4350298345088959, + "learning_rate": 3.081984065175292e-05, + "loss": 0.0799, + "step": 31706 + }, + { + "epoch": 0.7432620123419892, + "grad_norm": 0.6581299901008606, + "learning_rate": 3.08145218279434e-05, + "loss": 0.1242, + "step": 31707 + }, + { + "epoch": 0.7432854539167941, + "grad_norm": 0.36731767654418945, + "learning_rate": 3.080920337953623e-05, + "loss": 0.3019, + "step": 31708 + }, + { + "epoch": 0.7433088954915992, + "grad_norm": 0.16049368679523468, + "learning_rate": 3.080388530656024e-05, + "loss": 0.0145, + "step": 31709 + }, + { + "epoch": 0.7433323370664041, + "grad_norm": 0.462789922952652, + "learning_rate": 3.079856760904434e-05, + "loss": 0.0305, + "step": 31710 + }, + { + "epoch": 0.7433557786412092, + "grad_norm": 0.3292247951030731, + "learning_rate": 3.0793250287017364e-05, + "loss": 0.0503, + "step": 31711 + }, + { + "epoch": 0.7433792202160141, + "grad_norm": 0.12998096644878387, + "learning_rate": 3.0787933340508126e-05, + "loss": 0.0157, + "step": 31712 + }, + { + "epoch": 0.7434026617908192, + "grad_norm": 0.4906131327152252, + "learning_rate": 3.0782616769545536e-05, + "loss": 0.0457, + "step": 31713 + }, + { + "epoch": 0.7434261033656241, + "grad_norm": 0.9806424975395203, + "learning_rate": 3.0777300574158404e-05, + "loss": 0.2598, + "step": 31714 + }, + { + "epoch": 0.7434495449404291, + "grad_norm": 0.17348793148994446, + "learning_rate": 3.077198475437559e-05, + "loss": 0.0311, + "step": 31715 + }, + { + "epoch": 0.7434729865152341, + "grad_norm": 0.5237704515457153, + "learning_rate": 3.076666931022592e-05, + "loss": 0.3706, + "step": 31716 + }, + { + "epoch": 0.7434964280900391, + "grad_norm": 0.23172515630722046, + "learning_rate": 3.076135424173826e-05, + "loss": 0.054, + "step": 31717 + }, + { + "epoch": 0.7435198696648441, + "grad_norm": 0.6546432971954346, + "learning_rate": 3.075603954894145e-05, + "loss": 0.6567, + "step": 31718 + }, + { + "epoch": 0.7435433112396491, + "grad_norm": 0.3694620132446289, + "learning_rate": 3.075072523186431e-05, + "loss": 0.0774, + "step": 31719 + }, + { + "epoch": 0.7435667528144541, + "grad_norm": 0.44730374217033386, + "learning_rate": 3.0745411290535684e-05, + "loss": 0.4356, + "step": 31720 + }, + { + "epoch": 0.7435901943892591, + "grad_norm": 0.20065917074680328, + "learning_rate": 3.074009772498441e-05, + "loss": 0.0285, + "step": 31721 + }, + { + "epoch": 0.743613635964064, + "grad_norm": 0.45513617992401123, + "learning_rate": 3.073478453523927e-05, + "loss": 0.1202, + "step": 31722 + }, + { + "epoch": 0.7436370775388691, + "grad_norm": 0.27915653586387634, + "learning_rate": 3.072947172132917e-05, + "loss": 0.0259, + "step": 31723 + }, + { + "epoch": 0.743660519113674, + "grad_norm": 0.21070115268230438, + "learning_rate": 3.0724159283282886e-05, + "loss": 0.053, + "step": 31724 + }, + { + "epoch": 0.7436839606884791, + "grad_norm": 0.5538550615310669, + "learning_rate": 3.071884722112928e-05, + "loss": 0.6262, + "step": 31725 + }, + { + "epoch": 0.743707402263284, + "grad_norm": 0.8184198141098022, + "learning_rate": 3.071353553489715e-05, + "loss": 0.1179, + "step": 31726 + }, + { + "epoch": 0.7437308438380891, + "grad_norm": 0.24551644921302795, + "learning_rate": 3.070822422461535e-05, + "loss": 0.022, + "step": 31727 + }, + { + "epoch": 0.743754285412894, + "grad_norm": 0.8931947350502014, + "learning_rate": 3.070291329031265e-05, + "loss": 0.1805, + "step": 31728 + }, + { + "epoch": 0.7437777269876991, + "grad_norm": 0.19295500218868256, + "learning_rate": 3.0697602732017875e-05, + "loss": 0.0309, + "step": 31729 + }, + { + "epoch": 0.743801168562504, + "grad_norm": 0.3797351121902466, + "learning_rate": 3.069229254975987e-05, + "loss": 0.0665, + "step": 31730 + }, + { + "epoch": 0.743824610137309, + "grad_norm": 0.549286961555481, + "learning_rate": 3.068698274356745e-05, + "loss": 0.6525, + "step": 31731 + }, + { + "epoch": 0.743848051712114, + "grad_norm": 0.1560349315404892, + "learning_rate": 3.0681673313469405e-05, + "loss": 0.043, + "step": 31732 + }, + { + "epoch": 0.743871493286919, + "grad_norm": 0.4483853578567505, + "learning_rate": 3.067636425949455e-05, + "loss": 0.6429, + "step": 31733 + }, + { + "epoch": 0.743894934861724, + "grad_norm": 0.2878493666648865, + "learning_rate": 3.067105558167166e-05, + "loss": 0.0895, + "step": 31734 + }, + { + "epoch": 0.743918376436529, + "grad_norm": 0.4051832854747772, + "learning_rate": 3.06657472800296e-05, + "loss": 0.083, + "step": 31735 + }, + { + "epoch": 0.743941818011334, + "grad_norm": 0.12260071188211441, + "learning_rate": 3.066043935459711e-05, + "loss": 0.0204, + "step": 31736 + }, + { + "epoch": 0.743965259586139, + "grad_norm": 0.292366623878479, + "learning_rate": 3.065513180540307e-05, + "loss": 0.0685, + "step": 31737 + }, + { + "epoch": 0.743988701160944, + "grad_norm": 0.6006960868835449, + "learning_rate": 3.0649824632476234e-05, + "loss": 0.5508, + "step": 31738 + }, + { + "epoch": 0.744012142735749, + "grad_norm": 0.12399212270975113, + "learning_rate": 3.0644517835845384e-05, + "loss": 0.0202, + "step": 31739 + }, + { + "epoch": 0.7440355843105539, + "grad_norm": 0.09657887369394302, + "learning_rate": 3.063921141553935e-05, + "loss": 0.0174, + "step": 31740 + }, + { + "epoch": 0.744059025885359, + "grad_norm": 0.27620968222618103, + "learning_rate": 3.0633905371586854e-05, + "loss": 0.0401, + "step": 31741 + }, + { + "epoch": 0.7440824674601639, + "grad_norm": 0.20419016480445862, + "learning_rate": 3.062859970401679e-05, + "loss": 0.0458, + "step": 31742 + }, + { + "epoch": 0.744105909034969, + "grad_norm": 0.44533616304397583, + "learning_rate": 3.0623294412857884e-05, + "loss": 0.0249, + "step": 31743 + }, + { + "epoch": 0.7441293506097739, + "grad_norm": 0.2939581573009491, + "learning_rate": 3.061798949813893e-05, + "loss": 0.0552, + "step": 31744 + }, + { + "epoch": 0.744152792184579, + "grad_norm": 0.4867713451385498, + "learning_rate": 3.061268495988871e-05, + "loss": 0.0854, + "step": 31745 + }, + { + "epoch": 0.7441762337593839, + "grad_norm": 0.28812864422798157, + "learning_rate": 3.060738079813603e-05, + "loss": 0.0582, + "step": 31746 + }, + { + "epoch": 0.744199675334189, + "grad_norm": 0.36774930357933044, + "learning_rate": 3.0602077012909616e-05, + "loss": 0.0591, + "step": 31747 + }, + { + "epoch": 0.7442231169089939, + "grad_norm": 0.12953582406044006, + "learning_rate": 3.059677360423828e-05, + "loss": 0.0183, + "step": 31748 + }, + { + "epoch": 0.744246558483799, + "grad_norm": 0.5670375823974609, + "learning_rate": 3.059147057215084e-05, + "loss": 0.0896, + "step": 31749 + }, + { + "epoch": 0.7442700000586039, + "grad_norm": 0.5446216464042664, + "learning_rate": 3.058616791667602e-05, + "loss": 0.7004, + "step": 31750 + }, + { + "epoch": 0.7442934416334089, + "grad_norm": 0.48844146728515625, + "learning_rate": 3.0580865637842604e-05, + "loss": 0.0736, + "step": 31751 + }, + { + "epoch": 0.7443168832082139, + "grad_norm": 0.6149383187294006, + "learning_rate": 3.057556373567938e-05, + "loss": 0.7444, + "step": 31752 + }, + { + "epoch": 0.7443403247830189, + "grad_norm": 0.6691286563873291, + "learning_rate": 3.057026221021508e-05, + "loss": 0.1468, + "step": 31753 + }, + { + "epoch": 0.7443637663578239, + "grad_norm": 0.22041335701942444, + "learning_rate": 3.056496106147847e-05, + "loss": 0.0313, + "step": 31754 + }, + { + "epoch": 0.7443872079326289, + "grad_norm": 0.4041055142879486, + "learning_rate": 3.055966028949836e-05, + "loss": 0.061, + "step": 31755 + }, + { + "epoch": 0.744410649507434, + "grad_norm": 0.4232904314994812, + "learning_rate": 3.0554359894303486e-05, + "loss": 0.0287, + "step": 31756 + }, + { + "epoch": 0.7444340910822389, + "grad_norm": 0.4140641987323761, + "learning_rate": 3.0549059875922615e-05, + "loss": 0.0911, + "step": 31757 + }, + { + "epoch": 0.744457532657044, + "grad_norm": 0.1476898342370987, + "learning_rate": 3.054376023438449e-05, + "loss": 0.0153, + "step": 31758 + }, + { + "epoch": 0.7444809742318489, + "grad_norm": 0.29382991790771484, + "learning_rate": 3.053846096971787e-05, + "loss": 0.083, + "step": 31759 + }, + { + "epoch": 0.7445044158066539, + "grad_norm": 0.5190749764442444, + "learning_rate": 3.053316208195149e-05, + "loss": 0.1042, + "step": 31760 + }, + { + "epoch": 0.7445278573814589, + "grad_norm": 0.5219392776489258, + "learning_rate": 3.0527863571114135e-05, + "loss": 0.1222, + "step": 31761 + }, + { + "epoch": 0.7445512989562639, + "grad_norm": 0.6460249423980713, + "learning_rate": 3.052256543723456e-05, + "loss": 0.4604, + "step": 31762 + }, + { + "epoch": 0.7445747405310689, + "grad_norm": 0.3719387352466583, + "learning_rate": 3.051726768034151e-05, + "loss": 0.0388, + "step": 31763 + }, + { + "epoch": 0.7445981821058739, + "grad_norm": 0.5240890979766846, + "learning_rate": 3.051197030046372e-05, + "loss": 0.0973, + "step": 31764 + }, + { + "epoch": 0.7446216236806789, + "grad_norm": 0.3752622604370117, + "learning_rate": 3.050667329762993e-05, + "loss": 0.0622, + "step": 31765 + }, + { + "epoch": 0.7446450652554839, + "grad_norm": 0.40310370922088623, + "learning_rate": 3.0501376671868875e-05, + "loss": 0.0729, + "step": 31766 + }, + { + "epoch": 0.7446685068302888, + "grad_norm": 0.3041723072528839, + "learning_rate": 3.0496080423209285e-05, + "loss": 0.0371, + "step": 31767 + }, + { + "epoch": 0.7446919484050939, + "grad_norm": 0.4590606987476349, + "learning_rate": 3.049078455167994e-05, + "loss": 0.073, + "step": 31768 + }, + { + "epoch": 0.7447153899798988, + "grad_norm": 0.41271859407424927, + "learning_rate": 3.0485489057309556e-05, + "loss": 0.0526, + "step": 31769 + }, + { + "epoch": 0.7447388315547039, + "grad_norm": 0.35224494338035583, + "learning_rate": 3.0480193940126867e-05, + "loss": 0.0542, + "step": 31770 + }, + { + "epoch": 0.7447622731295088, + "grad_norm": 0.7674646377563477, + "learning_rate": 3.04748992001606e-05, + "loss": 0.1421, + "step": 31771 + }, + { + "epoch": 0.7447857147043139, + "grad_norm": 0.3439701795578003, + "learning_rate": 3.0469604837439447e-05, + "loss": 0.039, + "step": 31772 + }, + { + "epoch": 0.7448091562791188, + "grad_norm": 0.2677305042743683, + "learning_rate": 3.0464310851992206e-05, + "loss": 0.0373, + "step": 31773 + }, + { + "epoch": 0.7448325978539239, + "grad_norm": 0.2953377366065979, + "learning_rate": 3.0459017243847542e-05, + "loss": 0.047, + "step": 31774 + }, + { + "epoch": 0.7448560394287288, + "grad_norm": 0.3129563331604004, + "learning_rate": 3.0453724013034236e-05, + "loss": 0.0601, + "step": 31775 + }, + { + "epoch": 0.7448794810035338, + "grad_norm": 0.3508114218711853, + "learning_rate": 3.044843115958098e-05, + "loss": 0.0552, + "step": 31776 + }, + { + "epoch": 0.7449029225783388, + "grad_norm": 0.19272451102733612, + "learning_rate": 3.0443138683516493e-05, + "loss": 0.0633, + "step": 31777 + }, + { + "epoch": 0.7449263641531438, + "grad_norm": 0.4482474625110626, + "learning_rate": 3.0437846584869478e-05, + "loss": 0.0781, + "step": 31778 + }, + { + "epoch": 0.7449498057279488, + "grad_norm": 0.06504885107278824, + "learning_rate": 3.043255486366865e-05, + "loss": 0.0093, + "step": 31779 + }, + { + "epoch": 0.7449732473027538, + "grad_norm": 0.4301554262638092, + "learning_rate": 3.042726351994276e-05, + "loss": 0.0654, + "step": 31780 + }, + { + "epoch": 0.7449966888775588, + "grad_norm": 0.393638014793396, + "learning_rate": 3.042197255372049e-05, + "loss": 0.0718, + "step": 31781 + }, + { + "epoch": 0.7450201304523638, + "grad_norm": 0.3319648504257202, + "learning_rate": 3.0416681965030557e-05, + "loss": 0.0474, + "step": 31782 + }, + { + "epoch": 0.7450435720271688, + "grad_norm": 0.15808045864105225, + "learning_rate": 3.0411391753901665e-05, + "loss": 0.0151, + "step": 31783 + }, + { + "epoch": 0.7450670136019738, + "grad_norm": 0.5788270235061646, + "learning_rate": 3.0406101920362483e-05, + "loss": 0.1006, + "step": 31784 + }, + { + "epoch": 0.7450904551767787, + "grad_norm": 0.3280344605445862, + "learning_rate": 3.0400812464441787e-05, + "loss": 0.0285, + "step": 31785 + }, + { + "epoch": 0.7451138967515838, + "grad_norm": 0.46745532751083374, + "learning_rate": 3.03955233861682e-05, + "loss": 0.4282, + "step": 31786 + }, + { + "epoch": 0.7451373383263887, + "grad_norm": 0.3586511015892029, + "learning_rate": 3.0390234685570496e-05, + "loss": 0.0367, + "step": 31787 + }, + { + "epoch": 0.7451607799011938, + "grad_norm": 0.16360822319984436, + "learning_rate": 3.0384946362677336e-05, + "loss": 0.02, + "step": 31788 + }, + { + "epoch": 0.7451842214759987, + "grad_norm": 0.3678152859210968, + "learning_rate": 3.037965841751741e-05, + "loss": 0.0578, + "step": 31789 + }, + { + "epoch": 0.7452076630508038, + "grad_norm": 0.30612167716026306, + "learning_rate": 3.0374370850119427e-05, + "loss": 0.0289, + "step": 31790 + }, + { + "epoch": 0.7452311046256087, + "grad_norm": 0.28188204765319824, + "learning_rate": 3.0369083660512054e-05, + "loss": 0.3463, + "step": 31791 + }, + { + "epoch": 0.7452545462004138, + "grad_norm": 0.30838486552238464, + "learning_rate": 3.036379684872396e-05, + "loss": 0.0488, + "step": 31792 + }, + { + "epoch": 0.7452779877752187, + "grad_norm": 0.49664002656936646, + "learning_rate": 3.0358510414783904e-05, + "loss": 0.0912, + "step": 31793 + }, + { + "epoch": 0.7453014293500237, + "grad_norm": 0.24704445898532867, + "learning_rate": 3.035322435872051e-05, + "loss": 0.0361, + "step": 31794 + }, + { + "epoch": 0.7453248709248287, + "grad_norm": 0.22328217327594757, + "learning_rate": 3.0347938680562494e-05, + "loss": 0.0546, + "step": 31795 + }, + { + "epoch": 0.7453483124996337, + "grad_norm": 0.46408823132514954, + "learning_rate": 3.0342653380338513e-05, + "loss": 0.1236, + "step": 31796 + }, + { + "epoch": 0.7453717540744387, + "grad_norm": 0.31978514790534973, + "learning_rate": 3.033736845807723e-05, + "loss": 0.0497, + "step": 31797 + }, + { + "epoch": 0.7453951956492437, + "grad_norm": 0.9174191951751709, + "learning_rate": 3.0332083913807374e-05, + "loss": 0.1385, + "step": 31798 + }, + { + "epoch": 0.7454186372240487, + "grad_norm": 0.36448895931243896, + "learning_rate": 3.0326799747557553e-05, + "loss": 0.0768, + "step": 31799 + }, + { + "epoch": 0.7454420787988537, + "grad_norm": 0.4339261054992676, + "learning_rate": 3.0321515959356518e-05, + "loss": 0.0794, + "step": 31800 + }, + { + "epoch": 0.7454655203736587, + "grad_norm": 0.43921560049057007, + "learning_rate": 3.031623254923288e-05, + "loss": 0.0809, + "step": 31801 + }, + { + "epoch": 0.7454889619484637, + "grad_norm": 0.5375577807426453, + "learning_rate": 3.031094951721534e-05, + "loss": 0.0889, + "step": 31802 + }, + { + "epoch": 0.7455124035232686, + "grad_norm": 0.13901951909065247, + "learning_rate": 3.030566686333254e-05, + "loss": 0.0319, + "step": 31803 + }, + { + "epoch": 0.7455358450980737, + "grad_norm": 0.5259523391723633, + "learning_rate": 3.030038458761315e-05, + "loss": 0.0988, + "step": 31804 + }, + { + "epoch": 0.7455592866728786, + "grad_norm": 0.32824406027793884, + "learning_rate": 3.02951026900858e-05, + "loss": 0.0285, + "step": 31805 + }, + { + "epoch": 0.7455827282476837, + "grad_norm": 0.23045594990253448, + "learning_rate": 3.028982117077922e-05, + "loss": 0.0268, + "step": 31806 + }, + { + "epoch": 0.7456061698224887, + "grad_norm": 0.3641943633556366, + "learning_rate": 3.028454002972202e-05, + "loss": 0.0476, + "step": 31807 + }, + { + "epoch": 0.7456296113972937, + "grad_norm": 0.3922012746334076, + "learning_rate": 3.027925926694288e-05, + "loss": 0.065, + "step": 31808 + }, + { + "epoch": 0.7456530529720987, + "grad_norm": 0.2903176546096802, + "learning_rate": 3.02739788824704e-05, + "loss": 0.0495, + "step": 31809 + }, + { + "epoch": 0.7456764945469037, + "grad_norm": 0.3061348795890808, + "learning_rate": 3.0268698876333303e-05, + "loss": 0.0559, + "step": 31810 + }, + { + "epoch": 0.7456999361217087, + "grad_norm": 0.5462215542793274, + "learning_rate": 3.026341924856021e-05, + "loss": 0.0413, + "step": 31811 + }, + { + "epoch": 0.7457233776965136, + "grad_norm": 2.618170738220215, + "learning_rate": 3.0258139999179723e-05, + "loss": 0.1376, + "step": 31812 + }, + { + "epoch": 0.7457468192713187, + "grad_norm": 0.3637050986289978, + "learning_rate": 3.0252861128220578e-05, + "loss": 0.0445, + "step": 31813 + }, + { + "epoch": 0.7457702608461236, + "grad_norm": 0.5422897338867188, + "learning_rate": 3.0247582635711358e-05, + "loss": 0.1196, + "step": 31814 + }, + { + "epoch": 0.7457937024209287, + "grad_norm": 0.19879423081874847, + "learning_rate": 3.024230452168072e-05, + "loss": 0.0617, + "step": 31815 + }, + { + "epoch": 0.7458171439957336, + "grad_norm": 0.5111644864082336, + "learning_rate": 3.0237026786157296e-05, + "loss": 0.0818, + "step": 31816 + }, + { + "epoch": 0.7458405855705387, + "grad_norm": 0.5799282193183899, + "learning_rate": 3.0231749429169698e-05, + "loss": 0.1035, + "step": 31817 + }, + { + "epoch": 0.7458640271453436, + "grad_norm": 0.2741325795650482, + "learning_rate": 3.0226472450746614e-05, + "loss": 0.0465, + "step": 31818 + }, + { + "epoch": 0.7458874687201487, + "grad_norm": 0.4193410575389862, + "learning_rate": 3.0221195850916662e-05, + "loss": 0.0781, + "step": 31819 + }, + { + "epoch": 0.7459109102949536, + "grad_norm": 0.35864749550819397, + "learning_rate": 3.0215919629708456e-05, + "loss": 0.038, + "step": 31820 + }, + { + "epoch": 0.7459343518697586, + "grad_norm": 0.3722599744796753, + "learning_rate": 3.0210643787150604e-05, + "loss": 0.0532, + "step": 31821 + }, + { + "epoch": 0.7459577934445636, + "grad_norm": 0.14858461916446686, + "learning_rate": 3.0205368323271788e-05, + "loss": 0.0377, + "step": 31822 + }, + { + "epoch": 0.7459812350193686, + "grad_norm": 0.14529980719089508, + "learning_rate": 3.0200093238100612e-05, + "loss": 0.018, + "step": 31823 + }, + { + "epoch": 0.7460046765941736, + "grad_norm": 0.059752367436885834, + "learning_rate": 3.0194818531665657e-05, + "loss": 0.0056, + "step": 31824 + }, + { + "epoch": 0.7460281181689786, + "grad_norm": 0.1999962329864502, + "learning_rate": 3.0189544203995613e-05, + "loss": 0.0369, + "step": 31825 + }, + { + "epoch": 0.7460515597437836, + "grad_norm": 0.49692490696907043, + "learning_rate": 3.0184270255119063e-05, + "loss": 0.0929, + "step": 31826 + }, + { + "epoch": 0.7460750013185886, + "grad_norm": 0.09125708043575287, + "learning_rate": 3.017899668506463e-05, + "loss": 0.0118, + "step": 31827 + }, + { + "epoch": 0.7460984428933936, + "grad_norm": 0.3780296742916107, + "learning_rate": 3.017372349386093e-05, + "loss": 0.0775, + "step": 31828 + }, + { + "epoch": 0.7461218844681986, + "grad_norm": 0.31351178884506226, + "learning_rate": 3.0168450681536554e-05, + "loss": 0.0675, + "step": 31829 + }, + { + "epoch": 0.7461453260430035, + "grad_norm": 0.2478463053703308, + "learning_rate": 3.0163178248120107e-05, + "loss": 0.0504, + "step": 31830 + }, + { + "epoch": 0.7461687676178086, + "grad_norm": 0.5320703387260437, + "learning_rate": 3.015790619364024e-05, + "loss": 0.1095, + "step": 31831 + }, + { + "epoch": 0.7461922091926135, + "grad_norm": 0.261844277381897, + "learning_rate": 3.0152634518125544e-05, + "loss": 0.0502, + "step": 31832 + }, + { + "epoch": 0.7462156507674186, + "grad_norm": 0.35544419288635254, + "learning_rate": 3.0147363221604596e-05, + "loss": 0.0776, + "step": 31833 + }, + { + "epoch": 0.7462390923422235, + "grad_norm": 0.35477307438850403, + "learning_rate": 3.014209230410604e-05, + "loss": 0.0539, + "step": 31834 + }, + { + "epoch": 0.7462625339170286, + "grad_norm": 0.32445308566093445, + "learning_rate": 3.013682176565845e-05, + "loss": 0.0537, + "step": 31835 + }, + { + "epoch": 0.7462859754918335, + "grad_norm": 0.6267589330673218, + "learning_rate": 3.0131551606290443e-05, + "loss": 0.6724, + "step": 31836 + }, + { + "epoch": 0.7463094170666386, + "grad_norm": 0.13798679411411285, + "learning_rate": 3.0126281826030557e-05, + "loss": 0.0235, + "step": 31837 + }, + { + "epoch": 0.7463328586414435, + "grad_norm": 0.05404122546315193, + "learning_rate": 3.012101242490747e-05, + "loss": 0.0031, + "step": 31838 + }, + { + "epoch": 0.7463563002162485, + "grad_norm": 0.38175904750823975, + "learning_rate": 3.0115743402949736e-05, + "loss": 0.0689, + "step": 31839 + }, + { + "epoch": 0.7463797417910535, + "grad_norm": 0.4486825466156006, + "learning_rate": 3.0110474760185935e-05, + "loss": 0.0499, + "step": 31840 + }, + { + "epoch": 0.7464031833658585, + "grad_norm": 0.07815741002559662, + "learning_rate": 3.0105206496644667e-05, + "loss": 0.0078, + "step": 31841 + }, + { + "epoch": 0.7464266249406635, + "grad_norm": 0.28789621591567993, + "learning_rate": 3.0099938612354517e-05, + "loss": 0.0371, + "step": 31842 + }, + { + "epoch": 0.7464500665154685, + "grad_norm": 0.1995568871498108, + "learning_rate": 3.009467110734403e-05, + "loss": 0.0398, + "step": 31843 + }, + { + "epoch": 0.7464735080902735, + "grad_norm": 0.3481040596961975, + "learning_rate": 3.0089403981641863e-05, + "loss": 0.0469, + "step": 31844 + }, + { + "epoch": 0.7464969496650785, + "grad_norm": 0.23816938698291779, + "learning_rate": 3.008413723527651e-05, + "loss": 0.0248, + "step": 31845 + }, + { + "epoch": 0.7465203912398835, + "grad_norm": 0.22501511871814728, + "learning_rate": 3.007887086827663e-05, + "loss": 0.033, + "step": 31846 + }, + { + "epoch": 0.7465438328146885, + "grad_norm": 0.47624534368515015, + "learning_rate": 3.007360488067077e-05, + "loss": 0.0889, + "step": 31847 + }, + { + "epoch": 0.7465672743894934, + "grad_norm": 0.4703413248062134, + "learning_rate": 3.006833927248749e-05, + "loss": 0.0839, + "step": 31848 + }, + { + "epoch": 0.7465907159642985, + "grad_norm": 0.33057647943496704, + "learning_rate": 3.006307404375537e-05, + "loss": 0.0281, + "step": 31849 + }, + { + "epoch": 0.7466141575391034, + "grad_norm": 0.45826423168182373, + "learning_rate": 3.005780919450295e-05, + "loss": 0.0459, + "step": 31850 + }, + { + "epoch": 0.7466375991139085, + "grad_norm": 0.17705298960208893, + "learning_rate": 3.0052544724758847e-05, + "loss": 0.0226, + "step": 31851 + }, + { + "epoch": 0.7466610406887134, + "grad_norm": 0.5410919785499573, + "learning_rate": 3.00472806345516e-05, + "loss": 0.5255, + "step": 31852 + }, + { + "epoch": 0.7466844822635185, + "grad_norm": 0.4386042654514313, + "learning_rate": 3.0042016923909788e-05, + "loss": 0.3547, + "step": 31853 + }, + { + "epoch": 0.7467079238383234, + "grad_norm": 0.4337126612663269, + "learning_rate": 3.0036753592861956e-05, + "loss": 0.0573, + "step": 31854 + }, + { + "epoch": 0.7467313654131285, + "grad_norm": 0.5280452966690063, + "learning_rate": 3.0031490641436622e-05, + "loss": 0.084, + "step": 31855 + }, + { + "epoch": 0.7467548069879334, + "grad_norm": 0.5609124302864075, + "learning_rate": 3.0026228069662433e-05, + "loss": 0.1122, + "step": 31856 + }, + { + "epoch": 0.7467782485627384, + "grad_norm": 0.2651471197605133, + "learning_rate": 3.0020965877567852e-05, + "loss": 0.0461, + "step": 31857 + }, + { + "epoch": 0.7468016901375435, + "grad_norm": 0.4400993883609772, + "learning_rate": 3.0015704065181525e-05, + "loss": 0.0701, + "step": 31858 + }, + { + "epoch": 0.7468251317123484, + "grad_norm": 0.6159844994544983, + "learning_rate": 3.0010442632531945e-05, + "loss": 0.1647, + "step": 31859 + }, + { + "epoch": 0.7468485732871535, + "grad_norm": 0.16201899945735931, + "learning_rate": 3.000518157964768e-05, + "loss": 0.0291, + "step": 31860 + }, + { + "epoch": 0.7468720148619584, + "grad_norm": 0.6017575263977051, + "learning_rate": 2.9999920906557265e-05, + "loss": 0.127, + "step": 31861 + }, + { + "epoch": 0.7468954564367635, + "grad_norm": 0.2441680133342743, + "learning_rate": 2.9994660613289204e-05, + "loss": 0.06, + "step": 31862 + }, + { + "epoch": 0.7469188980115684, + "grad_norm": 0.30262696743011475, + "learning_rate": 2.9989400699872128e-05, + "loss": 0.0692, + "step": 31863 + }, + { + "epoch": 0.7469423395863735, + "grad_norm": 0.636476457118988, + "learning_rate": 2.9984141166334535e-05, + "loss": 0.1246, + "step": 31864 + }, + { + "epoch": 0.7469657811611784, + "grad_norm": 0.41344505548477173, + "learning_rate": 2.9978882012704945e-05, + "loss": 0.0483, + "step": 31865 + }, + { + "epoch": 0.7469892227359834, + "grad_norm": 0.32670438289642334, + "learning_rate": 2.997362323901193e-05, + "loss": 0.059, + "step": 31866 + }, + { + "epoch": 0.7470126643107884, + "grad_norm": 0.4118999242782593, + "learning_rate": 2.996836484528399e-05, + "loss": 0.0741, + "step": 31867 + }, + { + "epoch": 0.7470361058855934, + "grad_norm": 0.3680069148540497, + "learning_rate": 2.9963106831549637e-05, + "loss": 0.0793, + "step": 31868 + }, + { + "epoch": 0.7470595474603984, + "grad_norm": 0.4333973228931427, + "learning_rate": 2.9957849197837474e-05, + "loss": 0.0466, + "step": 31869 + }, + { + "epoch": 0.7470829890352034, + "grad_norm": 0.5153944492340088, + "learning_rate": 2.9952591944175968e-05, + "loss": 0.1093, + "step": 31870 + }, + { + "epoch": 0.7471064306100084, + "grad_norm": 0.16059952974319458, + "learning_rate": 2.9947335070593684e-05, + "loss": 0.0134, + "step": 31871 + }, + { + "epoch": 0.7471298721848134, + "grad_norm": 0.08217816799879074, + "learning_rate": 2.9942078577119138e-05, + "loss": 0.0107, + "step": 31872 + }, + { + "epoch": 0.7471533137596184, + "grad_norm": 0.5580101013183594, + "learning_rate": 2.9936822463780843e-05, + "loss": 0.0921, + "step": 31873 + }, + { + "epoch": 0.7471767553344234, + "grad_norm": 0.19997891783714294, + "learning_rate": 2.993156673060732e-05, + "loss": 0.0317, + "step": 31874 + }, + { + "epoch": 0.7472001969092283, + "grad_norm": 0.14981666207313538, + "learning_rate": 2.9926311377627046e-05, + "loss": 0.018, + "step": 31875 + }, + { + "epoch": 0.7472236384840334, + "grad_norm": 0.3147900402545929, + "learning_rate": 2.992105640486862e-05, + "loss": 0.0425, + "step": 31876 + }, + { + "epoch": 0.7472470800588383, + "grad_norm": 0.6181470155715942, + "learning_rate": 2.99158018123605e-05, + "loss": 0.616, + "step": 31877 + }, + { + "epoch": 0.7472705216336434, + "grad_norm": 0.4806985855102539, + "learning_rate": 2.991054760013122e-05, + "loss": 0.3446, + "step": 31878 + }, + { + "epoch": 0.7472939632084483, + "grad_norm": 0.18578343093395233, + "learning_rate": 2.9905293768209275e-05, + "loss": 0.0322, + "step": 31879 + }, + { + "epoch": 0.7473174047832534, + "grad_norm": 0.6141202449798584, + "learning_rate": 2.990004031662318e-05, + "loss": 0.1097, + "step": 31880 + }, + { + "epoch": 0.7473408463580583, + "grad_norm": 0.31034356355667114, + "learning_rate": 2.98947872454014e-05, + "loss": 0.0141, + "step": 31881 + }, + { + "epoch": 0.7473642879328634, + "grad_norm": 0.12558437883853912, + "learning_rate": 2.9889534554572473e-05, + "loss": 0.0113, + "step": 31882 + }, + { + "epoch": 0.7473877295076683, + "grad_norm": 0.4329705536365509, + "learning_rate": 2.988428224416495e-05, + "loss": 0.0909, + "step": 31883 + }, + { + "epoch": 0.7474111710824733, + "grad_norm": 0.08117284625768661, + "learning_rate": 2.9879030314207267e-05, + "loss": 0.0071, + "step": 31884 + }, + { + "epoch": 0.7474346126572783, + "grad_norm": 0.3787256181240082, + "learning_rate": 2.9873778764727943e-05, + "loss": 0.0547, + "step": 31885 + }, + { + "epoch": 0.7474580542320833, + "grad_norm": 0.49572476744651794, + "learning_rate": 2.9868527595755468e-05, + "loss": 0.1318, + "step": 31886 + }, + { + "epoch": 0.7474814958068883, + "grad_norm": 0.2708002030849457, + "learning_rate": 2.986327680731834e-05, + "loss": 0.0271, + "step": 31887 + }, + { + "epoch": 0.7475049373816933, + "grad_norm": 0.5346424579620361, + "learning_rate": 2.9858026399445006e-05, + "loss": 0.1189, + "step": 31888 + }, + { + "epoch": 0.7475283789564983, + "grad_norm": 0.11899523437023163, + "learning_rate": 2.9852776372164025e-05, + "loss": 0.0186, + "step": 31889 + }, + { + "epoch": 0.7475518205313033, + "grad_norm": 0.5124374032020569, + "learning_rate": 2.9847526725503848e-05, + "loss": 0.0433, + "step": 31890 + }, + { + "epoch": 0.7475752621061083, + "grad_norm": 0.2641180455684662, + "learning_rate": 2.9842277459492973e-05, + "loss": 0.0539, + "step": 31891 + }, + { + "epoch": 0.7475987036809133, + "grad_norm": 0.3969751000404358, + "learning_rate": 2.9837028574159863e-05, + "loss": 0.0765, + "step": 31892 + }, + { + "epoch": 0.7476221452557182, + "grad_norm": 0.33407914638519287, + "learning_rate": 2.983178006953299e-05, + "loss": 0.0361, + "step": 31893 + }, + { + "epoch": 0.7476455868305233, + "grad_norm": 0.551764965057373, + "learning_rate": 2.982653194564087e-05, + "loss": 0.0408, + "step": 31894 + }, + { + "epoch": 0.7476690284053282, + "grad_norm": 0.8011298775672913, + "learning_rate": 2.9821284202511935e-05, + "loss": 0.1303, + "step": 31895 + }, + { + "epoch": 0.7476924699801333, + "grad_norm": 0.22103866934776306, + "learning_rate": 2.981603684017471e-05, + "loss": 0.0306, + "step": 31896 + }, + { + "epoch": 0.7477159115549382, + "grad_norm": 0.5387338995933533, + "learning_rate": 2.981078985865765e-05, + "loss": 0.0763, + "step": 31897 + }, + { + "epoch": 0.7477393531297433, + "grad_norm": 0.2207091897726059, + "learning_rate": 2.9805543257989223e-05, + "loss": 0.0254, + "step": 31898 + }, + { + "epoch": 0.7477627947045482, + "grad_norm": 0.18390192091464996, + "learning_rate": 2.9800297038197877e-05, + "loss": 0.0278, + "step": 31899 + }, + { + "epoch": 0.7477862362793533, + "grad_norm": 0.6780083179473877, + "learning_rate": 2.9795051199312075e-05, + "loss": 0.1695, + "step": 31900 + }, + { + "epoch": 0.7478096778541582, + "grad_norm": 0.290567547082901, + "learning_rate": 2.9789805741360323e-05, + "loss": 0.0567, + "step": 31901 + }, + { + "epoch": 0.7478331194289632, + "grad_norm": 0.3423796594142914, + "learning_rate": 2.9784560664371065e-05, + "loss": 0.4732, + "step": 31902 + }, + { + "epoch": 0.7478565610037682, + "grad_norm": 0.4021366536617279, + "learning_rate": 2.9779315968372744e-05, + "loss": 0.0709, + "step": 31903 + }, + { + "epoch": 0.7478800025785732, + "grad_norm": 0.4481271803379059, + "learning_rate": 2.9774071653393844e-05, + "loss": 0.0559, + "step": 31904 + }, + { + "epoch": 0.7479034441533782, + "grad_norm": 0.5060120224952698, + "learning_rate": 2.9768827719462798e-05, + "loss": 0.0918, + "step": 31905 + }, + { + "epoch": 0.7479268857281832, + "grad_norm": 0.40903595089912415, + "learning_rate": 2.9763584166608037e-05, + "loss": 0.0647, + "step": 31906 + }, + { + "epoch": 0.7479503273029882, + "grad_norm": 0.40789955854415894, + "learning_rate": 2.9758340994858037e-05, + "loss": 0.4463, + "step": 31907 + }, + { + "epoch": 0.7479737688777932, + "grad_norm": 0.14388956129550934, + "learning_rate": 2.9753098204241293e-05, + "loss": 0.0455, + "step": 31908 + }, + { + "epoch": 0.7479972104525983, + "grad_norm": 0.18581633269786835, + "learning_rate": 2.9747855794786207e-05, + "loss": 0.0178, + "step": 31909 + }, + { + "epoch": 0.7480206520274032, + "grad_norm": 0.5698232054710388, + "learning_rate": 2.9742613766521244e-05, + "loss": 0.0716, + "step": 31910 + }, + { + "epoch": 0.7480440936022082, + "grad_norm": 0.27920615673065186, + "learning_rate": 2.9737372119474815e-05, + "loss": 0.0144, + "step": 31911 + }, + { + "epoch": 0.7480675351770132, + "grad_norm": 0.5035266876220703, + "learning_rate": 2.9732130853675398e-05, + "loss": 0.4127, + "step": 31912 + }, + { + "epoch": 0.7480909767518182, + "grad_norm": 0.2951366603374481, + "learning_rate": 2.9726889969151373e-05, + "loss": 0.0269, + "step": 31913 + }, + { + "epoch": 0.7481144183266232, + "grad_norm": 0.25299105048179626, + "learning_rate": 2.9721649465931256e-05, + "loss": 0.0361, + "step": 31914 + }, + { + "epoch": 0.7481378599014282, + "grad_norm": 0.6366047263145447, + "learning_rate": 2.9716409344043438e-05, + "loss": 0.7479, + "step": 31915 + }, + { + "epoch": 0.7481613014762332, + "grad_norm": 0.4950771629810333, + "learning_rate": 2.9711169603516365e-05, + "loss": 0.0668, + "step": 31916 + }, + { + "epoch": 0.7481847430510382, + "grad_norm": 0.32230424880981445, + "learning_rate": 2.9705930244378467e-05, + "loss": 0.0507, + "step": 31917 + }, + { + "epoch": 0.7482081846258432, + "grad_norm": 0.32412347197532654, + "learning_rate": 2.9700691266658122e-05, + "loss": 0.0668, + "step": 31918 + }, + { + "epoch": 0.7482316262006482, + "grad_norm": 0.7760697603225708, + "learning_rate": 2.969545267038384e-05, + "loss": 0.141, + "step": 31919 + }, + { + "epoch": 0.7482550677754531, + "grad_norm": 0.49393367767333984, + "learning_rate": 2.9690214455583986e-05, + "loss": 0.0832, + "step": 31920 + }, + { + "epoch": 0.7482785093502582, + "grad_norm": 0.6924063563346863, + "learning_rate": 2.968497662228703e-05, + "loss": 0.0591, + "step": 31921 + }, + { + "epoch": 0.7483019509250631, + "grad_norm": 0.450749933719635, + "learning_rate": 2.967973917052137e-05, + "loss": 0.0835, + "step": 31922 + }, + { + "epoch": 0.7483253924998682, + "grad_norm": 0.6949143409729004, + "learning_rate": 2.9674502100315417e-05, + "loss": 0.6241, + "step": 31923 + }, + { + "epoch": 0.7483488340746731, + "grad_norm": 0.19052891433238983, + "learning_rate": 2.96692654116976e-05, + "loss": 0.0285, + "step": 31924 + }, + { + "epoch": 0.7483722756494782, + "grad_norm": 0.5145760178565979, + "learning_rate": 2.9664029104696324e-05, + "loss": 0.0948, + "step": 31925 + }, + { + "epoch": 0.7483957172242831, + "grad_norm": 0.18741512298583984, + "learning_rate": 2.965879317933996e-05, + "loss": 0.0232, + "step": 31926 + }, + { + "epoch": 0.7484191587990882, + "grad_norm": 0.4907032549381256, + "learning_rate": 2.9653557635657013e-05, + "loss": 0.1074, + "step": 31927 + }, + { + "epoch": 0.7484426003738931, + "grad_norm": 2.1524438858032227, + "learning_rate": 2.9648322473675826e-05, + "loss": 0.085, + "step": 31928 + }, + { + "epoch": 0.7484660419486981, + "grad_norm": 0.6302312016487122, + "learning_rate": 2.9643087693424832e-05, + "loss": 0.0813, + "step": 31929 + }, + { + "epoch": 0.7484894835235031, + "grad_norm": 0.2864644229412079, + "learning_rate": 2.9637853294932384e-05, + "loss": 0.0338, + "step": 31930 + }, + { + "epoch": 0.7485129250983081, + "grad_norm": 0.5882852673530579, + "learning_rate": 2.963261927822695e-05, + "loss": 0.3841, + "step": 31931 + }, + { + "epoch": 0.7485363666731131, + "grad_norm": 0.5992704629898071, + "learning_rate": 2.9627385643336915e-05, + "loss": 0.0749, + "step": 31932 + }, + { + "epoch": 0.7485598082479181, + "grad_norm": 0.4671054482460022, + "learning_rate": 2.9622152390290615e-05, + "loss": 0.0825, + "step": 31933 + }, + { + "epoch": 0.7485832498227231, + "grad_norm": 0.15144242346286774, + "learning_rate": 2.961691951911654e-05, + "loss": 0.0109, + "step": 31934 + }, + { + "epoch": 0.7486066913975281, + "grad_norm": 0.31178414821624756, + "learning_rate": 2.961168702984304e-05, + "loss": 0.0588, + "step": 31935 + }, + { + "epoch": 0.748630132972333, + "grad_norm": 0.45245450735092163, + "learning_rate": 2.9606454922498493e-05, + "loss": 0.0942, + "step": 31936 + }, + { + "epoch": 0.7486535745471381, + "grad_norm": 0.6356929540634155, + "learning_rate": 2.9601223197111306e-05, + "loss": 0.1712, + "step": 31937 + }, + { + "epoch": 0.748677016121943, + "grad_norm": 0.14260511100292206, + "learning_rate": 2.9595991853709838e-05, + "loss": 0.0176, + "step": 31938 + }, + { + "epoch": 0.7487004576967481, + "grad_norm": 0.48780104517936707, + "learning_rate": 2.9590760892322524e-05, + "loss": 0.0972, + "step": 31939 + }, + { + "epoch": 0.748723899271553, + "grad_norm": 0.3853543698787689, + "learning_rate": 2.9585530312977717e-05, + "loss": 0.0505, + "step": 31940 + }, + { + "epoch": 0.7487473408463581, + "grad_norm": 0.09814980626106262, + "learning_rate": 2.9580300115703808e-05, + "loss": 0.0116, + "step": 31941 + }, + { + "epoch": 0.748770782421163, + "grad_norm": 0.5986500382423401, + "learning_rate": 2.9575070300529162e-05, + "loss": 0.5839, + "step": 31942 + }, + { + "epoch": 0.7487942239959681, + "grad_norm": 0.22370921075344086, + "learning_rate": 2.9569840867482134e-05, + "loss": 0.0226, + "step": 31943 + }, + { + "epoch": 0.748817665570773, + "grad_norm": 0.089035265147686, + "learning_rate": 2.956461181659117e-05, + "loss": 0.0045, + "step": 31944 + }, + { + "epoch": 0.748841107145578, + "grad_norm": 0.43799328804016113, + "learning_rate": 2.9559383147884567e-05, + "loss": 0.0567, + "step": 31945 + }, + { + "epoch": 0.748864548720383, + "grad_norm": 0.49772560596466064, + "learning_rate": 2.955415486139077e-05, + "loss": 0.0745, + "step": 31946 + }, + { + "epoch": 0.748887990295188, + "grad_norm": 0.16094335913658142, + "learning_rate": 2.95489269571381e-05, + "loss": 0.0244, + "step": 31947 + }, + { + "epoch": 0.748911431869993, + "grad_norm": 0.1398497372865677, + "learning_rate": 2.9543699435154937e-05, + "loss": 0.0282, + "step": 31948 + }, + { + "epoch": 0.748934873444798, + "grad_norm": 0.49278396368026733, + "learning_rate": 2.9538472295469644e-05, + "loss": 0.0507, + "step": 31949 + }, + { + "epoch": 0.748958315019603, + "grad_norm": 0.7718290090560913, + "learning_rate": 2.9533245538110575e-05, + "loss": 0.1218, + "step": 31950 + }, + { + "epoch": 0.748981756594408, + "grad_norm": 0.3748892843723297, + "learning_rate": 2.9528019163106068e-05, + "loss": 0.0661, + "step": 31951 + }, + { + "epoch": 0.749005198169213, + "grad_norm": 0.41996753215789795, + "learning_rate": 2.9522793170484543e-05, + "loss": 0.0588, + "step": 31952 + }, + { + "epoch": 0.749028639744018, + "grad_norm": 0.1510503739118576, + "learning_rate": 2.9517567560274327e-05, + "loss": 0.0259, + "step": 31953 + }, + { + "epoch": 0.749052081318823, + "grad_norm": 0.4768840968608856, + "learning_rate": 2.9512342332503773e-05, + "loss": 0.0766, + "step": 31954 + }, + { + "epoch": 0.749075522893628, + "grad_norm": 0.7484465837478638, + "learning_rate": 2.9507117487201195e-05, + "loss": 0.1201, + "step": 31955 + }, + { + "epoch": 0.7490989644684329, + "grad_norm": 0.5453270673751831, + "learning_rate": 2.950189302439501e-05, + "loss": 0.0621, + "step": 31956 + }, + { + "epoch": 0.749122406043238, + "grad_norm": 0.08993956446647644, + "learning_rate": 2.9496668944113535e-05, + "loss": 0.0099, + "step": 31957 + }, + { + "epoch": 0.7491458476180429, + "grad_norm": 0.10699938237667084, + "learning_rate": 2.9491445246385086e-05, + "loss": 0.0154, + "step": 31958 + }, + { + "epoch": 0.749169289192848, + "grad_norm": 0.41676560044288635, + "learning_rate": 2.9486221931238078e-05, + "loss": 0.0759, + "step": 31959 + }, + { + "epoch": 0.749192730767653, + "grad_norm": 0.5273450613021851, + "learning_rate": 2.9480998998700802e-05, + "loss": 0.1101, + "step": 31960 + }, + { + "epoch": 0.749216172342458, + "grad_norm": 0.17665621638298035, + "learning_rate": 2.9475776448801607e-05, + "loss": 0.0158, + "step": 31961 + }, + { + "epoch": 0.749239613917263, + "grad_norm": 0.5317404270172119, + "learning_rate": 2.9470554281568842e-05, + "loss": 0.112, + "step": 31962 + }, + { + "epoch": 0.749263055492068, + "grad_norm": 0.408258318901062, + "learning_rate": 2.9465332497030827e-05, + "loss": 0.0932, + "step": 31963 + }, + { + "epoch": 0.749286497066873, + "grad_norm": 0.3002016544342041, + "learning_rate": 2.946011109521587e-05, + "loss": 0.0555, + "step": 31964 + }, + { + "epoch": 0.7493099386416779, + "grad_norm": 0.6806029081344604, + "learning_rate": 2.9454890076152362e-05, + "loss": 0.074, + "step": 31965 + }, + { + "epoch": 0.749333380216483, + "grad_norm": 0.519865095615387, + "learning_rate": 2.944966943986861e-05, + "loss": 0.5736, + "step": 31966 + }, + { + "epoch": 0.7493568217912879, + "grad_norm": 0.29277530312538147, + "learning_rate": 2.9444449186392896e-05, + "loss": 0.0347, + "step": 31967 + }, + { + "epoch": 0.749380263366093, + "grad_norm": 0.7434586882591248, + "learning_rate": 2.9439229315753625e-05, + "loss": 0.1725, + "step": 31968 + }, + { + "epoch": 0.7494037049408979, + "grad_norm": 0.17512774467468262, + "learning_rate": 2.943400982797907e-05, + "loss": 0.0249, + "step": 31969 + }, + { + "epoch": 0.749427146515703, + "grad_norm": 0.23876875638961792, + "learning_rate": 2.9428790723097564e-05, + "loss": 0.0394, + "step": 31970 + }, + { + "epoch": 0.7494505880905079, + "grad_norm": 0.23412171006202698, + "learning_rate": 2.94235720011374e-05, + "loss": 0.0518, + "step": 31971 + }, + { + "epoch": 0.749474029665313, + "grad_norm": 0.1804700493812561, + "learning_rate": 2.9418353662126953e-05, + "loss": 0.0307, + "step": 31972 + }, + { + "epoch": 0.7494974712401179, + "grad_norm": 0.8175259232521057, + "learning_rate": 2.9413135706094485e-05, + "loss": 0.1117, + "step": 31973 + }, + { + "epoch": 0.749520912814923, + "grad_norm": 0.26296231150627136, + "learning_rate": 2.9407918133068347e-05, + "loss": 0.0344, + "step": 31974 + }, + { + "epoch": 0.7495443543897279, + "grad_norm": 0.5697599649429321, + "learning_rate": 2.9402700943076823e-05, + "loss": 0.1037, + "step": 31975 + }, + { + "epoch": 0.7495677959645329, + "grad_norm": 0.2535419166088104, + "learning_rate": 2.939748413614819e-05, + "loss": 0.0362, + "step": 31976 + }, + { + "epoch": 0.7495912375393379, + "grad_norm": 0.35006216168403625, + "learning_rate": 2.939226771231083e-05, + "loss": 0.0642, + "step": 31977 + }, + { + "epoch": 0.7496146791141429, + "grad_norm": 0.4443950355052948, + "learning_rate": 2.9387051671593014e-05, + "loss": 0.0357, + "step": 31978 + }, + { + "epoch": 0.7496381206889479, + "grad_norm": 0.24153177440166473, + "learning_rate": 2.9381836014023012e-05, + "loss": 0.0431, + "step": 31979 + }, + { + "epoch": 0.7496615622637529, + "grad_norm": 0.40045058727264404, + "learning_rate": 2.937662073962918e-05, + "loss": 0.108, + "step": 31980 + }, + { + "epoch": 0.7496850038385579, + "grad_norm": 0.46154916286468506, + "learning_rate": 2.9371405848439783e-05, + "loss": 0.0931, + "step": 31981 + }, + { + "epoch": 0.7497084454133629, + "grad_norm": 0.7109189033508301, + "learning_rate": 2.936619134048313e-05, + "loss": 0.1495, + "step": 31982 + }, + { + "epoch": 0.7497318869881678, + "grad_norm": 0.47446635365486145, + "learning_rate": 2.9360977215787477e-05, + "loss": 0.1006, + "step": 31983 + }, + { + "epoch": 0.7497553285629729, + "grad_norm": 0.4688689112663269, + "learning_rate": 2.935576347438118e-05, + "loss": 0.628, + "step": 31984 + }, + { + "epoch": 0.7497787701377778, + "grad_norm": 0.8229082822799683, + "learning_rate": 2.9350550116292497e-05, + "loss": 0.2085, + "step": 31985 + }, + { + "epoch": 0.7498022117125829, + "grad_norm": 0.6910369992256165, + "learning_rate": 2.9345337141549722e-05, + "loss": 0.1214, + "step": 31986 + }, + { + "epoch": 0.7498256532873878, + "grad_norm": 0.29091158509254456, + "learning_rate": 2.9340124550181126e-05, + "loss": 0.0564, + "step": 31987 + }, + { + "epoch": 0.7498490948621929, + "grad_norm": 0.20886826515197754, + "learning_rate": 2.9334912342215004e-05, + "loss": 0.0276, + "step": 31988 + }, + { + "epoch": 0.7498725364369978, + "grad_norm": 0.4769321084022522, + "learning_rate": 2.9329700517679603e-05, + "loss": 0.1118, + "step": 31989 + }, + { + "epoch": 0.7498959780118029, + "grad_norm": 0.32004937529563904, + "learning_rate": 2.9324489076603258e-05, + "loss": 0.0483, + "step": 31990 + }, + { + "epoch": 0.7499194195866078, + "grad_norm": 0.1808752864599228, + "learning_rate": 2.9319278019014194e-05, + "loss": 0.0166, + "step": 31991 + }, + { + "epoch": 0.7499428611614128, + "grad_norm": 0.338688462972641, + "learning_rate": 2.9314067344940755e-05, + "loss": 0.4539, + "step": 31992 + }, + { + "epoch": 0.7499663027362178, + "grad_norm": 0.29465997219085693, + "learning_rate": 2.9308857054411166e-05, + "loss": 0.0303, + "step": 31993 + }, + { + "epoch": 0.7499897443110228, + "grad_norm": 0.1140250489115715, + "learning_rate": 2.9303647147453715e-05, + "loss": 0.0057, + "step": 31994 + }, + { + "epoch": 0.7500131858858278, + "grad_norm": 0.2328081876039505, + "learning_rate": 2.9298437624096652e-05, + "loss": 0.0518, + "step": 31995 + }, + { + "epoch": 0.7500366274606328, + "grad_norm": 0.23403994739055634, + "learning_rate": 2.9293228484368228e-05, + "loss": 0.0526, + "step": 31996 + }, + { + "epoch": 0.7500600690354378, + "grad_norm": 0.7425317764282227, + "learning_rate": 2.9288019728296767e-05, + "loss": 0.2061, + "step": 31997 + }, + { + "epoch": 0.7500835106102428, + "grad_norm": 0.12401463836431503, + "learning_rate": 2.92828113559105e-05, + "loss": 0.0161, + "step": 31998 + }, + { + "epoch": 0.7501069521850477, + "grad_norm": 0.07252620160579681, + "learning_rate": 2.9277603367237684e-05, + "loss": 0.0048, + "step": 31999 + }, + { + "epoch": 0.7501303937598528, + "grad_norm": 0.31705498695373535, + "learning_rate": 2.927239576230658e-05, + "loss": 0.0288, + "step": 32000 + }, + { + "epoch": 0.7501538353346577, + "grad_norm": 0.342799574136734, + "learning_rate": 2.926718854114544e-05, + "loss": 0.4467, + "step": 32001 + }, + { + "epoch": 0.7501772769094628, + "grad_norm": 0.33767271041870117, + "learning_rate": 2.9261981703782505e-05, + "loss": 0.0326, + "step": 32002 + }, + { + "epoch": 0.7502007184842677, + "grad_norm": 0.5252028107643127, + "learning_rate": 2.9256775250246037e-05, + "loss": 0.0778, + "step": 32003 + }, + { + "epoch": 0.7502241600590728, + "grad_norm": 0.46819812059402466, + "learning_rate": 2.9251569180564332e-05, + "loss": 0.0501, + "step": 32004 + }, + { + "epoch": 0.7502476016338777, + "grad_norm": 0.5127869844436646, + "learning_rate": 2.9246363494765592e-05, + "loss": 0.1188, + "step": 32005 + }, + { + "epoch": 0.7502710432086828, + "grad_norm": 0.7212106585502625, + "learning_rate": 2.924115819287808e-05, + "loss": 0.1588, + "step": 32006 + }, + { + "epoch": 0.7502944847834877, + "grad_norm": 0.39693647623062134, + "learning_rate": 2.923595327493004e-05, + "loss": 0.0605, + "step": 32007 + }, + { + "epoch": 0.7503179263582928, + "grad_norm": 0.3603202700614929, + "learning_rate": 2.9230748740949666e-05, + "loss": 0.0721, + "step": 32008 + }, + { + "epoch": 0.7503413679330977, + "grad_norm": 0.4753865599632263, + "learning_rate": 2.9225544590965282e-05, + "loss": 0.1017, + "step": 32009 + }, + { + "epoch": 0.7503648095079027, + "grad_norm": 0.5393516421318054, + "learning_rate": 2.9220340825005068e-05, + "loss": 0.1632, + "step": 32010 + }, + { + "epoch": 0.7503882510827078, + "grad_norm": 0.34912657737731934, + "learning_rate": 2.9215137443097286e-05, + "loss": 0.0278, + "step": 32011 + }, + { + "epoch": 0.7504116926575127, + "grad_norm": 0.33186715841293335, + "learning_rate": 2.920993444527015e-05, + "loss": 0.0911, + "step": 32012 + }, + { + "epoch": 0.7504351342323178, + "grad_norm": 0.9662271738052368, + "learning_rate": 2.9204731831551912e-05, + "loss": 0.1076, + "step": 32013 + }, + { + "epoch": 0.7504585758071227, + "grad_norm": 0.33024492859840393, + "learning_rate": 2.9199529601970744e-05, + "loss": 0.0421, + "step": 32014 + }, + { + "epoch": 0.7504820173819278, + "grad_norm": 0.0756092518568039, + "learning_rate": 2.9194327756554927e-05, + "loss": 0.0081, + "step": 32015 + }, + { + "epoch": 0.7505054589567327, + "grad_norm": 0.33127835392951965, + "learning_rate": 2.9189126295332714e-05, + "loss": 0.0578, + "step": 32016 + }, + { + "epoch": 0.7505289005315378, + "grad_norm": 0.4805730879306793, + "learning_rate": 2.9183925218332285e-05, + "loss": 0.1112, + "step": 32017 + }, + { + "epoch": 0.7505523421063427, + "grad_norm": 0.40252071619033813, + "learning_rate": 2.9178724525581868e-05, + "loss": 0.0627, + "step": 32018 + }, + { + "epoch": 0.7505757836811477, + "grad_norm": 0.32139724493026733, + "learning_rate": 2.9173524217109693e-05, + "loss": 0.0862, + "step": 32019 + }, + { + "epoch": 0.7505992252559527, + "grad_norm": 0.2627830505371094, + "learning_rate": 2.9168324292943962e-05, + "loss": 0.0429, + "step": 32020 + }, + { + "epoch": 0.7506226668307577, + "grad_norm": 0.14967580139636993, + "learning_rate": 2.9163124753112858e-05, + "loss": 0.0336, + "step": 32021 + }, + { + "epoch": 0.7506461084055627, + "grad_norm": 0.5018420219421387, + "learning_rate": 2.9157925597644675e-05, + "loss": 0.082, + "step": 32022 + }, + { + "epoch": 0.7506695499803677, + "grad_norm": 0.6337891221046448, + "learning_rate": 2.9152726826567568e-05, + "loss": 0.1075, + "step": 32023 + }, + { + "epoch": 0.7506929915551727, + "grad_norm": 0.4111677408218384, + "learning_rate": 2.9147528439909754e-05, + "loss": 0.0822, + "step": 32024 + }, + { + "epoch": 0.7507164331299777, + "grad_norm": 0.5455731749534607, + "learning_rate": 2.914233043769945e-05, + "loss": 0.5723, + "step": 32025 + }, + { + "epoch": 0.7507398747047827, + "grad_norm": 0.34211066365242004, + "learning_rate": 2.9137132819964864e-05, + "loss": 0.0679, + "step": 32026 + }, + { + "epoch": 0.7507633162795877, + "grad_norm": 0.4844132363796234, + "learning_rate": 2.913193558673414e-05, + "loss": 0.0763, + "step": 32027 + }, + { + "epoch": 0.7507867578543926, + "grad_norm": 0.3676425814628601, + "learning_rate": 2.9126738738035532e-05, + "loss": 0.4384, + "step": 32028 + }, + { + "epoch": 0.7508101994291977, + "grad_norm": 0.3842182755470276, + "learning_rate": 2.9121542273897262e-05, + "loss": 0.0522, + "step": 32029 + }, + { + "epoch": 0.7508336410040026, + "grad_norm": 0.3975045084953308, + "learning_rate": 2.91163461943475e-05, + "loss": 0.048, + "step": 32030 + }, + { + "epoch": 0.7508570825788077, + "grad_norm": 0.12556354701519012, + "learning_rate": 2.911115049941443e-05, + "loss": 0.0192, + "step": 32031 + }, + { + "epoch": 0.7508805241536126, + "grad_norm": 0.5792248845100403, + "learning_rate": 2.9105955189126254e-05, + "loss": 0.7006, + "step": 32032 + }, + { + "epoch": 0.7509039657284177, + "grad_norm": 0.13994301855564117, + "learning_rate": 2.9100760263511162e-05, + "loss": 0.0245, + "step": 32033 + }, + { + "epoch": 0.7509274073032226, + "grad_norm": 0.8039514422416687, + "learning_rate": 2.9095565722597306e-05, + "loss": 0.1263, + "step": 32034 + }, + { + "epoch": 0.7509508488780277, + "grad_norm": 0.3883284330368042, + "learning_rate": 2.9090371566412933e-05, + "loss": 0.0422, + "step": 32035 + }, + { + "epoch": 0.7509742904528326, + "grad_norm": 0.42501163482666016, + "learning_rate": 2.9085177794986196e-05, + "loss": 0.0529, + "step": 32036 + }, + { + "epoch": 0.7509977320276376, + "grad_norm": 0.43193304538726807, + "learning_rate": 2.9079984408345286e-05, + "loss": 0.0854, + "step": 32037 + }, + { + "epoch": 0.7510211736024426, + "grad_norm": 0.32817143201828003, + "learning_rate": 2.9074791406518366e-05, + "loss": 0.0476, + "step": 32038 + }, + { + "epoch": 0.7510446151772476, + "grad_norm": 0.3327732980251312, + "learning_rate": 2.9069598789533625e-05, + "loss": 0.0872, + "step": 32039 + }, + { + "epoch": 0.7510680567520526, + "grad_norm": 0.36197763681411743, + "learning_rate": 2.9064406557419212e-05, + "loss": 0.0705, + "step": 32040 + }, + { + "epoch": 0.7510914983268576, + "grad_norm": 0.5438422560691833, + "learning_rate": 2.905921471020331e-05, + "loss": 0.0882, + "step": 32041 + }, + { + "epoch": 0.7511149399016626, + "grad_norm": 0.5191016793251038, + "learning_rate": 2.9054023247914143e-05, + "loss": 0.0946, + "step": 32042 + }, + { + "epoch": 0.7511383814764676, + "grad_norm": 0.5026108026504517, + "learning_rate": 2.904883217057984e-05, + "loss": 0.0557, + "step": 32043 + }, + { + "epoch": 0.7511618230512725, + "grad_norm": 0.507207453250885, + "learning_rate": 2.9043641478228568e-05, + "loss": 0.1066, + "step": 32044 + }, + { + "epoch": 0.7511852646260776, + "grad_norm": 0.4483104944229126, + "learning_rate": 2.9038451170888493e-05, + "loss": 0.0381, + "step": 32045 + }, + { + "epoch": 0.7512087062008825, + "grad_norm": 0.5152592658996582, + "learning_rate": 2.903326124858774e-05, + "loss": 0.076, + "step": 32046 + }, + { + "epoch": 0.7512321477756876, + "grad_norm": 0.2954416275024414, + "learning_rate": 2.9028071711354544e-05, + "loss": 0.0412, + "step": 32047 + }, + { + "epoch": 0.7512555893504925, + "grad_norm": 0.3793093264102936, + "learning_rate": 2.9022882559217024e-05, + "loss": 0.0575, + "step": 32048 + }, + { + "epoch": 0.7512790309252976, + "grad_norm": 0.30939000844955444, + "learning_rate": 2.901769379220335e-05, + "loss": 0.0528, + "step": 32049 + }, + { + "epoch": 0.7513024725001025, + "grad_norm": 0.5211654901504517, + "learning_rate": 2.901250541034165e-05, + "loss": 0.0959, + "step": 32050 + }, + { + "epoch": 0.7513259140749076, + "grad_norm": 0.4253501296043396, + "learning_rate": 2.9007317413660096e-05, + "loss": 0.0765, + "step": 32051 + }, + { + "epoch": 0.7513493556497125, + "grad_norm": 0.6216554045677185, + "learning_rate": 2.9002129802186806e-05, + "loss": 0.091, + "step": 32052 + }, + { + "epoch": 0.7513727972245176, + "grad_norm": 0.9689485430717468, + "learning_rate": 2.8996942575949958e-05, + "loss": 0.0912, + "step": 32053 + }, + { + "epoch": 0.7513962387993225, + "grad_norm": 0.2900626063346863, + "learning_rate": 2.899175573497772e-05, + "loss": 0.0535, + "step": 32054 + }, + { + "epoch": 0.7514196803741275, + "grad_norm": 0.29351019859313965, + "learning_rate": 2.8986569279298215e-05, + "loss": 0.0293, + "step": 32055 + }, + { + "epoch": 0.7514431219489325, + "grad_norm": 0.32649609446525574, + "learning_rate": 2.8981383208939582e-05, + "loss": 0.0379, + "step": 32056 + }, + { + "epoch": 0.7514665635237375, + "grad_norm": 0.4124945402145386, + "learning_rate": 2.897619752392996e-05, + "loss": 0.0629, + "step": 32057 + }, + { + "epoch": 0.7514900050985425, + "grad_norm": 0.4031147062778473, + "learning_rate": 2.897101222429749e-05, + "loss": 0.069, + "step": 32058 + }, + { + "epoch": 0.7515134466733475, + "grad_norm": 0.443880170583725, + "learning_rate": 2.896582731007027e-05, + "loss": 0.0512, + "step": 32059 + }, + { + "epoch": 0.7515368882481525, + "grad_norm": 0.426033079624176, + "learning_rate": 2.8960642781276502e-05, + "loss": 0.0514, + "step": 32060 + }, + { + "epoch": 0.7515603298229575, + "grad_norm": 0.3337339162826538, + "learning_rate": 2.8955458637944276e-05, + "loss": 0.0542, + "step": 32061 + }, + { + "epoch": 0.7515837713977626, + "grad_norm": 0.2911967933177948, + "learning_rate": 2.8950274880101746e-05, + "loss": 0.0314, + "step": 32062 + }, + { + "epoch": 0.7516072129725675, + "grad_norm": 0.4202089309692383, + "learning_rate": 2.894509150777701e-05, + "loss": 0.1067, + "step": 32063 + }, + { + "epoch": 0.7516306545473725, + "grad_norm": 0.7125017046928406, + "learning_rate": 2.893990852099817e-05, + "loss": 0.1365, + "step": 32064 + }, + { + "epoch": 0.7516540961221775, + "grad_norm": 0.3329467177391052, + "learning_rate": 2.8934725919793427e-05, + "loss": 0.0914, + "step": 32065 + }, + { + "epoch": 0.7516775376969825, + "grad_norm": 0.7658994793891907, + "learning_rate": 2.892954370419082e-05, + "loss": 0.136, + "step": 32066 + }, + { + "epoch": 0.7517009792717875, + "grad_norm": 0.31076687574386597, + "learning_rate": 2.892436187421854e-05, + "loss": 0.0993, + "step": 32067 + }, + { + "epoch": 0.7517244208465925, + "grad_norm": 0.6465831398963928, + "learning_rate": 2.891918042990468e-05, + "loss": 0.07, + "step": 32068 + }, + { + "epoch": 0.7517478624213975, + "grad_norm": 0.3926612436771393, + "learning_rate": 2.8913999371277335e-05, + "loss": 0.0529, + "step": 32069 + }, + { + "epoch": 0.7517713039962025, + "grad_norm": 0.493039608001709, + "learning_rate": 2.8908818698364627e-05, + "loss": 0.0627, + "step": 32070 + }, + { + "epoch": 0.7517947455710075, + "grad_norm": 0.41425982117652893, + "learning_rate": 2.8903638411194678e-05, + "loss": 0.05, + "step": 32071 + }, + { + "epoch": 0.7518181871458125, + "grad_norm": 0.37459951639175415, + "learning_rate": 2.8898458509795544e-05, + "loss": 0.0539, + "step": 32072 + }, + { + "epoch": 0.7518416287206174, + "grad_norm": 0.10570627450942993, + "learning_rate": 2.8893278994195404e-05, + "loss": 0.018, + "step": 32073 + }, + { + "epoch": 0.7518650702954225, + "grad_norm": 0.09155001491308212, + "learning_rate": 2.8888099864422347e-05, + "loss": 0.0153, + "step": 32074 + }, + { + "epoch": 0.7518885118702274, + "grad_norm": 0.11536643654108047, + "learning_rate": 2.888292112050445e-05, + "loss": 0.0189, + "step": 32075 + }, + { + "epoch": 0.7519119534450325, + "grad_norm": 0.47033387422561646, + "learning_rate": 2.8877742762469784e-05, + "loss": 0.0829, + "step": 32076 + }, + { + "epoch": 0.7519353950198374, + "grad_norm": 0.5404831767082214, + "learning_rate": 2.8872564790346524e-05, + "loss": 0.4675, + "step": 32077 + }, + { + "epoch": 0.7519588365946425, + "grad_norm": 0.7467077374458313, + "learning_rate": 2.886738720416272e-05, + "loss": 0.2162, + "step": 32078 + }, + { + "epoch": 0.7519822781694474, + "grad_norm": 0.4480263888835907, + "learning_rate": 2.8862210003946456e-05, + "loss": 0.0702, + "step": 32079 + }, + { + "epoch": 0.7520057197442525, + "grad_norm": 0.40448081493377686, + "learning_rate": 2.8857033189725868e-05, + "loss": 0.0348, + "step": 32080 + }, + { + "epoch": 0.7520291613190574, + "grad_norm": 0.3561602532863617, + "learning_rate": 2.8851856761529007e-05, + "loss": 0.0614, + "step": 32081 + }, + { + "epoch": 0.7520526028938624, + "grad_norm": 0.4564451575279236, + "learning_rate": 2.884668071938398e-05, + "loss": 0.5255, + "step": 32082 + }, + { + "epoch": 0.7520760444686674, + "grad_norm": 0.6380407214164734, + "learning_rate": 2.8841505063318862e-05, + "loss": 0.0915, + "step": 32083 + }, + { + "epoch": 0.7520994860434724, + "grad_norm": 0.38489776849746704, + "learning_rate": 2.8836329793361717e-05, + "loss": 0.0508, + "step": 32084 + }, + { + "epoch": 0.7521229276182774, + "grad_norm": 0.7288286089897156, + "learning_rate": 2.883115490954067e-05, + "loss": 0.1463, + "step": 32085 + }, + { + "epoch": 0.7521463691930824, + "grad_norm": 0.30030184984207153, + "learning_rate": 2.882598041188378e-05, + "loss": 0.0198, + "step": 32086 + }, + { + "epoch": 0.7521698107678874, + "grad_norm": 0.30111175775527954, + "learning_rate": 2.8820806300419113e-05, + "loss": 0.0403, + "step": 32087 + }, + { + "epoch": 0.7521932523426924, + "grad_norm": 0.23298700153827667, + "learning_rate": 2.8815632575174734e-05, + "loss": 0.038, + "step": 32088 + }, + { + "epoch": 0.7522166939174973, + "grad_norm": 0.32860761880874634, + "learning_rate": 2.8810459236178755e-05, + "loss": 0.0517, + "step": 32089 + }, + { + "epoch": 0.7522401354923024, + "grad_norm": 0.2004210203886032, + "learning_rate": 2.8805286283459232e-05, + "loss": 0.0119, + "step": 32090 + }, + { + "epoch": 0.7522635770671073, + "grad_norm": 0.444496750831604, + "learning_rate": 2.8800113717044186e-05, + "loss": 0.0515, + "step": 32091 + }, + { + "epoch": 0.7522870186419124, + "grad_norm": 0.3847750723361969, + "learning_rate": 2.879494153696177e-05, + "loss": 0.0791, + "step": 32092 + }, + { + "epoch": 0.7523104602167173, + "grad_norm": 0.45034199953079224, + "learning_rate": 2.8789769743240002e-05, + "loss": 0.0754, + "step": 32093 + }, + { + "epoch": 0.7523339017915224, + "grad_norm": 0.25824013352394104, + "learning_rate": 2.878459833590693e-05, + "loss": 0.074, + "step": 32094 + }, + { + "epoch": 0.7523573433663273, + "grad_norm": 0.5051159858703613, + "learning_rate": 2.8779427314990646e-05, + "loss": 0.5162, + "step": 32095 + }, + { + "epoch": 0.7523807849411324, + "grad_norm": 0.2588030993938446, + "learning_rate": 2.8774256680519185e-05, + "loss": 0.0435, + "step": 32096 + }, + { + "epoch": 0.7524042265159373, + "grad_norm": 0.4755012094974518, + "learning_rate": 2.8769086432520587e-05, + "loss": 0.0684, + "step": 32097 + }, + { + "epoch": 0.7524276680907424, + "grad_norm": 0.7494131326675415, + "learning_rate": 2.8763916571022953e-05, + "loss": 0.1251, + "step": 32098 + }, + { + "epoch": 0.7524511096655473, + "grad_norm": 0.6205427050590515, + "learning_rate": 2.8758747096054307e-05, + "loss": 0.1112, + "step": 32099 + }, + { + "epoch": 0.7524745512403523, + "grad_norm": 0.4113822281360626, + "learning_rate": 2.8753578007642712e-05, + "loss": 0.0494, + "step": 32100 + }, + { + "epoch": 0.7524979928151573, + "grad_norm": 0.2925722897052765, + "learning_rate": 2.8748409305816172e-05, + "loss": 0.048, + "step": 32101 + }, + { + "epoch": 0.7525214343899623, + "grad_norm": 0.16033808887004852, + "learning_rate": 2.8743240990602795e-05, + "loss": 0.0316, + "step": 32102 + }, + { + "epoch": 0.7525448759647673, + "grad_norm": 0.45572054386138916, + "learning_rate": 2.873807306203059e-05, + "loss": 0.0468, + "step": 32103 + }, + { + "epoch": 0.7525683175395723, + "grad_norm": 0.20348547399044037, + "learning_rate": 2.8732905520127586e-05, + "loss": 0.0319, + "step": 32104 + }, + { + "epoch": 0.7525917591143773, + "grad_norm": 0.29896166920661926, + "learning_rate": 2.8727738364921865e-05, + "loss": 0.0408, + "step": 32105 + }, + { + "epoch": 0.7526152006891823, + "grad_norm": 0.6816180348396301, + "learning_rate": 2.8722571596441437e-05, + "loss": 0.099, + "step": 32106 + }, + { + "epoch": 0.7526386422639872, + "grad_norm": 0.39593181014060974, + "learning_rate": 2.8717405214714342e-05, + "loss": 0.0631, + "step": 32107 + }, + { + "epoch": 0.7526620838387923, + "grad_norm": 0.2569828927516937, + "learning_rate": 2.871223921976861e-05, + "loss": 0.0268, + "step": 32108 + }, + { + "epoch": 0.7526855254135972, + "grad_norm": 0.46294504404067993, + "learning_rate": 2.8707073611632274e-05, + "loss": 0.0491, + "step": 32109 + }, + { + "epoch": 0.7527089669884023, + "grad_norm": 0.5403158068656921, + "learning_rate": 2.8701908390333333e-05, + "loss": 0.1115, + "step": 32110 + }, + { + "epoch": 0.7527324085632072, + "grad_norm": 0.09105290472507477, + "learning_rate": 2.8696743555899863e-05, + "loss": 0.0097, + "step": 32111 + }, + { + "epoch": 0.7527558501380123, + "grad_norm": 0.36481669545173645, + "learning_rate": 2.869157910835988e-05, + "loss": 0.0504, + "step": 32112 + }, + { + "epoch": 0.7527792917128173, + "grad_norm": 0.18963629007339478, + "learning_rate": 2.8686415047741357e-05, + "loss": 0.0386, + "step": 32113 + }, + { + "epoch": 0.7528027332876223, + "grad_norm": 0.4841170907020569, + "learning_rate": 2.868125137407238e-05, + "loss": 0.1385, + "step": 32114 + }, + { + "epoch": 0.7528261748624273, + "grad_norm": 0.5180361270904541, + "learning_rate": 2.8676088087380937e-05, + "loss": 0.0933, + "step": 32115 + }, + { + "epoch": 0.7528496164372322, + "grad_norm": 0.30689653754234314, + "learning_rate": 2.8670925187695053e-05, + "loss": 0.0293, + "step": 32116 + }, + { + "epoch": 0.7528730580120373, + "grad_norm": 0.4165189266204834, + "learning_rate": 2.8665762675042696e-05, + "loss": 0.0849, + "step": 32117 + }, + { + "epoch": 0.7528964995868422, + "grad_norm": 0.6543183922767639, + "learning_rate": 2.8660600549451955e-05, + "loss": 0.5311, + "step": 32118 + }, + { + "epoch": 0.7529199411616473, + "grad_norm": 0.32394546270370483, + "learning_rate": 2.8655438810950798e-05, + "loss": 0.0465, + "step": 32119 + }, + { + "epoch": 0.7529433827364522, + "grad_norm": 0.38749364018440247, + "learning_rate": 2.8650277459567232e-05, + "loss": 0.0374, + "step": 32120 + }, + { + "epoch": 0.7529668243112573, + "grad_norm": 0.5261133909225464, + "learning_rate": 2.864511649532926e-05, + "loss": 0.063, + "step": 32121 + }, + { + "epoch": 0.7529902658860622, + "grad_norm": 0.510644257068634, + "learning_rate": 2.863995591826487e-05, + "loss": 0.1121, + "step": 32122 + }, + { + "epoch": 0.7530137074608673, + "grad_norm": 0.3829723298549652, + "learning_rate": 2.863479572840212e-05, + "loss": 0.0533, + "step": 32123 + }, + { + "epoch": 0.7530371490356722, + "grad_norm": 0.7372503876686096, + "learning_rate": 2.8629635925768973e-05, + "loss": 0.1572, + "step": 32124 + }, + { + "epoch": 0.7530605906104773, + "grad_norm": 0.1814192533493042, + "learning_rate": 2.8624476510393395e-05, + "loss": 0.0111, + "step": 32125 + }, + { + "epoch": 0.7530840321852822, + "grad_norm": 0.5869743824005127, + "learning_rate": 2.8619317482303452e-05, + "loss": 0.1473, + "step": 32126 + }, + { + "epoch": 0.7531074737600872, + "grad_norm": 0.11081141233444214, + "learning_rate": 2.8614158841527094e-05, + "loss": 0.0241, + "step": 32127 + }, + { + "epoch": 0.7531309153348922, + "grad_norm": 0.150969460606575, + "learning_rate": 2.8609000588092316e-05, + "loss": 0.0246, + "step": 32128 + }, + { + "epoch": 0.7531543569096972, + "grad_norm": 0.23843920230865479, + "learning_rate": 2.8603842722027087e-05, + "loss": 0.0296, + "step": 32129 + }, + { + "epoch": 0.7531777984845022, + "grad_norm": 0.11544256657361984, + "learning_rate": 2.8598685243359437e-05, + "loss": 0.0242, + "step": 32130 + }, + { + "epoch": 0.7532012400593072, + "grad_norm": 0.4764936864376068, + "learning_rate": 2.8593528152117322e-05, + "loss": 0.0832, + "step": 32131 + }, + { + "epoch": 0.7532246816341122, + "grad_norm": 0.598206639289856, + "learning_rate": 2.8588371448328745e-05, + "loss": 0.0888, + "step": 32132 + }, + { + "epoch": 0.7532481232089172, + "grad_norm": 0.544338583946228, + "learning_rate": 2.8583215132021657e-05, + "loss": 0.1005, + "step": 32133 + }, + { + "epoch": 0.7532715647837221, + "grad_norm": 0.18998321890830994, + "learning_rate": 2.8578059203224063e-05, + "loss": 0.0442, + "step": 32134 + }, + { + "epoch": 0.7532950063585272, + "grad_norm": 0.45041388273239136, + "learning_rate": 2.857290366196389e-05, + "loss": 0.0581, + "step": 32135 + }, + { + "epoch": 0.7533184479333321, + "grad_norm": 0.18397268652915955, + "learning_rate": 2.8567748508269187e-05, + "loss": 0.0175, + "step": 32136 + }, + { + "epoch": 0.7533418895081372, + "grad_norm": 0.3938104510307312, + "learning_rate": 2.8562593742167864e-05, + "loss": 0.509, + "step": 32137 + }, + { + "epoch": 0.7533653310829421, + "grad_norm": 0.4226592481136322, + "learning_rate": 2.8557439363687933e-05, + "loss": 0.0765, + "step": 32138 + }, + { + "epoch": 0.7533887726577472, + "grad_norm": 0.444295734167099, + "learning_rate": 2.8552285372857358e-05, + "loss": 0.0932, + "step": 32139 + }, + { + "epoch": 0.7534122142325521, + "grad_norm": 0.6940943598747253, + "learning_rate": 2.8547131769704082e-05, + "loss": 0.107, + "step": 32140 + }, + { + "epoch": 0.7534356558073572, + "grad_norm": 0.11635105311870575, + "learning_rate": 2.8541978554256075e-05, + "loss": 0.0204, + "step": 32141 + }, + { + "epoch": 0.7534590973821621, + "grad_norm": 0.44770756363868713, + "learning_rate": 2.8536825726541283e-05, + "loss": 0.0843, + "step": 32142 + }, + { + "epoch": 0.7534825389569672, + "grad_norm": 0.5918171405792236, + "learning_rate": 2.8531673286587702e-05, + "loss": 0.0945, + "step": 32143 + }, + { + "epoch": 0.7535059805317721, + "grad_norm": 0.5626821517944336, + "learning_rate": 2.8526521234423286e-05, + "loss": 0.0783, + "step": 32144 + }, + { + "epoch": 0.7535294221065771, + "grad_norm": 0.24637144804000854, + "learning_rate": 2.852136957007596e-05, + "loss": 0.0267, + "step": 32145 + }, + { + "epoch": 0.7535528636813821, + "grad_norm": 0.055857256054878235, + "learning_rate": 2.8516218293573704e-05, + "loss": 0.0083, + "step": 32146 + }, + { + "epoch": 0.7535763052561871, + "grad_norm": 0.4665274918079376, + "learning_rate": 2.8511067404944448e-05, + "loss": 0.0336, + "step": 32147 + }, + { + "epoch": 0.7535997468309921, + "grad_norm": 0.18095220625400543, + "learning_rate": 2.8505916904216123e-05, + "loss": 0.0339, + "step": 32148 + }, + { + "epoch": 0.7536231884057971, + "grad_norm": 0.27137288451194763, + "learning_rate": 2.850076679141671e-05, + "loss": 0.0548, + "step": 32149 + }, + { + "epoch": 0.7536466299806021, + "grad_norm": 0.6439425945281982, + "learning_rate": 2.8495617066574176e-05, + "loss": 0.0781, + "step": 32150 + }, + { + "epoch": 0.7536700715554071, + "grad_norm": 0.6393788456916809, + "learning_rate": 2.8490467729716432e-05, + "loss": 0.0927, + "step": 32151 + }, + { + "epoch": 0.753693513130212, + "grad_norm": 0.4721868932247162, + "learning_rate": 2.8485318780871416e-05, + "loss": 0.0754, + "step": 32152 + }, + { + "epoch": 0.7537169547050171, + "grad_norm": 0.2795189619064331, + "learning_rate": 2.848017022006707e-05, + "loss": 0.0645, + "step": 32153 + }, + { + "epoch": 0.753740396279822, + "grad_norm": 0.44092586636543274, + "learning_rate": 2.8475022047331335e-05, + "loss": 0.128, + "step": 32154 + }, + { + "epoch": 0.7537638378546271, + "grad_norm": 0.37920740246772766, + "learning_rate": 2.846987426269212e-05, + "loss": 0.0808, + "step": 32155 + }, + { + "epoch": 0.753787279429432, + "grad_norm": 0.3319895267486572, + "learning_rate": 2.8464726866177393e-05, + "loss": 0.0407, + "step": 32156 + }, + { + "epoch": 0.7538107210042371, + "grad_norm": 0.3678751289844513, + "learning_rate": 2.8459579857815077e-05, + "loss": 0.0751, + "step": 32157 + }, + { + "epoch": 0.753834162579042, + "grad_norm": 0.5233991146087646, + "learning_rate": 2.8454433237633094e-05, + "loss": 0.0926, + "step": 32158 + }, + { + "epoch": 0.7538576041538471, + "grad_norm": 0.5347973704338074, + "learning_rate": 2.8449287005659362e-05, + "loss": 0.0763, + "step": 32159 + }, + { + "epoch": 0.753881045728652, + "grad_norm": 0.45908084511756897, + "learning_rate": 2.844414116192179e-05, + "loss": 0.1364, + "step": 32160 + }, + { + "epoch": 0.753904487303457, + "grad_norm": 0.42487838864326477, + "learning_rate": 2.8438995706448348e-05, + "loss": 0.0336, + "step": 32161 + }, + { + "epoch": 0.753927928878262, + "grad_norm": 0.21932075917720795, + "learning_rate": 2.843385063926689e-05, + "loss": 0.0411, + "step": 32162 + }, + { + "epoch": 0.753951370453067, + "grad_norm": 0.39393237233161926, + "learning_rate": 2.8428705960405412e-05, + "loss": 0.0848, + "step": 32163 + }, + { + "epoch": 0.7539748120278721, + "grad_norm": 0.12106145173311234, + "learning_rate": 2.8423561669891775e-05, + "loss": 0.0181, + "step": 32164 + }, + { + "epoch": 0.753998253602677, + "grad_norm": 0.718953013420105, + "learning_rate": 2.8418417767753915e-05, + "loss": 0.3872, + "step": 32165 + }, + { + "epoch": 0.7540216951774821, + "grad_norm": 0.7245931625366211, + "learning_rate": 2.841327425401973e-05, + "loss": 0.1284, + "step": 32166 + }, + { + "epoch": 0.754045136752287, + "grad_norm": 0.3259882926940918, + "learning_rate": 2.8408131128717097e-05, + "loss": 0.0375, + "step": 32167 + }, + { + "epoch": 0.7540685783270921, + "grad_norm": 0.1513916254043579, + "learning_rate": 2.8402988391873985e-05, + "loss": 0.0086, + "step": 32168 + }, + { + "epoch": 0.754092019901897, + "grad_norm": 0.25300636887550354, + "learning_rate": 2.8397846043518283e-05, + "loss": 0.0482, + "step": 32169 + }, + { + "epoch": 0.754115461476702, + "grad_norm": 0.6264988780021667, + "learning_rate": 2.8392704083677878e-05, + "loss": 0.5072, + "step": 32170 + }, + { + "epoch": 0.754138903051507, + "grad_norm": 0.14535027742385864, + "learning_rate": 2.838756251238067e-05, + "loss": 0.0193, + "step": 32171 + }, + { + "epoch": 0.754162344626312, + "grad_norm": 0.1297474205493927, + "learning_rate": 2.8382421329654563e-05, + "loss": 0.0285, + "step": 32172 + }, + { + "epoch": 0.754185786201117, + "grad_norm": 0.2541470229625702, + "learning_rate": 2.8377280535527418e-05, + "loss": 0.0412, + "step": 32173 + }, + { + "epoch": 0.754209227775922, + "grad_norm": 0.12793666124343872, + "learning_rate": 2.837214013002717e-05, + "loss": 0.0229, + "step": 32174 + }, + { + "epoch": 0.754232669350727, + "grad_norm": 0.47178807854652405, + "learning_rate": 2.836700011318173e-05, + "loss": 0.1242, + "step": 32175 + }, + { + "epoch": 0.754256110925532, + "grad_norm": 0.3931722044944763, + "learning_rate": 2.8361860485018954e-05, + "loss": 0.0846, + "step": 32176 + }, + { + "epoch": 0.754279552500337, + "grad_norm": 0.1429535150527954, + "learning_rate": 2.8356721245566754e-05, + "loss": 0.0242, + "step": 32177 + }, + { + "epoch": 0.754302994075142, + "grad_norm": 0.3398716449737549, + "learning_rate": 2.8351582394852983e-05, + "loss": 0.3426, + "step": 32178 + }, + { + "epoch": 0.754326435649947, + "grad_norm": 0.5702504515647888, + "learning_rate": 2.8346443932905552e-05, + "loss": 0.0638, + "step": 32179 + }, + { + "epoch": 0.754349877224752, + "grad_norm": 0.3393533229827881, + "learning_rate": 2.83413058597523e-05, + "loss": 0.0334, + "step": 32180 + }, + { + "epoch": 0.7543733187995569, + "grad_norm": 0.4547434151172638, + "learning_rate": 2.833616817542115e-05, + "loss": 0.0911, + "step": 32181 + }, + { + "epoch": 0.754396760374362, + "grad_norm": 0.27564990520477295, + "learning_rate": 2.833103087993999e-05, + "loss": 0.0477, + "step": 32182 + }, + { + "epoch": 0.7544202019491669, + "grad_norm": 0.2049802839756012, + "learning_rate": 2.832589397333666e-05, + "loss": 0.0476, + "step": 32183 + }, + { + "epoch": 0.754443643523972, + "grad_norm": 0.6325712203979492, + "learning_rate": 2.8320757455639047e-05, + "loss": 0.1196, + "step": 32184 + }, + { + "epoch": 0.7544670850987769, + "grad_norm": 0.5798960328102112, + "learning_rate": 2.831562132687502e-05, + "loss": 0.082, + "step": 32185 + }, + { + "epoch": 0.754490526673582, + "grad_norm": 0.321825236082077, + "learning_rate": 2.8310485587072412e-05, + "loss": 0.0368, + "step": 32186 + }, + { + "epoch": 0.7545139682483869, + "grad_norm": 0.31085219979286194, + "learning_rate": 2.830535023625913e-05, + "loss": 0.0612, + "step": 32187 + }, + { + "epoch": 0.754537409823192, + "grad_norm": 0.5551847815513611, + "learning_rate": 2.830021527446306e-05, + "loss": 0.0411, + "step": 32188 + }, + { + "epoch": 0.7545608513979969, + "grad_norm": 0.38540661334991455, + "learning_rate": 2.829508070171204e-05, + "loss": 0.0338, + "step": 32189 + }, + { + "epoch": 0.7545842929728019, + "grad_norm": 0.4187399744987488, + "learning_rate": 2.8289946518033927e-05, + "loss": 0.0712, + "step": 32190 + }, + { + "epoch": 0.7546077345476069, + "grad_norm": 0.4659649729728699, + "learning_rate": 2.828481272345659e-05, + "loss": 0.05, + "step": 32191 + }, + { + "epoch": 0.7546311761224119, + "grad_norm": 0.19589456915855408, + "learning_rate": 2.8279679318007867e-05, + "loss": 0.0094, + "step": 32192 + }, + { + "epoch": 0.7546546176972169, + "grad_norm": 0.3521682024002075, + "learning_rate": 2.8274546301715586e-05, + "loss": 0.0597, + "step": 32193 + }, + { + "epoch": 0.7546780592720219, + "grad_norm": 0.2528625726699829, + "learning_rate": 2.826941367460767e-05, + "loss": 0.0456, + "step": 32194 + }, + { + "epoch": 0.7547015008468269, + "grad_norm": 0.5650308132171631, + "learning_rate": 2.8264281436711927e-05, + "loss": 0.1121, + "step": 32195 + }, + { + "epoch": 0.7547249424216319, + "grad_norm": 0.10056787729263306, + "learning_rate": 2.825914958805622e-05, + "loss": 0.0216, + "step": 32196 + }, + { + "epoch": 0.7547483839964368, + "grad_norm": 0.7401906251907349, + "learning_rate": 2.825401812866838e-05, + "loss": 0.1405, + "step": 32197 + }, + { + "epoch": 0.7547718255712419, + "grad_norm": 0.14399203658103943, + "learning_rate": 2.8248887058576222e-05, + "loss": 0.0275, + "step": 32198 + }, + { + "epoch": 0.7547952671460468, + "grad_norm": 0.16433008015155792, + "learning_rate": 2.824375637780765e-05, + "loss": 0.0345, + "step": 32199 + }, + { + "epoch": 0.7548187087208519, + "grad_norm": 0.04556908458471298, + "learning_rate": 2.823862608639044e-05, + "loss": 0.0046, + "step": 32200 + }, + { + "epoch": 0.7548421502956568, + "grad_norm": 0.41792720556259155, + "learning_rate": 2.8233496184352505e-05, + "loss": 0.0618, + "step": 32201 + }, + { + "epoch": 0.7548655918704619, + "grad_norm": 0.14103420078754425, + "learning_rate": 2.822836667172163e-05, + "loss": 0.0211, + "step": 32202 + }, + { + "epoch": 0.7548890334452668, + "grad_norm": 0.1008622869849205, + "learning_rate": 2.822323754852565e-05, + "loss": 0.0148, + "step": 32203 + }, + { + "epoch": 0.7549124750200719, + "grad_norm": 0.05754714086651802, + "learning_rate": 2.821810881479241e-05, + "loss": 0.0091, + "step": 32204 + }, + { + "epoch": 0.7549359165948768, + "grad_norm": 0.535920262336731, + "learning_rate": 2.82129804705497e-05, + "loss": 0.6523, + "step": 32205 + }, + { + "epoch": 0.7549593581696818, + "grad_norm": 0.5402105450630188, + "learning_rate": 2.8207852515825406e-05, + "loss": 0.0855, + "step": 32206 + }, + { + "epoch": 0.7549827997444868, + "grad_norm": 0.3208090662956238, + "learning_rate": 2.820272495064732e-05, + "loss": 0.0434, + "step": 32207 + }, + { + "epoch": 0.7550062413192918, + "grad_norm": 0.2803140878677368, + "learning_rate": 2.8197597775043262e-05, + "loss": 0.0637, + "step": 32208 + }, + { + "epoch": 0.7550296828940968, + "grad_norm": 0.32471081614494324, + "learning_rate": 2.8192470989041063e-05, + "loss": 0.292, + "step": 32209 + }, + { + "epoch": 0.7550531244689018, + "grad_norm": 0.5560569167137146, + "learning_rate": 2.8187344592668507e-05, + "loss": 0.061, + "step": 32210 + }, + { + "epoch": 0.7550765660437068, + "grad_norm": 0.3820473551750183, + "learning_rate": 2.8182218585953468e-05, + "loss": 0.0763, + "step": 32211 + }, + { + "epoch": 0.7551000076185118, + "grad_norm": 0.4248409867286682, + "learning_rate": 2.8177092968923702e-05, + "loss": 0.0603, + "step": 32212 + }, + { + "epoch": 0.7551234491933168, + "grad_norm": 0.22258025407791138, + "learning_rate": 2.8171967741607077e-05, + "loss": 0.0279, + "step": 32213 + }, + { + "epoch": 0.7551468907681218, + "grad_norm": 0.42178791761398315, + "learning_rate": 2.8166842904031364e-05, + "loss": 0.0981, + "step": 32214 + }, + { + "epoch": 0.7551703323429269, + "grad_norm": 0.5165613889694214, + "learning_rate": 2.8161718456224394e-05, + "loss": 0.1092, + "step": 32215 + }, + { + "epoch": 0.7551937739177318, + "grad_norm": 0.8131556510925293, + "learning_rate": 2.8156594398213954e-05, + "loss": 0.0955, + "step": 32216 + }, + { + "epoch": 0.7552172154925368, + "grad_norm": 0.20387375354766846, + "learning_rate": 2.8151470730027862e-05, + "loss": 0.0342, + "step": 32217 + }, + { + "epoch": 0.7552406570673418, + "grad_norm": 0.05826921761035919, + "learning_rate": 2.8146347451693865e-05, + "loss": 0.0061, + "step": 32218 + }, + { + "epoch": 0.7552640986421468, + "grad_norm": 0.20668870210647583, + "learning_rate": 2.8141224563239854e-05, + "loss": 0.0124, + "step": 32219 + }, + { + "epoch": 0.7552875402169518, + "grad_norm": 0.564822256565094, + "learning_rate": 2.813610206469357e-05, + "loss": 0.6812, + "step": 32220 + }, + { + "epoch": 0.7553109817917568, + "grad_norm": 0.47595566511154175, + "learning_rate": 2.8130979956082816e-05, + "loss": 0.0442, + "step": 32221 + }, + { + "epoch": 0.7553344233665618, + "grad_norm": 0.4539026618003845, + "learning_rate": 2.8125858237435352e-05, + "loss": 0.4742, + "step": 32222 + }, + { + "epoch": 0.7553578649413668, + "grad_norm": 0.6256133913993835, + "learning_rate": 2.8120736908779045e-05, + "loss": 0.5704, + "step": 32223 + }, + { + "epoch": 0.7553813065161717, + "grad_norm": 0.3741570711135864, + "learning_rate": 2.8115615970141627e-05, + "loss": 0.0955, + "step": 32224 + }, + { + "epoch": 0.7554047480909768, + "grad_norm": 0.0714862048625946, + "learning_rate": 2.8110495421550874e-05, + "loss": 0.0139, + "step": 32225 + }, + { + "epoch": 0.7554281896657817, + "grad_norm": 0.2671331465244293, + "learning_rate": 2.810537526303463e-05, + "loss": 0.042, + "step": 32226 + }, + { + "epoch": 0.7554516312405868, + "grad_norm": 0.4802936017513275, + "learning_rate": 2.810025549462063e-05, + "loss": 0.0385, + "step": 32227 + }, + { + "epoch": 0.7554750728153917, + "grad_norm": 0.46651336550712585, + "learning_rate": 2.8095136116336672e-05, + "loss": 0.0679, + "step": 32228 + }, + { + "epoch": 0.7554985143901968, + "grad_norm": 0.4471438527107239, + "learning_rate": 2.8090017128210522e-05, + "loss": 0.0829, + "step": 32229 + }, + { + "epoch": 0.7555219559650017, + "grad_norm": 0.2844991385936737, + "learning_rate": 2.8084898530269964e-05, + "loss": 0.0494, + "step": 32230 + }, + { + "epoch": 0.7555453975398068, + "grad_norm": 0.46673130989074707, + "learning_rate": 2.8079780322542747e-05, + "loss": 0.0702, + "step": 32231 + }, + { + "epoch": 0.7555688391146117, + "grad_norm": 0.17975950241088867, + "learning_rate": 2.807466250505668e-05, + "loss": 0.0122, + "step": 32232 + }, + { + "epoch": 0.7555922806894168, + "grad_norm": 0.34007883071899414, + "learning_rate": 2.8069545077839522e-05, + "loss": 0.0559, + "step": 32233 + }, + { + "epoch": 0.7556157222642217, + "grad_norm": 0.299244225025177, + "learning_rate": 2.806442804091901e-05, + "loss": 0.0489, + "step": 32234 + }, + { + "epoch": 0.7556391638390267, + "grad_norm": 0.4624220132827759, + "learning_rate": 2.8059311394322962e-05, + "loss": 0.0785, + "step": 32235 + }, + { + "epoch": 0.7556626054138317, + "grad_norm": 0.30986711382865906, + "learning_rate": 2.805419513807911e-05, + "loss": 0.0539, + "step": 32236 + }, + { + "epoch": 0.7556860469886367, + "grad_norm": 0.6167994141578674, + "learning_rate": 2.8049079272215218e-05, + "loss": 0.0997, + "step": 32237 + }, + { + "epoch": 0.7557094885634417, + "grad_norm": 0.6716575622558594, + "learning_rate": 2.804396379675902e-05, + "loss": 0.1056, + "step": 32238 + }, + { + "epoch": 0.7557329301382467, + "grad_norm": 0.3362285792827606, + "learning_rate": 2.803884871173832e-05, + "loss": 0.0484, + "step": 32239 + }, + { + "epoch": 0.7557563717130517, + "grad_norm": 0.626589834690094, + "learning_rate": 2.8033734017180856e-05, + "loss": 0.5822, + "step": 32240 + }, + { + "epoch": 0.7557798132878567, + "grad_norm": 0.3481437861919403, + "learning_rate": 2.802861971311437e-05, + "loss": 0.0183, + "step": 32241 + }, + { + "epoch": 0.7558032548626616, + "grad_norm": 0.4269075095653534, + "learning_rate": 2.8023505799566617e-05, + "loss": 0.0916, + "step": 32242 + }, + { + "epoch": 0.7558266964374667, + "grad_norm": 0.30056703090667725, + "learning_rate": 2.8018392276565318e-05, + "loss": 0.0429, + "step": 32243 + }, + { + "epoch": 0.7558501380122716, + "grad_norm": 0.36030152440071106, + "learning_rate": 2.8013279144138283e-05, + "loss": 0.0468, + "step": 32244 + }, + { + "epoch": 0.7558735795870767, + "grad_norm": 0.09786234050989151, + "learning_rate": 2.8008166402313207e-05, + "loss": 0.0101, + "step": 32245 + }, + { + "epoch": 0.7558970211618816, + "grad_norm": 0.16634297370910645, + "learning_rate": 2.800305405111783e-05, + "loss": 0.0281, + "step": 32246 + }, + { + "epoch": 0.7559204627366867, + "grad_norm": 0.5478066802024841, + "learning_rate": 2.799794209057992e-05, + "loss": 0.0364, + "step": 32247 + }, + { + "epoch": 0.7559439043114916, + "grad_norm": 0.496864378452301, + "learning_rate": 2.799283052072721e-05, + "loss": 0.0588, + "step": 32248 + }, + { + "epoch": 0.7559673458862967, + "grad_norm": 0.2688884735107422, + "learning_rate": 2.7987719341587427e-05, + "loss": 0.0391, + "step": 32249 + }, + { + "epoch": 0.7559907874611016, + "grad_norm": 0.18079422414302826, + "learning_rate": 2.798260855318827e-05, + "loss": 0.0409, + "step": 32250 + }, + { + "epoch": 0.7560142290359066, + "grad_norm": 0.6120263338088989, + "learning_rate": 2.7977498155557535e-05, + "loss": 0.0961, + "step": 32251 + }, + { + "epoch": 0.7560376706107116, + "grad_norm": 0.7674747109413147, + "learning_rate": 2.7972388148722916e-05, + "loss": 0.5913, + "step": 32252 + }, + { + "epoch": 0.7560611121855166, + "grad_norm": 0.5675082206726074, + "learning_rate": 2.796727853271215e-05, + "loss": 0.1562, + "step": 32253 + }, + { + "epoch": 0.7560845537603216, + "grad_norm": 0.6552280783653259, + "learning_rate": 2.7962169307552954e-05, + "loss": 0.7339, + "step": 32254 + }, + { + "epoch": 0.7561079953351266, + "grad_norm": 0.3924207389354706, + "learning_rate": 2.7957060473273057e-05, + "loss": 0.0417, + "step": 32255 + }, + { + "epoch": 0.7561314369099316, + "grad_norm": 0.044021494686603546, + "learning_rate": 2.795195202990013e-05, + "loss": 0.0022, + "step": 32256 + }, + { + "epoch": 0.7561548784847366, + "grad_norm": 0.36982691287994385, + "learning_rate": 2.794684397746198e-05, + "loss": 0.0761, + "step": 32257 + }, + { + "epoch": 0.7561783200595416, + "grad_norm": 0.5784990787506104, + "learning_rate": 2.7941736315986277e-05, + "loss": 0.6044, + "step": 32258 + }, + { + "epoch": 0.7562017616343466, + "grad_norm": 0.23083443939685822, + "learning_rate": 2.7936629045500718e-05, + "loss": 0.0461, + "step": 32259 + }, + { + "epoch": 0.7562252032091515, + "grad_norm": 0.15278398990631104, + "learning_rate": 2.7931522166033052e-05, + "loss": 0.0112, + "step": 32260 + }, + { + "epoch": 0.7562486447839566, + "grad_norm": 0.3761284053325653, + "learning_rate": 2.7926415677610984e-05, + "loss": 0.0484, + "step": 32261 + }, + { + "epoch": 0.7562720863587615, + "grad_norm": 0.09136633574962616, + "learning_rate": 2.7921309580262202e-05, + "loss": 0.005, + "step": 32262 + }, + { + "epoch": 0.7562955279335666, + "grad_norm": 0.5522545576095581, + "learning_rate": 2.7916203874014403e-05, + "loss": 0.1173, + "step": 32263 + }, + { + "epoch": 0.7563189695083715, + "grad_norm": 0.36213573813438416, + "learning_rate": 2.7911098558895333e-05, + "loss": 0.05, + "step": 32264 + }, + { + "epoch": 0.7563424110831766, + "grad_norm": 0.3899804949760437, + "learning_rate": 2.790599363493266e-05, + "loss": 0.0756, + "step": 32265 + }, + { + "epoch": 0.7563658526579815, + "grad_norm": 0.574831485748291, + "learning_rate": 2.7900889102154104e-05, + "loss": 0.6424, + "step": 32266 + }, + { + "epoch": 0.7563892942327866, + "grad_norm": 0.15004533529281616, + "learning_rate": 2.7895784960587344e-05, + "loss": 0.0127, + "step": 32267 + }, + { + "epoch": 0.7564127358075916, + "grad_norm": 0.46212685108184814, + "learning_rate": 2.7890681210260085e-05, + "loss": 0.0756, + "step": 32268 + }, + { + "epoch": 0.7564361773823965, + "grad_norm": 0.41734594106674194, + "learning_rate": 2.7885577851199983e-05, + "loss": 0.0679, + "step": 32269 + }, + { + "epoch": 0.7564596189572016, + "grad_norm": 0.6649256348609924, + "learning_rate": 2.7880474883434805e-05, + "loss": 0.1278, + "step": 32270 + }, + { + "epoch": 0.7564830605320065, + "grad_norm": 0.2118970900774002, + "learning_rate": 2.7875372306992154e-05, + "loss": 0.0385, + "step": 32271 + }, + { + "epoch": 0.7565065021068116, + "grad_norm": 0.391120582818985, + "learning_rate": 2.78702701218998e-05, + "loss": 0.0698, + "step": 32272 + }, + { + "epoch": 0.7565299436816165, + "grad_norm": 0.28556162118911743, + "learning_rate": 2.7865168328185387e-05, + "loss": 0.0268, + "step": 32273 + }, + { + "epoch": 0.7565533852564216, + "grad_norm": 0.22851034998893738, + "learning_rate": 2.7860066925876605e-05, + "loss": 0.0388, + "step": 32274 + }, + { + "epoch": 0.7565768268312265, + "grad_norm": 0.5239654183387756, + "learning_rate": 2.785496591500112e-05, + "loss": 0.1096, + "step": 32275 + }, + { + "epoch": 0.7566002684060316, + "grad_norm": 0.3652835190296173, + "learning_rate": 2.7849865295586587e-05, + "loss": 0.0581, + "step": 32276 + }, + { + "epoch": 0.7566237099808365, + "grad_norm": 0.11591716855764389, + "learning_rate": 2.784476506766075e-05, + "loss": 0.0263, + "step": 32277 + }, + { + "epoch": 0.7566471515556416, + "grad_norm": 0.6333203911781311, + "learning_rate": 2.7839665231251234e-05, + "loss": 0.0977, + "step": 32278 + }, + { + "epoch": 0.7566705931304465, + "grad_norm": 0.3078766465187073, + "learning_rate": 2.7834565786385734e-05, + "loss": 0.0405, + "step": 32279 + }, + { + "epoch": 0.7566940347052515, + "grad_norm": 0.3996046185493469, + "learning_rate": 2.7829466733091914e-05, + "loss": 0.0354, + "step": 32280 + }, + { + "epoch": 0.7567174762800565, + "grad_norm": 0.4434506893157959, + "learning_rate": 2.78243680713974e-05, + "loss": 0.0968, + "step": 32281 + }, + { + "epoch": 0.7567409178548615, + "grad_norm": 0.3790617287158966, + "learning_rate": 2.7819269801329917e-05, + "loss": 0.0338, + "step": 32282 + }, + { + "epoch": 0.7567643594296665, + "grad_norm": 0.6043798327445984, + "learning_rate": 2.7814171922917087e-05, + "loss": 0.4777, + "step": 32283 + }, + { + "epoch": 0.7567878010044715, + "grad_norm": 0.09333351999521255, + "learning_rate": 2.780907443618661e-05, + "loss": 0.0077, + "step": 32284 + }, + { + "epoch": 0.7568112425792765, + "grad_norm": 0.1897062361240387, + "learning_rate": 2.7803977341166132e-05, + "loss": 0.0407, + "step": 32285 + }, + { + "epoch": 0.7568346841540815, + "grad_norm": 0.22782182693481445, + "learning_rate": 2.7798880637883305e-05, + "loss": 0.0451, + "step": 32286 + }, + { + "epoch": 0.7568581257288864, + "grad_norm": 0.5079787969589233, + "learning_rate": 2.779378432636578e-05, + "loss": 0.1226, + "step": 32287 + }, + { + "epoch": 0.7568815673036915, + "grad_norm": 0.101408950984478, + "learning_rate": 2.7788688406641183e-05, + "loss": 0.0137, + "step": 32288 + }, + { + "epoch": 0.7569050088784964, + "grad_norm": 0.41539546847343445, + "learning_rate": 2.778359287873722e-05, + "loss": 0.3864, + "step": 32289 + }, + { + "epoch": 0.7569284504533015, + "grad_norm": 0.44613730907440186, + "learning_rate": 2.7778497742681507e-05, + "loss": 0.0839, + "step": 32290 + }, + { + "epoch": 0.7569518920281064, + "grad_norm": 0.12167032063007355, + "learning_rate": 2.7773402998501706e-05, + "loss": 0.0089, + "step": 32291 + }, + { + "epoch": 0.7569753336029115, + "grad_norm": 0.1025753766298294, + "learning_rate": 2.776830864622545e-05, + "loss": 0.0186, + "step": 32292 + }, + { + "epoch": 0.7569987751777164, + "grad_norm": 0.20314526557922363, + "learning_rate": 2.776321468588038e-05, + "loss": 0.0338, + "step": 32293 + }, + { + "epoch": 0.7570222167525215, + "grad_norm": 0.45323118567466736, + "learning_rate": 2.7758121117494107e-05, + "loss": 0.0608, + "step": 32294 + }, + { + "epoch": 0.7570456583273264, + "grad_norm": 0.24623194336891174, + "learning_rate": 2.77530279410943e-05, + "loss": 0.0528, + "step": 32295 + }, + { + "epoch": 0.7570690999021314, + "grad_norm": 0.3261827528476715, + "learning_rate": 2.7747935156708625e-05, + "loss": 0.071, + "step": 32296 + }, + { + "epoch": 0.7570925414769364, + "grad_norm": 0.24639195203781128, + "learning_rate": 2.7742842764364686e-05, + "loss": 0.0333, + "step": 32297 + }, + { + "epoch": 0.7571159830517414, + "grad_norm": 0.09339816868305206, + "learning_rate": 2.773775076409011e-05, + "loss": 0.0127, + "step": 32298 + }, + { + "epoch": 0.7571394246265464, + "grad_norm": 0.09287728369235992, + "learning_rate": 2.7732659155912532e-05, + "loss": 0.0131, + "step": 32299 + }, + { + "epoch": 0.7571628662013514, + "grad_norm": 0.493102490901947, + "learning_rate": 2.7727567939859578e-05, + "loss": 0.0337, + "step": 32300 + }, + { + "epoch": 0.7571863077761564, + "grad_norm": 0.2809796929359436, + "learning_rate": 2.7722477115958846e-05, + "loss": 0.0361, + "step": 32301 + }, + { + "epoch": 0.7572097493509614, + "grad_norm": 0.22788631916046143, + "learning_rate": 2.771738668423801e-05, + "loss": 0.0326, + "step": 32302 + }, + { + "epoch": 0.7572331909257664, + "grad_norm": 0.5273977518081665, + "learning_rate": 2.7712296644724665e-05, + "loss": 0.0996, + "step": 32303 + }, + { + "epoch": 0.7572566325005714, + "grad_norm": 0.37330344319343567, + "learning_rate": 2.7707206997446443e-05, + "loss": 0.044, + "step": 32304 + }, + { + "epoch": 0.7572800740753763, + "grad_norm": 0.5338608622550964, + "learning_rate": 2.7702117742430945e-05, + "loss": 0.0593, + "step": 32305 + }, + { + "epoch": 0.7573035156501814, + "grad_norm": 0.20421014726161957, + "learning_rate": 2.769702887970578e-05, + "loss": 0.038, + "step": 32306 + }, + { + "epoch": 0.7573269572249863, + "grad_norm": 0.4152624309062958, + "learning_rate": 2.7691940409298546e-05, + "loss": 0.0414, + "step": 32307 + }, + { + "epoch": 0.7573503987997914, + "grad_norm": 0.22266939282417297, + "learning_rate": 2.768685233123687e-05, + "loss": 0.0426, + "step": 32308 + }, + { + "epoch": 0.7573738403745963, + "grad_norm": 0.4733195900917053, + "learning_rate": 2.76817646455484e-05, + "loss": 0.0896, + "step": 32309 + }, + { + "epoch": 0.7573972819494014, + "grad_norm": 0.16498368978500366, + "learning_rate": 2.7676677352260717e-05, + "loss": 0.0291, + "step": 32310 + }, + { + "epoch": 0.7574207235242063, + "grad_norm": 0.3226242959499359, + "learning_rate": 2.767159045140141e-05, + "loss": 0.0947, + "step": 32311 + }, + { + "epoch": 0.7574441650990114, + "grad_norm": 0.10486482083797455, + "learning_rate": 2.766650394299808e-05, + "loss": 0.0168, + "step": 32312 + }, + { + "epoch": 0.7574676066738163, + "grad_norm": 0.4205012917518616, + "learning_rate": 2.7661417827078307e-05, + "loss": 0.0438, + "step": 32313 + }, + { + "epoch": 0.7574910482486213, + "grad_norm": 0.4672446846961975, + "learning_rate": 2.7656332103669747e-05, + "loss": 0.0722, + "step": 32314 + }, + { + "epoch": 0.7575144898234263, + "grad_norm": 0.23635047674179077, + "learning_rate": 2.7651246772799965e-05, + "loss": 0.032, + "step": 32315 + }, + { + "epoch": 0.7575379313982313, + "grad_norm": 0.5456935167312622, + "learning_rate": 2.7646161834496554e-05, + "loss": 0.6021, + "step": 32316 + }, + { + "epoch": 0.7575613729730363, + "grad_norm": 0.13426326215267181, + "learning_rate": 2.7641077288787097e-05, + "loss": 0.0143, + "step": 32317 + }, + { + "epoch": 0.7575848145478413, + "grad_norm": 0.31882449984550476, + "learning_rate": 2.763599313569919e-05, + "loss": 0.0327, + "step": 32318 + }, + { + "epoch": 0.7576082561226464, + "grad_norm": 0.317935585975647, + "learning_rate": 2.7630909375260395e-05, + "loss": 0.0592, + "step": 32319 + }, + { + "epoch": 0.7576316976974513, + "grad_norm": 0.19233131408691406, + "learning_rate": 2.7625826007498312e-05, + "loss": 0.035, + "step": 32320 + }, + { + "epoch": 0.7576551392722564, + "grad_norm": 0.4868326485157013, + "learning_rate": 2.7620743032440565e-05, + "loss": 0.58, + "step": 32321 + }, + { + "epoch": 0.7576785808470613, + "grad_norm": 0.4768809676170349, + "learning_rate": 2.76156604501147e-05, + "loss": 0.3413, + "step": 32322 + }, + { + "epoch": 0.7577020224218664, + "grad_norm": 0.333034873008728, + "learning_rate": 2.7610578260548303e-05, + "loss": 0.078, + "step": 32323 + }, + { + "epoch": 0.7577254639966713, + "grad_norm": 0.3296387195587158, + "learning_rate": 2.7605496463768944e-05, + "loss": 0.078, + "step": 32324 + }, + { + "epoch": 0.7577489055714763, + "grad_norm": 0.5954675674438477, + "learning_rate": 2.7600415059804187e-05, + "loss": 0.1545, + "step": 32325 + }, + { + "epoch": 0.7577723471462813, + "grad_norm": 0.06970236450433731, + "learning_rate": 2.759533404868159e-05, + "loss": 0.0078, + "step": 32326 + }, + { + "epoch": 0.7577957887210863, + "grad_norm": 0.18129627406597137, + "learning_rate": 2.759025343042878e-05, + "loss": 0.0472, + "step": 32327 + }, + { + "epoch": 0.7578192302958913, + "grad_norm": 0.5778946876525879, + "learning_rate": 2.7585173205073288e-05, + "loss": 0.6198, + "step": 32328 + }, + { + "epoch": 0.7578426718706963, + "grad_norm": 0.3718940019607544, + "learning_rate": 2.758009337264268e-05, + "loss": 0.0866, + "step": 32329 + }, + { + "epoch": 0.7578661134455013, + "grad_norm": 0.615795373916626, + "learning_rate": 2.7575013933164517e-05, + "loss": 0.0664, + "step": 32330 + }, + { + "epoch": 0.7578895550203063, + "grad_norm": 0.10656807571649551, + "learning_rate": 2.7569934886666372e-05, + "loss": 0.0182, + "step": 32331 + }, + { + "epoch": 0.7579129965951112, + "grad_norm": 0.2730749845504761, + "learning_rate": 2.7564856233175764e-05, + "loss": 0.0484, + "step": 32332 + }, + { + "epoch": 0.7579364381699163, + "grad_norm": 0.3121178448200226, + "learning_rate": 2.755977797272029e-05, + "loss": 0.0758, + "step": 32333 + }, + { + "epoch": 0.7579598797447212, + "grad_norm": 0.19122672080993652, + "learning_rate": 2.7554700105327513e-05, + "loss": 0.0247, + "step": 32334 + }, + { + "epoch": 0.7579833213195263, + "grad_norm": 0.24573390185832977, + "learning_rate": 2.7549622631024985e-05, + "loss": 0.0286, + "step": 32335 + }, + { + "epoch": 0.7580067628943312, + "grad_norm": 0.399326890707016, + "learning_rate": 2.754454554984024e-05, + "loss": 0.086, + "step": 32336 + }, + { + "epoch": 0.7580302044691363, + "grad_norm": 0.5920359492301941, + "learning_rate": 2.753946886180082e-05, + "loss": 0.0952, + "step": 32337 + }, + { + "epoch": 0.7580536460439412, + "grad_norm": 0.3736629784107208, + "learning_rate": 2.7534392566934286e-05, + "loss": 0.0476, + "step": 32338 + }, + { + "epoch": 0.7580770876187463, + "grad_norm": 0.33376169204711914, + "learning_rate": 2.7529316665268135e-05, + "loss": 0.0499, + "step": 32339 + }, + { + "epoch": 0.7581005291935512, + "grad_norm": 0.300292432308197, + "learning_rate": 2.752424115682999e-05, + "loss": 0.0388, + "step": 32340 + }, + { + "epoch": 0.7581239707683562, + "grad_norm": 0.21565937995910645, + "learning_rate": 2.751916604164735e-05, + "loss": 0.0501, + "step": 32341 + }, + { + "epoch": 0.7581474123431612, + "grad_norm": 0.34891748428344727, + "learning_rate": 2.7514091319747748e-05, + "loss": 0.0509, + "step": 32342 + }, + { + "epoch": 0.7581708539179662, + "grad_norm": 1.0952844619750977, + "learning_rate": 2.750901699115873e-05, + "loss": 0.0759, + "step": 32343 + }, + { + "epoch": 0.7581942954927712, + "grad_norm": 0.48635491728782654, + "learning_rate": 2.7503943055907787e-05, + "loss": 0.0695, + "step": 32344 + }, + { + "epoch": 0.7582177370675762, + "grad_norm": 0.3681378662586212, + "learning_rate": 2.7498869514022517e-05, + "loss": 0.0554, + "step": 32345 + }, + { + "epoch": 0.7582411786423812, + "grad_norm": 0.3182293474674225, + "learning_rate": 2.74937963655304e-05, + "loss": 0.047, + "step": 32346 + }, + { + "epoch": 0.7582646202171862, + "grad_norm": 0.5258598923683167, + "learning_rate": 2.748872361045901e-05, + "loss": 0.0561, + "step": 32347 + }, + { + "epoch": 0.7582880617919912, + "grad_norm": 0.5189141631126404, + "learning_rate": 2.748365124883584e-05, + "loss": 0.6632, + "step": 32348 + }, + { + "epoch": 0.7583115033667962, + "grad_norm": 0.5466301441192627, + "learning_rate": 2.747857928068842e-05, + "loss": 0.1154, + "step": 32349 + }, + { + "epoch": 0.7583349449416011, + "grad_norm": 0.5340304374694824, + "learning_rate": 2.7473507706044265e-05, + "loss": 0.0881, + "step": 32350 + }, + { + "epoch": 0.7583583865164062, + "grad_norm": 0.2621305584907532, + "learning_rate": 2.7468436524930862e-05, + "loss": 0.0244, + "step": 32351 + }, + { + "epoch": 0.7583818280912111, + "grad_norm": 0.19067974388599396, + "learning_rate": 2.74633657373758e-05, + "loss": 0.0131, + "step": 32352 + }, + { + "epoch": 0.7584052696660162, + "grad_norm": 0.4451001286506653, + "learning_rate": 2.7458295343406548e-05, + "loss": 0.1342, + "step": 32353 + }, + { + "epoch": 0.7584287112408211, + "grad_norm": 0.6840852499008179, + "learning_rate": 2.7453225343050636e-05, + "loss": 0.0989, + "step": 32354 + }, + { + "epoch": 0.7584521528156262, + "grad_norm": 0.3491012156009674, + "learning_rate": 2.7448155736335556e-05, + "loss": 0.4126, + "step": 32355 + }, + { + "epoch": 0.7584755943904311, + "grad_norm": 0.5508949756622314, + "learning_rate": 2.7443086523288796e-05, + "loss": 0.0785, + "step": 32356 + }, + { + "epoch": 0.7584990359652362, + "grad_norm": 0.20158223807811737, + "learning_rate": 2.7438017703937912e-05, + "loss": 0.0302, + "step": 32357 + }, + { + "epoch": 0.7585224775400411, + "grad_norm": 0.2233533263206482, + "learning_rate": 2.743294927831036e-05, + "loss": 0.0382, + "step": 32358 + }, + { + "epoch": 0.7585459191148461, + "grad_norm": 0.4755454957485199, + "learning_rate": 2.7427881246433696e-05, + "loss": 0.0512, + "step": 32359 + }, + { + "epoch": 0.7585693606896511, + "grad_norm": 0.6738930940628052, + "learning_rate": 2.7422813608335383e-05, + "loss": 0.0433, + "step": 32360 + }, + { + "epoch": 0.7585928022644561, + "grad_norm": 0.36197429895401, + "learning_rate": 2.741774636404293e-05, + "loss": 0.0675, + "step": 32361 + }, + { + "epoch": 0.7586162438392611, + "grad_norm": 0.16859950125217438, + "learning_rate": 2.7412679513583827e-05, + "loss": 0.0365, + "step": 32362 + }, + { + "epoch": 0.7586396854140661, + "grad_norm": 0.2904989421367645, + "learning_rate": 2.7407613056985558e-05, + "loss": 0.0111, + "step": 32363 + }, + { + "epoch": 0.7586631269888711, + "grad_norm": 0.39703482389450073, + "learning_rate": 2.74025469942756e-05, + "loss": 0.0346, + "step": 32364 + }, + { + "epoch": 0.7586865685636761, + "grad_norm": 0.4067385792732239, + "learning_rate": 2.7397481325481488e-05, + "loss": 0.0349, + "step": 32365 + }, + { + "epoch": 0.758710010138481, + "grad_norm": 0.21953445672988892, + "learning_rate": 2.7392416050630686e-05, + "loss": 0.0256, + "step": 32366 + }, + { + "epoch": 0.7587334517132861, + "grad_norm": 0.12138979136943817, + "learning_rate": 2.7387351169750685e-05, + "loss": 0.0065, + "step": 32367 + }, + { + "epoch": 0.758756893288091, + "grad_norm": 0.3610655665397644, + "learning_rate": 2.7382286682868918e-05, + "loss": 0.0687, + "step": 32368 + }, + { + "epoch": 0.7587803348628961, + "grad_norm": 0.4065236747264862, + "learning_rate": 2.737722259001293e-05, + "loss": 0.0924, + "step": 32369 + }, + { + "epoch": 0.7588037764377011, + "grad_norm": 0.4557937681674957, + "learning_rate": 2.7372158891210188e-05, + "loss": 0.1109, + "step": 32370 + }, + { + "epoch": 0.7588272180125061, + "grad_norm": 0.748550295829773, + "learning_rate": 2.736709558648811e-05, + "loss": 0.125, + "step": 32371 + }, + { + "epoch": 0.7588506595873111, + "grad_norm": 0.4279494285583496, + "learning_rate": 2.736203267587425e-05, + "loss": 0.0377, + "step": 32372 + }, + { + "epoch": 0.7588741011621161, + "grad_norm": 0.1548629105091095, + "learning_rate": 2.7356970159396046e-05, + "loss": 0.0217, + "step": 32373 + }, + { + "epoch": 0.7588975427369211, + "grad_norm": 0.30900970101356506, + "learning_rate": 2.7351908037080965e-05, + "loss": 0.0694, + "step": 32374 + }, + { + "epoch": 0.758920984311726, + "grad_norm": 0.10466300696134567, + "learning_rate": 2.7346846308956477e-05, + "loss": 0.014, + "step": 32375 + }, + { + "epoch": 0.7589444258865311, + "grad_norm": 0.3507434129714966, + "learning_rate": 2.7341784975050033e-05, + "loss": 0.0577, + "step": 32376 + }, + { + "epoch": 0.758967867461336, + "grad_norm": 0.5279468297958374, + "learning_rate": 2.733672403538908e-05, + "loss": 0.1187, + "step": 32377 + }, + { + "epoch": 0.7589913090361411, + "grad_norm": 0.17545168101787567, + "learning_rate": 2.7331663490001137e-05, + "loss": 0.0275, + "step": 32378 + }, + { + "epoch": 0.759014750610946, + "grad_norm": 0.49431952834129333, + "learning_rate": 2.7326603338913625e-05, + "loss": 0.0655, + "step": 32379 + }, + { + "epoch": 0.7590381921857511, + "grad_norm": 0.39818519353866577, + "learning_rate": 2.7321543582153985e-05, + "loss": 0.0831, + "step": 32380 + }, + { + "epoch": 0.759061633760556, + "grad_norm": 0.5594670176506042, + "learning_rate": 2.7316484219749716e-05, + "loss": 0.0854, + "step": 32381 + }, + { + "epoch": 0.7590850753353611, + "grad_norm": 0.3499971926212311, + "learning_rate": 2.731142525172826e-05, + "loss": 0.0628, + "step": 32382 + }, + { + "epoch": 0.759108516910166, + "grad_norm": 0.4051789939403534, + "learning_rate": 2.7306366678117044e-05, + "loss": 0.0568, + "step": 32383 + }, + { + "epoch": 0.7591319584849711, + "grad_norm": 0.4503076672554016, + "learning_rate": 2.73013084989435e-05, + "loss": 0.0923, + "step": 32384 + }, + { + "epoch": 0.759155400059776, + "grad_norm": 0.5455703139305115, + "learning_rate": 2.7296250714235127e-05, + "loss": 0.5999, + "step": 32385 + }, + { + "epoch": 0.759178841634581, + "grad_norm": 0.23206375539302826, + "learning_rate": 2.729119332401935e-05, + "loss": 0.0355, + "step": 32386 + }, + { + "epoch": 0.759202283209386, + "grad_norm": 0.4521082937717438, + "learning_rate": 2.7286136328323585e-05, + "loss": 0.0578, + "step": 32387 + }, + { + "epoch": 0.759225724784191, + "grad_norm": 0.34451812505722046, + "learning_rate": 2.7281079727175306e-05, + "loss": 0.0586, + "step": 32388 + }, + { + "epoch": 0.759249166358996, + "grad_norm": 0.3490321636199951, + "learning_rate": 2.727602352060189e-05, + "loss": 0.0789, + "step": 32389 + }, + { + "epoch": 0.759272607933801, + "grad_norm": 0.237986758351326, + "learning_rate": 2.7270967708630857e-05, + "loss": 0.023, + "step": 32390 + }, + { + "epoch": 0.759296049508606, + "grad_norm": 0.4592054784297943, + "learning_rate": 2.7265912291289598e-05, + "loss": 0.4333, + "step": 32391 + }, + { + "epoch": 0.759319491083411, + "grad_norm": 0.8702009320259094, + "learning_rate": 2.72608572686055e-05, + "loss": 0.1046, + "step": 32392 + }, + { + "epoch": 0.759342932658216, + "grad_norm": 0.21403473615646362, + "learning_rate": 2.7255802640606078e-05, + "loss": 0.0226, + "step": 32393 + }, + { + "epoch": 0.759366374233021, + "grad_norm": 0.27044981718063354, + "learning_rate": 2.7250748407318715e-05, + "loss": 0.0492, + "step": 32394 + }, + { + "epoch": 0.7593898158078259, + "grad_norm": 0.20960962772369385, + "learning_rate": 2.724569456877084e-05, + "loss": 0.019, + "step": 32395 + }, + { + "epoch": 0.759413257382631, + "grad_norm": 0.533394992351532, + "learning_rate": 2.724064112498983e-05, + "loss": 0.5157, + "step": 32396 + }, + { + "epoch": 0.7594366989574359, + "grad_norm": 0.6949336528778076, + "learning_rate": 2.7235588076003183e-05, + "loss": 0.6375, + "step": 32397 + }, + { + "epoch": 0.759460140532241, + "grad_norm": 0.3439003527164459, + "learning_rate": 2.7230535421838287e-05, + "loss": 0.0514, + "step": 32398 + }, + { + "epoch": 0.7594835821070459, + "grad_norm": 0.47883692383766174, + "learning_rate": 2.722548316252255e-05, + "loss": 0.1032, + "step": 32399 + }, + { + "epoch": 0.759507023681851, + "grad_norm": 0.5947659015655518, + "learning_rate": 2.7220431298083382e-05, + "loss": 0.1248, + "step": 32400 + }, + { + "epoch": 0.7595304652566559, + "grad_norm": 0.3564891517162323, + "learning_rate": 2.7215379828548204e-05, + "loss": 0.0496, + "step": 32401 + }, + { + "epoch": 0.759553906831461, + "grad_norm": 0.4029134511947632, + "learning_rate": 2.721032875394439e-05, + "loss": 0.075, + "step": 32402 + }, + { + "epoch": 0.7595773484062659, + "grad_norm": 0.2989536225795746, + "learning_rate": 2.72052780742994e-05, + "loss": 0.0508, + "step": 32403 + }, + { + "epoch": 0.759600789981071, + "grad_norm": 0.9290332794189453, + "learning_rate": 2.7200227789640632e-05, + "loss": 0.1453, + "step": 32404 + }, + { + "epoch": 0.7596242315558759, + "grad_norm": 0.34709805250167847, + "learning_rate": 2.7195177899995438e-05, + "loss": 0.0642, + "step": 32405 + }, + { + "epoch": 0.7596476731306809, + "grad_norm": 0.12815193831920624, + "learning_rate": 2.7190128405391292e-05, + "loss": 0.017, + "step": 32406 + }, + { + "epoch": 0.7596711147054859, + "grad_norm": 0.5217496752738953, + "learning_rate": 2.7185079305855555e-05, + "loss": 0.485, + "step": 32407 + }, + { + "epoch": 0.7596945562802909, + "grad_norm": 0.488912433385849, + "learning_rate": 2.7180030601415618e-05, + "loss": 0.0399, + "step": 32408 + }, + { + "epoch": 0.7597179978550959, + "grad_norm": 0.5131219029426575, + "learning_rate": 2.7174982292098862e-05, + "loss": 0.0898, + "step": 32409 + }, + { + "epoch": 0.7597414394299009, + "grad_norm": 0.4361029267311096, + "learning_rate": 2.7169934377932716e-05, + "loss": 0.0573, + "step": 32410 + }, + { + "epoch": 0.7597648810047058, + "grad_norm": 0.6757389307022095, + "learning_rate": 2.716488685894457e-05, + "loss": 0.0881, + "step": 32411 + }, + { + "epoch": 0.7597883225795109, + "grad_norm": 0.2370813637971878, + "learning_rate": 2.7159839735161786e-05, + "loss": 0.062, + "step": 32412 + }, + { + "epoch": 0.7598117641543158, + "grad_norm": 0.080764539539814, + "learning_rate": 2.7154793006611757e-05, + "loss": 0.0186, + "step": 32413 + }, + { + "epoch": 0.7598352057291209, + "grad_norm": 0.508858859539032, + "learning_rate": 2.7149746673321873e-05, + "loss": 0.1339, + "step": 32414 + }, + { + "epoch": 0.7598586473039258, + "grad_norm": 0.1626778393983841, + "learning_rate": 2.7144700735319484e-05, + "loss": 0.024, + "step": 32415 + }, + { + "epoch": 0.7598820888787309, + "grad_norm": 0.5866027474403381, + "learning_rate": 2.713965519263203e-05, + "loss": 0.1434, + "step": 32416 + }, + { + "epoch": 0.7599055304535358, + "grad_norm": 0.453233003616333, + "learning_rate": 2.7134610045286823e-05, + "loss": 0.4992, + "step": 32417 + }, + { + "epoch": 0.7599289720283409, + "grad_norm": 0.45854753255844116, + "learning_rate": 2.71295652933113e-05, + "loss": 0.1215, + "step": 32418 + }, + { + "epoch": 0.7599524136031458, + "grad_norm": 0.25332510471343994, + "learning_rate": 2.71245209367328e-05, + "loss": 0.0415, + "step": 32419 + }, + { + "epoch": 0.7599758551779509, + "grad_norm": 0.47867846488952637, + "learning_rate": 2.7119476975578706e-05, + "loss": 0.1185, + "step": 32420 + }, + { + "epoch": 0.7599992967527559, + "grad_norm": 0.1876935362815857, + "learning_rate": 2.7114433409876383e-05, + "loss": 0.0184, + "step": 32421 + }, + { + "epoch": 0.7600227383275608, + "grad_norm": 0.449190229177475, + "learning_rate": 2.7109390239653154e-05, + "loss": 0.0539, + "step": 32422 + }, + { + "epoch": 0.7600461799023659, + "grad_norm": 0.22244402766227722, + "learning_rate": 2.7104347464936452e-05, + "loss": 0.0297, + "step": 32423 + }, + { + "epoch": 0.7600696214771708, + "grad_norm": 0.5819535255432129, + "learning_rate": 2.7099305085753612e-05, + "loss": 0.1242, + "step": 32424 + }, + { + "epoch": 0.7600930630519759, + "grad_norm": 0.33338239789009094, + "learning_rate": 2.7094263102132002e-05, + "loss": 0.0482, + "step": 32425 + }, + { + "epoch": 0.7601165046267808, + "grad_norm": 0.15871945023536682, + "learning_rate": 2.7089221514098962e-05, + "loss": 0.0147, + "step": 32426 + }, + { + "epoch": 0.7601399462015859, + "grad_norm": 0.5111619234085083, + "learning_rate": 2.708418032168183e-05, + "loss": 0.1016, + "step": 32427 + }, + { + "epoch": 0.7601633877763908, + "grad_norm": 0.0853787511587143, + "learning_rate": 2.7079139524908004e-05, + "loss": 0.0152, + "step": 32428 + }, + { + "epoch": 0.7601868293511959, + "grad_norm": 0.5114628076553345, + "learning_rate": 2.7074099123804798e-05, + "loss": 0.0656, + "step": 32429 + }, + { + "epoch": 0.7602102709260008, + "grad_norm": 0.3266765773296356, + "learning_rate": 2.706905911839961e-05, + "loss": 0.0586, + "step": 32430 + }, + { + "epoch": 0.7602337125008058, + "grad_norm": 0.07727673649787903, + "learning_rate": 2.706401950871975e-05, + "loss": 0.0081, + "step": 32431 + }, + { + "epoch": 0.7602571540756108, + "grad_norm": 0.3404232859611511, + "learning_rate": 2.705898029479258e-05, + "loss": 0.0982, + "step": 32432 + }, + { + "epoch": 0.7602805956504158, + "grad_norm": 0.14237616956233978, + "learning_rate": 2.705394147664544e-05, + "loss": 0.0273, + "step": 32433 + }, + { + "epoch": 0.7603040372252208, + "grad_norm": 0.21028846502304077, + "learning_rate": 2.704890305430562e-05, + "loss": 0.0456, + "step": 32434 + }, + { + "epoch": 0.7603274788000258, + "grad_norm": 0.2779904305934906, + "learning_rate": 2.704386502780054e-05, + "loss": 0.0448, + "step": 32435 + }, + { + "epoch": 0.7603509203748308, + "grad_norm": 0.5265330076217651, + "learning_rate": 2.7038827397157496e-05, + "loss": 0.0522, + "step": 32436 + }, + { + "epoch": 0.7603743619496358, + "grad_norm": 0.23711784183979034, + "learning_rate": 2.7033790162403837e-05, + "loss": 0.0432, + "step": 32437 + }, + { + "epoch": 0.7603978035244408, + "grad_norm": 0.3287695348262787, + "learning_rate": 2.702875332356687e-05, + "loss": 0.0377, + "step": 32438 + }, + { + "epoch": 0.7604212450992458, + "grad_norm": 0.3611995279788971, + "learning_rate": 2.7023716880673942e-05, + "loss": 0.0782, + "step": 32439 + }, + { + "epoch": 0.7604446866740507, + "grad_norm": 0.13393665850162506, + "learning_rate": 2.7018680833752353e-05, + "loss": 0.0353, + "step": 32440 + }, + { + "epoch": 0.7604681282488558, + "grad_norm": 0.3047625720500946, + "learning_rate": 2.701364518282946e-05, + "loss": 0.0371, + "step": 32441 + }, + { + "epoch": 0.7604915698236607, + "grad_norm": 0.2532539963722229, + "learning_rate": 2.7008609927932605e-05, + "loss": 0.0487, + "step": 32442 + }, + { + "epoch": 0.7605150113984658, + "grad_norm": 0.24237583577632904, + "learning_rate": 2.700357506908908e-05, + "loss": 0.2241, + "step": 32443 + }, + { + "epoch": 0.7605384529732707, + "grad_norm": 0.2833719253540039, + "learning_rate": 2.6998540606326206e-05, + "loss": 0.0433, + "step": 32444 + }, + { + "epoch": 0.7605618945480758, + "grad_norm": 0.29082968831062317, + "learning_rate": 2.699350653967131e-05, + "loss": 0.4266, + "step": 32445 + }, + { + "epoch": 0.7605853361228807, + "grad_norm": 0.6680054068565369, + "learning_rate": 2.6988472869151703e-05, + "loss": 0.1236, + "step": 32446 + }, + { + "epoch": 0.7606087776976858, + "grad_norm": 0.3738376796245575, + "learning_rate": 2.6983439594794657e-05, + "loss": 0.0462, + "step": 32447 + }, + { + "epoch": 0.7606322192724907, + "grad_norm": 0.510192334651947, + "learning_rate": 2.697840671662756e-05, + "loss": 0.4995, + "step": 32448 + }, + { + "epoch": 0.7606556608472957, + "grad_norm": 0.15282398462295532, + "learning_rate": 2.697337423467767e-05, + "loss": 0.0117, + "step": 32449 + }, + { + "epoch": 0.7606791024221007, + "grad_norm": 0.4892330765724182, + "learning_rate": 2.6968342148972315e-05, + "loss": 0.0566, + "step": 32450 + }, + { + "epoch": 0.7607025439969057, + "grad_norm": 0.6243522763252258, + "learning_rate": 2.6963310459538783e-05, + "loss": 0.1108, + "step": 32451 + }, + { + "epoch": 0.7607259855717107, + "grad_norm": 0.11303175240755081, + "learning_rate": 2.6958279166404388e-05, + "loss": 0.0109, + "step": 32452 + }, + { + "epoch": 0.7607494271465157, + "grad_norm": 0.2292434126138687, + "learning_rate": 2.6953248269596387e-05, + "loss": 0.0419, + "step": 32453 + }, + { + "epoch": 0.7607728687213207, + "grad_norm": 0.3582462668418884, + "learning_rate": 2.6948217769142126e-05, + "loss": 0.0364, + "step": 32454 + }, + { + "epoch": 0.7607963102961257, + "grad_norm": 0.41266030073165894, + "learning_rate": 2.6943187665068902e-05, + "loss": 0.0351, + "step": 32455 + }, + { + "epoch": 0.7608197518709306, + "grad_norm": 0.46868789196014404, + "learning_rate": 2.6938157957404008e-05, + "loss": 0.1005, + "step": 32456 + }, + { + "epoch": 0.7608431934457357, + "grad_norm": 0.5051397085189819, + "learning_rate": 2.693312864617472e-05, + "loss": 0.7199, + "step": 32457 + }, + { + "epoch": 0.7608666350205406, + "grad_norm": 0.43113410472869873, + "learning_rate": 2.692809973140834e-05, + "loss": 0.0713, + "step": 32458 + }, + { + "epoch": 0.7608900765953457, + "grad_norm": 0.32165226340293884, + "learning_rate": 2.6923071213132135e-05, + "loss": 0.0596, + "step": 32459 + }, + { + "epoch": 0.7609135181701506, + "grad_norm": 0.19878406822681427, + "learning_rate": 2.6918043091373367e-05, + "loss": 0.0353, + "step": 32460 + }, + { + "epoch": 0.7609369597449557, + "grad_norm": 0.09249850362539291, + "learning_rate": 2.691301536615939e-05, + "loss": 0.0158, + "step": 32461 + }, + { + "epoch": 0.7609604013197606, + "grad_norm": 0.9938367605209351, + "learning_rate": 2.6907988037517438e-05, + "loss": 0.1606, + "step": 32462 + }, + { + "epoch": 0.7609838428945657, + "grad_norm": 1.3697230815887451, + "learning_rate": 2.6902961105474812e-05, + "loss": 0.0581, + "step": 32463 + }, + { + "epoch": 0.7610072844693706, + "grad_norm": 0.11693922430276871, + "learning_rate": 2.689793457005876e-05, + "loss": 0.0072, + "step": 32464 + }, + { + "epoch": 0.7610307260441757, + "grad_norm": 0.3330000936985016, + "learning_rate": 2.689290843129655e-05, + "loss": 0.063, + "step": 32465 + }, + { + "epoch": 0.7610541676189806, + "grad_norm": 0.31748121976852417, + "learning_rate": 2.6887882689215505e-05, + "loss": 0.0503, + "step": 32466 + }, + { + "epoch": 0.7610776091937856, + "grad_norm": 0.1562109738588333, + "learning_rate": 2.6882857343842825e-05, + "loss": 0.0276, + "step": 32467 + }, + { + "epoch": 0.7611010507685906, + "grad_norm": 0.6380119323730469, + "learning_rate": 2.6877832395205847e-05, + "loss": 0.0512, + "step": 32468 + }, + { + "epoch": 0.7611244923433956, + "grad_norm": 0.2372322976589203, + "learning_rate": 2.6872807843331817e-05, + "loss": 0.0557, + "step": 32469 + }, + { + "epoch": 0.7611479339182006, + "grad_norm": 0.3423498570919037, + "learning_rate": 2.6867783688247982e-05, + "loss": 0.0714, + "step": 32470 + }, + { + "epoch": 0.7611713754930056, + "grad_norm": 0.12790842354297638, + "learning_rate": 2.6862759929981607e-05, + "loss": 0.013, + "step": 32471 + }, + { + "epoch": 0.7611948170678107, + "grad_norm": 0.7564119100570679, + "learning_rate": 2.6857736568559933e-05, + "loss": 0.1735, + "step": 32472 + }, + { + "epoch": 0.7612182586426156, + "grad_norm": 0.33009156584739685, + "learning_rate": 2.685271360401026e-05, + "loss": 0.0524, + "step": 32473 + }, + { + "epoch": 0.7612417002174207, + "grad_norm": 0.10870582610368729, + "learning_rate": 2.6847691036359813e-05, + "loss": 0.0182, + "step": 32474 + }, + { + "epoch": 0.7612651417922256, + "grad_norm": 0.6572216749191284, + "learning_rate": 2.6842668865635856e-05, + "loss": 0.1126, + "step": 32475 + }, + { + "epoch": 0.7612885833670306, + "grad_norm": 0.4577542245388031, + "learning_rate": 2.6837647091865637e-05, + "loss": 0.0651, + "step": 32476 + }, + { + "epoch": 0.7613120249418356, + "grad_norm": 0.297740638256073, + "learning_rate": 2.6832625715076378e-05, + "loss": 0.0316, + "step": 32477 + }, + { + "epoch": 0.7613354665166406, + "grad_norm": 0.5114248394966125, + "learning_rate": 2.6827604735295364e-05, + "loss": 0.06, + "step": 32478 + }, + { + "epoch": 0.7613589080914456, + "grad_norm": 0.12257380038499832, + "learning_rate": 2.6822584152549813e-05, + "loss": 0.017, + "step": 32479 + }, + { + "epoch": 0.7613823496662506, + "grad_norm": 0.5809882879257202, + "learning_rate": 2.681756396686699e-05, + "loss": 0.0941, + "step": 32480 + }, + { + "epoch": 0.7614057912410556, + "grad_norm": 0.22451186180114746, + "learning_rate": 2.6812544178274135e-05, + "loss": 0.0406, + "step": 32481 + }, + { + "epoch": 0.7614292328158606, + "grad_norm": 0.7025724649429321, + "learning_rate": 2.6807524786798465e-05, + "loss": 0.6195, + "step": 32482 + }, + { + "epoch": 0.7614526743906656, + "grad_norm": 0.3890184462070465, + "learning_rate": 2.6802505792467224e-05, + "loss": 0.0677, + "step": 32483 + }, + { + "epoch": 0.7614761159654706, + "grad_norm": 0.4693637490272522, + "learning_rate": 2.6797487195307647e-05, + "loss": 0.0822, + "step": 32484 + }, + { + "epoch": 0.7614995575402755, + "grad_norm": 0.1754664182662964, + "learning_rate": 2.679246899534693e-05, + "loss": 0.0258, + "step": 32485 + }, + { + "epoch": 0.7615229991150806, + "grad_norm": 0.38245171308517456, + "learning_rate": 2.6787451192612378e-05, + "loss": 0.0467, + "step": 32486 + }, + { + "epoch": 0.7615464406898855, + "grad_norm": 0.3670119345188141, + "learning_rate": 2.6782433787131157e-05, + "loss": 0.0917, + "step": 32487 + }, + { + "epoch": 0.7615698822646906, + "grad_norm": 0.4005720019340515, + "learning_rate": 2.677741677893052e-05, + "loss": 0.056, + "step": 32488 + }, + { + "epoch": 0.7615933238394955, + "grad_norm": 0.10226808488368988, + "learning_rate": 2.677240016803767e-05, + "loss": 0.0124, + "step": 32489 + }, + { + "epoch": 0.7616167654143006, + "grad_norm": 0.49202707409858704, + "learning_rate": 2.676738395447982e-05, + "loss": 0.1041, + "step": 32490 + }, + { + "epoch": 0.7616402069891055, + "grad_norm": 0.2230633795261383, + "learning_rate": 2.6762368138284223e-05, + "loss": 0.0597, + "step": 32491 + }, + { + "epoch": 0.7616636485639106, + "grad_norm": 0.14369440078735352, + "learning_rate": 2.6757352719478047e-05, + "loss": 0.0125, + "step": 32492 + }, + { + "epoch": 0.7616870901387155, + "grad_norm": 0.15005923807621002, + "learning_rate": 2.675233769808856e-05, + "loss": 0.0243, + "step": 32493 + }, + { + "epoch": 0.7617105317135205, + "grad_norm": 0.39308005571365356, + "learning_rate": 2.674732307414295e-05, + "loss": 0.0847, + "step": 32494 + }, + { + "epoch": 0.7617339732883255, + "grad_norm": 0.378325879573822, + "learning_rate": 2.674230884766843e-05, + "loss": 0.057, + "step": 32495 + }, + { + "epoch": 0.7617574148631305, + "grad_norm": 0.5436826348304749, + "learning_rate": 2.67372950186922e-05, + "loss": 0.1113, + "step": 32496 + }, + { + "epoch": 0.7617808564379355, + "grad_norm": 0.43132275342941284, + "learning_rate": 2.673228158724147e-05, + "loss": 0.0684, + "step": 32497 + }, + { + "epoch": 0.7618042980127405, + "grad_norm": 0.802645742893219, + "learning_rate": 2.67272685533434e-05, + "loss": 0.1233, + "step": 32498 + }, + { + "epoch": 0.7618277395875455, + "grad_norm": 0.22539985179901123, + "learning_rate": 2.6722255917025275e-05, + "loss": 0.0179, + "step": 32499 + }, + { + "epoch": 0.7618511811623505, + "grad_norm": 0.8129948973655701, + "learning_rate": 2.671724367831424e-05, + "loss": 0.1925, + "step": 32500 + }, + { + "epoch": 0.7618746227371554, + "grad_norm": 0.38264453411102295, + "learning_rate": 2.67122318372375e-05, + "loss": 0.09, + "step": 32501 + }, + { + "epoch": 0.7618980643119605, + "grad_norm": 0.7460920214653015, + "learning_rate": 2.670722039382223e-05, + "loss": 0.0668, + "step": 32502 + }, + { + "epoch": 0.7619215058867654, + "grad_norm": 0.30825573205947876, + "learning_rate": 2.6702209348095665e-05, + "loss": 0.0312, + "step": 32503 + }, + { + "epoch": 0.7619449474615705, + "grad_norm": 0.38258352875709534, + "learning_rate": 2.669719870008497e-05, + "loss": 0.05, + "step": 32504 + }, + { + "epoch": 0.7619683890363754, + "grad_norm": 0.5077416896820068, + "learning_rate": 2.6692188449817313e-05, + "loss": 0.0957, + "step": 32505 + }, + { + "epoch": 0.7619918306111805, + "grad_norm": 0.32712727785110474, + "learning_rate": 2.6687178597319927e-05, + "loss": 0.0577, + "step": 32506 + }, + { + "epoch": 0.7620152721859854, + "grad_norm": 0.21606463193893433, + "learning_rate": 2.6682169142619984e-05, + "loss": 0.0418, + "step": 32507 + }, + { + "epoch": 0.7620387137607905, + "grad_norm": 0.4479966461658478, + "learning_rate": 2.6677160085744647e-05, + "loss": 0.0982, + "step": 32508 + }, + { + "epoch": 0.7620621553355954, + "grad_norm": 0.43718811869621277, + "learning_rate": 2.6672151426721092e-05, + "loss": 0.0424, + "step": 32509 + }, + { + "epoch": 0.7620855969104005, + "grad_norm": 0.5658136606216431, + "learning_rate": 2.666714316557648e-05, + "loss": 0.0401, + "step": 32510 + }, + { + "epoch": 0.7621090384852054, + "grad_norm": 0.48382413387298584, + "learning_rate": 2.6662135302338042e-05, + "loss": 0.1192, + "step": 32511 + }, + { + "epoch": 0.7621324800600104, + "grad_norm": 0.4940923750400543, + "learning_rate": 2.6657127837032924e-05, + "loss": 0.0621, + "step": 32512 + }, + { + "epoch": 0.7621559216348154, + "grad_norm": 0.5583873987197876, + "learning_rate": 2.6652120769688295e-05, + "loss": 0.1111, + "step": 32513 + }, + { + "epoch": 0.7621793632096204, + "grad_norm": 0.16683268547058105, + "learning_rate": 2.6647114100331282e-05, + "loss": 0.0335, + "step": 32514 + }, + { + "epoch": 0.7622028047844254, + "grad_norm": 0.5275370478630066, + "learning_rate": 2.6642107828989115e-05, + "loss": 0.1036, + "step": 32515 + }, + { + "epoch": 0.7622262463592304, + "grad_norm": 0.2540106773376465, + "learning_rate": 2.6637101955688947e-05, + "loss": 0.0286, + "step": 32516 + }, + { + "epoch": 0.7622496879340354, + "grad_norm": 0.28470245003700256, + "learning_rate": 2.6632096480457892e-05, + "loss": 0.0423, + "step": 32517 + }, + { + "epoch": 0.7622731295088404, + "grad_norm": 0.3274291753768921, + "learning_rate": 2.6627091403323178e-05, + "loss": 0.0554, + "step": 32518 + }, + { + "epoch": 0.7622965710836453, + "grad_norm": 0.39180582761764526, + "learning_rate": 2.6622086724311923e-05, + "loss": 0.06, + "step": 32519 + }, + { + "epoch": 0.7623200126584504, + "grad_norm": 0.18279770016670227, + "learning_rate": 2.661708244345129e-05, + "loss": 0.0268, + "step": 32520 + }, + { + "epoch": 0.7623434542332553, + "grad_norm": 0.1747642457485199, + "learning_rate": 2.6612078560768427e-05, + "loss": 0.0224, + "step": 32521 + }, + { + "epoch": 0.7623668958080604, + "grad_norm": 0.32576993107795715, + "learning_rate": 2.66070750762905e-05, + "loss": 0.363, + "step": 32522 + }, + { + "epoch": 0.7623903373828654, + "grad_norm": 0.4371490776538849, + "learning_rate": 2.660207199004461e-05, + "loss": 0.05, + "step": 32523 + }, + { + "epoch": 0.7624137789576704, + "grad_norm": 0.47019970417022705, + "learning_rate": 2.6597069302057975e-05, + "loss": 0.5531, + "step": 32524 + }, + { + "epoch": 0.7624372205324754, + "grad_norm": 0.9092203378677368, + "learning_rate": 2.65920670123577e-05, + "loss": 0.2042, + "step": 32525 + }, + { + "epoch": 0.7624606621072804, + "grad_norm": 0.41803374886512756, + "learning_rate": 2.658706512097091e-05, + "loss": 0.0635, + "step": 32526 + }, + { + "epoch": 0.7624841036820854, + "grad_norm": 0.35882848501205444, + "learning_rate": 2.6582063627924792e-05, + "loss": 0.0399, + "step": 32527 + }, + { + "epoch": 0.7625075452568904, + "grad_norm": 0.278572678565979, + "learning_rate": 2.6577062533246468e-05, + "loss": 0.042, + "step": 32528 + }, + { + "epoch": 0.7625309868316954, + "grad_norm": 0.33343857526779175, + "learning_rate": 2.657206183696306e-05, + "loss": 0.0435, + "step": 32529 + }, + { + "epoch": 0.7625544284065003, + "grad_norm": 0.0924874097108841, + "learning_rate": 2.656706153910169e-05, + "loss": 0.013, + "step": 32530 + }, + { + "epoch": 0.7625778699813054, + "grad_norm": 0.5253579616546631, + "learning_rate": 2.656206163968953e-05, + "loss": 0.0605, + "step": 32531 + }, + { + "epoch": 0.7626013115561103, + "grad_norm": 0.25106510519981384, + "learning_rate": 2.6557062138753686e-05, + "loss": 0.0482, + "step": 32532 + }, + { + "epoch": 0.7626247531309154, + "grad_norm": 0.22852978110313416, + "learning_rate": 2.6552063036321296e-05, + "loss": 0.0351, + "step": 32533 + }, + { + "epoch": 0.7626481947057203, + "grad_norm": 0.32454949617385864, + "learning_rate": 2.654706433241947e-05, + "loss": 0.0564, + "step": 32534 + }, + { + "epoch": 0.7626716362805254, + "grad_norm": 0.09793080389499664, + "learning_rate": 2.6542066027075342e-05, + "loss": 0.0169, + "step": 32535 + }, + { + "epoch": 0.7626950778553303, + "grad_norm": 0.26743727922439575, + "learning_rate": 2.6537068120316e-05, + "loss": 0.0245, + "step": 32536 + }, + { + "epoch": 0.7627185194301354, + "grad_norm": 0.045577626675367355, + "learning_rate": 2.6532070612168615e-05, + "loss": 0.0031, + "step": 32537 + }, + { + "epoch": 0.7627419610049403, + "grad_norm": 0.26788330078125, + "learning_rate": 2.652707350266026e-05, + "loss": 0.0372, + "step": 32538 + }, + { + "epoch": 0.7627654025797453, + "grad_norm": 0.18898910284042358, + "learning_rate": 2.6522076791818097e-05, + "loss": 0.0239, + "step": 32539 + }, + { + "epoch": 0.7627888441545503, + "grad_norm": 0.3705241084098816, + "learning_rate": 2.651708047966921e-05, + "loss": 0.0728, + "step": 32540 + }, + { + "epoch": 0.7628122857293553, + "grad_norm": 0.5195115208625793, + "learning_rate": 2.6512084566240715e-05, + "loss": 0.0707, + "step": 32541 + }, + { + "epoch": 0.7628357273041603, + "grad_norm": 0.27033382654190063, + "learning_rate": 2.6507089051559708e-05, + "loss": 0.0391, + "step": 32542 + }, + { + "epoch": 0.7628591688789653, + "grad_norm": 0.15971748530864716, + "learning_rate": 2.650209393565327e-05, + "loss": 0.0155, + "step": 32543 + }, + { + "epoch": 0.7628826104537703, + "grad_norm": 0.4036498963832855, + "learning_rate": 2.6497099218548583e-05, + "loss": 0.0704, + "step": 32544 + }, + { + "epoch": 0.7629060520285753, + "grad_norm": 0.6886852979660034, + "learning_rate": 2.6492104900272686e-05, + "loss": 0.1534, + "step": 32545 + }, + { + "epoch": 0.7629294936033802, + "grad_norm": 0.21716025471687317, + "learning_rate": 2.6487110980852704e-05, + "loss": 0.0186, + "step": 32546 + }, + { + "epoch": 0.7629529351781853, + "grad_norm": 0.6930222511291504, + "learning_rate": 2.6482117460315713e-05, + "loss": 0.1428, + "step": 32547 + }, + { + "epoch": 0.7629763767529902, + "grad_norm": 0.5842614769935608, + "learning_rate": 2.6477124338688807e-05, + "loss": 0.1472, + "step": 32548 + }, + { + "epoch": 0.7629998183277953, + "grad_norm": 0.4353879392147064, + "learning_rate": 2.647213161599911e-05, + "loss": 0.0786, + "step": 32549 + }, + { + "epoch": 0.7630232599026002, + "grad_norm": 0.18486511707305908, + "learning_rate": 2.6467139292273668e-05, + "loss": 0.0546, + "step": 32550 + }, + { + "epoch": 0.7630467014774053, + "grad_norm": 0.2639697790145874, + "learning_rate": 2.6462147367539624e-05, + "loss": 0.0167, + "step": 32551 + }, + { + "epoch": 0.7630701430522102, + "grad_norm": 0.13514532148838043, + "learning_rate": 2.6457155841824032e-05, + "loss": 0.0222, + "step": 32552 + }, + { + "epoch": 0.7630935846270153, + "grad_norm": 0.30636516213417053, + "learning_rate": 2.6452164715153992e-05, + "loss": 0.0564, + "step": 32553 + }, + { + "epoch": 0.7631170262018202, + "grad_norm": 0.5757426023483276, + "learning_rate": 2.6447173987556563e-05, + "loss": 0.0723, + "step": 32554 + }, + { + "epoch": 0.7631404677766253, + "grad_norm": 0.5345349311828613, + "learning_rate": 2.6442183659058815e-05, + "loss": 0.0656, + "step": 32555 + }, + { + "epoch": 0.7631639093514302, + "grad_norm": 0.18614384531974792, + "learning_rate": 2.643719372968787e-05, + "loss": 0.0155, + "step": 32556 + }, + { + "epoch": 0.7631873509262352, + "grad_norm": 0.4507094621658325, + "learning_rate": 2.643220419947079e-05, + "loss": 0.0461, + "step": 32557 + }, + { + "epoch": 0.7632107925010402, + "grad_norm": 0.36709967255592346, + "learning_rate": 2.642721506843463e-05, + "loss": 0.0863, + "step": 32558 + }, + { + "epoch": 0.7632342340758452, + "grad_norm": 0.06322699040174484, + "learning_rate": 2.6422226336606472e-05, + "loss": 0.0042, + "step": 32559 + }, + { + "epoch": 0.7632576756506502, + "grad_norm": 0.3618397116661072, + "learning_rate": 2.6417238004013377e-05, + "loss": 0.0523, + "step": 32560 + }, + { + "epoch": 0.7632811172254552, + "grad_norm": 0.2280530035495758, + "learning_rate": 2.6412250070682387e-05, + "loss": 0.0551, + "step": 32561 + }, + { + "epoch": 0.7633045588002602, + "grad_norm": 0.31178802251815796, + "learning_rate": 2.640726253664063e-05, + "loss": 0.0313, + "step": 32562 + }, + { + "epoch": 0.7633280003750652, + "grad_norm": 0.4937344491481781, + "learning_rate": 2.6402275401915112e-05, + "loss": 0.0846, + "step": 32563 + }, + { + "epoch": 0.7633514419498701, + "grad_norm": 0.34145382046699524, + "learning_rate": 2.639728866653295e-05, + "loss": 0.0754, + "step": 32564 + }, + { + "epoch": 0.7633748835246752, + "grad_norm": 0.16011177003383636, + "learning_rate": 2.6392302330521157e-05, + "loss": 0.013, + "step": 32565 + }, + { + "epoch": 0.7633983250994801, + "grad_norm": 0.5139769911766052, + "learning_rate": 2.63873163939068e-05, + "loss": 0.0646, + "step": 32566 + }, + { + "epoch": 0.7634217666742852, + "grad_norm": 0.281842440366745, + "learning_rate": 2.638233085671695e-05, + "loss": 0.0229, + "step": 32567 + }, + { + "epoch": 0.7634452082490901, + "grad_norm": 0.33368009328842163, + "learning_rate": 2.6377345718978597e-05, + "loss": 0.0321, + "step": 32568 + }, + { + "epoch": 0.7634686498238952, + "grad_norm": 0.3947332203388214, + "learning_rate": 2.637236098071888e-05, + "loss": 0.0558, + "step": 32569 + }, + { + "epoch": 0.7634920913987001, + "grad_norm": 0.48347315192222595, + "learning_rate": 2.6367376641964792e-05, + "loss": 0.0667, + "step": 32570 + }, + { + "epoch": 0.7635155329735052, + "grad_norm": 0.3454737961292267, + "learning_rate": 2.6362392702743387e-05, + "loss": 0.068, + "step": 32571 + }, + { + "epoch": 0.7635389745483101, + "grad_norm": 0.5283222198486328, + "learning_rate": 2.6357409163081714e-05, + "loss": 0.1106, + "step": 32572 + }, + { + "epoch": 0.7635624161231152, + "grad_norm": 0.4976631999015808, + "learning_rate": 2.6352426023006814e-05, + "loss": 0.0677, + "step": 32573 + }, + { + "epoch": 0.7635858576979202, + "grad_norm": 0.4585740864276886, + "learning_rate": 2.634744328254569e-05, + "loss": 0.0886, + "step": 32574 + }, + { + "epoch": 0.7636092992727251, + "grad_norm": 0.09843441098928452, + "learning_rate": 2.634246094172541e-05, + "loss": 0.0111, + "step": 32575 + }, + { + "epoch": 0.7636327408475302, + "grad_norm": 0.4213763177394867, + "learning_rate": 2.633747900057303e-05, + "loss": 0.1069, + "step": 32576 + }, + { + "epoch": 0.7636561824223351, + "grad_norm": 0.2678213119506836, + "learning_rate": 2.633249745911557e-05, + "loss": 0.0195, + "step": 32577 + }, + { + "epoch": 0.7636796239971402, + "grad_norm": 0.3445814847946167, + "learning_rate": 2.6327516317380042e-05, + "loss": 0.0473, + "step": 32578 + }, + { + "epoch": 0.7637030655719451, + "grad_norm": 0.31462886929512024, + "learning_rate": 2.6322535575393482e-05, + "loss": 0.0446, + "step": 32579 + }, + { + "epoch": 0.7637265071467502, + "grad_norm": 0.3354867696762085, + "learning_rate": 2.6317555233182923e-05, + "loss": 0.0475, + "step": 32580 + }, + { + "epoch": 0.7637499487215551, + "grad_norm": 0.5554512143135071, + "learning_rate": 2.6312575290775343e-05, + "loss": 0.1049, + "step": 32581 + }, + { + "epoch": 0.7637733902963602, + "grad_norm": 0.40754976868629456, + "learning_rate": 2.6307595748197834e-05, + "loss": 0.0544, + "step": 32582 + }, + { + "epoch": 0.7637968318711651, + "grad_norm": 0.47551679611206055, + "learning_rate": 2.630261660547738e-05, + "loss": 0.1324, + "step": 32583 + }, + { + "epoch": 0.7638202734459701, + "grad_norm": 0.2084614634513855, + "learning_rate": 2.629763786264101e-05, + "loss": 0.0254, + "step": 32584 + }, + { + "epoch": 0.7638437150207751, + "grad_norm": 0.6304846405982971, + "learning_rate": 2.6292659519715724e-05, + "loss": 0.0965, + "step": 32585 + }, + { + "epoch": 0.7638671565955801, + "grad_norm": 0.5391427874565125, + "learning_rate": 2.6287681576728505e-05, + "loss": 0.5429, + "step": 32586 + }, + { + "epoch": 0.7638905981703851, + "grad_norm": 0.1442636251449585, + "learning_rate": 2.6282704033706408e-05, + "loss": 0.0297, + "step": 32587 + }, + { + "epoch": 0.7639140397451901, + "grad_norm": 0.4476216435432434, + "learning_rate": 2.6277726890676457e-05, + "loss": 0.0516, + "step": 32588 + }, + { + "epoch": 0.7639374813199951, + "grad_norm": 0.3663976490497589, + "learning_rate": 2.6272750147665626e-05, + "loss": 0.0503, + "step": 32589 + }, + { + "epoch": 0.7639609228948001, + "grad_norm": 0.7338818311691284, + "learning_rate": 2.6267773804700936e-05, + "loss": 0.0456, + "step": 32590 + }, + { + "epoch": 0.763984364469605, + "grad_norm": 0.3372522294521332, + "learning_rate": 2.6262797861809373e-05, + "loss": 0.2665, + "step": 32591 + }, + { + "epoch": 0.7640078060444101, + "grad_norm": 0.483355313539505, + "learning_rate": 2.6257822319017942e-05, + "loss": 0.0754, + "step": 32592 + }, + { + "epoch": 0.764031247619215, + "grad_norm": 0.5695862174034119, + "learning_rate": 2.625284717635361e-05, + "loss": 0.4251, + "step": 32593 + }, + { + "epoch": 0.7640546891940201, + "grad_norm": 0.17235763370990753, + "learning_rate": 2.6247872433843445e-05, + "loss": 0.0189, + "step": 32594 + }, + { + "epoch": 0.764078130768825, + "grad_norm": 0.27337193489074707, + "learning_rate": 2.624289809151438e-05, + "loss": 0.0177, + "step": 32595 + }, + { + "epoch": 0.7641015723436301, + "grad_norm": 0.40386253595352173, + "learning_rate": 2.6237924149393434e-05, + "loss": 0.0872, + "step": 32596 + }, + { + "epoch": 0.764125013918435, + "grad_norm": 0.5598697066307068, + "learning_rate": 2.623295060750759e-05, + "loss": 0.0442, + "step": 32597 + }, + { + "epoch": 0.7641484554932401, + "grad_norm": 0.24961042404174805, + "learning_rate": 2.6227977465883812e-05, + "loss": 0.0227, + "step": 32598 + }, + { + "epoch": 0.764171897068045, + "grad_norm": 0.22072666883468628, + "learning_rate": 2.6223004724549093e-05, + "loss": 0.0185, + "step": 32599 + }, + { + "epoch": 0.76419533864285, + "grad_norm": 0.16471515595912933, + "learning_rate": 2.621803238353041e-05, + "loss": 0.037, + "step": 32600 + }, + { + "epoch": 0.764218780217655, + "grad_norm": 0.25083985924720764, + "learning_rate": 2.6213060442854797e-05, + "loss": 0.0565, + "step": 32601 + }, + { + "epoch": 0.76424222179246, + "grad_norm": 0.5687932968139648, + "learning_rate": 2.620808890254919e-05, + "loss": 0.0899, + "step": 32602 + }, + { + "epoch": 0.764265663367265, + "grad_norm": 0.42575693130493164, + "learning_rate": 2.620311776264056e-05, + "loss": 0.4749, + "step": 32603 + }, + { + "epoch": 0.76428910494207, + "grad_norm": 0.15171830356121063, + "learning_rate": 2.6198147023155893e-05, + "loss": 0.0246, + "step": 32604 + }, + { + "epoch": 0.764312546516875, + "grad_norm": 0.6867642402648926, + "learning_rate": 2.619317668412217e-05, + "loss": 0.0589, + "step": 32605 + }, + { + "epoch": 0.76433598809168, + "grad_norm": 0.18168357014656067, + "learning_rate": 2.6188206745566302e-05, + "loss": 0.0306, + "step": 32606 + }, + { + "epoch": 0.764359429666485, + "grad_norm": 0.06457477807998657, + "learning_rate": 2.618323720751532e-05, + "loss": 0.0052, + "step": 32607 + }, + { + "epoch": 0.76438287124129, + "grad_norm": 0.5063832402229309, + "learning_rate": 2.6178268069996183e-05, + "loss": 0.1018, + "step": 32608 + }, + { + "epoch": 0.764406312816095, + "grad_norm": 0.5755625367164612, + "learning_rate": 2.6173299333035838e-05, + "loss": 0.6132, + "step": 32609 + }, + { + "epoch": 0.7644297543909, + "grad_norm": 0.36043864488601685, + "learning_rate": 2.616833099666124e-05, + "loss": 0.0778, + "step": 32610 + }, + { + "epoch": 0.7644531959657049, + "grad_norm": 0.2497245967388153, + "learning_rate": 2.6163363060899327e-05, + "loss": 0.0164, + "step": 32611 + }, + { + "epoch": 0.76447663754051, + "grad_norm": 0.49264976382255554, + "learning_rate": 2.6158395525777115e-05, + "loss": 0.0484, + "step": 32612 + }, + { + "epoch": 0.7645000791153149, + "grad_norm": 0.32286369800567627, + "learning_rate": 2.615342839132149e-05, + "loss": 0.0512, + "step": 32613 + }, + { + "epoch": 0.76452352069012, + "grad_norm": 0.20369207859039307, + "learning_rate": 2.6148461657559475e-05, + "loss": 0.0399, + "step": 32614 + }, + { + "epoch": 0.7645469622649249, + "grad_norm": 0.586796760559082, + "learning_rate": 2.6143495324517974e-05, + "loss": 0.1236, + "step": 32615 + }, + { + "epoch": 0.76457040383973, + "grad_norm": 0.6527544260025024, + "learning_rate": 2.6138529392223953e-05, + "loss": 0.0826, + "step": 32616 + }, + { + "epoch": 0.7645938454145349, + "grad_norm": 0.3508760631084442, + "learning_rate": 2.6133563860704346e-05, + "loss": 0.0308, + "step": 32617 + }, + { + "epoch": 0.76461728698934, + "grad_norm": 0.6491065621376038, + "learning_rate": 2.6128598729986096e-05, + "loss": 0.0808, + "step": 32618 + }, + { + "epoch": 0.7646407285641449, + "grad_norm": 0.15647612512111664, + "learning_rate": 2.612363400009612e-05, + "loss": 0.0149, + "step": 32619 + }, + { + "epoch": 0.7646641701389499, + "grad_norm": 0.6262555718421936, + "learning_rate": 2.611866967106141e-05, + "loss": 0.1284, + "step": 32620 + }, + { + "epoch": 0.7646876117137549, + "grad_norm": 0.5814237594604492, + "learning_rate": 2.611370574290888e-05, + "loss": 0.0727, + "step": 32621 + }, + { + "epoch": 0.7647110532885599, + "grad_norm": 0.584290623664856, + "learning_rate": 2.6108742215665462e-05, + "loss": 0.0657, + "step": 32622 + }, + { + "epoch": 0.7647344948633649, + "grad_norm": 0.13497500121593475, + "learning_rate": 2.610377908935805e-05, + "loss": 0.012, + "step": 32623 + }, + { + "epoch": 0.7647579364381699, + "grad_norm": 0.18858736753463745, + "learning_rate": 2.6098816364013644e-05, + "loss": 0.0196, + "step": 32624 + }, + { + "epoch": 0.764781378012975, + "grad_norm": 0.6539552211761475, + "learning_rate": 2.6093854039659104e-05, + "loss": 0.0934, + "step": 32625 + }, + { + "epoch": 0.7648048195877799, + "grad_norm": 0.275601327419281, + "learning_rate": 2.6088892116321417e-05, + "loss": 0.0462, + "step": 32626 + }, + { + "epoch": 0.764828261162585, + "grad_norm": 0.45463818311691284, + "learning_rate": 2.6083930594027494e-05, + "loss": 0.0786, + "step": 32627 + }, + { + "epoch": 0.7648517027373899, + "grad_norm": 0.21142445504665375, + "learning_rate": 2.6078969472804227e-05, + "loss": 0.0162, + "step": 32628 + }, + { + "epoch": 0.7648751443121949, + "grad_norm": 0.20918533205986023, + "learning_rate": 2.607400875267856e-05, + "loss": 0.037, + "step": 32629 + }, + { + "epoch": 0.7648985858869999, + "grad_norm": 0.4283890724182129, + "learning_rate": 2.606904843367739e-05, + "loss": 0.0693, + "step": 32630 + }, + { + "epoch": 0.7649220274618049, + "grad_norm": 2.0934560298919678, + "learning_rate": 2.6064088515827622e-05, + "loss": 0.1857, + "step": 32631 + }, + { + "epoch": 0.7649454690366099, + "grad_norm": 0.36860132217407227, + "learning_rate": 2.605912899915621e-05, + "loss": 0.0533, + "step": 32632 + }, + { + "epoch": 0.7649689106114149, + "grad_norm": 0.17257675528526306, + "learning_rate": 2.605416988369005e-05, + "loss": 0.0182, + "step": 32633 + }, + { + "epoch": 0.7649923521862199, + "grad_norm": 0.18092115223407745, + "learning_rate": 2.6049211169456044e-05, + "loss": 0.0209, + "step": 32634 + }, + { + "epoch": 0.7650157937610249, + "grad_norm": 0.74555504322052, + "learning_rate": 2.60442528564811e-05, + "loss": 0.1136, + "step": 32635 + }, + { + "epoch": 0.7650392353358298, + "grad_norm": 0.41450828313827515, + "learning_rate": 2.603929494479208e-05, + "loss": 0.0296, + "step": 32636 + }, + { + "epoch": 0.7650626769106349, + "grad_norm": 0.531719982624054, + "learning_rate": 2.6034337434415955e-05, + "loss": 0.0937, + "step": 32637 + }, + { + "epoch": 0.7650861184854398, + "grad_norm": 0.16109268367290497, + "learning_rate": 2.6029380325379572e-05, + "loss": 0.0177, + "step": 32638 + }, + { + "epoch": 0.7651095600602449, + "grad_norm": 0.7263495922088623, + "learning_rate": 2.602442361770988e-05, + "loss": 0.0904, + "step": 32639 + }, + { + "epoch": 0.7651330016350498, + "grad_norm": 0.5065631866455078, + "learning_rate": 2.6019467311433744e-05, + "loss": 0.0439, + "step": 32640 + }, + { + "epoch": 0.7651564432098549, + "grad_norm": 0.34261488914489746, + "learning_rate": 2.6014511406578057e-05, + "loss": 0.0529, + "step": 32641 + }, + { + "epoch": 0.7651798847846598, + "grad_norm": 0.39095285534858704, + "learning_rate": 2.6009555903169713e-05, + "loss": 0.0614, + "step": 32642 + }, + { + "epoch": 0.7652033263594649, + "grad_norm": 0.1829264611005783, + "learning_rate": 2.6004600801235602e-05, + "loss": 0.0516, + "step": 32643 + }, + { + "epoch": 0.7652267679342698, + "grad_norm": 0.33359548449516296, + "learning_rate": 2.5999646100802576e-05, + "loss": 0.0639, + "step": 32644 + }, + { + "epoch": 0.7652502095090749, + "grad_norm": 0.15931889414787292, + "learning_rate": 2.5994691801897576e-05, + "loss": 0.04, + "step": 32645 + }, + { + "epoch": 0.7652736510838798, + "grad_norm": 0.26893743872642517, + "learning_rate": 2.598973790454746e-05, + "loss": 0.0452, + "step": 32646 + }, + { + "epoch": 0.7652970926586848, + "grad_norm": 0.03810034692287445, + "learning_rate": 2.598478440877912e-05, + "loss": 0.0055, + "step": 32647 + }, + { + "epoch": 0.7653205342334898, + "grad_norm": 0.2367354929447174, + "learning_rate": 2.5979831314619373e-05, + "loss": 0.0549, + "step": 32648 + }, + { + "epoch": 0.7653439758082948, + "grad_norm": 0.7071672081947327, + "learning_rate": 2.5974878622095177e-05, + "loss": 0.1138, + "step": 32649 + }, + { + "epoch": 0.7653674173830998, + "grad_norm": 0.3217625916004181, + "learning_rate": 2.5969926331233374e-05, + "loss": 0.038, + "step": 32650 + }, + { + "epoch": 0.7653908589579048, + "grad_norm": 0.46768060326576233, + "learning_rate": 2.596497444206081e-05, + "loss": 0.1005, + "step": 32651 + }, + { + "epoch": 0.7654143005327098, + "grad_norm": 0.33274757862091064, + "learning_rate": 2.5960022954604403e-05, + "loss": 0.2341, + "step": 32652 + }, + { + "epoch": 0.7654377421075148, + "grad_norm": 0.3906935453414917, + "learning_rate": 2.595507186889098e-05, + "loss": 0.0697, + "step": 32653 + }, + { + "epoch": 0.7654611836823197, + "grad_norm": 0.4467635154724121, + "learning_rate": 2.595012118494743e-05, + "loss": 0.6561, + "step": 32654 + }, + { + "epoch": 0.7654846252571248, + "grad_norm": 0.13060544431209564, + "learning_rate": 2.594517090280061e-05, + "loss": 0.0204, + "step": 32655 + }, + { + "epoch": 0.7655080668319297, + "grad_norm": 0.2273750901222229, + "learning_rate": 2.594022102247733e-05, + "loss": 0.0459, + "step": 32656 + }, + { + "epoch": 0.7655315084067348, + "grad_norm": 0.4021090269088745, + "learning_rate": 2.5935271544004535e-05, + "loss": 0.0558, + "step": 32657 + }, + { + "epoch": 0.7655549499815397, + "grad_norm": 0.5944648385047913, + "learning_rate": 2.5930322467409028e-05, + "loss": 0.0709, + "step": 32658 + }, + { + "epoch": 0.7655783915563448, + "grad_norm": 0.8554658889770508, + "learning_rate": 2.5925373792717688e-05, + "loss": 0.4914, + "step": 32659 + }, + { + "epoch": 0.7656018331311497, + "grad_norm": 0.16187918186187744, + "learning_rate": 2.5920425519957316e-05, + "loss": 0.0321, + "step": 32660 + }, + { + "epoch": 0.7656252747059548, + "grad_norm": 0.4627043604850769, + "learning_rate": 2.591547764915483e-05, + "loss": 0.084, + "step": 32661 + }, + { + "epoch": 0.7656487162807597, + "grad_norm": 0.38806718587875366, + "learning_rate": 2.5910530180337035e-05, + "loss": 0.0479, + "step": 32662 + }, + { + "epoch": 0.7656721578555648, + "grad_norm": 0.41587451100349426, + "learning_rate": 2.590558311353076e-05, + "loss": 0.0583, + "step": 32663 + }, + { + "epoch": 0.7656955994303697, + "grad_norm": 0.5136216282844543, + "learning_rate": 2.5900636448762917e-05, + "loss": 0.0986, + "step": 32664 + }, + { + "epoch": 0.7657190410051747, + "grad_norm": 0.1965438723564148, + "learning_rate": 2.5895690186060297e-05, + "loss": 0.0271, + "step": 32665 + }, + { + "epoch": 0.7657424825799797, + "grad_norm": 0.4622262716293335, + "learning_rate": 2.5890744325449746e-05, + "loss": 0.0691, + "step": 32666 + }, + { + "epoch": 0.7657659241547847, + "grad_norm": 0.7041794061660767, + "learning_rate": 2.58857988669581e-05, + "loss": 0.2016, + "step": 32667 + }, + { + "epoch": 0.7657893657295897, + "grad_norm": 0.49423545598983765, + "learning_rate": 2.5880853810612184e-05, + "loss": 0.0836, + "step": 32668 + }, + { + "epoch": 0.7658128073043947, + "grad_norm": 0.7448984980583191, + "learning_rate": 2.5875909156438815e-05, + "loss": 0.1256, + "step": 32669 + }, + { + "epoch": 0.7658362488791997, + "grad_norm": 0.5412852764129639, + "learning_rate": 2.587096490446489e-05, + "loss": 0.571, + "step": 32670 + }, + { + "epoch": 0.7658596904540047, + "grad_norm": 0.38625478744506836, + "learning_rate": 2.5866021054717183e-05, + "loss": 0.0353, + "step": 32671 + }, + { + "epoch": 0.7658831320288096, + "grad_norm": 0.42040255665779114, + "learning_rate": 2.5861077607222496e-05, + "loss": 0.0538, + "step": 32672 + }, + { + "epoch": 0.7659065736036147, + "grad_norm": 0.32574594020843506, + "learning_rate": 2.5856134562007717e-05, + "loss": 0.0474, + "step": 32673 + }, + { + "epoch": 0.7659300151784196, + "grad_norm": 0.39193248748779297, + "learning_rate": 2.5851191919099638e-05, + "loss": 0.0549, + "step": 32674 + }, + { + "epoch": 0.7659534567532247, + "grad_norm": 0.1766601800918579, + "learning_rate": 2.584624967852508e-05, + "loss": 0.0223, + "step": 32675 + }, + { + "epoch": 0.7659768983280297, + "grad_norm": 0.2869073450565338, + "learning_rate": 2.5841307840310825e-05, + "loss": 0.0354, + "step": 32676 + }, + { + "epoch": 0.7660003399028347, + "grad_norm": 0.2545715272426605, + "learning_rate": 2.5836366404483747e-05, + "loss": 0.0366, + "step": 32677 + }, + { + "epoch": 0.7660237814776397, + "grad_norm": 0.16344177722930908, + "learning_rate": 2.583142537107064e-05, + "loss": 0.0303, + "step": 32678 + }, + { + "epoch": 0.7660472230524447, + "grad_norm": 0.16292786598205566, + "learning_rate": 2.582648474009829e-05, + "loss": 0.0572, + "step": 32679 + }, + { + "epoch": 0.7660706646272497, + "grad_norm": 0.21906109154224396, + "learning_rate": 2.582154451159353e-05, + "loss": 0.0431, + "step": 32680 + }, + { + "epoch": 0.7660941062020546, + "grad_norm": 0.28394830226898193, + "learning_rate": 2.5816604685583145e-05, + "loss": 0.0439, + "step": 32681 + }, + { + "epoch": 0.7661175477768597, + "grad_norm": 0.4710542559623718, + "learning_rate": 2.5811665262093932e-05, + "loss": 0.0711, + "step": 32682 + }, + { + "epoch": 0.7661409893516646, + "grad_norm": 0.3806874454021454, + "learning_rate": 2.5806726241152736e-05, + "loss": 0.0591, + "step": 32683 + }, + { + "epoch": 0.7661644309264697, + "grad_norm": 0.52066969871521, + "learning_rate": 2.580178762278629e-05, + "loss": 0.049, + "step": 32684 + }, + { + "epoch": 0.7661878725012746, + "grad_norm": 0.11350807547569275, + "learning_rate": 2.5796849407021463e-05, + "loss": 0.0204, + "step": 32685 + }, + { + "epoch": 0.7662113140760797, + "grad_norm": 0.17442293465137482, + "learning_rate": 2.579191159388502e-05, + "loss": 0.0253, + "step": 32686 + }, + { + "epoch": 0.7662347556508846, + "grad_norm": 0.2944415509700775, + "learning_rate": 2.5786974183403756e-05, + "loss": 0.0518, + "step": 32687 + }, + { + "epoch": 0.7662581972256897, + "grad_norm": 0.28719648718833923, + "learning_rate": 2.5782037175604458e-05, + "loss": 0.0544, + "step": 32688 + }, + { + "epoch": 0.7662816388004946, + "grad_norm": 0.3713674545288086, + "learning_rate": 2.577710057051387e-05, + "loss": 0.0532, + "step": 32689 + }, + { + "epoch": 0.7663050803752997, + "grad_norm": 0.4822627007961273, + "learning_rate": 2.5772164368158857e-05, + "loss": 0.1046, + "step": 32690 + }, + { + "epoch": 0.7663285219501046, + "grad_norm": 0.3148927092552185, + "learning_rate": 2.576722856856617e-05, + "loss": 0.0466, + "step": 32691 + }, + { + "epoch": 0.7663519635249096, + "grad_norm": 0.6196739673614502, + "learning_rate": 2.5762293171762598e-05, + "loss": 0.5752, + "step": 32692 + }, + { + "epoch": 0.7663754050997146, + "grad_norm": 0.2629557251930237, + "learning_rate": 2.5757358177774892e-05, + "loss": 0.0333, + "step": 32693 + }, + { + "epoch": 0.7663988466745196, + "grad_norm": 0.19740840792655945, + "learning_rate": 2.5752423586629826e-05, + "loss": 0.0392, + "step": 32694 + }, + { + "epoch": 0.7664222882493246, + "grad_norm": 0.4059380888938904, + "learning_rate": 2.574748939835423e-05, + "loss": 0.0565, + "step": 32695 + }, + { + "epoch": 0.7664457298241296, + "grad_norm": 0.6092957854270935, + "learning_rate": 2.5742555612974816e-05, + "loss": 0.109, + "step": 32696 + }, + { + "epoch": 0.7664691713989346, + "grad_norm": 0.5995429158210754, + "learning_rate": 2.573762223051841e-05, + "loss": 0.4686, + "step": 32697 + }, + { + "epoch": 0.7664926129737396, + "grad_norm": 0.7385517954826355, + "learning_rate": 2.5732689251011744e-05, + "loss": 0.1936, + "step": 32698 + }, + { + "epoch": 0.7665160545485445, + "grad_norm": 0.4473274350166321, + "learning_rate": 2.5727756674481607e-05, + "loss": 0.1078, + "step": 32699 + }, + { + "epoch": 0.7665394961233496, + "grad_norm": 0.5695046186447144, + "learning_rate": 2.5722824500954746e-05, + "loss": 0.4095, + "step": 32700 + }, + { + "epoch": 0.7665629376981545, + "grad_norm": 0.44538483023643494, + "learning_rate": 2.5717892730457905e-05, + "loss": 0.0919, + "step": 32701 + }, + { + "epoch": 0.7665863792729596, + "grad_norm": 0.23932397365570068, + "learning_rate": 2.5712961363017883e-05, + "loss": 0.0221, + "step": 32702 + }, + { + "epoch": 0.7666098208477645, + "grad_norm": 0.38572943210601807, + "learning_rate": 2.570803039866143e-05, + "loss": 0.063, + "step": 32703 + }, + { + "epoch": 0.7666332624225696, + "grad_norm": 0.5736715793609619, + "learning_rate": 2.570309983741529e-05, + "loss": 0.1048, + "step": 32704 + }, + { + "epoch": 0.7666567039973745, + "grad_norm": 0.35038408637046814, + "learning_rate": 2.5698169679306227e-05, + "loss": 0.0355, + "step": 32705 + }, + { + "epoch": 0.7666801455721796, + "grad_norm": 0.37657177448272705, + "learning_rate": 2.569323992436098e-05, + "loss": 0.0653, + "step": 32706 + }, + { + "epoch": 0.7667035871469845, + "grad_norm": 0.6886297464370728, + "learning_rate": 2.5688310572606266e-05, + "loss": 0.1269, + "step": 32707 + }, + { + "epoch": 0.7667270287217895, + "grad_norm": 0.4672578275203705, + "learning_rate": 2.568338162406887e-05, + "loss": 0.0658, + "step": 32708 + }, + { + "epoch": 0.7667504702965945, + "grad_norm": 0.5092548131942749, + "learning_rate": 2.5678453078775565e-05, + "loss": 0.0349, + "step": 32709 + }, + { + "epoch": 0.7667739118713995, + "grad_norm": 0.2434268593788147, + "learning_rate": 2.567352493675307e-05, + "loss": 0.0225, + "step": 32710 + }, + { + "epoch": 0.7667973534462045, + "grad_norm": 0.45117291808128357, + "learning_rate": 2.5668597198028123e-05, + "loss": 0.0824, + "step": 32711 + }, + { + "epoch": 0.7668207950210095, + "grad_norm": 0.33960989117622375, + "learning_rate": 2.566366986262745e-05, + "loss": 0.0715, + "step": 32712 + }, + { + "epoch": 0.7668442365958145, + "grad_norm": 0.28530389070510864, + "learning_rate": 2.5658742930577783e-05, + "loss": 0.0364, + "step": 32713 + }, + { + "epoch": 0.7668676781706195, + "grad_norm": 0.4174478352069855, + "learning_rate": 2.565381640190585e-05, + "loss": 0.063, + "step": 32714 + }, + { + "epoch": 0.7668911197454245, + "grad_norm": 0.4622770845890045, + "learning_rate": 2.5648890276638427e-05, + "loss": 0.1021, + "step": 32715 + }, + { + "epoch": 0.7669145613202295, + "grad_norm": 0.49716830253601074, + "learning_rate": 2.5643964554802225e-05, + "loss": 0.101, + "step": 32716 + }, + { + "epoch": 0.7669380028950344, + "grad_norm": 0.6654854416847229, + "learning_rate": 2.5639039236423944e-05, + "loss": 0.1451, + "step": 32717 + }, + { + "epoch": 0.7669614444698395, + "grad_norm": 0.15056899189949036, + "learning_rate": 2.5634114321530343e-05, + "loss": 0.0152, + "step": 32718 + }, + { + "epoch": 0.7669848860446444, + "grad_norm": 0.3521624505519867, + "learning_rate": 2.5629189810148123e-05, + "loss": 0.0503, + "step": 32719 + }, + { + "epoch": 0.7670083276194495, + "grad_norm": 0.44733113050460815, + "learning_rate": 2.5624265702303974e-05, + "loss": 0.0887, + "step": 32720 + }, + { + "epoch": 0.7670317691942544, + "grad_norm": 0.45198020339012146, + "learning_rate": 2.561934199802466e-05, + "loss": 0.0804, + "step": 32721 + }, + { + "epoch": 0.7670552107690595, + "grad_norm": 0.4879099726676941, + "learning_rate": 2.561441869733692e-05, + "loss": 0.0952, + "step": 32722 + }, + { + "epoch": 0.7670786523438644, + "grad_norm": 0.3165232539176941, + "learning_rate": 2.5609495800267424e-05, + "loss": 0.0449, + "step": 32723 + }, + { + "epoch": 0.7671020939186695, + "grad_norm": 1.1170016527175903, + "learning_rate": 2.5604573306842904e-05, + "loss": 0.0843, + "step": 32724 + }, + { + "epoch": 0.7671255354934744, + "grad_norm": 0.4094764292240143, + "learning_rate": 2.559965121709006e-05, + "loss": 0.0715, + "step": 32725 + }, + { + "epoch": 0.7671489770682794, + "grad_norm": 0.385486364364624, + "learning_rate": 2.5594729531035587e-05, + "loss": 0.3585, + "step": 32726 + }, + { + "epoch": 0.7671724186430845, + "grad_norm": 0.2779618501663208, + "learning_rate": 2.5589808248706192e-05, + "loss": 0.0394, + "step": 32727 + }, + { + "epoch": 0.7671958602178894, + "grad_norm": 0.09394483268260956, + "learning_rate": 2.5584887370128608e-05, + "loss": 0.023, + "step": 32728 + }, + { + "epoch": 0.7672193017926945, + "grad_norm": 0.1903492659330368, + "learning_rate": 2.5579966895329532e-05, + "loss": 0.0283, + "step": 32729 + }, + { + "epoch": 0.7672427433674994, + "grad_norm": 0.14377138018608093, + "learning_rate": 2.5575046824335636e-05, + "loss": 0.0222, + "step": 32730 + }, + { + "epoch": 0.7672661849423045, + "grad_norm": 0.3945811688899994, + "learning_rate": 2.557012715717363e-05, + "loss": 0.0664, + "step": 32731 + }, + { + "epoch": 0.7672896265171094, + "grad_norm": 0.3941899240016937, + "learning_rate": 2.556520789387018e-05, + "loss": 0.0592, + "step": 32732 + }, + { + "epoch": 0.7673130680919145, + "grad_norm": 0.24021591246128082, + "learning_rate": 2.5560289034452035e-05, + "loss": 0.0355, + "step": 32733 + }, + { + "epoch": 0.7673365096667194, + "grad_norm": 0.38973268866539, + "learning_rate": 2.5555370578945814e-05, + "loss": 0.0666, + "step": 32734 + }, + { + "epoch": 0.7673599512415245, + "grad_norm": 0.47758030891418457, + "learning_rate": 2.555045252737829e-05, + "loss": 0.0285, + "step": 32735 + }, + { + "epoch": 0.7673833928163294, + "grad_norm": 0.4543592035770416, + "learning_rate": 2.55455348797761e-05, + "loss": 0.0498, + "step": 32736 + }, + { + "epoch": 0.7674068343911344, + "grad_norm": 0.14041537046432495, + "learning_rate": 2.554061763616592e-05, + "loss": 0.0221, + "step": 32737 + }, + { + "epoch": 0.7674302759659394, + "grad_norm": 0.4806353449821472, + "learning_rate": 2.5535700796574456e-05, + "loss": 0.5407, + "step": 32738 + }, + { + "epoch": 0.7674537175407444, + "grad_norm": 0.5773870348930359, + "learning_rate": 2.5530784361028338e-05, + "loss": 0.283, + "step": 32739 + }, + { + "epoch": 0.7674771591155494, + "grad_norm": 0.5121378898620605, + "learning_rate": 2.5525868329554314e-05, + "loss": 0.3495, + "step": 32740 + }, + { + "epoch": 0.7675006006903544, + "grad_norm": 0.4917890131473541, + "learning_rate": 2.5520952702179014e-05, + "loss": 0.0558, + "step": 32741 + }, + { + "epoch": 0.7675240422651594, + "grad_norm": 0.38766273856163025, + "learning_rate": 2.5516037478929124e-05, + "loss": 0.0675, + "step": 32742 + }, + { + "epoch": 0.7675474838399644, + "grad_norm": 0.32596486806869507, + "learning_rate": 2.55111226598313e-05, + "loss": 0.0363, + "step": 32743 + }, + { + "epoch": 0.7675709254147693, + "grad_norm": 0.288659930229187, + "learning_rate": 2.5506208244912233e-05, + "loss": 0.0318, + "step": 32744 + }, + { + "epoch": 0.7675943669895744, + "grad_norm": 0.12234362959861755, + "learning_rate": 2.5501294234198535e-05, + "loss": 0.0256, + "step": 32745 + }, + { + "epoch": 0.7676178085643793, + "grad_norm": 0.6459991931915283, + "learning_rate": 2.549638062771691e-05, + "loss": 0.0565, + "step": 32746 + }, + { + "epoch": 0.7676412501391844, + "grad_norm": 0.31328123807907104, + "learning_rate": 2.5491467425494053e-05, + "loss": 0.0542, + "step": 32747 + }, + { + "epoch": 0.7676646917139893, + "grad_norm": 0.459137886762619, + "learning_rate": 2.5486554627556582e-05, + "loss": 0.0832, + "step": 32748 + }, + { + "epoch": 0.7676881332887944, + "grad_norm": 0.5817931294441223, + "learning_rate": 2.548164223393117e-05, + "loss": 0.1428, + "step": 32749 + }, + { + "epoch": 0.7677115748635993, + "grad_norm": 0.1914348155260086, + "learning_rate": 2.5476730244644452e-05, + "loss": 0.0381, + "step": 32750 + }, + { + "epoch": 0.7677350164384044, + "grad_norm": 0.5315142273902893, + "learning_rate": 2.54718186597231e-05, + "loss": 0.651, + "step": 32751 + }, + { + "epoch": 0.7677584580132093, + "grad_norm": 0.36051082611083984, + "learning_rate": 2.5466907479193726e-05, + "loss": 0.0775, + "step": 32752 + }, + { + "epoch": 0.7677818995880143, + "grad_norm": 0.3729764223098755, + "learning_rate": 2.5461996703083025e-05, + "loss": 0.0947, + "step": 32753 + }, + { + "epoch": 0.7678053411628193, + "grad_norm": 0.38993239402770996, + "learning_rate": 2.5457086331417635e-05, + "loss": 0.0573, + "step": 32754 + }, + { + "epoch": 0.7678287827376243, + "grad_norm": 0.35509535670280457, + "learning_rate": 2.545217636422419e-05, + "loss": 0.0403, + "step": 32755 + }, + { + "epoch": 0.7678522243124293, + "grad_norm": 0.567554235458374, + "learning_rate": 2.5447266801529334e-05, + "loss": 0.1105, + "step": 32756 + }, + { + "epoch": 0.7678756658872343, + "grad_norm": 0.13768358528614044, + "learning_rate": 2.544235764335967e-05, + "loss": 0.0233, + "step": 32757 + }, + { + "epoch": 0.7678991074620393, + "grad_norm": 0.7808454036712646, + "learning_rate": 2.54374488897419e-05, + "loss": 0.6894, + "step": 32758 + }, + { + "epoch": 0.7679225490368443, + "grad_norm": 0.5921339392662048, + "learning_rate": 2.5432540540702608e-05, + "loss": 0.1228, + "step": 32759 + }, + { + "epoch": 0.7679459906116493, + "grad_norm": 0.4752545952796936, + "learning_rate": 2.542763259626847e-05, + "loss": 0.0991, + "step": 32760 + }, + { + "epoch": 0.7679694321864543, + "grad_norm": 0.531378984451294, + "learning_rate": 2.542272505646609e-05, + "loss": 0.7069, + "step": 32761 + }, + { + "epoch": 0.7679928737612592, + "grad_norm": 0.3428219258785248, + "learning_rate": 2.5417817921322106e-05, + "loss": 0.0601, + "step": 32762 + }, + { + "epoch": 0.7680163153360643, + "grad_norm": 0.6007969379425049, + "learning_rate": 2.5412911190863142e-05, + "loss": 0.1195, + "step": 32763 + }, + { + "epoch": 0.7680397569108692, + "grad_norm": 0.18512983620166779, + "learning_rate": 2.540800486511582e-05, + "loss": 0.0166, + "step": 32764 + }, + { + "epoch": 0.7680631984856743, + "grad_norm": 0.31295594573020935, + "learning_rate": 2.5403098944106716e-05, + "loss": 0.0321, + "step": 32765 + }, + { + "epoch": 0.7680866400604792, + "grad_norm": 0.566062331199646, + "learning_rate": 2.5398193427862536e-05, + "loss": 0.0888, + "step": 32766 + }, + { + "epoch": 0.7681100816352843, + "grad_norm": 0.14565344154834747, + "learning_rate": 2.539328831640986e-05, + "loss": 0.0149, + "step": 32767 + }, + { + "epoch": 0.7681335232100892, + "grad_norm": 0.3393747806549072, + "learning_rate": 2.53883836097753e-05, + "loss": 0.0806, + "step": 32768 + }, + { + "epoch": 0.7681569647848943, + "grad_norm": 0.4682791233062744, + "learning_rate": 2.5383479307985426e-05, + "loss": 0.5048, + "step": 32769 + }, + { + "epoch": 0.7681804063596992, + "grad_norm": 0.303722620010376, + "learning_rate": 2.537857541106693e-05, + "loss": 0.0346, + "step": 32770 + }, + { + "epoch": 0.7682038479345042, + "grad_norm": 0.45698466897010803, + "learning_rate": 2.5373671919046382e-05, + "loss": 0.0488, + "step": 32771 + }, + { + "epoch": 0.7682272895093092, + "grad_norm": 0.39880985021591187, + "learning_rate": 2.5368768831950363e-05, + "loss": 0.0571, + "step": 32772 + }, + { + "epoch": 0.7682507310841142, + "grad_norm": 0.13338038325309753, + "learning_rate": 2.5363866149805525e-05, + "loss": 0.0131, + "step": 32773 + }, + { + "epoch": 0.7682741726589192, + "grad_norm": 0.4838261306285858, + "learning_rate": 2.5358963872638453e-05, + "loss": 0.0673, + "step": 32774 + }, + { + "epoch": 0.7682976142337242, + "grad_norm": 0.5345944166183472, + "learning_rate": 2.5354062000475732e-05, + "loss": 0.0742, + "step": 32775 + }, + { + "epoch": 0.7683210558085292, + "grad_norm": 0.2601739168167114, + "learning_rate": 2.5349160533343974e-05, + "loss": 0.0398, + "step": 32776 + }, + { + "epoch": 0.7683444973833342, + "grad_norm": 0.2872433066368103, + "learning_rate": 2.534425947126975e-05, + "loss": 0.032, + "step": 32777 + }, + { + "epoch": 0.7683679389581393, + "grad_norm": 0.6287412047386169, + "learning_rate": 2.533935881427969e-05, + "loss": 0.1206, + "step": 32778 + }, + { + "epoch": 0.7683913805329442, + "grad_norm": 0.18761903047561646, + "learning_rate": 2.5334458562400378e-05, + "loss": 0.0281, + "step": 32779 + }, + { + "epoch": 0.7684148221077493, + "grad_norm": 0.3558823764324188, + "learning_rate": 2.5329558715658385e-05, + "loss": 0.0612, + "step": 32780 + }, + { + "epoch": 0.7684382636825542, + "grad_norm": 0.2712797522544861, + "learning_rate": 2.5324659274080274e-05, + "loss": 0.0486, + "step": 32781 + }, + { + "epoch": 0.7684617052573592, + "grad_norm": 0.22287559509277344, + "learning_rate": 2.5319760237692703e-05, + "loss": 0.02, + "step": 32782 + }, + { + "epoch": 0.7684851468321642, + "grad_norm": 0.32657602429389954, + "learning_rate": 2.531486160652221e-05, + "loss": 0.037, + "step": 32783 + }, + { + "epoch": 0.7685085884069692, + "grad_norm": 0.3187412619590759, + "learning_rate": 2.5309963380595338e-05, + "loss": 0.0525, + "step": 32784 + }, + { + "epoch": 0.7685320299817742, + "grad_norm": 0.47150784730911255, + "learning_rate": 2.5305065559938747e-05, + "loss": 0.105, + "step": 32785 + }, + { + "epoch": 0.7685554715565792, + "grad_norm": 0.5778295993804932, + "learning_rate": 2.530016814457896e-05, + "loss": 0.581, + "step": 32786 + }, + { + "epoch": 0.7685789131313842, + "grad_norm": 0.2358601838350296, + "learning_rate": 2.5295271134542563e-05, + "loss": 0.051, + "step": 32787 + }, + { + "epoch": 0.7686023547061892, + "grad_norm": 0.2893798351287842, + "learning_rate": 2.529037452985613e-05, + "loss": 0.014, + "step": 32788 + }, + { + "epoch": 0.7686257962809941, + "grad_norm": 0.43798086047172546, + "learning_rate": 2.528547833054623e-05, + "loss": 0.0506, + "step": 32789 + }, + { + "epoch": 0.7686492378557992, + "grad_norm": 0.2990276515483856, + "learning_rate": 2.5280582536639385e-05, + "loss": 0.0407, + "step": 32790 + }, + { + "epoch": 0.7686726794306041, + "grad_norm": 0.6670650243759155, + "learning_rate": 2.527568714816224e-05, + "loss": 0.1981, + "step": 32791 + }, + { + "epoch": 0.7686961210054092, + "grad_norm": 0.3761550486087799, + "learning_rate": 2.527079216514131e-05, + "loss": 0.0661, + "step": 32792 + }, + { + "epoch": 0.7687195625802141, + "grad_norm": 0.31098809838294983, + "learning_rate": 2.526589758760316e-05, + "loss": 0.0313, + "step": 32793 + }, + { + "epoch": 0.7687430041550192, + "grad_norm": 0.237198144197464, + "learning_rate": 2.5261003415574326e-05, + "loss": 0.024, + "step": 32794 + }, + { + "epoch": 0.7687664457298241, + "grad_norm": 0.504197359085083, + "learning_rate": 2.5256109649081427e-05, + "loss": 0.1114, + "step": 32795 + }, + { + "epoch": 0.7687898873046292, + "grad_norm": 0.317211776971817, + "learning_rate": 2.525121628815097e-05, + "loss": 0.0532, + "step": 32796 + }, + { + "epoch": 0.7688133288794341, + "grad_norm": 0.37124204635620117, + "learning_rate": 2.5246323332809486e-05, + "loss": 0.0838, + "step": 32797 + }, + { + "epoch": 0.7688367704542391, + "grad_norm": 0.3721838891506195, + "learning_rate": 2.5241430783083577e-05, + "loss": 0.0534, + "step": 32798 + }, + { + "epoch": 0.7688602120290441, + "grad_norm": 0.14667938649654388, + "learning_rate": 2.5236538638999774e-05, + "loss": 0.0321, + "step": 32799 + }, + { + "epoch": 0.7688836536038491, + "grad_norm": 0.6184219717979431, + "learning_rate": 2.5231646900584616e-05, + "loss": 0.0818, + "step": 32800 + }, + { + "epoch": 0.7689070951786541, + "grad_norm": 0.47462937235832214, + "learning_rate": 2.5226755567864634e-05, + "loss": 0.0945, + "step": 32801 + }, + { + "epoch": 0.7689305367534591, + "grad_norm": 0.2719367742538452, + "learning_rate": 2.522186464086639e-05, + "loss": 0.065, + "step": 32802 + }, + { + "epoch": 0.7689539783282641, + "grad_norm": 0.43263137340545654, + "learning_rate": 2.5216974119616378e-05, + "loss": 0.0679, + "step": 32803 + }, + { + "epoch": 0.7689774199030691, + "grad_norm": 0.45957353711128235, + "learning_rate": 2.52120840041412e-05, + "loss": 0.5543, + "step": 32804 + }, + { + "epoch": 0.769000861477874, + "grad_norm": 0.26355311274528503, + "learning_rate": 2.520719429446735e-05, + "loss": 0.0222, + "step": 32805 + }, + { + "epoch": 0.7690243030526791, + "grad_norm": 0.5034597516059875, + "learning_rate": 2.520230499062134e-05, + "loss": 0.5501, + "step": 32806 + }, + { + "epoch": 0.769047744627484, + "grad_norm": 0.5947183966636658, + "learning_rate": 2.5197416092629755e-05, + "loss": 0.1079, + "step": 32807 + }, + { + "epoch": 0.7690711862022891, + "grad_norm": 0.5326249003410339, + "learning_rate": 2.51925276005191e-05, + "loss": 0.0711, + "step": 32808 + }, + { + "epoch": 0.769094627777094, + "grad_norm": 0.44106993079185486, + "learning_rate": 2.5187639514315887e-05, + "loss": 0.0516, + "step": 32809 + }, + { + "epoch": 0.7691180693518991, + "grad_norm": 0.3632589280605316, + "learning_rate": 2.5182751834046612e-05, + "loss": 0.0601, + "step": 32810 + }, + { + "epoch": 0.769141510926704, + "grad_norm": 0.18234287202358246, + "learning_rate": 2.5177864559737874e-05, + "loss": 0.0406, + "step": 32811 + }, + { + "epoch": 0.7691649525015091, + "grad_norm": 0.38955721259117126, + "learning_rate": 2.5172977691416134e-05, + "loss": 0.061, + "step": 32812 + }, + { + "epoch": 0.769188394076314, + "grad_norm": 0.5510233640670776, + "learning_rate": 2.5168091229107927e-05, + "loss": 0.1128, + "step": 32813 + }, + { + "epoch": 0.7692118356511191, + "grad_norm": 0.4390137493610382, + "learning_rate": 2.5163205172839753e-05, + "loss": 0.0589, + "step": 32814 + }, + { + "epoch": 0.769235277225924, + "grad_norm": 0.525722086429596, + "learning_rate": 2.5158319522638107e-05, + "loss": 0.0577, + "step": 32815 + }, + { + "epoch": 0.769258718800729, + "grad_norm": 0.2847917675971985, + "learning_rate": 2.5153434278529554e-05, + "loss": 0.0232, + "step": 32816 + }, + { + "epoch": 0.769282160375534, + "grad_norm": 0.18498928844928741, + "learning_rate": 2.514854944054057e-05, + "loss": 0.0376, + "step": 32817 + }, + { + "epoch": 0.769305601950339, + "grad_norm": 0.558309018611908, + "learning_rate": 2.5143665008697638e-05, + "loss": 0.0919, + "step": 32818 + }, + { + "epoch": 0.769329043525144, + "grad_norm": 0.2586272358894348, + "learning_rate": 2.5138780983027298e-05, + "loss": 0.0176, + "step": 32819 + }, + { + "epoch": 0.769352485099949, + "grad_norm": 0.33638739585876465, + "learning_rate": 2.5133897363556057e-05, + "loss": 0.0577, + "step": 32820 + }, + { + "epoch": 0.769375926674754, + "grad_norm": 0.49964550137519836, + "learning_rate": 2.5129014150310382e-05, + "loss": 0.459, + "step": 32821 + }, + { + "epoch": 0.769399368249559, + "grad_norm": 0.298303484916687, + "learning_rate": 2.512413134331676e-05, + "loss": 0.0436, + "step": 32822 + }, + { + "epoch": 0.769422809824364, + "grad_norm": 0.3538145124912262, + "learning_rate": 2.5119248942601726e-05, + "loss": 0.0233, + "step": 32823 + }, + { + "epoch": 0.769446251399169, + "grad_norm": 0.48125553131103516, + "learning_rate": 2.511436694819176e-05, + "loss": 0.0894, + "step": 32824 + }, + { + "epoch": 0.7694696929739739, + "grad_norm": 0.9412350654602051, + "learning_rate": 2.510948536011334e-05, + "loss": 0.0629, + "step": 32825 + }, + { + "epoch": 0.769493134548779, + "grad_norm": 0.418941468000412, + "learning_rate": 2.5104604178392953e-05, + "loss": 0.0709, + "step": 32826 + }, + { + "epoch": 0.7695165761235839, + "grad_norm": 0.478388249874115, + "learning_rate": 2.5099723403057107e-05, + "loss": 0.0422, + "step": 32827 + }, + { + "epoch": 0.769540017698389, + "grad_norm": 0.21836259961128235, + "learning_rate": 2.509484303413222e-05, + "loss": 0.0136, + "step": 32828 + }, + { + "epoch": 0.769563459273194, + "grad_norm": 0.17212705314159393, + "learning_rate": 2.5089963071644863e-05, + "loss": 0.0454, + "step": 32829 + }, + { + "epoch": 0.769586900847999, + "grad_norm": 0.2594846487045288, + "learning_rate": 2.5085083515621442e-05, + "loss": 0.031, + "step": 32830 + }, + { + "epoch": 0.769610342422804, + "grad_norm": 0.19120661914348602, + "learning_rate": 2.508020436608849e-05, + "loss": 0.0412, + "step": 32831 + }, + { + "epoch": 0.769633783997609, + "grad_norm": 0.4746236801147461, + "learning_rate": 2.5075325623072454e-05, + "loss": 0.049, + "step": 32832 + }, + { + "epoch": 0.769657225572414, + "grad_norm": 0.5801844596862793, + "learning_rate": 2.5070447286599806e-05, + "loss": 0.4421, + "step": 32833 + }, + { + "epoch": 0.7696806671472189, + "grad_norm": 0.19069123268127441, + "learning_rate": 2.5065569356697027e-05, + "loss": 0.0209, + "step": 32834 + }, + { + "epoch": 0.769704108722024, + "grad_norm": 0.05571182072162628, + "learning_rate": 2.5060691833390537e-05, + "loss": 0.0115, + "step": 32835 + }, + { + "epoch": 0.7697275502968289, + "grad_norm": 0.46258246898651123, + "learning_rate": 2.5055814716706872e-05, + "loss": 0.6601, + "step": 32836 + }, + { + "epoch": 0.769750991871634, + "grad_norm": 0.4582741856575012, + "learning_rate": 2.5050938006672463e-05, + "loss": 0.0724, + "step": 32837 + }, + { + "epoch": 0.7697744334464389, + "grad_norm": 0.5326187014579773, + "learning_rate": 2.5046061703313785e-05, + "loss": 0.0335, + "step": 32838 + }, + { + "epoch": 0.769797875021244, + "grad_norm": 0.11488323658704758, + "learning_rate": 2.5041185806657264e-05, + "loss": 0.0162, + "step": 32839 + }, + { + "epoch": 0.7698213165960489, + "grad_norm": 0.37201160192489624, + "learning_rate": 2.5036310316729384e-05, + "loss": 0.1048, + "step": 32840 + }, + { + "epoch": 0.769844758170854, + "grad_norm": 0.4066162109375, + "learning_rate": 2.5031435233556567e-05, + "loss": 0.0666, + "step": 32841 + }, + { + "epoch": 0.7698681997456589, + "grad_norm": 0.28951841592788696, + "learning_rate": 2.5026560557165292e-05, + "loss": 0.0328, + "step": 32842 + }, + { + "epoch": 0.769891641320464, + "grad_norm": 0.654625415802002, + "learning_rate": 2.5021686287582024e-05, + "loss": 0.5131, + "step": 32843 + }, + { + "epoch": 0.7699150828952689, + "grad_norm": 0.3390769958496094, + "learning_rate": 2.5016812424833204e-05, + "loss": 0.0625, + "step": 32844 + }, + { + "epoch": 0.7699385244700739, + "grad_norm": 0.6446809768676758, + "learning_rate": 2.501193896894527e-05, + "loss": 0.651, + "step": 32845 + }, + { + "epoch": 0.7699619660448789, + "grad_norm": 0.13731008768081665, + "learning_rate": 2.500706591994466e-05, + "loss": 0.0161, + "step": 32846 + }, + { + "epoch": 0.7699854076196839, + "grad_norm": 0.32472118735313416, + "learning_rate": 2.5002193277857822e-05, + "loss": 0.0447, + "step": 32847 + }, + { + "epoch": 0.7700088491944889, + "grad_norm": 0.15296316146850586, + "learning_rate": 2.499732104271115e-05, + "loss": 0.0162, + "step": 32848 + }, + { + "epoch": 0.7700322907692939, + "grad_norm": 0.3014877140522003, + "learning_rate": 2.4992449214531167e-05, + "loss": 0.0473, + "step": 32849 + }, + { + "epoch": 0.7700557323440989, + "grad_norm": 0.03649521619081497, + "learning_rate": 2.4987577793344263e-05, + "loss": 0.0028, + "step": 32850 + }, + { + "epoch": 0.7700791739189039, + "grad_norm": 0.3850228190422058, + "learning_rate": 2.4982706779176867e-05, + "loss": 0.0306, + "step": 32851 + }, + { + "epoch": 0.7701026154937088, + "grad_norm": 0.32239893078804016, + "learning_rate": 2.4977836172055412e-05, + "loss": 0.0531, + "step": 32852 + }, + { + "epoch": 0.7701260570685139, + "grad_norm": 0.4720321297645569, + "learning_rate": 2.4972965972006303e-05, + "loss": 0.057, + "step": 32853 + }, + { + "epoch": 0.7701494986433188, + "grad_norm": 0.336976021528244, + "learning_rate": 2.496809617905601e-05, + "loss": 0.0522, + "step": 32854 + }, + { + "epoch": 0.7701729402181239, + "grad_norm": 0.2597077488899231, + "learning_rate": 2.4963226793230908e-05, + "loss": 0.038, + "step": 32855 + }, + { + "epoch": 0.7701963817929288, + "grad_norm": 0.6116521954536438, + "learning_rate": 2.4958357814557488e-05, + "loss": 0.0466, + "step": 32856 + }, + { + "epoch": 0.7702198233677339, + "grad_norm": 0.47966158390045166, + "learning_rate": 2.4953489243062122e-05, + "loss": 0.0836, + "step": 32857 + }, + { + "epoch": 0.7702432649425388, + "grad_norm": 0.09333809465169907, + "learning_rate": 2.494862107877123e-05, + "loss": 0.0068, + "step": 32858 + }, + { + "epoch": 0.7702667065173439, + "grad_norm": 0.4778541624546051, + "learning_rate": 2.494375332171124e-05, + "loss": 0.0521, + "step": 32859 + }, + { + "epoch": 0.7702901480921488, + "grad_norm": 0.07617451250553131, + "learning_rate": 2.4938885971908523e-05, + "loss": 0.0235, + "step": 32860 + }, + { + "epoch": 0.7703135896669538, + "grad_norm": 0.7370039820671082, + "learning_rate": 2.4934019029389544e-05, + "loss": 0.1578, + "step": 32861 + }, + { + "epoch": 0.7703370312417588, + "grad_norm": 0.19497613608837128, + "learning_rate": 2.4929152494180686e-05, + "loss": 0.0362, + "step": 32862 + }, + { + "epoch": 0.7703604728165638, + "grad_norm": 0.2745797038078308, + "learning_rate": 2.492428636630837e-05, + "loss": 0.0531, + "step": 32863 + }, + { + "epoch": 0.7703839143913688, + "grad_norm": 0.48837846517562866, + "learning_rate": 2.4919420645798975e-05, + "loss": 0.0737, + "step": 32864 + }, + { + "epoch": 0.7704073559661738, + "grad_norm": 0.40908151865005493, + "learning_rate": 2.4914555332678923e-05, + "loss": 0.113, + "step": 32865 + }, + { + "epoch": 0.7704307975409788, + "grad_norm": 0.10105135291814804, + "learning_rate": 2.4909690426974576e-05, + "loss": 0.0166, + "step": 32866 + }, + { + "epoch": 0.7704542391157838, + "grad_norm": 0.25340160727500916, + "learning_rate": 2.490482592871235e-05, + "loss": 0.0471, + "step": 32867 + }, + { + "epoch": 0.7704776806905888, + "grad_norm": 0.5412267446517944, + "learning_rate": 2.4899961837918685e-05, + "loss": 0.0639, + "step": 32868 + }, + { + "epoch": 0.7705011222653938, + "grad_norm": 0.1674806773662567, + "learning_rate": 2.4895098154619945e-05, + "loss": 0.025, + "step": 32869 + }, + { + "epoch": 0.7705245638401987, + "grad_norm": 0.4385192394256592, + "learning_rate": 2.48902348788425e-05, + "loss": 0.0585, + "step": 32870 + }, + { + "epoch": 0.7705480054150038, + "grad_norm": 0.5532523393630981, + "learning_rate": 2.488537201061276e-05, + "loss": 0.0541, + "step": 32871 + }, + { + "epoch": 0.7705714469898087, + "grad_norm": 0.4273068308830261, + "learning_rate": 2.4880509549957108e-05, + "loss": 0.041, + "step": 32872 + }, + { + "epoch": 0.7705948885646138, + "grad_norm": 0.22454623878002167, + "learning_rate": 2.487564749690189e-05, + "loss": 0.0213, + "step": 32873 + }, + { + "epoch": 0.7706183301394187, + "grad_norm": 0.11767333000898361, + "learning_rate": 2.4870785851473556e-05, + "loss": 0.0166, + "step": 32874 + }, + { + "epoch": 0.7706417717142238, + "grad_norm": 0.2704617381095886, + "learning_rate": 2.4865924613698444e-05, + "loss": 0.0335, + "step": 32875 + }, + { + "epoch": 0.7706652132890287, + "grad_norm": 0.4206818640232086, + "learning_rate": 2.4861063783602933e-05, + "loss": 0.0834, + "step": 32876 + }, + { + "epoch": 0.7706886548638338, + "grad_norm": 0.3712913691997528, + "learning_rate": 2.4856203361213415e-05, + "loss": 0.0735, + "step": 32877 + }, + { + "epoch": 0.7707120964386387, + "grad_norm": 0.3311333954334259, + "learning_rate": 2.4851343346556244e-05, + "loss": 0.0487, + "step": 32878 + }, + { + "epoch": 0.7707355380134437, + "grad_norm": 0.41106247901916504, + "learning_rate": 2.4846483739657766e-05, + "loss": 0.1067, + "step": 32879 + }, + { + "epoch": 0.7707589795882488, + "grad_norm": 0.11545684933662415, + "learning_rate": 2.4841624540544384e-05, + "loss": 0.0118, + "step": 32880 + }, + { + "epoch": 0.7707824211630537, + "grad_norm": 0.305945485830307, + "learning_rate": 2.4836765749242485e-05, + "loss": 0.0946, + "step": 32881 + }, + { + "epoch": 0.7708058627378588, + "grad_norm": 0.6727683544158936, + "learning_rate": 2.4831907365778416e-05, + "loss": 0.145, + "step": 32882 + }, + { + "epoch": 0.7708293043126637, + "grad_norm": 0.14836207032203674, + "learning_rate": 2.4827049390178526e-05, + "loss": 0.0159, + "step": 32883 + }, + { + "epoch": 0.7708527458874688, + "grad_norm": 0.23068083822727203, + "learning_rate": 2.4822191822469178e-05, + "loss": 0.0314, + "step": 32884 + }, + { + "epoch": 0.7708761874622737, + "grad_norm": 0.13889721035957336, + "learning_rate": 2.4817334662676728e-05, + "loss": 0.0155, + "step": 32885 + }, + { + "epoch": 0.7708996290370788, + "grad_norm": 0.6014368534088135, + "learning_rate": 2.4812477910827515e-05, + "loss": 0.1031, + "step": 32886 + }, + { + "epoch": 0.7709230706118837, + "grad_norm": 0.2578219473361969, + "learning_rate": 2.4807621566947937e-05, + "loss": 0.0386, + "step": 32887 + }, + { + "epoch": 0.7709465121866887, + "grad_norm": 0.40225306153297424, + "learning_rate": 2.4802765631064316e-05, + "loss": 0.1005, + "step": 32888 + }, + { + "epoch": 0.7709699537614937, + "grad_norm": 0.3728142976760864, + "learning_rate": 2.4797910103203005e-05, + "loss": 0.0613, + "step": 32889 + }, + { + "epoch": 0.7709933953362987, + "grad_norm": 0.2406967282295227, + "learning_rate": 2.479305498339034e-05, + "loss": 0.0425, + "step": 32890 + }, + { + "epoch": 0.7710168369111037, + "grad_norm": 0.6692483425140381, + "learning_rate": 2.4788200271652652e-05, + "loss": 0.1209, + "step": 32891 + }, + { + "epoch": 0.7710402784859087, + "grad_norm": 0.40027546882629395, + "learning_rate": 2.4783345968016303e-05, + "loss": 0.048, + "step": 32892 + }, + { + "epoch": 0.7710637200607137, + "grad_norm": 0.5548657178878784, + "learning_rate": 2.4778492072507665e-05, + "loss": 0.1393, + "step": 32893 + }, + { + "epoch": 0.7710871616355187, + "grad_norm": 0.42185625433921814, + "learning_rate": 2.4773638585153047e-05, + "loss": 0.0761, + "step": 32894 + }, + { + "epoch": 0.7711106032103237, + "grad_norm": 0.9203552603721619, + "learning_rate": 2.4768785505978775e-05, + "loss": 0.6941, + "step": 32895 + }, + { + "epoch": 0.7711340447851287, + "grad_norm": 0.224221870303154, + "learning_rate": 2.4763932835011184e-05, + "loss": 0.0402, + "step": 32896 + }, + { + "epoch": 0.7711574863599336, + "grad_norm": 0.2019769847393036, + "learning_rate": 2.4759080572276615e-05, + "loss": 0.0352, + "step": 32897 + }, + { + "epoch": 0.7711809279347387, + "grad_norm": 0.24456541240215302, + "learning_rate": 2.4754228717801365e-05, + "loss": 0.0699, + "step": 32898 + }, + { + "epoch": 0.7712043695095436, + "grad_norm": 0.28702083230018616, + "learning_rate": 2.4749377271611805e-05, + "loss": 0.0223, + "step": 32899 + }, + { + "epoch": 0.7712278110843487, + "grad_norm": 0.3598672151565552, + "learning_rate": 2.4744526233734243e-05, + "loss": 0.0615, + "step": 32900 + }, + { + "epoch": 0.7712512526591536, + "grad_norm": 0.47581249475479126, + "learning_rate": 2.4739675604195013e-05, + "loss": 0.0514, + "step": 32901 + }, + { + "epoch": 0.7712746942339587, + "grad_norm": 0.5586250424385071, + "learning_rate": 2.47348253830204e-05, + "loss": 0.103, + "step": 32902 + }, + { + "epoch": 0.7712981358087636, + "grad_norm": 0.3847110867500305, + "learning_rate": 2.472997557023672e-05, + "loss": 0.0485, + "step": 32903 + }, + { + "epoch": 0.7713215773835687, + "grad_norm": 0.20283643901348114, + "learning_rate": 2.4725126165870338e-05, + "loss": 0.0493, + "step": 32904 + }, + { + "epoch": 0.7713450189583736, + "grad_norm": 0.12508106231689453, + "learning_rate": 2.4720277169947503e-05, + "loss": 0.0142, + "step": 32905 + }, + { + "epoch": 0.7713684605331786, + "grad_norm": 0.107810840010643, + "learning_rate": 2.4715428582494594e-05, + "loss": 0.0175, + "step": 32906 + }, + { + "epoch": 0.7713919021079836, + "grad_norm": 0.3405168950557709, + "learning_rate": 2.4710580403537886e-05, + "loss": 0.0394, + "step": 32907 + }, + { + "epoch": 0.7714153436827886, + "grad_norm": 0.6508251428604126, + "learning_rate": 2.4705732633103684e-05, + "loss": 0.1878, + "step": 32908 + }, + { + "epoch": 0.7714387852575936, + "grad_norm": 0.33597704768180847, + "learning_rate": 2.4700885271218298e-05, + "loss": 0.0859, + "step": 32909 + }, + { + "epoch": 0.7714622268323986, + "grad_norm": 0.6690005660057068, + "learning_rate": 2.4696038317908022e-05, + "loss": 0.122, + "step": 32910 + }, + { + "epoch": 0.7714856684072036, + "grad_norm": 0.3044871389865875, + "learning_rate": 2.469119177319913e-05, + "loss": 0.0533, + "step": 32911 + }, + { + "epoch": 0.7715091099820086, + "grad_norm": 0.18312284350395203, + "learning_rate": 2.468634563711798e-05, + "loss": 0.0386, + "step": 32912 + }, + { + "epoch": 0.7715325515568136, + "grad_norm": 0.6466910243034363, + "learning_rate": 2.4681499909690842e-05, + "loss": 0.0953, + "step": 32913 + }, + { + "epoch": 0.7715559931316186, + "grad_norm": 0.4382883906364441, + "learning_rate": 2.4676654590943992e-05, + "loss": 0.0839, + "step": 32914 + }, + { + "epoch": 0.7715794347064235, + "grad_norm": 0.46125006675720215, + "learning_rate": 2.4671809680903712e-05, + "loss": 0.1017, + "step": 32915 + }, + { + "epoch": 0.7716028762812286, + "grad_norm": 0.10066890716552734, + "learning_rate": 2.4666965179596335e-05, + "loss": 0.02, + "step": 32916 + }, + { + "epoch": 0.7716263178560335, + "grad_norm": 0.48513364791870117, + "learning_rate": 2.466212108704812e-05, + "loss": 0.0536, + "step": 32917 + }, + { + "epoch": 0.7716497594308386, + "grad_norm": 0.45647963881492615, + "learning_rate": 2.4657277403285328e-05, + "loss": 0.0919, + "step": 32918 + }, + { + "epoch": 0.7716732010056435, + "grad_norm": 0.2734310030937195, + "learning_rate": 2.46524341283343e-05, + "loss": 0.0313, + "step": 32919 + }, + { + "epoch": 0.7716966425804486, + "grad_norm": 0.2962345480918884, + "learning_rate": 2.464759126222128e-05, + "loss": 0.0357, + "step": 32920 + }, + { + "epoch": 0.7717200841552535, + "grad_norm": 0.37139368057250977, + "learning_rate": 2.464274880497255e-05, + "loss": 0.0512, + "step": 32921 + }, + { + "epoch": 0.7717435257300586, + "grad_norm": 0.43995851278305054, + "learning_rate": 2.4637906756614382e-05, + "loss": 0.1419, + "step": 32922 + }, + { + "epoch": 0.7717669673048635, + "grad_norm": 0.30151084065437317, + "learning_rate": 2.463306511717305e-05, + "loss": 0.0239, + "step": 32923 + }, + { + "epoch": 0.7717904088796685, + "grad_norm": 0.38246768712997437, + "learning_rate": 2.4628223886674806e-05, + "loss": 0.1114, + "step": 32924 + }, + { + "epoch": 0.7718138504544735, + "grad_norm": 0.5148017406463623, + "learning_rate": 2.4623383065145957e-05, + "loss": 0.1176, + "step": 32925 + }, + { + "epoch": 0.7718372920292785, + "grad_norm": 0.12917539477348328, + "learning_rate": 2.4618542652612753e-05, + "loss": 0.0276, + "step": 32926 + }, + { + "epoch": 0.7718607336040835, + "grad_norm": 0.14942115545272827, + "learning_rate": 2.4613702649101432e-05, + "loss": 0.0192, + "step": 32927 + }, + { + "epoch": 0.7718841751788885, + "grad_norm": 0.34963324666023254, + "learning_rate": 2.4608863054638297e-05, + "loss": 0.0833, + "step": 32928 + }, + { + "epoch": 0.7719076167536935, + "grad_norm": 0.42754074931144714, + "learning_rate": 2.4604023869249605e-05, + "loss": 0.0683, + "step": 32929 + }, + { + "epoch": 0.7719310583284985, + "grad_norm": 0.09625530242919922, + "learning_rate": 2.4599185092961563e-05, + "loss": 0.0086, + "step": 32930 + }, + { + "epoch": 0.7719544999033036, + "grad_norm": 0.6266574263572693, + "learning_rate": 2.4594346725800498e-05, + "loss": 0.116, + "step": 32931 + }, + { + "epoch": 0.7719779414781085, + "grad_norm": 0.9496042132377625, + "learning_rate": 2.4589508767792623e-05, + "loss": 0.1059, + "step": 32932 + }, + { + "epoch": 0.7720013830529135, + "grad_norm": 0.09975031763315201, + "learning_rate": 2.4584671218964196e-05, + "loss": 0.0131, + "step": 32933 + }, + { + "epoch": 0.7720248246277185, + "grad_norm": 0.40933358669281006, + "learning_rate": 2.457983407934146e-05, + "loss": 0.0518, + "step": 32934 + }, + { + "epoch": 0.7720482662025235, + "grad_norm": 0.21920610964298248, + "learning_rate": 2.457499734895066e-05, + "loss": 0.0272, + "step": 32935 + }, + { + "epoch": 0.7720717077773285, + "grad_norm": 0.556169331073761, + "learning_rate": 2.457016102781803e-05, + "loss": 0.1002, + "step": 32936 + }, + { + "epoch": 0.7720951493521335, + "grad_norm": 0.1530122011899948, + "learning_rate": 2.456532511596984e-05, + "loss": 0.0176, + "step": 32937 + }, + { + "epoch": 0.7721185909269385, + "grad_norm": 0.5106422305107117, + "learning_rate": 2.4560489613432326e-05, + "loss": 0.1196, + "step": 32938 + }, + { + "epoch": 0.7721420325017435, + "grad_norm": 0.3707784414291382, + "learning_rate": 2.4555654520231685e-05, + "loss": 0.0873, + "step": 32939 + }, + { + "epoch": 0.7721654740765485, + "grad_norm": 0.5945984125137329, + "learning_rate": 2.4550819836394224e-05, + "loss": 0.5217, + "step": 32940 + }, + { + "epoch": 0.7721889156513535, + "grad_norm": 0.11959698051214218, + "learning_rate": 2.4545985561946127e-05, + "loss": 0.025, + "step": 32941 + }, + { + "epoch": 0.7722123572261584, + "grad_norm": 0.5357799530029297, + "learning_rate": 2.4541151696913634e-05, + "loss": 0.0746, + "step": 32942 + }, + { + "epoch": 0.7722357988009635, + "grad_norm": 0.5482131838798523, + "learning_rate": 2.4536318241322943e-05, + "loss": 0.1464, + "step": 32943 + }, + { + "epoch": 0.7722592403757684, + "grad_norm": 0.49135681986808777, + "learning_rate": 2.4531485195200345e-05, + "loss": 0.0379, + "step": 32944 + }, + { + "epoch": 0.7722826819505735, + "grad_norm": 0.5353066325187683, + "learning_rate": 2.4526652558572017e-05, + "loss": 0.0554, + "step": 32945 + }, + { + "epoch": 0.7723061235253784, + "grad_norm": 0.5724666714668274, + "learning_rate": 2.4521820331464207e-05, + "loss": 0.07, + "step": 32946 + }, + { + "epoch": 0.7723295651001835, + "grad_norm": 0.5161126255989075, + "learning_rate": 2.4516988513903115e-05, + "loss": 0.0546, + "step": 32947 + }, + { + "epoch": 0.7723530066749884, + "grad_norm": 0.11139655858278275, + "learning_rate": 2.451215710591498e-05, + "loss": 0.0061, + "step": 32948 + }, + { + "epoch": 0.7723764482497935, + "grad_norm": 0.1810474842786789, + "learning_rate": 2.4507326107525963e-05, + "loss": 0.026, + "step": 32949 + }, + { + "epoch": 0.7723998898245984, + "grad_norm": 0.3151656985282898, + "learning_rate": 2.4502495518762348e-05, + "loss": 0.0402, + "step": 32950 + }, + { + "epoch": 0.7724233313994034, + "grad_norm": 0.2618253529071808, + "learning_rate": 2.449766533965031e-05, + "loss": 0.0469, + "step": 32951 + }, + { + "epoch": 0.7724467729742084, + "grad_norm": 0.3961074948310852, + "learning_rate": 2.4492835570216044e-05, + "loss": 0.0657, + "step": 32952 + }, + { + "epoch": 0.7724702145490134, + "grad_norm": 0.45433294773101807, + "learning_rate": 2.4488006210485793e-05, + "loss": 0.0929, + "step": 32953 + }, + { + "epoch": 0.7724936561238184, + "grad_norm": 0.3234916627407074, + "learning_rate": 2.4483177260485755e-05, + "loss": 0.0627, + "step": 32954 + }, + { + "epoch": 0.7725170976986234, + "grad_norm": 0.15100982785224915, + "learning_rate": 2.4478348720242117e-05, + "loss": 0.0215, + "step": 32955 + }, + { + "epoch": 0.7725405392734284, + "grad_norm": 0.28186434507369995, + "learning_rate": 2.4473520589781042e-05, + "loss": 0.0755, + "step": 32956 + }, + { + "epoch": 0.7725639808482334, + "grad_norm": 0.39496976137161255, + "learning_rate": 2.4468692869128808e-05, + "loss": 0.0696, + "step": 32957 + }, + { + "epoch": 0.7725874224230384, + "grad_norm": 0.2132510393857956, + "learning_rate": 2.446386555831157e-05, + "loss": 0.0434, + "step": 32958 + }, + { + "epoch": 0.7726108639978434, + "grad_norm": 0.5325183272361755, + "learning_rate": 2.4459038657355516e-05, + "loss": 0.0842, + "step": 32959 + }, + { + "epoch": 0.7726343055726483, + "grad_norm": 0.6996211409568787, + "learning_rate": 2.445421216628685e-05, + "loss": 0.182, + "step": 32960 + }, + { + "epoch": 0.7726577471474534, + "grad_norm": 0.1097472533583641, + "learning_rate": 2.444938608513172e-05, + "loss": 0.0172, + "step": 32961 + }, + { + "epoch": 0.7726811887222583, + "grad_norm": 0.34936222434043884, + "learning_rate": 2.4444560413916372e-05, + "loss": 0.0445, + "step": 32962 + }, + { + "epoch": 0.7727046302970634, + "grad_norm": 0.16947539150714874, + "learning_rate": 2.4439735152666963e-05, + "loss": 0.0288, + "step": 32963 + }, + { + "epoch": 0.7727280718718683, + "grad_norm": 0.3612683117389679, + "learning_rate": 2.4434910301409654e-05, + "loss": 0.0471, + "step": 32964 + }, + { + "epoch": 0.7727515134466734, + "grad_norm": 0.5258063673973083, + "learning_rate": 2.443008586017067e-05, + "loss": 0.1316, + "step": 32965 + }, + { + "epoch": 0.7727749550214783, + "grad_norm": 0.4340016543865204, + "learning_rate": 2.4425261828976165e-05, + "loss": 0.0917, + "step": 32966 + }, + { + "epoch": 0.7727983965962834, + "grad_norm": 0.48062947392463684, + "learning_rate": 2.4420438207852315e-05, + "loss": 0.0928, + "step": 32967 + }, + { + "epoch": 0.7728218381710883, + "grad_norm": 0.17706425487995148, + "learning_rate": 2.4415614996825264e-05, + "loss": 0.0509, + "step": 32968 + }, + { + "epoch": 0.7728452797458933, + "grad_norm": 0.4832294285297394, + "learning_rate": 2.441079219592124e-05, + "loss": 0.0224, + "step": 32969 + }, + { + "epoch": 0.7728687213206983, + "grad_norm": 0.2705754339694977, + "learning_rate": 2.4405969805166384e-05, + "loss": 0.0162, + "step": 32970 + }, + { + "epoch": 0.7728921628955033, + "grad_norm": 0.12596383690834045, + "learning_rate": 2.4401147824586856e-05, + "loss": 0.0239, + "step": 32971 + }, + { + "epoch": 0.7729156044703083, + "grad_norm": 0.2966402769088745, + "learning_rate": 2.4396326254208835e-05, + "loss": 0.0753, + "step": 32972 + }, + { + "epoch": 0.7729390460451133, + "grad_norm": 0.08663848787546158, + "learning_rate": 2.4391505094058464e-05, + "loss": 0.0136, + "step": 32973 + }, + { + "epoch": 0.7729624876199183, + "grad_norm": 0.6998372077941895, + "learning_rate": 2.438668434416189e-05, + "loss": 0.1129, + "step": 32974 + }, + { + "epoch": 0.7729859291947233, + "grad_norm": 0.7691296339035034, + "learning_rate": 2.4381864004545317e-05, + "loss": 0.0905, + "step": 32975 + }, + { + "epoch": 0.7730093707695282, + "grad_norm": 0.6048228144645691, + "learning_rate": 2.4377044075234856e-05, + "loss": 0.4974, + "step": 32976 + }, + { + "epoch": 0.7730328123443333, + "grad_norm": 0.3127167820930481, + "learning_rate": 2.4372224556256697e-05, + "loss": 0.0419, + "step": 32977 + }, + { + "epoch": 0.7730562539191382, + "grad_norm": 0.34567365050315857, + "learning_rate": 2.436740544763698e-05, + "loss": 0.0914, + "step": 32978 + }, + { + "epoch": 0.7730796954939433, + "grad_norm": 0.6844581961631775, + "learning_rate": 2.436258674940185e-05, + "loss": 0.1224, + "step": 32979 + }, + { + "epoch": 0.7731031370687482, + "grad_norm": 0.4947446882724762, + "learning_rate": 2.4357768461577447e-05, + "loss": 0.0729, + "step": 32980 + }, + { + "epoch": 0.7731265786435533, + "grad_norm": 0.46475860476493835, + "learning_rate": 2.4352950584189895e-05, + "loss": 0.1018, + "step": 32981 + }, + { + "epoch": 0.7731500202183583, + "grad_norm": 0.5059689283370972, + "learning_rate": 2.434813311726538e-05, + "loss": 0.0628, + "step": 32982 + }, + { + "epoch": 0.7731734617931633, + "grad_norm": 0.36766019463539124, + "learning_rate": 2.4343316060830025e-05, + "loss": 0.0574, + "step": 32983 + }, + { + "epoch": 0.7731969033679683, + "grad_norm": 0.12913747131824493, + "learning_rate": 2.4338499414909966e-05, + "loss": 0.0259, + "step": 32984 + }, + { + "epoch": 0.7732203449427733, + "grad_norm": 0.5160906910896301, + "learning_rate": 2.433368317953134e-05, + "loss": 0.1147, + "step": 32985 + }, + { + "epoch": 0.7732437865175783, + "grad_norm": 0.45223507285118103, + "learning_rate": 2.4328867354720275e-05, + "loss": 0.0863, + "step": 32986 + }, + { + "epoch": 0.7732672280923832, + "grad_norm": 0.5111016631126404, + "learning_rate": 2.432405194050287e-05, + "loss": 0.1041, + "step": 32987 + }, + { + "epoch": 0.7732906696671883, + "grad_norm": 0.5744023323059082, + "learning_rate": 2.43192369369053e-05, + "loss": 0.0943, + "step": 32988 + }, + { + "epoch": 0.7733141112419932, + "grad_norm": 0.2450990080833435, + "learning_rate": 2.4314422343953692e-05, + "loss": 0.0435, + "step": 32989 + }, + { + "epoch": 0.7733375528167983, + "grad_norm": 0.8468413352966309, + "learning_rate": 2.4309608161674158e-05, + "loss": 0.0403, + "step": 32990 + }, + { + "epoch": 0.7733609943916032, + "grad_norm": 0.43308329582214355, + "learning_rate": 2.4304794390092823e-05, + "loss": 0.5284, + "step": 32991 + }, + { + "epoch": 0.7733844359664083, + "grad_norm": 0.4341014325618744, + "learning_rate": 2.429998102923581e-05, + "loss": 0.106, + "step": 32992 + }, + { + "epoch": 0.7734078775412132, + "grad_norm": 0.25123101472854614, + "learning_rate": 2.4295168079129215e-05, + "loss": 0.0448, + "step": 32993 + }, + { + "epoch": 0.7734313191160183, + "grad_norm": 0.15637865662574768, + "learning_rate": 2.4290355539799136e-05, + "loss": 0.0113, + "step": 32994 + }, + { + "epoch": 0.7734547606908232, + "grad_norm": 0.5796982645988464, + "learning_rate": 2.428554341127176e-05, + "loss": 0.0654, + "step": 32995 + }, + { + "epoch": 0.7734782022656282, + "grad_norm": 0.5267210006713867, + "learning_rate": 2.4280731693573143e-05, + "loss": 0.1084, + "step": 32996 + }, + { + "epoch": 0.7735016438404332, + "grad_norm": 0.17329050600528717, + "learning_rate": 2.4275920386729413e-05, + "loss": 0.0293, + "step": 32997 + }, + { + "epoch": 0.7735250854152382, + "grad_norm": 0.3445665240287781, + "learning_rate": 2.4271109490766665e-05, + "loss": 0.0751, + "step": 32998 + }, + { + "epoch": 0.7735485269900432, + "grad_norm": 0.1546470671892166, + "learning_rate": 2.426629900571097e-05, + "loss": 0.0246, + "step": 32999 + }, + { + "epoch": 0.7735719685648482, + "grad_norm": 0.2972759008407593, + "learning_rate": 2.4261488931588504e-05, + "loss": 0.0237, + "step": 33000 + }, + { + "epoch": 0.7735954101396532, + "grad_norm": 0.8174270987510681, + "learning_rate": 2.4256679268425297e-05, + "loss": 0.2164, + "step": 33001 + }, + { + "epoch": 0.7736188517144582, + "grad_norm": 0.6274361610412598, + "learning_rate": 2.4251870016247503e-05, + "loss": 0.6272, + "step": 33002 + }, + { + "epoch": 0.7736422932892631, + "grad_norm": 0.39241111278533936, + "learning_rate": 2.42470611750812e-05, + "loss": 0.0807, + "step": 33003 + }, + { + "epoch": 0.7736657348640682, + "grad_norm": 0.09997636824846268, + "learning_rate": 2.4242252744952465e-05, + "loss": 0.0114, + "step": 33004 + }, + { + "epoch": 0.7736891764388731, + "grad_norm": 0.37867671251296997, + "learning_rate": 2.4237444725887405e-05, + "loss": 0.0514, + "step": 33005 + }, + { + "epoch": 0.7737126180136782, + "grad_norm": 0.42066827416419983, + "learning_rate": 2.4232637117912062e-05, + "loss": 0.0733, + "step": 33006 + }, + { + "epoch": 0.7737360595884831, + "grad_norm": 0.25287672877311707, + "learning_rate": 2.4227829921052593e-05, + "loss": 0.0425, + "step": 33007 + }, + { + "epoch": 0.7737595011632882, + "grad_norm": 0.4482022225856781, + "learning_rate": 2.4223023135335053e-05, + "loss": 0.0516, + "step": 33008 + }, + { + "epoch": 0.7737829427380931, + "grad_norm": 0.5765380859375, + "learning_rate": 2.4218216760785517e-05, + "loss": 0.063, + "step": 33009 + }, + { + "epoch": 0.7738063843128982, + "grad_norm": 0.9618842005729675, + "learning_rate": 2.421341079743007e-05, + "loss": 0.141, + "step": 33010 + }, + { + "epoch": 0.7738298258877031, + "grad_norm": 0.21252669394016266, + "learning_rate": 2.420860524529478e-05, + "loss": 0.0227, + "step": 33011 + }, + { + "epoch": 0.7738532674625082, + "grad_norm": 0.2875346541404724, + "learning_rate": 2.4203800104405705e-05, + "loss": 0.0605, + "step": 33012 + }, + { + "epoch": 0.7738767090373131, + "grad_norm": 0.2762441337108612, + "learning_rate": 2.4198995374788948e-05, + "loss": 0.0574, + "step": 33013 + }, + { + "epoch": 0.7739001506121181, + "grad_norm": 0.4242306649684906, + "learning_rate": 2.4194191056470594e-05, + "loss": 0.073, + "step": 33014 + }, + { + "epoch": 0.7739235921869231, + "grad_norm": 0.13218826055526733, + "learning_rate": 2.4189387149476694e-05, + "loss": 0.0197, + "step": 33015 + }, + { + "epoch": 0.7739470337617281, + "grad_norm": 0.42193207144737244, + "learning_rate": 2.4184583653833303e-05, + "loss": 0.0516, + "step": 33016 + }, + { + "epoch": 0.7739704753365331, + "grad_norm": 0.24439454078674316, + "learning_rate": 2.4179780569566503e-05, + "loss": 0.0222, + "step": 33017 + }, + { + "epoch": 0.7739939169113381, + "grad_norm": 0.12440554797649384, + "learning_rate": 2.4174977896702345e-05, + "loss": 0.0195, + "step": 33018 + }, + { + "epoch": 0.7740173584861431, + "grad_norm": 0.09614025056362152, + "learning_rate": 2.4170175635266856e-05, + "loss": 0.0083, + "step": 33019 + }, + { + "epoch": 0.7740408000609481, + "grad_norm": 0.307078093290329, + "learning_rate": 2.416537378528615e-05, + "loss": 0.0342, + "step": 33020 + }, + { + "epoch": 0.774064241635753, + "grad_norm": 0.28855788707733154, + "learning_rate": 2.416057234678627e-05, + "loss": 0.0329, + "step": 33021 + }, + { + "epoch": 0.7740876832105581, + "grad_norm": 0.2733462154865265, + "learning_rate": 2.415577131979325e-05, + "loss": 0.047, + "step": 33022 + }, + { + "epoch": 0.774111124785363, + "grad_norm": 0.2537898123264313, + "learning_rate": 2.4150970704333155e-05, + "loss": 0.0288, + "step": 33023 + }, + { + "epoch": 0.7741345663601681, + "grad_norm": 0.5252448916435242, + "learning_rate": 2.414617050043201e-05, + "loss": 0.0779, + "step": 33024 + }, + { + "epoch": 0.774158007934973, + "grad_norm": 0.5885856747627258, + "learning_rate": 2.414137070811586e-05, + "loss": 0.1114, + "step": 33025 + }, + { + "epoch": 0.7741814495097781, + "grad_norm": 0.12050052732229233, + "learning_rate": 2.4136571327410763e-05, + "loss": 0.017, + "step": 33026 + }, + { + "epoch": 0.774204891084583, + "grad_norm": 0.38254058361053467, + "learning_rate": 2.4131772358342796e-05, + "loss": 0.0652, + "step": 33027 + }, + { + "epoch": 0.7742283326593881, + "grad_norm": 0.13402722775936127, + "learning_rate": 2.4126973800937957e-05, + "loss": 0.0172, + "step": 33028 + }, + { + "epoch": 0.774251774234193, + "grad_norm": 0.12493336945772171, + "learning_rate": 2.41221756552223e-05, + "loss": 0.0179, + "step": 33029 + }, + { + "epoch": 0.774275215808998, + "grad_norm": 0.5952774286270142, + "learning_rate": 2.411737792122184e-05, + "loss": 0.1362, + "step": 33030 + }, + { + "epoch": 0.774298657383803, + "grad_norm": 0.43867266178131104, + "learning_rate": 2.411258059896263e-05, + "loss": 0.065, + "step": 33031 + }, + { + "epoch": 0.774322098958608, + "grad_norm": 0.271820068359375, + "learning_rate": 2.4107783688470663e-05, + "loss": 0.2464, + "step": 33032 + }, + { + "epoch": 0.774345540533413, + "grad_norm": 0.42474088072776794, + "learning_rate": 2.410298718977202e-05, + "loss": 0.0751, + "step": 33033 + }, + { + "epoch": 0.774368982108218, + "grad_norm": 0.3330010771751404, + "learning_rate": 2.409819110289271e-05, + "loss": 0.0407, + "step": 33034 + }, + { + "epoch": 0.7743924236830231, + "grad_norm": 0.14121675491333008, + "learning_rate": 2.4093395427858732e-05, + "loss": 0.0197, + "step": 33035 + }, + { + "epoch": 0.774415865257828, + "grad_norm": 0.3880805969238281, + "learning_rate": 2.408860016469614e-05, + "loss": 0.0483, + "step": 33036 + }, + { + "epoch": 0.7744393068326331, + "grad_norm": 0.682062566280365, + "learning_rate": 2.4083805313430907e-05, + "loss": 0.7396, + "step": 33037 + }, + { + "epoch": 0.774462748407438, + "grad_norm": 0.08961552381515503, + "learning_rate": 2.4079010874089113e-05, + "loss": 0.0137, + "step": 33038 + }, + { + "epoch": 0.7744861899822431, + "grad_norm": 0.6317614316940308, + "learning_rate": 2.4074216846696707e-05, + "loss": 0.0802, + "step": 33039 + }, + { + "epoch": 0.774509631557048, + "grad_norm": 0.2872220277786255, + "learning_rate": 2.4069423231279764e-05, + "loss": 0.055, + "step": 33040 + }, + { + "epoch": 0.774533073131853, + "grad_norm": 0.2459275871515274, + "learning_rate": 2.4064630027864266e-05, + "loss": 0.0181, + "step": 33041 + }, + { + "epoch": 0.774556514706658, + "grad_norm": 0.29317134618759155, + "learning_rate": 2.4059837236476224e-05, + "loss": 0.0456, + "step": 33042 + }, + { + "epoch": 0.774579956281463, + "grad_norm": 0.21296726167201996, + "learning_rate": 2.4055044857141628e-05, + "loss": 0.0178, + "step": 33043 + }, + { + "epoch": 0.774603397856268, + "grad_norm": 0.6088600754737854, + "learning_rate": 2.4050252889886483e-05, + "loss": 0.4978, + "step": 33044 + }, + { + "epoch": 0.774626839431073, + "grad_norm": 0.2945340871810913, + "learning_rate": 2.4045461334736818e-05, + "loss": 0.0409, + "step": 33045 + }, + { + "epoch": 0.774650281005878, + "grad_norm": 0.5717827677726746, + "learning_rate": 2.4040670191718616e-05, + "loss": 0.0867, + "step": 33046 + }, + { + "epoch": 0.774673722580683, + "grad_norm": 0.6308701038360596, + "learning_rate": 2.403587946085788e-05, + "loss": 0.0987, + "step": 33047 + }, + { + "epoch": 0.774697164155488, + "grad_norm": 0.35694852471351624, + "learning_rate": 2.4031089142180587e-05, + "loss": 0.0632, + "step": 33048 + }, + { + "epoch": 0.774720605730293, + "grad_norm": 0.43827447295188904, + "learning_rate": 2.4026299235712723e-05, + "loss": 0.0896, + "step": 33049 + }, + { + "epoch": 0.7747440473050979, + "grad_norm": 0.8093343377113342, + "learning_rate": 2.4021509741480318e-05, + "loss": 0.1462, + "step": 33050 + }, + { + "epoch": 0.774767488879903, + "grad_norm": 0.6674714684486389, + "learning_rate": 2.4016720659509307e-05, + "loss": 0.4611, + "step": 33051 + }, + { + "epoch": 0.7747909304547079, + "grad_norm": 0.13274121284484863, + "learning_rate": 2.4011931989825742e-05, + "loss": 0.0244, + "step": 33052 + }, + { + "epoch": 0.774814372029513, + "grad_norm": 0.6325022578239441, + "learning_rate": 2.4007143732455573e-05, + "loss": 0.0694, + "step": 33053 + }, + { + "epoch": 0.7748378136043179, + "grad_norm": 0.27490416169166565, + "learning_rate": 2.4002355887424766e-05, + "loss": 0.0454, + "step": 33054 + }, + { + "epoch": 0.774861255179123, + "grad_norm": 0.6153264045715332, + "learning_rate": 2.3997568454759322e-05, + "loss": 0.1076, + "step": 33055 + }, + { + "epoch": 0.7748846967539279, + "grad_norm": 0.31247150897979736, + "learning_rate": 2.39927814344852e-05, + "loss": 0.0196, + "step": 33056 + }, + { + "epoch": 0.774908138328733, + "grad_norm": 0.47898048162460327, + "learning_rate": 2.398799482662837e-05, + "loss": 0.0813, + "step": 33057 + }, + { + "epoch": 0.7749315799035379, + "grad_norm": 0.6541131734848022, + "learning_rate": 2.3983208631214837e-05, + "loss": 0.5781, + "step": 33058 + }, + { + "epoch": 0.7749550214783429, + "grad_norm": 0.04768432304263115, + "learning_rate": 2.3978422848270545e-05, + "loss": 0.004, + "step": 33059 + }, + { + "epoch": 0.7749784630531479, + "grad_norm": 0.3932088315486908, + "learning_rate": 2.3973637477821475e-05, + "loss": 0.3268, + "step": 33060 + }, + { + "epoch": 0.7750019046279529, + "grad_norm": 0.6027613282203674, + "learning_rate": 2.396885251989356e-05, + "loss": 0.0943, + "step": 33061 + }, + { + "epoch": 0.7750253462027579, + "grad_norm": 0.5130828619003296, + "learning_rate": 2.3964067974512815e-05, + "loss": 0.0953, + "step": 33062 + }, + { + "epoch": 0.7750487877775629, + "grad_norm": 0.3869446814060211, + "learning_rate": 2.395928384170517e-05, + "loss": 0.0593, + "step": 33063 + }, + { + "epoch": 0.7750722293523679, + "grad_norm": 0.4539617896080017, + "learning_rate": 2.395450012149656e-05, + "loss": 0.0567, + "step": 33064 + }, + { + "epoch": 0.7750956709271729, + "grad_norm": 0.606184720993042, + "learning_rate": 2.3949716813913003e-05, + "loss": 0.0719, + "step": 33065 + }, + { + "epoch": 0.7751191125019778, + "grad_norm": 0.1966056525707245, + "learning_rate": 2.394493391898043e-05, + "loss": 0.0415, + "step": 33066 + }, + { + "epoch": 0.7751425540767829, + "grad_norm": 0.3648355305194855, + "learning_rate": 2.3940151436724767e-05, + "loss": 0.0501, + "step": 33067 + }, + { + "epoch": 0.7751659956515878, + "grad_norm": 0.2637077867984772, + "learning_rate": 2.3935369367172e-05, + "loss": 0.0426, + "step": 33068 + }, + { + "epoch": 0.7751894372263929, + "grad_norm": 0.11252820491790771, + "learning_rate": 2.393058771034804e-05, + "loss": 0.0149, + "step": 33069 + }, + { + "epoch": 0.7752128788011978, + "grad_norm": 0.15639708936214447, + "learning_rate": 2.3925806466278834e-05, + "loss": 0.0208, + "step": 33070 + }, + { + "epoch": 0.7752363203760029, + "grad_norm": 0.09084691107273102, + "learning_rate": 2.3921025634990357e-05, + "loss": 0.0092, + "step": 33071 + }, + { + "epoch": 0.7752597619508078, + "grad_norm": 0.45759323239326477, + "learning_rate": 2.391624521650855e-05, + "loss": 0.0669, + "step": 33072 + }, + { + "epoch": 0.7752832035256129, + "grad_norm": 0.2937871515750885, + "learning_rate": 2.3911465210859295e-05, + "loss": 0.0327, + "step": 33073 + }, + { + "epoch": 0.7753066451004178, + "grad_norm": 0.2397400438785553, + "learning_rate": 2.3906685618068603e-05, + "loss": 0.0159, + "step": 33074 + }, + { + "epoch": 0.7753300866752229, + "grad_norm": 0.26379531621932983, + "learning_rate": 2.390190643816237e-05, + "loss": 0.0455, + "step": 33075 + }, + { + "epoch": 0.7753535282500278, + "grad_norm": 0.3189132809638977, + "learning_rate": 2.389712767116653e-05, + "loss": 0.0281, + "step": 33076 + }, + { + "epoch": 0.7753769698248328, + "grad_norm": 0.4259895384311676, + "learning_rate": 2.3892349317107e-05, + "loss": 0.095, + "step": 33077 + }, + { + "epoch": 0.7754004113996378, + "grad_norm": 0.21168451011180878, + "learning_rate": 2.3887571376009732e-05, + "loss": 0.0311, + "step": 33078 + }, + { + "epoch": 0.7754238529744428, + "grad_norm": 0.29201260209083557, + "learning_rate": 2.3882793847900653e-05, + "loss": 0.225, + "step": 33079 + }, + { + "epoch": 0.7754472945492478, + "grad_norm": 0.7828070521354675, + "learning_rate": 2.3878016732805673e-05, + "loss": 0.1199, + "step": 33080 + }, + { + "epoch": 0.7754707361240528, + "grad_norm": 0.4119817018508911, + "learning_rate": 2.387324003075072e-05, + "loss": 0.0729, + "step": 33081 + }, + { + "epoch": 0.7754941776988578, + "grad_norm": 0.09557618945837021, + "learning_rate": 2.3868463741761672e-05, + "loss": 0.023, + "step": 33082 + }, + { + "epoch": 0.7755176192736628, + "grad_norm": 0.33662495017051697, + "learning_rate": 2.3863687865864513e-05, + "loss": 0.046, + "step": 33083 + }, + { + "epoch": 0.7755410608484677, + "grad_norm": 0.22128771245479584, + "learning_rate": 2.385891240308512e-05, + "loss": 0.0494, + "step": 33084 + }, + { + "epoch": 0.7755645024232728, + "grad_norm": 0.2544949948787689, + "learning_rate": 2.385413735344939e-05, + "loss": 0.0445, + "step": 33085 + }, + { + "epoch": 0.7755879439980778, + "grad_norm": 0.4623587131500244, + "learning_rate": 2.384936271698327e-05, + "loss": 0.1027, + "step": 33086 + }, + { + "epoch": 0.7756113855728828, + "grad_norm": 0.7584068775177002, + "learning_rate": 2.384458849371265e-05, + "loss": 0.0786, + "step": 33087 + }, + { + "epoch": 0.7756348271476878, + "grad_norm": 0.38544759154319763, + "learning_rate": 2.3839814683663443e-05, + "loss": 0.0417, + "step": 33088 + }, + { + "epoch": 0.7756582687224928, + "grad_norm": 0.5122594833374023, + "learning_rate": 2.3835041286861504e-05, + "loss": 0.5788, + "step": 33089 + }, + { + "epoch": 0.7756817102972978, + "grad_norm": 0.5938555598258972, + "learning_rate": 2.3830268303332802e-05, + "loss": 0.0708, + "step": 33090 + }, + { + "epoch": 0.7757051518721028, + "grad_norm": 0.31040945649147034, + "learning_rate": 2.3825495733103208e-05, + "loss": 0.0303, + "step": 33091 + }, + { + "epoch": 0.7757285934469078, + "grad_norm": 0.33615127205848694, + "learning_rate": 2.382072357619862e-05, + "loss": 0.0439, + "step": 33092 + }, + { + "epoch": 0.7757520350217127, + "grad_norm": 0.2305091768503189, + "learning_rate": 2.3815951832644923e-05, + "loss": 0.0458, + "step": 33093 + }, + { + "epoch": 0.7757754765965178, + "grad_norm": 0.4939791262149811, + "learning_rate": 2.3811180502468024e-05, + "loss": 0.0553, + "step": 33094 + }, + { + "epoch": 0.7757989181713227, + "grad_norm": 0.5934353470802307, + "learning_rate": 2.3806409585693756e-05, + "loss": 0.486, + "step": 33095 + }, + { + "epoch": 0.7758223597461278, + "grad_norm": 0.09581349045038223, + "learning_rate": 2.3801639082348092e-05, + "loss": 0.0193, + "step": 33096 + }, + { + "epoch": 0.7758458013209327, + "grad_norm": 0.6068215370178223, + "learning_rate": 2.3796868992456878e-05, + "loss": 0.4577, + "step": 33097 + }, + { + "epoch": 0.7758692428957378, + "grad_norm": 0.10275857150554657, + "learning_rate": 2.3792099316045956e-05, + "loss": 0.0236, + "step": 33098 + }, + { + "epoch": 0.7758926844705427, + "grad_norm": 0.3001680374145508, + "learning_rate": 2.378733005314129e-05, + "loss": 0.027, + "step": 33099 + }, + { + "epoch": 0.7759161260453478, + "grad_norm": 0.28738078474998474, + "learning_rate": 2.3782561203768705e-05, + "loss": 0.2902, + "step": 33100 + }, + { + "epoch": 0.7759395676201527, + "grad_norm": 0.6696820855140686, + "learning_rate": 2.377779276795409e-05, + "loss": 0.1082, + "step": 33101 + }, + { + "epoch": 0.7759630091949578, + "grad_norm": 0.3141622245311737, + "learning_rate": 2.377302474572327e-05, + "loss": 0.0387, + "step": 33102 + }, + { + "epoch": 0.7759864507697627, + "grad_norm": 0.34134113788604736, + "learning_rate": 2.3768257137102213e-05, + "loss": 0.0416, + "step": 33103 + }, + { + "epoch": 0.7760098923445677, + "grad_norm": 0.13431546092033386, + "learning_rate": 2.3763489942116723e-05, + "loss": 0.0336, + "step": 33104 + }, + { + "epoch": 0.7760333339193727, + "grad_norm": 0.5023419260978699, + "learning_rate": 2.3758723160792684e-05, + "loss": 0.0728, + "step": 33105 + }, + { + "epoch": 0.7760567754941777, + "grad_norm": 0.1721077263355255, + "learning_rate": 2.3753956793155953e-05, + "loss": 0.0203, + "step": 33106 + }, + { + "epoch": 0.7760802170689827, + "grad_norm": 0.33054623007774353, + "learning_rate": 2.3749190839232393e-05, + "loss": 0.0261, + "step": 33107 + }, + { + "epoch": 0.7761036586437877, + "grad_norm": 0.20632141828536987, + "learning_rate": 2.3744425299047846e-05, + "loss": 0.0329, + "step": 33108 + }, + { + "epoch": 0.7761271002185927, + "grad_norm": 0.39627188444137573, + "learning_rate": 2.373966017262822e-05, + "loss": 0.0711, + "step": 33109 + }, + { + "epoch": 0.7761505417933977, + "grad_norm": 0.6587981581687927, + "learning_rate": 2.3734895459999306e-05, + "loss": 0.1294, + "step": 33110 + }, + { + "epoch": 0.7761739833682026, + "grad_norm": 0.13740013539791107, + "learning_rate": 2.373013116118703e-05, + "loss": 0.0171, + "step": 33111 + }, + { + "epoch": 0.7761974249430077, + "grad_norm": 0.44798174500465393, + "learning_rate": 2.3725367276217215e-05, + "loss": 0.0866, + "step": 33112 + }, + { + "epoch": 0.7762208665178126, + "grad_norm": 0.10103408992290497, + "learning_rate": 2.372060380511568e-05, + "loss": 0.021, + "step": 33113 + }, + { + "epoch": 0.7762443080926177, + "grad_norm": 0.116401806473732, + "learning_rate": 2.3715840747908314e-05, + "loss": 0.0143, + "step": 33114 + }, + { + "epoch": 0.7762677496674226, + "grad_norm": 0.19868330657482147, + "learning_rate": 2.37110781046209e-05, + "loss": 0.0315, + "step": 33115 + }, + { + "epoch": 0.7762911912422277, + "grad_norm": 0.6048821806907654, + "learning_rate": 2.370631587527935e-05, + "loss": 0.6276, + "step": 33116 + }, + { + "epoch": 0.7763146328170326, + "grad_norm": 0.6939118504524231, + "learning_rate": 2.370155405990948e-05, + "loss": 0.0896, + "step": 33117 + }, + { + "epoch": 0.7763380743918377, + "grad_norm": 0.4090295135974884, + "learning_rate": 2.3696792658537127e-05, + "loss": 0.0802, + "step": 33118 + }, + { + "epoch": 0.7763615159666426, + "grad_norm": 0.36251071095466614, + "learning_rate": 2.369203167118812e-05, + "loss": 0.4453, + "step": 33119 + }, + { + "epoch": 0.7763849575414477, + "grad_norm": 0.4346683919429779, + "learning_rate": 2.3687271097888264e-05, + "loss": 0.0638, + "step": 33120 + }, + { + "epoch": 0.7764083991162526, + "grad_norm": 0.3666906952857971, + "learning_rate": 2.368251093866345e-05, + "loss": 0.0657, + "step": 33121 + }, + { + "epoch": 0.7764318406910576, + "grad_norm": 0.5047769546508789, + "learning_rate": 2.3677751193539444e-05, + "loss": 0.0678, + "step": 33122 + }, + { + "epoch": 0.7764552822658626, + "grad_norm": 0.4682651162147522, + "learning_rate": 2.3672991862542138e-05, + "loss": 0.0848, + "step": 33123 + }, + { + "epoch": 0.7764787238406676, + "grad_norm": 0.584696352481842, + "learning_rate": 2.3668232945697323e-05, + "loss": 0.153, + "step": 33124 + }, + { + "epoch": 0.7765021654154726, + "grad_norm": 0.4712836742401123, + "learning_rate": 2.3663474443030832e-05, + "loss": 0.144, + "step": 33125 + }, + { + "epoch": 0.7765256069902776, + "grad_norm": 0.48100122809410095, + "learning_rate": 2.3658716354568468e-05, + "loss": 0.4758, + "step": 33126 + }, + { + "epoch": 0.7765490485650826, + "grad_norm": 0.43611636757850647, + "learning_rate": 2.365395868033602e-05, + "loss": 0.0404, + "step": 33127 + }, + { + "epoch": 0.7765724901398876, + "grad_norm": 0.34757474064826965, + "learning_rate": 2.3649201420359368e-05, + "loss": 0.0234, + "step": 33128 + }, + { + "epoch": 0.7765959317146925, + "grad_norm": 0.5853269100189209, + "learning_rate": 2.3644444574664303e-05, + "loss": 0.5533, + "step": 33129 + }, + { + "epoch": 0.7766193732894976, + "grad_norm": 0.6258560419082642, + "learning_rate": 2.363968814327663e-05, + "loss": 0.1384, + "step": 33130 + }, + { + "epoch": 0.7766428148643025, + "grad_norm": 0.34587299823760986, + "learning_rate": 2.3634932126222153e-05, + "loss": 0.0516, + "step": 33131 + }, + { + "epoch": 0.7766662564391076, + "grad_norm": 0.28390443325042725, + "learning_rate": 2.363017652352668e-05, + "loss": 0.0529, + "step": 33132 + }, + { + "epoch": 0.7766896980139125, + "grad_norm": 0.9968768954277039, + "learning_rate": 2.3625421335215993e-05, + "loss": 0.2047, + "step": 33133 + }, + { + "epoch": 0.7767131395887176, + "grad_norm": 0.0721016600728035, + "learning_rate": 2.362066656131592e-05, + "loss": 0.0106, + "step": 33134 + }, + { + "epoch": 0.7767365811635225, + "grad_norm": 0.19809922575950623, + "learning_rate": 2.3615912201852276e-05, + "loss": 0.0434, + "step": 33135 + }, + { + "epoch": 0.7767600227383276, + "grad_norm": 0.3888848125934601, + "learning_rate": 2.361115825685084e-05, + "loss": 0.1187, + "step": 33136 + }, + { + "epoch": 0.7767834643131326, + "grad_norm": 0.2581871449947357, + "learning_rate": 2.3606404726337416e-05, + "loss": 0.0523, + "step": 33137 + }, + { + "epoch": 0.7768069058879375, + "grad_norm": 0.11301887035369873, + "learning_rate": 2.3601651610337783e-05, + "loss": 0.0234, + "step": 33138 + }, + { + "epoch": 0.7768303474627426, + "grad_norm": 0.39738911390304565, + "learning_rate": 2.359689890887774e-05, + "loss": 0.0846, + "step": 33139 + }, + { + "epoch": 0.7768537890375475, + "grad_norm": 0.7024115324020386, + "learning_rate": 2.3592146621983035e-05, + "loss": 0.1372, + "step": 33140 + }, + { + "epoch": 0.7768772306123526, + "grad_norm": 0.1638713926076889, + "learning_rate": 2.3587394749679525e-05, + "loss": 0.0523, + "step": 33141 + }, + { + "epoch": 0.7769006721871575, + "grad_norm": 0.35578885674476624, + "learning_rate": 2.3582643291992956e-05, + "loss": 0.068, + "step": 33142 + }, + { + "epoch": 0.7769241137619626, + "grad_norm": 0.34781506657600403, + "learning_rate": 2.3577892248949117e-05, + "loss": 0.0319, + "step": 33143 + }, + { + "epoch": 0.7769475553367675, + "grad_norm": 0.3968997299671173, + "learning_rate": 2.357314162057379e-05, + "loss": 0.0514, + "step": 33144 + }, + { + "epoch": 0.7769709969115726, + "grad_norm": 0.28830385208129883, + "learning_rate": 2.3568391406892732e-05, + "loss": 0.057, + "step": 33145 + }, + { + "epoch": 0.7769944384863775, + "grad_norm": 0.4810985326766968, + "learning_rate": 2.356364160793171e-05, + "loss": 0.09, + "step": 33146 + }, + { + "epoch": 0.7770178800611826, + "grad_norm": 0.15110275149345398, + "learning_rate": 2.3558892223716523e-05, + "loss": 0.0198, + "step": 33147 + }, + { + "epoch": 0.7770413216359875, + "grad_norm": 0.7280561327934265, + "learning_rate": 2.3554143254272955e-05, + "loss": 0.1043, + "step": 33148 + }, + { + "epoch": 0.7770647632107925, + "grad_norm": 0.10152097791433334, + "learning_rate": 2.354939469962676e-05, + "loss": 0.0116, + "step": 33149 + }, + { + "epoch": 0.7770882047855975, + "grad_norm": 0.6583475470542908, + "learning_rate": 2.3544646559803696e-05, + "loss": 0.1169, + "step": 33150 + }, + { + "epoch": 0.7771116463604025, + "grad_norm": 0.4988606870174408, + "learning_rate": 2.3539898834829533e-05, + "loss": 0.0644, + "step": 33151 + }, + { + "epoch": 0.7771350879352075, + "grad_norm": 0.13508613407611847, + "learning_rate": 2.3535151524730027e-05, + "loss": 0.0094, + "step": 33152 + }, + { + "epoch": 0.7771585295100125, + "grad_norm": 0.4058132469654083, + "learning_rate": 2.353040462953091e-05, + "loss": 0.06, + "step": 33153 + }, + { + "epoch": 0.7771819710848175, + "grad_norm": 0.5309843420982361, + "learning_rate": 2.3525658149258002e-05, + "loss": 0.1167, + "step": 33154 + }, + { + "epoch": 0.7772054126596225, + "grad_norm": 0.45904621481895447, + "learning_rate": 2.3520912083937018e-05, + "loss": 0.0758, + "step": 33155 + }, + { + "epoch": 0.7772288542344274, + "grad_norm": 0.40421366691589355, + "learning_rate": 2.351616643359371e-05, + "loss": 0.0864, + "step": 33156 + }, + { + "epoch": 0.7772522958092325, + "grad_norm": 0.24047788977622986, + "learning_rate": 2.3511421198253835e-05, + "loss": 0.033, + "step": 33157 + }, + { + "epoch": 0.7772757373840374, + "grad_norm": 0.9788522720336914, + "learning_rate": 2.3506676377943117e-05, + "loss": 0.1047, + "step": 33158 + }, + { + "epoch": 0.7772991789588425, + "grad_norm": 0.48370376229286194, + "learning_rate": 2.3501931972687352e-05, + "loss": 0.0944, + "step": 33159 + }, + { + "epoch": 0.7773226205336474, + "grad_norm": 0.42391836643218994, + "learning_rate": 2.3497187982512225e-05, + "loss": 0.1004, + "step": 33160 + }, + { + "epoch": 0.7773460621084525, + "grad_norm": 0.2552887201309204, + "learning_rate": 2.3492444407443538e-05, + "loss": 0.0158, + "step": 33161 + }, + { + "epoch": 0.7773695036832574, + "grad_norm": 0.30030736327171326, + "learning_rate": 2.348770124750699e-05, + "loss": 0.051, + "step": 33162 + }, + { + "epoch": 0.7773929452580625, + "grad_norm": 0.37250933051109314, + "learning_rate": 2.3482958502728336e-05, + "loss": 0.0397, + "step": 33163 + }, + { + "epoch": 0.7774163868328674, + "grad_norm": 0.717059850692749, + "learning_rate": 2.3478216173133293e-05, + "loss": 0.2242, + "step": 33164 + }, + { + "epoch": 0.7774398284076725, + "grad_norm": 0.19203658401966095, + "learning_rate": 2.3473474258747573e-05, + "loss": 0.0576, + "step": 33165 + }, + { + "epoch": 0.7774632699824774, + "grad_norm": 0.26353421807289124, + "learning_rate": 2.3468732759596967e-05, + "loss": 0.0505, + "step": 33166 + }, + { + "epoch": 0.7774867115572824, + "grad_norm": 0.30752187967300415, + "learning_rate": 2.3463991675707164e-05, + "loss": 0.0425, + "step": 33167 + }, + { + "epoch": 0.7775101531320874, + "grad_norm": 0.30528023838996887, + "learning_rate": 2.34592510071039e-05, + "loss": 0.0248, + "step": 33168 + }, + { + "epoch": 0.7775335947068924, + "grad_norm": 0.22902914881706238, + "learning_rate": 2.3454510753812887e-05, + "loss": 0.0421, + "step": 33169 + }, + { + "epoch": 0.7775570362816974, + "grad_norm": 0.3562028110027313, + "learning_rate": 2.3449770915859824e-05, + "loss": 0.1096, + "step": 33170 + }, + { + "epoch": 0.7775804778565024, + "grad_norm": 0.17300784587860107, + "learning_rate": 2.3445031493270485e-05, + "loss": 0.0278, + "step": 33171 + }, + { + "epoch": 0.7776039194313074, + "grad_norm": 0.1922154724597931, + "learning_rate": 2.3440292486070536e-05, + "loss": 0.0355, + "step": 33172 + }, + { + "epoch": 0.7776273610061124, + "grad_norm": 0.3218775689601898, + "learning_rate": 2.3435553894285732e-05, + "loss": 0.0612, + "step": 33173 + }, + { + "epoch": 0.7776508025809173, + "grad_norm": 0.41108521819114685, + "learning_rate": 2.343081571794177e-05, + "loss": 0.0926, + "step": 33174 + }, + { + "epoch": 0.7776742441557224, + "grad_norm": 0.3910740613937378, + "learning_rate": 2.3426077957064353e-05, + "loss": 0.0693, + "step": 33175 + }, + { + "epoch": 0.7776976857305273, + "grad_norm": 0.519804835319519, + "learning_rate": 2.3421340611679186e-05, + "loss": 0.0315, + "step": 33176 + }, + { + "epoch": 0.7777211273053324, + "grad_norm": 0.8212093114852905, + "learning_rate": 2.341660368181199e-05, + "loss": 0.1379, + "step": 33177 + }, + { + "epoch": 0.7777445688801373, + "grad_norm": 0.34818077087402344, + "learning_rate": 2.341186716748841e-05, + "loss": 0.0541, + "step": 33178 + }, + { + "epoch": 0.7777680104549424, + "grad_norm": 0.38939177989959717, + "learning_rate": 2.3407131068734233e-05, + "loss": 0.0278, + "step": 33179 + }, + { + "epoch": 0.7777914520297473, + "grad_norm": 0.44323477149009705, + "learning_rate": 2.3402395385575114e-05, + "loss": 0.0496, + "step": 33180 + }, + { + "epoch": 0.7778148936045524, + "grad_norm": 0.45840343832969666, + "learning_rate": 2.3397660118036747e-05, + "loss": 0.1062, + "step": 33181 + }, + { + "epoch": 0.7778383351793573, + "grad_norm": 0.28581181168556213, + "learning_rate": 2.339292526614484e-05, + "loss": 0.0332, + "step": 33182 + }, + { + "epoch": 0.7778617767541623, + "grad_norm": 0.40931978821754456, + "learning_rate": 2.338819082992504e-05, + "loss": 0.0519, + "step": 33183 + }, + { + "epoch": 0.7778852183289673, + "grad_norm": 0.29252177476882935, + "learning_rate": 2.3383456809403092e-05, + "loss": 0.0397, + "step": 33184 + }, + { + "epoch": 0.7779086599037723, + "grad_norm": 0.1174977719783783, + "learning_rate": 2.3378723204604647e-05, + "loss": 0.0517, + "step": 33185 + }, + { + "epoch": 0.7779321014785773, + "grad_norm": 0.2972216010093689, + "learning_rate": 2.3373990015555424e-05, + "loss": 0.0425, + "step": 33186 + }, + { + "epoch": 0.7779555430533823, + "grad_norm": 0.22077858448028564, + "learning_rate": 2.3369257242281085e-05, + "loss": 0.0116, + "step": 33187 + }, + { + "epoch": 0.7779789846281874, + "grad_norm": 0.4246441125869751, + "learning_rate": 2.3364524884807315e-05, + "loss": 0.049, + "step": 33188 + }, + { + "epoch": 0.7780024262029923, + "grad_norm": 0.3387884199619293, + "learning_rate": 2.3359792943159787e-05, + "loss": 0.0569, + "step": 33189 + }, + { + "epoch": 0.7780258677777974, + "grad_norm": 0.43152105808258057, + "learning_rate": 2.3355061417364178e-05, + "loss": 0.0801, + "step": 33190 + }, + { + "epoch": 0.7780493093526023, + "grad_norm": 0.13683874905109406, + "learning_rate": 2.3350330307446134e-05, + "loss": 0.0168, + "step": 33191 + }, + { + "epoch": 0.7780727509274074, + "grad_norm": 0.6383987069129944, + "learning_rate": 2.334559961343138e-05, + "loss": 0.1277, + "step": 33192 + }, + { + "epoch": 0.7780961925022123, + "grad_norm": 0.2228614091873169, + "learning_rate": 2.334086933534555e-05, + "loss": 0.0176, + "step": 33193 + }, + { + "epoch": 0.7781196340770173, + "grad_norm": 0.38017454743385315, + "learning_rate": 2.3336139473214324e-05, + "loss": 0.0456, + "step": 33194 + }, + { + "epoch": 0.7781430756518223, + "grad_norm": 0.3174436390399933, + "learning_rate": 2.333141002706334e-05, + "loss": 0.0551, + "step": 33195 + }, + { + "epoch": 0.7781665172266273, + "grad_norm": 0.45516830682754517, + "learning_rate": 2.3326680996918294e-05, + "loss": 0.552, + "step": 33196 + }, + { + "epoch": 0.7781899588014323, + "grad_norm": 0.3551017940044403, + "learning_rate": 2.3321952382804845e-05, + "loss": 0.3693, + "step": 33197 + }, + { + "epoch": 0.7782134003762373, + "grad_norm": 0.4076981842517853, + "learning_rate": 2.331722418474861e-05, + "loss": 0.076, + "step": 33198 + }, + { + "epoch": 0.7782368419510423, + "grad_norm": 0.34831470251083374, + "learning_rate": 2.3312496402775296e-05, + "loss": 0.0255, + "step": 33199 + }, + { + "epoch": 0.7782602835258473, + "grad_norm": 0.3864161968231201, + "learning_rate": 2.330776903691053e-05, + "loss": 0.0603, + "step": 33200 + }, + { + "epoch": 0.7782837251006522, + "grad_norm": 0.16910773515701294, + "learning_rate": 2.330304208717997e-05, + "loss": 0.012, + "step": 33201 + }, + { + "epoch": 0.7783071666754573, + "grad_norm": 0.34400346875190735, + "learning_rate": 2.3298315553609263e-05, + "loss": 0.0579, + "step": 33202 + }, + { + "epoch": 0.7783306082502622, + "grad_norm": 0.5374251008033752, + "learning_rate": 2.3293589436224018e-05, + "loss": 0.1015, + "step": 33203 + }, + { + "epoch": 0.7783540498250673, + "grad_norm": 0.18085408210754395, + "learning_rate": 2.328886373504995e-05, + "loss": 0.0308, + "step": 33204 + }, + { + "epoch": 0.7783774913998722, + "grad_norm": 0.5648230314254761, + "learning_rate": 2.3284138450112658e-05, + "loss": 0.11, + "step": 33205 + }, + { + "epoch": 0.7784009329746773, + "grad_norm": 0.5890770554542542, + "learning_rate": 2.3279413581437782e-05, + "loss": 0.062, + "step": 33206 + }, + { + "epoch": 0.7784243745494822, + "grad_norm": 0.22616253793239594, + "learning_rate": 2.3274689129050954e-05, + "loss": 0.0556, + "step": 33207 + }, + { + "epoch": 0.7784478161242873, + "grad_norm": 0.25095003843307495, + "learning_rate": 2.326996509297783e-05, + "loss": 0.0455, + "step": 33208 + }, + { + "epoch": 0.7784712576990922, + "grad_norm": 0.379905104637146, + "learning_rate": 2.3265241473244048e-05, + "loss": 0.0538, + "step": 33209 + }, + { + "epoch": 0.7784946992738973, + "grad_norm": 0.36952683329582214, + "learning_rate": 2.3260518269875188e-05, + "loss": 0.081, + "step": 33210 + }, + { + "epoch": 0.7785181408487022, + "grad_norm": 0.7567716836929321, + "learning_rate": 2.3255795482896937e-05, + "loss": 0.1363, + "step": 33211 + }, + { + "epoch": 0.7785415824235072, + "grad_norm": 0.7623702883720398, + "learning_rate": 2.3251073112334896e-05, + "loss": 0.2696, + "step": 33212 + }, + { + "epoch": 0.7785650239983122, + "grad_norm": 0.7141168117523193, + "learning_rate": 2.3246351158214707e-05, + "loss": 0.6745, + "step": 33213 + }, + { + "epoch": 0.7785884655731172, + "grad_norm": 0.15440796315670013, + "learning_rate": 2.3241629620561966e-05, + "loss": 0.023, + "step": 33214 + }, + { + "epoch": 0.7786119071479222, + "grad_norm": 0.4781302213668823, + "learning_rate": 2.3236908499402297e-05, + "loss": 0.0963, + "step": 33215 + }, + { + "epoch": 0.7786353487227272, + "grad_norm": 0.08290628343820572, + "learning_rate": 2.3232187794761296e-05, + "loss": 0.014, + "step": 33216 + }, + { + "epoch": 0.7786587902975322, + "grad_norm": 0.4136587977409363, + "learning_rate": 2.3227467506664623e-05, + "loss": 0.0517, + "step": 33217 + }, + { + "epoch": 0.7786822318723372, + "grad_norm": 0.2670080363750458, + "learning_rate": 2.3222747635137876e-05, + "loss": 0.0276, + "step": 33218 + }, + { + "epoch": 0.7787056734471421, + "grad_norm": 0.34541773796081543, + "learning_rate": 2.3218028180206643e-05, + "loss": 0.0437, + "step": 33219 + }, + { + "epoch": 0.7787291150219472, + "grad_norm": 0.2638394236564636, + "learning_rate": 2.3213309141896557e-05, + "loss": 0.0584, + "step": 33220 + }, + { + "epoch": 0.7787525565967521, + "grad_norm": 0.4274892508983612, + "learning_rate": 2.320859052023322e-05, + "loss": 0.0744, + "step": 33221 + }, + { + "epoch": 0.7787759981715572, + "grad_norm": 0.33669641613960266, + "learning_rate": 2.3203872315242237e-05, + "loss": 0.0352, + "step": 33222 + }, + { + "epoch": 0.7787994397463621, + "grad_norm": 0.49162396788597107, + "learning_rate": 2.3199154526949175e-05, + "loss": 0.0888, + "step": 33223 + }, + { + "epoch": 0.7788228813211672, + "grad_norm": 0.27340665459632874, + "learning_rate": 2.3194437155379677e-05, + "loss": 0.0754, + "step": 33224 + }, + { + "epoch": 0.7788463228959721, + "grad_norm": 0.3328840732574463, + "learning_rate": 2.3189720200559327e-05, + "loss": 0.0694, + "step": 33225 + }, + { + "epoch": 0.7788697644707772, + "grad_norm": 0.3708660304546356, + "learning_rate": 2.3185003662513715e-05, + "loss": 0.0571, + "step": 33226 + }, + { + "epoch": 0.7788932060455821, + "grad_norm": 0.14199486374855042, + "learning_rate": 2.3180287541268443e-05, + "loss": 0.0256, + "step": 33227 + }, + { + "epoch": 0.7789166476203871, + "grad_norm": 0.4231768846511841, + "learning_rate": 2.3175571836849076e-05, + "loss": 0.1044, + "step": 33228 + }, + { + "epoch": 0.7789400891951921, + "grad_norm": 0.6775490641593933, + "learning_rate": 2.3170856549281193e-05, + "loss": 0.0925, + "step": 33229 + }, + { + "epoch": 0.7789635307699971, + "grad_norm": 0.2793493866920471, + "learning_rate": 2.3166141678590438e-05, + "loss": 0.0559, + "step": 33230 + }, + { + "epoch": 0.7789869723448021, + "grad_norm": 0.18163955211639404, + "learning_rate": 2.3161427224802323e-05, + "loss": 0.0306, + "step": 33231 + }, + { + "epoch": 0.7790104139196071, + "grad_norm": 0.44866296648979187, + "learning_rate": 2.3156713187942493e-05, + "loss": 0.1085, + "step": 33232 + }, + { + "epoch": 0.7790338554944121, + "grad_norm": 0.21440261602401733, + "learning_rate": 2.31519995680365e-05, + "loss": 0.035, + "step": 33233 + }, + { + "epoch": 0.7790572970692171, + "grad_norm": 0.33630093932151794, + "learning_rate": 2.314728636510991e-05, + "loss": 0.0501, + "step": 33234 + }, + { + "epoch": 0.779080738644022, + "grad_norm": 0.19039854407310486, + "learning_rate": 2.3142573579188285e-05, + "loss": 0.0233, + "step": 33235 + }, + { + "epoch": 0.7791041802188271, + "grad_norm": 0.5110500454902649, + "learning_rate": 2.313786121029724e-05, + "loss": 0.1227, + "step": 33236 + }, + { + "epoch": 0.779127621793632, + "grad_norm": 0.3977409601211548, + "learning_rate": 2.3133149258462327e-05, + "loss": 0.0566, + "step": 33237 + }, + { + "epoch": 0.7791510633684371, + "grad_norm": 0.138833686709404, + "learning_rate": 2.3128437723709106e-05, + "loss": 0.0311, + "step": 33238 + }, + { + "epoch": 0.7791745049432421, + "grad_norm": 0.613950788974762, + "learning_rate": 2.3123726606063144e-05, + "loss": 0.1209, + "step": 33239 + }, + { + "epoch": 0.7791979465180471, + "grad_norm": 0.49598899483680725, + "learning_rate": 2.3119015905550003e-05, + "loss": 0.0795, + "step": 33240 + }, + { + "epoch": 0.7792213880928521, + "grad_norm": 0.3840363919734955, + "learning_rate": 2.311430562219522e-05, + "loss": 0.0665, + "step": 33241 + }, + { + "epoch": 0.7792448296676571, + "grad_norm": 0.421151727437973, + "learning_rate": 2.3109595756024393e-05, + "loss": 0.0643, + "step": 33242 + }, + { + "epoch": 0.7792682712424621, + "grad_norm": 0.19419565796852112, + "learning_rate": 2.3104886307063045e-05, + "loss": 0.0216, + "step": 33243 + }, + { + "epoch": 0.7792917128172671, + "grad_norm": 0.43420925736427307, + "learning_rate": 2.3100177275336764e-05, + "loss": 0.098, + "step": 33244 + }, + { + "epoch": 0.7793151543920721, + "grad_norm": 0.3838839828968048, + "learning_rate": 2.309546866087109e-05, + "loss": 0.0319, + "step": 33245 + }, + { + "epoch": 0.779338595966877, + "grad_norm": 0.5096038579940796, + "learning_rate": 2.3090760463691565e-05, + "loss": 0.0715, + "step": 33246 + }, + { + "epoch": 0.7793620375416821, + "grad_norm": 0.2859726846218109, + "learning_rate": 2.3086052683823733e-05, + "loss": 0.0422, + "step": 33247 + }, + { + "epoch": 0.779385479116487, + "grad_norm": 0.45422643423080444, + "learning_rate": 2.3081345321293114e-05, + "loss": 0.0939, + "step": 33248 + }, + { + "epoch": 0.7794089206912921, + "grad_norm": 0.4282105267047882, + "learning_rate": 2.3076638376125305e-05, + "loss": 0.0989, + "step": 33249 + }, + { + "epoch": 0.779432362266097, + "grad_norm": 0.09717395156621933, + "learning_rate": 2.3071931848345818e-05, + "loss": 0.0201, + "step": 33250 + }, + { + "epoch": 0.7794558038409021, + "grad_norm": 0.3922237455844879, + "learning_rate": 2.3067225737980193e-05, + "loss": 0.0523, + "step": 33251 + }, + { + "epoch": 0.779479245415707, + "grad_norm": 0.7516816854476929, + "learning_rate": 2.306252004505397e-05, + "loss": 0.1371, + "step": 33252 + }, + { + "epoch": 0.7795026869905121, + "grad_norm": 0.15319916605949402, + "learning_rate": 2.305781476959267e-05, + "loss": 0.0239, + "step": 33253 + }, + { + "epoch": 0.779526128565317, + "grad_norm": 0.13353602588176727, + "learning_rate": 2.3053109911621807e-05, + "loss": 0.0212, + "step": 33254 + }, + { + "epoch": 0.779549570140122, + "grad_norm": 0.23129133880138397, + "learning_rate": 2.3048405471166957e-05, + "loss": 0.0302, + "step": 33255 + }, + { + "epoch": 0.779573011714927, + "grad_norm": 0.3198515474796295, + "learning_rate": 2.3043701448253597e-05, + "loss": 0.0475, + "step": 33256 + }, + { + "epoch": 0.779596453289732, + "grad_norm": 0.5869880318641663, + "learning_rate": 2.3038997842907307e-05, + "loss": 0.1085, + "step": 33257 + }, + { + "epoch": 0.779619894864537, + "grad_norm": 0.42187047004699707, + "learning_rate": 2.3034294655153575e-05, + "loss": 0.0442, + "step": 33258 + }, + { + "epoch": 0.779643336439342, + "grad_norm": 0.3926181495189667, + "learning_rate": 2.3029591885017932e-05, + "loss": 0.0575, + "step": 33259 + }, + { + "epoch": 0.779666778014147, + "grad_norm": 0.542394757270813, + "learning_rate": 2.3024889532525874e-05, + "loss": 0.0636, + "step": 33260 + }, + { + "epoch": 0.779690219588952, + "grad_norm": 0.4319138824939728, + "learning_rate": 2.3020187597702912e-05, + "loss": 0.1059, + "step": 33261 + }, + { + "epoch": 0.779713661163757, + "grad_norm": 0.35561347007751465, + "learning_rate": 2.30154860805746e-05, + "loss": 0.0382, + "step": 33262 + }, + { + "epoch": 0.779737102738562, + "grad_norm": 0.5519180297851562, + "learning_rate": 2.301078498116642e-05, + "loss": 0.083, + "step": 33263 + }, + { + "epoch": 0.7797605443133669, + "grad_norm": 0.3837122321128845, + "learning_rate": 2.3006084299503904e-05, + "loss": 0.0885, + "step": 33264 + }, + { + "epoch": 0.779783985888172, + "grad_norm": 0.5602971911430359, + "learning_rate": 2.3001384035612527e-05, + "loss": 0.0664, + "step": 33265 + }, + { + "epoch": 0.7798074274629769, + "grad_norm": 0.05946247652173042, + "learning_rate": 2.2996684189517804e-05, + "loss": 0.0092, + "step": 33266 + }, + { + "epoch": 0.779830869037782, + "grad_norm": 0.16638946533203125, + "learning_rate": 2.2991984761245212e-05, + "loss": 0.0253, + "step": 33267 + }, + { + "epoch": 0.7798543106125869, + "grad_norm": 0.4977055490016937, + "learning_rate": 2.2987285750820275e-05, + "loss": 0.0664, + "step": 33268 + }, + { + "epoch": 0.779877752187392, + "grad_norm": 0.3927115499973297, + "learning_rate": 2.298258715826852e-05, + "loss": 0.0525, + "step": 33269 + }, + { + "epoch": 0.7799011937621969, + "grad_norm": 0.4240530729293823, + "learning_rate": 2.2977888983615414e-05, + "loss": 0.0511, + "step": 33270 + }, + { + "epoch": 0.779924635337002, + "grad_norm": 0.399312824010849, + "learning_rate": 2.2973191226886436e-05, + "loss": 0.0525, + "step": 33271 + }, + { + "epoch": 0.7799480769118069, + "grad_norm": 0.22198854386806488, + "learning_rate": 2.2968493888107102e-05, + "loss": 0.0339, + "step": 33272 + }, + { + "epoch": 0.779971518486612, + "grad_norm": 0.4601903259754181, + "learning_rate": 2.296379696730285e-05, + "loss": 0.0657, + "step": 33273 + }, + { + "epoch": 0.7799949600614169, + "grad_norm": 0.4504240155220032, + "learning_rate": 2.295910046449924e-05, + "loss": 0.0464, + "step": 33274 + }, + { + "epoch": 0.7800184016362219, + "grad_norm": 0.18154294788837433, + "learning_rate": 2.2954404379721715e-05, + "loss": 0.028, + "step": 33275 + }, + { + "epoch": 0.7800418432110269, + "grad_norm": 0.3870541453361511, + "learning_rate": 2.2949708712995753e-05, + "loss": 0.0937, + "step": 33276 + }, + { + "epoch": 0.7800652847858319, + "grad_norm": 0.48401588201522827, + "learning_rate": 2.2945013464346843e-05, + "loss": 0.0583, + "step": 33277 + }, + { + "epoch": 0.7800887263606369, + "grad_norm": 0.6363365650177002, + "learning_rate": 2.294031863380045e-05, + "loss": 0.0783, + "step": 33278 + }, + { + "epoch": 0.7801121679354419, + "grad_norm": 0.5543710589408875, + "learning_rate": 2.2935624221382025e-05, + "loss": 0.0832, + "step": 33279 + }, + { + "epoch": 0.7801356095102469, + "grad_norm": 0.1284535527229309, + "learning_rate": 2.2930930227117075e-05, + "loss": 0.0172, + "step": 33280 + }, + { + "epoch": 0.7801590510850519, + "grad_norm": 0.5458230972290039, + "learning_rate": 2.2926236651031098e-05, + "loss": 0.1558, + "step": 33281 + }, + { + "epoch": 0.7801824926598568, + "grad_norm": 0.21022507548332214, + "learning_rate": 2.2921543493149522e-05, + "loss": 0.0232, + "step": 33282 + }, + { + "epoch": 0.7802059342346619, + "grad_norm": 0.5746094584465027, + "learning_rate": 2.291685075349782e-05, + "loss": 0.1499, + "step": 33283 + }, + { + "epoch": 0.7802293758094668, + "grad_norm": 0.30103954672813416, + "learning_rate": 2.291215843210145e-05, + "loss": 0.0691, + "step": 33284 + }, + { + "epoch": 0.7802528173842719, + "grad_norm": 0.43583232164382935, + "learning_rate": 2.2907466528985876e-05, + "loss": 0.1079, + "step": 33285 + }, + { + "epoch": 0.7802762589590768, + "grad_norm": 0.7627067565917969, + "learning_rate": 2.2902775044176528e-05, + "loss": 0.0709, + "step": 33286 + }, + { + "epoch": 0.7802997005338819, + "grad_norm": 0.4181435704231262, + "learning_rate": 2.2898083977698927e-05, + "loss": 0.0589, + "step": 33287 + }, + { + "epoch": 0.7803231421086868, + "grad_norm": 0.34680667519569397, + "learning_rate": 2.289339332957847e-05, + "loss": 0.0474, + "step": 33288 + }, + { + "epoch": 0.7803465836834919, + "grad_norm": 0.49446988105773926, + "learning_rate": 2.2888703099840646e-05, + "loss": 0.6071, + "step": 33289 + }, + { + "epoch": 0.7803700252582969, + "grad_norm": 0.5500105619430542, + "learning_rate": 2.2884013288510886e-05, + "loss": 0.0931, + "step": 33290 + }, + { + "epoch": 0.7803934668331018, + "grad_norm": 0.6253570318222046, + "learning_rate": 2.2879323895614636e-05, + "loss": 0.0777, + "step": 33291 + }, + { + "epoch": 0.7804169084079069, + "grad_norm": 0.4537813365459442, + "learning_rate": 2.2874634921177318e-05, + "loss": 0.3782, + "step": 33292 + }, + { + "epoch": 0.7804403499827118, + "grad_norm": 0.31748855113983154, + "learning_rate": 2.2869946365224393e-05, + "loss": 0.0274, + "step": 33293 + }, + { + "epoch": 0.7804637915575169, + "grad_norm": 0.4136195182800293, + "learning_rate": 2.286525822778134e-05, + "loss": 0.0415, + "step": 33294 + }, + { + "epoch": 0.7804872331323218, + "grad_norm": 0.5350306630134583, + "learning_rate": 2.2860570508873556e-05, + "loss": 0.0982, + "step": 33295 + }, + { + "epoch": 0.7805106747071269, + "grad_norm": 0.24481701850891113, + "learning_rate": 2.2855883208526497e-05, + "loss": 0.0427, + "step": 33296 + }, + { + "epoch": 0.7805341162819318, + "grad_norm": 0.39278343319892883, + "learning_rate": 2.2851196326765577e-05, + "loss": 0.0502, + "step": 33297 + }, + { + "epoch": 0.7805575578567369, + "grad_norm": 0.3795228600502014, + "learning_rate": 2.284650986361625e-05, + "loss": 0.0615, + "step": 33298 + }, + { + "epoch": 0.7805809994315418, + "grad_norm": 0.32458925247192383, + "learning_rate": 2.284182381910389e-05, + "loss": 0.0402, + "step": 33299 + }, + { + "epoch": 0.7806044410063468, + "grad_norm": 0.17453965544700623, + "learning_rate": 2.283713819325399e-05, + "loss": 0.0381, + "step": 33300 + }, + { + "epoch": 0.7806278825811518, + "grad_norm": 0.26250195503234863, + "learning_rate": 2.2832452986091946e-05, + "loss": 0.0394, + "step": 33301 + }, + { + "epoch": 0.7806513241559568, + "grad_norm": 0.3834723234176636, + "learning_rate": 2.2827768197643186e-05, + "loss": 0.0655, + "step": 33302 + }, + { + "epoch": 0.7806747657307618, + "grad_norm": 0.6224939823150635, + "learning_rate": 2.2823083827933122e-05, + "loss": 0.0926, + "step": 33303 + }, + { + "epoch": 0.7806982073055668, + "grad_norm": 0.3254983425140381, + "learning_rate": 2.281839987698715e-05, + "loss": 0.0476, + "step": 33304 + }, + { + "epoch": 0.7807216488803718, + "grad_norm": 0.48366841673851013, + "learning_rate": 2.281371634483075e-05, + "loss": 0.0592, + "step": 33305 + }, + { + "epoch": 0.7807450904551768, + "grad_norm": 0.30695414543151855, + "learning_rate": 2.280903323148925e-05, + "loss": 0.0289, + "step": 33306 + }, + { + "epoch": 0.7807685320299818, + "grad_norm": 0.9609414339065552, + "learning_rate": 2.280435053698814e-05, + "loss": 0.1969, + "step": 33307 + }, + { + "epoch": 0.7807919736047868, + "grad_norm": 0.24457387626171112, + "learning_rate": 2.2799668261352792e-05, + "loss": 0.0527, + "step": 33308 + }, + { + "epoch": 0.7808154151795917, + "grad_norm": 0.4607332944869995, + "learning_rate": 2.279498640460862e-05, + "loss": 0.4091, + "step": 33309 + }, + { + "epoch": 0.7808388567543968, + "grad_norm": 0.76023268699646, + "learning_rate": 2.279030496678102e-05, + "loss": 0.0988, + "step": 33310 + }, + { + "epoch": 0.7808622983292017, + "grad_norm": 0.5138198733329773, + "learning_rate": 2.2785623947895375e-05, + "loss": 0.0981, + "step": 33311 + }, + { + "epoch": 0.7808857399040068, + "grad_norm": 0.4681309759616852, + "learning_rate": 2.2780943347977125e-05, + "loss": 0.0629, + "step": 33312 + }, + { + "epoch": 0.7809091814788117, + "grad_norm": 0.1189272478222847, + "learning_rate": 2.2776263167051648e-05, + "loss": 0.0205, + "step": 33313 + }, + { + "epoch": 0.7809326230536168, + "grad_norm": 0.27471286058425903, + "learning_rate": 2.2771583405144337e-05, + "loss": 0.0396, + "step": 33314 + }, + { + "epoch": 0.7809560646284217, + "grad_norm": 0.43689754605293274, + "learning_rate": 2.2766904062280593e-05, + "loss": 0.0765, + "step": 33315 + }, + { + "epoch": 0.7809795062032268, + "grad_norm": 0.4935551583766937, + "learning_rate": 2.276222513848576e-05, + "loss": 0.0774, + "step": 33316 + }, + { + "epoch": 0.7810029477780317, + "grad_norm": 0.4023558497428894, + "learning_rate": 2.275754663378531e-05, + "loss": 0.0655, + "step": 33317 + }, + { + "epoch": 0.7810263893528367, + "grad_norm": 0.08200123906135559, + "learning_rate": 2.275286854820454e-05, + "loss": 0.0128, + "step": 33318 + }, + { + "epoch": 0.7810498309276417, + "grad_norm": 0.32443052530288696, + "learning_rate": 2.2748190881768917e-05, + "loss": 0.0417, + "step": 33319 + }, + { + "epoch": 0.7810732725024467, + "grad_norm": 0.10194157809019089, + "learning_rate": 2.2743513634503777e-05, + "loss": 0.0092, + "step": 33320 + }, + { + "epoch": 0.7810967140772517, + "grad_norm": 0.28901588916778564, + "learning_rate": 2.2738836806434494e-05, + "loss": 0.0608, + "step": 33321 + }, + { + "epoch": 0.7811201556520567, + "grad_norm": 0.5728350281715393, + "learning_rate": 2.2734160397586468e-05, + "loss": 0.6689, + "step": 33322 + }, + { + "epoch": 0.7811435972268617, + "grad_norm": 0.4358947277069092, + "learning_rate": 2.2729484407985057e-05, + "loss": 0.0609, + "step": 33323 + }, + { + "epoch": 0.7811670388016667, + "grad_norm": 0.6376729607582092, + "learning_rate": 2.272480883765561e-05, + "loss": 0.5891, + "step": 33324 + }, + { + "epoch": 0.7811904803764717, + "grad_norm": 0.32338663935661316, + "learning_rate": 2.2720133686623545e-05, + "loss": 0.0479, + "step": 33325 + }, + { + "epoch": 0.7812139219512767, + "grad_norm": 0.13315345346927643, + "learning_rate": 2.2715458954914216e-05, + "loss": 0.0207, + "step": 33326 + }, + { + "epoch": 0.7812373635260816, + "grad_norm": 0.35391929745674133, + "learning_rate": 2.2710784642552972e-05, + "loss": 0.358, + "step": 33327 + }, + { + "epoch": 0.7812608051008867, + "grad_norm": 0.4286124110221863, + "learning_rate": 2.270611074956519e-05, + "loss": 0.0957, + "step": 33328 + }, + { + "epoch": 0.7812842466756916, + "grad_norm": 0.2168768346309662, + "learning_rate": 2.270143727597619e-05, + "loss": 0.0371, + "step": 33329 + }, + { + "epoch": 0.7813076882504967, + "grad_norm": 0.5818126797676086, + "learning_rate": 2.2696764221811407e-05, + "loss": 0.1081, + "step": 33330 + }, + { + "epoch": 0.7813311298253016, + "grad_norm": 0.5341825485229492, + "learning_rate": 2.2692091587096108e-05, + "loss": 0.0724, + "step": 33331 + }, + { + "epoch": 0.7813545714001067, + "grad_norm": 0.42784440517425537, + "learning_rate": 2.2687419371855735e-05, + "loss": 0.0952, + "step": 33332 + }, + { + "epoch": 0.7813780129749116, + "grad_norm": 0.38459160923957825, + "learning_rate": 2.2682747576115592e-05, + "loss": 0.0667, + "step": 33333 + }, + { + "epoch": 0.7814014545497167, + "grad_norm": 0.3915186822414398, + "learning_rate": 2.2678076199901032e-05, + "loss": 0.0765, + "step": 33334 + }, + { + "epoch": 0.7814248961245216, + "grad_norm": 0.10186270624399185, + "learning_rate": 2.2673405243237412e-05, + "loss": 0.0231, + "step": 33335 + }, + { + "epoch": 0.7814483376993266, + "grad_norm": 0.30467185378074646, + "learning_rate": 2.2668734706150062e-05, + "loss": 0.0738, + "step": 33336 + }, + { + "epoch": 0.7814717792741316, + "grad_norm": 0.17377729713916779, + "learning_rate": 2.2664064588664314e-05, + "loss": 0.0181, + "step": 33337 + }, + { + "epoch": 0.7814952208489366, + "grad_norm": 0.2506134808063507, + "learning_rate": 2.2659394890805552e-05, + "loss": 0.0324, + "step": 33338 + }, + { + "epoch": 0.7815186624237416, + "grad_norm": 0.24295631051063538, + "learning_rate": 2.2654725612599083e-05, + "loss": 0.0375, + "step": 33339 + }, + { + "epoch": 0.7815421039985466, + "grad_norm": 0.36388522386550903, + "learning_rate": 2.265005675407025e-05, + "loss": 0.0567, + "step": 33340 + }, + { + "epoch": 0.7815655455733517, + "grad_norm": 0.6131036877632141, + "learning_rate": 2.2645388315244354e-05, + "loss": 0.0912, + "step": 33341 + }, + { + "epoch": 0.7815889871481566, + "grad_norm": 0.3098984658718109, + "learning_rate": 2.2640720296146777e-05, + "loss": 0.032, + "step": 33342 + }, + { + "epoch": 0.7816124287229617, + "grad_norm": 0.2856868505477905, + "learning_rate": 2.2636052696802835e-05, + "loss": 0.0498, + "step": 33343 + }, + { + "epoch": 0.7816358702977666, + "grad_norm": 0.4290897250175476, + "learning_rate": 2.2631385517237824e-05, + "loss": 0.5129, + "step": 33344 + }, + { + "epoch": 0.7816593118725716, + "grad_norm": 0.5427061915397644, + "learning_rate": 2.2626718757477106e-05, + "loss": 0.4589, + "step": 33345 + }, + { + "epoch": 0.7816827534473766, + "grad_norm": 0.795653223991394, + "learning_rate": 2.2622052417545993e-05, + "loss": 0.1468, + "step": 33346 + }, + { + "epoch": 0.7817061950221816, + "grad_norm": 0.5144878029823303, + "learning_rate": 2.2617386497469805e-05, + "loss": 0.0446, + "step": 33347 + }, + { + "epoch": 0.7817296365969866, + "grad_norm": 0.4721277058124542, + "learning_rate": 2.261272099727385e-05, + "loss": 0.4369, + "step": 33348 + }, + { + "epoch": 0.7817530781717916, + "grad_norm": 0.5817933678627014, + "learning_rate": 2.2608055916983416e-05, + "loss": 0.0766, + "step": 33349 + }, + { + "epoch": 0.7817765197465966, + "grad_norm": 0.09206418693065643, + "learning_rate": 2.2603391256623874e-05, + "loss": 0.0052, + "step": 33350 + }, + { + "epoch": 0.7817999613214016, + "grad_norm": 0.6359918713569641, + "learning_rate": 2.259872701622051e-05, + "loss": 0.0609, + "step": 33351 + }, + { + "epoch": 0.7818234028962066, + "grad_norm": 0.44828352332115173, + "learning_rate": 2.259406319579863e-05, + "loss": 0.0736, + "step": 33352 + }, + { + "epoch": 0.7818468444710116, + "grad_norm": 0.10960403084754944, + "learning_rate": 2.2589399795383502e-05, + "loss": 0.0249, + "step": 33353 + }, + { + "epoch": 0.7818702860458165, + "grad_norm": 0.516456663608551, + "learning_rate": 2.258473681500051e-05, + "loss": 0.6181, + "step": 33354 + }, + { + "epoch": 0.7818937276206216, + "grad_norm": 0.6010251045227051, + "learning_rate": 2.258007425467491e-05, + "loss": 0.0652, + "step": 33355 + }, + { + "epoch": 0.7819171691954265, + "grad_norm": 0.6782840490341187, + "learning_rate": 2.257541211443197e-05, + "loss": 0.1265, + "step": 33356 + }, + { + "epoch": 0.7819406107702316, + "grad_norm": 0.4107378125190735, + "learning_rate": 2.2570750394297046e-05, + "loss": 0.0305, + "step": 33357 + }, + { + "epoch": 0.7819640523450365, + "grad_norm": 0.5783552527427673, + "learning_rate": 2.2566089094295418e-05, + "loss": 0.0634, + "step": 33358 + }, + { + "epoch": 0.7819874939198416, + "grad_norm": 0.4556746482849121, + "learning_rate": 2.256142821445236e-05, + "loss": 0.0622, + "step": 33359 + }, + { + "epoch": 0.7820109354946465, + "grad_norm": 0.14900319278240204, + "learning_rate": 2.2556767754793174e-05, + "loss": 0.0189, + "step": 33360 + }, + { + "epoch": 0.7820343770694516, + "grad_norm": 0.3167215585708618, + "learning_rate": 2.255210771534314e-05, + "loss": 0.0298, + "step": 33361 + }, + { + "epoch": 0.7820578186442565, + "grad_norm": 0.3141610324382782, + "learning_rate": 2.254744809612752e-05, + "loss": 0.0488, + "step": 33362 + }, + { + "epoch": 0.7820812602190615, + "grad_norm": 0.13901099562644958, + "learning_rate": 2.2542788897171664e-05, + "loss": 0.0177, + "step": 33363 + }, + { + "epoch": 0.7821047017938665, + "grad_norm": 0.6109400391578674, + "learning_rate": 2.25381301185008e-05, + "loss": 0.0658, + "step": 33364 + }, + { + "epoch": 0.7821281433686715, + "grad_norm": 0.3831736445426941, + "learning_rate": 2.253347176014019e-05, + "loss": 0.0455, + "step": 33365 + }, + { + "epoch": 0.7821515849434765, + "grad_norm": 0.7588112950325012, + "learning_rate": 2.252881382211518e-05, + "loss": 0.0628, + "step": 33366 + }, + { + "epoch": 0.7821750265182815, + "grad_norm": 0.10851919651031494, + "learning_rate": 2.2524156304450994e-05, + "loss": 0.0133, + "step": 33367 + }, + { + "epoch": 0.7821984680930865, + "grad_norm": 0.10272049158811569, + "learning_rate": 2.2519499207172923e-05, + "loss": 0.0146, + "step": 33368 + }, + { + "epoch": 0.7822219096678915, + "grad_norm": 0.1368762105703354, + "learning_rate": 2.251484253030619e-05, + "loss": 0.028, + "step": 33369 + }, + { + "epoch": 0.7822453512426965, + "grad_norm": 0.4808826148509979, + "learning_rate": 2.251018627387613e-05, + "loss": 0.0446, + "step": 33370 + }, + { + "epoch": 0.7822687928175015, + "grad_norm": 0.9775944948196411, + "learning_rate": 2.2505530437907986e-05, + "loss": 0.1467, + "step": 33371 + }, + { + "epoch": 0.7822922343923064, + "grad_norm": 0.17649058997631073, + "learning_rate": 2.2500875022427015e-05, + "loss": 0.0345, + "step": 33372 + }, + { + "epoch": 0.7823156759671115, + "grad_norm": 0.5913797616958618, + "learning_rate": 2.2496220027458458e-05, + "loss": 0.1568, + "step": 33373 + }, + { + "epoch": 0.7823391175419164, + "grad_norm": 0.36703890562057495, + "learning_rate": 2.2491565453027607e-05, + "loss": 0.0756, + "step": 33374 + }, + { + "epoch": 0.7823625591167215, + "grad_norm": 0.7480735778808594, + "learning_rate": 2.2486911299159663e-05, + "loss": 0.1243, + "step": 33375 + }, + { + "epoch": 0.7823860006915264, + "grad_norm": 0.2897324562072754, + "learning_rate": 2.248225756587995e-05, + "loss": 0.0506, + "step": 33376 + }, + { + "epoch": 0.7824094422663315, + "grad_norm": 0.6065152287483215, + "learning_rate": 2.2477604253213656e-05, + "loss": 0.0888, + "step": 33377 + }, + { + "epoch": 0.7824328838411364, + "grad_norm": 0.5395263433456421, + "learning_rate": 2.24729513611861e-05, + "loss": 0.0717, + "step": 33378 + }, + { + "epoch": 0.7824563254159415, + "grad_norm": 0.23604540526866913, + "learning_rate": 2.246829888982248e-05, + "loss": 0.0379, + "step": 33379 + }, + { + "epoch": 0.7824797669907464, + "grad_norm": 0.39296233654022217, + "learning_rate": 2.2463646839148055e-05, + "loss": 0.1092, + "step": 33380 + }, + { + "epoch": 0.7825032085655514, + "grad_norm": 0.6545621752738953, + "learning_rate": 2.245899520918806e-05, + "loss": 0.0817, + "step": 33381 + }, + { + "epoch": 0.7825266501403564, + "grad_norm": 0.18403953313827515, + "learning_rate": 2.2454343999967708e-05, + "loss": 0.0261, + "step": 33382 + }, + { + "epoch": 0.7825500917151614, + "grad_norm": 0.44152429699897766, + "learning_rate": 2.244969321151229e-05, + "loss": 0.0563, + "step": 33383 + }, + { + "epoch": 0.7825735332899664, + "grad_norm": 0.2756267189979553, + "learning_rate": 2.244504284384702e-05, + "loss": 0.0466, + "step": 33384 + }, + { + "epoch": 0.7825969748647714, + "grad_norm": 0.8681884407997131, + "learning_rate": 2.244039289699713e-05, + "loss": 0.1162, + "step": 33385 + }, + { + "epoch": 0.7826204164395764, + "grad_norm": 0.5339304208755493, + "learning_rate": 2.243574337098785e-05, + "loss": 0.0748, + "step": 33386 + }, + { + "epoch": 0.7826438580143814, + "grad_norm": 0.5814312696456909, + "learning_rate": 2.2431094265844377e-05, + "loss": 0.7268, + "step": 33387 + }, + { + "epoch": 0.7826672995891863, + "grad_norm": 0.426479697227478, + "learning_rate": 2.2426445581592004e-05, + "loss": 0.0797, + "step": 33388 + }, + { + "epoch": 0.7826907411639914, + "grad_norm": 0.523745596408844, + "learning_rate": 2.2421797318255877e-05, + "loss": 0.1139, + "step": 33389 + }, + { + "epoch": 0.7827141827387963, + "grad_norm": 0.2511613965034485, + "learning_rate": 2.24171494758613e-05, + "loss": 0.0434, + "step": 33390 + }, + { + "epoch": 0.7827376243136014, + "grad_norm": 0.3249050974845886, + "learning_rate": 2.241250205443344e-05, + "loss": 0.0521, + "step": 33391 + }, + { + "epoch": 0.7827610658884064, + "grad_norm": 0.41339489817619324, + "learning_rate": 2.240785505399753e-05, + "loss": 0.0991, + "step": 33392 + }, + { + "epoch": 0.7827845074632114, + "grad_norm": 0.42728495597839355, + "learning_rate": 2.2403208474578785e-05, + "loss": 0.0725, + "step": 33393 + }, + { + "epoch": 0.7828079490380164, + "grad_norm": 0.4170713722705841, + "learning_rate": 2.2398562316202375e-05, + "loss": 0.0484, + "step": 33394 + }, + { + "epoch": 0.7828313906128214, + "grad_norm": 0.42664867639541626, + "learning_rate": 2.239391657889358e-05, + "loss": 0.0585, + "step": 33395 + }, + { + "epoch": 0.7828548321876264, + "grad_norm": 0.3990420699119568, + "learning_rate": 2.2389271262677582e-05, + "loss": 0.0551, + "step": 33396 + }, + { + "epoch": 0.7828782737624314, + "grad_norm": 0.42080944776535034, + "learning_rate": 2.2384626367579577e-05, + "loss": 0.0505, + "step": 33397 + }, + { + "epoch": 0.7829017153372364, + "grad_norm": 0.23446491360664368, + "learning_rate": 2.2379981893624757e-05, + "loss": 0.0257, + "step": 33398 + }, + { + "epoch": 0.7829251569120413, + "grad_norm": 0.5293718576431274, + "learning_rate": 2.237533784083835e-05, + "loss": 0.0762, + "step": 33399 + }, + { + "epoch": 0.7829485984868464, + "grad_norm": 0.562160849571228, + "learning_rate": 2.2370694209245523e-05, + "loss": 0.4026, + "step": 33400 + }, + { + "epoch": 0.7829720400616513, + "grad_norm": 0.414773553609848, + "learning_rate": 2.236605099887148e-05, + "loss": 0.0408, + "step": 33401 + }, + { + "epoch": 0.7829954816364564, + "grad_norm": 0.12284966558218002, + "learning_rate": 2.2361408209741462e-05, + "loss": 0.0231, + "step": 33402 + }, + { + "epoch": 0.7830189232112613, + "grad_norm": 0.16013608872890472, + "learning_rate": 2.2356765841880622e-05, + "loss": 0.0152, + "step": 33403 + }, + { + "epoch": 0.7830423647860664, + "grad_norm": 0.6301215887069702, + "learning_rate": 2.235212389531416e-05, + "loss": 0.076, + "step": 33404 + }, + { + "epoch": 0.7830658063608713, + "grad_norm": 0.6189089417457581, + "learning_rate": 2.234748237006724e-05, + "loss": 0.1621, + "step": 33405 + }, + { + "epoch": 0.7830892479356764, + "grad_norm": 0.342327356338501, + "learning_rate": 2.2342841266165082e-05, + "loss": 0.0807, + "step": 33406 + }, + { + "epoch": 0.7831126895104813, + "grad_norm": 0.43715760111808777, + "learning_rate": 2.2338200583632818e-05, + "loss": 0.0618, + "step": 33407 + }, + { + "epoch": 0.7831361310852863, + "grad_norm": 0.6234573721885681, + "learning_rate": 2.2333560322495685e-05, + "loss": 0.0975, + "step": 33408 + }, + { + "epoch": 0.7831595726600913, + "grad_norm": 0.574923574924469, + "learning_rate": 2.2328920482778836e-05, + "loss": 0.085, + "step": 33409 + }, + { + "epoch": 0.7831830142348963, + "grad_norm": 0.20937544107437134, + "learning_rate": 2.2324281064507458e-05, + "loss": 0.0467, + "step": 33410 + }, + { + "epoch": 0.7832064558097013, + "grad_norm": 0.2240760773420334, + "learning_rate": 2.23196420677067e-05, + "loss": 0.0425, + "step": 33411 + }, + { + "epoch": 0.7832298973845063, + "grad_norm": 0.28630325198173523, + "learning_rate": 2.2315003492401764e-05, + "loss": 0.0437, + "step": 33412 + }, + { + "epoch": 0.7832533389593113, + "grad_norm": 0.35613101720809937, + "learning_rate": 2.231036533861777e-05, + "loss": 0.0392, + "step": 33413 + }, + { + "epoch": 0.7832767805341163, + "grad_norm": 0.415239542722702, + "learning_rate": 2.2305727606379922e-05, + "loss": 0.0441, + "step": 33414 + }, + { + "epoch": 0.7833002221089213, + "grad_norm": 0.1269746869802475, + "learning_rate": 2.23010902957134e-05, + "loss": 0.0138, + "step": 33415 + }, + { + "epoch": 0.7833236636837263, + "grad_norm": 0.4119649827480316, + "learning_rate": 2.2296453406643347e-05, + "loss": 0.0541, + "step": 33416 + }, + { + "epoch": 0.7833471052585312, + "grad_norm": 0.5448541045188904, + "learning_rate": 2.2291816939194932e-05, + "loss": 0.0637, + "step": 33417 + }, + { + "epoch": 0.7833705468333363, + "grad_norm": 0.17287786304950714, + "learning_rate": 2.22871808933933e-05, + "loss": 0.0234, + "step": 33418 + }, + { + "epoch": 0.7833939884081412, + "grad_norm": 0.14270949363708496, + "learning_rate": 2.22825452692636e-05, + "loss": 0.0169, + "step": 33419 + }, + { + "epoch": 0.7834174299829463, + "grad_norm": 0.13615983724594116, + "learning_rate": 2.2277910066830977e-05, + "loss": 0.0163, + "step": 33420 + }, + { + "epoch": 0.7834408715577512, + "grad_norm": 0.44357308745384216, + "learning_rate": 2.2273275286120622e-05, + "loss": 0.0689, + "step": 33421 + }, + { + "epoch": 0.7834643131325563, + "grad_norm": 0.480685830116272, + "learning_rate": 2.2268640927157668e-05, + "loss": 0.0984, + "step": 33422 + }, + { + "epoch": 0.7834877547073612, + "grad_norm": 0.1954517513513565, + "learning_rate": 2.2264006989967245e-05, + "loss": 0.0264, + "step": 33423 + }, + { + "epoch": 0.7835111962821663, + "grad_norm": 0.5548635721206665, + "learning_rate": 2.2259373474574507e-05, + "loss": 0.1179, + "step": 33424 + }, + { + "epoch": 0.7835346378569712, + "grad_norm": 0.1390700489282608, + "learning_rate": 2.225474038100457e-05, + "loss": 0.0253, + "step": 33425 + }, + { + "epoch": 0.7835580794317762, + "grad_norm": 0.4434804320335388, + "learning_rate": 2.225010770928263e-05, + "loss": 0.0559, + "step": 33426 + }, + { + "epoch": 0.7835815210065812, + "grad_norm": 0.6058152914047241, + "learning_rate": 2.224547545943375e-05, + "loss": 0.0868, + "step": 33427 + }, + { + "epoch": 0.7836049625813862, + "grad_norm": 0.40757352113723755, + "learning_rate": 2.2240843631483143e-05, + "loss": 0.0329, + "step": 33428 + }, + { + "epoch": 0.7836284041561912, + "grad_norm": 0.07830198854207993, + "learning_rate": 2.223621222545591e-05, + "loss": 0.0088, + "step": 33429 + }, + { + "epoch": 0.7836518457309962, + "grad_norm": 0.5014049410820007, + "learning_rate": 2.2231581241377163e-05, + "loss": 0.1114, + "step": 33430 + }, + { + "epoch": 0.7836752873058012, + "grad_norm": 0.3340464234352112, + "learning_rate": 2.222695067927205e-05, + "loss": 0.0677, + "step": 33431 + }, + { + "epoch": 0.7836987288806062, + "grad_norm": 0.6539596319198608, + "learning_rate": 2.2222320539165663e-05, + "loss": 0.0865, + "step": 33432 + }, + { + "epoch": 0.7837221704554111, + "grad_norm": 0.5142173171043396, + "learning_rate": 2.221769082108318e-05, + "loss": 0.1259, + "step": 33433 + }, + { + "epoch": 0.7837456120302162, + "grad_norm": 0.6831764578819275, + "learning_rate": 2.2213061525049695e-05, + "loss": 0.129, + "step": 33434 + }, + { + "epoch": 0.7837690536050211, + "grad_norm": 0.5579103231430054, + "learning_rate": 2.2208432651090327e-05, + "loss": 0.0743, + "step": 33435 + }, + { + "epoch": 0.7837924951798262, + "grad_norm": 0.2632879316806793, + "learning_rate": 2.2203804199230182e-05, + "loss": 0.0365, + "step": 33436 + }, + { + "epoch": 0.7838159367546311, + "grad_norm": 0.6671556830406189, + "learning_rate": 2.2199176169494383e-05, + "loss": 0.1553, + "step": 33437 + }, + { + "epoch": 0.7838393783294362, + "grad_norm": 0.17583461105823517, + "learning_rate": 2.2194548561908024e-05, + "loss": 0.0327, + "step": 33438 + }, + { + "epoch": 0.7838628199042411, + "grad_norm": 0.3357074558734894, + "learning_rate": 2.2189921376496224e-05, + "loss": 0.0428, + "step": 33439 + }, + { + "epoch": 0.7838862614790462, + "grad_norm": 0.5329142808914185, + "learning_rate": 2.2185294613284125e-05, + "loss": 0.7106, + "step": 33440 + }, + { + "epoch": 0.7839097030538511, + "grad_norm": 0.4722532331943512, + "learning_rate": 2.218066827229681e-05, + "loss": 0.0765, + "step": 33441 + }, + { + "epoch": 0.7839331446286562, + "grad_norm": 0.4366622269153595, + "learning_rate": 2.217604235355938e-05, + "loss": 0.0684, + "step": 33442 + }, + { + "epoch": 0.7839565862034612, + "grad_norm": 0.6229112148284912, + "learning_rate": 2.2171416857096928e-05, + "loss": 0.1205, + "step": 33443 + }, + { + "epoch": 0.7839800277782661, + "grad_norm": 0.46637460589408875, + "learning_rate": 2.216679178293456e-05, + "loss": 0.0515, + "step": 33444 + }, + { + "epoch": 0.7840034693530712, + "grad_norm": 0.4271913468837738, + "learning_rate": 2.2162167131097335e-05, + "loss": 0.0745, + "step": 33445 + }, + { + "epoch": 0.7840269109278761, + "grad_norm": 0.2161787748336792, + "learning_rate": 2.2157542901610405e-05, + "loss": 0.0439, + "step": 33446 + }, + { + "epoch": 0.7840503525026812, + "grad_norm": 0.5179349780082703, + "learning_rate": 2.2152919094498846e-05, + "loss": 0.0708, + "step": 33447 + }, + { + "epoch": 0.7840737940774861, + "grad_norm": 0.5630332827568054, + "learning_rate": 2.2148295709787727e-05, + "loss": 0.1295, + "step": 33448 + }, + { + "epoch": 0.7840972356522912, + "grad_norm": 0.2925724387168884, + "learning_rate": 2.2143672747502154e-05, + "loss": 0.0467, + "step": 33449 + }, + { + "epoch": 0.7841206772270961, + "grad_norm": 0.7399718165397644, + "learning_rate": 2.2139050207667165e-05, + "loss": 0.0888, + "step": 33450 + }, + { + "epoch": 0.7841441188019012, + "grad_norm": 0.36487525701522827, + "learning_rate": 2.213442809030791e-05, + "loss": 0.0721, + "step": 33451 + }, + { + "epoch": 0.7841675603767061, + "grad_norm": 0.16915558278560638, + "learning_rate": 2.2129806395449416e-05, + "loss": 0.0139, + "step": 33452 + }, + { + "epoch": 0.7841910019515111, + "grad_norm": 0.6466168761253357, + "learning_rate": 2.21251851231168e-05, + "loss": 0.1138, + "step": 33453 + }, + { + "epoch": 0.7842144435263161, + "grad_norm": 0.5918063521385193, + "learning_rate": 2.212056427333512e-05, + "loss": 0.5776, + "step": 33454 + }, + { + "epoch": 0.7842378851011211, + "grad_norm": 0.6613453030586243, + "learning_rate": 2.211594384612945e-05, + "loss": 0.0786, + "step": 33455 + }, + { + "epoch": 0.7842613266759261, + "grad_norm": 0.3866861164569855, + "learning_rate": 2.2111323841524867e-05, + "loss": 0.0892, + "step": 33456 + }, + { + "epoch": 0.7842847682507311, + "grad_norm": 0.37968793511390686, + "learning_rate": 2.2106704259546417e-05, + "loss": 0.0515, + "step": 33457 + }, + { + "epoch": 0.7843082098255361, + "grad_norm": 0.5527629256248474, + "learning_rate": 2.210208510021916e-05, + "loss": 0.4128, + "step": 33458 + }, + { + "epoch": 0.7843316514003411, + "grad_norm": 0.5935451984405518, + "learning_rate": 2.2097466363568208e-05, + "loss": 0.1858, + "step": 33459 + }, + { + "epoch": 0.784355092975146, + "grad_norm": 0.39098307490348816, + "learning_rate": 2.2092848049618598e-05, + "loss": 0.0303, + "step": 33460 + }, + { + "epoch": 0.7843785345499511, + "grad_norm": 0.511823296546936, + "learning_rate": 2.2088230158395373e-05, + "loss": 0.0947, + "step": 33461 + }, + { + "epoch": 0.784401976124756, + "grad_norm": 0.36892926692962646, + "learning_rate": 2.208361268992358e-05, + "loss": 0.0541, + "step": 33462 + }, + { + "epoch": 0.7844254176995611, + "grad_norm": 0.32652950286865234, + "learning_rate": 2.207899564422833e-05, + "loss": 0.3678, + "step": 33463 + }, + { + "epoch": 0.784448859274366, + "grad_norm": 0.43447259068489075, + "learning_rate": 2.2074379021334635e-05, + "loss": 0.0777, + "step": 33464 + }, + { + "epoch": 0.7844723008491711, + "grad_norm": 0.37865114212036133, + "learning_rate": 2.2069762821267516e-05, + "loss": 0.0538, + "step": 33465 + }, + { + "epoch": 0.784495742423976, + "grad_norm": 0.32427841424942017, + "learning_rate": 2.206514704405209e-05, + "loss": 0.066, + "step": 33466 + }, + { + "epoch": 0.7845191839987811, + "grad_norm": 0.6426756978034973, + "learning_rate": 2.206053168971337e-05, + "loss": 0.6213, + "step": 33467 + }, + { + "epoch": 0.784542625573586, + "grad_norm": 0.4018085300922394, + "learning_rate": 2.2055916758276384e-05, + "loss": 0.0624, + "step": 33468 + }, + { + "epoch": 0.784566067148391, + "grad_norm": 0.6665343642234802, + "learning_rate": 2.2051302249766203e-05, + "loss": 0.1224, + "step": 33469 + }, + { + "epoch": 0.784589508723196, + "grad_norm": 0.8172735571861267, + "learning_rate": 2.2046688164207807e-05, + "loss": 0.1018, + "step": 33470 + }, + { + "epoch": 0.784612950298001, + "grad_norm": 0.2333301603794098, + "learning_rate": 2.2042074501626307e-05, + "loss": 0.0281, + "step": 33471 + }, + { + "epoch": 0.784636391872806, + "grad_norm": 0.4026302993297577, + "learning_rate": 2.20374612620467e-05, + "loss": 0.0617, + "step": 33472 + }, + { + "epoch": 0.784659833447611, + "grad_norm": 0.5373356938362122, + "learning_rate": 2.2032848445494016e-05, + "loss": 0.1026, + "step": 33473 + }, + { + "epoch": 0.784683275022416, + "grad_norm": 0.44629716873168945, + "learning_rate": 2.2028236051993266e-05, + "loss": 0.0585, + "step": 33474 + }, + { + "epoch": 0.784706716597221, + "grad_norm": 0.1749071180820465, + "learning_rate": 2.2023624081569526e-05, + "loss": 0.0384, + "step": 33475 + }, + { + "epoch": 0.784730158172026, + "grad_norm": 0.5068687796592712, + "learning_rate": 2.2019012534247796e-05, + "loss": 0.1254, + "step": 33476 + }, + { + "epoch": 0.784753599746831, + "grad_norm": 0.5425533056259155, + "learning_rate": 2.2014401410053064e-05, + "loss": 0.0633, + "step": 33477 + }, + { + "epoch": 0.784777041321636, + "grad_norm": 0.11550435423851013, + "learning_rate": 2.2009790709010412e-05, + "loss": 0.0101, + "step": 33478 + }, + { + "epoch": 0.784800482896441, + "grad_norm": 0.7130046486854553, + "learning_rate": 2.2005180431144835e-05, + "loss": 0.1382, + "step": 33479 + }, + { + "epoch": 0.7848239244712459, + "grad_norm": 0.08129320293664932, + "learning_rate": 2.200057057648134e-05, + "loss": 0.0116, + "step": 33480 + }, + { + "epoch": 0.784847366046051, + "grad_norm": 0.36952805519104004, + "learning_rate": 2.199596114504494e-05, + "loss": 0.0213, + "step": 33481 + }, + { + "epoch": 0.7848708076208559, + "grad_norm": 0.1438869833946228, + "learning_rate": 2.1991352136860642e-05, + "loss": 0.0287, + "step": 33482 + }, + { + "epoch": 0.784894249195661, + "grad_norm": 0.32985156774520874, + "learning_rate": 2.1986743551953447e-05, + "loss": 0.0433, + "step": 33483 + }, + { + "epoch": 0.7849176907704659, + "grad_norm": 0.7206581234931946, + "learning_rate": 2.198213539034839e-05, + "loss": 0.1703, + "step": 33484 + }, + { + "epoch": 0.784941132345271, + "grad_norm": 0.48923733830451965, + "learning_rate": 2.1977527652070474e-05, + "loss": 0.0906, + "step": 33485 + }, + { + "epoch": 0.7849645739200759, + "grad_norm": 0.5304007530212402, + "learning_rate": 2.1972920337144675e-05, + "loss": 0.0712, + "step": 33486 + }, + { + "epoch": 0.784988015494881, + "grad_norm": 0.5850328207015991, + "learning_rate": 2.196831344559598e-05, + "loss": 0.0789, + "step": 33487 + }, + { + "epoch": 0.7850114570696859, + "grad_norm": 0.6811527609825134, + "learning_rate": 2.196370697744943e-05, + "loss": 0.1468, + "step": 33488 + }, + { + "epoch": 0.7850348986444909, + "grad_norm": 0.2000693529844284, + "learning_rate": 2.1959100932730016e-05, + "loss": 0.0173, + "step": 33489 + }, + { + "epoch": 0.7850583402192959, + "grad_norm": 0.19817408919334412, + "learning_rate": 2.1954495311462676e-05, + "loss": 0.0279, + "step": 33490 + }, + { + "epoch": 0.7850817817941009, + "grad_norm": 0.48304179310798645, + "learning_rate": 2.1949890113672466e-05, + "loss": 0.0988, + "step": 33491 + }, + { + "epoch": 0.7851052233689059, + "grad_norm": 0.4107488989830017, + "learning_rate": 2.1945285339384347e-05, + "loss": 0.0889, + "step": 33492 + }, + { + "epoch": 0.7851286649437109, + "grad_norm": 0.5695686936378479, + "learning_rate": 2.1940680988623307e-05, + "loss": 0.541, + "step": 33493 + }, + { + "epoch": 0.785152106518516, + "grad_norm": 0.18163728713989258, + "learning_rate": 2.193607706141433e-05, + "loss": 0.023, + "step": 33494 + }, + { + "epoch": 0.7851755480933209, + "grad_norm": 0.43961143493652344, + "learning_rate": 2.1931473557782388e-05, + "loss": 0.1666, + "step": 33495 + }, + { + "epoch": 0.785198989668126, + "grad_norm": 0.34914231300354004, + "learning_rate": 2.1926870477752437e-05, + "loss": 0.0731, + "step": 33496 + }, + { + "epoch": 0.7852224312429309, + "grad_norm": 0.6164426207542419, + "learning_rate": 2.1922267821349517e-05, + "loss": 0.0672, + "step": 33497 + }, + { + "epoch": 0.785245872817736, + "grad_norm": 0.35680291056632996, + "learning_rate": 2.191766558859857e-05, + "loss": 0.5096, + "step": 33498 + }, + { + "epoch": 0.7852693143925409, + "grad_norm": 0.2995893061161041, + "learning_rate": 2.1913063779524524e-05, + "loss": 0.0525, + "step": 33499 + }, + { + "epoch": 0.7852927559673459, + "grad_norm": 0.5498282313346863, + "learning_rate": 2.1908462394152428e-05, + "loss": 0.1647, + "step": 33500 + }, + { + "epoch": 0.7853161975421509, + "grad_norm": 0.480635404586792, + "learning_rate": 2.19038614325072e-05, + "loss": 0.1054, + "step": 33501 + }, + { + "epoch": 0.7853396391169559, + "grad_norm": 0.31702369451522827, + "learning_rate": 2.1899260894613826e-05, + "loss": 0.0236, + "step": 33502 + }, + { + "epoch": 0.7853630806917609, + "grad_norm": 0.24154949188232422, + "learning_rate": 2.189466078049722e-05, + "loss": 0.0244, + "step": 33503 + }, + { + "epoch": 0.7853865222665659, + "grad_norm": 0.15493334829807281, + "learning_rate": 2.1890061090182425e-05, + "loss": 0.0301, + "step": 33504 + }, + { + "epoch": 0.7854099638413709, + "grad_norm": 0.5004531145095825, + "learning_rate": 2.1885461823694343e-05, + "loss": 0.0966, + "step": 33505 + }, + { + "epoch": 0.7854334054161759, + "grad_norm": 0.4422878623008728, + "learning_rate": 2.188086298105795e-05, + "loss": 0.0401, + "step": 33506 + }, + { + "epoch": 0.7854568469909808, + "grad_norm": 0.42366138100624084, + "learning_rate": 2.1876264562298178e-05, + "loss": 0.4628, + "step": 33507 + }, + { + "epoch": 0.7854802885657859, + "grad_norm": 0.36869776248931885, + "learning_rate": 2.1871666567439975e-05, + "loss": 0.4577, + "step": 33508 + }, + { + "epoch": 0.7855037301405908, + "grad_norm": 0.5053325295448303, + "learning_rate": 2.186706899650832e-05, + "loss": 0.1149, + "step": 33509 + }, + { + "epoch": 0.7855271717153959, + "grad_norm": 0.25640618801116943, + "learning_rate": 2.1862471849528154e-05, + "loss": 0.0585, + "step": 33510 + }, + { + "epoch": 0.7855506132902008, + "grad_norm": 0.6016006469726562, + "learning_rate": 2.1857875126524386e-05, + "loss": 0.1446, + "step": 33511 + }, + { + "epoch": 0.7855740548650059, + "grad_norm": 0.41520705819129944, + "learning_rate": 2.1853278827522017e-05, + "loss": 0.0729, + "step": 33512 + }, + { + "epoch": 0.7855974964398108, + "grad_norm": 0.2904743254184723, + "learning_rate": 2.184868295254595e-05, + "loss": 0.0339, + "step": 33513 + }, + { + "epoch": 0.7856209380146159, + "grad_norm": 0.4778503477573395, + "learning_rate": 2.1844087501621125e-05, + "loss": 0.0395, + "step": 33514 + }, + { + "epoch": 0.7856443795894208, + "grad_norm": 0.4247760772705078, + "learning_rate": 2.1839492474772448e-05, + "loss": 0.0466, + "step": 33515 + }, + { + "epoch": 0.7856678211642258, + "grad_norm": 0.5204753279685974, + "learning_rate": 2.1834897872024917e-05, + "loss": 0.1129, + "step": 33516 + }, + { + "epoch": 0.7856912627390308, + "grad_norm": 0.5325473546981812, + "learning_rate": 2.1830303693403432e-05, + "loss": 0.1448, + "step": 33517 + }, + { + "epoch": 0.7857147043138358, + "grad_norm": 0.3873327970504761, + "learning_rate": 2.1825709938932903e-05, + "loss": 0.0571, + "step": 33518 + }, + { + "epoch": 0.7857381458886408, + "grad_norm": 0.5167956352233887, + "learning_rate": 2.1821116608638283e-05, + "loss": 0.0646, + "step": 33519 + }, + { + "epoch": 0.7857615874634458, + "grad_norm": 0.12087476998567581, + "learning_rate": 2.1816523702544478e-05, + "loss": 0.0227, + "step": 33520 + }, + { + "epoch": 0.7857850290382508, + "grad_norm": 0.4761104881763458, + "learning_rate": 2.1811931220676384e-05, + "loss": 0.0794, + "step": 33521 + }, + { + "epoch": 0.7858084706130558, + "grad_norm": 0.07101323455572128, + "learning_rate": 2.1807339163058972e-05, + "loss": 0.0121, + "step": 33522 + }, + { + "epoch": 0.7858319121878607, + "grad_norm": 0.2885656952857971, + "learning_rate": 2.1802747529717115e-05, + "loss": 0.0475, + "step": 33523 + }, + { + "epoch": 0.7858553537626658, + "grad_norm": 0.3610973060131073, + "learning_rate": 2.1798156320675778e-05, + "loss": 0.0435, + "step": 33524 + }, + { + "epoch": 0.7858787953374707, + "grad_norm": 0.573111891746521, + "learning_rate": 2.1793565535959838e-05, + "loss": 0.7684, + "step": 33525 + }, + { + "epoch": 0.7859022369122758, + "grad_norm": 0.2631193697452545, + "learning_rate": 2.178897517559422e-05, + "loss": 0.0321, + "step": 33526 + }, + { + "epoch": 0.7859256784870807, + "grad_norm": 0.43740496039390564, + "learning_rate": 2.178438523960381e-05, + "loss": 0.1428, + "step": 33527 + }, + { + "epoch": 0.7859491200618858, + "grad_norm": 0.2319382280111313, + "learning_rate": 2.177979572801351e-05, + "loss": 0.0176, + "step": 33528 + }, + { + "epoch": 0.7859725616366907, + "grad_norm": 0.5189213156700134, + "learning_rate": 2.1775206640848256e-05, + "loss": 0.0574, + "step": 33529 + }, + { + "epoch": 0.7859960032114958, + "grad_norm": 0.8967977166175842, + "learning_rate": 2.1770617978132934e-05, + "loss": 0.1088, + "step": 33530 + }, + { + "epoch": 0.7860194447863007, + "grad_norm": 0.21010006964206696, + "learning_rate": 2.1766029739892435e-05, + "loss": 0.0268, + "step": 33531 + }, + { + "epoch": 0.7860428863611058, + "grad_norm": 0.39463526010513306, + "learning_rate": 2.176144192615166e-05, + "loss": 0.4499, + "step": 33532 + }, + { + "epoch": 0.7860663279359107, + "grad_norm": 0.39714547991752625, + "learning_rate": 2.17568545369355e-05, + "loss": 0.0679, + "step": 33533 + }, + { + "epoch": 0.7860897695107157, + "grad_norm": 0.5867975950241089, + "learning_rate": 2.1752267572268827e-05, + "loss": 0.4995, + "step": 33534 + }, + { + "epoch": 0.7861132110855207, + "grad_norm": 0.06771831214427948, + "learning_rate": 2.1747681032176535e-05, + "loss": 0.021, + "step": 33535 + }, + { + "epoch": 0.7861366526603257, + "grad_norm": 0.5992892980575562, + "learning_rate": 2.174309491668357e-05, + "loss": 0.6207, + "step": 33536 + }, + { + "epoch": 0.7861600942351307, + "grad_norm": 0.29939907789230347, + "learning_rate": 2.1738509225814762e-05, + "loss": 0.0319, + "step": 33537 + }, + { + "epoch": 0.7861835358099357, + "grad_norm": 0.09609147906303406, + "learning_rate": 2.1733923959595015e-05, + "loss": 0.0121, + "step": 33538 + }, + { + "epoch": 0.7862069773847407, + "grad_norm": 0.4257752001285553, + "learning_rate": 2.1729339118049185e-05, + "loss": 0.1218, + "step": 33539 + }, + { + "epoch": 0.7862304189595457, + "grad_norm": 0.4105648100376129, + "learning_rate": 2.1724754701202144e-05, + "loss": 0.3465, + "step": 33540 + }, + { + "epoch": 0.7862538605343506, + "grad_norm": 0.12768319249153137, + "learning_rate": 2.1720170709078802e-05, + "loss": 0.0144, + "step": 33541 + }, + { + "epoch": 0.7862773021091557, + "grad_norm": 0.3785734474658966, + "learning_rate": 2.1715587141704018e-05, + "loss": 0.0507, + "step": 33542 + }, + { + "epoch": 0.7863007436839606, + "grad_norm": 0.32199034094810486, + "learning_rate": 2.1711003999102664e-05, + "loss": 0.0542, + "step": 33543 + }, + { + "epoch": 0.7863241852587657, + "grad_norm": 0.10835100710391998, + "learning_rate": 2.1706421281299595e-05, + "loss": 0.015, + "step": 33544 + }, + { + "epoch": 0.7863476268335707, + "grad_norm": 0.3827160596847534, + "learning_rate": 2.170183898831969e-05, + "loss": 0.0793, + "step": 33545 + }, + { + "epoch": 0.7863710684083757, + "grad_norm": 0.3826553225517273, + "learning_rate": 2.1697257120187787e-05, + "loss": 0.0526, + "step": 33546 + }, + { + "epoch": 0.7863945099831807, + "grad_norm": 0.77959144115448, + "learning_rate": 2.1692675676928763e-05, + "loss": 0.1515, + "step": 33547 + }, + { + "epoch": 0.7864179515579857, + "grad_norm": 0.4747293293476105, + "learning_rate": 2.1688094658567514e-05, + "loss": 0.0606, + "step": 33548 + }, + { + "epoch": 0.7864413931327907, + "grad_norm": 0.16058118641376495, + "learning_rate": 2.1683514065128863e-05, + "loss": 0.0257, + "step": 33549 + }, + { + "epoch": 0.7864648347075957, + "grad_norm": 0.10573848336935043, + "learning_rate": 2.1678933896637667e-05, + "loss": 0.0183, + "step": 33550 + }, + { + "epoch": 0.7864882762824007, + "grad_norm": 0.33260515332221985, + "learning_rate": 2.167435415311877e-05, + "loss": 0.0456, + "step": 33551 + }, + { + "epoch": 0.7865117178572056, + "grad_norm": 0.2922401428222656, + "learning_rate": 2.1669774834597036e-05, + "loss": 0.045, + "step": 33552 + }, + { + "epoch": 0.7865351594320107, + "grad_norm": 0.38677945733070374, + "learning_rate": 2.166519594109728e-05, + "loss": 0.0655, + "step": 33553 + }, + { + "epoch": 0.7865586010068156, + "grad_norm": 0.5528755187988281, + "learning_rate": 2.1660617472644395e-05, + "loss": 0.112, + "step": 33554 + }, + { + "epoch": 0.7865820425816207, + "grad_norm": 0.6497868895530701, + "learning_rate": 2.165603942926321e-05, + "loss": 0.1355, + "step": 33555 + }, + { + "epoch": 0.7866054841564256, + "grad_norm": 0.3432541787624359, + "learning_rate": 2.1651461810978557e-05, + "loss": 0.0465, + "step": 33556 + }, + { + "epoch": 0.7866289257312307, + "grad_norm": 0.3763248920440674, + "learning_rate": 2.1646884617815265e-05, + "loss": 0.0703, + "step": 33557 + }, + { + "epoch": 0.7866523673060356, + "grad_norm": 0.42272937297821045, + "learning_rate": 2.1642307849798195e-05, + "loss": 0.0394, + "step": 33558 + }, + { + "epoch": 0.7866758088808407, + "grad_norm": 0.30021604895591736, + "learning_rate": 2.163773150695213e-05, + "loss": 0.0628, + "step": 33559 + }, + { + "epoch": 0.7866992504556456, + "grad_norm": 0.32699549198150635, + "learning_rate": 2.1633155589301935e-05, + "loss": 0.0345, + "step": 33560 + }, + { + "epoch": 0.7867226920304506, + "grad_norm": 0.4339250922203064, + "learning_rate": 2.1628580096872475e-05, + "loss": 0.0498, + "step": 33561 + }, + { + "epoch": 0.7867461336052556, + "grad_norm": 0.2452067881822586, + "learning_rate": 2.1624005029688544e-05, + "loss": 0.0271, + "step": 33562 + }, + { + "epoch": 0.7867695751800606, + "grad_norm": 0.49098634719848633, + "learning_rate": 2.1619430387774963e-05, + "loss": 0.0886, + "step": 33563 + }, + { + "epoch": 0.7867930167548656, + "grad_norm": 0.14514464139938354, + "learning_rate": 2.161485617115655e-05, + "loss": 0.021, + "step": 33564 + }, + { + "epoch": 0.7868164583296706, + "grad_norm": 0.3827359080314636, + "learning_rate": 2.161028237985814e-05, + "loss": 0.0185, + "step": 33565 + }, + { + "epoch": 0.7868398999044756, + "grad_norm": 0.6709076762199402, + "learning_rate": 2.1605709013904507e-05, + "loss": 0.1448, + "step": 33566 + }, + { + "epoch": 0.7868633414792806, + "grad_norm": 0.12152868509292603, + "learning_rate": 2.1601136073320538e-05, + "loss": 0.019, + "step": 33567 + }, + { + "epoch": 0.7868867830540855, + "grad_norm": 0.1361647993326187, + "learning_rate": 2.1596563558131e-05, + "loss": 0.0141, + "step": 33568 + }, + { + "epoch": 0.7869102246288906, + "grad_norm": 0.46063289046287537, + "learning_rate": 2.1591991468360717e-05, + "loss": 0.0689, + "step": 33569 + }, + { + "epoch": 0.7869336662036955, + "grad_norm": 0.22467400133609772, + "learning_rate": 2.1587419804034493e-05, + "loss": 0.0496, + "step": 33570 + }, + { + "epoch": 0.7869571077785006, + "grad_norm": 0.8763799667358398, + "learning_rate": 2.1582848565177138e-05, + "loss": 0.1152, + "step": 33571 + }, + { + "epoch": 0.7869805493533055, + "grad_norm": 0.14499981701374054, + "learning_rate": 2.1578277751813413e-05, + "loss": 0.0257, + "step": 33572 + }, + { + "epoch": 0.7870039909281106, + "grad_norm": 0.43533387780189514, + "learning_rate": 2.1573707363968166e-05, + "loss": 0.0949, + "step": 33573 + }, + { + "epoch": 0.7870274325029155, + "grad_norm": 0.3476375639438629, + "learning_rate": 2.1569137401666216e-05, + "loss": 0.0758, + "step": 33574 + }, + { + "epoch": 0.7870508740777206, + "grad_norm": 0.13357442617416382, + "learning_rate": 2.156456786493233e-05, + "loss": 0.0255, + "step": 33575 + }, + { + "epoch": 0.7870743156525255, + "grad_norm": 0.1640813797712326, + "learning_rate": 2.15599987537913e-05, + "loss": 0.0402, + "step": 33576 + }, + { + "epoch": 0.7870977572273306, + "grad_norm": 0.47820910811424255, + "learning_rate": 2.1555430068267934e-05, + "loss": 0.1003, + "step": 33577 + }, + { + "epoch": 0.7871211988021355, + "grad_norm": 0.5681387782096863, + "learning_rate": 2.1550861808386968e-05, + "loss": 0.6335, + "step": 33578 + }, + { + "epoch": 0.7871446403769405, + "grad_norm": 0.3508279025554657, + "learning_rate": 2.1546293974173272e-05, + "loss": 0.0686, + "step": 33579 + }, + { + "epoch": 0.7871680819517455, + "grad_norm": 0.18817077577114105, + "learning_rate": 2.154172656565159e-05, + "loss": 0.0278, + "step": 33580 + }, + { + "epoch": 0.7871915235265505, + "grad_norm": 0.2280024141073227, + "learning_rate": 2.153715958284669e-05, + "loss": 0.0275, + "step": 33581 + }, + { + "epoch": 0.7872149651013555, + "grad_norm": 0.5969537496566772, + "learning_rate": 2.1532593025783377e-05, + "loss": 0.0771, + "step": 33582 + }, + { + "epoch": 0.7872384066761605, + "grad_norm": 0.25054511427879333, + "learning_rate": 2.152802689448643e-05, + "loss": 0.0393, + "step": 33583 + }, + { + "epoch": 0.7872618482509655, + "grad_norm": 0.13957765698432922, + "learning_rate": 2.1523461188980577e-05, + "loss": 0.0217, + "step": 33584 + }, + { + "epoch": 0.7872852898257705, + "grad_norm": 0.3294655382633209, + "learning_rate": 2.1518895909290636e-05, + "loss": 0.0848, + "step": 33585 + }, + { + "epoch": 0.7873087314005754, + "grad_norm": 0.17791059613227844, + "learning_rate": 2.1514331055441396e-05, + "loss": 0.0179, + "step": 33586 + }, + { + "epoch": 0.7873321729753805, + "grad_norm": 0.12718644738197327, + "learning_rate": 2.1509766627457605e-05, + "loss": 0.0291, + "step": 33587 + }, + { + "epoch": 0.7873556145501854, + "grad_norm": 0.2533433437347412, + "learning_rate": 2.150520262536403e-05, + "loss": 0.3012, + "step": 33588 + }, + { + "epoch": 0.7873790561249905, + "grad_norm": 0.3099662959575653, + "learning_rate": 2.1500639049185423e-05, + "loss": 0.0544, + "step": 33589 + }, + { + "epoch": 0.7874024976997954, + "grad_norm": 0.2244311273097992, + "learning_rate": 2.1496075898946567e-05, + "loss": 0.0376, + "step": 33590 + }, + { + "epoch": 0.7874259392746005, + "grad_norm": 0.3132827579975128, + "learning_rate": 2.1491513174672175e-05, + "loss": 0.0626, + "step": 33591 + }, + { + "epoch": 0.7874493808494054, + "grad_norm": 0.5428159832954407, + "learning_rate": 2.148695087638707e-05, + "loss": 0.0887, + "step": 33592 + }, + { + "epoch": 0.7874728224242105, + "grad_norm": 0.16461701691150665, + "learning_rate": 2.1482389004115976e-05, + "loss": 0.0353, + "step": 33593 + }, + { + "epoch": 0.7874962639990154, + "grad_norm": 0.5594035983085632, + "learning_rate": 2.1477827557883645e-05, + "loss": 0.4854, + "step": 33594 + }, + { + "epoch": 0.7875197055738204, + "grad_norm": 0.5050202012062073, + "learning_rate": 2.147326653771483e-05, + "loss": 0.073, + "step": 33595 + }, + { + "epoch": 0.7875431471486255, + "grad_norm": 0.16668613255023956, + "learning_rate": 2.146870594363426e-05, + "loss": 0.0212, + "step": 33596 + }, + { + "epoch": 0.7875665887234304, + "grad_norm": 0.5506035685539246, + "learning_rate": 2.146414577566672e-05, + "loss": 0.0522, + "step": 33597 + }, + { + "epoch": 0.7875900302982355, + "grad_norm": 0.4560471177101135, + "learning_rate": 2.145958603383691e-05, + "loss": 0.1098, + "step": 33598 + }, + { + "epoch": 0.7876134718730404, + "grad_norm": 0.40391451120376587, + "learning_rate": 2.1455026718169612e-05, + "loss": 0.0471, + "step": 33599 + }, + { + "epoch": 0.7876369134478455, + "grad_norm": 0.18984651565551758, + "learning_rate": 2.1450467828689558e-05, + "loss": 0.0447, + "step": 33600 + }, + { + "epoch": 0.7876603550226504, + "grad_norm": 0.4501417279243469, + "learning_rate": 2.1445909365421467e-05, + "loss": 0.5005, + "step": 33601 + }, + { + "epoch": 0.7876837965974555, + "grad_norm": 0.5949227809906006, + "learning_rate": 2.144135132839009e-05, + "loss": 0.0858, + "step": 33602 + }, + { + "epoch": 0.7877072381722604, + "grad_norm": 0.5995817184448242, + "learning_rate": 2.1436793717620142e-05, + "loss": 0.1301, + "step": 33603 + }, + { + "epoch": 0.7877306797470655, + "grad_norm": 0.30728527903556824, + "learning_rate": 2.143223653313633e-05, + "loss": 0.0305, + "step": 33604 + }, + { + "epoch": 0.7877541213218704, + "grad_norm": 0.26767468452453613, + "learning_rate": 2.1427679774963448e-05, + "loss": 0.0199, + "step": 33605 + }, + { + "epoch": 0.7877775628966754, + "grad_norm": 0.18263018131256104, + "learning_rate": 2.1423123443126182e-05, + "loss": 0.0257, + "step": 33606 + }, + { + "epoch": 0.7878010044714804, + "grad_norm": 0.3865322172641754, + "learning_rate": 2.141856753764926e-05, + "loss": 0.0225, + "step": 33607 + }, + { + "epoch": 0.7878244460462854, + "grad_norm": 0.12194695323705673, + "learning_rate": 2.1414012058557364e-05, + "loss": 0.0118, + "step": 33608 + }, + { + "epoch": 0.7878478876210904, + "grad_norm": 0.4136655330657959, + "learning_rate": 2.1409457005875276e-05, + "loss": 0.0819, + "step": 33609 + }, + { + "epoch": 0.7878713291958954, + "grad_norm": 0.5413221716880798, + "learning_rate": 2.140490237962769e-05, + "loss": 0.1014, + "step": 33610 + }, + { + "epoch": 0.7878947707707004, + "grad_norm": 0.29935720562934875, + "learning_rate": 2.140034817983929e-05, + "loss": 0.0441, + "step": 33611 + }, + { + "epoch": 0.7879182123455054, + "grad_norm": 0.29588544368743896, + "learning_rate": 2.139579440653483e-05, + "loss": 0.0413, + "step": 33612 + }, + { + "epoch": 0.7879416539203103, + "grad_norm": 0.4792070686817169, + "learning_rate": 2.1391241059739e-05, + "loss": 0.0897, + "step": 33613 + }, + { + "epoch": 0.7879650954951154, + "grad_norm": 0.1607985943555832, + "learning_rate": 2.1386688139476508e-05, + "loss": 0.0264, + "step": 33614 + }, + { + "epoch": 0.7879885370699203, + "grad_norm": 0.38725051283836365, + "learning_rate": 2.138213564577206e-05, + "loss": 0.0667, + "step": 33615 + }, + { + "epoch": 0.7880119786447254, + "grad_norm": 0.12631535530090332, + "learning_rate": 2.1377583578650328e-05, + "loss": 0.0205, + "step": 33616 + }, + { + "epoch": 0.7880354202195303, + "grad_norm": 0.267238587141037, + "learning_rate": 2.1373031938136055e-05, + "loss": 0.0477, + "step": 33617 + }, + { + "epoch": 0.7880588617943354, + "grad_norm": 0.7524234056472778, + "learning_rate": 2.1368480724253935e-05, + "loss": 0.0709, + "step": 33618 + }, + { + "epoch": 0.7880823033691403, + "grad_norm": 0.5363144874572754, + "learning_rate": 2.1363929937028636e-05, + "loss": 0.0816, + "step": 33619 + }, + { + "epoch": 0.7881057449439454, + "grad_norm": 0.4505622386932373, + "learning_rate": 2.135937957648485e-05, + "loss": 0.0978, + "step": 33620 + }, + { + "epoch": 0.7881291865187503, + "grad_norm": 0.4242682456970215, + "learning_rate": 2.13548296426473e-05, + "loss": 0.0666, + "step": 33621 + }, + { + "epoch": 0.7881526280935554, + "grad_norm": 0.5269678831100464, + "learning_rate": 2.1350280135540657e-05, + "loss": 0.0704, + "step": 33622 + }, + { + "epoch": 0.7881760696683603, + "grad_norm": 0.23021748661994934, + "learning_rate": 2.1345731055189576e-05, + "loss": 0.0188, + "step": 33623 + }, + { + "epoch": 0.7881995112431653, + "grad_norm": 0.0828556939959526, + "learning_rate": 2.1341182401618807e-05, + "loss": 0.0141, + "step": 33624 + }, + { + "epoch": 0.7882229528179703, + "grad_norm": 0.4017128348350525, + "learning_rate": 2.133663417485299e-05, + "loss": 0.0995, + "step": 33625 + }, + { + "epoch": 0.7882463943927753, + "grad_norm": 0.5602160692214966, + "learning_rate": 2.1332086374916803e-05, + "loss": 0.4461, + "step": 33626 + }, + { + "epoch": 0.7882698359675803, + "grad_norm": 0.2913568913936615, + "learning_rate": 2.132753900183493e-05, + "loss": 0.0653, + "step": 33627 + }, + { + "epoch": 0.7882932775423853, + "grad_norm": 0.5117395520210266, + "learning_rate": 2.132299205563204e-05, + "loss": 0.1236, + "step": 33628 + }, + { + "epoch": 0.7883167191171903, + "grad_norm": 0.20266997814178467, + "learning_rate": 2.1318445536332787e-05, + "loss": 0.0336, + "step": 33629 + }, + { + "epoch": 0.7883401606919953, + "grad_norm": 0.25559893250465393, + "learning_rate": 2.1313899443961884e-05, + "loss": 0.043, + "step": 33630 + }, + { + "epoch": 0.7883636022668002, + "grad_norm": 0.32077646255493164, + "learning_rate": 2.130935377854397e-05, + "loss": 0.0489, + "step": 33631 + }, + { + "epoch": 0.7883870438416053, + "grad_norm": 0.047984328120946884, + "learning_rate": 2.1304808540103693e-05, + "loss": 0.0035, + "step": 33632 + }, + { + "epoch": 0.7884104854164102, + "grad_norm": 0.37080442905426025, + "learning_rate": 2.130026372866577e-05, + "loss": 0.0644, + "step": 33633 + }, + { + "epoch": 0.7884339269912153, + "grad_norm": 0.308165967464447, + "learning_rate": 2.1295719344254815e-05, + "loss": 0.0436, + "step": 33634 + }, + { + "epoch": 0.7884573685660202, + "grad_norm": 0.344938188791275, + "learning_rate": 2.129117538689551e-05, + "loss": 0.0415, + "step": 33635 + }, + { + "epoch": 0.7884808101408253, + "grad_norm": 0.14660422503948212, + "learning_rate": 2.1286631856612472e-05, + "loss": 0.0244, + "step": 33636 + }, + { + "epoch": 0.7885042517156302, + "grad_norm": 0.5657244324684143, + "learning_rate": 2.1282088753430407e-05, + "loss": 0.0459, + "step": 33637 + }, + { + "epoch": 0.7885276932904353, + "grad_norm": 0.5694532990455627, + "learning_rate": 2.1277546077373946e-05, + "loss": 0.0774, + "step": 33638 + }, + { + "epoch": 0.7885511348652402, + "grad_norm": 0.3524153530597687, + "learning_rate": 2.127300382846773e-05, + "loss": 0.0508, + "step": 33639 + }, + { + "epoch": 0.7885745764400452, + "grad_norm": 0.12163325399160385, + "learning_rate": 2.1268462006736413e-05, + "loss": 0.0133, + "step": 33640 + }, + { + "epoch": 0.7885980180148502, + "grad_norm": 0.08657706528902054, + "learning_rate": 2.1263920612204626e-05, + "loss": 0.0118, + "step": 33641 + }, + { + "epoch": 0.7886214595896552, + "grad_norm": 0.4997928738594055, + "learning_rate": 2.1259379644897002e-05, + "loss": 0.043, + "step": 33642 + }, + { + "epoch": 0.7886449011644602, + "grad_norm": 0.25193288922309875, + "learning_rate": 2.125483910483822e-05, + "loss": 0.0515, + "step": 33643 + }, + { + "epoch": 0.7886683427392652, + "grad_norm": 0.12519796192646027, + "learning_rate": 2.1250298992052896e-05, + "loss": 0.015, + "step": 33644 + }, + { + "epoch": 0.7886917843140702, + "grad_norm": 0.3222901523113251, + "learning_rate": 2.124575930656564e-05, + "loss": 0.0299, + "step": 33645 + }, + { + "epoch": 0.7887152258888752, + "grad_norm": 0.12963007390499115, + "learning_rate": 2.1241220048401134e-05, + "loss": 0.0294, + "step": 33646 + }, + { + "epoch": 0.7887386674636803, + "grad_norm": 0.44248977303504944, + "learning_rate": 2.1236681217583974e-05, + "loss": 0.1209, + "step": 33647 + }, + { + "epoch": 0.7887621090384852, + "grad_norm": 0.2963268458843231, + "learning_rate": 2.123214281413881e-05, + "loss": 0.0515, + "step": 33648 + }, + { + "epoch": 0.7887855506132903, + "grad_norm": 0.3602486252784729, + "learning_rate": 2.122760483809022e-05, + "loss": 0.0595, + "step": 33649 + }, + { + "epoch": 0.7888089921880952, + "grad_norm": 0.5750418901443481, + "learning_rate": 2.1223067289462884e-05, + "loss": 0.1, + "step": 33650 + }, + { + "epoch": 0.7888324337629002, + "grad_norm": 0.6200474500656128, + "learning_rate": 2.1218530168281414e-05, + "loss": 0.073, + "step": 33651 + }, + { + "epoch": 0.7888558753377052, + "grad_norm": 0.11895602941513062, + "learning_rate": 2.1213993474570403e-05, + "loss": 0.022, + "step": 33652 + }, + { + "epoch": 0.7888793169125102, + "grad_norm": 0.29471665620803833, + "learning_rate": 2.120945720835448e-05, + "loss": 0.0385, + "step": 33653 + }, + { + "epoch": 0.7889027584873152, + "grad_norm": 0.17753787338733673, + "learning_rate": 2.1204921369658228e-05, + "loss": 0.0264, + "step": 33654 + }, + { + "epoch": 0.7889262000621202, + "grad_norm": 0.48689597845077515, + "learning_rate": 2.120038595850632e-05, + "loss": 0.6626, + "step": 33655 + }, + { + "epoch": 0.7889496416369252, + "grad_norm": 0.311201810836792, + "learning_rate": 2.1195850974923336e-05, + "loss": 0.0228, + "step": 33656 + }, + { + "epoch": 0.7889730832117302, + "grad_norm": 0.41060715913772583, + "learning_rate": 2.1191316418933848e-05, + "loss": 0.0674, + "step": 33657 + }, + { + "epoch": 0.7889965247865351, + "grad_norm": 0.279457688331604, + "learning_rate": 2.1186782290562522e-05, + "loss": 0.2105, + "step": 33658 + }, + { + "epoch": 0.7890199663613402, + "grad_norm": 0.17916448414325714, + "learning_rate": 2.118224858983393e-05, + "loss": 0.0279, + "step": 33659 + }, + { + "epoch": 0.7890434079361451, + "grad_norm": 0.353679895401001, + "learning_rate": 2.1177715316772673e-05, + "loss": 0.0683, + "step": 33660 + }, + { + "epoch": 0.7890668495109502, + "grad_norm": 0.26557448506355286, + "learning_rate": 2.117318247140332e-05, + "loss": 0.0591, + "step": 33661 + }, + { + "epoch": 0.7890902910857551, + "grad_norm": 0.36952775716781616, + "learning_rate": 2.1168650053750528e-05, + "loss": 0.0589, + "step": 33662 + }, + { + "epoch": 0.7891137326605602, + "grad_norm": 0.4669050872325897, + "learning_rate": 2.1164118063838856e-05, + "loss": 0.0854, + "step": 33663 + }, + { + "epoch": 0.7891371742353651, + "grad_norm": 0.7702537775039673, + "learning_rate": 2.1159586501692897e-05, + "loss": 0.1143, + "step": 33664 + }, + { + "epoch": 0.7891606158101702, + "grad_norm": 0.3752058446407318, + "learning_rate": 2.1155055367337227e-05, + "loss": 0.0735, + "step": 33665 + }, + { + "epoch": 0.7891840573849751, + "grad_norm": 0.6634159684181213, + "learning_rate": 2.1150524660796456e-05, + "loss": 0.1295, + "step": 33666 + }, + { + "epoch": 0.7892074989597802, + "grad_norm": 0.2839606702327728, + "learning_rate": 2.1145994382095123e-05, + "loss": 0.0554, + "step": 33667 + }, + { + "epoch": 0.7892309405345851, + "grad_norm": 0.13018092513084412, + "learning_rate": 2.114146453125787e-05, + "loss": 0.0229, + "step": 33668 + }, + { + "epoch": 0.7892543821093901, + "grad_norm": 0.8713784217834473, + "learning_rate": 2.113693510830922e-05, + "loss": 0.1015, + "step": 33669 + }, + { + "epoch": 0.7892778236841951, + "grad_norm": 0.4448651969432831, + "learning_rate": 2.1132406113273805e-05, + "loss": 0.0744, + "step": 33670 + }, + { + "epoch": 0.7893012652590001, + "grad_norm": 0.6359485387802124, + "learning_rate": 2.1127877546176177e-05, + "loss": 0.0823, + "step": 33671 + }, + { + "epoch": 0.7893247068338051, + "grad_norm": 0.2547585070133209, + "learning_rate": 2.1123349407040892e-05, + "loss": 0.0777, + "step": 33672 + }, + { + "epoch": 0.7893481484086101, + "grad_norm": 0.42561325430870056, + "learning_rate": 2.1118821695892545e-05, + "loss": 0.0534, + "step": 33673 + }, + { + "epoch": 0.7893715899834151, + "grad_norm": 0.45208561420440674, + "learning_rate": 2.1114294412755654e-05, + "loss": 0.4645, + "step": 33674 + }, + { + "epoch": 0.7893950315582201, + "grad_norm": 0.13324333727359772, + "learning_rate": 2.1109767557654846e-05, + "loss": 0.0256, + "step": 33675 + }, + { + "epoch": 0.789418473133025, + "grad_norm": 0.1716730147600174, + "learning_rate": 2.110524113061466e-05, + "loss": 0.02, + "step": 33676 + }, + { + "epoch": 0.7894419147078301, + "grad_norm": 0.17394937574863434, + "learning_rate": 2.1100715131659654e-05, + "loss": 0.0221, + "step": 33677 + }, + { + "epoch": 0.789465356282635, + "grad_norm": 0.8895725011825562, + "learning_rate": 2.1096189560814385e-05, + "loss": 0.1662, + "step": 33678 + }, + { + "epoch": 0.7894887978574401, + "grad_norm": 0.5568966865539551, + "learning_rate": 2.1091664418103406e-05, + "loss": 0.0862, + "step": 33679 + }, + { + "epoch": 0.789512239432245, + "grad_norm": 0.13056904077529907, + "learning_rate": 2.108713970355126e-05, + "loss": 0.0181, + "step": 33680 + }, + { + "epoch": 0.7895356810070501, + "grad_norm": 0.19621095061302185, + "learning_rate": 2.1082615417182505e-05, + "loss": 0.0199, + "step": 33681 + }, + { + "epoch": 0.789559122581855, + "grad_norm": 0.16159026324748993, + "learning_rate": 2.1078091559021717e-05, + "loss": 0.0175, + "step": 33682 + }, + { + "epoch": 0.7895825641566601, + "grad_norm": 0.3589235842227936, + "learning_rate": 2.1073568129093435e-05, + "loss": 0.0536, + "step": 33683 + }, + { + "epoch": 0.789606005731465, + "grad_norm": 0.250016450881958, + "learning_rate": 2.1069045127422193e-05, + "loss": 0.0365, + "step": 33684 + }, + { + "epoch": 0.78962944730627, + "grad_norm": 0.5801045298576355, + "learning_rate": 2.1064522554032516e-05, + "loss": 0.051, + "step": 33685 + }, + { + "epoch": 0.789652888881075, + "grad_norm": 0.3451993763446808, + "learning_rate": 2.1060000408948976e-05, + "loss": 0.0598, + "step": 33686 + }, + { + "epoch": 0.78967633045588, + "grad_norm": 0.5098230838775635, + "learning_rate": 2.105547869219605e-05, + "loss": 0.119, + "step": 33687 + }, + { + "epoch": 0.789699772030685, + "grad_norm": 0.4914032220840454, + "learning_rate": 2.1050957403798354e-05, + "loss": 0.4443, + "step": 33688 + }, + { + "epoch": 0.78972321360549, + "grad_norm": 0.6492440700531006, + "learning_rate": 2.104643654378039e-05, + "loss": 0.1355, + "step": 33689 + }, + { + "epoch": 0.789746655180295, + "grad_norm": 0.5464791655540466, + "learning_rate": 2.104191611216667e-05, + "loss": 0.0428, + "step": 33690 + }, + { + "epoch": 0.7897700967551, + "grad_norm": 0.9819116592407227, + "learning_rate": 2.1037396108981734e-05, + "loss": 0.0835, + "step": 33691 + }, + { + "epoch": 0.789793538329905, + "grad_norm": 0.1846611052751541, + "learning_rate": 2.1032876534250078e-05, + "loss": 0.0327, + "step": 33692 + }, + { + "epoch": 0.78981697990471, + "grad_norm": 0.11301612854003906, + "learning_rate": 2.102835738799629e-05, + "loss": 0.0138, + "step": 33693 + }, + { + "epoch": 0.7898404214795149, + "grad_norm": 0.17723312973976135, + "learning_rate": 2.1023838670244822e-05, + "loss": 0.0374, + "step": 33694 + }, + { + "epoch": 0.78986386305432, + "grad_norm": 0.4678628146648407, + "learning_rate": 2.1019320381020257e-05, + "loss": 0.0157, + "step": 33695 + }, + { + "epoch": 0.7898873046291249, + "grad_norm": 0.3319489657878876, + "learning_rate": 2.1014802520347077e-05, + "loss": 0.0375, + "step": 33696 + }, + { + "epoch": 0.78991074620393, + "grad_norm": 0.29172244668006897, + "learning_rate": 2.1010285088249805e-05, + "loss": 0.0363, + "step": 33697 + }, + { + "epoch": 0.789934187778735, + "grad_norm": 0.19439202547073364, + "learning_rate": 2.100576808475294e-05, + "loss": 0.0331, + "step": 33698 + }, + { + "epoch": 0.78995762935354, + "grad_norm": 0.3921303451061249, + "learning_rate": 2.100125150988097e-05, + "loss": 0.0561, + "step": 33699 + }, + { + "epoch": 0.789981070928345, + "grad_norm": 0.24052736163139343, + "learning_rate": 2.099673536365846e-05, + "loss": 0.0181, + "step": 33700 + }, + { + "epoch": 0.79000451250315, + "grad_norm": 0.7584171295166016, + "learning_rate": 2.099221964610989e-05, + "loss": 0.1595, + "step": 33701 + }, + { + "epoch": 0.790027954077955, + "grad_norm": 0.6713016033172607, + "learning_rate": 2.098770435725975e-05, + "loss": 0.1749, + "step": 33702 + }, + { + "epoch": 0.79005139565276, + "grad_norm": 0.1634693592786789, + "learning_rate": 2.098318949713256e-05, + "loss": 0.0315, + "step": 33703 + }, + { + "epoch": 0.790074837227565, + "grad_norm": 0.3129090368747711, + "learning_rate": 2.0978675065752794e-05, + "loss": 0.0582, + "step": 33704 + }, + { + "epoch": 0.7900982788023699, + "grad_norm": 0.410521537065506, + "learning_rate": 2.0974161063144937e-05, + "loss": 0.1005, + "step": 33705 + }, + { + "epoch": 0.790121720377175, + "grad_norm": 0.3070673942565918, + "learning_rate": 2.09696474893335e-05, + "loss": 0.0579, + "step": 33706 + }, + { + "epoch": 0.7901451619519799, + "grad_norm": 0.2894454002380371, + "learning_rate": 2.0965134344343006e-05, + "loss": 0.0301, + "step": 33707 + }, + { + "epoch": 0.790168603526785, + "grad_norm": 0.2882939279079437, + "learning_rate": 2.0960621628197918e-05, + "loss": 0.0508, + "step": 33708 + }, + { + "epoch": 0.7901920451015899, + "grad_norm": 0.4025862216949463, + "learning_rate": 2.0956109340922723e-05, + "loss": 0.0743, + "step": 33709 + }, + { + "epoch": 0.790215486676395, + "grad_norm": 0.4470635950565338, + "learning_rate": 2.0951597482541896e-05, + "loss": 0.6516, + "step": 33710 + }, + { + "epoch": 0.7902389282511999, + "grad_norm": 0.22456598281860352, + "learning_rate": 2.0947086053079922e-05, + "loss": 0.0292, + "step": 33711 + }, + { + "epoch": 0.790262369826005, + "grad_norm": 0.10378328710794449, + "learning_rate": 2.0942575052561264e-05, + "loss": 0.021, + "step": 33712 + }, + { + "epoch": 0.7902858114008099, + "grad_norm": 0.3384835720062256, + "learning_rate": 2.0938064481010434e-05, + "loss": 0.037, + "step": 33713 + }, + { + "epoch": 0.7903092529756149, + "grad_norm": 0.1238008663058281, + "learning_rate": 2.093355433845189e-05, + "loss": 0.0076, + "step": 33714 + }, + { + "epoch": 0.7903326945504199, + "grad_norm": 0.19067546725273132, + "learning_rate": 2.092904462491011e-05, + "loss": 0.0205, + "step": 33715 + }, + { + "epoch": 0.7903561361252249, + "grad_norm": 0.27627307176589966, + "learning_rate": 2.0924535340409546e-05, + "loss": 0.0385, + "step": 33716 + }, + { + "epoch": 0.7903795777000299, + "grad_norm": 0.26239481568336487, + "learning_rate": 2.0920026484974688e-05, + "loss": 0.04, + "step": 33717 + }, + { + "epoch": 0.7904030192748349, + "grad_norm": 0.5797544717788696, + "learning_rate": 2.0915518058629956e-05, + "loss": 0.5025, + "step": 33718 + }, + { + "epoch": 0.7904264608496399, + "grad_norm": 0.4030584692955017, + "learning_rate": 2.091101006139985e-05, + "loss": 0.0439, + "step": 33719 + }, + { + "epoch": 0.7904499024244449, + "grad_norm": 0.18450920283794403, + "learning_rate": 2.0906502493308854e-05, + "loss": 0.0366, + "step": 33720 + }, + { + "epoch": 0.7904733439992498, + "grad_norm": 0.6445318460464478, + "learning_rate": 2.0901995354381398e-05, + "loss": 0.0606, + "step": 33721 + }, + { + "epoch": 0.7904967855740549, + "grad_norm": 0.29856380820274353, + "learning_rate": 2.0897488644641938e-05, + "loss": 0.058, + "step": 33722 + }, + { + "epoch": 0.7905202271488598, + "grad_norm": 0.26105958223342896, + "learning_rate": 2.0892982364114933e-05, + "loss": 0.0626, + "step": 33723 + }, + { + "epoch": 0.7905436687236649, + "grad_norm": 0.3348713517189026, + "learning_rate": 2.088847651282483e-05, + "loss": 0.0593, + "step": 33724 + }, + { + "epoch": 0.7905671102984698, + "grad_norm": 0.575232982635498, + "learning_rate": 2.088397109079604e-05, + "loss": 0.0305, + "step": 33725 + }, + { + "epoch": 0.7905905518732749, + "grad_norm": 0.4035387337207794, + "learning_rate": 2.0879466098053092e-05, + "loss": 0.1637, + "step": 33726 + }, + { + "epoch": 0.7906139934480798, + "grad_norm": 0.5186634659767151, + "learning_rate": 2.0874961534620375e-05, + "loss": 0.0556, + "step": 33727 + }, + { + "epoch": 0.7906374350228849, + "grad_norm": 0.44005662202835083, + "learning_rate": 2.0870457400522336e-05, + "loss": 0.0431, + "step": 33728 + }, + { + "epoch": 0.7906608765976898, + "grad_norm": 0.2303503006696701, + "learning_rate": 2.0865953695783425e-05, + "loss": 0.0516, + "step": 33729 + }, + { + "epoch": 0.7906843181724948, + "grad_norm": 0.5865662097930908, + "learning_rate": 2.0861450420428053e-05, + "loss": 0.0753, + "step": 33730 + }, + { + "epoch": 0.7907077597472998, + "grad_norm": 0.2900909185409546, + "learning_rate": 2.08569475744807e-05, + "loss": 0.0321, + "step": 33731 + }, + { + "epoch": 0.7907312013221048, + "grad_norm": 0.14577174186706543, + "learning_rate": 2.085244515796575e-05, + "loss": 0.0146, + "step": 33732 + }, + { + "epoch": 0.7907546428969098, + "grad_norm": 0.4662059545516968, + "learning_rate": 2.084794317090768e-05, + "loss": 0.0457, + "step": 33733 + }, + { + "epoch": 0.7907780844717148, + "grad_norm": 0.9135267734527588, + "learning_rate": 2.0843441613330895e-05, + "loss": 0.1256, + "step": 33734 + }, + { + "epoch": 0.7908015260465198, + "grad_norm": 0.3517574369907379, + "learning_rate": 2.0838940485259826e-05, + "loss": 0.0485, + "step": 33735 + }, + { + "epoch": 0.7908249676213248, + "grad_norm": 0.29108884930610657, + "learning_rate": 2.0834439786718897e-05, + "loss": 0.0515, + "step": 33736 + }, + { + "epoch": 0.7908484091961298, + "grad_norm": 0.6440610289573669, + "learning_rate": 2.082993951773249e-05, + "loss": 0.0855, + "step": 33737 + }, + { + "epoch": 0.7908718507709348, + "grad_norm": 0.3047025203704834, + "learning_rate": 2.0825439678325078e-05, + "loss": 0.0486, + "step": 33738 + }, + { + "epoch": 0.7908952923457397, + "grad_norm": 0.09861162304878235, + "learning_rate": 2.082094026852106e-05, + "loss": 0.0115, + "step": 33739 + }, + { + "epoch": 0.7909187339205448, + "grad_norm": 0.45681533217430115, + "learning_rate": 2.0816441288344857e-05, + "loss": 0.0751, + "step": 33740 + }, + { + "epoch": 0.7909421754953497, + "grad_norm": 0.569602906703949, + "learning_rate": 2.0811942737820854e-05, + "loss": 0.079, + "step": 33741 + }, + { + "epoch": 0.7909656170701548, + "grad_norm": 0.3507155478000641, + "learning_rate": 2.080744461697346e-05, + "loss": 0.1057, + "step": 33742 + }, + { + "epoch": 0.7909890586449597, + "grad_norm": 0.7205074429512024, + "learning_rate": 2.0802946925827126e-05, + "loss": 0.0703, + "step": 33743 + }, + { + "epoch": 0.7910125002197648, + "grad_norm": 0.13731256127357483, + "learning_rate": 2.0798449664406194e-05, + "loss": 0.0131, + "step": 33744 + }, + { + "epoch": 0.7910359417945697, + "grad_norm": 0.6808082461357117, + "learning_rate": 2.079395283273513e-05, + "loss": 0.5569, + "step": 33745 + }, + { + "epoch": 0.7910593833693748, + "grad_norm": 0.19651107490062714, + "learning_rate": 2.0789456430838305e-05, + "loss": 0.0196, + "step": 33746 + }, + { + "epoch": 0.7910828249441797, + "grad_norm": 0.12432966381311417, + "learning_rate": 2.078496045874012e-05, + "loss": 0.024, + "step": 33747 + }, + { + "epoch": 0.7911062665189847, + "grad_norm": 0.12406724691390991, + "learning_rate": 2.078046491646497e-05, + "loss": 0.0157, + "step": 33748 + }, + { + "epoch": 0.7911297080937897, + "grad_norm": 0.2993672788143158, + "learning_rate": 2.0775969804037244e-05, + "loss": 0.0186, + "step": 33749 + }, + { + "epoch": 0.7911531496685947, + "grad_norm": 0.12335746735334396, + "learning_rate": 2.07714751214813e-05, + "loss": 0.0227, + "step": 33750 + }, + { + "epoch": 0.7911765912433998, + "grad_norm": 0.2711329460144043, + "learning_rate": 2.0766980868821594e-05, + "loss": 0.0483, + "step": 33751 + }, + { + "epoch": 0.7912000328182047, + "grad_norm": 0.41504916548728943, + "learning_rate": 2.076248704608247e-05, + "loss": 0.0807, + "step": 33752 + }, + { + "epoch": 0.7912234743930098, + "grad_norm": 0.32906675338745117, + "learning_rate": 2.075799365328832e-05, + "loss": 0.0477, + "step": 33753 + }, + { + "epoch": 0.7912469159678147, + "grad_norm": 0.270438551902771, + "learning_rate": 2.0753500690463502e-05, + "loss": 0.0507, + "step": 33754 + }, + { + "epoch": 0.7912703575426198, + "grad_norm": 0.504517138004303, + "learning_rate": 2.0749008157632444e-05, + "loss": 0.0891, + "step": 33755 + }, + { + "epoch": 0.7912937991174247, + "grad_norm": 0.1656917929649353, + "learning_rate": 2.0744516054819496e-05, + "loss": 0.0353, + "step": 33756 + }, + { + "epoch": 0.7913172406922298, + "grad_norm": 0.5220736265182495, + "learning_rate": 2.0740024382049007e-05, + "loss": 0.0974, + "step": 33757 + }, + { + "epoch": 0.7913406822670347, + "grad_norm": 0.5483742952346802, + "learning_rate": 2.07355331393454e-05, + "loss": 0.0707, + "step": 33758 + }, + { + "epoch": 0.7913641238418397, + "grad_norm": 0.3329554796218872, + "learning_rate": 2.073104232673301e-05, + "loss": 0.054, + "step": 33759 + }, + { + "epoch": 0.7913875654166447, + "grad_norm": 0.587151825428009, + "learning_rate": 2.0726551944236227e-05, + "loss": 0.0681, + "step": 33760 + }, + { + "epoch": 0.7914110069914497, + "grad_norm": 0.17189735174179077, + "learning_rate": 2.0722061991879392e-05, + "loss": 0.0106, + "step": 33761 + }, + { + "epoch": 0.7914344485662547, + "grad_norm": 0.3554172217845917, + "learning_rate": 2.0717572469686884e-05, + "loss": 0.0245, + "step": 33762 + }, + { + "epoch": 0.7914578901410597, + "grad_norm": 0.5629000067710876, + "learning_rate": 2.071308337768303e-05, + "loss": 0.0934, + "step": 33763 + }, + { + "epoch": 0.7914813317158647, + "grad_norm": 0.4515696167945862, + "learning_rate": 2.0708594715892237e-05, + "loss": 0.0387, + "step": 33764 + }, + { + "epoch": 0.7915047732906697, + "grad_norm": 0.4645843803882599, + "learning_rate": 2.0704106484338837e-05, + "loss": 0.0882, + "step": 33765 + }, + { + "epoch": 0.7915282148654746, + "grad_norm": 0.3438495397567749, + "learning_rate": 2.0699618683047162e-05, + "loss": 0.0633, + "step": 33766 + }, + { + "epoch": 0.7915516564402797, + "grad_norm": 0.41580602526664734, + "learning_rate": 2.0695131312041614e-05, + "loss": 0.0809, + "step": 33767 + }, + { + "epoch": 0.7915750980150846, + "grad_norm": 0.6071038842201233, + "learning_rate": 2.0690644371346513e-05, + "loss": 0.1399, + "step": 33768 + }, + { + "epoch": 0.7915985395898897, + "grad_norm": 0.3163117468357086, + "learning_rate": 2.06861578609862e-05, + "loss": 0.0623, + "step": 33769 + }, + { + "epoch": 0.7916219811646946, + "grad_norm": 0.1464906483888626, + "learning_rate": 2.0681671780985e-05, + "loss": 0.012, + "step": 33770 + }, + { + "epoch": 0.7916454227394997, + "grad_norm": 0.4091370403766632, + "learning_rate": 2.067718613136731e-05, + "loss": 0.4864, + "step": 33771 + }, + { + "epoch": 0.7916688643143046, + "grad_norm": 0.5058548450469971, + "learning_rate": 2.0672700912157437e-05, + "loss": 0.6037, + "step": 33772 + }, + { + "epoch": 0.7916923058891097, + "grad_norm": 0.3678569793701172, + "learning_rate": 2.0668216123379724e-05, + "loss": 0.3792, + "step": 33773 + }, + { + "epoch": 0.7917157474639146, + "grad_norm": 0.14271722733974457, + "learning_rate": 2.0663731765058493e-05, + "loss": 0.0104, + "step": 33774 + }, + { + "epoch": 0.7917391890387196, + "grad_norm": 0.31291836500167847, + "learning_rate": 2.0659247837218054e-05, + "loss": 0.0401, + "step": 33775 + }, + { + "epoch": 0.7917626306135246, + "grad_norm": 0.571869969367981, + "learning_rate": 2.0654764339882804e-05, + "loss": 0.0571, + "step": 33776 + }, + { + "epoch": 0.7917860721883296, + "grad_norm": 0.4236292243003845, + "learning_rate": 2.065028127307703e-05, + "loss": 0.0422, + "step": 33777 + }, + { + "epoch": 0.7918095137631346, + "grad_norm": 0.16041724383831024, + "learning_rate": 2.064579863682503e-05, + "loss": 0.0208, + "step": 33778 + }, + { + "epoch": 0.7918329553379396, + "grad_norm": 0.7850202322006226, + "learning_rate": 2.064131643115119e-05, + "loss": 0.1476, + "step": 33779 + }, + { + "epoch": 0.7918563969127446, + "grad_norm": 0.38198158144950867, + "learning_rate": 2.0636834656079806e-05, + "loss": 0.0419, + "step": 33780 + }, + { + "epoch": 0.7918798384875496, + "grad_norm": 0.6390038728713989, + "learning_rate": 2.063235331163518e-05, + "loss": 0.0882, + "step": 33781 + }, + { + "epoch": 0.7919032800623546, + "grad_norm": 0.13125859200954437, + "learning_rate": 2.0627872397841607e-05, + "loss": 0.0166, + "step": 33782 + }, + { + "epoch": 0.7919267216371596, + "grad_norm": 0.12762431800365448, + "learning_rate": 2.0623391914723455e-05, + "loss": 0.0298, + "step": 33783 + }, + { + "epoch": 0.7919501632119645, + "grad_norm": 0.3178447484970093, + "learning_rate": 2.0618911862305012e-05, + "loss": 0.0623, + "step": 33784 + }, + { + "epoch": 0.7919736047867696, + "grad_norm": 0.11421056091785431, + "learning_rate": 2.0614432240610593e-05, + "loss": 0.0175, + "step": 33785 + }, + { + "epoch": 0.7919970463615745, + "grad_norm": 0.42077311873435974, + "learning_rate": 2.060995304966449e-05, + "loss": 0.0435, + "step": 33786 + }, + { + "epoch": 0.7920204879363796, + "grad_norm": 0.6461148262023926, + "learning_rate": 2.0605474289491013e-05, + "loss": 0.1624, + "step": 33787 + }, + { + "epoch": 0.7920439295111845, + "grad_norm": 0.09448303282260895, + "learning_rate": 2.0600995960114446e-05, + "loss": 0.0091, + "step": 33788 + }, + { + "epoch": 0.7920673710859896, + "grad_norm": 0.3113292157649994, + "learning_rate": 2.0596518061559123e-05, + "loss": 0.047, + "step": 33789 + }, + { + "epoch": 0.7920908126607945, + "grad_norm": 0.13189521431922913, + "learning_rate": 2.059204059384934e-05, + "loss": 0.0261, + "step": 33790 + }, + { + "epoch": 0.7921142542355996, + "grad_norm": 0.19603200256824493, + "learning_rate": 2.058756355700934e-05, + "loss": 0.0332, + "step": 33791 + }, + { + "epoch": 0.7921376958104045, + "grad_norm": 0.4512452781200409, + "learning_rate": 2.058308695106348e-05, + "loss": 0.0693, + "step": 33792 + }, + { + "epoch": 0.7921611373852095, + "grad_norm": 0.5296308398246765, + "learning_rate": 2.057861077603602e-05, + "loss": 0.1021, + "step": 33793 + }, + { + "epoch": 0.7921845789600145, + "grad_norm": 0.5567145943641663, + "learning_rate": 2.0574135031951258e-05, + "loss": 0.061, + "step": 33794 + }, + { + "epoch": 0.7922080205348195, + "grad_norm": 0.3906368613243103, + "learning_rate": 2.0569659718833445e-05, + "loss": 0.076, + "step": 33795 + }, + { + "epoch": 0.7922314621096245, + "grad_norm": 0.3435029983520508, + "learning_rate": 2.0565184836706918e-05, + "loss": 0.0443, + "step": 33796 + }, + { + "epoch": 0.7922549036844295, + "grad_norm": 0.31284889578819275, + "learning_rate": 2.056071038559594e-05, + "loss": 0.0435, + "step": 33797 + }, + { + "epoch": 0.7922783452592345, + "grad_norm": 0.3949955105781555, + "learning_rate": 2.0556236365524763e-05, + "loss": 0.0395, + "step": 33798 + }, + { + "epoch": 0.7923017868340395, + "grad_norm": 0.40192317962646484, + "learning_rate": 2.055176277651769e-05, + "loss": 0.0594, + "step": 33799 + }, + { + "epoch": 0.7923252284088445, + "grad_norm": 0.5405805706977844, + "learning_rate": 2.0547289618598985e-05, + "loss": 0.0375, + "step": 33800 + }, + { + "epoch": 0.7923486699836495, + "grad_norm": 0.383279412984848, + "learning_rate": 2.0542816891792905e-05, + "loss": 0.3257, + "step": 33801 + }, + { + "epoch": 0.7923721115584546, + "grad_norm": 0.2712993025779724, + "learning_rate": 2.053834459612375e-05, + "loss": 0.0496, + "step": 33802 + }, + { + "epoch": 0.7923955531332595, + "grad_norm": 0.4481247365474701, + "learning_rate": 2.0533872731615743e-05, + "loss": 0.0676, + "step": 33803 + }, + { + "epoch": 0.7924189947080645, + "grad_norm": 0.40678080916404724, + "learning_rate": 2.0529401298293206e-05, + "loss": 0.0616, + "step": 33804 + }, + { + "epoch": 0.7924424362828695, + "grad_norm": 0.3004608452320099, + "learning_rate": 2.0524930296180377e-05, + "loss": 0.0314, + "step": 33805 + }, + { + "epoch": 0.7924658778576745, + "grad_norm": 0.22094403207302094, + "learning_rate": 2.052045972530151e-05, + "loss": 0.0356, + "step": 33806 + }, + { + "epoch": 0.7924893194324795, + "grad_norm": 0.4823892414569855, + "learning_rate": 2.051598958568086e-05, + "loss": 0.0705, + "step": 33807 + }, + { + "epoch": 0.7925127610072845, + "grad_norm": 0.06237201392650604, + "learning_rate": 2.0511519877342656e-05, + "loss": 0.0102, + "step": 33808 + }, + { + "epoch": 0.7925362025820895, + "grad_norm": 0.2652184069156647, + "learning_rate": 2.0507050600311205e-05, + "loss": 0.0396, + "step": 33809 + }, + { + "epoch": 0.7925596441568945, + "grad_norm": 0.6150851845741272, + "learning_rate": 2.050258175461074e-05, + "loss": 0.615, + "step": 33810 + }, + { + "epoch": 0.7925830857316994, + "grad_norm": 0.494537889957428, + "learning_rate": 2.04981133402655e-05, + "loss": 0.1243, + "step": 33811 + }, + { + "epoch": 0.7926065273065045, + "grad_norm": 0.09251010417938232, + "learning_rate": 2.049364535729973e-05, + "loss": 0.0109, + "step": 33812 + }, + { + "epoch": 0.7926299688813094, + "grad_norm": 0.12479745596647263, + "learning_rate": 2.048917780573766e-05, + "loss": 0.0179, + "step": 33813 + }, + { + "epoch": 0.7926534104561145, + "grad_norm": 0.4620220363140106, + "learning_rate": 2.048471068560356e-05, + "loss": 0.063, + "step": 33814 + }, + { + "epoch": 0.7926768520309194, + "grad_norm": 0.4491897523403168, + "learning_rate": 2.0480243996921633e-05, + "loss": 0.0618, + "step": 33815 + }, + { + "epoch": 0.7927002936057245, + "grad_norm": 0.5477237105369568, + "learning_rate": 2.047577773971616e-05, + "loss": 0.0663, + "step": 33816 + }, + { + "epoch": 0.7927237351805294, + "grad_norm": 0.6436837911605835, + "learning_rate": 2.047131191401136e-05, + "loss": 0.1011, + "step": 33817 + }, + { + "epoch": 0.7927471767553345, + "grad_norm": 0.3302847146987915, + "learning_rate": 2.0466846519831463e-05, + "loss": 0.0316, + "step": 33818 + }, + { + "epoch": 0.7927706183301394, + "grad_norm": 0.1022721529006958, + "learning_rate": 2.0462381557200682e-05, + "loss": 0.0135, + "step": 33819 + }, + { + "epoch": 0.7927940599049444, + "grad_norm": 0.3359731435775757, + "learning_rate": 2.0457917026143237e-05, + "loss": 0.3668, + "step": 33820 + }, + { + "epoch": 0.7928175014797494, + "grad_norm": 0.6458839774131775, + "learning_rate": 2.045345292668339e-05, + "loss": 0.0527, + "step": 33821 + }, + { + "epoch": 0.7928409430545544, + "grad_norm": 0.539136528968811, + "learning_rate": 2.044898925884534e-05, + "loss": 0.1341, + "step": 33822 + }, + { + "epoch": 0.7928643846293594, + "grad_norm": 0.83982914686203, + "learning_rate": 2.0444526022653322e-05, + "loss": 0.0945, + "step": 33823 + }, + { + "epoch": 0.7928878262041644, + "grad_norm": 0.10221879929304123, + "learning_rate": 2.0440063218131533e-05, + "loss": 0.0139, + "step": 33824 + }, + { + "epoch": 0.7929112677789694, + "grad_norm": 0.34832772612571716, + "learning_rate": 2.0435600845304205e-05, + "loss": 0.0242, + "step": 33825 + }, + { + "epoch": 0.7929347093537744, + "grad_norm": 0.35929450392723083, + "learning_rate": 2.0431138904195524e-05, + "loss": 0.0518, + "step": 33826 + }, + { + "epoch": 0.7929581509285794, + "grad_norm": 0.4014359712600708, + "learning_rate": 2.0426677394829706e-05, + "loss": 0.0583, + "step": 33827 + }, + { + "epoch": 0.7929815925033844, + "grad_norm": 0.3991183638572693, + "learning_rate": 2.0422216317231013e-05, + "loss": 0.0692, + "step": 33828 + }, + { + "epoch": 0.7930050340781893, + "grad_norm": 0.4910520911216736, + "learning_rate": 2.04177556714236e-05, + "loss": 0.0433, + "step": 33829 + }, + { + "epoch": 0.7930284756529944, + "grad_norm": 0.27079564332962036, + "learning_rate": 2.0413295457431693e-05, + "loss": 0.2551, + "step": 33830 + }, + { + "epoch": 0.7930519172277993, + "grad_norm": 0.3484254479408264, + "learning_rate": 2.040883567527947e-05, + "loss": 0.0986, + "step": 33831 + }, + { + "epoch": 0.7930753588026044, + "grad_norm": 0.21968974173069, + "learning_rate": 2.0404376324991148e-05, + "loss": 0.0308, + "step": 33832 + }, + { + "epoch": 0.7930988003774093, + "grad_norm": 0.317947655916214, + "learning_rate": 2.0399917406590896e-05, + "loss": 0.0387, + "step": 33833 + }, + { + "epoch": 0.7931222419522144, + "grad_norm": 0.5662028789520264, + "learning_rate": 2.0395458920102963e-05, + "loss": 0.0786, + "step": 33834 + }, + { + "epoch": 0.7931456835270193, + "grad_norm": 0.41698306798934937, + "learning_rate": 2.039100086555149e-05, + "loss": 0.0814, + "step": 33835 + }, + { + "epoch": 0.7931691251018244, + "grad_norm": 0.3074251711368561, + "learning_rate": 2.0386543242960697e-05, + "loss": 0.0625, + "step": 33836 + }, + { + "epoch": 0.7931925666766293, + "grad_norm": 0.793358564376831, + "learning_rate": 2.0382086052354755e-05, + "loss": 0.1605, + "step": 33837 + }, + { + "epoch": 0.7932160082514343, + "grad_norm": 0.41386017203330994, + "learning_rate": 2.0377629293757862e-05, + "loss": 0.0376, + "step": 33838 + }, + { + "epoch": 0.7932394498262393, + "grad_norm": 0.4792758822441101, + "learning_rate": 2.0373172967194153e-05, + "loss": 0.0505, + "step": 33839 + }, + { + "epoch": 0.7932628914010443, + "grad_norm": 0.20403259992599487, + "learning_rate": 2.0368717072687848e-05, + "loss": 0.0333, + "step": 33840 + }, + { + "epoch": 0.7932863329758493, + "grad_norm": 0.41707518696784973, + "learning_rate": 2.0364261610263156e-05, + "loss": 0.5084, + "step": 33841 + }, + { + "epoch": 0.7933097745506543, + "grad_norm": 0.12920601665973663, + "learning_rate": 2.035980657994421e-05, + "loss": 0.018, + "step": 33842 + }, + { + "epoch": 0.7933332161254593, + "grad_norm": 0.49138906598091125, + "learning_rate": 2.03553519817552e-05, + "loss": 0.0571, + "step": 33843 + }, + { + "epoch": 0.7933566577002643, + "grad_norm": 0.475478857755661, + "learning_rate": 2.0350897815720294e-05, + "loss": 0.1223, + "step": 33844 + }, + { + "epoch": 0.7933800992750693, + "grad_norm": 0.09989234805107117, + "learning_rate": 2.0346444081863646e-05, + "loss": 0.0179, + "step": 33845 + }, + { + "epoch": 0.7934035408498743, + "grad_norm": 0.39199113845825195, + "learning_rate": 2.034199078020941e-05, + "loss": 0.0546, + "step": 33846 + }, + { + "epoch": 0.7934269824246792, + "grad_norm": 0.33149072527885437, + "learning_rate": 2.03375379107818e-05, + "loss": 0.0694, + "step": 33847 + }, + { + "epoch": 0.7934504239994843, + "grad_norm": 0.2039509266614914, + "learning_rate": 2.033308547360494e-05, + "loss": 0.034, + "step": 33848 + }, + { + "epoch": 0.7934738655742892, + "grad_norm": 0.48776480555534363, + "learning_rate": 2.0328633468703006e-05, + "loss": 0.1059, + "step": 33849 + }, + { + "epoch": 0.7934973071490943, + "grad_norm": 0.6092759370803833, + "learning_rate": 2.032418189610015e-05, + "loss": 0.1171, + "step": 33850 + }, + { + "epoch": 0.7935207487238992, + "grad_norm": 0.14939327538013458, + "learning_rate": 2.031973075582049e-05, + "loss": 0.0095, + "step": 33851 + }, + { + "epoch": 0.7935441902987043, + "grad_norm": 0.5295272469520569, + "learning_rate": 2.0315280047888218e-05, + "loss": 0.0602, + "step": 33852 + }, + { + "epoch": 0.7935676318735093, + "grad_norm": 0.5219801664352417, + "learning_rate": 2.03108297723275e-05, + "loss": 0.0835, + "step": 33853 + }, + { + "epoch": 0.7935910734483143, + "grad_norm": 0.23878011107444763, + "learning_rate": 2.0306379929162456e-05, + "loss": 0.0435, + "step": 33854 + }, + { + "epoch": 0.7936145150231193, + "grad_norm": 0.31187739968299866, + "learning_rate": 2.0301930518417244e-05, + "loss": 0.038, + "step": 33855 + }, + { + "epoch": 0.7936379565979242, + "grad_norm": 0.5686957240104675, + "learning_rate": 2.029748154011599e-05, + "loss": 0.1264, + "step": 33856 + }, + { + "epoch": 0.7936613981727293, + "grad_norm": 0.2864932715892792, + "learning_rate": 2.0293032994282847e-05, + "loss": 0.046, + "step": 33857 + }, + { + "epoch": 0.7936848397475342, + "grad_norm": 0.10421642661094666, + "learning_rate": 2.028858488094192e-05, + "loss": 0.0118, + "step": 33858 + }, + { + "epoch": 0.7937082813223393, + "grad_norm": 0.44046953320503235, + "learning_rate": 2.0284137200117404e-05, + "loss": 0.0767, + "step": 33859 + }, + { + "epoch": 0.7937317228971442, + "grad_norm": 0.4950665831565857, + "learning_rate": 2.027968995183339e-05, + "loss": 0.0474, + "step": 33860 + }, + { + "epoch": 0.7937551644719493, + "grad_norm": 0.48457834124565125, + "learning_rate": 2.027524313611403e-05, + "loss": 0.0461, + "step": 33861 + }, + { + "epoch": 0.7937786060467542, + "grad_norm": 0.2820471227169037, + "learning_rate": 2.0270796752983435e-05, + "loss": 0.025, + "step": 33862 + }, + { + "epoch": 0.7938020476215593, + "grad_norm": 0.31444069743156433, + "learning_rate": 2.0266350802465715e-05, + "loss": 0.032, + "step": 33863 + }, + { + "epoch": 0.7938254891963642, + "grad_norm": 0.15506869554519653, + "learning_rate": 2.026190528458505e-05, + "loss": 0.0226, + "step": 33864 + }, + { + "epoch": 0.7938489307711692, + "grad_norm": 0.8507525324821472, + "learning_rate": 2.0257460199365495e-05, + "loss": 0.1389, + "step": 33865 + }, + { + "epoch": 0.7938723723459742, + "grad_norm": 0.3853793740272522, + "learning_rate": 2.0253015546831232e-05, + "loss": 0.0674, + "step": 33866 + }, + { + "epoch": 0.7938958139207792, + "grad_norm": 0.4964674413204193, + "learning_rate": 2.0248571327006348e-05, + "loss": 0.0935, + "step": 33867 + }, + { + "epoch": 0.7939192554955842, + "grad_norm": 0.2103308141231537, + "learning_rate": 2.0244127539914957e-05, + "loss": 0.0455, + "step": 33868 + }, + { + "epoch": 0.7939426970703892, + "grad_norm": 0.1273474097251892, + "learning_rate": 2.0239684185581176e-05, + "loss": 0.0157, + "step": 33869 + }, + { + "epoch": 0.7939661386451942, + "grad_norm": 0.7456743717193604, + "learning_rate": 2.0235241264029115e-05, + "loss": 0.1541, + "step": 33870 + }, + { + "epoch": 0.7939895802199992, + "grad_norm": 0.11877831816673279, + "learning_rate": 2.0230798775282844e-05, + "loss": 0.0147, + "step": 33871 + }, + { + "epoch": 0.7940130217948042, + "grad_norm": 0.5082494020462036, + "learning_rate": 2.0226356719366525e-05, + "loss": 0.061, + "step": 33872 + }, + { + "epoch": 0.7940364633696092, + "grad_norm": 0.1403103470802307, + "learning_rate": 2.0221915096304246e-05, + "loss": 0.0133, + "step": 33873 + }, + { + "epoch": 0.7940599049444141, + "grad_norm": 0.09457015246152878, + "learning_rate": 2.02174739061201e-05, + "loss": 0.0168, + "step": 33874 + }, + { + "epoch": 0.7940833465192192, + "grad_norm": 0.09147144109010696, + "learning_rate": 2.0213033148838178e-05, + "loss": 0.0073, + "step": 33875 + }, + { + "epoch": 0.7941067880940241, + "grad_norm": 0.10343801975250244, + "learning_rate": 2.020859282448255e-05, + "loss": 0.0107, + "step": 33876 + }, + { + "epoch": 0.7941302296688292, + "grad_norm": 0.11997964233160019, + "learning_rate": 2.020415293307738e-05, + "loss": 0.0225, + "step": 33877 + }, + { + "epoch": 0.7941536712436341, + "grad_norm": 0.3442535102367401, + "learning_rate": 2.0199713474646687e-05, + "loss": 0.0655, + "step": 33878 + }, + { + "epoch": 0.7941771128184392, + "grad_norm": 0.13314519822597504, + "learning_rate": 2.0195274449214618e-05, + "loss": 0.0209, + "step": 33879 + }, + { + "epoch": 0.7942005543932441, + "grad_norm": 0.09078554064035416, + "learning_rate": 2.0190835856805233e-05, + "loss": 0.0083, + "step": 33880 + }, + { + "epoch": 0.7942239959680492, + "grad_norm": 0.08513526618480682, + "learning_rate": 2.0186397697442615e-05, + "loss": 0.0147, + "step": 33881 + }, + { + "epoch": 0.7942474375428541, + "grad_norm": 0.48020461201667786, + "learning_rate": 2.0181959971150856e-05, + "loss": 0.6395, + "step": 33882 + }, + { + "epoch": 0.7942708791176591, + "grad_norm": 0.3994525671005249, + "learning_rate": 2.0177522677954e-05, + "loss": 0.0461, + "step": 33883 + }, + { + "epoch": 0.7942943206924641, + "grad_norm": 0.3925583064556122, + "learning_rate": 2.017308581787617e-05, + "loss": 0.035, + "step": 33884 + }, + { + "epoch": 0.7943177622672691, + "grad_norm": 0.33699050545692444, + "learning_rate": 2.016864939094142e-05, + "loss": 0.0357, + "step": 33885 + }, + { + "epoch": 0.7943412038420741, + "grad_norm": 0.48979222774505615, + "learning_rate": 2.0164213397173825e-05, + "loss": 0.1152, + "step": 33886 + }, + { + "epoch": 0.7943646454168791, + "grad_norm": 0.24933260679244995, + "learning_rate": 2.0159777836597447e-05, + "loss": 0.0485, + "step": 33887 + }, + { + "epoch": 0.7943880869916841, + "grad_norm": 0.21749849617481232, + "learning_rate": 2.0155342709236348e-05, + "loss": 0.0268, + "step": 33888 + }, + { + "epoch": 0.7944115285664891, + "grad_norm": 0.26542946696281433, + "learning_rate": 2.015090801511462e-05, + "loss": 0.0306, + "step": 33889 + }, + { + "epoch": 0.794434970141294, + "grad_norm": 0.9733597040176392, + "learning_rate": 2.014647375425629e-05, + "loss": 0.168, + "step": 33890 + }, + { + "epoch": 0.7944584117160991, + "grad_norm": 0.14866454899311066, + "learning_rate": 2.0142039926685454e-05, + "loss": 0.0181, + "step": 33891 + }, + { + "epoch": 0.794481853290904, + "grad_norm": 0.6728268265724182, + "learning_rate": 2.013760653242617e-05, + "loss": 0.1183, + "step": 33892 + }, + { + "epoch": 0.7945052948657091, + "grad_norm": 0.6342041492462158, + "learning_rate": 2.0133173571502462e-05, + "loss": 0.0478, + "step": 33893 + }, + { + "epoch": 0.794528736440514, + "grad_norm": 0.2954198718070984, + "learning_rate": 2.0128741043938416e-05, + "loss": 0.0338, + "step": 33894 + }, + { + "epoch": 0.7945521780153191, + "grad_norm": 0.13356952369213104, + "learning_rate": 2.0124308949758052e-05, + "loss": 0.0307, + "step": 33895 + }, + { + "epoch": 0.794575619590124, + "grad_norm": 0.4214622378349304, + "learning_rate": 2.0119877288985423e-05, + "loss": 0.088, + "step": 33896 + }, + { + "epoch": 0.7945990611649291, + "grad_norm": 0.503334105014801, + "learning_rate": 2.0115446061644594e-05, + "loss": 0.0426, + "step": 33897 + }, + { + "epoch": 0.794622502739734, + "grad_norm": 0.6583929061889648, + "learning_rate": 2.0111015267759616e-05, + "loss": 0.1178, + "step": 33898 + }, + { + "epoch": 0.794645944314539, + "grad_norm": 0.10974713414907455, + "learning_rate": 2.0106584907354508e-05, + "loss": 0.0063, + "step": 33899 + }, + { + "epoch": 0.794669385889344, + "grad_norm": 0.5850854516029358, + "learning_rate": 2.010215498045329e-05, + "loss": 0.143, + "step": 33900 + }, + { + "epoch": 0.794692827464149, + "grad_norm": 0.4674558639526367, + "learning_rate": 2.0097725487080056e-05, + "loss": 0.0667, + "step": 33901 + }, + { + "epoch": 0.794716269038954, + "grad_norm": 0.2971431314945221, + "learning_rate": 2.0093296427258813e-05, + "loss": 0.0615, + "step": 33902 + }, + { + "epoch": 0.794739710613759, + "grad_norm": 0.15278558433055878, + "learning_rate": 2.0088867801013556e-05, + "loss": 0.02, + "step": 33903 + }, + { + "epoch": 0.7947631521885641, + "grad_norm": 0.17193418741226196, + "learning_rate": 2.0084439608368377e-05, + "loss": 0.0362, + "step": 33904 + }, + { + "epoch": 0.794786593763369, + "grad_norm": 0.44207292795181274, + "learning_rate": 2.008001184934728e-05, + "loss": 0.4731, + "step": 33905 + }, + { + "epoch": 0.7948100353381741, + "grad_norm": 0.5439582467079163, + "learning_rate": 2.0075584523974287e-05, + "loss": 0.1098, + "step": 33906 + }, + { + "epoch": 0.794833476912979, + "grad_norm": 0.4340592920780182, + "learning_rate": 2.007115763227342e-05, + "loss": 0.0965, + "step": 33907 + }, + { + "epoch": 0.7948569184877841, + "grad_norm": 0.44986531138420105, + "learning_rate": 2.0066731174268694e-05, + "loss": 0.0687, + "step": 33908 + }, + { + "epoch": 0.794880360062589, + "grad_norm": 0.39971232414245605, + "learning_rate": 2.0062305149984107e-05, + "loss": 0.1, + "step": 33909 + }, + { + "epoch": 0.794903801637394, + "grad_norm": 0.308437317609787, + "learning_rate": 2.0057879559443726e-05, + "loss": 0.0478, + "step": 33910 + }, + { + "epoch": 0.794927243212199, + "grad_norm": 0.14646443724632263, + "learning_rate": 2.0053454402671533e-05, + "loss": 0.0191, + "step": 33911 + }, + { + "epoch": 0.794950684787004, + "grad_norm": 0.34157824516296387, + "learning_rate": 2.0049029679691522e-05, + "loss": 0.0736, + "step": 33912 + }, + { + "epoch": 0.794974126361809, + "grad_norm": 0.5415790677070618, + "learning_rate": 2.004460539052776e-05, + "loss": 0.0436, + "step": 33913 + }, + { + "epoch": 0.794997567936614, + "grad_norm": 0.46159934997558594, + "learning_rate": 2.0040181535204205e-05, + "loss": 0.0709, + "step": 33914 + }, + { + "epoch": 0.795021009511419, + "grad_norm": 0.3659953773021698, + "learning_rate": 2.0035758113744874e-05, + "loss": 0.0481, + "step": 33915 + }, + { + "epoch": 0.795044451086224, + "grad_norm": 0.3697003126144409, + "learning_rate": 2.0031335126173745e-05, + "loss": 0.1053, + "step": 33916 + }, + { + "epoch": 0.795067892661029, + "grad_norm": 0.2325124740600586, + "learning_rate": 2.0026912572514867e-05, + "loss": 0.0312, + "step": 33917 + }, + { + "epoch": 0.795091334235834, + "grad_norm": 0.5421911478042603, + "learning_rate": 2.002249045279221e-05, + "loss": 0.1252, + "step": 33918 + }, + { + "epoch": 0.7951147758106389, + "grad_norm": 0.25113269686698914, + "learning_rate": 2.0018068767029774e-05, + "loss": 0.0548, + "step": 33919 + }, + { + "epoch": 0.795138217385444, + "grad_norm": 0.3848015367984772, + "learning_rate": 2.0013647515251544e-05, + "loss": 0.0727, + "step": 33920 + }, + { + "epoch": 0.7951616589602489, + "grad_norm": 0.2476864606142044, + "learning_rate": 2.0009226697481477e-05, + "loss": 0.0426, + "step": 33921 + }, + { + "epoch": 0.795185100535054, + "grad_norm": 0.39637094736099243, + "learning_rate": 2.0004806313743628e-05, + "loss": 0.0703, + "step": 33922 + }, + { + "epoch": 0.7952085421098589, + "grad_norm": 0.1891053020954132, + "learning_rate": 2.000038636406194e-05, + "loss": 0.0235, + "step": 33923 + }, + { + "epoch": 0.795231983684664, + "grad_norm": 0.49017396569252014, + "learning_rate": 1.999596684846039e-05, + "loss": 0.0707, + "step": 33924 + }, + { + "epoch": 0.7952554252594689, + "grad_norm": 0.0819351077079773, + "learning_rate": 1.9991547766962992e-05, + "loss": 0.0188, + "step": 33925 + }, + { + "epoch": 0.795278866834274, + "grad_norm": 0.3704732060432434, + "learning_rate": 1.9987129119593706e-05, + "loss": 0.0658, + "step": 33926 + }, + { + "epoch": 0.7953023084090789, + "grad_norm": 0.30105307698249817, + "learning_rate": 1.998271090637651e-05, + "loss": 0.0443, + "step": 33927 + }, + { + "epoch": 0.795325749983884, + "grad_norm": 0.5396103262901306, + "learning_rate": 1.9978293127335356e-05, + "loss": 0.5577, + "step": 33928 + }, + { + "epoch": 0.7953491915586889, + "grad_norm": 0.10293859988451004, + "learning_rate": 1.9973875782494244e-05, + "loss": 0.0146, + "step": 33929 + }, + { + "epoch": 0.7953726331334939, + "grad_norm": 0.3340744078159332, + "learning_rate": 1.9969458871877146e-05, + "loss": 0.0528, + "step": 33930 + }, + { + "epoch": 0.7953960747082989, + "grad_norm": 0.28828299045562744, + "learning_rate": 1.9965042395508005e-05, + "loss": 0.0609, + "step": 33931 + }, + { + "epoch": 0.7954195162831039, + "grad_norm": 0.41668352484703064, + "learning_rate": 1.9960626353410804e-05, + "loss": 0.0758, + "step": 33932 + }, + { + "epoch": 0.7954429578579089, + "grad_norm": 0.4327545166015625, + "learning_rate": 1.9956210745609484e-05, + "loss": 0.0662, + "step": 33933 + }, + { + "epoch": 0.7954663994327139, + "grad_norm": 0.16846737265586853, + "learning_rate": 1.9951795572127995e-05, + "loss": 0.0474, + "step": 33934 + }, + { + "epoch": 0.7954898410075188, + "grad_norm": 0.2975846529006958, + "learning_rate": 1.9947380832990337e-05, + "loss": 0.0434, + "step": 33935 + }, + { + "epoch": 0.7955132825823239, + "grad_norm": 0.29166093468666077, + "learning_rate": 1.994296652822042e-05, + "loss": 0.0461, + "step": 33936 + }, + { + "epoch": 0.7955367241571288, + "grad_norm": 0.37028518319129944, + "learning_rate": 1.9938552657842234e-05, + "loss": 0.0714, + "step": 33937 + }, + { + "epoch": 0.7955601657319339, + "grad_norm": 0.9593228101730347, + "learning_rate": 1.993413922187972e-05, + "loss": 0.118, + "step": 33938 + }, + { + "epoch": 0.7955836073067388, + "grad_norm": 0.40661680698394775, + "learning_rate": 1.9929726220356825e-05, + "loss": 0.0892, + "step": 33939 + }, + { + "epoch": 0.7956070488815439, + "grad_norm": 0.5129885673522949, + "learning_rate": 1.9925313653297474e-05, + "loss": 0.0751, + "step": 33940 + }, + { + "epoch": 0.7956304904563488, + "grad_norm": 0.6323990225791931, + "learning_rate": 1.99209015207256e-05, + "loss": 0.1225, + "step": 33941 + }, + { + "epoch": 0.7956539320311539, + "grad_norm": 0.1411912590265274, + "learning_rate": 1.9916489822665197e-05, + "loss": 0.0238, + "step": 33942 + }, + { + "epoch": 0.7956773736059588, + "grad_norm": 0.10020402818918228, + "learning_rate": 1.991207855914017e-05, + "loss": 0.0193, + "step": 33943 + }, + { + "epoch": 0.7957008151807639, + "grad_norm": 0.3668126165866852, + "learning_rate": 1.990766773017445e-05, + "loss": 0.0423, + "step": 33944 + }, + { + "epoch": 0.7957242567555688, + "grad_norm": 0.3553812503814697, + "learning_rate": 1.990325733579198e-05, + "loss": 0.5191, + "step": 33945 + }, + { + "epoch": 0.7957476983303738, + "grad_norm": 0.48381271958351135, + "learning_rate": 1.9898847376016694e-05, + "loss": 0.6184, + "step": 33946 + }, + { + "epoch": 0.7957711399051788, + "grad_norm": 0.7764918804168701, + "learning_rate": 1.989443785087248e-05, + "loss": 0.0939, + "step": 33947 + }, + { + "epoch": 0.7957945814799838, + "grad_norm": 0.327950656414032, + "learning_rate": 1.9890028760383338e-05, + "loss": 0.0557, + "step": 33948 + }, + { + "epoch": 0.7958180230547888, + "grad_norm": 0.2364341765642166, + "learning_rate": 1.9885620104573112e-05, + "loss": 0.039, + "step": 33949 + }, + { + "epoch": 0.7958414646295938, + "grad_norm": 0.5352154970169067, + "learning_rate": 1.98812118834658e-05, + "loss": 0.0283, + "step": 33950 + }, + { + "epoch": 0.7958649062043988, + "grad_norm": 0.16093161702156067, + "learning_rate": 1.9876804097085277e-05, + "loss": 0.0146, + "step": 33951 + }, + { + "epoch": 0.7958883477792038, + "grad_norm": 0.382239431142807, + "learning_rate": 1.9872396745455467e-05, + "loss": 0.0527, + "step": 33952 + }, + { + "epoch": 0.7959117893540087, + "grad_norm": 0.6261469721794128, + "learning_rate": 1.9867989828600297e-05, + "loss": 0.5494, + "step": 33953 + }, + { + "epoch": 0.7959352309288138, + "grad_norm": 0.3127865493297577, + "learning_rate": 1.9863583346543636e-05, + "loss": 0.0643, + "step": 33954 + }, + { + "epoch": 0.7959586725036188, + "grad_norm": 0.19777564704418182, + "learning_rate": 1.985917729930944e-05, + "loss": 0.0192, + "step": 33955 + }, + { + "epoch": 0.7959821140784238, + "grad_norm": 0.402495414018631, + "learning_rate": 1.9854771686921615e-05, + "loss": 0.0568, + "step": 33956 + }, + { + "epoch": 0.7960055556532288, + "grad_norm": 0.5119594931602478, + "learning_rate": 1.9850366509404052e-05, + "loss": 0.0875, + "step": 33957 + }, + { + "epoch": 0.7960289972280338, + "grad_norm": 0.4258561432361603, + "learning_rate": 1.984596176678064e-05, + "loss": 0.0395, + "step": 33958 + }, + { + "epoch": 0.7960524388028388, + "grad_norm": 0.1528392732143402, + "learning_rate": 1.9841557459075278e-05, + "loss": 0.0223, + "step": 33959 + }, + { + "epoch": 0.7960758803776438, + "grad_norm": 0.1749771386384964, + "learning_rate": 1.983715358631191e-05, + "loss": 0.0309, + "step": 33960 + }, + { + "epoch": 0.7960993219524488, + "grad_norm": 0.405112087726593, + "learning_rate": 1.9832750148514367e-05, + "loss": 0.0818, + "step": 33961 + }, + { + "epoch": 0.7961227635272538, + "grad_norm": 0.2591998279094696, + "learning_rate": 1.9828347145706594e-05, + "loss": 0.0264, + "step": 33962 + }, + { + "epoch": 0.7961462051020588, + "grad_norm": 0.5403984785079956, + "learning_rate": 1.982394457791248e-05, + "loss": 0.0912, + "step": 33963 + }, + { + "epoch": 0.7961696466768637, + "grad_norm": 0.34846413135528564, + "learning_rate": 1.9819542445155882e-05, + "loss": 0.0858, + "step": 33964 + }, + { + "epoch": 0.7961930882516688, + "grad_norm": 0.4565144181251526, + "learning_rate": 1.9815140747460713e-05, + "loss": 0.0596, + "step": 33965 + }, + { + "epoch": 0.7962165298264737, + "grad_norm": 0.13894663751125336, + "learning_rate": 1.9810739484850817e-05, + "loss": 0.0242, + "step": 33966 + }, + { + "epoch": 0.7962399714012788, + "grad_norm": 0.5551672577857971, + "learning_rate": 1.980633865735012e-05, + "loss": 0.0513, + "step": 33967 + }, + { + "epoch": 0.7962634129760837, + "grad_norm": 0.22833049297332764, + "learning_rate": 1.98019382649825e-05, + "loss": 0.0355, + "step": 33968 + }, + { + "epoch": 0.7962868545508888, + "grad_norm": 0.703934371471405, + "learning_rate": 1.9797538307771802e-05, + "loss": 0.0796, + "step": 33969 + }, + { + "epoch": 0.7963102961256937, + "grad_norm": 0.1470489203929901, + "learning_rate": 1.9793138785741927e-05, + "loss": 0.0137, + "step": 33970 + }, + { + "epoch": 0.7963337377004988, + "grad_norm": 0.28020599484443665, + "learning_rate": 1.9788739698916737e-05, + "loss": 0.0694, + "step": 33971 + }, + { + "epoch": 0.7963571792753037, + "grad_norm": 0.5224365592002869, + "learning_rate": 1.9784341047320067e-05, + "loss": 0.1181, + "step": 33972 + }, + { + "epoch": 0.7963806208501087, + "grad_norm": 0.29856395721435547, + "learning_rate": 1.9779942830975817e-05, + "loss": 0.0457, + "step": 33973 + }, + { + "epoch": 0.7964040624249137, + "grad_norm": 0.4109741747379303, + "learning_rate": 1.9775545049907883e-05, + "loss": 0.0369, + "step": 33974 + }, + { + "epoch": 0.7964275039997187, + "grad_norm": 0.5362470746040344, + "learning_rate": 1.97711477041401e-05, + "loss": 0.5338, + "step": 33975 + }, + { + "epoch": 0.7964509455745237, + "grad_norm": 0.42150720953941345, + "learning_rate": 1.976675079369632e-05, + "loss": 0.0524, + "step": 33976 + }, + { + "epoch": 0.7964743871493287, + "grad_norm": 0.3532564640045166, + "learning_rate": 1.9762354318600408e-05, + "loss": 0.062, + "step": 33977 + }, + { + "epoch": 0.7964978287241337, + "grad_norm": 0.434015691280365, + "learning_rate": 1.975795827887622e-05, + "loss": 0.0546, + "step": 33978 + }, + { + "epoch": 0.7965212702989387, + "grad_norm": 0.509768009185791, + "learning_rate": 1.9753562674547575e-05, + "loss": 0.4874, + "step": 33979 + }, + { + "epoch": 0.7965447118737436, + "grad_norm": 0.16954462230205536, + "learning_rate": 1.974916750563839e-05, + "loss": 0.0235, + "step": 33980 + }, + { + "epoch": 0.7965681534485487, + "grad_norm": 0.544037401676178, + "learning_rate": 1.974477277217247e-05, + "loss": 0.0651, + "step": 33981 + }, + { + "epoch": 0.7965915950233536, + "grad_norm": 0.15036599338054657, + "learning_rate": 1.974037847417368e-05, + "loss": 0.0099, + "step": 33982 + }, + { + "epoch": 0.7966150365981587, + "grad_norm": 0.578269898891449, + "learning_rate": 1.9735984611665847e-05, + "loss": 0.0817, + "step": 33983 + }, + { + "epoch": 0.7966384781729636, + "grad_norm": 0.2976333200931549, + "learning_rate": 1.9731591184672815e-05, + "loss": 0.0485, + "step": 33984 + }, + { + "epoch": 0.7966619197477687, + "grad_norm": 0.0476374551653862, + "learning_rate": 1.9727198193218398e-05, + "loss": 0.0052, + "step": 33985 + }, + { + "epoch": 0.7966853613225736, + "grad_norm": 0.28357887268066406, + "learning_rate": 1.972280563732647e-05, + "loss": 0.0886, + "step": 33986 + }, + { + "epoch": 0.7967088028973787, + "grad_norm": 0.5377218723297119, + "learning_rate": 1.9718413517020884e-05, + "loss": 0.776, + "step": 33987 + }, + { + "epoch": 0.7967322444721836, + "grad_norm": 0.39291682839393616, + "learning_rate": 1.9714021832325435e-05, + "loss": 0.0511, + "step": 33988 + }, + { + "epoch": 0.7967556860469887, + "grad_norm": 0.5463482141494751, + "learning_rate": 1.970963058326396e-05, + "loss": 0.0828, + "step": 33989 + }, + { + "epoch": 0.7967791276217936, + "grad_norm": 0.41756319999694824, + "learning_rate": 1.9705239769860296e-05, + "loss": 0.0601, + "step": 33990 + }, + { + "epoch": 0.7968025691965986, + "grad_norm": 0.5150948166847229, + "learning_rate": 1.9700849392138243e-05, + "loss": 0.0921, + "step": 33991 + }, + { + "epoch": 0.7968260107714036, + "grad_norm": 0.5964739918708801, + "learning_rate": 1.969645945012163e-05, + "loss": 0.0858, + "step": 33992 + }, + { + "epoch": 0.7968494523462086, + "grad_norm": 0.31838855147361755, + "learning_rate": 1.96920699438343e-05, + "loss": 0.0398, + "step": 33993 + }, + { + "epoch": 0.7968728939210136, + "grad_norm": 0.5015832185745239, + "learning_rate": 1.9687680873300062e-05, + "loss": 0.0872, + "step": 33994 + }, + { + "epoch": 0.7968963354958186, + "grad_norm": 0.727954089641571, + "learning_rate": 1.9683292238542726e-05, + "loss": 0.1541, + "step": 33995 + }, + { + "epoch": 0.7969197770706236, + "grad_norm": 0.6256834864616394, + "learning_rate": 1.9678904039586098e-05, + "loss": 0.0998, + "step": 33996 + }, + { + "epoch": 0.7969432186454286, + "grad_norm": 0.4266400933265686, + "learning_rate": 1.967451627645398e-05, + "loss": 0.0932, + "step": 33997 + }, + { + "epoch": 0.7969666602202335, + "grad_norm": 0.4013099670410156, + "learning_rate": 1.967012894917022e-05, + "loss": 0.0783, + "step": 33998 + }, + { + "epoch": 0.7969901017950386, + "grad_norm": 0.32690492272377014, + "learning_rate": 1.9665742057758562e-05, + "loss": 0.0533, + "step": 33999 + }, + { + "epoch": 0.7970135433698435, + "grad_norm": 0.5465080738067627, + "learning_rate": 1.966135560224288e-05, + "loss": 0.0324, + "step": 34000 + }, + { + "epoch": 0.7970369849446486, + "grad_norm": 0.16419188678264618, + "learning_rate": 1.9656969582646944e-05, + "loss": 0.0197, + "step": 34001 + }, + { + "epoch": 0.7970604265194535, + "grad_norm": 0.15294843912124634, + "learning_rate": 1.9652583998994543e-05, + "loss": 0.0118, + "step": 34002 + }, + { + "epoch": 0.7970838680942586, + "grad_norm": 0.24485498666763306, + "learning_rate": 1.9648198851309485e-05, + "loss": 0.0326, + "step": 34003 + }, + { + "epoch": 0.7971073096690635, + "grad_norm": 0.06741133332252502, + "learning_rate": 1.9643814139615534e-05, + "loss": 0.0086, + "step": 34004 + }, + { + "epoch": 0.7971307512438686, + "grad_norm": 0.13291503489017487, + "learning_rate": 1.9639429863936533e-05, + "loss": 0.0039, + "step": 34005 + }, + { + "epoch": 0.7971541928186736, + "grad_norm": 0.4001046121120453, + "learning_rate": 1.9635046024296246e-05, + "loss": 0.0402, + "step": 34006 + }, + { + "epoch": 0.7971776343934786, + "grad_norm": 0.5547946095466614, + "learning_rate": 1.9630662620718454e-05, + "loss": 0.1159, + "step": 34007 + }, + { + "epoch": 0.7972010759682836, + "grad_norm": 0.557915985584259, + "learning_rate": 1.9626279653226952e-05, + "loss": 0.7267, + "step": 34008 + }, + { + "epoch": 0.7972245175430885, + "grad_norm": 0.6300258040428162, + "learning_rate": 1.962189712184549e-05, + "loss": 0.041, + "step": 34009 + }, + { + "epoch": 0.7972479591178936, + "grad_norm": 0.3681977093219757, + "learning_rate": 1.961751502659789e-05, + "loss": 0.0341, + "step": 34010 + }, + { + "epoch": 0.7972714006926985, + "grad_norm": 0.6774725317955017, + "learning_rate": 1.9613133367507898e-05, + "loss": 0.0631, + "step": 34011 + }, + { + "epoch": 0.7972948422675036, + "grad_norm": 0.2647043764591217, + "learning_rate": 1.960875214459933e-05, + "loss": 0.0205, + "step": 34012 + }, + { + "epoch": 0.7973182838423085, + "grad_norm": 0.39889299869537354, + "learning_rate": 1.9604371357895922e-05, + "loss": 0.0768, + "step": 34013 + }, + { + "epoch": 0.7973417254171136, + "grad_norm": 0.16093656420707703, + "learning_rate": 1.959999100742147e-05, + "loss": 0.0287, + "step": 34014 + }, + { + "epoch": 0.7973651669919185, + "grad_norm": 0.5198678374290466, + "learning_rate": 1.959561109319973e-05, + "loss": 0.0753, + "step": 34015 + }, + { + "epoch": 0.7973886085667236, + "grad_norm": 0.5883759260177612, + "learning_rate": 1.9591231615254448e-05, + "loss": 0.1231, + "step": 34016 + }, + { + "epoch": 0.7974120501415285, + "grad_norm": 0.09876321256160736, + "learning_rate": 1.9586852573609393e-05, + "loss": 0.0087, + "step": 34017 + }, + { + "epoch": 0.7974354917163335, + "grad_norm": 0.3527567982673645, + "learning_rate": 1.958247396828835e-05, + "loss": 0.0609, + "step": 34018 + }, + { + "epoch": 0.7974589332911385, + "grad_norm": 0.6007838249206543, + "learning_rate": 1.9578095799315076e-05, + "loss": 0.0536, + "step": 34019 + }, + { + "epoch": 0.7974823748659435, + "grad_norm": 0.36164048314094543, + "learning_rate": 1.957371806671331e-05, + "loss": 0.084, + "step": 34020 + }, + { + "epoch": 0.7975058164407485, + "grad_norm": 0.6031360030174255, + "learning_rate": 1.9569340770506782e-05, + "loss": 0.4621, + "step": 34021 + }, + { + "epoch": 0.7975292580155535, + "grad_norm": 0.5382558107376099, + "learning_rate": 1.9564963910719303e-05, + "loss": 0.0597, + "step": 34022 + }, + { + "epoch": 0.7975526995903585, + "grad_norm": 0.14664162695407867, + "learning_rate": 1.9560587487374594e-05, + "loss": 0.0384, + "step": 34023 + }, + { + "epoch": 0.7975761411651635, + "grad_norm": 0.45173999667167664, + "learning_rate": 1.9556211500496367e-05, + "loss": 0.0442, + "step": 34024 + }, + { + "epoch": 0.7975995827399684, + "grad_norm": 0.4684939384460449, + "learning_rate": 1.955183595010843e-05, + "loss": 0.0271, + "step": 34025 + }, + { + "epoch": 0.7976230243147735, + "grad_norm": 0.11934885382652283, + "learning_rate": 1.95474608362345e-05, + "loss": 0.0113, + "step": 34026 + }, + { + "epoch": 0.7976464658895784, + "grad_norm": 0.7344972491264343, + "learning_rate": 1.9543086158898293e-05, + "loss": 0.0795, + "step": 34027 + }, + { + "epoch": 0.7976699074643835, + "grad_norm": 0.3950115442276001, + "learning_rate": 1.9538711918123575e-05, + "loss": 0.0849, + "step": 34028 + }, + { + "epoch": 0.7976933490391884, + "grad_norm": 0.5165933966636658, + "learning_rate": 1.9534338113934057e-05, + "loss": 0.5609, + "step": 34029 + }, + { + "epoch": 0.7977167906139935, + "grad_norm": 0.5571342706680298, + "learning_rate": 1.952996474635347e-05, + "loss": 0.0567, + "step": 34030 + }, + { + "epoch": 0.7977402321887984, + "grad_norm": 0.58207768201828, + "learning_rate": 1.9525591815405574e-05, + "loss": 0.178, + "step": 34031 + }, + { + "epoch": 0.7977636737636035, + "grad_norm": 0.42112526297569275, + "learning_rate": 1.952121932111408e-05, + "loss": 0.0352, + "step": 34032 + }, + { + "epoch": 0.7977871153384084, + "grad_norm": 0.11304270476102829, + "learning_rate": 1.9516847263502715e-05, + "loss": 0.0108, + "step": 34033 + }, + { + "epoch": 0.7978105569132135, + "grad_norm": 0.4382070302963257, + "learning_rate": 1.9512475642595174e-05, + "loss": 0.0849, + "step": 34034 + }, + { + "epoch": 0.7978339984880184, + "grad_norm": 0.44553640484809875, + "learning_rate": 1.9508104458415232e-05, + "loss": 0.056, + "step": 34035 + }, + { + "epoch": 0.7978574400628234, + "grad_norm": 0.6106178164482117, + "learning_rate": 1.9503733710986572e-05, + "loss": 0.0967, + "step": 34036 + }, + { + "epoch": 0.7978808816376284, + "grad_norm": 0.6817134022712708, + "learning_rate": 1.9499363400332894e-05, + "loss": 0.0669, + "step": 34037 + }, + { + "epoch": 0.7979043232124334, + "grad_norm": 0.4626786708831787, + "learning_rate": 1.9494993526477956e-05, + "loss": 0.0355, + "step": 34038 + }, + { + "epoch": 0.7979277647872384, + "grad_norm": 0.6975715160369873, + "learning_rate": 1.9490624089445454e-05, + "loss": 0.0443, + "step": 34039 + }, + { + "epoch": 0.7979512063620434, + "grad_norm": 0.572166383266449, + "learning_rate": 1.9486255089259085e-05, + "loss": 0.0696, + "step": 34040 + }, + { + "epoch": 0.7979746479368484, + "grad_norm": 0.2746383547782898, + "learning_rate": 1.948188652594256e-05, + "loss": 0.0241, + "step": 34041 + }, + { + "epoch": 0.7979980895116534, + "grad_norm": 0.4010014832019806, + "learning_rate": 1.9477518399519556e-05, + "loss": 0.068, + "step": 34042 + }, + { + "epoch": 0.7980215310864583, + "grad_norm": 0.18542812764644623, + "learning_rate": 1.9473150710013833e-05, + "loss": 0.0325, + "step": 34043 + }, + { + "epoch": 0.7980449726612634, + "grad_norm": 0.5176634192466736, + "learning_rate": 1.9468783457449046e-05, + "loss": 0.0997, + "step": 34044 + }, + { + "epoch": 0.7980684142360683, + "grad_norm": 0.500238835811615, + "learning_rate": 1.9464416641848916e-05, + "loss": 0.4747, + "step": 34045 + }, + { + "epoch": 0.7980918558108734, + "grad_norm": 0.5340089201927185, + "learning_rate": 1.9460050263237095e-05, + "loss": 0.1234, + "step": 34046 + }, + { + "epoch": 0.7981152973856783, + "grad_norm": 0.24040618538856506, + "learning_rate": 1.9455684321637337e-05, + "loss": 0.0363, + "step": 34047 + }, + { + "epoch": 0.7981387389604834, + "grad_norm": 0.6270033717155457, + "learning_rate": 1.9451318817073294e-05, + "loss": 0.1329, + "step": 34048 + }, + { + "epoch": 0.7981621805352883, + "grad_norm": 0.43513867259025574, + "learning_rate": 1.944695374956863e-05, + "loss": 0.4337, + "step": 34049 + }, + { + "epoch": 0.7981856221100934, + "grad_norm": 0.8481200933456421, + "learning_rate": 1.9442589119147092e-05, + "loss": 0.1825, + "step": 34050 + }, + { + "epoch": 0.7982090636848983, + "grad_norm": 0.1397368609905243, + "learning_rate": 1.9438224925832326e-05, + "loss": 0.0119, + "step": 34051 + }, + { + "epoch": 0.7982325052597034, + "grad_norm": 0.06908737123012543, + "learning_rate": 1.9433861169648028e-05, + "loss": 0.0117, + "step": 34052 + }, + { + "epoch": 0.7982559468345083, + "grad_norm": 0.9996734261512756, + "learning_rate": 1.942949785061785e-05, + "loss": 0.0411, + "step": 34053 + }, + { + "epoch": 0.7982793884093133, + "grad_norm": 0.5889319777488708, + "learning_rate": 1.942513496876549e-05, + "loss": 0.0748, + "step": 34054 + }, + { + "epoch": 0.7983028299841183, + "grad_norm": 0.5847730040550232, + "learning_rate": 1.9420772524114595e-05, + "loss": 0.0722, + "step": 34055 + }, + { + "epoch": 0.7983262715589233, + "grad_norm": 0.46521177887916565, + "learning_rate": 1.9416410516688865e-05, + "loss": 0.0676, + "step": 34056 + }, + { + "epoch": 0.7983497131337284, + "grad_norm": 0.3176164925098419, + "learning_rate": 1.9412048946511964e-05, + "loss": 0.0679, + "step": 34057 + }, + { + "epoch": 0.7983731547085333, + "grad_norm": 0.49215397238731384, + "learning_rate": 1.9407687813607535e-05, + "loss": 0.0738, + "step": 34058 + }, + { + "epoch": 0.7983965962833384, + "grad_norm": 0.2563806474208832, + "learning_rate": 1.940332711799927e-05, + "loss": 0.0337, + "step": 34059 + }, + { + "epoch": 0.7984200378581433, + "grad_norm": 0.2895713746547699, + "learning_rate": 1.939896685971082e-05, + "loss": 0.0503, + "step": 34060 + }, + { + "epoch": 0.7984434794329484, + "grad_norm": 0.13246609270572662, + "learning_rate": 1.9394607038765843e-05, + "loss": 0.0334, + "step": 34061 + }, + { + "epoch": 0.7984669210077533, + "grad_norm": 0.323760986328125, + "learning_rate": 1.9390247655187975e-05, + "loss": 0.0508, + "step": 34062 + }, + { + "epoch": 0.7984903625825583, + "grad_norm": 0.3938685357570648, + "learning_rate": 1.938588870900091e-05, + "loss": 0.0568, + "step": 34063 + }, + { + "epoch": 0.7985138041573633, + "grad_norm": 0.5696830153465271, + "learning_rate": 1.938153020022828e-05, + "loss": 0.1153, + "step": 34064 + }, + { + "epoch": 0.7985372457321683, + "grad_norm": 0.5177119970321655, + "learning_rate": 1.937717212889374e-05, + "loss": 0.1628, + "step": 34065 + }, + { + "epoch": 0.7985606873069733, + "grad_norm": 0.48102548718452454, + "learning_rate": 1.937281449502092e-05, + "loss": 0.1073, + "step": 34066 + }, + { + "epoch": 0.7985841288817783, + "grad_norm": 0.09900985658168793, + "learning_rate": 1.936845729863348e-05, + "loss": 0.02, + "step": 34067 + }, + { + "epoch": 0.7986075704565833, + "grad_norm": 0.1728137582540512, + "learning_rate": 1.9364100539755035e-05, + "loss": 0.0255, + "step": 34068 + }, + { + "epoch": 0.7986310120313883, + "grad_norm": 0.48090946674346924, + "learning_rate": 1.935974421840927e-05, + "loss": 0.0523, + "step": 34069 + }, + { + "epoch": 0.7986544536061932, + "grad_norm": 0.09927273541688919, + "learning_rate": 1.9355388334619785e-05, + "loss": 0.015, + "step": 34070 + }, + { + "epoch": 0.7986778951809983, + "grad_norm": 0.3655858635902405, + "learning_rate": 1.9351032888410238e-05, + "loss": 0.0755, + "step": 34071 + }, + { + "epoch": 0.7987013367558032, + "grad_norm": 0.34144267439842224, + "learning_rate": 1.9346677879804265e-05, + "loss": 0.046, + "step": 34072 + }, + { + "epoch": 0.7987247783306083, + "grad_norm": 0.3967948257923126, + "learning_rate": 1.9342323308825484e-05, + "loss": 0.3733, + "step": 34073 + }, + { + "epoch": 0.7987482199054132, + "grad_norm": 0.26752451062202454, + "learning_rate": 1.9337969175497528e-05, + "loss": 0.0371, + "step": 34074 + }, + { + "epoch": 0.7987716614802183, + "grad_norm": 0.6535712480545044, + "learning_rate": 1.933361547984399e-05, + "loss": 0.5383, + "step": 34075 + }, + { + "epoch": 0.7987951030550232, + "grad_norm": 0.33068856596946716, + "learning_rate": 1.9329262221888545e-05, + "loss": 0.0392, + "step": 34076 + }, + { + "epoch": 0.7988185446298283, + "grad_norm": 0.3371865153312683, + "learning_rate": 1.93249094016548e-05, + "loss": 0.0466, + "step": 34077 + }, + { + "epoch": 0.7988419862046332, + "grad_norm": 0.5064812302589417, + "learning_rate": 1.9320557019166362e-05, + "loss": 0.1006, + "step": 34078 + }, + { + "epoch": 0.7988654277794383, + "grad_norm": 0.4539395272731781, + "learning_rate": 1.9316205074446847e-05, + "loss": 0.0974, + "step": 34079 + }, + { + "epoch": 0.7988888693542432, + "grad_norm": 0.16788998246192932, + "learning_rate": 1.9311853567519846e-05, + "loss": 0.0421, + "step": 34080 + }, + { + "epoch": 0.7989123109290482, + "grad_norm": 0.5477123856544495, + "learning_rate": 1.9307502498409026e-05, + "loss": 0.0864, + "step": 34081 + }, + { + "epoch": 0.7989357525038532, + "grad_norm": 0.6125577688217163, + "learning_rate": 1.9303151867137936e-05, + "loss": 0.0672, + "step": 34082 + }, + { + "epoch": 0.7989591940786582, + "grad_norm": 0.20921824872493744, + "learning_rate": 1.929880167373024e-05, + "loss": 0.0492, + "step": 34083 + }, + { + "epoch": 0.7989826356534632, + "grad_norm": 0.40163174271583557, + "learning_rate": 1.929445191820952e-05, + "loss": 0.0529, + "step": 34084 + }, + { + "epoch": 0.7990060772282682, + "grad_norm": 0.22944669425487518, + "learning_rate": 1.929010260059937e-05, + "loss": 0.0239, + "step": 34085 + }, + { + "epoch": 0.7990295188030732, + "grad_norm": 0.18520763516426086, + "learning_rate": 1.9285753720923384e-05, + "loss": 0.0507, + "step": 34086 + }, + { + "epoch": 0.7990529603778782, + "grad_norm": 0.5494467616081238, + "learning_rate": 1.9281405279205155e-05, + "loss": 0.0803, + "step": 34087 + }, + { + "epoch": 0.7990764019526831, + "grad_norm": 0.31212711334228516, + "learning_rate": 1.9277057275468312e-05, + "loss": 0.0224, + "step": 34088 + }, + { + "epoch": 0.7990998435274882, + "grad_norm": 0.4122965931892395, + "learning_rate": 1.9272709709736425e-05, + "loss": 0.0667, + "step": 34089 + }, + { + "epoch": 0.7991232851022931, + "grad_norm": 0.7434238791465759, + "learning_rate": 1.9268362582033085e-05, + "loss": 0.149, + "step": 34090 + }, + { + "epoch": 0.7991467266770982, + "grad_norm": 0.5205325484275818, + "learning_rate": 1.926401589238188e-05, + "loss": 0.0915, + "step": 34091 + }, + { + "epoch": 0.7991701682519031, + "grad_norm": 0.4923035204410553, + "learning_rate": 1.9259669640806388e-05, + "loss": 0.5545, + "step": 34092 + }, + { + "epoch": 0.7991936098267082, + "grad_norm": 0.6643503308296204, + "learning_rate": 1.9255323827330186e-05, + "loss": 0.1093, + "step": 34093 + }, + { + "epoch": 0.7992170514015131, + "grad_norm": 0.2016552835702896, + "learning_rate": 1.9250978451976854e-05, + "loss": 0.0183, + "step": 34094 + }, + { + "epoch": 0.7992404929763182, + "grad_norm": 0.1370140016078949, + "learning_rate": 1.924663351477002e-05, + "loss": 0.0112, + "step": 34095 + }, + { + "epoch": 0.7992639345511231, + "grad_norm": 0.38913610577583313, + "learning_rate": 1.9242289015733217e-05, + "loss": 0.0667, + "step": 34096 + }, + { + "epoch": 0.7992873761259282, + "grad_norm": 0.5223564505577087, + "learning_rate": 1.923794495489003e-05, + "loss": 0.0914, + "step": 34097 + }, + { + "epoch": 0.7993108177007331, + "grad_norm": 0.5645748972892761, + "learning_rate": 1.9233601332264016e-05, + "loss": 0.0833, + "step": 34098 + }, + { + "epoch": 0.7993342592755381, + "grad_norm": 0.379803329706192, + "learning_rate": 1.922925814787876e-05, + "loss": 0.0552, + "step": 34099 + }, + { + "epoch": 0.7993577008503431, + "grad_norm": 0.3628944754600525, + "learning_rate": 1.9224915401757794e-05, + "loss": 0.0789, + "step": 34100 + }, + { + "epoch": 0.7993811424251481, + "grad_norm": 0.31511715054512024, + "learning_rate": 1.9220573093924732e-05, + "loss": 0.0493, + "step": 34101 + }, + { + "epoch": 0.7994045839999531, + "grad_norm": 0.3778521418571472, + "learning_rate": 1.9216231224403113e-05, + "loss": 0.0564, + "step": 34102 + }, + { + "epoch": 0.7994280255747581, + "grad_norm": 0.6846614480018616, + "learning_rate": 1.9211889793216487e-05, + "loss": 0.1887, + "step": 34103 + }, + { + "epoch": 0.799451467149563, + "grad_norm": 0.6439220905303955, + "learning_rate": 1.9207548800388432e-05, + "loss": 0.1654, + "step": 34104 + }, + { + "epoch": 0.7994749087243681, + "grad_norm": 0.16838189959526062, + "learning_rate": 1.9203208245942482e-05, + "loss": 0.0169, + "step": 34105 + }, + { + "epoch": 0.799498350299173, + "grad_norm": 0.4510754942893982, + "learning_rate": 1.919886812990217e-05, + "loss": 0.0429, + "step": 34106 + }, + { + "epoch": 0.7995217918739781, + "grad_norm": 0.15762346982955933, + "learning_rate": 1.919452845229107e-05, + "loss": 0.0368, + "step": 34107 + }, + { + "epoch": 0.7995452334487831, + "grad_norm": 0.575692355632782, + "learning_rate": 1.919018921313276e-05, + "loss": 0.0748, + "step": 34108 + }, + { + "epoch": 0.7995686750235881, + "grad_norm": 0.5380964875221252, + "learning_rate": 1.9185850412450746e-05, + "loss": 0.0899, + "step": 34109 + }, + { + "epoch": 0.7995921165983931, + "grad_norm": 0.4629177749156952, + "learning_rate": 1.9181512050268592e-05, + "loss": 0.1298, + "step": 34110 + }, + { + "epoch": 0.7996155581731981, + "grad_norm": 0.33335378766059875, + "learning_rate": 1.9177174126609822e-05, + "loss": 0.0417, + "step": 34111 + }, + { + "epoch": 0.7996389997480031, + "grad_norm": 0.45780399441719055, + "learning_rate": 1.9172836641497972e-05, + "loss": 0.0506, + "step": 34112 + }, + { + "epoch": 0.7996624413228081, + "grad_norm": 0.2273339182138443, + "learning_rate": 1.9168499594956567e-05, + "loss": 0.0425, + "step": 34113 + }, + { + "epoch": 0.7996858828976131, + "grad_norm": 0.1686350405216217, + "learning_rate": 1.916416298700917e-05, + "loss": 0.0205, + "step": 34114 + }, + { + "epoch": 0.799709324472418, + "grad_norm": 0.3918706774711609, + "learning_rate": 1.9159826817679305e-05, + "loss": 0.0661, + "step": 34115 + }, + { + "epoch": 0.7997327660472231, + "grad_norm": 0.5717669129371643, + "learning_rate": 1.9155491086990496e-05, + "loss": 0.7432, + "step": 34116 + }, + { + "epoch": 0.799756207622028, + "grad_norm": 0.5086197853088379, + "learning_rate": 1.9151155794966256e-05, + "loss": 0.0974, + "step": 34117 + }, + { + "epoch": 0.7997796491968331, + "grad_norm": 0.23621797561645508, + "learning_rate": 1.9146820941630104e-05, + "loss": 0.018, + "step": 34118 + }, + { + "epoch": 0.799803090771638, + "grad_norm": 0.5373913049697876, + "learning_rate": 1.9142486527005567e-05, + "loss": 0.0769, + "step": 34119 + }, + { + "epoch": 0.7998265323464431, + "grad_norm": 0.3876760005950928, + "learning_rate": 1.9138152551116195e-05, + "loss": 0.0493, + "step": 34120 + }, + { + "epoch": 0.799849973921248, + "grad_norm": 0.23788346350193024, + "learning_rate": 1.9133819013985488e-05, + "loss": 0.0241, + "step": 34121 + }, + { + "epoch": 0.7998734154960531, + "grad_norm": 0.3832484185695648, + "learning_rate": 1.9129485915636957e-05, + "loss": 0.0558, + "step": 34122 + }, + { + "epoch": 0.799896857070858, + "grad_norm": 0.324484646320343, + "learning_rate": 1.91251532560941e-05, + "loss": 0.03, + "step": 34123 + }, + { + "epoch": 0.799920298645663, + "grad_norm": 0.08605165779590607, + "learning_rate": 1.9120821035380433e-05, + "loss": 0.0151, + "step": 34124 + }, + { + "epoch": 0.799943740220468, + "grad_norm": 0.09049301594495773, + "learning_rate": 1.9116489253519444e-05, + "loss": 0.0161, + "step": 34125 + }, + { + "epoch": 0.799967181795273, + "grad_norm": 0.213335320353508, + "learning_rate": 1.911215791053469e-05, + "loss": 0.0369, + "step": 34126 + }, + { + "epoch": 0.799990623370078, + "grad_norm": 0.21471761167049408, + "learning_rate": 1.910782700644963e-05, + "loss": 0.0428, + "step": 34127 + }, + { + "epoch": 0.800014064944883, + "grad_norm": 0.3615885078907013, + "learning_rate": 1.910349654128778e-05, + "loss": 0.0394, + "step": 34128 + }, + { + "epoch": 0.800037506519688, + "grad_norm": 0.41955453157424927, + "learning_rate": 1.9099166515072643e-05, + "loss": 0.0774, + "step": 34129 + }, + { + "epoch": 0.800060948094493, + "grad_norm": 0.36625728011131287, + "learning_rate": 1.9094836927827687e-05, + "loss": 0.0505, + "step": 34130 + }, + { + "epoch": 0.800084389669298, + "grad_norm": 0.25484052300453186, + "learning_rate": 1.9090507779576404e-05, + "loss": 0.0374, + "step": 34131 + }, + { + "epoch": 0.800107831244103, + "grad_norm": 0.4651392102241516, + "learning_rate": 1.908617907034229e-05, + "loss": 0.103, + "step": 34132 + }, + { + "epoch": 0.800131272818908, + "grad_norm": 0.1792866587638855, + "learning_rate": 1.9081850800148882e-05, + "loss": 0.0265, + "step": 34133 + }, + { + "epoch": 0.800154714393713, + "grad_norm": 0.11624031513929367, + "learning_rate": 1.9077522969019622e-05, + "loss": 0.0237, + "step": 34134 + }, + { + "epoch": 0.8001781559685179, + "grad_norm": 0.5260119438171387, + "learning_rate": 1.9073195576978e-05, + "loss": 0.6816, + "step": 34135 + }, + { + "epoch": 0.800201597543323, + "grad_norm": 0.5259126424789429, + "learning_rate": 1.9068868624047498e-05, + "loss": 0.0991, + "step": 34136 + }, + { + "epoch": 0.8002250391181279, + "grad_norm": 0.3850935995578766, + "learning_rate": 1.9064542110251582e-05, + "loss": 0.0166, + "step": 34137 + }, + { + "epoch": 0.800248480692933, + "grad_norm": 0.38211750984191895, + "learning_rate": 1.9060216035613708e-05, + "loss": 0.0866, + "step": 34138 + }, + { + "epoch": 0.8002719222677379, + "grad_norm": 0.5089958906173706, + "learning_rate": 1.9055890400157417e-05, + "loss": 0.0506, + "step": 34139 + }, + { + "epoch": 0.800295363842543, + "grad_norm": 0.5158923864364624, + "learning_rate": 1.905156520390613e-05, + "loss": 0.0504, + "step": 34140 + }, + { + "epoch": 0.8003188054173479, + "grad_norm": 0.31313884258270264, + "learning_rate": 1.904724044688333e-05, + "loss": 0.2525, + "step": 34141 + }, + { + "epoch": 0.800342246992153, + "grad_norm": 0.21192792057991028, + "learning_rate": 1.9042916129112477e-05, + "loss": 0.0214, + "step": 34142 + }, + { + "epoch": 0.8003656885669579, + "grad_norm": 0.35474494099617004, + "learning_rate": 1.9038592250617004e-05, + "loss": 0.0524, + "step": 34143 + }, + { + "epoch": 0.8003891301417629, + "grad_norm": 0.12184768170118332, + "learning_rate": 1.903426881142044e-05, + "loss": 0.0068, + "step": 34144 + }, + { + "epoch": 0.8004125717165679, + "grad_norm": 0.48516350984573364, + "learning_rate": 1.902994581154619e-05, + "loss": 0.0802, + "step": 34145 + }, + { + "epoch": 0.8004360132913729, + "grad_norm": 0.30414879322052, + "learning_rate": 1.9025623251017756e-05, + "loss": 0.0332, + "step": 34146 + }, + { + "epoch": 0.8004594548661779, + "grad_norm": 0.08976183086633682, + "learning_rate": 1.902130112985856e-05, + "loss": 0.0151, + "step": 34147 + }, + { + "epoch": 0.8004828964409829, + "grad_norm": 0.4243966341018677, + "learning_rate": 1.9016979448092064e-05, + "loss": 0.0739, + "step": 34148 + }, + { + "epoch": 0.8005063380157879, + "grad_norm": 0.2282593846321106, + "learning_rate": 1.9012658205741708e-05, + "loss": 0.0174, + "step": 34149 + }, + { + "epoch": 0.8005297795905929, + "grad_norm": 0.12792664766311646, + "learning_rate": 1.9008337402830957e-05, + "loss": 0.0188, + "step": 34150 + }, + { + "epoch": 0.8005532211653978, + "grad_norm": 0.253398060798645, + "learning_rate": 1.9004017039383203e-05, + "loss": 0.0344, + "step": 34151 + }, + { + "epoch": 0.8005766627402029, + "grad_norm": 0.28829389810562134, + "learning_rate": 1.8999697115421967e-05, + "loss": 0.0501, + "step": 34152 + }, + { + "epoch": 0.8006001043150078, + "grad_norm": 0.3385864198207855, + "learning_rate": 1.8995377630970656e-05, + "loss": 0.029, + "step": 34153 + }, + { + "epoch": 0.8006235458898129, + "grad_norm": 0.45948678255081177, + "learning_rate": 1.89910585860527e-05, + "loss": 0.0506, + "step": 34154 + }, + { + "epoch": 0.8006469874646178, + "grad_norm": 0.30849549174308777, + "learning_rate": 1.8986739980691503e-05, + "loss": 0.0649, + "step": 34155 + }, + { + "epoch": 0.8006704290394229, + "grad_norm": 0.45302844047546387, + "learning_rate": 1.898242181491057e-05, + "loss": 0.0762, + "step": 34156 + }, + { + "epoch": 0.8006938706142278, + "grad_norm": 0.48842811584472656, + "learning_rate": 1.8978104088733262e-05, + "loss": 0.0919, + "step": 34157 + }, + { + "epoch": 0.8007173121890329, + "grad_norm": 0.5849717259407043, + "learning_rate": 1.8973786802183067e-05, + "loss": 0.5205, + "step": 34158 + }, + { + "epoch": 0.8007407537638379, + "grad_norm": 0.3292417824268341, + "learning_rate": 1.8969469955283392e-05, + "loss": 0.0491, + "step": 34159 + }, + { + "epoch": 0.8007641953386428, + "grad_norm": 0.6394347548484802, + "learning_rate": 1.8965153548057636e-05, + "loss": 0.0702, + "step": 34160 + }, + { + "epoch": 0.8007876369134479, + "grad_norm": 0.5539954900741577, + "learning_rate": 1.8960837580529244e-05, + "loss": 0.1019, + "step": 34161 + }, + { + "epoch": 0.8008110784882528, + "grad_norm": 0.31695443391799927, + "learning_rate": 1.8956522052721625e-05, + "loss": 0.0753, + "step": 34162 + }, + { + "epoch": 0.8008345200630579, + "grad_norm": 0.09798817336559296, + "learning_rate": 1.895220696465817e-05, + "loss": 0.018, + "step": 34163 + }, + { + "epoch": 0.8008579616378628, + "grad_norm": 0.6291114091873169, + "learning_rate": 1.8947892316362347e-05, + "loss": 0.114, + "step": 34164 + }, + { + "epoch": 0.8008814032126679, + "grad_norm": 0.4443124234676361, + "learning_rate": 1.894357810785753e-05, + "loss": 0.041, + "step": 34165 + }, + { + "epoch": 0.8009048447874728, + "grad_norm": 0.4671596884727478, + "learning_rate": 1.8939264339167152e-05, + "loss": 0.12, + "step": 34166 + }, + { + "epoch": 0.8009282863622779, + "grad_norm": 0.4783782958984375, + "learning_rate": 1.893495101031457e-05, + "loss": 0.0801, + "step": 34167 + }, + { + "epoch": 0.8009517279370828, + "grad_norm": 0.48139527440071106, + "learning_rate": 1.893063812132325e-05, + "loss": 0.5394, + "step": 34168 + }, + { + "epoch": 0.8009751695118879, + "grad_norm": 0.11701072752475739, + "learning_rate": 1.8926325672216562e-05, + "loss": 0.0125, + "step": 34169 + }, + { + "epoch": 0.8009986110866928, + "grad_norm": 0.31677794456481934, + "learning_rate": 1.892201366301789e-05, + "loss": 0.034, + "step": 34170 + }, + { + "epoch": 0.8010220526614978, + "grad_norm": 0.3839840888977051, + "learning_rate": 1.8917702093750676e-05, + "loss": 0.0575, + "step": 34171 + }, + { + "epoch": 0.8010454942363028, + "grad_norm": 0.3428306579589844, + "learning_rate": 1.8913390964438294e-05, + "loss": 0.0794, + "step": 34172 + }, + { + "epoch": 0.8010689358111078, + "grad_norm": 0.16461926698684692, + "learning_rate": 1.8909080275104128e-05, + "loss": 0.031, + "step": 34173 + }, + { + "epoch": 0.8010923773859128, + "grad_norm": 0.24857699871063232, + "learning_rate": 1.8904770025771577e-05, + "loss": 0.0216, + "step": 34174 + }, + { + "epoch": 0.8011158189607178, + "grad_norm": 0.20742565393447876, + "learning_rate": 1.8900460216464024e-05, + "loss": 0.0184, + "step": 34175 + }, + { + "epoch": 0.8011392605355228, + "grad_norm": 0.15645553171634674, + "learning_rate": 1.8896150847204817e-05, + "loss": 0.0176, + "step": 34176 + }, + { + "epoch": 0.8011627021103278, + "grad_norm": 0.5145191550254822, + "learning_rate": 1.8891841918017418e-05, + "loss": 0.0874, + "step": 34177 + }, + { + "epoch": 0.8011861436851327, + "grad_norm": 0.7603044509887695, + "learning_rate": 1.8887533428925154e-05, + "loss": 0.7809, + "step": 34178 + }, + { + "epoch": 0.8012095852599378, + "grad_norm": 0.11246559023857117, + "learning_rate": 1.888322537995142e-05, + "loss": 0.0138, + "step": 34179 + }, + { + "epoch": 0.8012330268347427, + "grad_norm": 0.7461226582527161, + "learning_rate": 1.8878917771119554e-05, + "loss": 0.126, + "step": 34180 + }, + { + "epoch": 0.8012564684095478, + "grad_norm": 0.7625440955162048, + "learning_rate": 1.887461060245299e-05, + "loss": 0.1337, + "step": 34181 + }, + { + "epoch": 0.8012799099843527, + "grad_norm": 0.19686460494995117, + "learning_rate": 1.8870303873975083e-05, + "loss": 0.0237, + "step": 34182 + }, + { + "epoch": 0.8013033515591578, + "grad_norm": 0.22718650102615356, + "learning_rate": 1.8865997585709138e-05, + "loss": 0.0224, + "step": 34183 + }, + { + "epoch": 0.8013267931339627, + "grad_norm": 0.3317948877811432, + "learning_rate": 1.886169173767861e-05, + "loss": 0.0369, + "step": 34184 + }, + { + "epoch": 0.8013502347087678, + "grad_norm": 0.614273190498352, + "learning_rate": 1.8857386329906836e-05, + "loss": 0.1255, + "step": 34185 + }, + { + "epoch": 0.8013736762835727, + "grad_norm": 0.07060734182596207, + "learning_rate": 1.8853081362417147e-05, + "loss": 0.0108, + "step": 34186 + }, + { + "epoch": 0.8013971178583777, + "grad_norm": 0.3764779567718506, + "learning_rate": 1.8848776835232927e-05, + "loss": 0.0179, + "step": 34187 + }, + { + "epoch": 0.8014205594331827, + "grad_norm": 0.08244488388299942, + "learning_rate": 1.8844472748377494e-05, + "loss": 0.0119, + "step": 34188 + }, + { + "epoch": 0.8014440010079877, + "grad_norm": 0.28883010149002075, + "learning_rate": 1.8840169101874274e-05, + "loss": 0.0302, + "step": 34189 + }, + { + "epoch": 0.8014674425827927, + "grad_norm": 0.3738764226436615, + "learning_rate": 1.8835865895746562e-05, + "loss": 0.0744, + "step": 34190 + }, + { + "epoch": 0.8014908841575977, + "grad_norm": 0.5464391112327576, + "learning_rate": 1.8831563130017738e-05, + "loss": 0.0855, + "step": 34191 + }, + { + "epoch": 0.8015143257324027, + "grad_norm": 0.22983765602111816, + "learning_rate": 1.88272608047111e-05, + "loss": 0.0413, + "step": 34192 + }, + { + "epoch": 0.8015377673072077, + "grad_norm": 0.12452863156795502, + "learning_rate": 1.8822958919850053e-05, + "loss": 0.0114, + "step": 34193 + }, + { + "epoch": 0.8015612088820127, + "grad_norm": 0.5471017360687256, + "learning_rate": 1.8818657475457912e-05, + "loss": 0.0783, + "step": 34194 + }, + { + "epoch": 0.8015846504568177, + "grad_norm": 0.5403804779052734, + "learning_rate": 1.8814356471558004e-05, + "loss": 0.0995, + "step": 34195 + }, + { + "epoch": 0.8016080920316226, + "grad_norm": 0.4928397834300995, + "learning_rate": 1.8810055908173695e-05, + "loss": 0.0992, + "step": 34196 + }, + { + "epoch": 0.8016315336064277, + "grad_norm": 0.34455782175064087, + "learning_rate": 1.8805755785328304e-05, + "loss": 0.0429, + "step": 34197 + }, + { + "epoch": 0.8016549751812326, + "grad_norm": 0.09215959161520004, + "learning_rate": 1.880145610304517e-05, + "loss": 0.0115, + "step": 34198 + }, + { + "epoch": 0.8016784167560377, + "grad_norm": 0.37901216745376587, + "learning_rate": 1.879715686134762e-05, + "loss": 0.0674, + "step": 34199 + }, + { + "epoch": 0.8017018583308426, + "grad_norm": 0.40833693742752075, + "learning_rate": 1.8792858060258977e-05, + "loss": 0.1231, + "step": 34200 + }, + { + "epoch": 0.8017252999056477, + "grad_norm": 0.47295480966567993, + "learning_rate": 1.8788559699802544e-05, + "loss": 0.0422, + "step": 34201 + }, + { + "epoch": 0.8017487414804526, + "grad_norm": 0.2028752863407135, + "learning_rate": 1.87842617800017e-05, + "loss": 0.0137, + "step": 34202 + }, + { + "epoch": 0.8017721830552577, + "grad_norm": 0.18165738880634308, + "learning_rate": 1.8779964300879728e-05, + "loss": 0.0243, + "step": 34203 + }, + { + "epoch": 0.8017956246300626, + "grad_norm": 0.45635467767715454, + "learning_rate": 1.877566726245993e-05, + "loss": 0.0679, + "step": 34204 + }, + { + "epoch": 0.8018190662048676, + "grad_norm": 0.5290080904960632, + "learning_rate": 1.8771370664765676e-05, + "loss": 0.0454, + "step": 34205 + }, + { + "epoch": 0.8018425077796726, + "grad_norm": 0.5538951754570007, + "learning_rate": 1.8767074507820237e-05, + "loss": 0.0856, + "step": 34206 + }, + { + "epoch": 0.8018659493544776, + "grad_norm": 0.37073609232902527, + "learning_rate": 1.8762778791646953e-05, + "loss": 0.0444, + "step": 34207 + }, + { + "epoch": 0.8018893909292826, + "grad_norm": 0.6473044157028198, + "learning_rate": 1.8758483516269075e-05, + "loss": 0.1164, + "step": 34208 + }, + { + "epoch": 0.8019128325040876, + "grad_norm": 0.5721173882484436, + "learning_rate": 1.8754188681709983e-05, + "loss": 0.132, + "step": 34209 + }, + { + "epoch": 0.8019362740788927, + "grad_norm": 0.5317147970199585, + "learning_rate": 1.8749894287992942e-05, + "loss": 0.059, + "step": 34210 + }, + { + "epoch": 0.8019597156536976, + "grad_norm": 0.25009188055992126, + "learning_rate": 1.8745600335141266e-05, + "loss": 0.0376, + "step": 34211 + }, + { + "epoch": 0.8019831572285027, + "grad_norm": 0.22530794143676758, + "learning_rate": 1.874130682317824e-05, + "loss": 0.0329, + "step": 34212 + }, + { + "epoch": 0.8020065988033076, + "grad_norm": 0.37700027227401733, + "learning_rate": 1.8737013752127174e-05, + "loss": 0.0765, + "step": 34213 + }, + { + "epoch": 0.8020300403781127, + "grad_norm": 0.993575394153595, + "learning_rate": 1.873272112201133e-05, + "loss": 0.1526, + "step": 34214 + }, + { + "epoch": 0.8020534819529176, + "grad_norm": 0.42859527468681335, + "learning_rate": 1.8728428932854046e-05, + "loss": 0.1035, + "step": 34215 + }, + { + "epoch": 0.8020769235277226, + "grad_norm": 0.7465382218360901, + "learning_rate": 1.872413718467857e-05, + "loss": 0.1532, + "step": 34216 + }, + { + "epoch": 0.8021003651025276, + "grad_norm": 0.3029557764530182, + "learning_rate": 1.8719845877508237e-05, + "loss": 0.0111, + "step": 34217 + }, + { + "epoch": 0.8021238066773326, + "grad_norm": 0.09742260724306107, + "learning_rate": 1.8715555011366302e-05, + "loss": 0.0115, + "step": 34218 + }, + { + "epoch": 0.8021472482521376, + "grad_norm": 0.26526933908462524, + "learning_rate": 1.8711264586276056e-05, + "loss": 0.0355, + "step": 34219 + }, + { + "epoch": 0.8021706898269426, + "grad_norm": 0.48734334111213684, + "learning_rate": 1.8706974602260772e-05, + "loss": 0.1013, + "step": 34220 + }, + { + "epoch": 0.8021941314017476, + "grad_norm": 0.6651628017425537, + "learning_rate": 1.8702685059343706e-05, + "loss": 0.1705, + "step": 34221 + }, + { + "epoch": 0.8022175729765526, + "grad_norm": 0.4694649577140808, + "learning_rate": 1.869839595754819e-05, + "loss": 0.0498, + "step": 34222 + }, + { + "epoch": 0.8022410145513575, + "grad_norm": 0.3956429660320282, + "learning_rate": 1.8694107296897445e-05, + "loss": 0.0741, + "step": 34223 + }, + { + "epoch": 0.8022644561261626, + "grad_norm": 0.6106464266777039, + "learning_rate": 1.8689819077414773e-05, + "loss": 0.161, + "step": 34224 + }, + { + "epoch": 0.8022878977009675, + "grad_norm": 0.42782139778137207, + "learning_rate": 1.8685531299123428e-05, + "loss": 0.0408, + "step": 34225 + }, + { + "epoch": 0.8023113392757726, + "grad_norm": 0.2406064122915268, + "learning_rate": 1.8681243962046646e-05, + "loss": 0.0295, + "step": 34226 + }, + { + "epoch": 0.8023347808505775, + "grad_norm": 0.5856611132621765, + "learning_rate": 1.8676957066207744e-05, + "loss": 0.0768, + "step": 34227 + }, + { + "epoch": 0.8023582224253826, + "grad_norm": 0.09576143324375153, + "learning_rate": 1.8672670611629938e-05, + "loss": 0.0153, + "step": 34228 + }, + { + "epoch": 0.8023816640001875, + "grad_norm": 0.6680250763893127, + "learning_rate": 1.866838459833654e-05, + "loss": 0.1052, + "step": 34229 + }, + { + "epoch": 0.8024051055749926, + "grad_norm": 0.6268686652183533, + "learning_rate": 1.866409902635077e-05, + "loss": 0.1055, + "step": 34230 + }, + { + "epoch": 0.8024285471497975, + "grad_norm": 0.5060523152351379, + "learning_rate": 1.8659813895695877e-05, + "loss": 0.1102, + "step": 34231 + }, + { + "epoch": 0.8024519887246025, + "grad_norm": 0.1482478380203247, + "learning_rate": 1.865552920639513e-05, + "loss": 0.0293, + "step": 34232 + }, + { + "epoch": 0.8024754302994075, + "grad_norm": 0.3758324980735779, + "learning_rate": 1.8651244958471735e-05, + "loss": 0.053, + "step": 34233 + }, + { + "epoch": 0.8024988718742125, + "grad_norm": 0.32326942682266235, + "learning_rate": 1.8646961151949006e-05, + "loss": 0.0591, + "step": 34234 + }, + { + "epoch": 0.8025223134490175, + "grad_norm": 0.36617833375930786, + "learning_rate": 1.864267778685015e-05, + "loss": 0.0757, + "step": 34235 + }, + { + "epoch": 0.8025457550238225, + "grad_norm": 0.2446012794971466, + "learning_rate": 1.8638394863198416e-05, + "loss": 0.0238, + "step": 34236 + }, + { + "epoch": 0.8025691965986275, + "grad_norm": 0.49589744210243225, + "learning_rate": 1.8634112381017034e-05, + "loss": 0.0782, + "step": 34237 + }, + { + "epoch": 0.8025926381734325, + "grad_norm": 0.14731523394584656, + "learning_rate": 1.8629830340329247e-05, + "loss": 0.0224, + "step": 34238 + }, + { + "epoch": 0.8026160797482375, + "grad_norm": 0.14852334558963776, + "learning_rate": 1.862554874115826e-05, + "loss": 0.0272, + "step": 34239 + }, + { + "epoch": 0.8026395213230425, + "grad_norm": 0.23056451976299286, + "learning_rate": 1.8621267583527347e-05, + "loss": 0.0418, + "step": 34240 + }, + { + "epoch": 0.8026629628978474, + "grad_norm": 0.569015622138977, + "learning_rate": 1.861698686745974e-05, + "loss": 0.0902, + "step": 34241 + }, + { + "epoch": 0.8026864044726525, + "grad_norm": 0.3784351348876953, + "learning_rate": 1.8612706592978647e-05, + "loss": 0.0565, + "step": 34242 + }, + { + "epoch": 0.8027098460474574, + "grad_norm": 0.5407015681266785, + "learning_rate": 1.86084267601073e-05, + "loss": 0.6327, + "step": 34243 + }, + { + "epoch": 0.8027332876222625, + "grad_norm": 0.33606091141700745, + "learning_rate": 1.860414736886892e-05, + "loss": 0.0748, + "step": 34244 + }, + { + "epoch": 0.8027567291970674, + "grad_norm": 0.560910701751709, + "learning_rate": 1.8599868419286727e-05, + "loss": 0.0644, + "step": 34245 + }, + { + "epoch": 0.8027801707718725, + "grad_norm": 0.15967239439487457, + "learning_rate": 1.8595589911383902e-05, + "loss": 0.025, + "step": 34246 + }, + { + "epoch": 0.8028036123466774, + "grad_norm": 0.4938678741455078, + "learning_rate": 1.8591311845183725e-05, + "loss": 0.1368, + "step": 34247 + }, + { + "epoch": 0.8028270539214825, + "grad_norm": 0.11090510338544846, + "learning_rate": 1.8587034220709376e-05, + "loss": 0.0236, + "step": 34248 + }, + { + "epoch": 0.8028504954962874, + "grad_norm": 0.45239365100860596, + "learning_rate": 1.8582757037984078e-05, + "loss": 0.0536, + "step": 34249 + }, + { + "epoch": 0.8028739370710924, + "grad_norm": 0.4563693404197693, + "learning_rate": 1.857848029703102e-05, + "loss": 0.0607, + "step": 34250 + }, + { + "epoch": 0.8028973786458974, + "grad_norm": 0.19740207493305206, + "learning_rate": 1.857420399787342e-05, + "loss": 0.0376, + "step": 34251 + }, + { + "epoch": 0.8029208202207024, + "grad_norm": 0.25277695059776306, + "learning_rate": 1.8569928140534455e-05, + "loss": 0.0335, + "step": 34252 + }, + { + "epoch": 0.8029442617955074, + "grad_norm": 0.4087314307689667, + "learning_rate": 1.8565652725037353e-05, + "loss": 0.0489, + "step": 34253 + }, + { + "epoch": 0.8029677033703124, + "grad_norm": 0.4732700288295746, + "learning_rate": 1.8561377751405328e-05, + "loss": 0.0866, + "step": 34254 + }, + { + "epoch": 0.8029911449451174, + "grad_norm": 0.11533928662538528, + "learning_rate": 1.8557103219661553e-05, + "loss": 0.0251, + "step": 34255 + }, + { + "epoch": 0.8030145865199224, + "grad_norm": 0.8840770721435547, + "learning_rate": 1.855282912982924e-05, + "loss": 0.1686, + "step": 34256 + }, + { + "epoch": 0.8030380280947274, + "grad_norm": 0.4215922951698303, + "learning_rate": 1.854855548193155e-05, + "loss": 0.0529, + "step": 34257 + }, + { + "epoch": 0.8030614696695324, + "grad_norm": 0.24767287075519562, + "learning_rate": 1.85442822759917e-05, + "loss": 0.0348, + "step": 34258 + }, + { + "epoch": 0.8030849112443373, + "grad_norm": 0.4014054238796234, + "learning_rate": 1.854000951203284e-05, + "loss": 0.0634, + "step": 34259 + }, + { + "epoch": 0.8031083528191424, + "grad_norm": 0.26981112360954285, + "learning_rate": 1.8535737190078195e-05, + "loss": 0.0529, + "step": 34260 + }, + { + "epoch": 0.8031317943939474, + "grad_norm": 0.20823431015014648, + "learning_rate": 1.853146531015094e-05, + "loss": 0.0296, + "step": 34261 + }, + { + "epoch": 0.8031552359687524, + "grad_norm": 0.43128982186317444, + "learning_rate": 1.852719387227424e-05, + "loss": 0.0713, + "step": 34262 + }, + { + "epoch": 0.8031786775435574, + "grad_norm": 0.4948301613330841, + "learning_rate": 1.852292287647128e-05, + "loss": 0.1257, + "step": 34263 + }, + { + "epoch": 0.8032021191183624, + "grad_norm": 0.36096182465553284, + "learning_rate": 1.8518652322765206e-05, + "loss": 0.0448, + "step": 34264 + }, + { + "epoch": 0.8032255606931674, + "grad_norm": 0.5714245438575745, + "learning_rate": 1.8514382211179236e-05, + "loss": 0.0978, + "step": 34265 + }, + { + "epoch": 0.8032490022679724, + "grad_norm": 0.14019666612148285, + "learning_rate": 1.851011254173649e-05, + "loss": 0.0171, + "step": 34266 + }, + { + "epoch": 0.8032724438427774, + "grad_norm": 0.49546903371810913, + "learning_rate": 1.85058433144602e-05, + "loss": 0.1067, + "step": 34267 + }, + { + "epoch": 0.8032958854175823, + "grad_norm": 0.1808004230260849, + "learning_rate": 1.850157452937349e-05, + "loss": 0.0293, + "step": 34268 + }, + { + "epoch": 0.8033193269923874, + "grad_norm": 0.4751138389110565, + "learning_rate": 1.849730618649953e-05, + "loss": 0.0589, + "step": 34269 + }, + { + "epoch": 0.8033427685671923, + "grad_norm": 0.274291455745697, + "learning_rate": 1.8493038285861485e-05, + "loss": 0.0372, + "step": 34270 + }, + { + "epoch": 0.8033662101419974, + "grad_norm": 0.3900502920150757, + "learning_rate": 1.848877082748247e-05, + "loss": 0.0575, + "step": 34271 + }, + { + "epoch": 0.8033896517168023, + "grad_norm": 0.37518492341041565, + "learning_rate": 1.8484503811385702e-05, + "loss": 0.0594, + "step": 34272 + }, + { + "epoch": 0.8034130932916074, + "grad_norm": 0.43649742007255554, + "learning_rate": 1.848023723759431e-05, + "loss": 0.1067, + "step": 34273 + }, + { + "epoch": 0.8034365348664123, + "grad_norm": 0.41683652997016907, + "learning_rate": 1.847597110613145e-05, + "loss": 0.0643, + "step": 34274 + }, + { + "epoch": 0.8034599764412174, + "grad_norm": 0.7936809659004211, + "learning_rate": 1.8471705417020258e-05, + "loss": 0.126, + "step": 34275 + }, + { + "epoch": 0.8034834180160223, + "grad_norm": 0.08534763008356094, + "learning_rate": 1.846744017028389e-05, + "loss": 0.0146, + "step": 34276 + }, + { + "epoch": 0.8035068595908273, + "grad_norm": 0.4148710072040558, + "learning_rate": 1.846317536594545e-05, + "loss": 0.0804, + "step": 34277 + }, + { + "epoch": 0.8035303011656323, + "grad_norm": 0.37014177441596985, + "learning_rate": 1.8458911004028113e-05, + "loss": 0.0659, + "step": 34278 + }, + { + "epoch": 0.8035537427404373, + "grad_norm": 0.4127615988254547, + "learning_rate": 1.8454647084555053e-05, + "loss": 0.5389, + "step": 34279 + }, + { + "epoch": 0.8035771843152423, + "grad_norm": 0.39666110277175903, + "learning_rate": 1.845038360754936e-05, + "loss": 0.0774, + "step": 34280 + }, + { + "epoch": 0.8036006258900473, + "grad_norm": 0.40164944529533386, + "learning_rate": 1.8446120573034188e-05, + "loss": 0.0343, + "step": 34281 + }, + { + "epoch": 0.8036240674648523, + "grad_norm": 0.2918759882450104, + "learning_rate": 1.844185798103265e-05, + "loss": 0.0365, + "step": 34282 + }, + { + "epoch": 0.8036475090396573, + "grad_norm": 0.3151133358478546, + "learning_rate": 1.8437595831567888e-05, + "loss": 0.0294, + "step": 34283 + }, + { + "epoch": 0.8036709506144623, + "grad_norm": 0.9661442041397095, + "learning_rate": 1.8433334124662994e-05, + "loss": 0.1435, + "step": 34284 + }, + { + "epoch": 0.8036943921892673, + "grad_norm": 0.2853556275367737, + "learning_rate": 1.8429072860341156e-05, + "loss": 0.0554, + "step": 34285 + }, + { + "epoch": 0.8037178337640722, + "grad_norm": 0.47681868076324463, + "learning_rate": 1.842481203862546e-05, + "loss": 0.1035, + "step": 34286 + }, + { + "epoch": 0.8037412753388773, + "grad_norm": 0.1581914871931076, + "learning_rate": 1.842055165953903e-05, + "loss": 0.0295, + "step": 34287 + }, + { + "epoch": 0.8037647169136822, + "grad_norm": 0.5439826250076294, + "learning_rate": 1.8416291723104974e-05, + "loss": 0.0766, + "step": 34288 + }, + { + "epoch": 0.8037881584884873, + "grad_norm": 0.5073341727256775, + "learning_rate": 1.8412032229346388e-05, + "loss": 0.0995, + "step": 34289 + }, + { + "epoch": 0.8038116000632922, + "grad_norm": 0.519900918006897, + "learning_rate": 1.8407773178286435e-05, + "loss": 0.1245, + "step": 34290 + }, + { + "epoch": 0.8038350416380973, + "grad_norm": 0.5830562710762024, + "learning_rate": 1.8403514569948178e-05, + "loss": 0.092, + "step": 34291 + }, + { + "epoch": 0.8038584832129022, + "grad_norm": 0.8178461194038391, + "learning_rate": 1.839925640435477e-05, + "loss": 0.0423, + "step": 34292 + }, + { + "epoch": 0.8038819247877073, + "grad_norm": 0.29898813366889954, + "learning_rate": 1.8394998681529286e-05, + "loss": 0.0227, + "step": 34293 + }, + { + "epoch": 0.8039053663625122, + "grad_norm": 0.4767317771911621, + "learning_rate": 1.839074140149484e-05, + "loss": 0.5014, + "step": 34294 + }, + { + "epoch": 0.8039288079373172, + "grad_norm": 0.3358484208583832, + "learning_rate": 1.838648456427452e-05, + "loss": 0.0432, + "step": 34295 + }, + { + "epoch": 0.8039522495121222, + "grad_norm": 0.3443777859210968, + "learning_rate": 1.8382228169891436e-05, + "loss": 0.0286, + "step": 34296 + }, + { + "epoch": 0.8039756910869272, + "grad_norm": 0.33610352873802185, + "learning_rate": 1.8377972218368644e-05, + "loss": 0.0597, + "step": 34297 + }, + { + "epoch": 0.8039991326617322, + "grad_norm": 0.15824568271636963, + "learning_rate": 1.8373716709729307e-05, + "loss": 0.0438, + "step": 34298 + }, + { + "epoch": 0.8040225742365372, + "grad_norm": 0.5926572680473328, + "learning_rate": 1.8369461643996467e-05, + "loss": 0.5142, + "step": 34299 + }, + { + "epoch": 0.8040460158113422, + "grad_norm": 0.5615009069442749, + "learning_rate": 1.836520702119322e-05, + "loss": 0.0816, + "step": 34300 + }, + { + "epoch": 0.8040694573861472, + "grad_norm": 0.07619286328554153, + "learning_rate": 1.836095284134264e-05, + "loss": 0.0177, + "step": 34301 + }, + { + "epoch": 0.8040928989609522, + "grad_norm": 0.2663669288158417, + "learning_rate": 1.8356699104467843e-05, + "loss": 0.0417, + "step": 34302 + }, + { + "epoch": 0.8041163405357572, + "grad_norm": 0.5568036437034607, + "learning_rate": 1.83524458105919e-05, + "loss": 0.1641, + "step": 34303 + }, + { + "epoch": 0.8041397821105621, + "grad_norm": 0.0992436632514, + "learning_rate": 1.8348192959737853e-05, + "loss": 0.0208, + "step": 34304 + }, + { + "epoch": 0.8041632236853672, + "grad_norm": 0.847632884979248, + "learning_rate": 1.834394055192883e-05, + "loss": 0.0833, + "step": 34305 + }, + { + "epoch": 0.8041866652601721, + "grad_norm": 0.16926313936710358, + "learning_rate": 1.8339688587187888e-05, + "loss": 0.0236, + "step": 34306 + }, + { + "epoch": 0.8042101068349772, + "grad_norm": 0.18067240715026855, + "learning_rate": 1.8335437065538085e-05, + "loss": 0.0173, + "step": 34307 + }, + { + "epoch": 0.8042335484097821, + "grad_norm": 0.47323283553123474, + "learning_rate": 1.8331185987002496e-05, + "loss": 0.0586, + "step": 34308 + }, + { + "epoch": 0.8042569899845872, + "grad_norm": 0.6625989675521851, + "learning_rate": 1.8326935351604167e-05, + "loss": 0.1176, + "step": 34309 + }, + { + "epoch": 0.8042804315593921, + "grad_norm": 0.3667525351047516, + "learning_rate": 1.8322685159366205e-05, + "loss": 0.0482, + "step": 34310 + }, + { + "epoch": 0.8043038731341972, + "grad_norm": 0.28348833322525024, + "learning_rate": 1.8318435410311653e-05, + "loss": 0.394, + "step": 34311 + }, + { + "epoch": 0.8043273147090022, + "grad_norm": 0.7376950979232788, + "learning_rate": 1.831418610446357e-05, + "loss": 0.0961, + "step": 34312 + }, + { + "epoch": 0.8043507562838071, + "grad_norm": 0.43619683384895325, + "learning_rate": 1.8309937241844976e-05, + "loss": 0.397, + "step": 34313 + }, + { + "epoch": 0.8043741978586122, + "grad_norm": 0.3768080174922943, + "learning_rate": 1.8305688822478994e-05, + "loss": 0.0925, + "step": 34314 + }, + { + "epoch": 0.8043976394334171, + "grad_norm": 0.4220537841320038, + "learning_rate": 1.830144084638864e-05, + "loss": 0.0372, + "step": 34315 + }, + { + "epoch": 0.8044210810082222, + "grad_norm": 0.5155550241470337, + "learning_rate": 1.829719331359694e-05, + "loss": 0.1225, + "step": 34316 + }, + { + "epoch": 0.8044445225830271, + "grad_norm": 0.61102294921875, + "learning_rate": 1.8292946224126983e-05, + "loss": 0.1157, + "step": 34317 + }, + { + "epoch": 0.8044679641578322, + "grad_norm": 0.1906900852918625, + "learning_rate": 1.82886995780018e-05, + "loss": 0.0266, + "step": 34318 + }, + { + "epoch": 0.8044914057326371, + "grad_norm": 0.17246456444263458, + "learning_rate": 1.8284453375244438e-05, + "loss": 0.0199, + "step": 34319 + }, + { + "epoch": 0.8045148473074422, + "grad_norm": 0.9033164978027344, + "learning_rate": 1.828020761587792e-05, + "loss": 0.5333, + "step": 34320 + }, + { + "epoch": 0.8045382888822471, + "grad_norm": 0.6403535604476929, + "learning_rate": 1.82759622999253e-05, + "loss": 0.1396, + "step": 34321 + }, + { + "epoch": 0.8045617304570521, + "grad_norm": 0.19403903186321259, + "learning_rate": 1.8271717427409586e-05, + "loss": 0.0336, + "step": 34322 + }, + { + "epoch": 0.8045851720318571, + "grad_norm": 0.37713634967803955, + "learning_rate": 1.826747299835384e-05, + "loss": 0.0559, + "step": 34323 + }, + { + "epoch": 0.8046086136066621, + "grad_norm": 0.4421235918998718, + "learning_rate": 1.82632290127811e-05, + "loss": 0.5492, + "step": 34324 + }, + { + "epoch": 0.8046320551814671, + "grad_norm": 0.33002838492393494, + "learning_rate": 1.825898547071434e-05, + "loss": 0.0275, + "step": 34325 + }, + { + "epoch": 0.8046554967562721, + "grad_norm": 0.14031904935836792, + "learning_rate": 1.825474237217666e-05, + "loss": 0.0381, + "step": 34326 + }, + { + "epoch": 0.8046789383310771, + "grad_norm": 0.29538971185684204, + "learning_rate": 1.8250499717191038e-05, + "loss": 0.0394, + "step": 34327 + }, + { + "epoch": 0.8047023799058821, + "grad_norm": 0.591105043888092, + "learning_rate": 1.8246257505780507e-05, + "loss": 0.0886, + "step": 34328 + }, + { + "epoch": 0.804725821480687, + "grad_norm": 0.32423239946365356, + "learning_rate": 1.8242015737968053e-05, + "loss": 0.076, + "step": 34329 + }, + { + "epoch": 0.8047492630554921, + "grad_norm": 0.39694809913635254, + "learning_rate": 1.8237774413776752e-05, + "loss": 0.0594, + "step": 34330 + }, + { + "epoch": 0.804772704630297, + "grad_norm": 0.35896119475364685, + "learning_rate": 1.8233533533229586e-05, + "loss": 0.0618, + "step": 34331 + }, + { + "epoch": 0.8047961462051021, + "grad_norm": 0.35498130321502686, + "learning_rate": 1.8229293096349553e-05, + "loss": 0.0382, + "step": 34332 + }, + { + "epoch": 0.804819587779907, + "grad_norm": 0.560732364654541, + "learning_rate": 1.8225053103159683e-05, + "loss": 0.6084, + "step": 34333 + }, + { + "epoch": 0.8048430293547121, + "grad_norm": 0.3887743353843689, + "learning_rate": 1.822081355368298e-05, + "loss": 0.0628, + "step": 34334 + }, + { + "epoch": 0.804866470929517, + "grad_norm": 0.4646737277507782, + "learning_rate": 1.8216574447942413e-05, + "loss": 0.0674, + "step": 34335 + }, + { + "epoch": 0.8048899125043221, + "grad_norm": 0.302718847990036, + "learning_rate": 1.8212335785961032e-05, + "loss": 0.0326, + "step": 34336 + }, + { + "epoch": 0.804913354079127, + "grad_norm": 0.5092176198959351, + "learning_rate": 1.820809756776183e-05, + "loss": 0.067, + "step": 34337 + }, + { + "epoch": 0.8049367956539321, + "grad_norm": 0.1805776059627533, + "learning_rate": 1.820385979336775e-05, + "loss": 0.0331, + "step": 34338 + }, + { + "epoch": 0.804960237228737, + "grad_norm": 0.6850404739379883, + "learning_rate": 1.8199622462801856e-05, + "loss": 0.0816, + "step": 34339 + }, + { + "epoch": 0.804983678803542, + "grad_norm": 0.13519318401813507, + "learning_rate": 1.8195385576087122e-05, + "loss": 0.015, + "step": 34340 + }, + { + "epoch": 0.805007120378347, + "grad_norm": 0.7594640254974365, + "learning_rate": 1.819114913324651e-05, + "loss": 0.189, + "step": 34341 + }, + { + "epoch": 0.805030561953152, + "grad_norm": 0.207699716091156, + "learning_rate": 1.818691313430301e-05, + "loss": 0.052, + "step": 34342 + }, + { + "epoch": 0.805054003527957, + "grad_norm": 0.7450758814811707, + "learning_rate": 1.8182677579279637e-05, + "loss": 0.066, + "step": 34343 + }, + { + "epoch": 0.805077445102762, + "grad_norm": 0.15530434250831604, + "learning_rate": 1.8178442468199363e-05, + "loss": 0.018, + "step": 34344 + }, + { + "epoch": 0.805100886677567, + "grad_norm": 0.6238073110580444, + "learning_rate": 1.8174207801085152e-05, + "loss": 0.1088, + "step": 34345 + }, + { + "epoch": 0.805124328252372, + "grad_norm": 0.14673195779323578, + "learning_rate": 1.816997357796e-05, + "loss": 0.0087, + "step": 34346 + }, + { + "epoch": 0.805147769827177, + "grad_norm": 0.1205885261297226, + "learning_rate": 1.8165739798846837e-05, + "loss": 0.0214, + "step": 34347 + }, + { + "epoch": 0.805171211401982, + "grad_norm": 0.5267468690872192, + "learning_rate": 1.8161506463768698e-05, + "loss": 0.0494, + "step": 34348 + }, + { + "epoch": 0.8051946529767869, + "grad_norm": 0.5554106831550598, + "learning_rate": 1.8157273572748533e-05, + "loss": 0.1314, + "step": 34349 + }, + { + "epoch": 0.805218094551592, + "grad_norm": 0.42406517267227173, + "learning_rate": 1.8153041125809268e-05, + "loss": 0.0983, + "step": 34350 + }, + { + "epoch": 0.8052415361263969, + "grad_norm": 0.39725005626678467, + "learning_rate": 1.8148809122973944e-05, + "loss": 0.046, + "step": 34351 + }, + { + "epoch": 0.805264977701202, + "grad_norm": 0.44878968596458435, + "learning_rate": 1.8144577564265465e-05, + "loss": 0.4633, + "step": 34352 + }, + { + "epoch": 0.8052884192760069, + "grad_norm": 0.5233448147773743, + "learning_rate": 1.8140346449706813e-05, + "loss": 0.1021, + "step": 34353 + }, + { + "epoch": 0.805311860850812, + "grad_norm": 0.5354406237602234, + "learning_rate": 1.813611577932093e-05, + "loss": 0.081, + "step": 34354 + }, + { + "epoch": 0.8053353024256169, + "grad_norm": 0.2701408863067627, + "learning_rate": 1.81318855531308e-05, + "loss": 0.0458, + "step": 34355 + }, + { + "epoch": 0.805358744000422, + "grad_norm": 0.5268895626068115, + "learning_rate": 1.8127655771159357e-05, + "loss": 0.0801, + "step": 34356 + }, + { + "epoch": 0.8053821855752269, + "grad_norm": 0.2797057628631592, + "learning_rate": 1.812342643342957e-05, + "loss": 0.0453, + "step": 34357 + }, + { + "epoch": 0.8054056271500319, + "grad_norm": 0.3666597604751587, + "learning_rate": 1.811919753996436e-05, + "loss": 0.0363, + "step": 34358 + }, + { + "epoch": 0.8054290687248369, + "grad_norm": 0.24006906151771545, + "learning_rate": 1.8114969090786693e-05, + "loss": 0.0395, + "step": 34359 + }, + { + "epoch": 0.8054525102996419, + "grad_norm": 0.5106396079063416, + "learning_rate": 1.811074108591948e-05, + "loss": 0.0172, + "step": 34360 + }, + { + "epoch": 0.8054759518744469, + "grad_norm": 0.4061492383480072, + "learning_rate": 1.8106513525385717e-05, + "loss": 0.079, + "step": 34361 + }, + { + "epoch": 0.8054993934492519, + "grad_norm": 0.4821522831916809, + "learning_rate": 1.810228640920828e-05, + "loss": 0.6592, + "step": 34362 + }, + { + "epoch": 0.805522835024057, + "grad_norm": 0.2855095863342285, + "learning_rate": 1.8098059737410167e-05, + "loss": 0.0302, + "step": 34363 + }, + { + "epoch": 0.8055462765988619, + "grad_norm": 0.3539699912071228, + "learning_rate": 1.8093833510014292e-05, + "loss": 0.0595, + "step": 34364 + }, + { + "epoch": 0.805569718173667, + "grad_norm": 0.23960405588150024, + "learning_rate": 1.8089607727043577e-05, + "loss": 0.186, + "step": 34365 + }, + { + "epoch": 0.8055931597484719, + "grad_norm": 0.15331140160560608, + "learning_rate": 1.8085382388520954e-05, + "loss": 0.0159, + "step": 34366 + }, + { + "epoch": 0.805616601323277, + "grad_norm": 0.2768864929676056, + "learning_rate": 1.8081157494469315e-05, + "loss": 0.0306, + "step": 34367 + }, + { + "epoch": 0.8056400428980819, + "grad_norm": 0.3138168156147003, + "learning_rate": 1.807693304491166e-05, + "loss": 0.0527, + "step": 34368 + }, + { + "epoch": 0.8056634844728869, + "grad_norm": 0.3946826756000519, + "learning_rate": 1.807270903987085e-05, + "loss": 0.0735, + "step": 34369 + }, + { + "epoch": 0.8056869260476919, + "grad_norm": 0.3465336561203003, + "learning_rate": 1.8068485479369847e-05, + "loss": 0.0582, + "step": 34370 + }, + { + "epoch": 0.8057103676224969, + "grad_norm": 0.877050518989563, + "learning_rate": 1.806426236343153e-05, + "loss": 0.251, + "step": 34371 + }, + { + "epoch": 0.8057338091973019, + "grad_norm": 0.14424347877502441, + "learning_rate": 1.8060039692078834e-05, + "loss": 0.0193, + "step": 34372 + }, + { + "epoch": 0.8057572507721069, + "grad_norm": 0.16897240281105042, + "learning_rate": 1.8055817465334633e-05, + "loss": 0.0352, + "step": 34373 + }, + { + "epoch": 0.8057806923469119, + "grad_norm": 0.3435041308403015, + "learning_rate": 1.805159568322188e-05, + "loss": 0.0585, + "step": 34374 + }, + { + "epoch": 0.8058041339217169, + "grad_norm": 0.3539445400238037, + "learning_rate": 1.8047374345763503e-05, + "loss": 0.0428, + "step": 34375 + }, + { + "epoch": 0.8058275754965218, + "grad_norm": 0.6207783222198486, + "learning_rate": 1.8043153452982363e-05, + "loss": 0.1061, + "step": 34376 + }, + { + "epoch": 0.8058510170713269, + "grad_norm": 0.27729150652885437, + "learning_rate": 1.803893300490138e-05, + "loss": 0.0405, + "step": 34377 + }, + { + "epoch": 0.8058744586461318, + "grad_norm": 0.4861139953136444, + "learning_rate": 1.8034713001543458e-05, + "loss": 0.0785, + "step": 34378 + }, + { + "epoch": 0.8058979002209369, + "grad_norm": 0.7021895051002502, + "learning_rate": 1.8030493442931486e-05, + "loss": 0.0678, + "step": 34379 + }, + { + "epoch": 0.8059213417957418, + "grad_norm": 0.40804365277290344, + "learning_rate": 1.8026274329088332e-05, + "loss": 0.0972, + "step": 34380 + }, + { + "epoch": 0.8059447833705469, + "grad_norm": 0.2876838445663452, + "learning_rate": 1.802205566003695e-05, + "loss": 0.0336, + "step": 34381 + }, + { + "epoch": 0.8059682249453518, + "grad_norm": 0.6734774112701416, + "learning_rate": 1.80178374358002e-05, + "loss": 0.1414, + "step": 34382 + }, + { + "epoch": 0.8059916665201569, + "grad_norm": 0.5052988529205322, + "learning_rate": 1.801361965640096e-05, + "loss": 0.0568, + "step": 34383 + }, + { + "epoch": 0.8060151080949618, + "grad_norm": 0.11803604662418365, + "learning_rate": 1.8009402321862135e-05, + "loss": 0.015, + "step": 34384 + }, + { + "epoch": 0.8060385496697668, + "grad_norm": 0.2182644009590149, + "learning_rate": 1.800518543220656e-05, + "loss": 0.0329, + "step": 34385 + }, + { + "epoch": 0.8060619912445718, + "grad_norm": 0.4885023236274719, + "learning_rate": 1.8000968987457188e-05, + "loss": 0.093, + "step": 34386 + }, + { + "epoch": 0.8060854328193768, + "grad_norm": 0.670387327671051, + "learning_rate": 1.7996752987636844e-05, + "loss": 0.1654, + "step": 34387 + }, + { + "epoch": 0.8061088743941818, + "grad_norm": 0.18566541373729706, + "learning_rate": 1.799253743276844e-05, + "loss": 0.0271, + "step": 34388 + }, + { + "epoch": 0.8061323159689868, + "grad_norm": 0.22700385749340057, + "learning_rate": 1.798832232287484e-05, + "loss": 0.035, + "step": 34389 + }, + { + "epoch": 0.8061557575437918, + "grad_norm": 0.3465404808521271, + "learning_rate": 1.79841076579789e-05, + "loss": 0.038, + "step": 34390 + }, + { + "epoch": 0.8061791991185968, + "grad_norm": 0.45991379022598267, + "learning_rate": 1.79798934381035e-05, + "loss": 0.0637, + "step": 34391 + }, + { + "epoch": 0.8062026406934018, + "grad_norm": 0.4900152087211609, + "learning_rate": 1.7975679663271495e-05, + "loss": 0.0723, + "step": 34392 + }, + { + "epoch": 0.8062260822682068, + "grad_norm": 0.4554785192012787, + "learning_rate": 1.797146633350577e-05, + "loss": 0.0761, + "step": 34393 + }, + { + "epoch": 0.8062495238430117, + "grad_norm": 0.2860134541988373, + "learning_rate": 1.796725344882917e-05, + "loss": 0.0913, + "step": 34394 + }, + { + "epoch": 0.8062729654178168, + "grad_norm": 0.29732558131217957, + "learning_rate": 1.7963041009264568e-05, + "loss": 0.0456, + "step": 34395 + }, + { + "epoch": 0.8062964069926217, + "grad_norm": 0.18437731266021729, + "learning_rate": 1.795882901483481e-05, + "loss": 0.0212, + "step": 34396 + }, + { + "epoch": 0.8063198485674268, + "grad_norm": 0.4627033472061157, + "learning_rate": 1.795461746556275e-05, + "loss": 0.0626, + "step": 34397 + }, + { + "epoch": 0.8063432901422317, + "grad_norm": 0.8568124771118164, + "learning_rate": 1.795040636147123e-05, + "loss": 0.1121, + "step": 34398 + }, + { + "epoch": 0.8063667317170368, + "grad_norm": 0.30483171343803406, + "learning_rate": 1.79461957025831e-05, + "loss": 0.0398, + "step": 34399 + }, + { + "epoch": 0.8063901732918417, + "grad_norm": 0.3707248568534851, + "learning_rate": 1.7941985488921242e-05, + "loss": 0.069, + "step": 34400 + }, + { + "epoch": 0.8064136148666468, + "grad_norm": 0.46982890367507935, + "learning_rate": 1.7937775720508486e-05, + "loss": 0.0531, + "step": 34401 + }, + { + "epoch": 0.8064370564414517, + "grad_norm": 0.2055768072605133, + "learning_rate": 1.7933566397367663e-05, + "loss": 0.0343, + "step": 34402 + }, + { + "epoch": 0.8064604980162567, + "grad_norm": 0.1400367170572281, + "learning_rate": 1.7929357519521618e-05, + "loss": 0.0447, + "step": 34403 + }, + { + "epoch": 0.8064839395910617, + "grad_norm": 0.1645985245704651, + "learning_rate": 1.7925149086993187e-05, + "loss": 0.0227, + "step": 34404 + }, + { + "epoch": 0.8065073811658667, + "grad_norm": 0.19544865190982819, + "learning_rate": 1.7920941099805176e-05, + "loss": 0.0174, + "step": 34405 + }, + { + "epoch": 0.8065308227406717, + "grad_norm": 0.3453769385814667, + "learning_rate": 1.7916733557980468e-05, + "loss": 0.0711, + "step": 34406 + }, + { + "epoch": 0.8065542643154767, + "grad_norm": 0.48768889904022217, + "learning_rate": 1.791252646154188e-05, + "loss": 0.0737, + "step": 34407 + }, + { + "epoch": 0.8065777058902817, + "grad_norm": 0.483122855424881, + "learning_rate": 1.7908319810512232e-05, + "loss": 0.5307, + "step": 34408 + }, + { + "epoch": 0.8066011474650867, + "grad_norm": 0.3468589782714844, + "learning_rate": 1.790411360491434e-05, + "loss": 0.0405, + "step": 34409 + }, + { + "epoch": 0.8066245890398916, + "grad_norm": 0.5518803000450134, + "learning_rate": 1.7899907844771048e-05, + "loss": 0.0801, + "step": 34410 + }, + { + "epoch": 0.8066480306146967, + "grad_norm": 0.619484007358551, + "learning_rate": 1.789570253010513e-05, + "loss": 0.141, + "step": 34411 + }, + { + "epoch": 0.8066714721895016, + "grad_norm": 0.37114059925079346, + "learning_rate": 1.7891497660939438e-05, + "loss": 0.0458, + "step": 34412 + }, + { + "epoch": 0.8066949137643067, + "grad_norm": 0.25031816959381104, + "learning_rate": 1.788729323729682e-05, + "loss": 0.0564, + "step": 34413 + }, + { + "epoch": 0.8067183553391117, + "grad_norm": 0.607538640499115, + "learning_rate": 1.788308925920006e-05, + "loss": 0.1237, + "step": 34414 + }, + { + "epoch": 0.8067417969139167, + "grad_norm": 0.44622862339019775, + "learning_rate": 1.7878885726671956e-05, + "loss": 0.0449, + "step": 34415 + }, + { + "epoch": 0.8067652384887217, + "grad_norm": 0.28274548053741455, + "learning_rate": 1.787468263973533e-05, + "loss": 0.2128, + "step": 34416 + }, + { + "epoch": 0.8067886800635267, + "grad_norm": 0.20318683981895447, + "learning_rate": 1.7870479998412992e-05, + "loss": 0.0327, + "step": 34417 + }, + { + "epoch": 0.8068121216383317, + "grad_norm": 0.7657132744789124, + "learning_rate": 1.7866277802727706e-05, + "loss": 0.1642, + "step": 34418 + }, + { + "epoch": 0.8068355632131367, + "grad_norm": 0.1446579247713089, + "learning_rate": 1.7862076052702336e-05, + "loss": 0.016, + "step": 34419 + }, + { + "epoch": 0.8068590047879417, + "grad_norm": 0.39497891068458557, + "learning_rate": 1.785787474835965e-05, + "loss": 0.0628, + "step": 34420 + }, + { + "epoch": 0.8068824463627466, + "grad_norm": 0.2940082550048828, + "learning_rate": 1.7853673889722443e-05, + "loss": 0.0326, + "step": 34421 + }, + { + "epoch": 0.8069058879375517, + "grad_norm": 0.158392533659935, + "learning_rate": 1.7849473476813517e-05, + "loss": 0.0292, + "step": 34422 + }, + { + "epoch": 0.8069293295123566, + "grad_norm": 0.4088016450405121, + "learning_rate": 1.7845273509655624e-05, + "loss": 0.0511, + "step": 34423 + }, + { + "epoch": 0.8069527710871617, + "grad_norm": 0.22729212045669556, + "learning_rate": 1.7841073988271616e-05, + "loss": 0.0175, + "step": 34424 + }, + { + "epoch": 0.8069762126619666, + "grad_norm": 0.5419312715530396, + "learning_rate": 1.7836874912684232e-05, + "loss": 0.5212, + "step": 34425 + }, + { + "epoch": 0.8069996542367717, + "grad_norm": 0.3995082974433899, + "learning_rate": 1.7832676282916293e-05, + "loss": 0.0674, + "step": 34426 + }, + { + "epoch": 0.8070230958115766, + "grad_norm": 0.1228233352303505, + "learning_rate": 1.7828478098990566e-05, + "loss": 0.0136, + "step": 34427 + }, + { + "epoch": 0.8070465373863817, + "grad_norm": 0.37958115339279175, + "learning_rate": 1.7824280360929822e-05, + "loss": 0.062, + "step": 34428 + }, + { + "epoch": 0.8070699789611866, + "grad_norm": 0.46808522939682007, + "learning_rate": 1.782008306875684e-05, + "loss": 0.1047, + "step": 34429 + }, + { + "epoch": 0.8070934205359916, + "grad_norm": 0.09859217703342438, + "learning_rate": 1.781588622249438e-05, + "loss": 0.0215, + "step": 34430 + }, + { + "epoch": 0.8071168621107966, + "grad_norm": 0.5692812204360962, + "learning_rate": 1.7811689822165257e-05, + "loss": 0.1312, + "step": 34431 + }, + { + "epoch": 0.8071403036856016, + "grad_norm": 0.14889788627624512, + "learning_rate": 1.7807493867792223e-05, + "loss": 0.0192, + "step": 34432 + }, + { + "epoch": 0.8071637452604066, + "grad_norm": 0.3384445309638977, + "learning_rate": 1.7803298359398025e-05, + "loss": 0.0541, + "step": 34433 + }, + { + "epoch": 0.8071871868352116, + "grad_norm": 0.24821488559246063, + "learning_rate": 1.7799103297005447e-05, + "loss": 0.0402, + "step": 34434 + }, + { + "epoch": 0.8072106284100166, + "grad_norm": 0.055598799139261246, + "learning_rate": 1.7794908680637224e-05, + "loss": 0.0101, + "step": 34435 + }, + { + "epoch": 0.8072340699848216, + "grad_norm": 0.46029552817344666, + "learning_rate": 1.7790714510316165e-05, + "loss": 0.0881, + "step": 34436 + }, + { + "epoch": 0.8072575115596266, + "grad_norm": 0.131507009267807, + "learning_rate": 1.7786520786064964e-05, + "loss": 0.0219, + "step": 34437 + }, + { + "epoch": 0.8072809531344316, + "grad_norm": 0.3827849328517914, + "learning_rate": 1.778232750790645e-05, + "loss": 0.0445, + "step": 34438 + }, + { + "epoch": 0.8073043947092365, + "grad_norm": 0.5900744795799255, + "learning_rate": 1.7778134675863335e-05, + "loss": 0.0892, + "step": 34439 + }, + { + "epoch": 0.8073278362840416, + "grad_norm": 0.4085703194141388, + "learning_rate": 1.7773942289958366e-05, + "loss": 0.0647, + "step": 34440 + }, + { + "epoch": 0.8073512778588465, + "grad_norm": 0.45272940397262573, + "learning_rate": 1.7769750350214297e-05, + "loss": 0.0869, + "step": 34441 + }, + { + "epoch": 0.8073747194336516, + "grad_norm": 0.29652467370033264, + "learning_rate": 1.7765558856653886e-05, + "loss": 0.0801, + "step": 34442 + }, + { + "epoch": 0.8073981610084565, + "grad_norm": 0.5093584656715393, + "learning_rate": 1.7761367809299834e-05, + "loss": 0.5936, + "step": 34443 + }, + { + "epoch": 0.8074216025832616, + "grad_norm": 0.4143105745315552, + "learning_rate": 1.7757177208174936e-05, + "loss": 0.5813, + "step": 34444 + }, + { + "epoch": 0.8074450441580665, + "grad_norm": 0.5494718551635742, + "learning_rate": 1.775298705330192e-05, + "loss": 0.0914, + "step": 34445 + }, + { + "epoch": 0.8074684857328716, + "grad_norm": 0.12742771208286285, + "learning_rate": 1.7748797344703494e-05, + "loss": 0.0237, + "step": 34446 + }, + { + "epoch": 0.8074919273076765, + "grad_norm": 0.23157386481761932, + "learning_rate": 1.7744608082402382e-05, + "loss": 0.0173, + "step": 34447 + }, + { + "epoch": 0.8075153688824815, + "grad_norm": 0.3965262174606323, + "learning_rate": 1.774041926642137e-05, + "loss": 0.0376, + "step": 34448 + }, + { + "epoch": 0.8075388104572865, + "grad_norm": 0.27790218591690063, + "learning_rate": 1.7736230896783156e-05, + "loss": 0.3791, + "step": 34449 + }, + { + "epoch": 0.8075622520320915, + "grad_norm": 0.44885700941085815, + "learning_rate": 1.7732042973510434e-05, + "loss": 0.1044, + "step": 34450 + }, + { + "epoch": 0.8075856936068965, + "grad_norm": 0.32469791173934937, + "learning_rate": 1.772785549662599e-05, + "loss": 0.071, + "step": 34451 + }, + { + "epoch": 0.8076091351817015, + "grad_norm": 0.5945317149162292, + "learning_rate": 1.772366846615252e-05, + "loss": 0.0986, + "step": 34452 + }, + { + "epoch": 0.8076325767565065, + "grad_norm": 0.46799373626708984, + "learning_rate": 1.7719481882112732e-05, + "loss": 0.0408, + "step": 34453 + }, + { + "epoch": 0.8076560183313115, + "grad_norm": 0.3080897331237793, + "learning_rate": 1.771529574452935e-05, + "loss": 0.0316, + "step": 34454 + }, + { + "epoch": 0.8076794599061164, + "grad_norm": 0.4903825521469116, + "learning_rate": 1.7711110053425083e-05, + "loss": 0.0963, + "step": 34455 + }, + { + "epoch": 0.8077029014809215, + "grad_norm": 0.30028530955314636, + "learning_rate": 1.7706924808822634e-05, + "loss": 0.0316, + "step": 34456 + }, + { + "epoch": 0.8077263430557264, + "grad_norm": 0.16057617962360382, + "learning_rate": 1.7702740010744733e-05, + "loss": 0.0178, + "step": 34457 + }, + { + "epoch": 0.8077497846305315, + "grad_norm": 0.38229134678840637, + "learning_rate": 1.7698555659214088e-05, + "loss": 0.0511, + "step": 34458 + }, + { + "epoch": 0.8077732262053364, + "grad_norm": 0.3890351951122284, + "learning_rate": 1.7694371754253368e-05, + "loss": 0.0917, + "step": 34459 + }, + { + "epoch": 0.8077966677801415, + "grad_norm": 0.5027632117271423, + "learning_rate": 1.7690188295885335e-05, + "loss": 0.1062, + "step": 34460 + }, + { + "epoch": 0.8078201093549464, + "grad_norm": 0.23147684335708618, + "learning_rate": 1.7686005284132645e-05, + "loss": 0.0379, + "step": 34461 + }, + { + "epoch": 0.8078435509297515, + "grad_norm": 0.6101475954055786, + "learning_rate": 1.7681822719017982e-05, + "loss": 0.1398, + "step": 34462 + }, + { + "epoch": 0.8078669925045564, + "grad_norm": 0.3969908356666565, + "learning_rate": 1.7677640600564084e-05, + "loss": 0.044, + "step": 34463 + }, + { + "epoch": 0.8078904340793615, + "grad_norm": 0.5589504241943359, + "learning_rate": 1.767345892879363e-05, + "loss": 0.0884, + "step": 34464 + }, + { + "epoch": 0.8079138756541665, + "grad_norm": 0.4292849600315094, + "learning_rate": 1.7669277703729302e-05, + "loss": 0.0653, + "step": 34465 + }, + { + "epoch": 0.8079373172289714, + "grad_norm": 0.28990811109542847, + "learning_rate": 1.766509692539379e-05, + "loss": 0.0451, + "step": 34466 + }, + { + "epoch": 0.8079607588037765, + "grad_norm": 0.4260377585887909, + "learning_rate": 1.7660916593809764e-05, + "loss": 0.1037, + "step": 34467 + }, + { + "epoch": 0.8079842003785814, + "grad_norm": 0.20284609496593475, + "learning_rate": 1.76567367089999e-05, + "loss": 0.0271, + "step": 34468 + }, + { + "epoch": 0.8080076419533865, + "grad_norm": 0.16051016747951508, + "learning_rate": 1.7652557270986934e-05, + "loss": 0.02, + "step": 34469 + }, + { + "epoch": 0.8080310835281914, + "grad_norm": 0.25463518500328064, + "learning_rate": 1.7648378279793497e-05, + "loss": 0.0342, + "step": 34470 + }, + { + "epoch": 0.8080545251029965, + "grad_norm": 0.5145891904830933, + "learning_rate": 1.764419973544226e-05, + "loss": 0.0796, + "step": 34471 + }, + { + "epoch": 0.8080779666778014, + "grad_norm": 0.6965653300285339, + "learning_rate": 1.7640021637955917e-05, + "loss": 0.1398, + "step": 34472 + }, + { + "epoch": 0.8081014082526065, + "grad_norm": 0.09295236319303513, + "learning_rate": 1.7635843987357147e-05, + "loss": 0.0088, + "step": 34473 + }, + { + "epoch": 0.8081248498274114, + "grad_norm": 0.3382127583026886, + "learning_rate": 1.7631666783668597e-05, + "loss": 0.0252, + "step": 34474 + }, + { + "epoch": 0.8081482914022164, + "grad_norm": 0.4318200647830963, + "learning_rate": 1.762749002691292e-05, + "loss": 0.0377, + "step": 34475 + }, + { + "epoch": 0.8081717329770214, + "grad_norm": 0.42945289611816406, + "learning_rate": 1.762331371711282e-05, + "loss": 0.066, + "step": 34476 + }, + { + "epoch": 0.8081951745518264, + "grad_norm": 0.2837563455104828, + "learning_rate": 1.7619137854290935e-05, + "loss": 0.0456, + "step": 34477 + }, + { + "epoch": 0.8082186161266314, + "grad_norm": 0.1310713291168213, + "learning_rate": 1.7614962438469928e-05, + "loss": 0.0226, + "step": 34478 + }, + { + "epoch": 0.8082420577014364, + "grad_norm": 0.31404227018356323, + "learning_rate": 1.7610787469672453e-05, + "loss": 0.0398, + "step": 34479 + }, + { + "epoch": 0.8082654992762414, + "grad_norm": 1.2513911724090576, + "learning_rate": 1.7606612947921152e-05, + "loss": 0.0676, + "step": 34480 + }, + { + "epoch": 0.8082889408510464, + "grad_norm": 0.1966153085231781, + "learning_rate": 1.760243887323867e-05, + "loss": 0.0221, + "step": 34481 + }, + { + "epoch": 0.8083123824258514, + "grad_norm": 0.4037853479385376, + "learning_rate": 1.7598265245647694e-05, + "loss": 0.1022, + "step": 34482 + }, + { + "epoch": 0.8083358240006564, + "grad_norm": 0.30348342657089233, + "learning_rate": 1.759409206517082e-05, + "loss": 0.0538, + "step": 34483 + }, + { + "epoch": 0.8083592655754613, + "grad_norm": 0.2467043101787567, + "learning_rate": 1.7589919331830752e-05, + "loss": 0.0422, + "step": 34484 + }, + { + "epoch": 0.8083827071502664, + "grad_norm": 0.4570336639881134, + "learning_rate": 1.758574704565009e-05, + "loss": 0.0938, + "step": 34485 + }, + { + "epoch": 0.8084061487250713, + "grad_norm": 0.36444514989852905, + "learning_rate": 1.758157520665149e-05, + "loss": 0.0869, + "step": 34486 + }, + { + "epoch": 0.8084295902998764, + "grad_norm": 0.39291253685951233, + "learning_rate": 1.7577403814857574e-05, + "loss": 0.0524, + "step": 34487 + }, + { + "epoch": 0.8084530318746813, + "grad_norm": 0.49111106991767883, + "learning_rate": 1.7573232870290958e-05, + "loss": 0.1109, + "step": 34488 + }, + { + "epoch": 0.8084764734494864, + "grad_norm": 0.5483818054199219, + "learning_rate": 1.756906237297432e-05, + "loss": 0.0788, + "step": 34489 + }, + { + "epoch": 0.8084999150242913, + "grad_norm": 0.32952433824539185, + "learning_rate": 1.756489232293027e-05, + "loss": 0.3801, + "step": 34490 + }, + { + "epoch": 0.8085233565990964, + "grad_norm": 0.5470524430274963, + "learning_rate": 1.7560722720181445e-05, + "loss": 0.6395, + "step": 34491 + }, + { + "epoch": 0.8085467981739013, + "grad_norm": 0.35215699672698975, + "learning_rate": 1.7556553564750444e-05, + "loss": 0.0891, + "step": 34492 + }, + { + "epoch": 0.8085702397487063, + "grad_norm": 0.284948468208313, + "learning_rate": 1.7552384856659898e-05, + "loss": 0.0569, + "step": 34493 + }, + { + "epoch": 0.8085936813235113, + "grad_norm": 0.6531988978385925, + "learning_rate": 1.7548216595932408e-05, + "loss": 0.0828, + "step": 34494 + }, + { + "epoch": 0.8086171228983163, + "grad_norm": 0.4653722643852234, + "learning_rate": 1.7544048782590638e-05, + "loss": 0.0532, + "step": 34495 + }, + { + "epoch": 0.8086405644731213, + "grad_norm": 0.3125365972518921, + "learning_rate": 1.7539881416657156e-05, + "loss": 0.0476, + "step": 34496 + }, + { + "epoch": 0.8086640060479263, + "grad_norm": 0.22608676552772522, + "learning_rate": 1.7535714498154622e-05, + "loss": 0.053, + "step": 34497 + }, + { + "epoch": 0.8086874476227313, + "grad_norm": 0.12710285186767578, + "learning_rate": 1.7531548027105614e-05, + "loss": 0.0325, + "step": 34498 + }, + { + "epoch": 0.8087108891975363, + "grad_norm": 0.38010522723197937, + "learning_rate": 1.7527382003532756e-05, + "loss": 0.0534, + "step": 34499 + }, + { + "epoch": 0.8087343307723412, + "grad_norm": 0.45916712284088135, + "learning_rate": 1.752321642745861e-05, + "loss": 0.0479, + "step": 34500 + }, + { + "epoch": 0.8087577723471463, + "grad_norm": 0.2573206126689911, + "learning_rate": 1.751905129890583e-05, + "loss": 0.0427, + "step": 34501 + }, + { + "epoch": 0.8087812139219512, + "grad_norm": 0.4252554178237915, + "learning_rate": 1.751488661789701e-05, + "loss": 0.074, + "step": 34502 + }, + { + "epoch": 0.8088046554967563, + "grad_norm": 0.5073342323303223, + "learning_rate": 1.7510722384454724e-05, + "loss": 0.0872, + "step": 34503 + }, + { + "epoch": 0.8088280970715612, + "grad_norm": 0.24515779316425323, + "learning_rate": 1.7506558598601573e-05, + "loss": 0.0346, + "step": 34504 + }, + { + "epoch": 0.8088515386463663, + "grad_norm": 0.1948549896478653, + "learning_rate": 1.750239526036016e-05, + "loss": 0.0207, + "step": 34505 + }, + { + "epoch": 0.8088749802211712, + "grad_norm": 0.42195841670036316, + "learning_rate": 1.7498232369753055e-05, + "loss": 0.0451, + "step": 34506 + }, + { + "epoch": 0.8088984217959763, + "grad_norm": 0.6242957711219788, + "learning_rate": 1.749406992680287e-05, + "loss": 0.2118, + "step": 34507 + }, + { + "epoch": 0.8089218633707812, + "grad_norm": 0.4955415725708008, + "learning_rate": 1.7489907931532167e-05, + "loss": 0.0615, + "step": 34508 + }, + { + "epoch": 0.8089453049455863, + "grad_norm": 0.4760837256908417, + "learning_rate": 1.7485746383963562e-05, + "loss": 0.0862, + "step": 34509 + }, + { + "epoch": 0.8089687465203912, + "grad_norm": 0.5354580879211426, + "learning_rate": 1.748158528411963e-05, + "loss": 0.0952, + "step": 34510 + }, + { + "epoch": 0.8089921880951962, + "grad_norm": 0.752704918384552, + "learning_rate": 1.747742463202292e-05, + "loss": 0.2033, + "step": 34511 + }, + { + "epoch": 0.8090156296700012, + "grad_norm": 0.42207667231559753, + "learning_rate": 1.7473264427696035e-05, + "loss": 0.0666, + "step": 34512 + }, + { + "epoch": 0.8090390712448062, + "grad_norm": 0.23166103661060333, + "learning_rate": 1.7469104671161507e-05, + "loss": 0.0431, + "step": 34513 + }, + { + "epoch": 0.8090625128196112, + "grad_norm": 0.14717727899551392, + "learning_rate": 1.7464945362441965e-05, + "loss": 0.0325, + "step": 34514 + }, + { + "epoch": 0.8090859543944162, + "grad_norm": 0.3541710078716278, + "learning_rate": 1.7460786501559957e-05, + "loss": 0.0459, + "step": 34515 + }, + { + "epoch": 0.8091093959692212, + "grad_norm": 0.2746870517730713, + "learning_rate": 1.745662808853803e-05, + "loss": 0.0425, + "step": 34516 + }, + { + "epoch": 0.8091328375440262, + "grad_norm": 0.607995867729187, + "learning_rate": 1.745247012339877e-05, + "loss": 0.1389, + "step": 34517 + }, + { + "epoch": 0.8091562791188313, + "grad_norm": 0.15653015673160553, + "learning_rate": 1.744831260616472e-05, + "loss": 0.0264, + "step": 34518 + }, + { + "epoch": 0.8091797206936362, + "grad_norm": 0.3937303423881531, + "learning_rate": 1.7444155536858432e-05, + "loss": 0.0666, + "step": 34519 + }, + { + "epoch": 0.8092031622684412, + "grad_norm": 0.4664512276649475, + "learning_rate": 1.7439998915502466e-05, + "loss": 0.1372, + "step": 34520 + }, + { + "epoch": 0.8092266038432462, + "grad_norm": 0.12787841260433197, + "learning_rate": 1.7435842742119414e-05, + "loss": 0.0247, + "step": 34521 + }, + { + "epoch": 0.8092500454180512, + "grad_norm": 0.24411779642105103, + "learning_rate": 1.74316870167318e-05, + "loss": 0.0191, + "step": 34522 + }, + { + "epoch": 0.8092734869928562, + "grad_norm": 0.41490453481674194, + "learning_rate": 1.7427531739362178e-05, + "loss": 0.0529, + "step": 34523 + }, + { + "epoch": 0.8092969285676612, + "grad_norm": 0.11244486272335052, + "learning_rate": 1.7423376910033097e-05, + "loss": 0.0096, + "step": 34524 + }, + { + "epoch": 0.8093203701424662, + "grad_norm": 0.3424408435821533, + "learning_rate": 1.7419222528767088e-05, + "loss": 0.0576, + "step": 34525 + }, + { + "epoch": 0.8093438117172712, + "grad_norm": 0.5017952919006348, + "learning_rate": 1.741506859558667e-05, + "loss": 0.0662, + "step": 34526 + }, + { + "epoch": 0.8093672532920761, + "grad_norm": 0.46552079916000366, + "learning_rate": 1.7410915110514446e-05, + "loss": 0.0674, + "step": 34527 + }, + { + "epoch": 0.8093906948668812, + "grad_norm": 0.1902908831834793, + "learning_rate": 1.7406762073572913e-05, + "loss": 0.0399, + "step": 34528 + }, + { + "epoch": 0.8094141364416861, + "grad_norm": 0.413686603307724, + "learning_rate": 1.7402609484784616e-05, + "loss": 0.0719, + "step": 34529 + }, + { + "epoch": 0.8094375780164912, + "grad_norm": 0.5392215251922607, + "learning_rate": 1.739845734417207e-05, + "loss": 0.6242, + "step": 34530 + }, + { + "epoch": 0.8094610195912961, + "grad_norm": 0.16402703523635864, + "learning_rate": 1.7394305651757803e-05, + "loss": 0.0173, + "step": 34531 + }, + { + "epoch": 0.8094844611661012, + "grad_norm": 0.11382540315389633, + "learning_rate": 1.7390154407564386e-05, + "loss": 0.0123, + "step": 34532 + }, + { + "epoch": 0.8095079027409061, + "grad_norm": 0.38811826705932617, + "learning_rate": 1.7386003611614276e-05, + "loss": 0.0414, + "step": 34533 + }, + { + "epoch": 0.8095313443157112, + "grad_norm": 0.30361440777778625, + "learning_rate": 1.738185326393007e-05, + "loss": 0.0442, + "step": 34534 + }, + { + "epoch": 0.8095547858905161, + "grad_norm": 0.3800484538078308, + "learning_rate": 1.737770336453424e-05, + "loss": 0.1128, + "step": 34535 + }, + { + "epoch": 0.8095782274653212, + "grad_norm": 0.4272419512271881, + "learning_rate": 1.7373553913449326e-05, + "loss": 0.0479, + "step": 34536 + }, + { + "epoch": 0.8096016690401261, + "grad_norm": 0.3940165638923645, + "learning_rate": 1.7369404910697818e-05, + "loss": 0.0601, + "step": 34537 + }, + { + "epoch": 0.8096251106149311, + "grad_norm": 0.3485601544380188, + "learning_rate": 1.7365256356302228e-05, + "loss": 0.0661, + "step": 34538 + }, + { + "epoch": 0.8096485521897361, + "grad_norm": 0.22506332397460938, + "learning_rate": 1.7361108250285108e-05, + "loss": 0.0277, + "step": 34539 + }, + { + "epoch": 0.8096719937645411, + "grad_norm": 0.07990321516990662, + "learning_rate": 1.7356960592668923e-05, + "loss": 0.0088, + "step": 34540 + }, + { + "epoch": 0.8096954353393461, + "grad_norm": 0.5016797780990601, + "learning_rate": 1.7352813383476206e-05, + "loss": 0.0643, + "step": 34541 + }, + { + "epoch": 0.8097188769141511, + "grad_norm": 0.43787550926208496, + "learning_rate": 1.7348666622729447e-05, + "loss": 0.0485, + "step": 34542 + }, + { + "epoch": 0.8097423184889561, + "grad_norm": 0.268320232629776, + "learning_rate": 1.734452031045114e-05, + "loss": 0.0337, + "step": 34543 + }, + { + "epoch": 0.8097657600637611, + "grad_norm": 0.3849506676197052, + "learning_rate": 1.7340374446663776e-05, + "loss": 0.06, + "step": 34544 + }, + { + "epoch": 0.809789201638566, + "grad_norm": 0.4010971486568451, + "learning_rate": 1.733622903138986e-05, + "loss": 0.0516, + "step": 34545 + }, + { + "epoch": 0.8098126432133711, + "grad_norm": 0.33742454648017883, + "learning_rate": 1.733208406465191e-05, + "loss": 0.0254, + "step": 34546 + }, + { + "epoch": 0.809836084788176, + "grad_norm": 0.5944324135780334, + "learning_rate": 1.7327939546472393e-05, + "loss": 0.0442, + "step": 34547 + }, + { + "epoch": 0.8098595263629811, + "grad_norm": 0.15139229595661163, + "learning_rate": 1.7323795476873818e-05, + "loss": 0.0092, + "step": 34548 + }, + { + "epoch": 0.809882967937786, + "grad_norm": 0.12909170985221863, + "learning_rate": 1.7319651855878636e-05, + "loss": 0.0175, + "step": 34549 + }, + { + "epoch": 0.8099064095125911, + "grad_norm": 0.3831847608089447, + "learning_rate": 1.7315508683509353e-05, + "loss": 0.0457, + "step": 34550 + }, + { + "epoch": 0.809929851087396, + "grad_norm": 0.22851626574993134, + "learning_rate": 1.7311365959788417e-05, + "loss": 0.0385, + "step": 34551 + }, + { + "epoch": 0.8099532926622011, + "grad_norm": 0.5554648637771606, + "learning_rate": 1.730722368473837e-05, + "loss": 0.0768, + "step": 34552 + }, + { + "epoch": 0.809976734237006, + "grad_norm": 0.6009315252304077, + "learning_rate": 1.7303081858381642e-05, + "loss": 0.1431, + "step": 34553 + }, + { + "epoch": 0.810000175811811, + "grad_norm": 0.09769581258296967, + "learning_rate": 1.7298940480740723e-05, + "loss": 0.0186, + "step": 34554 + }, + { + "epoch": 0.810023617386616, + "grad_norm": 0.7607023119926453, + "learning_rate": 1.729479955183807e-05, + "loss": 0.5487, + "step": 34555 + }, + { + "epoch": 0.810047058961421, + "grad_norm": 0.09453792870044708, + "learning_rate": 1.7290659071696146e-05, + "loss": 0.0073, + "step": 34556 + }, + { + "epoch": 0.810070500536226, + "grad_norm": 0.44081178307533264, + "learning_rate": 1.7286519040337447e-05, + "loss": 0.1122, + "step": 34557 + }, + { + "epoch": 0.810093942111031, + "grad_norm": 0.14587241411209106, + "learning_rate": 1.7282379457784403e-05, + "loss": 0.0128, + "step": 34558 + }, + { + "epoch": 0.810117383685836, + "grad_norm": 0.5245948433876038, + "learning_rate": 1.7278240324059514e-05, + "loss": 0.0434, + "step": 34559 + }, + { + "epoch": 0.810140825260641, + "grad_norm": 0.6133720874786377, + "learning_rate": 1.7274101639185225e-05, + "loss": 0.1151, + "step": 34560 + }, + { + "epoch": 0.810164266835446, + "grad_norm": 0.4638957679271698, + "learning_rate": 1.7269963403183974e-05, + "loss": 0.1006, + "step": 34561 + }, + { + "epoch": 0.810187708410251, + "grad_norm": 0.12852740287780762, + "learning_rate": 1.7265825616078245e-05, + "loss": 0.0196, + "step": 34562 + }, + { + "epoch": 0.810211149985056, + "grad_norm": 0.344451904296875, + "learning_rate": 1.7261688277890465e-05, + "loss": 0.0235, + "step": 34563 + }, + { + "epoch": 0.810234591559861, + "grad_norm": 0.11943665891885757, + "learning_rate": 1.7257551388643055e-05, + "loss": 0.0202, + "step": 34564 + }, + { + "epoch": 0.8102580331346659, + "grad_norm": 0.5255306363105774, + "learning_rate": 1.7253414948358537e-05, + "loss": 0.1058, + "step": 34565 + }, + { + "epoch": 0.810281474709471, + "grad_norm": 0.3289507031440735, + "learning_rate": 1.7249278957059313e-05, + "loss": 0.0416, + "step": 34566 + }, + { + "epoch": 0.8103049162842759, + "grad_norm": 0.10829752683639526, + "learning_rate": 1.724514341476783e-05, + "loss": 0.0164, + "step": 34567 + }, + { + "epoch": 0.810328357859081, + "grad_norm": 0.49698320031166077, + "learning_rate": 1.7241008321506525e-05, + "loss": 0.5067, + "step": 34568 + }, + { + "epoch": 0.810351799433886, + "grad_norm": 0.25286105275154114, + "learning_rate": 1.7236873677297804e-05, + "loss": 0.0423, + "step": 34569 + }, + { + "epoch": 0.810375241008691, + "grad_norm": 0.35491493344306946, + "learning_rate": 1.7232739482164163e-05, + "loss": 0.0865, + "step": 34570 + }, + { + "epoch": 0.810398682583496, + "grad_norm": 0.14003144204616547, + "learning_rate": 1.722860573612798e-05, + "loss": 0.0275, + "step": 34571 + }, + { + "epoch": 0.810422124158301, + "grad_norm": 0.34094077348709106, + "learning_rate": 1.7224472439211738e-05, + "loss": 0.041, + "step": 34572 + }, + { + "epoch": 0.810445565733106, + "grad_norm": 0.4280880391597748, + "learning_rate": 1.722033959143783e-05, + "loss": 0.0494, + "step": 34573 + }, + { + "epoch": 0.8104690073079109, + "grad_norm": 0.11943686008453369, + "learning_rate": 1.7216207192828682e-05, + "loss": 0.0111, + "step": 34574 + }, + { + "epoch": 0.810492448882716, + "grad_norm": 0.19123108685016632, + "learning_rate": 1.7212075243406734e-05, + "loss": 0.0421, + "step": 34575 + }, + { + "epoch": 0.8105158904575209, + "grad_norm": 0.21029695868492126, + "learning_rate": 1.7207943743194364e-05, + "loss": 0.0227, + "step": 34576 + }, + { + "epoch": 0.810539332032326, + "grad_norm": 0.321627676486969, + "learning_rate": 1.720381269221404e-05, + "loss": 0.0315, + "step": 34577 + }, + { + "epoch": 0.8105627736071309, + "grad_norm": 0.6476531028747559, + "learning_rate": 1.7199682090488155e-05, + "loss": 0.5626, + "step": 34578 + }, + { + "epoch": 0.810586215181936, + "grad_norm": 0.08539112657308578, + "learning_rate": 1.7195551938039134e-05, + "loss": 0.0098, + "step": 34579 + }, + { + "epoch": 0.8106096567567409, + "grad_norm": 0.3811820149421692, + "learning_rate": 1.7191422234889365e-05, + "loss": 0.0623, + "step": 34580 + }, + { + "epoch": 0.810633098331546, + "grad_norm": 0.22209005057811737, + "learning_rate": 1.7187292981061253e-05, + "loss": 0.0487, + "step": 34581 + }, + { + "epoch": 0.8106565399063509, + "grad_norm": 0.634394109249115, + "learning_rate": 1.7183164176577238e-05, + "loss": 0.155, + "step": 34582 + }, + { + "epoch": 0.8106799814811559, + "grad_norm": 0.35506683588027954, + "learning_rate": 1.7179035821459676e-05, + "loss": 0.0795, + "step": 34583 + }, + { + "epoch": 0.8107034230559609, + "grad_norm": 0.3375168740749359, + "learning_rate": 1.717490791573102e-05, + "loss": 0.058, + "step": 34584 + }, + { + "epoch": 0.8107268646307659, + "grad_norm": 0.45161303877830505, + "learning_rate": 1.717078045941364e-05, + "loss": 0.0967, + "step": 34585 + }, + { + "epoch": 0.8107503062055709, + "grad_norm": 0.4793623387813568, + "learning_rate": 1.7166653452529935e-05, + "loss": 0.0919, + "step": 34586 + }, + { + "epoch": 0.8107737477803759, + "grad_norm": 0.21525022387504578, + "learning_rate": 1.7162526895102306e-05, + "loss": 0.0323, + "step": 34587 + }, + { + "epoch": 0.8107971893551809, + "grad_norm": 0.38148725032806396, + "learning_rate": 1.7158400787153127e-05, + "loss": 0.0585, + "step": 34588 + }, + { + "epoch": 0.8108206309299859, + "grad_norm": 0.5017943382263184, + "learning_rate": 1.7154275128704767e-05, + "loss": 0.126, + "step": 34589 + }, + { + "epoch": 0.8108440725047908, + "grad_norm": 0.8457287549972534, + "learning_rate": 1.7150149919779668e-05, + "loss": 0.1412, + "step": 34590 + }, + { + "epoch": 0.8108675140795959, + "grad_norm": 0.4825197756290436, + "learning_rate": 1.7146025160400193e-05, + "loss": 0.0717, + "step": 34591 + }, + { + "epoch": 0.8108909556544008, + "grad_norm": 0.5394377708435059, + "learning_rate": 1.71419008505887e-05, + "loss": 0.1015, + "step": 34592 + }, + { + "epoch": 0.8109143972292059, + "grad_norm": 0.33713236451148987, + "learning_rate": 1.7137776990367573e-05, + "loss": 0.0715, + "step": 34593 + }, + { + "epoch": 0.8109378388040108, + "grad_norm": 0.4215202033519745, + "learning_rate": 1.7133653579759212e-05, + "loss": 0.0441, + "step": 34594 + }, + { + "epoch": 0.8109612803788159, + "grad_norm": 0.3830995559692383, + "learning_rate": 1.7129530618785982e-05, + "loss": 0.0804, + "step": 34595 + }, + { + "epoch": 0.8109847219536208, + "grad_norm": 0.6258466243743896, + "learning_rate": 1.712540810747022e-05, + "loss": 0.1239, + "step": 34596 + }, + { + "epoch": 0.8110081635284259, + "grad_norm": 0.11021937429904938, + "learning_rate": 1.7121286045834352e-05, + "loss": 0.0192, + "step": 34597 + }, + { + "epoch": 0.8110316051032308, + "grad_norm": 0.3822855055332184, + "learning_rate": 1.7117164433900712e-05, + "loss": 0.0586, + "step": 34598 + }, + { + "epoch": 0.8110550466780359, + "grad_norm": 0.25823643803596497, + "learning_rate": 1.7113043271691677e-05, + "loss": 0.0305, + "step": 34599 + }, + { + "epoch": 0.8110784882528408, + "grad_norm": 0.06331108510494232, + "learning_rate": 1.7108922559229588e-05, + "loss": 0.0062, + "step": 34600 + }, + { + "epoch": 0.8111019298276458, + "grad_norm": 0.48783648014068604, + "learning_rate": 1.7104802296536827e-05, + "loss": 0.0676, + "step": 34601 + }, + { + "epoch": 0.8111253714024508, + "grad_norm": 0.3023151755332947, + "learning_rate": 1.710068248363571e-05, + "loss": 0.0328, + "step": 34602 + }, + { + "epoch": 0.8111488129772558, + "grad_norm": 0.46651479601860046, + "learning_rate": 1.709656312054865e-05, + "loss": 0.0547, + "step": 34603 + }, + { + "epoch": 0.8111722545520608, + "grad_norm": 0.32983869314193726, + "learning_rate": 1.7092444207297964e-05, + "loss": 0.0615, + "step": 34604 + }, + { + "epoch": 0.8111956961268658, + "grad_norm": 0.12642645835876465, + "learning_rate": 1.7088325743905987e-05, + "loss": 0.0169, + "step": 34605 + }, + { + "epoch": 0.8112191377016708, + "grad_norm": 0.18896804749965668, + "learning_rate": 1.7084207730395095e-05, + "loss": 0.0425, + "step": 34606 + }, + { + "epoch": 0.8112425792764758, + "grad_norm": 0.5898551344871521, + "learning_rate": 1.7080090166787644e-05, + "loss": 0.1174, + "step": 34607 + }, + { + "epoch": 0.8112660208512807, + "grad_norm": 0.30272331833839417, + "learning_rate": 1.7075973053105944e-05, + "loss": 0.0538, + "step": 34608 + }, + { + "epoch": 0.8112894624260858, + "grad_norm": 0.48645034432411194, + "learning_rate": 1.707185638937232e-05, + "loss": 0.062, + "step": 34609 + }, + { + "epoch": 0.8113129040008907, + "grad_norm": 0.927429735660553, + "learning_rate": 1.706774017560917e-05, + "loss": 0.0968, + "step": 34610 + }, + { + "epoch": 0.8113363455756958, + "grad_norm": 0.4741672873497009, + "learning_rate": 1.7063624411838786e-05, + "loss": 0.1114, + "step": 34611 + }, + { + "epoch": 0.8113597871505007, + "grad_norm": 0.39137497544288635, + "learning_rate": 1.7059509098083504e-05, + "loss": 0.0807, + "step": 34612 + }, + { + "epoch": 0.8113832287253058, + "grad_norm": 0.09548425674438477, + "learning_rate": 1.705539423436566e-05, + "loss": 0.0131, + "step": 34613 + }, + { + "epoch": 0.8114066703001107, + "grad_norm": 0.4734938144683838, + "learning_rate": 1.7051279820707566e-05, + "loss": 0.0905, + "step": 34614 + }, + { + "epoch": 0.8114301118749158, + "grad_norm": 0.17728744447231293, + "learning_rate": 1.7047165857131575e-05, + "loss": 0.03, + "step": 34615 + }, + { + "epoch": 0.8114535534497207, + "grad_norm": 0.19477146863937378, + "learning_rate": 1.7043052343659994e-05, + "loss": 0.0451, + "step": 34616 + }, + { + "epoch": 0.8114769950245257, + "grad_norm": 0.38126733899116516, + "learning_rate": 1.7038939280315112e-05, + "loss": 0.0585, + "step": 34617 + }, + { + "epoch": 0.8115004365993307, + "grad_norm": 0.1284555196762085, + "learning_rate": 1.703482666711932e-05, + "loss": 0.031, + "step": 34618 + }, + { + "epoch": 0.8115238781741357, + "grad_norm": 0.40397343039512634, + "learning_rate": 1.703071450409488e-05, + "loss": 0.053, + "step": 34619 + }, + { + "epoch": 0.8115473197489408, + "grad_norm": 0.10734786838293076, + "learning_rate": 1.702660279126411e-05, + "loss": 0.0155, + "step": 34620 + }, + { + "epoch": 0.8115707613237457, + "grad_norm": 0.34661227464675903, + "learning_rate": 1.702249152864931e-05, + "loss": 0.0702, + "step": 34621 + }, + { + "epoch": 0.8115942028985508, + "grad_norm": 0.2656266391277313, + "learning_rate": 1.701838071627283e-05, + "loss": 0.0388, + "step": 34622 + }, + { + "epoch": 0.8116176444733557, + "grad_norm": 0.18486924469470978, + "learning_rate": 1.7014270354156946e-05, + "loss": 0.0125, + "step": 34623 + }, + { + "epoch": 0.8116410860481608, + "grad_norm": 0.22214196622371674, + "learning_rate": 1.7010160442323975e-05, + "loss": 0.0297, + "step": 34624 + }, + { + "epoch": 0.8116645276229657, + "grad_norm": 0.2306324988603592, + "learning_rate": 1.7006050980796197e-05, + "loss": 0.0394, + "step": 34625 + }, + { + "epoch": 0.8116879691977708, + "grad_norm": 0.3506318926811218, + "learning_rate": 1.7001941969595914e-05, + "loss": 0.0625, + "step": 34626 + }, + { + "epoch": 0.8117114107725757, + "grad_norm": 0.3431820869445801, + "learning_rate": 1.699783340874541e-05, + "loss": 0.0544, + "step": 34627 + }, + { + "epoch": 0.8117348523473807, + "grad_norm": 0.524743914604187, + "learning_rate": 1.6993725298267028e-05, + "loss": 0.5454, + "step": 34628 + }, + { + "epoch": 0.8117582939221857, + "grad_norm": 0.31569236516952515, + "learning_rate": 1.6989617638182985e-05, + "loss": 0.0511, + "step": 34629 + }, + { + "epoch": 0.8117817354969907, + "grad_norm": 0.5306389331817627, + "learning_rate": 1.698551042851565e-05, + "loss": 0.111, + "step": 34630 + }, + { + "epoch": 0.8118051770717957, + "grad_norm": 0.44277358055114746, + "learning_rate": 1.6981403669287266e-05, + "loss": 0.069, + "step": 34631 + }, + { + "epoch": 0.8118286186466007, + "grad_norm": 0.41722145676612854, + "learning_rate": 1.6977297360520115e-05, + "loss": 0.05, + "step": 34632 + }, + { + "epoch": 0.8118520602214057, + "grad_norm": 0.18690572679042816, + "learning_rate": 1.6973191502236484e-05, + "loss": 0.0321, + "step": 34633 + }, + { + "epoch": 0.8118755017962107, + "grad_norm": 0.6109552979469299, + "learning_rate": 1.696908609445863e-05, + "loss": 0.0893, + "step": 34634 + }, + { + "epoch": 0.8118989433710156, + "grad_norm": 0.48155760765075684, + "learning_rate": 1.6964981137208867e-05, + "loss": 0.536, + "step": 34635 + }, + { + "epoch": 0.8119223849458207, + "grad_norm": 0.16994966566562653, + "learning_rate": 1.696087663050946e-05, + "loss": 0.0224, + "step": 34636 + }, + { + "epoch": 0.8119458265206256, + "grad_norm": 0.4471379816532135, + "learning_rate": 1.6956772574382674e-05, + "loss": 0.0682, + "step": 34637 + }, + { + "epoch": 0.8119692680954307, + "grad_norm": 0.743590772151947, + "learning_rate": 1.695266896885076e-05, + "loss": 0.1195, + "step": 34638 + }, + { + "epoch": 0.8119927096702356, + "grad_norm": 0.31631284952163696, + "learning_rate": 1.6948565813936012e-05, + "loss": 0.0486, + "step": 34639 + }, + { + "epoch": 0.8120161512450407, + "grad_norm": 0.4124850034713745, + "learning_rate": 1.6944463109660645e-05, + "loss": 0.5189, + "step": 34640 + }, + { + "epoch": 0.8120395928198456, + "grad_norm": 0.25896042585372925, + "learning_rate": 1.694036085604699e-05, + "loss": 0.0554, + "step": 34641 + }, + { + "epoch": 0.8120630343946507, + "grad_norm": 0.1803169995546341, + "learning_rate": 1.6936259053117253e-05, + "loss": 0.0291, + "step": 34642 + }, + { + "epoch": 0.8120864759694556, + "grad_norm": 0.13935500383377075, + "learning_rate": 1.6932157700893725e-05, + "loss": 0.0079, + "step": 34643 + }, + { + "epoch": 0.8121099175442607, + "grad_norm": 0.49906831979751587, + "learning_rate": 1.692805679939865e-05, + "loss": 0.1054, + "step": 34644 + }, + { + "epoch": 0.8121333591190656, + "grad_norm": 0.6113277077674866, + "learning_rate": 1.6923956348654267e-05, + "loss": 0.1081, + "step": 34645 + }, + { + "epoch": 0.8121568006938706, + "grad_norm": 0.3429611623287201, + "learning_rate": 1.6919856348682837e-05, + "loss": 0.0659, + "step": 34646 + }, + { + "epoch": 0.8121802422686756, + "grad_norm": 0.8496502637863159, + "learning_rate": 1.6915756799506577e-05, + "loss": 0.1144, + "step": 34647 + }, + { + "epoch": 0.8122036838434806, + "grad_norm": 0.532950222492218, + "learning_rate": 1.6911657701147787e-05, + "loss": 0.7148, + "step": 34648 + }, + { + "epoch": 0.8122271254182856, + "grad_norm": 0.31653138995170593, + "learning_rate": 1.6907559053628673e-05, + "loss": 0.0656, + "step": 34649 + }, + { + "epoch": 0.8122505669930906, + "grad_norm": 0.164706289768219, + "learning_rate": 1.6903460856971485e-05, + "loss": 0.0203, + "step": 34650 + }, + { + "epoch": 0.8122740085678956, + "grad_norm": 0.29443466663360596, + "learning_rate": 1.6899363111198453e-05, + "loss": 0.0535, + "step": 34651 + }, + { + "epoch": 0.8122974501427006, + "grad_norm": 0.5817375183105469, + "learning_rate": 1.6895265816331786e-05, + "loss": 0.1031, + "step": 34652 + }, + { + "epoch": 0.8123208917175055, + "grad_norm": 0.44194164872169495, + "learning_rate": 1.6891168972393767e-05, + "loss": 0.0862, + "step": 34653 + }, + { + "epoch": 0.8123443332923106, + "grad_norm": 0.3323632478713989, + "learning_rate": 1.6887072579406582e-05, + "loss": 0.0692, + "step": 34654 + }, + { + "epoch": 0.8123677748671155, + "grad_norm": 0.134318545460701, + "learning_rate": 1.68829766373925e-05, + "loss": 0.022, + "step": 34655 + }, + { + "epoch": 0.8123912164419206, + "grad_norm": 0.3234715163707733, + "learning_rate": 1.6878881146373714e-05, + "loss": 0.053, + "step": 34656 + }, + { + "epoch": 0.8124146580167255, + "grad_norm": 0.25801458954811096, + "learning_rate": 1.6874786106372476e-05, + "loss": 0.0411, + "step": 34657 + }, + { + "epoch": 0.8124380995915306, + "grad_norm": 0.37571367621421814, + "learning_rate": 1.6870691517410973e-05, + "loss": 0.043, + "step": 34658 + }, + { + "epoch": 0.8124615411663355, + "grad_norm": 0.19417287409305573, + "learning_rate": 1.6866597379511405e-05, + "loss": 0.0288, + "step": 34659 + }, + { + "epoch": 0.8124849827411406, + "grad_norm": 0.27570465207099915, + "learning_rate": 1.686250369269605e-05, + "loss": 0.0441, + "step": 34660 + }, + { + "epoch": 0.8125084243159455, + "grad_norm": 0.4349946081638336, + "learning_rate": 1.6858410456987094e-05, + "loss": 0.0801, + "step": 34661 + }, + { + "epoch": 0.8125318658907505, + "grad_norm": 0.7399290204048157, + "learning_rate": 1.685431767240674e-05, + "loss": 0.1007, + "step": 34662 + }, + { + "epoch": 0.8125553074655555, + "grad_norm": 0.6715207099914551, + "learning_rate": 1.6850225338977187e-05, + "loss": 0.1246, + "step": 34663 + }, + { + "epoch": 0.8125787490403605, + "grad_norm": 0.19264616072177887, + "learning_rate": 1.6846133456720648e-05, + "loss": 0.0394, + "step": 34664 + }, + { + "epoch": 0.8126021906151655, + "grad_norm": 0.3369165062904358, + "learning_rate": 1.6842042025659312e-05, + "loss": 0.0447, + "step": 34665 + }, + { + "epoch": 0.8126256321899705, + "grad_norm": 0.611299455165863, + "learning_rate": 1.683795104581538e-05, + "loss": 0.0526, + "step": 34666 + }, + { + "epoch": 0.8126490737647755, + "grad_norm": 0.3065170347690582, + "learning_rate": 1.683386051721111e-05, + "loss": 0.0533, + "step": 34667 + }, + { + "epoch": 0.8126725153395805, + "grad_norm": 0.3220697343349457, + "learning_rate": 1.682977043986863e-05, + "loss": 0.0482, + "step": 34668 + }, + { + "epoch": 0.8126959569143855, + "grad_norm": 0.468046635389328, + "learning_rate": 1.6825680813810163e-05, + "loss": 0.0808, + "step": 34669 + }, + { + "epoch": 0.8127193984891905, + "grad_norm": 0.12731218338012695, + "learning_rate": 1.68215916390579e-05, + "loss": 0.0214, + "step": 34670 + }, + { + "epoch": 0.8127428400639956, + "grad_norm": 0.3307677209377289, + "learning_rate": 1.681750291563401e-05, + "loss": 0.0608, + "step": 34671 + }, + { + "epoch": 0.8127662816388005, + "grad_norm": 0.2550942301750183, + "learning_rate": 1.6813414643560653e-05, + "loss": 0.0413, + "step": 34672 + }, + { + "epoch": 0.8127897232136055, + "grad_norm": 0.46038636565208435, + "learning_rate": 1.6809326822860083e-05, + "loss": 0.0615, + "step": 34673 + }, + { + "epoch": 0.8128131647884105, + "grad_norm": 0.589209794998169, + "learning_rate": 1.6805239453554444e-05, + "loss": 0.127, + "step": 34674 + }, + { + "epoch": 0.8128366063632155, + "grad_norm": 0.11615202575922012, + "learning_rate": 1.680115253566592e-05, + "loss": 0.0096, + "step": 34675 + }, + { + "epoch": 0.8128600479380205, + "grad_norm": 0.13452312350273132, + "learning_rate": 1.6797066069216672e-05, + "loss": 0.0222, + "step": 34676 + }, + { + "epoch": 0.8128834895128255, + "grad_norm": 0.3467026352882385, + "learning_rate": 1.679298005422889e-05, + "loss": 0.0759, + "step": 34677 + }, + { + "epoch": 0.8129069310876305, + "grad_norm": 0.3364485204219818, + "learning_rate": 1.6788894490724714e-05, + "loss": 0.0446, + "step": 34678 + }, + { + "epoch": 0.8129303726624355, + "grad_norm": 0.507013201713562, + "learning_rate": 1.678480937872633e-05, + "loss": 0.1046, + "step": 34679 + }, + { + "epoch": 0.8129538142372404, + "grad_norm": 0.28502604365348816, + "learning_rate": 1.6780724718255935e-05, + "loss": 0.0308, + "step": 34680 + }, + { + "epoch": 0.8129772558120455, + "grad_norm": 0.419615775346756, + "learning_rate": 1.6776640509335674e-05, + "loss": 0.0798, + "step": 34681 + }, + { + "epoch": 0.8130006973868504, + "grad_norm": 0.40983259677886963, + "learning_rate": 1.67725567519877e-05, + "loss": 0.0992, + "step": 34682 + }, + { + "epoch": 0.8130241389616555, + "grad_norm": 0.45386379957199097, + "learning_rate": 1.6768473446234168e-05, + "loss": 0.5895, + "step": 34683 + }, + { + "epoch": 0.8130475805364604, + "grad_norm": 0.1299818605184555, + "learning_rate": 1.6764390592097244e-05, + "loss": 0.0178, + "step": 34684 + }, + { + "epoch": 0.8130710221112655, + "grad_norm": 0.34101763367652893, + "learning_rate": 1.676030818959905e-05, + "loss": 0.0383, + "step": 34685 + }, + { + "epoch": 0.8130944636860704, + "grad_norm": 0.12511132657527924, + "learning_rate": 1.6756226238761798e-05, + "loss": 0.0153, + "step": 34686 + }, + { + "epoch": 0.8131179052608755, + "grad_norm": 0.3859764337539673, + "learning_rate": 1.675214473960759e-05, + "loss": 0.0507, + "step": 34687 + }, + { + "epoch": 0.8131413468356804, + "grad_norm": 0.09305767714977264, + "learning_rate": 1.674806369215859e-05, + "loss": 0.0089, + "step": 34688 + }, + { + "epoch": 0.8131647884104855, + "grad_norm": 1.6787726879119873, + "learning_rate": 1.6743983096436945e-05, + "loss": 0.0574, + "step": 34689 + }, + { + "epoch": 0.8131882299852904, + "grad_norm": 0.2066514641046524, + "learning_rate": 1.673990295246475e-05, + "loss": 0.0336, + "step": 34690 + }, + { + "epoch": 0.8132116715600954, + "grad_norm": 0.4303053915500641, + "learning_rate": 1.6735823260264215e-05, + "loss": 0.0656, + "step": 34691 + }, + { + "epoch": 0.8132351131349004, + "grad_norm": 0.5478767156600952, + "learning_rate": 1.6731744019857422e-05, + "loss": 0.1079, + "step": 34692 + }, + { + "epoch": 0.8132585547097054, + "grad_norm": 0.3524964451789856, + "learning_rate": 1.6727665231266544e-05, + "loss": 0.0551, + "step": 34693 + }, + { + "epoch": 0.8132819962845104, + "grad_norm": 0.23489269614219666, + "learning_rate": 1.67235868945137e-05, + "loss": 0.0401, + "step": 34694 + }, + { + "epoch": 0.8133054378593154, + "grad_norm": 0.4499242603778839, + "learning_rate": 1.6719509009621016e-05, + "loss": 0.0499, + "step": 34695 + }, + { + "epoch": 0.8133288794341204, + "grad_norm": 0.453355997800827, + "learning_rate": 1.671543157661062e-05, + "loss": 0.0984, + "step": 34696 + }, + { + "epoch": 0.8133523210089254, + "grad_norm": 0.7660284638404846, + "learning_rate": 1.6711354595504604e-05, + "loss": 0.1585, + "step": 34697 + }, + { + "epoch": 0.8133757625837303, + "grad_norm": 0.4186258018016815, + "learning_rate": 1.6707278066325148e-05, + "loss": 0.0622, + "step": 34698 + }, + { + "epoch": 0.8133992041585354, + "grad_norm": 0.20525988936424255, + "learning_rate": 1.670320198909434e-05, + "loss": 0.0335, + "step": 34699 + }, + { + "epoch": 0.8134226457333403, + "grad_norm": 0.574553906917572, + "learning_rate": 1.6699126363834307e-05, + "loss": 0.1473, + "step": 34700 + }, + { + "epoch": 0.8134460873081454, + "grad_norm": 0.14942017197608948, + "learning_rate": 1.6695051190567144e-05, + "loss": 0.03, + "step": 34701 + }, + { + "epoch": 0.8134695288829503, + "grad_norm": 0.07540823519229889, + "learning_rate": 1.6690976469314967e-05, + "loss": 0.0054, + "step": 34702 + }, + { + "epoch": 0.8134929704577554, + "grad_norm": 0.520898699760437, + "learning_rate": 1.668690220009991e-05, + "loss": 0.089, + "step": 34703 + }, + { + "epoch": 0.8135164120325603, + "grad_norm": 0.21252553164958954, + "learning_rate": 1.6682828382944038e-05, + "loss": 0.0248, + "step": 34704 + }, + { + "epoch": 0.8135398536073654, + "grad_norm": 0.44230660796165466, + "learning_rate": 1.6678755017869506e-05, + "loss": 0.5585, + "step": 34705 + }, + { + "epoch": 0.8135632951821703, + "grad_norm": 0.3718281388282776, + "learning_rate": 1.66746821048984e-05, + "loss": 0.0589, + "step": 34706 + }, + { + "epoch": 0.8135867367569753, + "grad_norm": 0.45045918226242065, + "learning_rate": 1.6670609644052805e-05, + "loss": 0.0856, + "step": 34707 + }, + { + "epoch": 0.8136101783317803, + "grad_norm": 0.20004937052726746, + "learning_rate": 1.6666537635354818e-05, + "loss": 0.0428, + "step": 34708 + }, + { + "epoch": 0.8136336199065853, + "grad_norm": 0.5941588282585144, + "learning_rate": 1.6662466078826555e-05, + "loss": 0.0709, + "step": 34709 + }, + { + "epoch": 0.8136570614813903, + "grad_norm": 0.3287532925605774, + "learning_rate": 1.6658394974490067e-05, + "loss": 0.0304, + "step": 34710 + }, + { + "epoch": 0.8136805030561953, + "grad_norm": 0.5124742984771729, + "learning_rate": 1.665432432236749e-05, + "loss": 0.0397, + "step": 34711 + }, + { + "epoch": 0.8137039446310003, + "grad_norm": 0.3418819308280945, + "learning_rate": 1.665025412248089e-05, + "loss": 0.0643, + "step": 34712 + }, + { + "epoch": 0.8137273862058053, + "grad_norm": 0.4468291997909546, + "learning_rate": 1.664618437485237e-05, + "loss": 0.1034, + "step": 34713 + }, + { + "epoch": 0.8137508277806103, + "grad_norm": 0.28508260846138, + "learning_rate": 1.664211507950396e-05, + "loss": 0.0389, + "step": 34714 + }, + { + "epoch": 0.8137742693554153, + "grad_norm": 0.9769321084022522, + "learning_rate": 1.6638046236457804e-05, + "loss": 0.2324, + "step": 34715 + }, + { + "epoch": 0.8137977109302202, + "grad_norm": 0.40701985359191895, + "learning_rate": 1.6633977845735947e-05, + "loss": 0.0612, + "step": 34716 + }, + { + "epoch": 0.8138211525050253, + "grad_norm": 0.5477232933044434, + "learning_rate": 1.6629909907360463e-05, + "loss": 0.0928, + "step": 34717 + }, + { + "epoch": 0.8138445940798302, + "grad_norm": 0.34978896379470825, + "learning_rate": 1.6625842421353444e-05, + "loss": 0.0546, + "step": 34718 + }, + { + "epoch": 0.8138680356546353, + "grad_norm": 0.5742238163948059, + "learning_rate": 1.6621775387736947e-05, + "loss": 0.1052, + "step": 34719 + }, + { + "epoch": 0.8138914772294402, + "grad_norm": 0.49570462107658386, + "learning_rate": 1.661770880653305e-05, + "loss": 0.0945, + "step": 34720 + }, + { + "epoch": 0.8139149188042453, + "grad_norm": 0.5084497332572937, + "learning_rate": 1.6613642677763795e-05, + "loss": 0.0799, + "step": 34721 + }, + { + "epoch": 0.8139383603790503, + "grad_norm": 0.5599242448806763, + "learning_rate": 1.660957700145127e-05, + "loss": 0.0757, + "step": 34722 + }, + { + "epoch": 0.8139618019538553, + "grad_norm": 0.32700154185295105, + "learning_rate": 1.66055117776175e-05, + "loss": 0.0565, + "step": 34723 + }, + { + "epoch": 0.8139852435286603, + "grad_norm": 0.4975830614566803, + "learning_rate": 1.6601447006284597e-05, + "loss": 0.072, + "step": 34724 + }, + { + "epoch": 0.8140086851034652, + "grad_norm": 0.658464789390564, + "learning_rate": 1.659738268747458e-05, + "loss": 0.1046, + "step": 34725 + }, + { + "epoch": 0.8140321266782703, + "grad_norm": 0.19419635832309723, + "learning_rate": 1.6593318821209513e-05, + "loss": 0.0224, + "step": 34726 + }, + { + "epoch": 0.8140555682530752, + "grad_norm": 0.1600380539894104, + "learning_rate": 1.6589255407511418e-05, + "loss": 0.0249, + "step": 34727 + }, + { + "epoch": 0.8140790098278803, + "grad_norm": 0.3226635158061981, + "learning_rate": 1.6585192446402386e-05, + "loss": 0.0846, + "step": 34728 + }, + { + "epoch": 0.8141024514026852, + "grad_norm": 0.1796116828918457, + "learning_rate": 1.658112993790446e-05, + "loss": 0.0365, + "step": 34729 + }, + { + "epoch": 0.8141258929774903, + "grad_norm": 0.6605568528175354, + "learning_rate": 1.657706788203963e-05, + "loss": 0.1523, + "step": 34730 + }, + { + "epoch": 0.8141493345522952, + "grad_norm": 0.5317580103874207, + "learning_rate": 1.6573006278830007e-05, + "loss": 0.084, + "step": 34731 + }, + { + "epoch": 0.8141727761271003, + "grad_norm": 0.9497779011726379, + "learning_rate": 1.6568945128297597e-05, + "loss": 0.1349, + "step": 34732 + }, + { + "epoch": 0.8141962177019052, + "grad_norm": 0.6220201253890991, + "learning_rate": 1.6564884430464435e-05, + "loss": 0.1057, + "step": 34733 + }, + { + "epoch": 0.8142196592767103, + "grad_norm": 0.5953447818756104, + "learning_rate": 1.6560824185352552e-05, + "loss": 0.5318, + "step": 34734 + }, + { + "epoch": 0.8142431008515152, + "grad_norm": 0.4993845522403717, + "learning_rate": 1.655676439298397e-05, + "loss": 0.0995, + "step": 34735 + }, + { + "epoch": 0.8142665424263202, + "grad_norm": 0.4904788136482239, + "learning_rate": 1.6552705053380747e-05, + "loss": 0.4621, + "step": 34736 + }, + { + "epoch": 0.8142899840011252, + "grad_norm": 0.38415589928627014, + "learning_rate": 1.6548646166564898e-05, + "loss": 0.0523, + "step": 34737 + }, + { + "epoch": 0.8143134255759302, + "grad_norm": 0.44235894083976746, + "learning_rate": 1.6544587732558435e-05, + "loss": 0.0416, + "step": 34738 + }, + { + "epoch": 0.8143368671507352, + "grad_norm": 0.18059633672237396, + "learning_rate": 1.654052975138336e-05, + "loss": 0.0181, + "step": 34739 + }, + { + "epoch": 0.8143603087255402, + "grad_norm": 0.34495827555656433, + "learning_rate": 1.653647222306175e-05, + "loss": 0.034, + "step": 34740 + }, + { + "epoch": 0.8143837503003452, + "grad_norm": 0.1295383870601654, + "learning_rate": 1.6532415147615588e-05, + "loss": 0.0164, + "step": 34741 + }, + { + "epoch": 0.8144071918751502, + "grad_norm": 0.7026227116584778, + "learning_rate": 1.6528358525066855e-05, + "loss": 0.1691, + "step": 34742 + }, + { + "epoch": 0.8144306334499551, + "grad_norm": 0.36240604519844055, + "learning_rate": 1.6524302355437626e-05, + "loss": 0.1173, + "step": 34743 + }, + { + "epoch": 0.8144540750247602, + "grad_norm": 0.3206673562526703, + "learning_rate": 1.652024663874988e-05, + "loss": 0.0277, + "step": 34744 + }, + { + "epoch": 0.8144775165995651, + "grad_norm": 0.5556597113609314, + "learning_rate": 1.6516191375025614e-05, + "loss": 0.0837, + "step": 34745 + }, + { + "epoch": 0.8145009581743702, + "grad_norm": 0.19786444306373596, + "learning_rate": 1.651213656428685e-05, + "loss": 0.0304, + "step": 34746 + }, + { + "epoch": 0.8145243997491751, + "grad_norm": 0.08133866637945175, + "learning_rate": 1.6508082206555574e-05, + "loss": 0.0152, + "step": 34747 + }, + { + "epoch": 0.8145478413239802, + "grad_norm": 0.7087007761001587, + "learning_rate": 1.650402830185378e-05, + "loss": 0.0717, + "step": 34748 + }, + { + "epoch": 0.8145712828987851, + "grad_norm": 0.1986861377954483, + "learning_rate": 1.6499974850203493e-05, + "loss": 0.0306, + "step": 34749 + }, + { + "epoch": 0.8145947244735902, + "grad_norm": 0.815048336982727, + "learning_rate": 1.6495921851626685e-05, + "loss": 0.2778, + "step": 34750 + }, + { + "epoch": 0.8146181660483951, + "grad_norm": 0.11808722466230392, + "learning_rate": 1.6491869306145336e-05, + "loss": 0.0207, + "step": 34751 + }, + { + "epoch": 0.8146416076232001, + "grad_norm": 0.4220362603664398, + "learning_rate": 1.6487817213781475e-05, + "loss": 0.2841, + "step": 34752 + }, + { + "epoch": 0.8146650491980051, + "grad_norm": 0.16951167583465576, + "learning_rate": 1.6483765574557065e-05, + "loss": 0.0273, + "step": 34753 + }, + { + "epoch": 0.8146884907728101, + "grad_norm": 0.23528048396110535, + "learning_rate": 1.6479714388494093e-05, + "loss": 0.0166, + "step": 34754 + }, + { + "epoch": 0.8147119323476151, + "grad_norm": 0.31296825408935547, + "learning_rate": 1.6475663655614503e-05, + "loss": 0.0292, + "step": 34755 + }, + { + "epoch": 0.8147353739224201, + "grad_norm": 0.2159663736820221, + "learning_rate": 1.6471613375940353e-05, + "loss": 0.0383, + "step": 34756 + }, + { + "epoch": 0.8147588154972251, + "grad_norm": 0.29278823733329773, + "learning_rate": 1.646756354949357e-05, + "loss": 0.0617, + "step": 34757 + }, + { + "epoch": 0.8147822570720301, + "grad_norm": 0.9067873358726501, + "learning_rate": 1.6463514176296136e-05, + "loss": 0.0955, + "step": 34758 + }, + { + "epoch": 0.814805698646835, + "grad_norm": 1.0158028602600098, + "learning_rate": 1.645946525637002e-05, + "loss": 0.1488, + "step": 34759 + }, + { + "epoch": 0.8148291402216401, + "grad_norm": 0.11497650295495987, + "learning_rate": 1.6455416789737198e-05, + "loss": 0.0136, + "step": 34760 + }, + { + "epoch": 0.814852581796445, + "grad_norm": 0.17706026136875153, + "learning_rate": 1.6451368776419607e-05, + "loss": 0.0207, + "step": 34761 + }, + { + "epoch": 0.8148760233712501, + "grad_norm": 0.5069721937179565, + "learning_rate": 1.644732121643926e-05, + "loss": 0.0728, + "step": 34762 + }, + { + "epoch": 0.814899464946055, + "grad_norm": 0.4452235698699951, + "learning_rate": 1.6443274109818073e-05, + "loss": 0.0925, + "step": 34763 + }, + { + "epoch": 0.8149229065208601, + "grad_norm": 0.33249765634536743, + "learning_rate": 1.6439227456578054e-05, + "loss": 0.0536, + "step": 34764 + }, + { + "epoch": 0.814946348095665, + "grad_norm": 0.27916860580444336, + "learning_rate": 1.643518125674115e-05, + "loss": 0.0264, + "step": 34765 + }, + { + "epoch": 0.8149697896704701, + "grad_norm": 0.3637235164642334, + "learning_rate": 1.643113551032929e-05, + "loss": 0.0752, + "step": 34766 + }, + { + "epoch": 0.814993231245275, + "grad_norm": 0.5646817088127136, + "learning_rate": 1.6427090217364416e-05, + "loss": 0.042, + "step": 34767 + }, + { + "epoch": 0.8150166728200801, + "grad_norm": 0.7224993705749512, + "learning_rate": 1.6423045377868518e-05, + "loss": 0.1471, + "step": 34768 + }, + { + "epoch": 0.815040114394885, + "grad_norm": 0.17372800409793854, + "learning_rate": 1.6419000991863543e-05, + "loss": 0.0188, + "step": 34769 + }, + { + "epoch": 0.81506355596969, + "grad_norm": 0.08945964276790619, + "learning_rate": 1.6414957059371406e-05, + "loss": 0.0133, + "step": 34770 + }, + { + "epoch": 0.815086997544495, + "grad_norm": 0.5323563814163208, + "learning_rate": 1.6410913580414068e-05, + "loss": 0.1042, + "step": 34771 + }, + { + "epoch": 0.8151104391193, + "grad_norm": 0.598639726638794, + "learning_rate": 1.6406870555013455e-05, + "loss": 0.0888, + "step": 34772 + }, + { + "epoch": 0.8151338806941051, + "grad_norm": 0.3818499445915222, + "learning_rate": 1.6402827983191493e-05, + "loss": 0.0921, + "step": 34773 + }, + { + "epoch": 0.81515732226891, + "grad_norm": 0.5763763189315796, + "learning_rate": 1.6398785864970156e-05, + "loss": 0.0874, + "step": 34774 + }, + { + "epoch": 0.8151807638437151, + "grad_norm": 0.38037410378456116, + "learning_rate": 1.6394744200371347e-05, + "loss": 0.4199, + "step": 34775 + }, + { + "epoch": 0.81520420541852, + "grad_norm": 0.10734827816486359, + "learning_rate": 1.6390702989417018e-05, + "loss": 0.0148, + "step": 34776 + }, + { + "epoch": 0.8152276469933251, + "grad_norm": 0.0963943600654602, + "learning_rate": 1.638666223212909e-05, + "loss": 0.0103, + "step": 34777 + }, + { + "epoch": 0.81525108856813, + "grad_norm": 0.17806024849414825, + "learning_rate": 1.6382621928529486e-05, + "loss": 0.031, + "step": 34778 + }, + { + "epoch": 0.815274530142935, + "grad_norm": 0.08436320722103119, + "learning_rate": 1.6378582078640125e-05, + "loss": 0.0083, + "step": 34779 + }, + { + "epoch": 0.81529797171774, + "grad_norm": 0.2525162100791931, + "learning_rate": 1.6374542682482908e-05, + "loss": 0.0604, + "step": 34780 + }, + { + "epoch": 0.815321413292545, + "grad_norm": 0.3338004946708679, + "learning_rate": 1.637050374007979e-05, + "loss": 0.0396, + "step": 34781 + }, + { + "epoch": 0.81534485486735, + "grad_norm": 0.15464438498020172, + "learning_rate": 1.6366465251452678e-05, + "loss": 0.0306, + "step": 34782 + }, + { + "epoch": 0.815368296442155, + "grad_norm": 0.5124607086181641, + "learning_rate": 1.6362427216623467e-05, + "loss": 0.0915, + "step": 34783 + }, + { + "epoch": 0.81539173801696, + "grad_norm": 0.2609585225582123, + "learning_rate": 1.6358389635614078e-05, + "loss": 0.0546, + "step": 34784 + }, + { + "epoch": 0.815415179591765, + "grad_norm": 0.23510444164276123, + "learning_rate": 1.6354352508446437e-05, + "loss": 0.042, + "step": 34785 + }, + { + "epoch": 0.81543862116657, + "grad_norm": 0.2954455018043518, + "learning_rate": 1.6350315835142393e-05, + "loss": 0.0746, + "step": 34786 + }, + { + "epoch": 0.815462062741375, + "grad_norm": 0.4227551221847534, + "learning_rate": 1.634627961572388e-05, + "loss": 0.0709, + "step": 34787 + }, + { + "epoch": 0.8154855043161799, + "grad_norm": 0.7117533087730408, + "learning_rate": 1.6342243850212846e-05, + "loss": 0.1327, + "step": 34788 + }, + { + "epoch": 0.815508945890985, + "grad_norm": 0.6921875476837158, + "learning_rate": 1.633820853863114e-05, + "loss": 0.0974, + "step": 34789 + }, + { + "epoch": 0.8155323874657899, + "grad_norm": 0.3656403422355652, + "learning_rate": 1.6334173681000662e-05, + "loss": 0.038, + "step": 34790 + }, + { + "epoch": 0.815555829040595, + "grad_norm": 0.3856056332588196, + "learning_rate": 1.6330139277343325e-05, + "loss": 0.1173, + "step": 34791 + }, + { + "epoch": 0.8155792706153999, + "grad_norm": 0.6382459998130798, + "learning_rate": 1.6326105327680997e-05, + "loss": 0.0421, + "step": 34792 + }, + { + "epoch": 0.815602712190205, + "grad_norm": 0.26059624552726746, + "learning_rate": 1.6322071832035546e-05, + "loss": 0.0626, + "step": 34793 + }, + { + "epoch": 0.8156261537650099, + "grad_norm": 0.5461251139640808, + "learning_rate": 1.6318038790428913e-05, + "loss": 0.142, + "step": 34794 + }, + { + "epoch": 0.815649595339815, + "grad_norm": 0.2768266201019287, + "learning_rate": 1.6314006202882958e-05, + "loss": 0.0479, + "step": 34795 + }, + { + "epoch": 0.8156730369146199, + "grad_norm": 0.163561150431633, + "learning_rate": 1.6309974069419554e-05, + "loss": 0.0239, + "step": 34796 + }, + { + "epoch": 0.815696478489425, + "grad_norm": 0.2609463632106781, + "learning_rate": 1.630594239006059e-05, + "loss": 0.0224, + "step": 34797 + }, + { + "epoch": 0.8157199200642299, + "grad_norm": 0.40098199248313904, + "learning_rate": 1.6301911164827933e-05, + "loss": 0.0345, + "step": 34798 + }, + { + "epoch": 0.8157433616390349, + "grad_norm": 0.4466351568698883, + "learning_rate": 1.6297880393743436e-05, + "loss": 0.0821, + "step": 34799 + }, + { + "epoch": 0.8157668032138399, + "grad_norm": 0.47411438822746277, + "learning_rate": 1.629385007682899e-05, + "loss": 0.0751, + "step": 34800 + }, + { + "epoch": 0.8157902447886449, + "grad_norm": 0.2711835205554962, + "learning_rate": 1.628982021410649e-05, + "loss": 0.0353, + "step": 34801 + }, + { + "epoch": 0.8158136863634499, + "grad_norm": 0.6598461270332336, + "learning_rate": 1.6285790805597778e-05, + "loss": 0.0823, + "step": 34802 + }, + { + "epoch": 0.8158371279382549, + "grad_norm": 0.5657410025596619, + "learning_rate": 1.6281761851324728e-05, + "loss": 0.1166, + "step": 34803 + }, + { + "epoch": 0.8158605695130599, + "grad_norm": 0.49826666712760925, + "learning_rate": 1.6277733351309187e-05, + "loss": 0.481, + "step": 34804 + }, + { + "epoch": 0.8158840110878649, + "grad_norm": 0.48070600628852844, + "learning_rate": 1.6273705305572994e-05, + "loss": 0.1215, + "step": 34805 + }, + { + "epoch": 0.8159074526626698, + "grad_norm": 0.3005545437335968, + "learning_rate": 1.6269677714138053e-05, + "loss": 0.0764, + "step": 34806 + }, + { + "epoch": 0.8159308942374749, + "grad_norm": 0.4044812321662903, + "learning_rate": 1.6265650577026205e-05, + "loss": 0.0741, + "step": 34807 + }, + { + "epoch": 0.8159543358122798, + "grad_norm": 0.16194656491279602, + "learning_rate": 1.6261623894259283e-05, + "loss": 0.0339, + "step": 34808 + }, + { + "epoch": 0.8159777773870849, + "grad_norm": 0.3759737014770508, + "learning_rate": 1.6257597665859136e-05, + "loss": 0.0724, + "step": 34809 + }, + { + "epoch": 0.8160012189618898, + "grad_norm": 0.07091677933931351, + "learning_rate": 1.625357189184763e-05, + "loss": 0.0051, + "step": 34810 + }, + { + "epoch": 0.8160246605366949, + "grad_norm": 0.28213968873023987, + "learning_rate": 1.6249546572246576e-05, + "loss": 0.0628, + "step": 34811 + }, + { + "epoch": 0.8160481021114998, + "grad_norm": 0.25477346777915955, + "learning_rate": 1.6245521707077825e-05, + "loss": 0.0446, + "step": 34812 + }, + { + "epoch": 0.8160715436863049, + "grad_norm": 0.7900967597961426, + "learning_rate": 1.624149729636326e-05, + "loss": 0.1325, + "step": 34813 + }, + { + "epoch": 0.8160949852611098, + "grad_norm": 0.5669744610786438, + "learning_rate": 1.623747334012469e-05, + "loss": 0.723, + "step": 34814 + }, + { + "epoch": 0.8161184268359148, + "grad_norm": 0.46840518712997437, + "learning_rate": 1.623344983838394e-05, + "loss": 0.0774, + "step": 34815 + }, + { + "epoch": 0.8161418684107198, + "grad_norm": 0.3311254382133484, + "learning_rate": 1.622942679116285e-05, + "loss": 0.0468, + "step": 34816 + }, + { + "epoch": 0.8161653099855248, + "grad_norm": 0.1492847055196762, + "learning_rate": 1.622540419848324e-05, + "loss": 0.0189, + "step": 34817 + }, + { + "epoch": 0.8161887515603298, + "grad_norm": 0.08326203376054764, + "learning_rate": 1.6221382060366918e-05, + "loss": 0.0106, + "step": 34818 + }, + { + "epoch": 0.8162121931351348, + "grad_norm": 0.24330684542655945, + "learning_rate": 1.6217360376835755e-05, + "loss": 0.0237, + "step": 34819 + }, + { + "epoch": 0.8162356347099398, + "grad_norm": 0.20851486921310425, + "learning_rate": 1.6213339147911556e-05, + "loss": 0.0309, + "step": 34820 + }, + { + "epoch": 0.8162590762847448, + "grad_norm": 0.2806934416294098, + "learning_rate": 1.6209318373616133e-05, + "loss": 0.0444, + "step": 34821 + }, + { + "epoch": 0.8162825178595497, + "grad_norm": 0.7123078107833862, + "learning_rate": 1.6205298053971306e-05, + "loss": 0.4845, + "step": 34822 + }, + { + "epoch": 0.8163059594343548, + "grad_norm": 0.5224402546882629, + "learning_rate": 1.6201278188998892e-05, + "loss": 0.0849, + "step": 34823 + }, + { + "epoch": 0.8163294010091598, + "grad_norm": 0.79441899061203, + "learning_rate": 1.6197258778720658e-05, + "loss": 0.1613, + "step": 34824 + }, + { + "epoch": 0.8163528425839648, + "grad_norm": 0.10365215688943863, + "learning_rate": 1.619323982315847e-05, + "loss": 0.0129, + "step": 34825 + }, + { + "epoch": 0.8163762841587698, + "grad_norm": 0.4149973690509796, + "learning_rate": 1.6189221322334138e-05, + "loss": 0.0506, + "step": 34826 + }, + { + "epoch": 0.8163997257335748, + "grad_norm": 0.5589354038238525, + "learning_rate": 1.6185203276269446e-05, + "loss": 0.0314, + "step": 34827 + }, + { + "epoch": 0.8164231673083798, + "grad_norm": 0.5126256346702576, + "learning_rate": 1.61811856849862e-05, + "loss": 0.076, + "step": 34828 + }, + { + "epoch": 0.8164466088831848, + "grad_norm": 0.112644262611866, + "learning_rate": 1.6177168548506204e-05, + "loss": 0.0206, + "step": 34829 + }, + { + "epoch": 0.8164700504579898, + "grad_norm": 0.42646750807762146, + "learning_rate": 1.6173151866851244e-05, + "loss": 0.0684, + "step": 34830 + }, + { + "epoch": 0.8164934920327948, + "grad_norm": 0.419109582901001, + "learning_rate": 1.6169135640043086e-05, + "loss": 0.035, + "step": 34831 + }, + { + "epoch": 0.8165169336075998, + "grad_norm": 0.4623226821422577, + "learning_rate": 1.6165119868103584e-05, + "loss": 0.0495, + "step": 34832 + }, + { + "epoch": 0.8165403751824047, + "grad_norm": 0.16710366308689117, + "learning_rate": 1.6161104551054507e-05, + "loss": 0.0283, + "step": 34833 + }, + { + "epoch": 0.8165638167572098, + "grad_norm": 0.4296693503856659, + "learning_rate": 1.6157089688917627e-05, + "loss": 0.0635, + "step": 34834 + }, + { + "epoch": 0.8165872583320147, + "grad_norm": 0.5514533519744873, + "learning_rate": 1.6153075281714737e-05, + "loss": 0.5733, + "step": 34835 + }, + { + "epoch": 0.8166106999068198, + "grad_norm": 0.38750314712524414, + "learning_rate": 1.6149061329467584e-05, + "loss": 0.0736, + "step": 34836 + }, + { + "epoch": 0.8166341414816247, + "grad_norm": 0.5414901971817017, + "learning_rate": 1.6145047832198023e-05, + "loss": 0.1066, + "step": 34837 + }, + { + "epoch": 0.8166575830564298, + "grad_norm": 0.5384231209754944, + "learning_rate": 1.6141034789927754e-05, + "loss": 0.0971, + "step": 34838 + }, + { + "epoch": 0.8166810246312347, + "grad_norm": 0.23312267661094666, + "learning_rate": 1.613702220267862e-05, + "loss": 0.0328, + "step": 34839 + }, + { + "epoch": 0.8167044662060398, + "grad_norm": 0.7868351340293884, + "learning_rate": 1.6133010070472355e-05, + "loss": 0.089, + "step": 34840 + }, + { + "epoch": 0.8167279077808447, + "grad_norm": 0.279674768447876, + "learning_rate": 1.6128998393330742e-05, + "loss": 0.037, + "step": 34841 + }, + { + "epoch": 0.8167513493556497, + "grad_norm": 0.43279972672462463, + "learning_rate": 1.612498717127554e-05, + "loss": 0.0645, + "step": 34842 + }, + { + "epoch": 0.8167747909304547, + "grad_norm": 0.34309372305870056, + "learning_rate": 1.6120976404328492e-05, + "loss": 0.0621, + "step": 34843 + }, + { + "epoch": 0.8167982325052597, + "grad_norm": 0.6638780236244202, + "learning_rate": 1.61169660925114e-05, + "loss": 0.1088, + "step": 34844 + }, + { + "epoch": 0.8168216740800647, + "grad_norm": 0.3357860743999481, + "learning_rate": 1.611295623584602e-05, + "loss": 0.0663, + "step": 34845 + }, + { + "epoch": 0.8168451156548697, + "grad_norm": 0.5181999206542969, + "learning_rate": 1.6108946834354087e-05, + "loss": 0.0939, + "step": 34846 + }, + { + "epoch": 0.8168685572296747, + "grad_norm": 0.45725104212760925, + "learning_rate": 1.6104937888057382e-05, + "loss": 0.1138, + "step": 34847 + }, + { + "epoch": 0.8168919988044797, + "grad_norm": 0.5009312033653259, + "learning_rate": 1.6100929396977603e-05, + "loss": 0.5456, + "step": 34848 + }, + { + "epoch": 0.8169154403792847, + "grad_norm": 0.41788339614868164, + "learning_rate": 1.6096921361136575e-05, + "loss": 0.0504, + "step": 34849 + }, + { + "epoch": 0.8169388819540897, + "grad_norm": 0.07475647330284119, + "learning_rate": 1.6092913780555984e-05, + "loss": 0.0093, + "step": 34850 + }, + { + "epoch": 0.8169623235288946, + "grad_norm": 0.43339574337005615, + "learning_rate": 1.6088906655257628e-05, + "loss": 0.0769, + "step": 34851 + }, + { + "epoch": 0.8169857651036997, + "grad_norm": 0.18129529058933258, + "learning_rate": 1.6084899985263225e-05, + "loss": 0.0247, + "step": 34852 + }, + { + "epoch": 0.8170092066785046, + "grad_norm": 0.5075936317443848, + "learning_rate": 1.6080893770594507e-05, + "loss": 0.0709, + "step": 34853 + }, + { + "epoch": 0.8170326482533097, + "grad_norm": 0.36141684651374817, + "learning_rate": 1.6076888011273228e-05, + "loss": 0.0332, + "step": 34854 + }, + { + "epoch": 0.8170560898281146, + "grad_norm": 0.29725751280784607, + "learning_rate": 1.60728827073211e-05, + "loss": 0.0747, + "step": 34855 + }, + { + "epoch": 0.8170795314029197, + "grad_norm": 0.8283411860466003, + "learning_rate": 1.6068877858759856e-05, + "loss": 0.0855, + "step": 34856 + }, + { + "epoch": 0.8171029729777246, + "grad_norm": 0.18866169452667236, + "learning_rate": 1.6064873465611276e-05, + "loss": 0.0217, + "step": 34857 + }, + { + "epoch": 0.8171264145525297, + "grad_norm": 0.38748404383659363, + "learning_rate": 1.6060869527897037e-05, + "loss": 0.0625, + "step": 34858 + }, + { + "epoch": 0.8171498561273346, + "grad_norm": 0.5143520832061768, + "learning_rate": 1.6056866045638886e-05, + "loss": 0.0433, + "step": 34859 + }, + { + "epoch": 0.8171732977021396, + "grad_norm": 0.3184615969657898, + "learning_rate": 1.605286301885851e-05, + "loss": 0.0729, + "step": 34860 + }, + { + "epoch": 0.8171967392769446, + "grad_norm": 0.452812522649765, + "learning_rate": 1.6048860447577687e-05, + "loss": 0.0729, + "step": 34861 + }, + { + "epoch": 0.8172201808517496, + "grad_norm": 0.3208985924720764, + "learning_rate": 1.6044858331818114e-05, + "loss": 0.0304, + "step": 34862 + }, + { + "epoch": 0.8172436224265546, + "grad_norm": 0.20956707000732422, + "learning_rate": 1.6040856671601467e-05, + "loss": 0.0194, + "step": 34863 + }, + { + "epoch": 0.8172670640013596, + "grad_norm": 0.20879317820072174, + "learning_rate": 1.6036855466949517e-05, + "loss": 0.0333, + "step": 34864 + }, + { + "epoch": 0.8172905055761646, + "grad_norm": 0.39971908926963806, + "learning_rate": 1.603285471788396e-05, + "loss": 0.0555, + "step": 34865 + }, + { + "epoch": 0.8173139471509696, + "grad_norm": 0.5229238867759705, + "learning_rate": 1.6028854424426488e-05, + "loss": 0.1401, + "step": 34866 + }, + { + "epoch": 0.8173373887257745, + "grad_norm": 0.3311910331249237, + "learning_rate": 1.6024854586598814e-05, + "loss": 0.0454, + "step": 34867 + }, + { + "epoch": 0.8173608303005796, + "grad_norm": 0.5336331129074097, + "learning_rate": 1.6020855204422635e-05, + "loss": 0.0995, + "step": 34868 + }, + { + "epoch": 0.8173842718753845, + "grad_norm": 0.15489153563976288, + "learning_rate": 1.6016856277919633e-05, + "loss": 0.0174, + "step": 34869 + }, + { + "epoch": 0.8174077134501896, + "grad_norm": 0.16146168112754822, + "learning_rate": 1.601285780711156e-05, + "loss": 0.011, + "step": 34870 + }, + { + "epoch": 0.8174311550249945, + "grad_norm": 0.8640155792236328, + "learning_rate": 1.600885979202008e-05, + "loss": 0.1413, + "step": 34871 + }, + { + "epoch": 0.8174545965997996, + "grad_norm": 0.528633177280426, + "learning_rate": 1.6004862232666885e-05, + "loss": 0.0939, + "step": 34872 + }, + { + "epoch": 0.8174780381746045, + "grad_norm": 0.5496454834938049, + "learning_rate": 1.600086512907365e-05, + "loss": 0.615, + "step": 34873 + }, + { + "epoch": 0.8175014797494096, + "grad_norm": 0.7961952686309814, + "learning_rate": 1.5996868481262106e-05, + "loss": 0.1353, + "step": 34874 + }, + { + "epoch": 0.8175249213242146, + "grad_norm": 0.33025652170181274, + "learning_rate": 1.5992872289253912e-05, + "loss": 0.0484, + "step": 34875 + }, + { + "epoch": 0.8175483628990196, + "grad_norm": 0.4538460969924927, + "learning_rate": 1.5988876553070742e-05, + "loss": 0.0352, + "step": 34876 + }, + { + "epoch": 0.8175718044738246, + "grad_norm": 0.5023803114891052, + "learning_rate": 1.5984881272734297e-05, + "loss": 0.0945, + "step": 34877 + }, + { + "epoch": 0.8175952460486295, + "grad_norm": 0.3573865592479706, + "learning_rate": 1.5980886448266265e-05, + "loss": 0.4055, + "step": 34878 + }, + { + "epoch": 0.8176186876234346, + "grad_norm": 0.3808004558086395, + "learning_rate": 1.5976892079688298e-05, + "loss": 0.0469, + "step": 34879 + }, + { + "epoch": 0.8176421291982395, + "grad_norm": 0.3374634385108948, + "learning_rate": 1.5972898167022075e-05, + "loss": 0.0226, + "step": 34880 + }, + { + "epoch": 0.8176655707730446, + "grad_norm": 0.2738020718097687, + "learning_rate": 1.5968904710289246e-05, + "loss": 0.0623, + "step": 34881 + }, + { + "epoch": 0.8176890123478495, + "grad_norm": 0.27577143907546997, + "learning_rate": 1.596491170951153e-05, + "loss": 0.0426, + "step": 34882 + }, + { + "epoch": 0.8177124539226546, + "grad_norm": 0.4359705150127411, + "learning_rate": 1.5960919164710565e-05, + "loss": 0.429, + "step": 34883 + }, + { + "epoch": 0.8177358954974595, + "grad_norm": 0.4469667375087738, + "learning_rate": 1.5956927075908013e-05, + "loss": 0.0671, + "step": 34884 + }, + { + "epoch": 0.8177593370722646, + "grad_norm": 0.21900928020477295, + "learning_rate": 1.5952935443125517e-05, + "loss": 0.0185, + "step": 34885 + }, + { + "epoch": 0.8177827786470695, + "grad_norm": 0.32042163610458374, + "learning_rate": 1.594894426638478e-05, + "loss": 0.0319, + "step": 34886 + }, + { + "epoch": 0.8178062202218745, + "grad_norm": 0.4003197252750397, + "learning_rate": 1.594495354570743e-05, + "loss": 0.094, + "step": 34887 + }, + { + "epoch": 0.8178296617966795, + "grad_norm": 0.10168997198343277, + "learning_rate": 1.5940963281115108e-05, + "loss": 0.0139, + "step": 34888 + }, + { + "epoch": 0.8178531033714845, + "grad_norm": 0.5775301456451416, + "learning_rate": 1.59369734726295e-05, + "loss": 0.1178, + "step": 34889 + }, + { + "epoch": 0.8178765449462895, + "grad_norm": 0.6406936645507812, + "learning_rate": 1.5932984120272242e-05, + "loss": 0.455, + "step": 34890 + }, + { + "epoch": 0.8178999865210945, + "grad_norm": 0.1513940393924713, + "learning_rate": 1.592899522406498e-05, + "loss": 0.0317, + "step": 34891 + }, + { + "epoch": 0.8179234280958995, + "grad_norm": 0.09948492795228958, + "learning_rate": 1.592500678402935e-05, + "loss": 0.0171, + "step": 34892 + }, + { + "epoch": 0.8179468696707045, + "grad_norm": 0.41949692368507385, + "learning_rate": 1.5921018800186992e-05, + "loss": 0.6966, + "step": 34893 + }, + { + "epoch": 0.8179703112455095, + "grad_norm": 0.5280149579048157, + "learning_rate": 1.5917031272559534e-05, + "loss": 0.076, + "step": 34894 + }, + { + "epoch": 0.8179937528203145, + "grad_norm": 0.34807780385017395, + "learning_rate": 1.5913044201168646e-05, + "loss": 0.0505, + "step": 34895 + }, + { + "epoch": 0.8180171943951194, + "grad_norm": 0.18001097440719604, + "learning_rate": 1.5909057586035945e-05, + "loss": 0.0088, + "step": 34896 + }, + { + "epoch": 0.8180406359699245, + "grad_norm": 0.46885955333709717, + "learning_rate": 1.5905071427183037e-05, + "loss": 0.1025, + "step": 34897 + }, + { + "epoch": 0.8180640775447294, + "grad_norm": 0.3081279695034027, + "learning_rate": 1.5901085724631602e-05, + "loss": 0.0612, + "step": 34898 + }, + { + "epoch": 0.8180875191195345, + "grad_norm": 0.7716780304908752, + "learning_rate": 1.5897100478403247e-05, + "loss": 0.1347, + "step": 34899 + }, + { + "epoch": 0.8181109606943394, + "grad_norm": 0.37888434529304504, + "learning_rate": 1.5893115688519587e-05, + "loss": 0.0653, + "step": 34900 + }, + { + "epoch": 0.8181344022691445, + "grad_norm": 0.760593831539154, + "learning_rate": 1.588913135500222e-05, + "loss": 0.1137, + "step": 34901 + }, + { + "epoch": 0.8181578438439494, + "grad_norm": 0.1650351732969284, + "learning_rate": 1.5885147477872807e-05, + "loss": 0.0253, + "step": 34902 + }, + { + "epoch": 0.8181812854187545, + "grad_norm": 0.3647644817829132, + "learning_rate": 1.5881164057152963e-05, + "loss": 0.0437, + "step": 34903 + }, + { + "epoch": 0.8182047269935594, + "grad_norm": 0.4654847979545593, + "learning_rate": 1.587718109286428e-05, + "loss": 0.0734, + "step": 34904 + }, + { + "epoch": 0.8182281685683644, + "grad_norm": 0.2795773446559906, + "learning_rate": 1.587319858502838e-05, + "loss": 0.0527, + "step": 34905 + }, + { + "epoch": 0.8182516101431694, + "grad_norm": 0.4657801389694214, + "learning_rate": 1.5869216533666865e-05, + "loss": 0.4587, + "step": 34906 + }, + { + "epoch": 0.8182750517179744, + "grad_norm": 0.2122306078672409, + "learning_rate": 1.5865234938801332e-05, + "loss": 0.0233, + "step": 34907 + }, + { + "epoch": 0.8182984932927794, + "grad_norm": 0.38572147488594055, + "learning_rate": 1.5861253800453414e-05, + "loss": 0.0181, + "step": 34908 + }, + { + "epoch": 0.8183219348675844, + "grad_norm": 0.3105544447898865, + "learning_rate": 1.585727311864469e-05, + "loss": 0.0506, + "step": 34909 + }, + { + "epoch": 0.8183453764423894, + "grad_norm": 0.8735084533691406, + "learning_rate": 1.5853292893396777e-05, + "loss": 0.1389, + "step": 34910 + }, + { + "epoch": 0.8183688180171944, + "grad_norm": 0.3381781578063965, + "learning_rate": 1.584931312473127e-05, + "loss": 0.0317, + "step": 34911 + }, + { + "epoch": 0.8183922595919993, + "grad_norm": 0.3408201336860657, + "learning_rate": 1.5845333812669748e-05, + "loss": 0.0442, + "step": 34912 + }, + { + "epoch": 0.8184157011668044, + "grad_norm": 0.2963751554489136, + "learning_rate": 1.5841354957233812e-05, + "loss": 0.0659, + "step": 34913 + }, + { + "epoch": 0.8184391427416093, + "grad_norm": 0.4168306291103363, + "learning_rate": 1.5837376558445037e-05, + "loss": 0.0225, + "step": 34914 + }, + { + "epoch": 0.8184625843164144, + "grad_norm": 0.48378705978393555, + "learning_rate": 1.5833398616325036e-05, + "loss": 0.0768, + "step": 34915 + }, + { + "epoch": 0.8184860258912193, + "grad_norm": 0.38394251465797424, + "learning_rate": 1.5829421130895385e-05, + "loss": 0.052, + "step": 34916 + }, + { + "epoch": 0.8185094674660244, + "grad_norm": 0.1560477912425995, + "learning_rate": 1.5825444102177665e-05, + "loss": 0.0382, + "step": 34917 + }, + { + "epoch": 0.8185329090408293, + "grad_norm": 0.4734629988670349, + "learning_rate": 1.5821467530193447e-05, + "loss": 0.0953, + "step": 34918 + }, + { + "epoch": 0.8185563506156344, + "grad_norm": 0.3039783835411072, + "learning_rate": 1.581749141496429e-05, + "loss": 0.0605, + "step": 34919 + }, + { + "epoch": 0.8185797921904393, + "grad_norm": 0.46955016255378723, + "learning_rate": 1.581351575651182e-05, + "loss": 0.0794, + "step": 34920 + }, + { + "epoch": 0.8186032337652444, + "grad_norm": 0.35877132415771484, + "learning_rate": 1.580954055485755e-05, + "loss": 0.077, + "step": 34921 + }, + { + "epoch": 0.8186266753400493, + "grad_norm": 0.34365779161453247, + "learning_rate": 1.5805565810023114e-05, + "loss": 0.0429, + "step": 34922 + }, + { + "epoch": 0.8186501169148543, + "grad_norm": 0.2280755639076233, + "learning_rate": 1.5801591522030034e-05, + "loss": 0.0189, + "step": 34923 + }, + { + "epoch": 0.8186735584896593, + "grad_norm": 0.3003571629524231, + "learning_rate": 1.5797617690899892e-05, + "loss": 0.0696, + "step": 34924 + }, + { + "epoch": 0.8186970000644643, + "grad_norm": 0.7832307815551758, + "learning_rate": 1.5793644316654242e-05, + "loss": 0.1267, + "step": 34925 + }, + { + "epoch": 0.8187204416392694, + "grad_norm": 0.4395759105682373, + "learning_rate": 1.578967139931462e-05, + "loss": 0.0819, + "step": 34926 + }, + { + "epoch": 0.8187438832140743, + "grad_norm": 0.4284491539001465, + "learning_rate": 1.5785698938902637e-05, + "loss": 0.0497, + "step": 34927 + }, + { + "epoch": 0.8187673247888794, + "grad_norm": 0.3571758270263672, + "learning_rate": 1.5781726935439822e-05, + "loss": 0.0657, + "step": 34928 + }, + { + "epoch": 0.8187907663636843, + "grad_norm": 0.3812790811061859, + "learning_rate": 1.5777755388947724e-05, + "loss": 0.0592, + "step": 34929 + }, + { + "epoch": 0.8188142079384894, + "grad_norm": 0.508395254611969, + "learning_rate": 1.577378429944788e-05, + "loss": 0.0675, + "step": 34930 + }, + { + "epoch": 0.8188376495132943, + "grad_norm": 0.2896510660648346, + "learning_rate": 1.5769813666961862e-05, + "loss": 0.0336, + "step": 34931 + }, + { + "epoch": 0.8188610910880993, + "grad_norm": 0.3016001880168915, + "learning_rate": 1.5765843491511178e-05, + "loss": 0.036, + "step": 34932 + }, + { + "epoch": 0.8188845326629043, + "grad_norm": 0.10861867666244507, + "learning_rate": 1.5761873773117398e-05, + "loss": 0.0209, + "step": 34933 + }, + { + "epoch": 0.8189079742377093, + "grad_norm": 0.1557772159576416, + "learning_rate": 1.575790451180208e-05, + "loss": 0.0214, + "step": 34934 + }, + { + "epoch": 0.8189314158125143, + "grad_norm": 0.6432322263717651, + "learning_rate": 1.575393570758674e-05, + "loss": 0.0998, + "step": 34935 + }, + { + "epoch": 0.8189548573873193, + "grad_norm": 0.5138429999351501, + "learning_rate": 1.5749967360492912e-05, + "loss": 0.1003, + "step": 34936 + }, + { + "epoch": 0.8189782989621243, + "grad_norm": 0.2531256079673767, + "learning_rate": 1.574599947054214e-05, + "loss": 0.0129, + "step": 34937 + }, + { + "epoch": 0.8190017405369293, + "grad_norm": 0.42871251702308655, + "learning_rate": 1.5742032037755938e-05, + "loss": 0.0486, + "step": 34938 + }, + { + "epoch": 0.8190251821117343, + "grad_norm": 0.15477211773395538, + "learning_rate": 1.5738065062155815e-05, + "loss": 0.0362, + "step": 34939 + }, + { + "epoch": 0.8190486236865393, + "grad_norm": 1.3518213033676147, + "learning_rate": 1.573409854376334e-05, + "loss": 0.148, + "step": 34940 + }, + { + "epoch": 0.8190720652613442, + "grad_norm": 0.22689104080200195, + "learning_rate": 1.573013248260001e-05, + "loss": 0.0305, + "step": 34941 + }, + { + "epoch": 0.8190955068361493, + "grad_norm": 0.3185727000236511, + "learning_rate": 1.5726166878687364e-05, + "loss": 0.0623, + "step": 34942 + }, + { + "epoch": 0.8191189484109542, + "grad_norm": 0.07160449028015137, + "learning_rate": 1.5722201732046903e-05, + "loss": 0.0133, + "step": 34943 + }, + { + "epoch": 0.8191423899857593, + "grad_norm": 0.25172051787376404, + "learning_rate": 1.5718237042700146e-05, + "loss": 0.0428, + "step": 34944 + }, + { + "epoch": 0.8191658315605642, + "grad_norm": 0.5028994083404541, + "learning_rate": 1.5714272810668586e-05, + "loss": 0.0877, + "step": 34945 + }, + { + "epoch": 0.8191892731353693, + "grad_norm": 0.15038295090198517, + "learning_rate": 1.5710309035973736e-05, + "loss": 0.0293, + "step": 34946 + }, + { + "epoch": 0.8192127147101742, + "grad_norm": 0.2574305534362793, + "learning_rate": 1.570634571863715e-05, + "loss": 0.0364, + "step": 34947 + }, + { + "epoch": 0.8192361562849793, + "grad_norm": 0.36162465810775757, + "learning_rate": 1.570238285868031e-05, + "loss": 0.0473, + "step": 34948 + }, + { + "epoch": 0.8192595978597842, + "grad_norm": 0.4215855896472931, + "learning_rate": 1.5698420456124698e-05, + "loss": 0.0437, + "step": 34949 + }, + { + "epoch": 0.8192830394345892, + "grad_norm": 0.13284504413604736, + "learning_rate": 1.5694458510991838e-05, + "loss": 0.0162, + "step": 34950 + }, + { + "epoch": 0.8193064810093942, + "grad_norm": 0.13827258348464966, + "learning_rate": 1.5690497023303208e-05, + "loss": 0.016, + "step": 34951 + }, + { + "epoch": 0.8193299225841992, + "grad_norm": 0.30020564794540405, + "learning_rate": 1.5686535993080298e-05, + "loss": 0.0431, + "step": 34952 + }, + { + "epoch": 0.8193533641590042, + "grad_norm": 0.2117554098367691, + "learning_rate": 1.5682575420344626e-05, + "loss": 0.0445, + "step": 34953 + }, + { + "epoch": 0.8193768057338092, + "grad_norm": 0.5890825986862183, + "learning_rate": 1.5678615305117683e-05, + "loss": 0.0585, + "step": 34954 + }, + { + "epoch": 0.8194002473086142, + "grad_norm": 0.4291232228279114, + "learning_rate": 1.5674655647420934e-05, + "loss": 0.0644, + "step": 34955 + }, + { + "epoch": 0.8194236888834192, + "grad_norm": 0.5400636792182922, + "learning_rate": 1.5670696447275878e-05, + "loss": 0.7469, + "step": 34956 + }, + { + "epoch": 0.8194471304582241, + "grad_norm": 0.4454038143157959, + "learning_rate": 1.566673770470397e-05, + "loss": 0.0461, + "step": 34957 + }, + { + "epoch": 0.8194705720330292, + "grad_norm": 0.6462815403938293, + "learning_rate": 1.5662779419726738e-05, + "loss": 0.1061, + "step": 34958 + }, + { + "epoch": 0.8194940136078341, + "grad_norm": 0.4371855556964874, + "learning_rate": 1.5658821592365614e-05, + "loss": 0.0506, + "step": 34959 + }, + { + "epoch": 0.8195174551826392, + "grad_norm": 0.4295330047607422, + "learning_rate": 1.565486422264212e-05, + "loss": 0.0891, + "step": 34960 + }, + { + "epoch": 0.8195408967574441, + "grad_norm": 0.5000329613685608, + "learning_rate": 1.56509073105777e-05, + "loss": 0.0766, + "step": 34961 + }, + { + "epoch": 0.8195643383322492, + "grad_norm": 0.18193958699703217, + "learning_rate": 1.5646950856193833e-05, + "loss": 0.0327, + "step": 34962 + }, + { + "epoch": 0.8195877799070541, + "grad_norm": 0.28287988901138306, + "learning_rate": 1.5642994859511985e-05, + "loss": 0.0299, + "step": 34963 + }, + { + "epoch": 0.8196112214818592, + "grad_norm": 0.452303409576416, + "learning_rate": 1.5639039320553595e-05, + "loss": 0.1042, + "step": 34964 + }, + { + "epoch": 0.8196346630566641, + "grad_norm": 0.6186293363571167, + "learning_rate": 1.563508423934017e-05, + "loss": 0.0548, + "step": 34965 + }, + { + "epoch": 0.8196581046314692, + "grad_norm": 0.3538934290409088, + "learning_rate": 1.5631129615893146e-05, + "loss": 0.0538, + "step": 34966 + }, + { + "epoch": 0.8196815462062741, + "grad_norm": 0.47886601090431213, + "learning_rate": 1.562717545023399e-05, + "loss": 0.0994, + "step": 34967 + }, + { + "epoch": 0.8197049877810791, + "grad_norm": 0.4190092384815216, + "learning_rate": 1.562322174238415e-05, + "loss": 0.1003, + "step": 34968 + }, + { + "epoch": 0.8197284293558841, + "grad_norm": 0.4221567213535309, + "learning_rate": 1.561926849236508e-05, + "loss": 0.0641, + "step": 34969 + }, + { + "epoch": 0.8197518709306891, + "grad_norm": 0.05848166346549988, + "learning_rate": 1.5615315700198207e-05, + "loss": 0.0061, + "step": 34970 + }, + { + "epoch": 0.8197753125054941, + "grad_norm": 0.673250675201416, + "learning_rate": 1.5611363365905006e-05, + "loss": 0.1498, + "step": 34971 + }, + { + "epoch": 0.8197987540802991, + "grad_norm": 0.34006208181381226, + "learning_rate": 1.560741148950695e-05, + "loss": 0.0651, + "step": 34972 + }, + { + "epoch": 0.8198221956551041, + "grad_norm": 0.389114648103714, + "learning_rate": 1.5603460071025445e-05, + "loss": 0.0884, + "step": 34973 + }, + { + "epoch": 0.8198456372299091, + "grad_norm": 0.3908058702945709, + "learning_rate": 1.5599509110481946e-05, + "loss": 0.0589, + "step": 34974 + }, + { + "epoch": 0.819869078804714, + "grad_norm": 0.49206116795539856, + "learning_rate": 1.5595558607897876e-05, + "loss": 0.066, + "step": 34975 + }, + { + "epoch": 0.8198925203795191, + "grad_norm": 0.37165403366088867, + "learning_rate": 1.559160856329468e-05, + "loss": 0.0511, + "step": 34976 + }, + { + "epoch": 0.8199159619543241, + "grad_norm": 0.5294396877288818, + "learning_rate": 1.5587658976693764e-05, + "loss": 0.6955, + "step": 34977 + }, + { + "epoch": 0.8199394035291291, + "grad_norm": 0.5180808305740356, + "learning_rate": 1.5583709848116613e-05, + "loss": 0.0634, + "step": 34978 + }, + { + "epoch": 0.8199628451039341, + "grad_norm": 0.4868365526199341, + "learning_rate": 1.5579761177584617e-05, + "loss": 0.0796, + "step": 34979 + }, + { + "epoch": 0.8199862866787391, + "grad_norm": 0.6374804377555847, + "learning_rate": 1.557581296511922e-05, + "loss": 0.4061, + "step": 34980 + }, + { + "epoch": 0.8200097282535441, + "grad_norm": 0.2224045842885971, + "learning_rate": 1.557186521074183e-05, + "loss": 0.0404, + "step": 34981 + }, + { + "epoch": 0.8200331698283491, + "grad_norm": 0.2107713669538498, + "learning_rate": 1.556791791447385e-05, + "loss": 0.0377, + "step": 34982 + }, + { + "epoch": 0.8200566114031541, + "grad_norm": 0.4783908724784851, + "learning_rate": 1.5563971076336747e-05, + "loss": 0.0325, + "step": 34983 + }, + { + "epoch": 0.820080052977959, + "grad_norm": 0.9068606495857239, + "learning_rate": 1.5560024696351894e-05, + "loss": 0.0769, + "step": 34984 + }, + { + "epoch": 0.8201034945527641, + "grad_norm": 0.17811083793640137, + "learning_rate": 1.555607877454074e-05, + "loss": 0.041, + "step": 34985 + }, + { + "epoch": 0.820126936127569, + "grad_norm": 0.5034117698669434, + "learning_rate": 1.5552133310924667e-05, + "loss": 0.0977, + "step": 34986 + }, + { + "epoch": 0.8201503777023741, + "grad_norm": 0.20438715815544128, + "learning_rate": 1.55481883055251e-05, + "loss": 0.0386, + "step": 34987 + }, + { + "epoch": 0.820173819277179, + "grad_norm": 0.43614456057548523, + "learning_rate": 1.554424375836344e-05, + "loss": 0.1054, + "step": 34988 + }, + { + "epoch": 0.8201972608519841, + "grad_norm": 0.5609239935874939, + "learning_rate": 1.554029966946109e-05, + "loss": 0.1036, + "step": 34989 + }, + { + "epoch": 0.820220702426789, + "grad_norm": 0.20536504685878754, + "learning_rate": 1.5536356038839417e-05, + "loss": 0.0299, + "step": 34990 + }, + { + "epoch": 0.8202441440015941, + "grad_norm": 0.10127836465835571, + "learning_rate": 1.5532412866519886e-05, + "loss": 0.0098, + "step": 34991 + }, + { + "epoch": 0.820267585576399, + "grad_norm": 0.3398098349571228, + "learning_rate": 1.5528470152523844e-05, + "loss": 0.0431, + "step": 34992 + }, + { + "epoch": 0.820291027151204, + "grad_norm": 0.5503886342048645, + "learning_rate": 1.552452789687271e-05, + "loss": 0.6331, + "step": 34993 + }, + { + "epoch": 0.820314468726009, + "grad_norm": 0.1168818473815918, + "learning_rate": 1.5520586099587832e-05, + "loss": 0.0044, + "step": 34994 + }, + { + "epoch": 0.820337910300814, + "grad_norm": 0.4050067067146301, + "learning_rate": 1.5516644760690658e-05, + "loss": 0.0408, + "step": 34995 + }, + { + "epoch": 0.820361351875619, + "grad_norm": 0.4649584889411926, + "learning_rate": 1.551270388020254e-05, + "loss": 0.05, + "step": 34996 + }, + { + "epoch": 0.820384793450424, + "grad_norm": 0.5553595423698425, + "learning_rate": 1.550876345814485e-05, + "loss": 0.0846, + "step": 34997 + }, + { + "epoch": 0.820408235025229, + "grad_norm": 0.42533761262893677, + "learning_rate": 1.5504823494539e-05, + "loss": 0.0627, + "step": 34998 + }, + { + "epoch": 0.820431676600034, + "grad_norm": 0.5251067280769348, + "learning_rate": 1.5500883989406358e-05, + "loss": 0.0983, + "step": 34999 + }, + { + "epoch": 0.820455118174839, + "grad_norm": 0.2931448817253113, + "learning_rate": 1.5496944942768298e-05, + "loss": 0.0431, + "step": 35000 + }, + { + "epoch": 0.820478559749644, + "grad_norm": 0.667325496673584, + "learning_rate": 1.549300635464619e-05, + "loss": 0.0818, + "step": 35001 + }, + { + "epoch": 0.820502001324449, + "grad_norm": 0.6882172226905823, + "learning_rate": 1.5489068225061377e-05, + "loss": 0.1436, + "step": 35002 + }, + { + "epoch": 0.820525442899254, + "grad_norm": 0.8191399574279785, + "learning_rate": 1.5485130554035286e-05, + "loss": 0.1406, + "step": 35003 + }, + { + "epoch": 0.8205488844740589, + "grad_norm": 0.3284049332141876, + "learning_rate": 1.548119334158925e-05, + "loss": 0.0439, + "step": 35004 + }, + { + "epoch": 0.820572326048864, + "grad_norm": 0.38428181409835815, + "learning_rate": 1.5477256587744627e-05, + "loss": 0.0609, + "step": 35005 + }, + { + "epoch": 0.8205957676236689, + "grad_norm": 0.12899306416511536, + "learning_rate": 1.5473320292522775e-05, + "loss": 0.0248, + "step": 35006 + }, + { + "epoch": 0.820619209198474, + "grad_norm": 0.4850406348705292, + "learning_rate": 1.546938445594508e-05, + "loss": 0.1044, + "step": 35007 + }, + { + "epoch": 0.8206426507732789, + "grad_norm": 0.4817088842391968, + "learning_rate": 1.546544907803289e-05, + "loss": 0.0419, + "step": 35008 + }, + { + "epoch": 0.820666092348084, + "grad_norm": 0.4186793565750122, + "learning_rate": 1.5461514158807523e-05, + "loss": 0.0815, + "step": 35009 + }, + { + "epoch": 0.8206895339228889, + "grad_norm": 0.44027793407440186, + "learning_rate": 1.545757969829037e-05, + "loss": 0.0737, + "step": 35010 + }, + { + "epoch": 0.820712975497694, + "grad_norm": 0.4125116169452667, + "learning_rate": 1.5453645696502782e-05, + "loss": 0.0783, + "step": 35011 + }, + { + "epoch": 0.8207364170724989, + "grad_norm": 0.42500272393226624, + "learning_rate": 1.5449712153466078e-05, + "loss": 0.026, + "step": 35012 + }, + { + "epoch": 0.8207598586473039, + "grad_norm": 0.29720214009284973, + "learning_rate": 1.5445779069201626e-05, + "loss": 0.0319, + "step": 35013 + }, + { + "epoch": 0.8207833002221089, + "grad_norm": 0.4623205363750458, + "learning_rate": 1.5441846443730746e-05, + "loss": 0.0598, + "step": 35014 + }, + { + "epoch": 0.8208067417969139, + "grad_norm": 0.5576539635658264, + "learning_rate": 1.5437914277074762e-05, + "loss": 0.0865, + "step": 35015 + }, + { + "epoch": 0.8208301833717189, + "grad_norm": 0.08697790652513504, + "learning_rate": 1.5433982569255057e-05, + "loss": 0.0176, + "step": 35016 + }, + { + "epoch": 0.8208536249465239, + "grad_norm": 0.3914484977722168, + "learning_rate": 1.5430051320292938e-05, + "loss": 0.0868, + "step": 35017 + }, + { + "epoch": 0.8208770665213289, + "grad_norm": 0.5462064743041992, + "learning_rate": 1.542612053020971e-05, + "loss": 0.0824, + "step": 35018 + }, + { + "epoch": 0.8209005080961339, + "grad_norm": 0.4403412938117981, + "learning_rate": 1.5422190199026765e-05, + "loss": 0.0623, + "step": 35019 + }, + { + "epoch": 0.8209239496709388, + "grad_norm": 0.06042490527033806, + "learning_rate": 1.5418260326765387e-05, + "loss": 0.0107, + "step": 35020 + }, + { + "epoch": 0.8209473912457439, + "grad_norm": 0.5491812229156494, + "learning_rate": 1.5414330913446905e-05, + "loss": 0.0821, + "step": 35021 + }, + { + "epoch": 0.8209708328205488, + "grad_norm": 0.5824021100997925, + "learning_rate": 1.5410401959092625e-05, + "loss": 0.0741, + "step": 35022 + }, + { + "epoch": 0.8209942743953539, + "grad_norm": 0.4405493140220642, + "learning_rate": 1.5406473463723903e-05, + "loss": 0.058, + "step": 35023 + }, + { + "epoch": 0.8210177159701588, + "grad_norm": 0.0565822534263134, + "learning_rate": 1.5402545427362037e-05, + "loss": 0.0073, + "step": 35024 + }, + { + "epoch": 0.8210411575449639, + "grad_norm": 0.15461181104183197, + "learning_rate": 1.5398617850028342e-05, + "loss": 0.0221, + "step": 35025 + }, + { + "epoch": 0.8210645991197688, + "grad_norm": 0.1464395821094513, + "learning_rate": 1.539469073174411e-05, + "loss": 0.0155, + "step": 35026 + }, + { + "epoch": 0.8210880406945739, + "grad_norm": 0.45203107595443726, + "learning_rate": 1.5390764072530682e-05, + "loss": 0.091, + "step": 35027 + }, + { + "epoch": 0.8211114822693789, + "grad_norm": 0.43884986639022827, + "learning_rate": 1.538683787240931e-05, + "loss": 0.1177, + "step": 35028 + }, + { + "epoch": 0.8211349238441839, + "grad_norm": 0.2979639172554016, + "learning_rate": 1.538291213140136e-05, + "loss": 0.3234, + "step": 35029 + }, + { + "epoch": 0.8211583654189889, + "grad_norm": 0.15701617300510406, + "learning_rate": 1.537898684952812e-05, + "loss": 0.0173, + "step": 35030 + }, + { + "epoch": 0.8211818069937938, + "grad_norm": 0.38253235816955566, + "learning_rate": 1.537506202681084e-05, + "loss": 0.0717, + "step": 35031 + }, + { + "epoch": 0.8212052485685989, + "grad_norm": 0.25929149985313416, + "learning_rate": 1.5371137663270885e-05, + "loss": 0.0327, + "step": 35032 + }, + { + "epoch": 0.8212286901434038, + "grad_norm": 0.16659849882125854, + "learning_rate": 1.536721375892951e-05, + "loss": 0.0265, + "step": 35033 + }, + { + "epoch": 0.8212521317182089, + "grad_norm": 0.287371963262558, + "learning_rate": 1.536329031380802e-05, + "loss": 0.03, + "step": 35034 + }, + { + "epoch": 0.8212755732930138, + "grad_norm": 0.44752347469329834, + "learning_rate": 1.5359367327927664e-05, + "loss": 0.0683, + "step": 35035 + }, + { + "epoch": 0.8212990148678189, + "grad_norm": 0.156940296292305, + "learning_rate": 1.5355444801309793e-05, + "loss": 0.0279, + "step": 35036 + }, + { + "epoch": 0.8213224564426238, + "grad_norm": 0.5438907742500305, + "learning_rate": 1.5351522733975655e-05, + "loss": 0.1229, + "step": 35037 + }, + { + "epoch": 0.8213458980174289, + "grad_norm": 0.23774738609790802, + "learning_rate": 1.534760112594653e-05, + "loss": 0.0304, + "step": 35038 + }, + { + "epoch": 0.8213693395922338, + "grad_norm": 0.6003788113594055, + "learning_rate": 1.5343679977243706e-05, + "loss": 0.0889, + "step": 35039 + }, + { + "epoch": 0.8213927811670388, + "grad_norm": 0.5274109840393066, + "learning_rate": 1.5339759287888433e-05, + "loss": 0.0844, + "step": 35040 + }, + { + "epoch": 0.8214162227418438, + "grad_norm": 0.10704568773508072, + "learning_rate": 1.5335839057902034e-05, + "loss": 0.0228, + "step": 35041 + }, + { + "epoch": 0.8214396643166488, + "grad_norm": 0.6532295942306519, + "learning_rate": 1.5331919287305753e-05, + "loss": 0.1424, + "step": 35042 + }, + { + "epoch": 0.8214631058914538, + "grad_norm": 0.20700430870056152, + "learning_rate": 1.5327999976120823e-05, + "loss": 0.0266, + "step": 35043 + }, + { + "epoch": 0.8214865474662588, + "grad_norm": 0.6814290881156921, + "learning_rate": 1.5324081124368573e-05, + "loss": 0.0733, + "step": 35044 + }, + { + "epoch": 0.8215099890410638, + "grad_norm": 0.21732620894908905, + "learning_rate": 1.5320162732070243e-05, + "loss": 0.0454, + "step": 35045 + }, + { + "epoch": 0.8215334306158688, + "grad_norm": 0.7211192846298218, + "learning_rate": 1.5316244799247093e-05, + "loss": 0.1093, + "step": 35046 + }, + { + "epoch": 0.8215568721906737, + "grad_norm": 0.1787765473127365, + "learning_rate": 1.531232732592035e-05, + "loss": 0.0247, + "step": 35047 + }, + { + "epoch": 0.8215803137654788, + "grad_norm": 0.4185822606086731, + "learning_rate": 1.530841031211133e-05, + "loss": 0.0916, + "step": 35048 + }, + { + "epoch": 0.8216037553402837, + "grad_norm": 0.39911219477653503, + "learning_rate": 1.5304493757841255e-05, + "loss": 0.0392, + "step": 35049 + }, + { + "epoch": 0.8216271969150888, + "grad_norm": 0.13284555077552795, + "learning_rate": 1.5300577663131375e-05, + "loss": 0.0138, + "step": 35050 + }, + { + "epoch": 0.8216506384898937, + "grad_norm": 0.6705722808837891, + "learning_rate": 1.529666202800295e-05, + "loss": 0.1042, + "step": 35051 + }, + { + "epoch": 0.8216740800646988, + "grad_norm": 0.14033980667591095, + "learning_rate": 1.5292746852477214e-05, + "loss": 0.0196, + "step": 35052 + }, + { + "epoch": 0.8216975216395037, + "grad_norm": 0.5840402245521545, + "learning_rate": 1.5288832136575392e-05, + "loss": 0.127, + "step": 35053 + }, + { + "epoch": 0.8217209632143088, + "grad_norm": 0.9416020512580872, + "learning_rate": 1.5284917880318774e-05, + "loss": 0.1003, + "step": 35054 + }, + { + "epoch": 0.8217444047891137, + "grad_norm": 0.4115184247493744, + "learning_rate": 1.5281004083728545e-05, + "loss": 0.063, + "step": 35055 + }, + { + "epoch": 0.8217678463639188, + "grad_norm": 0.32033005356788635, + "learning_rate": 1.5277090746825994e-05, + "loss": 0.0507, + "step": 35056 + }, + { + "epoch": 0.8217912879387237, + "grad_norm": 0.18280357122421265, + "learning_rate": 1.5273177869632327e-05, + "loss": 0.0309, + "step": 35057 + }, + { + "epoch": 0.8218147295135287, + "grad_norm": 0.10883274674415588, + "learning_rate": 1.526926545216879e-05, + "loss": 0.0136, + "step": 35058 + }, + { + "epoch": 0.8218381710883337, + "grad_norm": 0.8578455448150635, + "learning_rate": 1.526535349445659e-05, + "loss": 0.1673, + "step": 35059 + }, + { + "epoch": 0.8218616126631387, + "grad_norm": 0.4317772686481476, + "learning_rate": 1.5261441996516957e-05, + "loss": 0.0469, + "step": 35060 + }, + { + "epoch": 0.8218850542379437, + "grad_norm": 0.4849815368652344, + "learning_rate": 1.5257530958371125e-05, + "loss": 0.5553, + "step": 35061 + }, + { + "epoch": 0.8219084958127487, + "grad_norm": 0.14480598270893097, + "learning_rate": 1.5253620380040323e-05, + "loss": 0.0246, + "step": 35062 + }, + { + "epoch": 0.8219319373875537, + "grad_norm": 0.29042157530784607, + "learning_rate": 1.524971026154577e-05, + "loss": 0.0358, + "step": 35063 + }, + { + "epoch": 0.8219553789623587, + "grad_norm": 0.5385611057281494, + "learning_rate": 1.5245800602908656e-05, + "loss": 0.1638, + "step": 35064 + }, + { + "epoch": 0.8219788205371636, + "grad_norm": 0.5010958909988403, + "learning_rate": 1.5241891404150222e-05, + "loss": 0.0449, + "step": 35065 + }, + { + "epoch": 0.8220022621119687, + "grad_norm": 0.6154279112815857, + "learning_rate": 1.5237982665291638e-05, + "loss": 0.1475, + "step": 35066 + }, + { + "epoch": 0.8220257036867736, + "grad_norm": 0.3143951892852783, + "learning_rate": 1.5234074386354147e-05, + "loss": 0.0652, + "step": 35067 + }, + { + "epoch": 0.8220491452615787, + "grad_norm": 0.18184177577495575, + "learning_rate": 1.5230166567358972e-05, + "loss": 0.0325, + "step": 35068 + }, + { + "epoch": 0.8220725868363836, + "grad_norm": 0.1720850020647049, + "learning_rate": 1.5226259208327297e-05, + "loss": 0.0337, + "step": 35069 + }, + { + "epoch": 0.8220960284111887, + "grad_norm": 0.06023966521024704, + "learning_rate": 1.522235230928033e-05, + "loss": 0.0088, + "step": 35070 + }, + { + "epoch": 0.8221194699859936, + "grad_norm": 0.5598320960998535, + "learning_rate": 1.521844587023925e-05, + "loss": 0.0763, + "step": 35071 + }, + { + "epoch": 0.8221429115607987, + "grad_norm": 0.33101001381874084, + "learning_rate": 1.5214539891225277e-05, + "loss": 0.0457, + "step": 35072 + }, + { + "epoch": 0.8221663531356036, + "grad_norm": 0.45286884903907776, + "learning_rate": 1.5210634372259568e-05, + "loss": 0.0522, + "step": 35073 + }, + { + "epoch": 0.8221897947104087, + "grad_norm": 0.454717218875885, + "learning_rate": 1.5206729313363366e-05, + "loss": 0.1057, + "step": 35074 + }, + { + "epoch": 0.8222132362852136, + "grad_norm": 0.33091890811920166, + "learning_rate": 1.520282471455784e-05, + "loss": 0.04, + "step": 35075 + }, + { + "epoch": 0.8222366778600186, + "grad_norm": 0.24712103605270386, + "learning_rate": 1.5198920575864163e-05, + "loss": 0.0648, + "step": 35076 + }, + { + "epoch": 0.8222601194348236, + "grad_norm": 0.3340274691581726, + "learning_rate": 1.5195016897303527e-05, + "loss": 0.0544, + "step": 35077 + }, + { + "epoch": 0.8222835610096286, + "grad_norm": 0.27793100476264954, + "learning_rate": 1.5191113678897095e-05, + "loss": 0.0492, + "step": 35078 + }, + { + "epoch": 0.8223070025844337, + "grad_norm": 0.4426319897174835, + "learning_rate": 1.5187210920666062e-05, + "loss": 0.0927, + "step": 35079 + }, + { + "epoch": 0.8223304441592386, + "grad_norm": 0.26347431540489197, + "learning_rate": 1.5183308622631632e-05, + "loss": 0.0362, + "step": 35080 + }, + { + "epoch": 0.8223538857340437, + "grad_norm": 0.47553950548171997, + "learning_rate": 1.5179406784814954e-05, + "loss": 0.0929, + "step": 35081 + }, + { + "epoch": 0.8223773273088486, + "grad_norm": 0.08403049409389496, + "learning_rate": 1.5175505407237201e-05, + "loss": 0.0113, + "step": 35082 + }, + { + "epoch": 0.8224007688836537, + "grad_norm": 0.4034770131111145, + "learning_rate": 1.5171604489919533e-05, + "loss": 0.0444, + "step": 35083 + }, + { + "epoch": 0.8224242104584586, + "grad_norm": 0.2866774797439575, + "learning_rate": 1.5167704032883135e-05, + "loss": 0.0423, + "step": 35084 + }, + { + "epoch": 0.8224476520332636, + "grad_norm": 0.20662954449653625, + "learning_rate": 1.5163804036149131e-05, + "loss": 0.0279, + "step": 35085 + }, + { + "epoch": 0.8224710936080686, + "grad_norm": 1.0708191394805908, + "learning_rate": 1.515990449973873e-05, + "loss": 0.0717, + "step": 35086 + }, + { + "epoch": 0.8224945351828736, + "grad_norm": 0.4556792378425598, + "learning_rate": 1.515600542367308e-05, + "loss": 0.0859, + "step": 35087 + }, + { + "epoch": 0.8225179767576786, + "grad_norm": 0.6275018453598022, + "learning_rate": 1.5152106807973332e-05, + "loss": 0.1312, + "step": 35088 + }, + { + "epoch": 0.8225414183324836, + "grad_norm": 0.4710232615470886, + "learning_rate": 1.5148208652660633e-05, + "loss": 0.063, + "step": 35089 + }, + { + "epoch": 0.8225648599072886, + "grad_norm": 0.5413904190063477, + "learning_rate": 1.5144310957756148e-05, + "loss": 0.6477, + "step": 35090 + }, + { + "epoch": 0.8225883014820936, + "grad_norm": 0.17458081245422363, + "learning_rate": 1.5140413723280989e-05, + "loss": 0.032, + "step": 35091 + }, + { + "epoch": 0.8226117430568985, + "grad_norm": 0.5330232381820679, + "learning_rate": 1.513651694925633e-05, + "loss": 0.0811, + "step": 35092 + }, + { + "epoch": 0.8226351846317036, + "grad_norm": 1.7913795709609985, + "learning_rate": 1.5132620635703332e-05, + "loss": 0.1658, + "step": 35093 + }, + { + "epoch": 0.8226586262065085, + "grad_norm": 0.7171405553817749, + "learning_rate": 1.5128724782643133e-05, + "loss": 0.5163, + "step": 35094 + }, + { + "epoch": 0.8226820677813136, + "grad_norm": 0.17656031250953674, + "learning_rate": 1.512482939009685e-05, + "loss": 0.0121, + "step": 35095 + }, + { + "epoch": 0.8227055093561185, + "grad_norm": 0.5708425641059875, + "learning_rate": 1.5120934458085634e-05, + "loss": 0.1201, + "step": 35096 + }, + { + "epoch": 0.8227289509309236, + "grad_norm": 0.4426698386669159, + "learning_rate": 1.511703998663062e-05, + "loss": 0.0595, + "step": 35097 + }, + { + "epoch": 0.8227523925057285, + "grad_norm": 0.5450360774993896, + "learning_rate": 1.5113145975752906e-05, + "loss": 0.0878, + "step": 35098 + }, + { + "epoch": 0.8227758340805336, + "grad_norm": 0.3382877707481384, + "learning_rate": 1.5109252425473675e-05, + "loss": 0.0721, + "step": 35099 + }, + { + "epoch": 0.8227992756553385, + "grad_norm": 0.509229838848114, + "learning_rate": 1.5105359335814017e-05, + "loss": 0.0944, + "step": 35100 + }, + { + "epoch": 0.8228227172301436, + "grad_norm": 0.2604774236679077, + "learning_rate": 1.5101466706795075e-05, + "loss": 0.0293, + "step": 35101 + }, + { + "epoch": 0.8228461588049485, + "grad_norm": 0.3280622661113739, + "learning_rate": 1.5097574538437963e-05, + "loss": 0.0427, + "step": 35102 + }, + { + "epoch": 0.8228696003797535, + "grad_norm": 0.17867536842823029, + "learning_rate": 1.5093682830763789e-05, + "loss": 0.0338, + "step": 35103 + }, + { + "epoch": 0.8228930419545585, + "grad_norm": 0.1521870493888855, + "learning_rate": 1.5089791583793667e-05, + "loss": 0.0296, + "step": 35104 + }, + { + "epoch": 0.8229164835293635, + "grad_norm": 0.8297504186630249, + "learning_rate": 1.5085900797548714e-05, + "loss": 0.091, + "step": 35105 + }, + { + "epoch": 0.8229399251041685, + "grad_norm": 0.5510829091072083, + "learning_rate": 1.5082010472050079e-05, + "loss": 0.5515, + "step": 35106 + }, + { + "epoch": 0.8229633666789735, + "grad_norm": 0.5101467370986938, + "learning_rate": 1.5078120607318846e-05, + "loss": 0.5982, + "step": 35107 + }, + { + "epoch": 0.8229868082537785, + "grad_norm": 0.16477888822555542, + "learning_rate": 1.507423120337611e-05, + "loss": 0.0281, + "step": 35108 + }, + { + "epoch": 0.8230102498285835, + "grad_norm": 0.3309018611907959, + "learning_rate": 1.5070342260242987e-05, + "loss": 0.0446, + "step": 35109 + }, + { + "epoch": 0.8230336914033884, + "grad_norm": 0.48424118757247925, + "learning_rate": 1.506645377794056e-05, + "loss": 0.0808, + "step": 35110 + }, + { + "epoch": 0.8230571329781935, + "grad_norm": 0.28808727860450745, + "learning_rate": 1.5062565756489955e-05, + "loss": 0.0528, + "step": 35111 + }, + { + "epoch": 0.8230805745529984, + "grad_norm": 0.2674234211444855, + "learning_rate": 1.5058678195912268e-05, + "loss": 0.0367, + "step": 35112 + }, + { + "epoch": 0.8231040161278035, + "grad_norm": 0.15929079055786133, + "learning_rate": 1.5054791096228571e-05, + "loss": 0.0162, + "step": 35113 + }, + { + "epoch": 0.8231274577026084, + "grad_norm": 0.33306685090065, + "learning_rate": 1.5050904457459968e-05, + "loss": 0.0387, + "step": 35114 + }, + { + "epoch": 0.8231508992774135, + "grad_norm": 0.37482622265815735, + "learning_rate": 1.5047018279627546e-05, + "loss": 0.0415, + "step": 35115 + }, + { + "epoch": 0.8231743408522184, + "grad_norm": 0.43298545479774475, + "learning_rate": 1.504313256275237e-05, + "loss": 0.0645, + "step": 35116 + }, + { + "epoch": 0.8231977824270235, + "grad_norm": 0.6952533721923828, + "learning_rate": 1.5039247306855542e-05, + "loss": 0.1312, + "step": 35117 + }, + { + "epoch": 0.8232212240018284, + "grad_norm": 0.406323105096817, + "learning_rate": 1.5035362511958173e-05, + "loss": 0.0555, + "step": 35118 + }, + { + "epoch": 0.8232446655766334, + "grad_norm": 0.3592770993709564, + "learning_rate": 1.5031478178081315e-05, + "loss": 0.0618, + "step": 35119 + }, + { + "epoch": 0.8232681071514384, + "grad_norm": 0.23685650527477264, + "learning_rate": 1.5027594305246051e-05, + "loss": 0.022, + "step": 35120 + }, + { + "epoch": 0.8232915487262434, + "grad_norm": 0.04825371876358986, + "learning_rate": 1.5023710893473442e-05, + "loss": 0.0079, + "step": 35121 + }, + { + "epoch": 0.8233149903010484, + "grad_norm": 0.18831397593021393, + "learning_rate": 1.501982794278457e-05, + "loss": 0.0347, + "step": 35122 + }, + { + "epoch": 0.8233384318758534, + "grad_norm": 0.4042835235595703, + "learning_rate": 1.5015945453200475e-05, + "loss": 0.0574, + "step": 35123 + }, + { + "epoch": 0.8233618734506584, + "grad_norm": 0.10563317686319351, + "learning_rate": 1.5012063424742285e-05, + "loss": 0.0175, + "step": 35124 + }, + { + "epoch": 0.8233853150254634, + "grad_norm": 0.3877677023410797, + "learning_rate": 1.5008181857431015e-05, + "loss": 0.0537, + "step": 35125 + }, + { + "epoch": 0.8234087566002684, + "grad_norm": 0.16549205780029297, + "learning_rate": 1.5004300751287748e-05, + "loss": 0.0476, + "step": 35126 + }, + { + "epoch": 0.8234321981750734, + "grad_norm": 0.4910600781440735, + "learning_rate": 1.5000420106333524e-05, + "loss": 0.5247, + "step": 35127 + }, + { + "epoch": 0.8234556397498783, + "grad_norm": 0.655716061592102, + "learning_rate": 1.4996539922589392e-05, + "loss": 0.0717, + "step": 35128 + }, + { + "epoch": 0.8234790813246834, + "grad_norm": 0.6766155362129211, + "learning_rate": 1.4992660200076458e-05, + "loss": 0.1178, + "step": 35129 + }, + { + "epoch": 0.8235025228994884, + "grad_norm": 0.22545161843299866, + "learning_rate": 1.4988780938815706e-05, + "loss": 0.0335, + "step": 35130 + }, + { + "epoch": 0.8235259644742934, + "grad_norm": 0.2595843970775604, + "learning_rate": 1.498490213882825e-05, + "loss": 0.0294, + "step": 35131 + }, + { + "epoch": 0.8235494060490984, + "grad_norm": 0.15887528657913208, + "learning_rate": 1.4981023800135097e-05, + "loss": 0.0157, + "step": 35132 + }, + { + "epoch": 0.8235728476239034, + "grad_norm": 0.3232748210430145, + "learning_rate": 1.4977145922757297e-05, + "loss": 0.0218, + "step": 35133 + }, + { + "epoch": 0.8235962891987084, + "grad_norm": 0.25851213932037354, + "learning_rate": 1.4973268506715899e-05, + "loss": 0.0349, + "step": 35134 + }, + { + "epoch": 0.8236197307735134, + "grad_norm": 0.5286219716072083, + "learning_rate": 1.496939155203194e-05, + "loss": 0.0858, + "step": 35135 + }, + { + "epoch": 0.8236431723483184, + "grad_norm": 0.4528387784957886, + "learning_rate": 1.4965515058726431e-05, + "loss": 0.055, + "step": 35136 + }, + { + "epoch": 0.8236666139231233, + "grad_norm": 0.2926042377948761, + "learning_rate": 1.496163902682045e-05, + "loss": 0.0365, + "step": 35137 + }, + { + "epoch": 0.8236900554979284, + "grad_norm": 0.27552130818367004, + "learning_rate": 1.4957763456335006e-05, + "loss": 0.0325, + "step": 35138 + }, + { + "epoch": 0.8237134970727333, + "grad_norm": 0.2139909714460373, + "learning_rate": 1.4953888347291123e-05, + "loss": 0.0272, + "step": 35139 + }, + { + "epoch": 0.8237369386475384, + "grad_norm": 0.3849559426307678, + "learning_rate": 1.4950013699709808e-05, + "loss": 0.0665, + "step": 35140 + }, + { + "epoch": 0.8237603802223433, + "grad_norm": 0.554022490978241, + "learning_rate": 1.4946139513612145e-05, + "loss": 0.7185, + "step": 35141 + }, + { + "epoch": 0.8237838217971484, + "grad_norm": 0.3310728073120117, + "learning_rate": 1.4942265789019117e-05, + "loss": 0.0619, + "step": 35142 + }, + { + "epoch": 0.8238072633719533, + "grad_norm": 0.30187106132507324, + "learning_rate": 1.4938392525951716e-05, + "loss": 0.0332, + "step": 35143 + }, + { + "epoch": 0.8238307049467584, + "grad_norm": 0.521657407283783, + "learning_rate": 1.4934519724431018e-05, + "loss": 0.0798, + "step": 35144 + }, + { + "epoch": 0.8238541465215633, + "grad_norm": 0.6362862586975098, + "learning_rate": 1.4930647384478013e-05, + "loss": 0.1565, + "step": 35145 + }, + { + "epoch": 0.8238775880963684, + "grad_norm": 0.6894063353538513, + "learning_rate": 1.49267755061137e-05, + "loss": 0.11, + "step": 35146 + }, + { + "epoch": 0.8239010296711733, + "grad_norm": 0.34168511629104614, + "learning_rate": 1.4922904089359091e-05, + "loss": 0.0631, + "step": 35147 + }, + { + "epoch": 0.8239244712459783, + "grad_norm": 0.8271443247795105, + "learning_rate": 1.4919033134235184e-05, + "loss": 0.093, + "step": 35148 + }, + { + "epoch": 0.8239479128207833, + "grad_norm": 0.5286281704902649, + "learning_rate": 1.4915162640763015e-05, + "loss": 0.0353, + "step": 35149 + }, + { + "epoch": 0.8239713543955883, + "grad_norm": 0.19737400114536285, + "learning_rate": 1.4911292608963557e-05, + "loss": 0.0397, + "step": 35150 + }, + { + "epoch": 0.8239947959703933, + "grad_norm": 0.6005486845970154, + "learning_rate": 1.4907423038857826e-05, + "loss": 0.0755, + "step": 35151 + }, + { + "epoch": 0.8240182375451983, + "grad_norm": 0.4919731914997101, + "learning_rate": 1.4903553930466784e-05, + "loss": 0.102, + "step": 35152 + }, + { + "epoch": 0.8240416791200033, + "grad_norm": 0.14594928920269012, + "learning_rate": 1.4899685283811482e-05, + "loss": 0.0352, + "step": 35153 + }, + { + "epoch": 0.8240651206948083, + "grad_norm": 0.2149895578622818, + "learning_rate": 1.489581709891288e-05, + "loss": 0.0263, + "step": 35154 + }, + { + "epoch": 0.8240885622696132, + "grad_norm": 0.4044245183467865, + "learning_rate": 1.4891949375791937e-05, + "loss": 0.0514, + "step": 35155 + }, + { + "epoch": 0.8241120038444183, + "grad_norm": 0.27950599789619446, + "learning_rate": 1.4888082114469692e-05, + "loss": 0.0239, + "step": 35156 + }, + { + "epoch": 0.8241354454192232, + "grad_norm": 0.2730030417442322, + "learning_rate": 1.488421531496711e-05, + "loss": 0.0351, + "step": 35157 + }, + { + "epoch": 0.8241588869940283, + "grad_norm": 0.09320724755525589, + "learning_rate": 1.488034897730517e-05, + "loss": 0.01, + "step": 35158 + }, + { + "epoch": 0.8241823285688332, + "grad_norm": 0.6359819769859314, + "learning_rate": 1.4876483101504846e-05, + "loss": 0.0747, + "step": 35159 + }, + { + "epoch": 0.8242057701436383, + "grad_norm": 0.30457159876823425, + "learning_rate": 1.487261768758712e-05, + "loss": 0.0316, + "step": 35160 + }, + { + "epoch": 0.8242292117184432, + "grad_norm": 0.22769930958747864, + "learning_rate": 1.4868752735572943e-05, + "loss": 0.0325, + "step": 35161 + }, + { + "epoch": 0.8242526532932483, + "grad_norm": 0.06392965465784073, + "learning_rate": 1.4864888245483322e-05, + "loss": 0.009, + "step": 35162 + }, + { + "epoch": 0.8242760948680532, + "grad_norm": 0.3012886047363281, + "learning_rate": 1.4861024217339226e-05, + "loss": 0.0411, + "step": 35163 + }, + { + "epoch": 0.8242995364428582, + "grad_norm": 0.32612740993499756, + "learning_rate": 1.4857160651161572e-05, + "loss": 0.0462, + "step": 35164 + }, + { + "epoch": 0.8243229780176632, + "grad_norm": 0.825121283531189, + "learning_rate": 1.485329754697139e-05, + "loss": 0.5587, + "step": 35165 + }, + { + "epoch": 0.8243464195924682, + "grad_norm": 0.28192374110221863, + "learning_rate": 1.4849434904789595e-05, + "loss": 0.0547, + "step": 35166 + }, + { + "epoch": 0.8243698611672732, + "grad_norm": 0.43047034740448, + "learning_rate": 1.4845572724637168e-05, + "loss": 0.0412, + "step": 35167 + }, + { + "epoch": 0.8243933027420782, + "grad_norm": 0.38708043098449707, + "learning_rate": 1.4841711006535041e-05, + "loss": 0.0797, + "step": 35168 + }, + { + "epoch": 0.8244167443168832, + "grad_norm": 0.5911995768547058, + "learning_rate": 1.4837849750504206e-05, + "loss": 0.127, + "step": 35169 + }, + { + "epoch": 0.8244401858916882, + "grad_norm": 0.2617676854133606, + "learning_rate": 1.4833988956565593e-05, + "loss": 0.0271, + "step": 35170 + }, + { + "epoch": 0.8244636274664932, + "grad_norm": 0.24246850609779358, + "learning_rate": 1.4830128624740147e-05, + "loss": 0.0294, + "step": 35171 + }, + { + "epoch": 0.8244870690412982, + "grad_norm": 0.6020262241363525, + "learning_rate": 1.4826268755048811e-05, + "loss": 0.1167, + "step": 35172 + }, + { + "epoch": 0.8245105106161031, + "grad_norm": 0.3329002857208252, + "learning_rate": 1.4822409347512545e-05, + "loss": 0.0453, + "step": 35173 + }, + { + "epoch": 0.8245339521909082, + "grad_norm": 0.3182561695575714, + "learning_rate": 1.4818550402152253e-05, + "loss": 0.0499, + "step": 35174 + }, + { + "epoch": 0.8245573937657131, + "grad_norm": 0.5948124527931213, + "learning_rate": 1.4814691918988932e-05, + "loss": 0.1129, + "step": 35175 + }, + { + "epoch": 0.8245808353405182, + "grad_norm": 0.36089715361595154, + "learning_rate": 1.4810833898043464e-05, + "loss": 0.0622, + "step": 35176 + }, + { + "epoch": 0.8246042769153231, + "grad_norm": 0.8950420618057251, + "learning_rate": 1.480697633933682e-05, + "loss": 0.1628, + "step": 35177 + }, + { + "epoch": 0.8246277184901282, + "grad_norm": 0.23150020837783813, + "learning_rate": 1.4803119242889929e-05, + "loss": 0.0475, + "step": 35178 + }, + { + "epoch": 0.8246511600649331, + "grad_norm": 0.13591602444648743, + "learning_rate": 1.4799262608723697e-05, + "loss": 0.0276, + "step": 35179 + }, + { + "epoch": 0.8246746016397382, + "grad_norm": 0.24087032675743103, + "learning_rate": 1.4795406436859071e-05, + "loss": 0.0155, + "step": 35180 + }, + { + "epoch": 0.8246980432145432, + "grad_norm": 0.37761905789375305, + "learning_rate": 1.4791550727316938e-05, + "loss": 0.0869, + "step": 35181 + }, + { + "epoch": 0.8247214847893481, + "grad_norm": 0.4043181836605072, + "learning_rate": 1.4787695480118268e-05, + "loss": 0.0607, + "step": 35182 + }, + { + "epoch": 0.8247449263641532, + "grad_norm": 0.29986968636512756, + "learning_rate": 1.4783840695283968e-05, + "loss": 0.028, + "step": 35183 + }, + { + "epoch": 0.8247683679389581, + "grad_norm": 0.29376569390296936, + "learning_rate": 1.477998637283493e-05, + "loss": 0.0328, + "step": 35184 + }, + { + "epoch": 0.8247918095137632, + "grad_norm": 0.17942410707473755, + "learning_rate": 1.4776132512792096e-05, + "loss": 0.0263, + "step": 35185 + }, + { + "epoch": 0.8248152510885681, + "grad_norm": 0.4754871129989624, + "learning_rate": 1.4772279115176335e-05, + "loss": 0.0559, + "step": 35186 + }, + { + "epoch": 0.8248386926633732, + "grad_norm": 0.6919564008712769, + "learning_rate": 1.47684261800086e-05, + "loss": 0.5433, + "step": 35187 + }, + { + "epoch": 0.8248621342381781, + "grad_norm": 0.12034102529287338, + "learning_rate": 1.4764573707309793e-05, + "loss": 0.0197, + "step": 35188 + }, + { + "epoch": 0.8248855758129832, + "grad_norm": 0.5786453485488892, + "learning_rate": 1.4760721697100765e-05, + "loss": 0.1602, + "step": 35189 + }, + { + "epoch": 0.8249090173877881, + "grad_norm": 0.570415198802948, + "learning_rate": 1.4756870149402503e-05, + "loss": 0.7049, + "step": 35190 + }, + { + "epoch": 0.8249324589625932, + "grad_norm": 0.4232686758041382, + "learning_rate": 1.475301906423584e-05, + "loss": 0.0616, + "step": 35191 + }, + { + "epoch": 0.8249559005373981, + "grad_norm": 0.42444494366645813, + "learning_rate": 1.4749168441621708e-05, + "loss": 0.0783, + "step": 35192 + }, + { + "epoch": 0.8249793421122031, + "grad_norm": 0.4085855185985565, + "learning_rate": 1.4745318281580955e-05, + "loss": 0.0577, + "step": 35193 + }, + { + "epoch": 0.8250027836870081, + "grad_norm": 0.4111630320549011, + "learning_rate": 1.4741468584134532e-05, + "loss": 0.043, + "step": 35194 + }, + { + "epoch": 0.8250262252618131, + "grad_norm": 0.26558852195739746, + "learning_rate": 1.4737619349303289e-05, + "loss": 0.058, + "step": 35195 + }, + { + "epoch": 0.8250496668366181, + "grad_norm": 0.6342557668685913, + "learning_rate": 1.4733770577108118e-05, + "loss": 0.7169, + "step": 35196 + }, + { + "epoch": 0.8250731084114231, + "grad_norm": 0.4154902696609497, + "learning_rate": 1.4729922267569918e-05, + "loss": 0.0508, + "step": 35197 + }, + { + "epoch": 0.8250965499862281, + "grad_norm": 0.34323737025260925, + "learning_rate": 1.4726074420709545e-05, + "loss": 0.0454, + "step": 35198 + }, + { + "epoch": 0.8251199915610331, + "grad_norm": 0.20012547075748444, + "learning_rate": 1.4722227036547876e-05, + "loss": 0.0357, + "step": 35199 + }, + { + "epoch": 0.825143433135838, + "grad_norm": 0.7991947531700134, + "learning_rate": 1.4718380115105823e-05, + "loss": 0.1657, + "step": 35200 + }, + { + "epoch": 0.8251668747106431, + "grad_norm": 0.13789288699626923, + "learning_rate": 1.4714533656404206e-05, + "loss": 0.0201, + "step": 35201 + }, + { + "epoch": 0.825190316285448, + "grad_norm": 0.1805848777294159, + "learning_rate": 1.4710687660463962e-05, + "loss": 0.0303, + "step": 35202 + }, + { + "epoch": 0.8252137578602531, + "grad_norm": 0.44417935609817505, + "learning_rate": 1.4706842127305919e-05, + "loss": 0.1145, + "step": 35203 + }, + { + "epoch": 0.825237199435058, + "grad_norm": 0.42238420248031616, + "learning_rate": 1.4702997056950951e-05, + "loss": 0.0912, + "step": 35204 + }, + { + "epoch": 0.8252606410098631, + "grad_norm": 0.4760516285896301, + "learning_rate": 1.4699152449419917e-05, + "loss": 0.037, + "step": 35205 + }, + { + "epoch": 0.825284082584668, + "grad_norm": 0.061603154987096786, + "learning_rate": 1.4695308304733657e-05, + "loss": 0.005, + "step": 35206 + }, + { + "epoch": 0.8253075241594731, + "grad_norm": 0.5082315802574158, + "learning_rate": 1.4691464622913076e-05, + "loss": 0.1117, + "step": 35207 + }, + { + "epoch": 0.825330965734278, + "grad_norm": 0.10677687078714371, + "learning_rate": 1.4687621403979002e-05, + "loss": 0.0137, + "step": 35208 + }, + { + "epoch": 0.825354407309083, + "grad_norm": 0.5786046385765076, + "learning_rate": 1.4683778647952295e-05, + "loss": 0.103, + "step": 35209 + }, + { + "epoch": 0.825377848883888, + "grad_norm": 0.28619781136512756, + "learning_rate": 1.4679936354853808e-05, + "loss": 0.1887, + "step": 35210 + }, + { + "epoch": 0.825401290458693, + "grad_norm": 0.2668987810611725, + "learning_rate": 1.4676094524704387e-05, + "loss": 0.0377, + "step": 35211 + }, + { + "epoch": 0.825424732033498, + "grad_norm": 0.46406689286231995, + "learning_rate": 1.4672253157524852e-05, + "loss": 0.0959, + "step": 35212 + }, + { + "epoch": 0.825448173608303, + "grad_norm": 0.17431126534938812, + "learning_rate": 1.4668412253336061e-05, + "loss": 0.0244, + "step": 35213 + }, + { + "epoch": 0.825471615183108, + "grad_norm": 0.23816600441932678, + "learning_rate": 1.46645718121589e-05, + "loss": 0.0436, + "step": 35214 + }, + { + "epoch": 0.825495056757913, + "grad_norm": 0.2842039167881012, + "learning_rate": 1.4660731834014163e-05, + "loss": 0.0623, + "step": 35215 + }, + { + "epoch": 0.825518498332718, + "grad_norm": 0.34662291407585144, + "learning_rate": 1.4656892318922699e-05, + "loss": 0.0639, + "step": 35216 + }, + { + "epoch": 0.825541939907523, + "grad_norm": 0.7574036121368408, + "learning_rate": 1.4653053266905337e-05, + "loss": 0.1435, + "step": 35217 + }, + { + "epoch": 0.8255653814823279, + "grad_norm": 0.6779253482818604, + "learning_rate": 1.4649214677982915e-05, + "loss": 0.0999, + "step": 35218 + }, + { + "epoch": 0.825588823057133, + "grad_norm": 0.4014638662338257, + "learning_rate": 1.4645376552176216e-05, + "loss": 0.0814, + "step": 35219 + }, + { + "epoch": 0.8256122646319379, + "grad_norm": 0.8082044720649719, + "learning_rate": 1.4641538889506135e-05, + "loss": 0.1301, + "step": 35220 + }, + { + "epoch": 0.825635706206743, + "grad_norm": 0.4390197694301605, + "learning_rate": 1.4637701689993466e-05, + "loss": 0.0851, + "step": 35221 + }, + { + "epoch": 0.8256591477815479, + "grad_norm": 0.472648948431015, + "learning_rate": 1.4633864953659027e-05, + "loss": 0.0991, + "step": 35222 + }, + { + "epoch": 0.825682589356353, + "grad_norm": 0.48421403765678406, + "learning_rate": 1.4630028680523633e-05, + "loss": 0.6232, + "step": 35223 + }, + { + "epoch": 0.8257060309311579, + "grad_norm": 0.3537391126155853, + "learning_rate": 1.462619287060808e-05, + "loss": 0.2985, + "step": 35224 + }, + { + "epoch": 0.825729472505963, + "grad_norm": 0.389454185962677, + "learning_rate": 1.4622357523933239e-05, + "loss": 0.5123, + "step": 35225 + }, + { + "epoch": 0.8257529140807679, + "grad_norm": 0.40249204635620117, + "learning_rate": 1.461852264051985e-05, + "loss": 0.0586, + "step": 35226 + }, + { + "epoch": 0.825776355655573, + "grad_norm": 0.3940393030643463, + "learning_rate": 1.4614688220388784e-05, + "loss": 0.0866, + "step": 35227 + }, + { + "epoch": 0.8257997972303779, + "grad_norm": 0.47830018401145935, + "learning_rate": 1.4610854263560826e-05, + "loss": 0.1233, + "step": 35228 + }, + { + "epoch": 0.8258232388051829, + "grad_norm": 0.20085598528385162, + "learning_rate": 1.4607020770056767e-05, + "loss": 0.0208, + "step": 35229 + }, + { + "epoch": 0.8258466803799879, + "grad_norm": 0.2801125645637512, + "learning_rate": 1.4603187739897417e-05, + "loss": 0.0256, + "step": 35230 + }, + { + "epoch": 0.8258701219547929, + "grad_norm": 0.3836151659488678, + "learning_rate": 1.4599355173103558e-05, + "loss": 0.0412, + "step": 35231 + }, + { + "epoch": 0.825893563529598, + "grad_norm": 0.3904697000980377, + "learning_rate": 1.4595523069696004e-05, + "loss": 0.0437, + "step": 35232 + }, + { + "epoch": 0.8259170051044029, + "grad_norm": 0.8657170534133911, + "learning_rate": 1.4591691429695554e-05, + "loss": 0.1101, + "step": 35233 + }, + { + "epoch": 0.825940446679208, + "grad_norm": 0.2919728755950928, + "learning_rate": 1.4587860253122986e-05, + "loss": 0.0494, + "step": 35234 + }, + { + "epoch": 0.8259638882540129, + "grad_norm": 0.2778853178024292, + "learning_rate": 1.4584029539999088e-05, + "loss": 0.0584, + "step": 35235 + }, + { + "epoch": 0.825987329828818, + "grad_norm": 0.23948267102241516, + "learning_rate": 1.4580199290344653e-05, + "loss": 0.04, + "step": 35236 + }, + { + "epoch": 0.8260107714036229, + "grad_norm": 0.4296962320804596, + "learning_rate": 1.4576369504180431e-05, + "loss": 0.0759, + "step": 35237 + }, + { + "epoch": 0.8260342129784279, + "grad_norm": 0.4796404242515564, + "learning_rate": 1.4572540181527239e-05, + "loss": 0.1271, + "step": 35238 + }, + { + "epoch": 0.8260576545532329, + "grad_norm": 0.3791734576225281, + "learning_rate": 1.456871132240586e-05, + "loss": 0.0628, + "step": 35239 + }, + { + "epoch": 0.8260810961280379, + "grad_norm": 0.4361400902271271, + "learning_rate": 1.4564882926837053e-05, + "loss": 0.0636, + "step": 35240 + }, + { + "epoch": 0.8261045377028429, + "grad_norm": 0.30905023217201233, + "learning_rate": 1.4561054994841594e-05, + "loss": 0.0402, + "step": 35241 + }, + { + "epoch": 0.8261279792776479, + "grad_norm": 0.2564297616481781, + "learning_rate": 1.4557227526440264e-05, + "loss": 0.0299, + "step": 35242 + }, + { + "epoch": 0.8261514208524529, + "grad_norm": 0.4199031591415405, + "learning_rate": 1.4553400521653804e-05, + "loss": 0.0911, + "step": 35243 + }, + { + "epoch": 0.8261748624272579, + "grad_norm": 0.3736790120601654, + "learning_rate": 1.4549573980502985e-05, + "loss": 0.0719, + "step": 35244 + }, + { + "epoch": 0.8261983040020628, + "grad_norm": 0.5284562706947327, + "learning_rate": 1.45457479030086e-05, + "loss": 0.5417, + "step": 35245 + }, + { + "epoch": 0.8262217455768679, + "grad_norm": 0.39542314410209656, + "learning_rate": 1.4541922289191389e-05, + "loss": 0.0456, + "step": 35246 + }, + { + "epoch": 0.8262451871516728, + "grad_norm": 0.23063528537750244, + "learning_rate": 1.4538097139072104e-05, + "loss": 0.0488, + "step": 35247 + }, + { + "epoch": 0.8262686287264779, + "grad_norm": 0.07415583729743958, + "learning_rate": 1.4534272452671516e-05, + "loss": 0.0112, + "step": 35248 + }, + { + "epoch": 0.8262920703012828, + "grad_norm": 0.2520829737186432, + "learning_rate": 1.453044823001034e-05, + "loss": 0.0322, + "step": 35249 + }, + { + "epoch": 0.8263155118760879, + "grad_norm": 0.5409162640571594, + "learning_rate": 1.4526624471109384e-05, + "loss": 0.1453, + "step": 35250 + }, + { + "epoch": 0.8263389534508928, + "grad_norm": 0.14678679406642914, + "learning_rate": 1.4522801175989332e-05, + "loss": 0.01, + "step": 35251 + }, + { + "epoch": 0.8263623950256979, + "grad_norm": 0.3434128165245056, + "learning_rate": 1.4518978344670997e-05, + "loss": 0.0672, + "step": 35252 + }, + { + "epoch": 0.8263858366005028, + "grad_norm": 0.38498350977897644, + "learning_rate": 1.4515155977175087e-05, + "loss": 0.0622, + "step": 35253 + }, + { + "epoch": 0.8264092781753078, + "grad_norm": 0.36619606614112854, + "learning_rate": 1.4511334073522342e-05, + "loss": 0.0633, + "step": 35254 + }, + { + "epoch": 0.8264327197501128, + "grad_norm": 0.11424613744020462, + "learning_rate": 1.4507512633733511e-05, + "loss": 0.0173, + "step": 35255 + }, + { + "epoch": 0.8264561613249178, + "grad_norm": 0.10988426208496094, + "learning_rate": 1.450369165782931e-05, + "loss": 0.0129, + "step": 35256 + }, + { + "epoch": 0.8264796028997228, + "grad_norm": 0.6177948713302612, + "learning_rate": 1.4499871145830468e-05, + "loss": 0.0632, + "step": 35257 + }, + { + "epoch": 0.8265030444745278, + "grad_norm": 0.5764116048812866, + "learning_rate": 1.4496051097757745e-05, + "loss": 0.0779, + "step": 35258 + }, + { + "epoch": 0.8265264860493328, + "grad_norm": 0.43953144550323486, + "learning_rate": 1.449223151363186e-05, + "loss": 0.086, + "step": 35259 + }, + { + "epoch": 0.8265499276241378, + "grad_norm": 0.1140022724866867, + "learning_rate": 1.4488412393473527e-05, + "loss": 0.0228, + "step": 35260 + }, + { + "epoch": 0.8265733691989428, + "grad_norm": 0.713699996471405, + "learning_rate": 1.4484593737303464e-05, + "loss": 0.1368, + "step": 35261 + }, + { + "epoch": 0.8265968107737478, + "grad_norm": 0.4501645863056183, + "learning_rate": 1.4480775545142389e-05, + "loss": 0.4288, + "step": 35262 + }, + { + "epoch": 0.8266202523485527, + "grad_norm": 0.5011765360832214, + "learning_rate": 1.447695781701105e-05, + "loss": 0.0927, + "step": 35263 + }, + { + "epoch": 0.8266436939233578, + "grad_norm": 0.5080888271331787, + "learning_rate": 1.447314055293012e-05, + "loss": 0.0782, + "step": 35264 + }, + { + "epoch": 0.8266671354981627, + "grad_norm": 0.3784431219100952, + "learning_rate": 1.446932375292035e-05, + "loss": 0.0778, + "step": 35265 + }, + { + "epoch": 0.8266905770729678, + "grad_norm": 0.34154608845710754, + "learning_rate": 1.4465507417002444e-05, + "loss": 0.0751, + "step": 35266 + }, + { + "epoch": 0.8267140186477727, + "grad_norm": 0.7289538979530334, + "learning_rate": 1.4461691545197098e-05, + "loss": 0.1643, + "step": 35267 + }, + { + "epoch": 0.8267374602225778, + "grad_norm": 0.35498929023742676, + "learning_rate": 1.4457876137525017e-05, + "loss": 0.0649, + "step": 35268 + }, + { + "epoch": 0.8267609017973827, + "grad_norm": 0.22825010120868683, + "learning_rate": 1.4454061194006885e-05, + "loss": 0.0328, + "step": 35269 + }, + { + "epoch": 0.8267843433721878, + "grad_norm": 0.280853271484375, + "learning_rate": 1.445024671466344e-05, + "loss": 0.0333, + "step": 35270 + }, + { + "epoch": 0.8268077849469927, + "grad_norm": 0.10634130984544754, + "learning_rate": 1.4446432699515367e-05, + "loss": 0.0133, + "step": 35271 + }, + { + "epoch": 0.8268312265217977, + "grad_norm": 0.34209203720092773, + "learning_rate": 1.4442619148583347e-05, + "loss": 0.0491, + "step": 35272 + }, + { + "epoch": 0.8268546680966027, + "grad_norm": 0.21498723328113556, + "learning_rate": 1.4438806061888089e-05, + "loss": 0.033, + "step": 35273 + }, + { + "epoch": 0.8268781096714077, + "grad_norm": 0.21611452102661133, + "learning_rate": 1.443499343945025e-05, + "loss": 0.0291, + "step": 35274 + }, + { + "epoch": 0.8269015512462127, + "grad_norm": 0.33048108220100403, + "learning_rate": 1.4431181281290563e-05, + "loss": 0.0679, + "step": 35275 + }, + { + "epoch": 0.8269249928210177, + "grad_norm": 0.2745516300201416, + "learning_rate": 1.4427369587429674e-05, + "loss": 0.0283, + "step": 35276 + }, + { + "epoch": 0.8269484343958227, + "grad_norm": 0.5051072835922241, + "learning_rate": 1.44235583578883e-05, + "loss": 0.0665, + "step": 35277 + }, + { + "epoch": 0.8269718759706277, + "grad_norm": 0.727778434753418, + "learning_rate": 1.4419747592687116e-05, + "loss": 0.119, + "step": 35278 + }, + { + "epoch": 0.8269953175454327, + "grad_norm": 0.5243810415267944, + "learning_rate": 1.4415937291846782e-05, + "loss": 0.0282, + "step": 35279 + }, + { + "epoch": 0.8270187591202377, + "grad_norm": 0.40603265166282654, + "learning_rate": 1.441212745538797e-05, + "loss": 0.0477, + "step": 35280 + }, + { + "epoch": 0.8270422006950426, + "grad_norm": 0.3643619418144226, + "learning_rate": 1.4408318083331374e-05, + "loss": 0.0788, + "step": 35281 + }, + { + "epoch": 0.8270656422698477, + "grad_norm": 0.23034565150737762, + "learning_rate": 1.4404509175697622e-05, + "loss": 0.0246, + "step": 35282 + }, + { + "epoch": 0.8270890838446526, + "grad_norm": 0.4091968834400177, + "learning_rate": 1.4400700732507433e-05, + "loss": 0.0398, + "step": 35283 + }, + { + "epoch": 0.8271125254194577, + "grad_norm": 0.07498151063919067, + "learning_rate": 1.4396892753781454e-05, + "loss": 0.0101, + "step": 35284 + }, + { + "epoch": 0.8271359669942627, + "grad_norm": 0.35386133193969727, + "learning_rate": 1.4393085239540326e-05, + "loss": 0.0426, + "step": 35285 + }, + { + "epoch": 0.8271594085690677, + "grad_norm": 0.2993687093257904, + "learning_rate": 1.4389278189804723e-05, + "loss": 0.0662, + "step": 35286 + }, + { + "epoch": 0.8271828501438727, + "grad_norm": 0.3069121837615967, + "learning_rate": 1.4385471604595313e-05, + "loss": 0.0277, + "step": 35287 + }, + { + "epoch": 0.8272062917186777, + "grad_norm": 0.5951676368713379, + "learning_rate": 1.4381665483932738e-05, + "loss": 0.1062, + "step": 35288 + }, + { + "epoch": 0.8272297332934827, + "grad_norm": 0.5721297860145569, + "learning_rate": 1.4377859827837647e-05, + "loss": 0.0917, + "step": 35289 + }, + { + "epoch": 0.8272531748682876, + "grad_norm": 0.5343432426452637, + "learning_rate": 1.4374054636330702e-05, + "loss": 0.0848, + "step": 35290 + }, + { + "epoch": 0.8272766164430927, + "grad_norm": 0.4793280363082886, + "learning_rate": 1.4370249909432554e-05, + "loss": 0.0758, + "step": 35291 + }, + { + "epoch": 0.8273000580178976, + "grad_norm": 0.10246779024600983, + "learning_rate": 1.436644564716384e-05, + "loss": 0.01, + "step": 35292 + }, + { + "epoch": 0.8273234995927027, + "grad_norm": 0.6383845210075378, + "learning_rate": 1.4362641849545189e-05, + "loss": 0.1969, + "step": 35293 + }, + { + "epoch": 0.8273469411675076, + "grad_norm": 0.37822604179382324, + "learning_rate": 1.4358838516597262e-05, + "loss": 0.0579, + "step": 35294 + }, + { + "epoch": 0.8273703827423127, + "grad_norm": 0.23954921960830688, + "learning_rate": 1.4355035648340654e-05, + "loss": 0.0558, + "step": 35295 + }, + { + "epoch": 0.8273938243171176, + "grad_norm": 0.08642179518938065, + "learning_rate": 1.4351233244796058e-05, + "loss": 0.0097, + "step": 35296 + }, + { + "epoch": 0.8274172658919227, + "grad_norm": 0.16640953719615936, + "learning_rate": 1.4347431305984071e-05, + "loss": 0.0142, + "step": 35297 + }, + { + "epoch": 0.8274407074667276, + "grad_norm": 0.24471719563007355, + "learning_rate": 1.434362983192532e-05, + "loss": 0.0327, + "step": 35298 + }, + { + "epoch": 0.8274641490415326, + "grad_norm": 0.20903730392456055, + "learning_rate": 1.4339828822640456e-05, + "loss": 0.0388, + "step": 35299 + }, + { + "epoch": 0.8274875906163376, + "grad_norm": 0.09487566351890564, + "learning_rate": 1.4336028278150093e-05, + "loss": 0.0217, + "step": 35300 + }, + { + "epoch": 0.8275110321911426, + "grad_norm": 0.28905370831489563, + "learning_rate": 1.433222819847485e-05, + "loss": 0.2961, + "step": 35301 + }, + { + "epoch": 0.8275344737659476, + "grad_norm": 0.5670478343963623, + "learning_rate": 1.4328428583635311e-05, + "loss": 0.0246, + "step": 35302 + }, + { + "epoch": 0.8275579153407526, + "grad_norm": 0.3493918776512146, + "learning_rate": 1.4324629433652171e-05, + "loss": 0.0859, + "step": 35303 + }, + { + "epoch": 0.8275813569155576, + "grad_norm": 0.5553448796272278, + "learning_rate": 1.4320830748545988e-05, + "loss": 0.0845, + "step": 35304 + }, + { + "epoch": 0.8276047984903626, + "grad_norm": 0.24059191346168518, + "learning_rate": 1.4317032528337381e-05, + "loss": 0.0327, + "step": 35305 + }, + { + "epoch": 0.8276282400651676, + "grad_norm": 0.36812883615493774, + "learning_rate": 1.431323477304698e-05, + "loss": 0.0706, + "step": 35306 + }, + { + "epoch": 0.8276516816399726, + "grad_norm": 0.33174338936805725, + "learning_rate": 1.4309437482695342e-05, + "loss": 0.0447, + "step": 35307 + }, + { + "epoch": 0.8276751232147775, + "grad_norm": 0.10280603915452957, + "learning_rate": 1.4305640657303132e-05, + "loss": 0.0143, + "step": 35308 + }, + { + "epoch": 0.8276985647895826, + "grad_norm": 0.5805962085723877, + "learning_rate": 1.4301844296890921e-05, + "loss": 0.1016, + "step": 35309 + }, + { + "epoch": 0.8277220063643875, + "grad_norm": 0.4135262370109558, + "learning_rate": 1.4298048401479291e-05, + "loss": 0.0632, + "step": 35310 + }, + { + "epoch": 0.8277454479391926, + "grad_norm": 0.3467923104763031, + "learning_rate": 1.4294252971088895e-05, + "loss": 0.0535, + "step": 35311 + }, + { + "epoch": 0.8277688895139975, + "grad_norm": 0.43751776218414307, + "learning_rate": 1.4290458005740281e-05, + "loss": 0.0871, + "step": 35312 + }, + { + "epoch": 0.8277923310888026, + "grad_norm": 0.38962310552597046, + "learning_rate": 1.4286663505454057e-05, + "loss": 0.0693, + "step": 35313 + }, + { + "epoch": 0.8278157726636075, + "grad_norm": 0.4617680013179779, + "learning_rate": 1.428286947025077e-05, + "loss": 0.1206, + "step": 35314 + }, + { + "epoch": 0.8278392142384126, + "grad_norm": 0.09193576127290726, + "learning_rate": 1.4279075900151085e-05, + "loss": 0.0152, + "step": 35315 + }, + { + "epoch": 0.8278626558132175, + "grad_norm": 0.5208391547203064, + "learning_rate": 1.4275282795175538e-05, + "loss": 0.0678, + "step": 35316 + }, + { + "epoch": 0.8278860973880225, + "grad_norm": 0.20469187200069427, + "learning_rate": 1.4271490155344714e-05, + "loss": 0.0273, + "step": 35317 + }, + { + "epoch": 0.8279095389628275, + "grad_norm": 0.4682497978210449, + "learning_rate": 1.4267697980679196e-05, + "loss": 0.0654, + "step": 35318 + }, + { + "epoch": 0.8279329805376325, + "grad_norm": 0.39311444759368896, + "learning_rate": 1.4263906271199557e-05, + "loss": 0.0844, + "step": 35319 + }, + { + "epoch": 0.8279564221124375, + "grad_norm": 0.3957187533378601, + "learning_rate": 1.4260115026926358e-05, + "loss": 0.0286, + "step": 35320 + }, + { + "epoch": 0.8279798636872425, + "grad_norm": 0.24449902772903442, + "learning_rate": 1.4256324247880192e-05, + "loss": 0.0655, + "step": 35321 + }, + { + "epoch": 0.8280033052620475, + "grad_norm": 0.6393737196922302, + "learning_rate": 1.4252533934081613e-05, + "loss": 0.1234, + "step": 35322 + }, + { + "epoch": 0.8280267468368525, + "grad_norm": 0.5020649433135986, + "learning_rate": 1.4248744085551214e-05, + "loss": 0.0948, + "step": 35323 + }, + { + "epoch": 0.8280501884116575, + "grad_norm": 0.7639071345329285, + "learning_rate": 1.4244954702309542e-05, + "loss": 0.1542, + "step": 35324 + }, + { + "epoch": 0.8280736299864625, + "grad_norm": 0.2993619441986084, + "learning_rate": 1.4241165784377164e-05, + "loss": 0.0421, + "step": 35325 + }, + { + "epoch": 0.8280970715612674, + "grad_norm": 0.1890122890472412, + "learning_rate": 1.4237377331774625e-05, + "loss": 0.0202, + "step": 35326 + }, + { + "epoch": 0.8281205131360725, + "grad_norm": 0.12481103837490082, + "learning_rate": 1.423358934452247e-05, + "loss": 0.0232, + "step": 35327 + }, + { + "epoch": 0.8281439547108774, + "grad_norm": 0.07266689091920853, + "learning_rate": 1.4229801822641287e-05, + "loss": 0.0095, + "step": 35328 + }, + { + "epoch": 0.8281673962856825, + "grad_norm": 0.47659268975257874, + "learning_rate": 1.4226014766151608e-05, + "loss": 0.181, + "step": 35329 + }, + { + "epoch": 0.8281908378604874, + "grad_norm": 0.3499574363231659, + "learning_rate": 1.4222228175073993e-05, + "loss": 0.0828, + "step": 35330 + }, + { + "epoch": 0.8282142794352925, + "grad_norm": 0.38627105951309204, + "learning_rate": 1.4218442049428971e-05, + "loss": 0.061, + "step": 35331 + }, + { + "epoch": 0.8282377210100974, + "grad_norm": 0.5777174234390259, + "learning_rate": 1.4214656389237102e-05, + "loss": 0.1136, + "step": 35332 + }, + { + "epoch": 0.8282611625849025, + "grad_norm": 0.2018909752368927, + "learning_rate": 1.4210871194518892e-05, + "loss": 0.0179, + "step": 35333 + }, + { + "epoch": 0.8282846041597074, + "grad_norm": 0.47601521015167236, + "learning_rate": 1.4207086465294939e-05, + "loss": 0.0555, + "step": 35334 + }, + { + "epoch": 0.8283080457345124, + "grad_norm": 0.25866127014160156, + "learning_rate": 1.420330220158571e-05, + "loss": 0.061, + "step": 35335 + }, + { + "epoch": 0.8283314873093175, + "grad_norm": 0.24555039405822754, + "learning_rate": 1.4199518403411816e-05, + "loss": 0.0325, + "step": 35336 + }, + { + "epoch": 0.8283549288841224, + "grad_norm": 0.2333049476146698, + "learning_rate": 1.4195735070793736e-05, + "loss": 0.0386, + "step": 35337 + }, + { + "epoch": 0.8283783704589275, + "grad_norm": 0.34591415524482727, + "learning_rate": 1.4191952203752013e-05, + "loss": 0.0363, + "step": 35338 + }, + { + "epoch": 0.8284018120337324, + "grad_norm": 0.5466766953468323, + "learning_rate": 1.418816980230716e-05, + "loss": 0.6586, + "step": 35339 + }, + { + "epoch": 0.8284252536085375, + "grad_norm": 0.39776912331581116, + "learning_rate": 1.4184387866479698e-05, + "loss": 0.0764, + "step": 35340 + }, + { + "epoch": 0.8284486951833424, + "grad_norm": 0.4160284399986267, + "learning_rate": 1.4180606396290186e-05, + "loss": 0.0873, + "step": 35341 + }, + { + "epoch": 0.8284721367581475, + "grad_norm": 0.21702127158641815, + "learning_rate": 1.4176825391759107e-05, + "loss": 0.0238, + "step": 35342 + }, + { + "epoch": 0.8284955783329524, + "grad_norm": 0.13905587792396545, + "learning_rate": 1.4173044852907002e-05, + "loss": 0.0339, + "step": 35343 + }, + { + "epoch": 0.8285190199077574, + "grad_norm": 0.5532408952713013, + "learning_rate": 1.4169264779754355e-05, + "loss": 0.136, + "step": 35344 + }, + { + "epoch": 0.8285424614825624, + "grad_norm": 0.5606169700622559, + "learning_rate": 1.416548517232167e-05, + "loss": 0.0482, + "step": 35345 + }, + { + "epoch": 0.8285659030573674, + "grad_norm": 0.3021167516708374, + "learning_rate": 1.4161706030629508e-05, + "loss": 0.0485, + "step": 35346 + }, + { + "epoch": 0.8285893446321724, + "grad_norm": 0.46460315585136414, + "learning_rate": 1.4157927354698309e-05, + "loss": 0.096, + "step": 35347 + }, + { + "epoch": 0.8286127862069774, + "grad_norm": 0.10042431950569153, + "learning_rate": 1.4154149144548634e-05, + "loss": 0.0128, + "step": 35348 + }, + { + "epoch": 0.8286362277817824, + "grad_norm": 0.2858497202396393, + "learning_rate": 1.4150371400200957e-05, + "loss": 0.0583, + "step": 35349 + }, + { + "epoch": 0.8286596693565874, + "grad_norm": 0.5366578102111816, + "learning_rate": 1.4146594121675783e-05, + "loss": 0.0611, + "step": 35350 + }, + { + "epoch": 0.8286831109313924, + "grad_norm": 0.28190919756889343, + "learning_rate": 1.4142817308993605e-05, + "loss": 0.0339, + "step": 35351 + }, + { + "epoch": 0.8287065525061974, + "grad_norm": 0.4879330098628998, + "learning_rate": 1.4139040962174888e-05, + "loss": 0.0644, + "step": 35352 + }, + { + "epoch": 0.8287299940810023, + "grad_norm": 0.09178797900676727, + "learning_rate": 1.413526508124018e-05, + "loss": 0.0071, + "step": 35353 + }, + { + "epoch": 0.8287534356558074, + "grad_norm": 0.21659624576568604, + "learning_rate": 1.413148966620993e-05, + "loss": 0.0314, + "step": 35354 + }, + { + "epoch": 0.8287768772306123, + "grad_norm": 0.47607600688934326, + "learning_rate": 1.4127714717104634e-05, + "loss": 0.0468, + "step": 35355 + }, + { + "epoch": 0.8288003188054174, + "grad_norm": 0.2990375757217407, + "learning_rate": 1.4123940233944777e-05, + "loss": 0.033, + "step": 35356 + }, + { + "epoch": 0.8288237603802223, + "grad_norm": 0.3475095331668854, + "learning_rate": 1.4120166216750818e-05, + "loss": 0.0434, + "step": 35357 + }, + { + "epoch": 0.8288472019550274, + "grad_norm": 0.7030065059661865, + "learning_rate": 1.4116392665543244e-05, + "loss": 0.1075, + "step": 35358 + }, + { + "epoch": 0.8288706435298323, + "grad_norm": 0.4434644281864166, + "learning_rate": 1.4112619580342535e-05, + "loss": 0.0864, + "step": 35359 + }, + { + "epoch": 0.8288940851046374, + "grad_norm": 0.7629445791244507, + "learning_rate": 1.4108846961169197e-05, + "loss": 0.0425, + "step": 35360 + }, + { + "epoch": 0.8289175266794423, + "grad_norm": 0.19997335970401764, + "learning_rate": 1.4105074808043662e-05, + "loss": 0.0188, + "step": 35361 + }, + { + "epoch": 0.8289409682542473, + "grad_norm": 0.21468693017959595, + "learning_rate": 1.4101303120986408e-05, + "loss": 0.0334, + "step": 35362 + }, + { + "epoch": 0.8289644098290523, + "grad_norm": 0.28125473856925964, + "learning_rate": 1.409753190001789e-05, + "loss": 0.0366, + "step": 35363 + }, + { + "epoch": 0.8289878514038573, + "grad_norm": 0.4878726601600647, + "learning_rate": 1.4093761145158591e-05, + "loss": 0.0908, + "step": 35364 + }, + { + "epoch": 0.8290112929786623, + "grad_norm": 0.3913080394268036, + "learning_rate": 1.4089990856428936e-05, + "loss": 0.0399, + "step": 35365 + }, + { + "epoch": 0.8290347345534673, + "grad_norm": 0.566903293132782, + "learning_rate": 1.4086221033849422e-05, + "loss": 0.0904, + "step": 35366 + }, + { + "epoch": 0.8290581761282723, + "grad_norm": 0.25914761424064636, + "learning_rate": 1.4082451677440489e-05, + "loss": 0.0341, + "step": 35367 + }, + { + "epoch": 0.8290816177030773, + "grad_norm": 0.12041620165109634, + "learning_rate": 1.4078682787222598e-05, + "loss": 0.0268, + "step": 35368 + }, + { + "epoch": 0.8291050592778823, + "grad_norm": 0.1501733362674713, + "learning_rate": 1.4074914363216184e-05, + "loss": 0.0338, + "step": 35369 + }, + { + "epoch": 0.8291285008526873, + "grad_norm": 1.093315601348877, + "learning_rate": 1.4071146405441704e-05, + "loss": 0.1223, + "step": 35370 + }, + { + "epoch": 0.8291519424274922, + "grad_norm": 0.36720412969589233, + "learning_rate": 1.4067378913919571e-05, + "loss": 0.0574, + "step": 35371 + }, + { + "epoch": 0.8291753840022973, + "grad_norm": 0.507374107837677, + "learning_rate": 1.4063611888670258e-05, + "loss": 0.142, + "step": 35372 + }, + { + "epoch": 0.8291988255771022, + "grad_norm": 0.5812376737594604, + "learning_rate": 1.4059845329714228e-05, + "loss": 0.5431, + "step": 35373 + }, + { + "epoch": 0.8292222671519073, + "grad_norm": 0.12405742704868317, + "learning_rate": 1.4056079237071894e-05, + "loss": 0.0201, + "step": 35374 + }, + { + "epoch": 0.8292457087267122, + "grad_norm": 0.516441822052002, + "learning_rate": 1.405231361076369e-05, + "loss": 0.0965, + "step": 35375 + }, + { + "epoch": 0.8292691503015173, + "grad_norm": 0.4046814441680908, + "learning_rate": 1.4048548450810051e-05, + "loss": 0.0961, + "step": 35376 + }, + { + "epoch": 0.8292925918763222, + "grad_norm": 0.739238977432251, + "learning_rate": 1.4044783757231395e-05, + "loss": 0.1344, + "step": 35377 + }, + { + "epoch": 0.8293160334511273, + "grad_norm": 0.38159582018852234, + "learning_rate": 1.404101953004815e-05, + "loss": 0.0709, + "step": 35378 + }, + { + "epoch": 0.8293394750259322, + "grad_norm": 0.35724782943725586, + "learning_rate": 1.4037255769280766e-05, + "loss": 0.0835, + "step": 35379 + }, + { + "epoch": 0.8293629166007372, + "grad_norm": 0.33322063088417053, + "learning_rate": 1.403349247494965e-05, + "loss": 0.0722, + "step": 35380 + }, + { + "epoch": 0.8293863581755422, + "grad_norm": 0.4899657368659973, + "learning_rate": 1.4029729647075218e-05, + "loss": 0.0755, + "step": 35381 + }, + { + "epoch": 0.8294097997503472, + "grad_norm": 0.30796122550964355, + "learning_rate": 1.4025967285677888e-05, + "loss": 0.0576, + "step": 35382 + }, + { + "epoch": 0.8294332413251522, + "grad_norm": 0.46329161524772644, + "learning_rate": 1.4022205390778065e-05, + "loss": 0.0892, + "step": 35383 + }, + { + "epoch": 0.8294566828999572, + "grad_norm": 0.4657871127128601, + "learning_rate": 1.4018443962396165e-05, + "loss": 0.0755, + "step": 35384 + }, + { + "epoch": 0.8294801244747622, + "grad_norm": 0.1859849989414215, + "learning_rate": 1.4014683000552631e-05, + "loss": 0.011, + "step": 35385 + }, + { + "epoch": 0.8295035660495672, + "grad_norm": 0.37605881690979004, + "learning_rate": 1.4010922505267843e-05, + "loss": 0.0198, + "step": 35386 + }, + { + "epoch": 0.8295270076243723, + "grad_norm": 0.14343063533306122, + "learning_rate": 1.4007162476562208e-05, + "loss": 0.0178, + "step": 35387 + }, + { + "epoch": 0.8295504491991772, + "grad_norm": 0.39095795154571533, + "learning_rate": 1.4003402914456132e-05, + "loss": 0.0356, + "step": 35388 + }, + { + "epoch": 0.8295738907739822, + "grad_norm": 0.4858228862285614, + "learning_rate": 1.3999643818969999e-05, + "loss": 0.0697, + "step": 35389 + }, + { + "epoch": 0.8295973323487872, + "grad_norm": 0.3637869656085968, + "learning_rate": 1.3995885190124202e-05, + "loss": 0.3136, + "step": 35390 + }, + { + "epoch": 0.8296207739235922, + "grad_norm": 0.37377703189849854, + "learning_rate": 1.3992127027939173e-05, + "loss": 0.0541, + "step": 35391 + }, + { + "epoch": 0.8296442154983972, + "grad_norm": 0.49297332763671875, + "learning_rate": 1.398836933243528e-05, + "loss": 0.0783, + "step": 35392 + }, + { + "epoch": 0.8296676570732022, + "grad_norm": 0.497037410736084, + "learning_rate": 1.398461210363291e-05, + "loss": 0.0683, + "step": 35393 + }, + { + "epoch": 0.8296910986480072, + "grad_norm": 0.3349895477294922, + "learning_rate": 1.3980855341552457e-05, + "loss": 0.0483, + "step": 35394 + }, + { + "epoch": 0.8297145402228122, + "grad_norm": 0.41544628143310547, + "learning_rate": 1.3977099046214281e-05, + "loss": 0.0573, + "step": 35395 + }, + { + "epoch": 0.8297379817976172, + "grad_norm": 0.33921515941619873, + "learning_rate": 1.3973343217638802e-05, + "loss": 0.0314, + "step": 35396 + }, + { + "epoch": 0.8297614233724222, + "grad_norm": 0.3519490659236908, + "learning_rate": 1.3969587855846366e-05, + "loss": 0.0701, + "step": 35397 + }, + { + "epoch": 0.8297848649472271, + "grad_norm": 0.3886312246322632, + "learning_rate": 1.3965832960857384e-05, + "loss": 0.0598, + "step": 35398 + }, + { + "epoch": 0.8298083065220322, + "grad_norm": 0.414034903049469, + "learning_rate": 1.3962078532692201e-05, + "loss": 0.081, + "step": 35399 + }, + { + "epoch": 0.8298317480968371, + "grad_norm": 0.15799374878406525, + "learning_rate": 1.3958324571371218e-05, + "loss": 0.0259, + "step": 35400 + }, + { + "epoch": 0.8298551896716422, + "grad_norm": 0.20487111806869507, + "learning_rate": 1.3954571076914768e-05, + "loss": 0.0227, + "step": 35401 + }, + { + "epoch": 0.8298786312464471, + "grad_norm": 0.3484320342540741, + "learning_rate": 1.395081804934325e-05, + "loss": 0.0483, + "step": 35402 + }, + { + "epoch": 0.8299020728212522, + "grad_norm": 0.44476068019866943, + "learning_rate": 1.394706548867698e-05, + "loss": 0.0519, + "step": 35403 + }, + { + "epoch": 0.8299255143960571, + "grad_norm": 0.2237134724855423, + "learning_rate": 1.3943313394936376e-05, + "loss": 0.0532, + "step": 35404 + }, + { + "epoch": 0.8299489559708622, + "grad_norm": 0.46254369616508484, + "learning_rate": 1.3939561768141774e-05, + "loss": 0.051, + "step": 35405 + }, + { + "epoch": 0.8299723975456671, + "grad_norm": 0.34474870562553406, + "learning_rate": 1.3935810608313527e-05, + "loss": 0.0468, + "step": 35406 + }, + { + "epoch": 0.8299958391204721, + "grad_norm": 0.07840907573699951, + "learning_rate": 1.3932059915471974e-05, + "loss": 0.0135, + "step": 35407 + }, + { + "epoch": 0.8300192806952771, + "grad_norm": 0.5009525418281555, + "learning_rate": 1.3928309689637508e-05, + "loss": 0.0789, + "step": 35408 + }, + { + "epoch": 0.8300427222700821, + "grad_norm": 0.35519877076148987, + "learning_rate": 1.392455993083045e-05, + "loss": 0.0347, + "step": 35409 + }, + { + "epoch": 0.8300661638448871, + "grad_norm": 0.11119483411312103, + "learning_rate": 1.3920810639071125e-05, + "loss": 0.0078, + "step": 35410 + }, + { + "epoch": 0.8300896054196921, + "grad_norm": 0.189711794257164, + "learning_rate": 1.3917061814379928e-05, + "loss": 0.0262, + "step": 35411 + }, + { + "epoch": 0.8301130469944971, + "grad_norm": 0.5201699137687683, + "learning_rate": 1.3913313456777177e-05, + "loss": 0.1394, + "step": 35412 + }, + { + "epoch": 0.8301364885693021, + "grad_norm": 0.3225127160549164, + "learning_rate": 1.39095655662832e-05, + "loss": 0.0366, + "step": 35413 + }, + { + "epoch": 0.830159930144107, + "grad_norm": 0.45620185136795044, + "learning_rate": 1.3905818142918347e-05, + "loss": 0.075, + "step": 35414 + }, + { + "epoch": 0.8301833717189121, + "grad_norm": 0.5125839710235596, + "learning_rate": 1.3902071186702913e-05, + "loss": 0.0856, + "step": 35415 + }, + { + "epoch": 0.830206813293717, + "grad_norm": 0.1468053013086319, + "learning_rate": 1.3898324697657294e-05, + "loss": 0.0197, + "step": 35416 + }, + { + "epoch": 0.8302302548685221, + "grad_norm": 0.1396772861480713, + "learning_rate": 1.3894578675801783e-05, + "loss": 0.008, + "step": 35417 + }, + { + "epoch": 0.830253696443327, + "grad_norm": 0.4228314757347107, + "learning_rate": 1.389083312115672e-05, + "loss": 0.0531, + "step": 35418 + }, + { + "epoch": 0.8302771380181321, + "grad_norm": 0.1338241994380951, + "learning_rate": 1.3887088033742401e-05, + "loss": 0.0278, + "step": 35419 + }, + { + "epoch": 0.830300579592937, + "grad_norm": 0.1950959861278534, + "learning_rate": 1.3883343413579153e-05, + "loss": 0.0269, + "step": 35420 + }, + { + "epoch": 0.8303240211677421, + "grad_norm": 0.5040866136550903, + "learning_rate": 1.3879599260687314e-05, + "loss": 0.6469, + "step": 35421 + }, + { + "epoch": 0.830347462742547, + "grad_norm": 0.6229488849639893, + "learning_rate": 1.3875855575087182e-05, + "loss": 0.5119, + "step": 35422 + }, + { + "epoch": 0.830370904317352, + "grad_norm": 0.2693333625793457, + "learning_rate": 1.3872112356799093e-05, + "loss": 0.0424, + "step": 35423 + }, + { + "epoch": 0.830394345892157, + "grad_norm": 0.37869906425476074, + "learning_rate": 1.3868369605843356e-05, + "loss": 0.0575, + "step": 35424 + }, + { + "epoch": 0.830417787466962, + "grad_norm": 0.24712151288986206, + "learning_rate": 1.3864627322240254e-05, + "loss": 0.0371, + "step": 35425 + }, + { + "epoch": 0.830441229041767, + "grad_norm": 0.380156546831131, + "learning_rate": 1.3860885506010112e-05, + "loss": 0.0648, + "step": 35426 + }, + { + "epoch": 0.830464670616572, + "grad_norm": 0.3727814257144928, + "learning_rate": 1.385714415717323e-05, + "loss": 0.0759, + "step": 35427 + }, + { + "epoch": 0.830488112191377, + "grad_norm": 0.20955941081047058, + "learning_rate": 1.3853403275749877e-05, + "loss": 0.0451, + "step": 35428 + }, + { + "epoch": 0.830511553766182, + "grad_norm": 0.42008572816848755, + "learning_rate": 1.3849662861760404e-05, + "loss": 0.0423, + "step": 35429 + }, + { + "epoch": 0.830534995340987, + "grad_norm": 0.23216724395751953, + "learning_rate": 1.3845922915225084e-05, + "loss": 0.0519, + "step": 35430 + }, + { + "epoch": 0.830558436915792, + "grad_norm": 0.4527747929096222, + "learning_rate": 1.3842183436164213e-05, + "loss": 0.0511, + "step": 35431 + }, + { + "epoch": 0.830581878490597, + "grad_norm": 0.0988416075706482, + "learning_rate": 1.3838444424598052e-05, + "loss": 0.0174, + "step": 35432 + }, + { + "epoch": 0.830605320065402, + "grad_norm": 0.3898790180683136, + "learning_rate": 1.3834705880546927e-05, + "loss": 0.0774, + "step": 35433 + }, + { + "epoch": 0.8306287616402069, + "grad_norm": 0.3166711628437042, + "learning_rate": 1.3830967804031126e-05, + "loss": 0.0274, + "step": 35434 + }, + { + "epoch": 0.830652203215012, + "grad_norm": 0.2965392768383026, + "learning_rate": 1.3827230195070884e-05, + "loss": 0.056, + "step": 35435 + }, + { + "epoch": 0.8306756447898169, + "grad_norm": 0.36850062012672424, + "learning_rate": 1.3823493053686542e-05, + "loss": 0.0507, + "step": 35436 + }, + { + "epoch": 0.830699086364622, + "grad_norm": 0.35453706979751587, + "learning_rate": 1.3819756379898342e-05, + "loss": 0.0413, + "step": 35437 + }, + { + "epoch": 0.830722527939427, + "grad_norm": 0.36084815859794617, + "learning_rate": 1.3816020173726574e-05, + "loss": 0.0842, + "step": 35438 + }, + { + "epoch": 0.830745969514232, + "grad_norm": 0.44966921210289, + "learning_rate": 1.3812284435191503e-05, + "loss": 0.1157, + "step": 35439 + }, + { + "epoch": 0.830769411089037, + "grad_norm": 0.08164346218109131, + "learning_rate": 1.38085491643134e-05, + "loss": 0.005, + "step": 35440 + }, + { + "epoch": 0.830792852663842, + "grad_norm": 0.44006872177124023, + "learning_rate": 1.3804814361112506e-05, + "loss": 0.1019, + "step": 35441 + }, + { + "epoch": 0.830816294238647, + "grad_norm": 0.23187242448329926, + "learning_rate": 1.3801080025609136e-05, + "loss": 0.0402, + "step": 35442 + }, + { + "epoch": 0.8308397358134519, + "grad_norm": 0.17754103243350983, + "learning_rate": 1.3797346157823533e-05, + "loss": 0.0273, + "step": 35443 + }, + { + "epoch": 0.830863177388257, + "grad_norm": 0.3834657073020935, + "learning_rate": 1.3793612757775931e-05, + "loss": 0.0765, + "step": 35444 + }, + { + "epoch": 0.8308866189630619, + "grad_norm": 0.13369639217853546, + "learning_rate": 1.378987982548664e-05, + "loss": 0.0196, + "step": 35445 + }, + { + "epoch": 0.830910060537867, + "grad_norm": 0.4611814022064209, + "learning_rate": 1.3786147360975876e-05, + "loss": 0.0731, + "step": 35446 + }, + { + "epoch": 0.8309335021126719, + "grad_norm": 0.3841375708580017, + "learning_rate": 1.3782415364263901e-05, + "loss": 0.048, + "step": 35447 + }, + { + "epoch": 0.830956943687477, + "grad_norm": 0.5724286437034607, + "learning_rate": 1.3778683835370943e-05, + "loss": 0.5988, + "step": 35448 + }, + { + "epoch": 0.8309803852622819, + "grad_norm": 0.39294877648353577, + "learning_rate": 1.3774952774317284e-05, + "loss": 0.072, + "step": 35449 + }, + { + "epoch": 0.831003826837087, + "grad_norm": 0.7193180918693542, + "learning_rate": 1.3771222181123167e-05, + "loss": 0.0865, + "step": 35450 + }, + { + "epoch": 0.8310272684118919, + "grad_norm": 0.3455347716808319, + "learning_rate": 1.3767492055808817e-05, + "loss": 0.0482, + "step": 35451 + }, + { + "epoch": 0.831050709986697, + "grad_norm": 0.27007919549942017, + "learning_rate": 1.3763762398394486e-05, + "loss": 0.0398, + "step": 35452 + }, + { + "epoch": 0.8310741515615019, + "grad_norm": 0.5692459940910339, + "learning_rate": 1.376003320890038e-05, + "loss": 0.1144, + "step": 35453 + }, + { + "epoch": 0.8310975931363069, + "grad_norm": 0.4003628194332123, + "learning_rate": 1.3756304487346783e-05, + "loss": 0.0707, + "step": 35454 + }, + { + "epoch": 0.8311210347111119, + "grad_norm": 0.5939705967903137, + "learning_rate": 1.37525762337539e-05, + "loss": 0.0619, + "step": 35455 + }, + { + "epoch": 0.8311444762859169, + "grad_norm": 0.3052457571029663, + "learning_rate": 1.3748848448141938e-05, + "loss": 0.0408, + "step": 35456 + }, + { + "epoch": 0.8311679178607219, + "grad_norm": 0.27870234847068787, + "learning_rate": 1.374512113053118e-05, + "loss": 0.0352, + "step": 35457 + }, + { + "epoch": 0.8311913594355269, + "grad_norm": 0.5113890767097473, + "learning_rate": 1.374139428094181e-05, + "loss": 0.1232, + "step": 35458 + }, + { + "epoch": 0.8312148010103318, + "grad_norm": 0.3860456645488739, + "learning_rate": 1.3737667899394068e-05, + "loss": 0.0741, + "step": 35459 + }, + { + "epoch": 0.8312382425851369, + "grad_norm": 0.5090420246124268, + "learning_rate": 1.3733941985908139e-05, + "loss": 0.6365, + "step": 35460 + }, + { + "epoch": 0.8312616841599418, + "grad_norm": 0.08494653552770615, + "learning_rate": 1.3730216540504292e-05, + "loss": 0.0157, + "step": 35461 + }, + { + "epoch": 0.8312851257347469, + "grad_norm": 0.27065157890319824, + "learning_rate": 1.3726491563202715e-05, + "loss": 0.0423, + "step": 35462 + }, + { + "epoch": 0.8313085673095518, + "grad_norm": 0.9405777454376221, + "learning_rate": 1.3722767054023622e-05, + "loss": 0.5625, + "step": 35463 + }, + { + "epoch": 0.8313320088843569, + "grad_norm": 0.4427320659160614, + "learning_rate": 1.371904301298722e-05, + "loss": 0.0434, + "step": 35464 + }, + { + "epoch": 0.8313554504591618, + "grad_norm": 0.1034708023071289, + "learning_rate": 1.3715319440113706e-05, + "loss": 0.007, + "step": 35465 + }, + { + "epoch": 0.8313788920339669, + "grad_norm": 0.3306187689304352, + "learning_rate": 1.3711596335423282e-05, + "loss": 0.0692, + "step": 35466 + }, + { + "epoch": 0.8314023336087718, + "grad_norm": 0.508343517780304, + "learning_rate": 1.3707873698936191e-05, + "loss": 0.6609, + "step": 35467 + }, + { + "epoch": 0.8314257751835769, + "grad_norm": 0.07969978451728821, + "learning_rate": 1.3704151530672571e-05, + "loss": 0.0048, + "step": 35468 + }, + { + "epoch": 0.8314492167583818, + "grad_norm": 0.5102785229682922, + "learning_rate": 1.3700429830652683e-05, + "loss": 0.068, + "step": 35469 + }, + { + "epoch": 0.8314726583331868, + "grad_norm": 0.9472704529762268, + "learning_rate": 1.3696708598896679e-05, + "loss": 0.0895, + "step": 35470 + }, + { + "epoch": 0.8314960999079918, + "grad_norm": 0.3369271755218506, + "learning_rate": 1.3692987835424764e-05, + "loss": 0.0753, + "step": 35471 + }, + { + "epoch": 0.8315195414827968, + "grad_norm": 0.2615366280078888, + "learning_rate": 1.3689267540257122e-05, + "loss": 0.0354, + "step": 35472 + }, + { + "epoch": 0.8315429830576018, + "grad_norm": 0.1198432445526123, + "learning_rate": 1.3685547713413927e-05, + "loss": 0.0191, + "step": 35473 + }, + { + "epoch": 0.8315664246324068, + "grad_norm": 0.16849718987941742, + "learning_rate": 1.3681828354915394e-05, + "loss": 0.0214, + "step": 35474 + }, + { + "epoch": 0.8315898662072118, + "grad_norm": 0.47509482502937317, + "learning_rate": 1.3678109464781686e-05, + "loss": 0.0447, + "step": 35475 + }, + { + "epoch": 0.8316133077820168, + "grad_norm": 0.18722963333129883, + "learning_rate": 1.3674391043032987e-05, + "loss": 0.0386, + "step": 35476 + }, + { + "epoch": 0.8316367493568217, + "grad_norm": 0.3167557716369629, + "learning_rate": 1.3670673089689468e-05, + "loss": 0.0574, + "step": 35477 + }, + { + "epoch": 0.8316601909316268, + "grad_norm": 0.15962769091129303, + "learning_rate": 1.3666955604771304e-05, + "loss": 0.0318, + "step": 35478 + }, + { + "epoch": 0.8316836325064317, + "grad_norm": 0.17636090517044067, + "learning_rate": 1.3663238588298643e-05, + "loss": 0.0103, + "step": 35479 + }, + { + "epoch": 0.8317070740812368, + "grad_norm": 0.3449031412601471, + "learning_rate": 1.3659522040291672e-05, + "loss": 0.0127, + "step": 35480 + }, + { + "epoch": 0.8317305156560417, + "grad_norm": 0.48531582951545715, + "learning_rate": 1.3655805960770596e-05, + "loss": 0.084, + "step": 35481 + }, + { + "epoch": 0.8317539572308468, + "grad_norm": 0.061737705022096634, + "learning_rate": 1.365209034975553e-05, + "loss": 0.0046, + "step": 35482 + }, + { + "epoch": 0.8317773988056517, + "grad_norm": 0.22576549649238586, + "learning_rate": 1.3648375207266661e-05, + "loss": 0.0291, + "step": 35483 + }, + { + "epoch": 0.8318008403804568, + "grad_norm": 0.5153520703315735, + "learning_rate": 1.3644660533324138e-05, + "loss": 0.0902, + "step": 35484 + }, + { + "epoch": 0.8318242819552617, + "grad_norm": 0.6076741814613342, + "learning_rate": 1.3640946327948112e-05, + "loss": 0.1105, + "step": 35485 + }, + { + "epoch": 0.8318477235300668, + "grad_norm": 0.20761893689632416, + "learning_rate": 1.363723259115871e-05, + "loss": 0.0173, + "step": 35486 + }, + { + "epoch": 0.8318711651048717, + "grad_norm": 0.486348956823349, + "learning_rate": 1.363351932297614e-05, + "loss": 0.6803, + "step": 35487 + }, + { + "epoch": 0.8318946066796767, + "grad_norm": 0.4402039647102356, + "learning_rate": 1.362980652342053e-05, + "loss": 0.0817, + "step": 35488 + }, + { + "epoch": 0.8319180482544818, + "grad_norm": 0.4542292356491089, + "learning_rate": 1.3626094192512007e-05, + "loss": 0.4511, + "step": 35489 + }, + { + "epoch": 0.8319414898292867, + "grad_norm": 0.3990617096424103, + "learning_rate": 1.3622382330270732e-05, + "loss": 0.3449, + "step": 35490 + }, + { + "epoch": 0.8319649314040918, + "grad_norm": 0.16719868779182434, + "learning_rate": 1.3618670936716815e-05, + "loss": 0.0259, + "step": 35491 + }, + { + "epoch": 0.8319883729788967, + "grad_norm": 0.3649204671382904, + "learning_rate": 1.3614960011870436e-05, + "loss": 0.0472, + "step": 35492 + }, + { + "epoch": 0.8320118145537018, + "grad_norm": 0.14758513867855072, + "learning_rate": 1.3611249555751693e-05, + "loss": 0.0316, + "step": 35493 + }, + { + "epoch": 0.8320352561285067, + "grad_norm": 0.29965755343437195, + "learning_rate": 1.3607539568380768e-05, + "loss": 0.0414, + "step": 35494 + }, + { + "epoch": 0.8320586977033118, + "grad_norm": 0.23182092607021332, + "learning_rate": 1.3603830049777755e-05, + "loss": 0.0459, + "step": 35495 + }, + { + "epoch": 0.8320821392781167, + "grad_norm": 0.4482352137565613, + "learning_rate": 1.3600120999962796e-05, + "loss": 0.0876, + "step": 35496 + }, + { + "epoch": 0.8321055808529217, + "grad_norm": 0.2568841278553009, + "learning_rate": 1.3596412418955994e-05, + "loss": 0.0467, + "step": 35497 + }, + { + "epoch": 0.8321290224277267, + "grad_norm": 0.6342647671699524, + "learning_rate": 1.3592704306777482e-05, + "loss": 0.1217, + "step": 35498 + }, + { + "epoch": 0.8321524640025317, + "grad_norm": 0.311161607503891, + "learning_rate": 1.3588996663447406e-05, + "loss": 0.0286, + "step": 35499 + }, + { + "epoch": 0.8321759055773367, + "grad_norm": 0.6083036661148071, + "learning_rate": 1.3585289488985864e-05, + "loss": 0.1884, + "step": 35500 + }, + { + "epoch": 0.8321993471521417, + "grad_norm": 0.37946197390556335, + "learning_rate": 1.3581582783412961e-05, + "loss": 0.0473, + "step": 35501 + }, + { + "epoch": 0.8322227887269467, + "grad_norm": 0.46100708842277527, + "learning_rate": 1.3577876546748824e-05, + "loss": 0.06, + "step": 35502 + }, + { + "epoch": 0.8322462303017517, + "grad_norm": 0.18626219034194946, + "learning_rate": 1.3574170779013561e-05, + "loss": 0.0461, + "step": 35503 + }, + { + "epoch": 0.8322696718765566, + "grad_norm": 0.18027539551258087, + "learning_rate": 1.3570465480227256e-05, + "loss": 0.0093, + "step": 35504 + }, + { + "epoch": 0.8322931134513617, + "grad_norm": 0.0716266930103302, + "learning_rate": 1.3566760650410038e-05, + "loss": 0.0069, + "step": 35505 + }, + { + "epoch": 0.8323165550261666, + "grad_norm": 0.6734405159950256, + "learning_rate": 1.3563056289582032e-05, + "loss": 0.151, + "step": 35506 + }, + { + "epoch": 0.8323399966009717, + "grad_norm": 0.2435084581375122, + "learning_rate": 1.3559352397763302e-05, + "loss": 0.0715, + "step": 35507 + }, + { + "epoch": 0.8323634381757766, + "grad_norm": 0.4970356523990631, + "learning_rate": 1.3555648974973966e-05, + "loss": 0.0286, + "step": 35508 + }, + { + "epoch": 0.8323868797505817, + "grad_norm": 0.12879861891269684, + "learning_rate": 1.3551946021234097e-05, + "loss": 0.0181, + "step": 35509 + }, + { + "epoch": 0.8324103213253866, + "grad_norm": 0.4317511022090912, + "learning_rate": 1.354824353656381e-05, + "loss": 0.0807, + "step": 35510 + }, + { + "epoch": 0.8324337629001917, + "grad_norm": 0.3479520380496979, + "learning_rate": 1.3544541520983167e-05, + "loss": 0.0273, + "step": 35511 + }, + { + "epoch": 0.8324572044749966, + "grad_norm": 0.4200077950954437, + "learning_rate": 1.3540839974512288e-05, + "loss": 0.0825, + "step": 35512 + }, + { + "epoch": 0.8324806460498017, + "grad_norm": 0.29765838384628296, + "learning_rate": 1.3537138897171231e-05, + "loss": 0.0578, + "step": 35513 + }, + { + "epoch": 0.8325040876246066, + "grad_norm": 0.397684246301651, + "learning_rate": 1.3533438288980105e-05, + "loss": 0.0589, + "step": 35514 + }, + { + "epoch": 0.8325275291994116, + "grad_norm": 0.08293971419334412, + "learning_rate": 1.352973814995897e-05, + "loss": 0.0074, + "step": 35515 + }, + { + "epoch": 0.8325509707742166, + "grad_norm": 0.4108508229255676, + "learning_rate": 1.3526038480127912e-05, + "loss": 0.0693, + "step": 35516 + }, + { + "epoch": 0.8325744123490216, + "grad_norm": 0.253200501203537, + "learning_rate": 1.3522339279506969e-05, + "loss": 0.051, + "step": 35517 + }, + { + "epoch": 0.8325978539238266, + "grad_norm": 0.45330190658569336, + "learning_rate": 1.3518640548116246e-05, + "loss": 0.0653, + "step": 35518 + }, + { + "epoch": 0.8326212954986316, + "grad_norm": 0.6145191192626953, + "learning_rate": 1.351494228597584e-05, + "loss": 0.5263, + "step": 35519 + }, + { + "epoch": 0.8326447370734366, + "grad_norm": 0.1990761160850525, + "learning_rate": 1.3511244493105779e-05, + "loss": 0.0387, + "step": 35520 + }, + { + "epoch": 0.8326681786482416, + "grad_norm": 0.4820731282234192, + "learning_rate": 1.3507547169526147e-05, + "loss": 0.1158, + "step": 35521 + }, + { + "epoch": 0.8326916202230465, + "grad_norm": 0.2359958440065384, + "learning_rate": 1.3503850315256994e-05, + "loss": 0.0299, + "step": 35522 + }, + { + "epoch": 0.8327150617978516, + "grad_norm": 0.17861661314964294, + "learning_rate": 1.3500153930318371e-05, + "loss": 0.0268, + "step": 35523 + }, + { + "epoch": 0.8327385033726565, + "grad_norm": 0.25880178809165955, + "learning_rate": 1.349645801473034e-05, + "loss": 0.0451, + "step": 35524 + }, + { + "epoch": 0.8327619449474616, + "grad_norm": 0.14332760870456696, + "learning_rate": 1.3492762568512973e-05, + "loss": 0.0218, + "step": 35525 + }, + { + "epoch": 0.8327853865222665, + "grad_norm": 0.25566017627716064, + "learning_rate": 1.3489067591686299e-05, + "loss": 0.0521, + "step": 35526 + }, + { + "epoch": 0.8328088280970716, + "grad_norm": 0.4819473624229431, + "learning_rate": 1.3485373084270392e-05, + "loss": 0.1259, + "step": 35527 + }, + { + "epoch": 0.8328322696718765, + "grad_norm": 0.4910340905189514, + "learning_rate": 1.3481679046285278e-05, + "loss": 0.0914, + "step": 35528 + }, + { + "epoch": 0.8328557112466816, + "grad_norm": 0.9187519550323486, + "learning_rate": 1.3477985477750988e-05, + "loss": 0.1223, + "step": 35529 + }, + { + "epoch": 0.8328791528214865, + "grad_norm": 0.2347835749387741, + "learning_rate": 1.3474292378687593e-05, + "loss": 0.0387, + "step": 35530 + }, + { + "epoch": 0.8329025943962916, + "grad_norm": 0.3122592568397522, + "learning_rate": 1.3470599749115108e-05, + "loss": 0.0226, + "step": 35531 + }, + { + "epoch": 0.8329260359710965, + "grad_norm": 0.33591777086257935, + "learning_rate": 1.34669075890536e-05, + "loss": 0.3395, + "step": 35532 + }, + { + "epoch": 0.8329494775459015, + "grad_norm": 0.4728482961654663, + "learning_rate": 1.3463215898523085e-05, + "loss": 0.0594, + "step": 35533 + }, + { + "epoch": 0.8329729191207065, + "grad_norm": 0.4047900438308716, + "learning_rate": 1.34595246775436e-05, + "loss": 0.051, + "step": 35534 + }, + { + "epoch": 0.8329963606955115, + "grad_norm": 0.3684054911136627, + "learning_rate": 1.3455833926135153e-05, + "loss": 0.0556, + "step": 35535 + }, + { + "epoch": 0.8330198022703165, + "grad_norm": 0.27078840136528015, + "learning_rate": 1.3452143644317772e-05, + "loss": 0.0363, + "step": 35536 + }, + { + "epoch": 0.8330432438451215, + "grad_norm": 0.34551215171813965, + "learning_rate": 1.3448453832111518e-05, + "loss": 0.0487, + "step": 35537 + }, + { + "epoch": 0.8330666854199265, + "grad_norm": 0.29994115233421326, + "learning_rate": 1.3444764489536377e-05, + "loss": 0.0323, + "step": 35538 + }, + { + "epoch": 0.8330901269947315, + "grad_norm": 0.5710854530334473, + "learning_rate": 1.3441075616612387e-05, + "loss": 0.0203, + "step": 35539 + }, + { + "epoch": 0.8331135685695366, + "grad_norm": 0.07895421981811523, + "learning_rate": 1.3437387213359553e-05, + "loss": 0.0137, + "step": 35540 + }, + { + "epoch": 0.8331370101443415, + "grad_norm": 0.471007764339447, + "learning_rate": 1.3433699279797863e-05, + "loss": 0.0748, + "step": 35541 + }, + { + "epoch": 0.8331604517191465, + "grad_norm": 0.1261948198080063, + "learning_rate": 1.3430011815947386e-05, + "loss": 0.0116, + "step": 35542 + }, + { + "epoch": 0.8331838932939515, + "grad_norm": 0.13150827586650848, + "learning_rate": 1.3426324821828062e-05, + "loss": 0.0238, + "step": 35543 + }, + { + "epoch": 0.8332073348687565, + "grad_norm": 0.16711558401584625, + "learning_rate": 1.3422638297459966e-05, + "loss": 0.0317, + "step": 35544 + }, + { + "epoch": 0.8332307764435615, + "grad_norm": 0.37172064185142517, + "learning_rate": 1.341895224286307e-05, + "loss": 0.0402, + "step": 35545 + }, + { + "epoch": 0.8332542180183665, + "grad_norm": 0.4172300100326538, + "learning_rate": 1.3415266658057368e-05, + "loss": 0.0938, + "step": 35546 + }, + { + "epoch": 0.8332776595931715, + "grad_norm": 0.5519633889198303, + "learning_rate": 1.3411581543062867e-05, + "loss": 0.0726, + "step": 35547 + }, + { + "epoch": 0.8333011011679765, + "grad_norm": 0.46008577942848206, + "learning_rate": 1.3407896897899563e-05, + "loss": 0.0567, + "step": 35548 + }, + { + "epoch": 0.8333245427427814, + "grad_norm": 0.1137651726603508, + "learning_rate": 1.3404212722587416e-05, + "loss": 0.0233, + "step": 35549 + }, + { + "epoch": 0.8333479843175865, + "grad_norm": 0.5009006857872009, + "learning_rate": 1.3400529017146479e-05, + "loss": 0.0448, + "step": 35550 + }, + { + "epoch": 0.8333714258923914, + "grad_norm": 0.417360782623291, + "learning_rate": 1.3396845781596689e-05, + "loss": 0.1, + "step": 35551 + }, + { + "epoch": 0.8333948674671965, + "grad_norm": 0.3864533603191376, + "learning_rate": 1.3393163015958066e-05, + "loss": 0.054, + "step": 35552 + }, + { + "epoch": 0.8334183090420014, + "grad_norm": 0.08659445494413376, + "learning_rate": 1.3389480720250546e-05, + "loss": 0.0094, + "step": 35553 + }, + { + "epoch": 0.8334417506168065, + "grad_norm": 0.7216535210609436, + "learning_rate": 1.3385798894494161e-05, + "loss": 0.3227, + "step": 35554 + }, + { + "epoch": 0.8334651921916114, + "grad_norm": 0.30728796124458313, + "learning_rate": 1.338211753870886e-05, + "loss": 0.0552, + "step": 35555 + }, + { + "epoch": 0.8334886337664165, + "grad_norm": 0.18736432492733002, + "learning_rate": 1.3378436652914616e-05, + "loss": 0.0377, + "step": 35556 + }, + { + "epoch": 0.8335120753412214, + "grad_norm": 0.5311655402183533, + "learning_rate": 1.3374756237131425e-05, + "loss": 0.086, + "step": 35557 + }, + { + "epoch": 0.8335355169160265, + "grad_norm": 0.5239221453666687, + "learning_rate": 1.3371076291379236e-05, + "loss": 0.5669, + "step": 35558 + }, + { + "epoch": 0.8335589584908314, + "grad_norm": 0.23268403112888336, + "learning_rate": 1.3367396815678035e-05, + "loss": 0.0486, + "step": 35559 + }, + { + "epoch": 0.8335824000656364, + "grad_norm": 0.3262293338775635, + "learning_rate": 1.336371781004777e-05, + "loss": 0.0626, + "step": 35560 + }, + { + "epoch": 0.8336058416404414, + "grad_norm": 0.39615511894226074, + "learning_rate": 1.3360039274508407e-05, + "loss": 0.0655, + "step": 35561 + }, + { + "epoch": 0.8336292832152464, + "grad_norm": 0.4397823214530945, + "learning_rate": 1.3356361209079881e-05, + "loss": 0.0855, + "step": 35562 + }, + { + "epoch": 0.8336527247900514, + "grad_norm": 0.20286068320274353, + "learning_rate": 1.3352683613782202e-05, + "loss": 0.0282, + "step": 35563 + }, + { + "epoch": 0.8336761663648564, + "grad_norm": 0.36020737886428833, + "learning_rate": 1.3349006488635296e-05, + "loss": 0.0769, + "step": 35564 + }, + { + "epoch": 0.8336996079396614, + "grad_norm": 0.39297574758529663, + "learning_rate": 1.3345329833659126e-05, + "loss": 0.1024, + "step": 35565 + }, + { + "epoch": 0.8337230495144664, + "grad_norm": 0.5501923561096191, + "learning_rate": 1.3341653648873599e-05, + "loss": 0.54, + "step": 35566 + }, + { + "epoch": 0.8337464910892713, + "grad_norm": 0.3062054514884949, + "learning_rate": 1.3337977934298729e-05, + "loss": 0.0411, + "step": 35567 + }, + { + "epoch": 0.8337699326640764, + "grad_norm": 0.18885886669158936, + "learning_rate": 1.3334302689954415e-05, + "loss": 0.0261, + "step": 35568 + }, + { + "epoch": 0.8337933742388813, + "grad_norm": 0.34625932574272156, + "learning_rate": 1.3330627915860594e-05, + "loss": 0.0459, + "step": 35569 + }, + { + "epoch": 0.8338168158136864, + "grad_norm": 0.2891314923763275, + "learning_rate": 1.3326953612037252e-05, + "loss": 0.054, + "step": 35570 + }, + { + "epoch": 0.8338402573884913, + "grad_norm": 0.6133825182914734, + "learning_rate": 1.3323279778504283e-05, + "loss": 0.1149, + "step": 35571 + }, + { + "epoch": 0.8338636989632964, + "grad_norm": 1.791858434677124, + "learning_rate": 1.331960641528165e-05, + "loss": 0.3842, + "step": 35572 + }, + { + "epoch": 0.8338871405381013, + "grad_norm": 0.1268724650144577, + "learning_rate": 1.3315933522389257e-05, + "loss": 0.0154, + "step": 35573 + }, + { + "epoch": 0.8339105821129064, + "grad_norm": 0.6485373973846436, + "learning_rate": 1.3312261099847034e-05, + "loss": 0.0959, + "step": 35574 + }, + { + "epoch": 0.8339340236877113, + "grad_norm": 0.44172969460487366, + "learning_rate": 1.3308589147674944e-05, + "loss": 0.0759, + "step": 35575 + }, + { + "epoch": 0.8339574652625164, + "grad_norm": 0.39894771575927734, + "learning_rate": 1.330491766589288e-05, + "loss": 0.0686, + "step": 35576 + }, + { + "epoch": 0.8339809068373213, + "grad_norm": 0.36195066571235657, + "learning_rate": 1.3301246654520771e-05, + "loss": 0.0601, + "step": 35577 + }, + { + "epoch": 0.8340043484121263, + "grad_norm": 0.5166406631469727, + "learning_rate": 1.3297576113578513e-05, + "loss": 0.0605, + "step": 35578 + }, + { + "epoch": 0.8340277899869313, + "grad_norm": 0.25029268860816956, + "learning_rate": 1.3293906043086057e-05, + "loss": 0.0517, + "step": 35579 + }, + { + "epoch": 0.8340512315617363, + "grad_norm": 0.02035854384303093, + "learning_rate": 1.329023644306332e-05, + "loss": 0.0013, + "step": 35580 + }, + { + "epoch": 0.8340746731365413, + "grad_norm": 0.3241576552391052, + "learning_rate": 1.3286567313530174e-05, + "loss": 0.0346, + "step": 35581 + }, + { + "epoch": 0.8340981147113463, + "grad_norm": 0.7473545074462891, + "learning_rate": 1.328289865450657e-05, + "loss": 0.131, + "step": 35582 + }, + { + "epoch": 0.8341215562861513, + "grad_norm": 0.2903096079826355, + "learning_rate": 1.3279230466012393e-05, + "loss": 0.0454, + "step": 35583 + }, + { + "epoch": 0.8341449978609563, + "grad_norm": 0.07253379374742508, + "learning_rate": 1.327556274806756e-05, + "loss": 0.0056, + "step": 35584 + }, + { + "epoch": 0.8341684394357612, + "grad_norm": 0.3716956675052643, + "learning_rate": 1.3271895500691955e-05, + "loss": 0.0524, + "step": 35585 + }, + { + "epoch": 0.8341918810105663, + "grad_norm": 0.1546209454536438, + "learning_rate": 1.3268228723905485e-05, + "loss": 0.0158, + "step": 35586 + }, + { + "epoch": 0.8342153225853712, + "grad_norm": 0.19136524200439453, + "learning_rate": 1.3264562417728021e-05, + "loss": 0.0259, + "step": 35587 + }, + { + "epoch": 0.8342387641601763, + "grad_norm": 0.31479179859161377, + "learning_rate": 1.3260896582179505e-05, + "loss": 0.0525, + "step": 35588 + }, + { + "epoch": 0.8342622057349812, + "grad_norm": 0.5666862726211548, + "learning_rate": 1.3257231217279798e-05, + "loss": 0.6533, + "step": 35589 + }, + { + "epoch": 0.8342856473097863, + "grad_norm": 0.347332239151001, + "learning_rate": 1.325356632304877e-05, + "loss": 0.0351, + "step": 35590 + }, + { + "epoch": 0.8343090888845913, + "grad_norm": 0.41434261202812195, + "learning_rate": 1.3249901899506367e-05, + "loss": 0.0999, + "step": 35591 + }, + { + "epoch": 0.8343325304593963, + "grad_norm": 0.1416088193655014, + "learning_rate": 1.3246237946672424e-05, + "loss": 0.0228, + "step": 35592 + }, + { + "epoch": 0.8343559720342013, + "grad_norm": 0.40760210156440735, + "learning_rate": 1.3242574464566838e-05, + "loss": 0.0473, + "step": 35593 + }, + { + "epoch": 0.8343794136090062, + "grad_norm": 0.47029170393943787, + "learning_rate": 1.3238911453209457e-05, + "loss": 0.0401, + "step": 35594 + }, + { + "epoch": 0.8344028551838113, + "grad_norm": 0.3848918080329895, + "learning_rate": 1.3235248912620202e-05, + "loss": 0.057, + "step": 35595 + }, + { + "epoch": 0.8344262967586162, + "grad_norm": 0.298030287027359, + "learning_rate": 1.3231586842818922e-05, + "loss": 0.0475, + "step": 35596 + }, + { + "epoch": 0.8344497383334213, + "grad_norm": 0.2970123887062073, + "learning_rate": 1.32279252438255e-05, + "loss": 0.1484, + "step": 35597 + }, + { + "epoch": 0.8344731799082262, + "grad_norm": 0.4484848380088806, + "learning_rate": 1.322426411565979e-05, + "loss": 0.4543, + "step": 35598 + }, + { + "epoch": 0.8344966214830313, + "grad_norm": 0.1911109983921051, + "learning_rate": 1.3220603458341674e-05, + "loss": 0.0209, + "step": 35599 + }, + { + "epoch": 0.8345200630578362, + "grad_norm": 0.5819787979125977, + "learning_rate": 1.3216943271890969e-05, + "loss": 0.0688, + "step": 35600 + }, + { + "epoch": 0.8345435046326413, + "grad_norm": 0.5028584003448486, + "learning_rate": 1.3213283556327594e-05, + "loss": 0.679, + "step": 35601 + }, + { + "epoch": 0.8345669462074462, + "grad_norm": 0.37438416481018066, + "learning_rate": 1.3209624311671365e-05, + "loss": 0.0292, + "step": 35602 + }, + { + "epoch": 0.8345903877822513, + "grad_norm": 0.35447949171066284, + "learning_rate": 1.3205965537942178e-05, + "loss": 0.0597, + "step": 35603 + }, + { + "epoch": 0.8346138293570562, + "grad_norm": 0.27309805154800415, + "learning_rate": 1.3202307235159861e-05, + "loss": 0.031, + "step": 35604 + }, + { + "epoch": 0.8346372709318612, + "grad_norm": 0.31650033593177795, + "learning_rate": 1.3198649403344265e-05, + "loss": 0.0444, + "step": 35605 + }, + { + "epoch": 0.8346607125066662, + "grad_norm": 0.41628944873809814, + "learning_rate": 1.3194992042515231e-05, + "loss": 0.0401, + "step": 35606 + }, + { + "epoch": 0.8346841540814712, + "grad_norm": 0.29619118571281433, + "learning_rate": 1.3191335152692596e-05, + "loss": 0.0619, + "step": 35607 + }, + { + "epoch": 0.8347075956562762, + "grad_norm": 0.1287471354007721, + "learning_rate": 1.3187678733896236e-05, + "loss": 0.0158, + "step": 35608 + }, + { + "epoch": 0.8347310372310812, + "grad_norm": 0.4510459899902344, + "learning_rate": 1.3184022786145977e-05, + "loss": 0.0907, + "step": 35609 + }, + { + "epoch": 0.8347544788058862, + "grad_norm": 0.45289328694343567, + "learning_rate": 1.3180367309461649e-05, + "loss": 0.082, + "step": 35610 + }, + { + "epoch": 0.8347779203806912, + "grad_norm": 0.12898236513137817, + "learning_rate": 1.3176712303863092e-05, + "loss": 0.0102, + "step": 35611 + }, + { + "epoch": 0.8348013619554961, + "grad_norm": 0.2870950400829315, + "learning_rate": 1.3173057769370112e-05, + "loss": 0.0377, + "step": 35612 + }, + { + "epoch": 0.8348248035303012, + "grad_norm": 0.5803459286689758, + "learning_rate": 1.3169403706002581e-05, + "loss": 0.1355, + "step": 35613 + }, + { + "epoch": 0.8348482451051061, + "grad_norm": 0.5488201379776001, + "learning_rate": 1.3165750113780284e-05, + "loss": 0.0853, + "step": 35614 + }, + { + "epoch": 0.8348716866799112, + "grad_norm": 0.4716263711452484, + "learning_rate": 1.3162096992723095e-05, + "loss": 0.0393, + "step": 35615 + }, + { + "epoch": 0.8348951282547161, + "grad_norm": 0.17401468753814697, + "learning_rate": 1.3158444342850807e-05, + "loss": 0.018, + "step": 35616 + }, + { + "epoch": 0.8349185698295212, + "grad_norm": 0.09089156985282898, + "learning_rate": 1.3154792164183238e-05, + "loss": 0.0144, + "step": 35617 + }, + { + "epoch": 0.8349420114043261, + "grad_norm": 0.449564665555954, + "learning_rate": 1.3151140456740207e-05, + "loss": 0.0578, + "step": 35618 + }, + { + "epoch": 0.8349654529791312, + "grad_norm": 0.28874021768569946, + "learning_rate": 1.314748922054152e-05, + "loss": 0.3512, + "step": 35619 + }, + { + "epoch": 0.8349888945539361, + "grad_norm": 0.332354336977005, + "learning_rate": 1.3143838455607017e-05, + "loss": 0.0296, + "step": 35620 + }, + { + "epoch": 0.8350123361287412, + "grad_norm": 0.23377886414527893, + "learning_rate": 1.314018816195648e-05, + "loss": 0.0444, + "step": 35621 + }, + { + "epoch": 0.8350357777035461, + "grad_norm": 0.4293239712715149, + "learning_rate": 1.3136538339609728e-05, + "loss": 0.0581, + "step": 35622 + }, + { + "epoch": 0.8350592192783511, + "grad_norm": 0.20873288810253143, + "learning_rate": 1.3132888988586567e-05, + "loss": 0.0314, + "step": 35623 + }, + { + "epoch": 0.8350826608531561, + "grad_norm": 0.5600927472114563, + "learning_rate": 1.3129240108906782e-05, + "loss": 0.054, + "step": 35624 + }, + { + "epoch": 0.8351061024279611, + "grad_norm": 0.27199798822402954, + "learning_rate": 1.3125591700590167e-05, + "loss": 0.0663, + "step": 35625 + }, + { + "epoch": 0.8351295440027661, + "grad_norm": 0.1489420384168625, + "learning_rate": 1.3121943763656541e-05, + "loss": 0.0293, + "step": 35626 + }, + { + "epoch": 0.8351529855775711, + "grad_norm": 0.3375772535800934, + "learning_rate": 1.311829629812571e-05, + "loss": 0.034, + "step": 35627 + }, + { + "epoch": 0.835176427152376, + "grad_norm": 0.4293399155139923, + "learning_rate": 1.3114649304017435e-05, + "loss": 0.0978, + "step": 35628 + }, + { + "epoch": 0.8351998687271811, + "grad_norm": 0.4277794063091278, + "learning_rate": 1.3111002781351523e-05, + "loss": 0.0666, + "step": 35629 + }, + { + "epoch": 0.835223310301986, + "grad_norm": 0.47195032238960266, + "learning_rate": 1.3107356730147758e-05, + "loss": 0.0948, + "step": 35630 + }, + { + "epoch": 0.8352467518767911, + "grad_norm": 0.09609385579824448, + "learning_rate": 1.3103711150425912e-05, + "loss": 0.0119, + "step": 35631 + }, + { + "epoch": 0.835270193451596, + "grad_norm": 0.6219946146011353, + "learning_rate": 1.310006604220575e-05, + "loss": 0.1319, + "step": 35632 + }, + { + "epoch": 0.8352936350264011, + "grad_norm": 0.6133585572242737, + "learning_rate": 1.3096421405507098e-05, + "loss": 0.0718, + "step": 35633 + }, + { + "epoch": 0.835317076601206, + "grad_norm": 0.2527383863925934, + "learning_rate": 1.3092777240349707e-05, + "loss": 0.056, + "step": 35634 + }, + { + "epoch": 0.8353405181760111, + "grad_norm": 0.2778409719467163, + "learning_rate": 1.3089133546753352e-05, + "loss": 0.0481, + "step": 35635 + }, + { + "epoch": 0.835363959750816, + "grad_norm": 0.19690266251564026, + "learning_rate": 1.3085490324737803e-05, + "loss": 0.0311, + "step": 35636 + }, + { + "epoch": 0.8353874013256211, + "grad_norm": 0.08126234263181686, + "learning_rate": 1.3081847574322825e-05, + "loss": 0.0092, + "step": 35637 + }, + { + "epoch": 0.835410842900426, + "grad_norm": 0.16154004633426666, + "learning_rate": 1.3078205295528167e-05, + "loss": 0.0201, + "step": 35638 + }, + { + "epoch": 0.835434284475231, + "grad_norm": 0.35504478216171265, + "learning_rate": 1.3074563488373604e-05, + "loss": 0.0353, + "step": 35639 + }, + { + "epoch": 0.835457726050036, + "grad_norm": 0.2501865327358246, + "learning_rate": 1.3070922152878928e-05, + "loss": 0.0284, + "step": 35640 + }, + { + "epoch": 0.835481167624841, + "grad_norm": 0.3552130460739136, + "learning_rate": 1.306728128906387e-05, + "loss": 0.0816, + "step": 35641 + }, + { + "epoch": 0.8355046091996461, + "grad_norm": 0.5250107049942017, + "learning_rate": 1.3063640896948193e-05, + "loss": 0.656, + "step": 35642 + }, + { + "epoch": 0.835528050774451, + "grad_norm": 0.364903062582016, + "learning_rate": 1.3060000976551646e-05, + "loss": 0.0547, + "step": 35643 + }, + { + "epoch": 0.8355514923492561, + "grad_norm": 0.2858818471431732, + "learning_rate": 1.3056361527893978e-05, + "loss": 0.0374, + "step": 35644 + }, + { + "epoch": 0.835574933924061, + "grad_norm": 0.567996084690094, + "learning_rate": 1.3052722550994912e-05, + "loss": 0.1375, + "step": 35645 + }, + { + "epoch": 0.8355983754988661, + "grad_norm": 0.5942785739898682, + "learning_rate": 1.3049084045874239e-05, + "loss": 0.6415, + "step": 35646 + }, + { + "epoch": 0.835621817073671, + "grad_norm": 0.25133493542671204, + "learning_rate": 1.3045446012551688e-05, + "loss": 0.0173, + "step": 35647 + }, + { + "epoch": 0.835645258648476, + "grad_norm": 0.09956663101911545, + "learning_rate": 1.3041808451046989e-05, + "loss": 0.0173, + "step": 35648 + }, + { + "epoch": 0.835668700223281, + "grad_norm": 0.15105216205120087, + "learning_rate": 1.303817136137988e-05, + "loss": 0.0234, + "step": 35649 + }, + { + "epoch": 0.835692141798086, + "grad_norm": 0.39821118116378784, + "learning_rate": 1.303453474357007e-05, + "loss": 0.1095, + "step": 35650 + }, + { + "epoch": 0.835715583372891, + "grad_norm": 0.6135725378990173, + "learning_rate": 1.3030898597637352e-05, + "loss": 0.17, + "step": 35651 + }, + { + "epoch": 0.835739024947696, + "grad_norm": 0.317283570766449, + "learning_rate": 1.3027262923601403e-05, + "loss": 0.0396, + "step": 35652 + }, + { + "epoch": 0.835762466522501, + "grad_norm": 0.2123904973268509, + "learning_rate": 1.3023627721481979e-05, + "loss": 0.0159, + "step": 35653 + }, + { + "epoch": 0.835785908097306, + "grad_norm": 0.6350081562995911, + "learning_rate": 1.3019992991298813e-05, + "loss": 0.1539, + "step": 35654 + }, + { + "epoch": 0.835809349672111, + "grad_norm": 0.15856729447841644, + "learning_rate": 1.30163587330716e-05, + "loss": 0.0341, + "step": 35655 + }, + { + "epoch": 0.835832791246916, + "grad_norm": 0.2751205563545227, + "learning_rate": 1.3012724946820077e-05, + "loss": 0.0466, + "step": 35656 + }, + { + "epoch": 0.835856232821721, + "grad_norm": 0.5207450985908508, + "learning_rate": 1.300909163256393e-05, + "loss": 0.0947, + "step": 35657 + }, + { + "epoch": 0.835879674396526, + "grad_norm": 0.4372881054878235, + "learning_rate": 1.300545879032291e-05, + "loss": 0.079, + "step": 35658 + }, + { + "epoch": 0.8359031159713309, + "grad_norm": 0.09197814762592316, + "learning_rate": 1.3001826420116725e-05, + "loss": 0.0128, + "step": 35659 + }, + { + "epoch": 0.835926557546136, + "grad_norm": 0.4345740079879761, + "learning_rate": 1.2998194521965079e-05, + "loss": 0.0788, + "step": 35660 + }, + { + "epoch": 0.8359499991209409, + "grad_norm": 0.10293491929769516, + "learning_rate": 1.2994563095887668e-05, + "loss": 0.0095, + "step": 35661 + }, + { + "epoch": 0.835973440695746, + "grad_norm": 0.34629449248313904, + "learning_rate": 1.2990932141904211e-05, + "loss": 0.0468, + "step": 35662 + }, + { + "epoch": 0.8359968822705509, + "grad_norm": 0.8247362375259399, + "learning_rate": 1.298730166003438e-05, + "loss": 0.112, + "step": 35663 + }, + { + "epoch": 0.836020323845356, + "grad_norm": 0.378999799489975, + "learning_rate": 1.2983671650297902e-05, + "loss": 0.0586, + "step": 35664 + }, + { + "epoch": 0.8360437654201609, + "grad_norm": 0.28394362330436707, + "learning_rate": 1.2980042112714474e-05, + "loss": 0.0484, + "step": 35665 + }, + { + "epoch": 0.836067206994966, + "grad_norm": 0.27152740955352783, + "learning_rate": 1.2976413047303804e-05, + "loss": 0.029, + "step": 35666 + }, + { + "epoch": 0.8360906485697709, + "grad_norm": 0.45444944500923157, + "learning_rate": 1.297278445408555e-05, + "loss": 0.0811, + "step": 35667 + }, + { + "epoch": 0.8361140901445759, + "grad_norm": 0.39697766304016113, + "learning_rate": 1.2969156333079414e-05, + "loss": 0.0481, + "step": 35668 + }, + { + "epoch": 0.8361375317193809, + "grad_norm": 0.058214545249938965, + "learning_rate": 1.2965528684305095e-05, + "loss": 0.0093, + "step": 35669 + }, + { + "epoch": 0.8361609732941859, + "grad_norm": 0.6320633292198181, + "learning_rate": 1.2961901507782226e-05, + "loss": 0.4381, + "step": 35670 + }, + { + "epoch": 0.8361844148689909, + "grad_norm": 0.3818911612033844, + "learning_rate": 1.2958274803530556e-05, + "loss": 0.0717, + "step": 35671 + }, + { + "epoch": 0.8362078564437959, + "grad_norm": 0.42812666296958923, + "learning_rate": 1.2954648571569738e-05, + "loss": 0.0744, + "step": 35672 + }, + { + "epoch": 0.8362312980186009, + "grad_norm": 0.6292544007301331, + "learning_rate": 1.2951022811919444e-05, + "loss": 0.1624, + "step": 35673 + }, + { + "epoch": 0.8362547395934059, + "grad_norm": 0.3334861099720001, + "learning_rate": 1.2947397524599347e-05, + "loss": 0.0338, + "step": 35674 + }, + { + "epoch": 0.8362781811682108, + "grad_norm": 0.4742435812950134, + "learning_rate": 1.2943772709629099e-05, + "loss": 0.0807, + "step": 35675 + }, + { + "epoch": 0.8363016227430159, + "grad_norm": 0.3329545557498932, + "learning_rate": 1.2940148367028403e-05, + "loss": 0.0709, + "step": 35676 + }, + { + "epoch": 0.8363250643178208, + "grad_norm": 0.6604623794555664, + "learning_rate": 1.2936524496816893e-05, + "loss": 0.1554, + "step": 35677 + }, + { + "epoch": 0.8363485058926259, + "grad_norm": 0.7028256058692932, + "learning_rate": 1.2932901099014273e-05, + "loss": 0.1404, + "step": 35678 + }, + { + "epoch": 0.8363719474674308, + "grad_norm": 0.5813405513763428, + "learning_rate": 1.2929278173640168e-05, + "loss": 0.0603, + "step": 35679 + }, + { + "epoch": 0.8363953890422359, + "grad_norm": 0.15287208557128906, + "learning_rate": 1.2925655720714259e-05, + "loss": 0.029, + "step": 35680 + }, + { + "epoch": 0.8364188306170408, + "grad_norm": 0.2673890292644501, + "learning_rate": 1.292203374025619e-05, + "loss": 0.0286, + "step": 35681 + }, + { + "epoch": 0.8364422721918459, + "grad_norm": 0.1742739975452423, + "learning_rate": 1.2918412232285615e-05, + "loss": 0.0295, + "step": 35682 + }, + { + "epoch": 0.8364657137666508, + "grad_norm": 0.5144211053848267, + "learning_rate": 1.2914791196822152e-05, + "loss": 0.1476, + "step": 35683 + }, + { + "epoch": 0.8364891553414558, + "grad_norm": 0.07821688055992126, + "learning_rate": 1.2911170633885516e-05, + "loss": 0.0112, + "step": 35684 + }, + { + "epoch": 0.8365125969162608, + "grad_norm": 0.6349289417266846, + "learning_rate": 1.2907550543495306e-05, + "loss": 0.4321, + "step": 35685 + }, + { + "epoch": 0.8365360384910658, + "grad_norm": 0.18728701770305634, + "learning_rate": 1.2903930925671182e-05, + "loss": 0.0213, + "step": 35686 + }, + { + "epoch": 0.8365594800658708, + "grad_norm": 0.3162713050842285, + "learning_rate": 1.290031178043275e-05, + "loss": 0.0473, + "step": 35687 + }, + { + "epoch": 0.8365829216406758, + "grad_norm": 0.5377033948898315, + "learning_rate": 1.2896693107799707e-05, + "loss": 0.1058, + "step": 35688 + }, + { + "epoch": 0.8366063632154808, + "grad_norm": 0.4106149971485138, + "learning_rate": 1.2893074907791624e-05, + "loss": 0.06, + "step": 35689 + }, + { + "epoch": 0.8366298047902858, + "grad_norm": 0.48441797494888306, + "learning_rate": 1.2889457180428189e-05, + "loss": 0.1472, + "step": 35690 + }, + { + "epoch": 0.8366532463650908, + "grad_norm": 0.5252445936203003, + "learning_rate": 1.2885839925729014e-05, + "loss": 0.5476, + "step": 35691 + }, + { + "epoch": 0.8366766879398958, + "grad_norm": 0.30589261651039124, + "learning_rate": 1.2882223143713723e-05, + "loss": 0.0518, + "step": 35692 + }, + { + "epoch": 0.8367001295147009, + "grad_norm": 0.3705272972583771, + "learning_rate": 1.287860683440193e-05, + "loss": 0.0638, + "step": 35693 + }, + { + "epoch": 0.8367235710895058, + "grad_norm": 0.3193657696247101, + "learning_rate": 1.2874990997813274e-05, + "loss": 0.0327, + "step": 35694 + }, + { + "epoch": 0.8367470126643108, + "grad_norm": 0.15417265892028809, + "learning_rate": 1.2871375633967342e-05, + "loss": 0.0229, + "step": 35695 + }, + { + "epoch": 0.8367704542391158, + "grad_norm": 0.329988956451416, + "learning_rate": 1.2867760742883795e-05, + "loss": 0.0398, + "step": 35696 + }, + { + "epoch": 0.8367938958139208, + "grad_norm": 0.5528362989425659, + "learning_rate": 1.2864146324582228e-05, + "loss": 0.1018, + "step": 35697 + }, + { + "epoch": 0.8368173373887258, + "grad_norm": 0.16760219633579254, + "learning_rate": 1.2860532379082258e-05, + "loss": 0.037, + "step": 35698 + }, + { + "epoch": 0.8368407789635308, + "grad_norm": 0.1356881558895111, + "learning_rate": 1.2856918906403458e-05, + "loss": 0.0244, + "step": 35699 + }, + { + "epoch": 0.8368642205383358, + "grad_norm": 0.2217017561197281, + "learning_rate": 1.2853305906565494e-05, + "loss": 0.0253, + "step": 35700 + }, + { + "epoch": 0.8368876621131408, + "grad_norm": 0.7951138019561768, + "learning_rate": 1.2849693379587947e-05, + "loss": 0.0871, + "step": 35701 + }, + { + "epoch": 0.8369111036879457, + "grad_norm": 0.3343460261821747, + "learning_rate": 1.284608132549039e-05, + "loss": 0.0592, + "step": 35702 + }, + { + "epoch": 0.8369345452627508, + "grad_norm": 0.1521981954574585, + "learning_rate": 1.2842469744292462e-05, + "loss": 0.0289, + "step": 35703 + }, + { + "epoch": 0.8369579868375557, + "grad_norm": 0.9051623344421387, + "learning_rate": 1.283885863601375e-05, + "loss": 0.14, + "step": 35704 + }, + { + "epoch": 0.8369814284123608, + "grad_norm": 0.2088419646024704, + "learning_rate": 1.2835248000673838e-05, + "loss": 0.0492, + "step": 35705 + }, + { + "epoch": 0.8370048699871657, + "grad_norm": 0.5002136826515198, + "learning_rate": 1.2831637838292322e-05, + "loss": 0.0702, + "step": 35706 + }, + { + "epoch": 0.8370283115619708, + "grad_norm": 0.537469208240509, + "learning_rate": 1.2828028148888793e-05, + "loss": 0.0707, + "step": 35707 + }, + { + "epoch": 0.8370517531367757, + "grad_norm": 0.48815926909446716, + "learning_rate": 1.2824418932482806e-05, + "loss": 0.37, + "step": 35708 + }, + { + "epoch": 0.8370751947115808, + "grad_norm": 0.5492148399353027, + "learning_rate": 1.2820810189094002e-05, + "loss": 0.4677, + "step": 35709 + }, + { + "epoch": 0.8370986362863857, + "grad_norm": 0.3487105667591095, + "learning_rate": 1.2817201918741927e-05, + "loss": 0.0531, + "step": 35710 + }, + { + "epoch": 0.8371220778611907, + "grad_norm": 0.3745720386505127, + "learning_rate": 1.2813594121446148e-05, + "loss": 0.0542, + "step": 35711 + }, + { + "epoch": 0.8371455194359957, + "grad_norm": 0.49042388796806335, + "learning_rate": 1.2809986797226281e-05, + "loss": 0.0636, + "step": 35712 + }, + { + "epoch": 0.8371689610108007, + "grad_norm": 0.34011906385421753, + "learning_rate": 1.2806379946101876e-05, + "loss": 0.249, + "step": 35713 + }, + { + "epoch": 0.8371924025856057, + "grad_norm": 0.331149697303772, + "learning_rate": 1.2802773568092507e-05, + "loss": 0.0655, + "step": 35714 + }, + { + "epoch": 0.8372158441604107, + "grad_norm": 0.5278723239898682, + "learning_rate": 1.2799167663217715e-05, + "loss": 0.3309, + "step": 35715 + }, + { + "epoch": 0.8372392857352157, + "grad_norm": 0.5718372464179993, + "learning_rate": 1.2795562231497116e-05, + "loss": 0.4407, + "step": 35716 + }, + { + "epoch": 0.8372627273100207, + "grad_norm": 0.35153812170028687, + "learning_rate": 1.279195727295025e-05, + "loss": 0.0391, + "step": 35717 + }, + { + "epoch": 0.8372861688848257, + "grad_norm": 0.3365703821182251, + "learning_rate": 1.278835278759668e-05, + "loss": 0.0502, + "step": 35718 + }, + { + "epoch": 0.8373096104596307, + "grad_norm": 0.1830347180366516, + "learning_rate": 1.2784748775455945e-05, + "loss": 0.0349, + "step": 35719 + }, + { + "epoch": 0.8373330520344356, + "grad_norm": 0.5167121291160583, + "learning_rate": 1.2781145236547632e-05, + "loss": 0.0837, + "step": 35720 + }, + { + "epoch": 0.8373564936092407, + "grad_norm": 0.3787849545478821, + "learning_rate": 1.2777542170891244e-05, + "loss": 0.0696, + "step": 35721 + }, + { + "epoch": 0.8373799351840456, + "grad_norm": 0.5002890825271606, + "learning_rate": 1.277393957850639e-05, + "loss": 0.0793, + "step": 35722 + }, + { + "epoch": 0.8374033767588507, + "grad_norm": 0.2349546253681183, + "learning_rate": 1.2770337459412585e-05, + "loss": 0.0356, + "step": 35723 + }, + { + "epoch": 0.8374268183336556, + "grad_norm": 0.7497057318687439, + "learning_rate": 1.2766735813629371e-05, + "loss": 0.0525, + "step": 35724 + }, + { + "epoch": 0.8374502599084607, + "grad_norm": 0.41090092062950134, + "learning_rate": 1.2763134641176322e-05, + "loss": 0.5763, + "step": 35725 + }, + { + "epoch": 0.8374737014832656, + "grad_norm": 0.08146756142377853, + "learning_rate": 1.2759533942072955e-05, + "loss": 0.0062, + "step": 35726 + }, + { + "epoch": 0.8374971430580707, + "grad_norm": 0.18572212755680084, + "learning_rate": 1.2755933716338785e-05, + "loss": 0.0328, + "step": 35727 + }, + { + "epoch": 0.8375205846328756, + "grad_norm": 0.7443866729736328, + "learning_rate": 1.275233396399339e-05, + "loss": 0.1211, + "step": 35728 + }, + { + "epoch": 0.8375440262076806, + "grad_norm": 0.3540724813938141, + "learning_rate": 1.2748734685056295e-05, + "loss": 0.0585, + "step": 35729 + }, + { + "epoch": 0.8375674677824856, + "grad_norm": 0.26124972105026245, + "learning_rate": 1.274513587954701e-05, + "loss": 0.0336, + "step": 35730 + }, + { + "epoch": 0.8375909093572906, + "grad_norm": 0.6246309876441956, + "learning_rate": 1.2741537547485083e-05, + "loss": 0.1257, + "step": 35731 + }, + { + "epoch": 0.8376143509320956, + "grad_norm": 0.4182003140449524, + "learning_rate": 1.2737939688890011e-05, + "loss": 0.0694, + "step": 35732 + }, + { + "epoch": 0.8376377925069006, + "grad_norm": 0.4227449893951416, + "learning_rate": 1.2734342303781322e-05, + "loss": 0.0471, + "step": 35733 + }, + { + "epoch": 0.8376612340817056, + "grad_norm": 0.41543644666671753, + "learning_rate": 1.2730745392178567e-05, + "loss": 0.0539, + "step": 35734 + }, + { + "epoch": 0.8376846756565106, + "grad_norm": 0.5207611918449402, + "learning_rate": 1.272714895410123e-05, + "loss": 0.0843, + "step": 35735 + }, + { + "epoch": 0.8377081172313156, + "grad_norm": 0.40755710005760193, + "learning_rate": 1.272355298956882e-05, + "loss": 0.1066, + "step": 35736 + }, + { + "epoch": 0.8377315588061206, + "grad_norm": 0.37673836946487427, + "learning_rate": 1.2719957498600899e-05, + "loss": 0.0133, + "step": 35737 + }, + { + "epoch": 0.8377550003809255, + "grad_norm": 0.5707099437713623, + "learning_rate": 1.2716362481216926e-05, + "loss": 0.1246, + "step": 35738 + }, + { + "epoch": 0.8377784419557306, + "grad_norm": 0.08996262401342392, + "learning_rate": 1.2712767937436431e-05, + "loss": 0.0113, + "step": 35739 + }, + { + "epoch": 0.8378018835305355, + "grad_norm": 0.40188542008399963, + "learning_rate": 1.270917386727889e-05, + "loss": 0.0509, + "step": 35740 + }, + { + "epoch": 0.8378253251053406, + "grad_norm": 0.6241859197616577, + "learning_rate": 1.270558027076385e-05, + "loss": 0.0925, + "step": 35741 + }, + { + "epoch": 0.8378487666801455, + "grad_norm": 0.5114964842796326, + "learning_rate": 1.2701987147910788e-05, + "loss": 0.0817, + "step": 35742 + }, + { + "epoch": 0.8378722082549506, + "grad_norm": 0.338608980178833, + "learning_rate": 1.2698394498739185e-05, + "loss": 0.0672, + "step": 35743 + }, + { + "epoch": 0.8378956498297556, + "grad_norm": 0.3947877883911133, + "learning_rate": 1.2694802323268562e-05, + "loss": 0.0532, + "step": 35744 + }, + { + "epoch": 0.8379190914045606, + "grad_norm": 0.4619762599468231, + "learning_rate": 1.2691210621518391e-05, + "loss": 0.0765, + "step": 35745 + }, + { + "epoch": 0.8379425329793656, + "grad_norm": 0.43684205412864685, + "learning_rate": 1.2687619393508143e-05, + "loss": 0.0942, + "step": 35746 + }, + { + "epoch": 0.8379659745541705, + "grad_norm": 0.2449486404657364, + "learning_rate": 1.268402863925735e-05, + "loss": 0.0402, + "step": 35747 + }, + { + "epoch": 0.8379894161289756, + "grad_norm": 0.31215983629226685, + "learning_rate": 1.2680438358785452e-05, + "loss": 0.0553, + "step": 35748 + }, + { + "epoch": 0.8380128577037805, + "grad_norm": 0.20965878665447235, + "learning_rate": 1.2676848552111975e-05, + "loss": 0.0325, + "step": 35749 + }, + { + "epoch": 0.8380362992785856, + "grad_norm": 0.27561506628990173, + "learning_rate": 1.2673259219256373e-05, + "loss": 0.0435, + "step": 35750 + }, + { + "epoch": 0.8380597408533905, + "grad_norm": 0.5180766582489014, + "learning_rate": 1.2669670360238118e-05, + "loss": 0.1071, + "step": 35751 + }, + { + "epoch": 0.8380831824281956, + "grad_norm": 0.5532217025756836, + "learning_rate": 1.2666081975076694e-05, + "loss": 0.1062, + "step": 35752 + }, + { + "epoch": 0.8381066240030005, + "grad_norm": 0.5139483213424683, + "learning_rate": 1.266249406379153e-05, + "loss": 0.0726, + "step": 35753 + }, + { + "epoch": 0.8381300655778056, + "grad_norm": 1.557496428489685, + "learning_rate": 1.2658906626402155e-05, + "loss": 0.3997, + "step": 35754 + }, + { + "epoch": 0.8381535071526105, + "grad_norm": 0.3442184031009674, + "learning_rate": 1.2655319662928011e-05, + "loss": 0.0527, + "step": 35755 + }, + { + "epoch": 0.8381769487274155, + "grad_norm": 0.30430492758750916, + "learning_rate": 1.2651733173388569e-05, + "loss": 0.056, + "step": 35756 + }, + { + "epoch": 0.8382003903022205, + "grad_norm": 0.34961506724357605, + "learning_rate": 1.264814715780327e-05, + "loss": 0.07, + "step": 35757 + }, + { + "epoch": 0.8382238318770255, + "grad_norm": 0.3286442160606384, + "learning_rate": 1.2644561616191552e-05, + "loss": 0.0455, + "step": 35758 + }, + { + "epoch": 0.8382472734518305, + "grad_norm": 0.6196757555007935, + "learning_rate": 1.2640976548572925e-05, + "loss": 0.1072, + "step": 35759 + }, + { + "epoch": 0.8382707150266355, + "grad_norm": 0.4427172541618347, + "learning_rate": 1.2637391954966792e-05, + "loss": 0.0326, + "step": 35760 + }, + { + "epoch": 0.8382941566014405, + "grad_norm": 0.523634135723114, + "learning_rate": 1.2633807835392653e-05, + "loss": 0.09, + "step": 35761 + }, + { + "epoch": 0.8383175981762455, + "grad_norm": 0.5109749436378479, + "learning_rate": 1.2630224189869932e-05, + "loss": 0.1154, + "step": 35762 + }, + { + "epoch": 0.8383410397510505, + "grad_norm": 0.38547617197036743, + "learning_rate": 1.2626641018418061e-05, + "loss": 0.3818, + "step": 35763 + }, + { + "epoch": 0.8383644813258555, + "grad_norm": 0.5243650674819946, + "learning_rate": 1.2623058321056502e-05, + "loss": 0.2635, + "step": 35764 + }, + { + "epoch": 0.8383879229006604, + "grad_norm": 0.13560554385185242, + "learning_rate": 1.2619476097804661e-05, + "loss": 0.0109, + "step": 35765 + }, + { + "epoch": 0.8384113644754655, + "grad_norm": 0.8564973473548889, + "learning_rate": 1.2615894348682022e-05, + "loss": 0.177, + "step": 35766 + }, + { + "epoch": 0.8384348060502704, + "grad_norm": 0.4429014027118683, + "learning_rate": 1.2612313073707992e-05, + "loss": 0.0768, + "step": 35767 + }, + { + "epoch": 0.8384582476250755, + "grad_norm": 0.27828770875930786, + "learning_rate": 1.2608732272902012e-05, + "loss": 0.0615, + "step": 35768 + }, + { + "epoch": 0.8384816891998804, + "grad_norm": 0.3513472378253937, + "learning_rate": 1.2605151946283511e-05, + "loss": 0.0827, + "step": 35769 + }, + { + "epoch": 0.8385051307746855, + "grad_norm": 0.3429851233959198, + "learning_rate": 1.2601572093871904e-05, + "loss": 0.0427, + "step": 35770 + }, + { + "epoch": 0.8385285723494904, + "grad_norm": 0.1592012643814087, + "learning_rate": 1.2597992715686613e-05, + "loss": 0.0251, + "step": 35771 + }, + { + "epoch": 0.8385520139242955, + "grad_norm": 0.43801283836364746, + "learning_rate": 1.2594413811747074e-05, + "loss": 0.5952, + "step": 35772 + }, + { + "epoch": 0.8385754554991004, + "grad_norm": 0.4430859088897705, + "learning_rate": 1.2590835382072707e-05, + "loss": 0.06, + "step": 35773 + }, + { + "epoch": 0.8385988970739054, + "grad_norm": 0.3508794903755188, + "learning_rate": 1.2587257426682942e-05, + "loss": 0.0586, + "step": 35774 + }, + { + "epoch": 0.8386223386487104, + "grad_norm": 0.4485441744327545, + "learning_rate": 1.258367994559717e-05, + "loss": 0.0577, + "step": 35775 + }, + { + "epoch": 0.8386457802235154, + "grad_norm": 0.5754098296165466, + "learning_rate": 1.2580102938834804e-05, + "loss": 0.1119, + "step": 35776 + }, + { + "epoch": 0.8386692217983204, + "grad_norm": 0.44721388816833496, + "learning_rate": 1.2576526406415268e-05, + "loss": 0.0867, + "step": 35777 + }, + { + "epoch": 0.8386926633731254, + "grad_norm": 0.6012942790985107, + "learning_rate": 1.2572950348357925e-05, + "loss": 0.1031, + "step": 35778 + }, + { + "epoch": 0.8387161049479304, + "grad_norm": 0.45197537541389465, + "learning_rate": 1.256937476468224e-05, + "loss": 0.0611, + "step": 35779 + }, + { + "epoch": 0.8387395465227354, + "grad_norm": 0.53327476978302, + "learning_rate": 1.2565799655407584e-05, + "loss": 0.0961, + "step": 35780 + }, + { + "epoch": 0.8387629880975404, + "grad_norm": 0.4977330267429352, + "learning_rate": 1.2562225020553353e-05, + "loss": 0.0853, + "step": 35781 + }, + { + "epoch": 0.8387864296723454, + "grad_norm": 0.123058021068573, + "learning_rate": 1.2558650860138955e-05, + "loss": 0.0343, + "step": 35782 + }, + { + "epoch": 0.8388098712471503, + "grad_norm": 0.17993277311325073, + "learning_rate": 1.2555077174183772e-05, + "loss": 0.014, + "step": 35783 + }, + { + "epoch": 0.8388333128219554, + "grad_norm": 0.3894564211368561, + "learning_rate": 1.2551503962707178e-05, + "loss": 0.0477, + "step": 35784 + }, + { + "epoch": 0.8388567543967603, + "grad_norm": 0.40489283204078674, + "learning_rate": 1.2547931225728582e-05, + "loss": 0.0814, + "step": 35785 + }, + { + "epoch": 0.8388801959715654, + "grad_norm": 0.3577871024608612, + "learning_rate": 1.2544358963267399e-05, + "loss": 0.0725, + "step": 35786 + }, + { + "epoch": 0.8389036375463703, + "grad_norm": 0.4581041932106018, + "learning_rate": 1.2540787175342972e-05, + "loss": 0.1279, + "step": 35787 + }, + { + "epoch": 0.8389270791211754, + "grad_norm": 0.5444358587265015, + "learning_rate": 1.2537215861974693e-05, + "loss": 0.0631, + "step": 35788 + }, + { + "epoch": 0.8389505206959803, + "grad_norm": 0.16174496710300446, + "learning_rate": 1.253364502318195e-05, + "loss": 0.0267, + "step": 35789 + }, + { + "epoch": 0.8389739622707854, + "grad_norm": 0.1803349405527115, + "learning_rate": 1.2530074658984104e-05, + "loss": 0.028, + "step": 35790 + }, + { + "epoch": 0.8389974038455903, + "grad_norm": 0.4793696999549866, + "learning_rate": 1.2526504769400504e-05, + "loss": 0.1068, + "step": 35791 + }, + { + "epoch": 0.8390208454203953, + "grad_norm": 0.35724836587905884, + "learning_rate": 1.2522935354450582e-05, + "loss": 0.0657, + "step": 35792 + }, + { + "epoch": 0.8390442869952003, + "grad_norm": 0.2773135006427765, + "learning_rate": 1.2519366414153666e-05, + "loss": 0.0323, + "step": 35793 + }, + { + "epoch": 0.8390677285700053, + "grad_norm": 0.23446901142597198, + "learning_rate": 1.2515797948529129e-05, + "loss": 0.0117, + "step": 35794 + }, + { + "epoch": 0.8390911701448104, + "grad_norm": 0.14762361347675323, + "learning_rate": 1.2512229957596334e-05, + "loss": 0.019, + "step": 35795 + }, + { + "epoch": 0.8391146117196153, + "grad_norm": 0.36208343505859375, + "learning_rate": 1.2508662441374609e-05, + "loss": 0.0567, + "step": 35796 + }, + { + "epoch": 0.8391380532944204, + "grad_norm": 0.6918135285377502, + "learning_rate": 1.2505095399883371e-05, + "loss": 0.6216, + "step": 35797 + }, + { + "epoch": 0.8391614948692253, + "grad_norm": 0.09954127669334412, + "learning_rate": 1.2501528833141917e-05, + "loss": 0.0053, + "step": 35798 + }, + { + "epoch": 0.8391849364440304, + "grad_norm": 0.6772552728652954, + "learning_rate": 1.2497962741169655e-05, + "loss": 0.1303, + "step": 35799 + }, + { + "epoch": 0.8392083780188353, + "grad_norm": 0.1882620006799698, + "learning_rate": 1.2494397123985913e-05, + "loss": 0.0327, + "step": 35800 + }, + { + "epoch": 0.8392318195936403, + "grad_norm": 0.2905590236186981, + "learning_rate": 1.2490831981610029e-05, + "loss": 0.034, + "step": 35801 + }, + { + "epoch": 0.8392552611684453, + "grad_norm": 0.6748525500297546, + "learning_rate": 1.2487267314061346e-05, + "loss": 0.0877, + "step": 35802 + }, + { + "epoch": 0.8392787027432503, + "grad_norm": 0.1393069177865982, + "learning_rate": 1.2483703121359191e-05, + "loss": 0.0183, + "step": 35803 + }, + { + "epoch": 0.8393021443180553, + "grad_norm": 0.4227527976036072, + "learning_rate": 1.248013940352295e-05, + "loss": 0.0922, + "step": 35804 + }, + { + "epoch": 0.8393255858928603, + "grad_norm": 0.1719624400138855, + "learning_rate": 1.2476576160571929e-05, + "loss": 0.0088, + "step": 35805 + }, + { + "epoch": 0.8393490274676653, + "grad_norm": 0.29290783405303955, + "learning_rate": 1.2473013392525467e-05, + "loss": 0.0433, + "step": 35806 + }, + { + "epoch": 0.8393724690424703, + "grad_norm": 0.28994283080101013, + "learning_rate": 1.2469451099402895e-05, + "loss": 0.0323, + "step": 35807 + }, + { + "epoch": 0.8393959106172753, + "grad_norm": 0.49208328127861023, + "learning_rate": 1.2465889281223541e-05, + "loss": 0.058, + "step": 35808 + }, + { + "epoch": 0.8394193521920803, + "grad_norm": 0.7295680046081543, + "learning_rate": 1.2462327938006713e-05, + "loss": 0.1539, + "step": 35809 + }, + { + "epoch": 0.8394427937668852, + "grad_norm": 0.5627496242523193, + "learning_rate": 1.2458767069771759e-05, + "loss": 0.1166, + "step": 35810 + }, + { + "epoch": 0.8394662353416903, + "grad_norm": 0.5438691973686218, + "learning_rate": 1.2455206676538012e-05, + "loss": 0.0384, + "step": 35811 + }, + { + "epoch": 0.8394896769164952, + "grad_norm": 0.2144843190908432, + "learning_rate": 1.2451646758324776e-05, + "loss": 0.0407, + "step": 35812 + }, + { + "epoch": 0.8395131184913003, + "grad_norm": 0.5453482866287231, + "learning_rate": 1.244808731515137e-05, + "loss": 0.0741, + "step": 35813 + }, + { + "epoch": 0.8395365600661052, + "grad_norm": 0.2785923182964325, + "learning_rate": 1.244452834703711e-05, + "loss": 0.0292, + "step": 35814 + }, + { + "epoch": 0.8395600016409103, + "grad_norm": 0.3307693302631378, + "learning_rate": 1.2440969854001283e-05, + "loss": 0.0415, + "step": 35815 + }, + { + "epoch": 0.8395834432157152, + "grad_norm": 0.4026378095149994, + "learning_rate": 1.2437411836063207e-05, + "loss": 0.4097, + "step": 35816 + }, + { + "epoch": 0.8396068847905203, + "grad_norm": 0.21159712970256805, + "learning_rate": 1.2433854293242219e-05, + "loss": 0.032, + "step": 35817 + }, + { + "epoch": 0.8396303263653252, + "grad_norm": 0.26762279868125916, + "learning_rate": 1.2430297225557597e-05, + "loss": 0.0315, + "step": 35818 + }, + { + "epoch": 0.8396537679401302, + "grad_norm": 0.17002148926258087, + "learning_rate": 1.2426740633028644e-05, + "loss": 0.0289, + "step": 35819 + }, + { + "epoch": 0.8396772095149352, + "grad_norm": 0.13212472200393677, + "learning_rate": 1.2423184515674657e-05, + "loss": 0.0145, + "step": 35820 + }, + { + "epoch": 0.8397006510897402, + "grad_norm": 0.35006511211395264, + "learning_rate": 1.2419628873514911e-05, + "loss": 0.0391, + "step": 35821 + }, + { + "epoch": 0.8397240926645452, + "grad_norm": 0.20449164509773254, + "learning_rate": 1.2416073706568743e-05, + "loss": 0.0277, + "step": 35822 + }, + { + "epoch": 0.8397475342393502, + "grad_norm": 0.2515605390071869, + "learning_rate": 1.2412519014855407e-05, + "loss": 0.0434, + "step": 35823 + }, + { + "epoch": 0.8397709758141552, + "grad_norm": 0.26448437571525574, + "learning_rate": 1.2408964798394217e-05, + "loss": 0.0499, + "step": 35824 + }, + { + "epoch": 0.8397944173889602, + "grad_norm": 0.6126567721366882, + "learning_rate": 1.2405411057204453e-05, + "loss": 0.1181, + "step": 35825 + }, + { + "epoch": 0.8398178589637652, + "grad_norm": 0.4560796022415161, + "learning_rate": 1.2401857791305394e-05, + "loss": 0.0756, + "step": 35826 + }, + { + "epoch": 0.8398413005385702, + "grad_norm": 0.3959432542324066, + "learning_rate": 1.2398305000716315e-05, + "loss": 0.0665, + "step": 35827 + }, + { + "epoch": 0.8398647421133751, + "grad_norm": 0.6284880042076111, + "learning_rate": 1.239475268545649e-05, + "loss": 0.1282, + "step": 35828 + }, + { + "epoch": 0.8398881836881802, + "grad_norm": 0.5744632482528687, + "learning_rate": 1.239120084554518e-05, + "loss": 0.0919, + "step": 35829 + }, + { + "epoch": 0.8399116252629851, + "grad_norm": 0.6179132461547852, + "learning_rate": 1.2387649481001707e-05, + "loss": 0.087, + "step": 35830 + }, + { + "epoch": 0.8399350668377902, + "grad_norm": 0.3500189185142517, + "learning_rate": 1.2384098591845305e-05, + "loss": 0.0191, + "step": 35831 + }, + { + "epoch": 0.8399585084125951, + "grad_norm": 0.4001926779747009, + "learning_rate": 1.2380548178095253e-05, + "loss": 0.4814, + "step": 35832 + }, + { + "epoch": 0.8399819499874002, + "grad_norm": 0.19890041649341583, + "learning_rate": 1.2376998239770787e-05, + "loss": 0.0155, + "step": 35833 + }, + { + "epoch": 0.8400053915622051, + "grad_norm": 0.5657269954681396, + "learning_rate": 1.2373448776891205e-05, + "loss": 0.0837, + "step": 35834 + }, + { + "epoch": 0.8400288331370102, + "grad_norm": 0.18896766006946564, + "learning_rate": 1.236989978947577e-05, + "loss": 0.0161, + "step": 35835 + }, + { + "epoch": 0.8400522747118151, + "grad_norm": 0.0857856348156929, + "learning_rate": 1.2366351277543686e-05, + "loss": 0.0185, + "step": 35836 + }, + { + "epoch": 0.8400757162866201, + "grad_norm": 0.7035707831382751, + "learning_rate": 1.2362803241114274e-05, + "loss": 0.7829, + "step": 35837 + }, + { + "epoch": 0.8400991578614251, + "grad_norm": 0.36240243911743164, + "learning_rate": 1.235925568020675e-05, + "loss": 0.0616, + "step": 35838 + }, + { + "epoch": 0.8401225994362301, + "grad_norm": 0.21824532747268677, + "learning_rate": 1.2355708594840376e-05, + "loss": 0.0257, + "step": 35839 + }, + { + "epoch": 0.8401460410110351, + "grad_norm": 0.2814922630786896, + "learning_rate": 1.2352161985034382e-05, + "loss": 0.0462, + "step": 35840 + }, + { + "epoch": 0.8401694825858401, + "grad_norm": 0.21992266178131104, + "learning_rate": 1.234861585080801e-05, + "loss": 0.0133, + "step": 35841 + }, + { + "epoch": 0.8401929241606451, + "grad_norm": 0.19224591553211212, + "learning_rate": 1.234507019218053e-05, + "loss": 0.0237, + "step": 35842 + }, + { + "epoch": 0.8402163657354501, + "grad_norm": 0.5515382289886475, + "learning_rate": 1.2341525009171162e-05, + "loss": 0.0926, + "step": 35843 + }, + { + "epoch": 0.840239807310255, + "grad_norm": 0.21167100965976715, + "learning_rate": 1.2337980301799145e-05, + "loss": 0.0364, + "step": 35844 + }, + { + "epoch": 0.8402632488850601, + "grad_norm": 0.4887891113758087, + "learning_rate": 1.23344360700837e-05, + "loss": 0.0288, + "step": 35845 + }, + { + "epoch": 0.8402866904598651, + "grad_norm": 0.18216572701931, + "learning_rate": 1.2330892314044085e-05, + "loss": 0.0327, + "step": 35846 + }, + { + "epoch": 0.8403101320346701, + "grad_norm": 0.42411357164382935, + "learning_rate": 1.232734903369951e-05, + "loss": 0.0987, + "step": 35847 + }, + { + "epoch": 0.8403335736094751, + "grad_norm": 0.13545982539653778, + "learning_rate": 1.232380622906918e-05, + "loss": 0.0244, + "step": 35848 + }, + { + "epoch": 0.8403570151842801, + "grad_norm": 0.41854143142700195, + "learning_rate": 1.2320263900172369e-05, + "loss": 0.0663, + "step": 35849 + }, + { + "epoch": 0.8403804567590851, + "grad_norm": 0.3125380277633667, + "learning_rate": 1.2316722047028272e-05, + "loss": 0.0484, + "step": 35850 + }, + { + "epoch": 0.8404038983338901, + "grad_norm": 0.3862565755844116, + "learning_rate": 1.23131806696561e-05, + "loss": 0.0681, + "step": 35851 + }, + { + "epoch": 0.8404273399086951, + "grad_norm": 0.2632560431957245, + "learning_rate": 1.2309639768075087e-05, + "loss": 0.063, + "step": 35852 + }, + { + "epoch": 0.8404507814835, + "grad_norm": 0.169081449508667, + "learning_rate": 1.2306099342304423e-05, + "loss": 0.03, + "step": 35853 + }, + { + "epoch": 0.8404742230583051, + "grad_norm": 0.8941329717636108, + "learning_rate": 1.2302559392363311e-05, + "loss": 0.1942, + "step": 35854 + }, + { + "epoch": 0.84049766463311, + "grad_norm": 0.5045083165168762, + "learning_rate": 1.2299019918271004e-05, + "loss": 0.0508, + "step": 35855 + }, + { + "epoch": 0.8405211062079151, + "grad_norm": 1.4847018718719482, + "learning_rate": 1.2295480920046675e-05, + "loss": 0.123, + "step": 35856 + }, + { + "epoch": 0.84054454778272, + "grad_norm": 0.2642614245414734, + "learning_rate": 1.229194239770951e-05, + "loss": 0.0556, + "step": 35857 + }, + { + "epoch": 0.8405679893575251, + "grad_norm": 0.6047650575637817, + "learning_rate": 1.228840435127876e-05, + "loss": 0.0414, + "step": 35858 + }, + { + "epoch": 0.84059143093233, + "grad_norm": 0.536892294883728, + "learning_rate": 1.2284866780773597e-05, + "loss": 0.0521, + "step": 35859 + }, + { + "epoch": 0.8406148725071351, + "grad_norm": 0.20035308599472046, + "learning_rate": 1.2281329686213205e-05, + "loss": 0.0348, + "step": 35860 + }, + { + "epoch": 0.84063831408194, + "grad_norm": 0.5492673516273499, + "learning_rate": 1.227779306761676e-05, + "loss": 0.1186, + "step": 35861 + }, + { + "epoch": 0.8406617556567451, + "grad_norm": 0.3694417476654053, + "learning_rate": 1.2274256925003502e-05, + "loss": 0.0483, + "step": 35862 + }, + { + "epoch": 0.84068519723155, + "grad_norm": 0.32346245646476746, + "learning_rate": 1.2270721258392603e-05, + "loss": 0.0605, + "step": 35863 + }, + { + "epoch": 0.840708638806355, + "grad_norm": 0.6811867952346802, + "learning_rate": 1.2267186067803227e-05, + "loss": 0.1234, + "step": 35864 + }, + { + "epoch": 0.84073208038116, + "grad_norm": 0.2920735478401184, + "learning_rate": 1.2263651353254569e-05, + "loss": 0.0539, + "step": 35865 + }, + { + "epoch": 0.840755521955965, + "grad_norm": 0.11088359355926514, + "learning_rate": 1.2260117114765801e-05, + "loss": 0.0134, + "step": 35866 + }, + { + "epoch": 0.84077896353077, + "grad_norm": 0.23147232830524445, + "learning_rate": 1.2256583352356088e-05, + "loss": 0.0193, + "step": 35867 + }, + { + "epoch": 0.840802405105575, + "grad_norm": 0.3545176088809967, + "learning_rate": 1.2253050066044636e-05, + "loss": 0.0361, + "step": 35868 + }, + { + "epoch": 0.84082584668038, + "grad_norm": 0.27724793553352356, + "learning_rate": 1.2249517255850584e-05, + "loss": 0.0129, + "step": 35869 + }, + { + "epoch": 0.840849288255185, + "grad_norm": 0.5520254969596863, + "learning_rate": 1.2245984921793141e-05, + "loss": 0.1736, + "step": 35870 + }, + { + "epoch": 0.84087272982999, + "grad_norm": 0.45519036054611206, + "learning_rate": 1.2242453063891446e-05, + "loss": 0.0944, + "step": 35871 + }, + { + "epoch": 0.840896171404795, + "grad_norm": 0.48561376333236694, + "learning_rate": 1.2238921682164684e-05, + "loss": 0.0753, + "step": 35872 + }, + { + "epoch": 0.8409196129795999, + "grad_norm": 0.4815542697906494, + "learning_rate": 1.2235390776631982e-05, + "loss": 0.061, + "step": 35873 + }, + { + "epoch": 0.840943054554405, + "grad_norm": 0.3096488416194916, + "learning_rate": 1.2231860347312507e-05, + "loss": 0.0533, + "step": 35874 + }, + { + "epoch": 0.8409664961292099, + "grad_norm": 0.11262445151805878, + "learning_rate": 1.222833039422544e-05, + "loss": 0.0183, + "step": 35875 + }, + { + "epoch": 0.840989937704015, + "grad_norm": 0.1315385401248932, + "learning_rate": 1.2224800917389923e-05, + "loss": 0.0161, + "step": 35876 + }, + { + "epoch": 0.8410133792788199, + "grad_norm": 0.14325056970119476, + "learning_rate": 1.2221271916825106e-05, + "loss": 0.0258, + "step": 35877 + }, + { + "epoch": 0.841036820853625, + "grad_norm": 0.6533817052841187, + "learning_rate": 1.2217743392550141e-05, + "loss": 0.5038, + "step": 35878 + }, + { + "epoch": 0.8410602624284299, + "grad_norm": 0.07099981606006622, + "learning_rate": 1.2214215344584135e-05, + "loss": 0.0099, + "step": 35879 + }, + { + "epoch": 0.841083704003235, + "grad_norm": 0.4685891270637512, + "learning_rate": 1.2210687772946295e-05, + "loss": 0.1283, + "step": 35880 + }, + { + "epoch": 0.8411071455780399, + "grad_norm": 0.14639392495155334, + "learning_rate": 1.2207160677655727e-05, + "loss": 0.0109, + "step": 35881 + }, + { + "epoch": 0.8411305871528449, + "grad_norm": 0.5671653151512146, + "learning_rate": 1.220363405873155e-05, + "loss": 0.0495, + "step": 35882 + }, + { + "epoch": 0.8411540287276499, + "grad_norm": 0.3765909969806671, + "learning_rate": 1.2200107916192948e-05, + "loss": 0.4588, + "step": 35883 + }, + { + "epoch": 0.8411774703024549, + "grad_norm": 0.34894174337387085, + "learning_rate": 1.219658225005903e-05, + "loss": 0.0911, + "step": 35884 + }, + { + "epoch": 0.8412009118772599, + "grad_norm": 0.4205215573310852, + "learning_rate": 1.2193057060348933e-05, + "loss": 0.0535, + "step": 35885 + }, + { + "epoch": 0.8412243534520649, + "grad_norm": 0.3665502965450287, + "learning_rate": 1.2189532347081745e-05, + "loss": 0.3634, + "step": 35886 + }, + { + "epoch": 0.8412477950268699, + "grad_norm": 0.27748745679855347, + "learning_rate": 1.2186008110276648e-05, + "loss": 0.0233, + "step": 35887 + }, + { + "epoch": 0.8412712366016749, + "grad_norm": 0.1808880716562271, + "learning_rate": 1.2182484349952738e-05, + "loss": 0.0297, + "step": 35888 + }, + { + "epoch": 0.8412946781764798, + "grad_norm": 0.1502416729927063, + "learning_rate": 1.2178961066129136e-05, + "loss": 0.0137, + "step": 35889 + }, + { + "epoch": 0.8413181197512849, + "grad_norm": 0.3229614198207855, + "learning_rate": 1.2175438258824967e-05, + "loss": 0.0301, + "step": 35890 + }, + { + "epoch": 0.8413415613260898, + "grad_norm": 0.09871308505535126, + "learning_rate": 1.217191592805934e-05, + "loss": 0.0184, + "step": 35891 + }, + { + "epoch": 0.8413650029008949, + "grad_norm": 0.3472765386104584, + "learning_rate": 1.2168394073851341e-05, + "loss": 0.0639, + "step": 35892 + }, + { + "epoch": 0.8413884444756998, + "grad_norm": 0.5166349411010742, + "learning_rate": 1.216487269622012e-05, + "loss": 0.1136, + "step": 35893 + }, + { + "epoch": 0.8414118860505049, + "grad_norm": 0.3467228412628174, + "learning_rate": 1.216135179518476e-05, + "loss": 0.0214, + "step": 35894 + }, + { + "epoch": 0.8414353276253098, + "grad_norm": 0.3870733976364136, + "learning_rate": 1.2157831370764394e-05, + "loss": 0.6519, + "step": 35895 + }, + { + "epoch": 0.8414587692001149, + "grad_norm": 0.3846779763698578, + "learning_rate": 1.2154311422978104e-05, + "loss": 0.0511, + "step": 35896 + }, + { + "epoch": 0.8414822107749199, + "grad_norm": 0.19571958482265472, + "learning_rate": 1.2150791951844986e-05, + "loss": 0.034, + "step": 35897 + }, + { + "epoch": 0.8415056523497249, + "grad_norm": 0.2101467400789261, + "learning_rate": 1.2147272957384149e-05, + "loss": 0.034, + "step": 35898 + }, + { + "epoch": 0.8415290939245299, + "grad_norm": 0.09796684235334396, + "learning_rate": 1.2143754439614651e-05, + "loss": 0.0156, + "step": 35899 + }, + { + "epoch": 0.8415525354993348, + "grad_norm": 0.2920183837413788, + "learning_rate": 1.2140236398555627e-05, + "loss": 0.0599, + "step": 35900 + }, + { + "epoch": 0.8415759770741399, + "grad_norm": 0.9499390721321106, + "learning_rate": 1.2136718834226157e-05, + "loss": 0.1647, + "step": 35901 + }, + { + "epoch": 0.8415994186489448, + "grad_norm": 0.3750404417514801, + "learning_rate": 1.2133201746645328e-05, + "loss": 0.1066, + "step": 35902 + }, + { + "epoch": 0.8416228602237499, + "grad_norm": 0.5187917351722717, + "learning_rate": 1.2129685135832213e-05, + "loss": 0.1077, + "step": 35903 + }, + { + "epoch": 0.8416463017985548, + "grad_norm": 0.49152520298957825, + "learning_rate": 1.2126169001805899e-05, + "loss": 0.1163, + "step": 35904 + }, + { + "epoch": 0.8416697433733599, + "grad_norm": 0.49474674463272095, + "learning_rate": 1.2122653344585432e-05, + "loss": 0.6166, + "step": 35905 + }, + { + "epoch": 0.8416931849481648, + "grad_norm": 0.4860999584197998, + "learning_rate": 1.2119138164189925e-05, + "loss": 0.0749, + "step": 35906 + }, + { + "epoch": 0.8417166265229699, + "grad_norm": 0.5161135792732239, + "learning_rate": 1.211562346063847e-05, + "loss": 0.0746, + "step": 35907 + }, + { + "epoch": 0.8417400680977748, + "grad_norm": 0.3717523217201233, + "learning_rate": 1.2112109233950109e-05, + "loss": 0.045, + "step": 35908 + }, + { + "epoch": 0.8417635096725798, + "grad_norm": 0.12638886272907257, + "learning_rate": 1.2108595484143915e-05, + "loss": 0.0226, + "step": 35909 + }, + { + "epoch": 0.8417869512473848, + "grad_norm": 0.1658090502023697, + "learning_rate": 1.210508221123895e-05, + "loss": 0.0286, + "step": 35910 + }, + { + "epoch": 0.8418103928221898, + "grad_norm": 0.30496910214424133, + "learning_rate": 1.2101569415254289e-05, + "loss": 0.0228, + "step": 35911 + }, + { + "epoch": 0.8418338343969948, + "grad_norm": 0.4759009778499603, + "learning_rate": 1.2098057096208947e-05, + "loss": 0.0727, + "step": 35912 + }, + { + "epoch": 0.8418572759717998, + "grad_norm": 0.47533273696899414, + "learning_rate": 1.2094545254122058e-05, + "loss": 0.0766, + "step": 35913 + }, + { + "epoch": 0.8418807175466048, + "grad_norm": 0.6096349954605103, + "learning_rate": 1.2091033889012625e-05, + "loss": 0.6828, + "step": 35914 + }, + { + "epoch": 0.8419041591214098, + "grad_norm": 0.1624639332294464, + "learning_rate": 1.2087523000899725e-05, + "loss": 0.0269, + "step": 35915 + }, + { + "epoch": 0.8419276006962148, + "grad_norm": 0.43946799635887146, + "learning_rate": 1.2084012589802385e-05, + "loss": 0.0496, + "step": 35916 + }, + { + "epoch": 0.8419510422710198, + "grad_norm": 0.24304459989070892, + "learning_rate": 1.2080502655739656e-05, + "loss": 0.1933, + "step": 35917 + }, + { + "epoch": 0.8419744838458247, + "grad_norm": 0.23443353176116943, + "learning_rate": 1.2076993198730601e-05, + "loss": 0.0424, + "step": 35918 + }, + { + "epoch": 0.8419979254206298, + "grad_norm": 0.3750724494457245, + "learning_rate": 1.207348421879424e-05, + "loss": 0.0372, + "step": 35919 + }, + { + "epoch": 0.8420213669954347, + "grad_norm": 0.4276534616947174, + "learning_rate": 1.2069975715949645e-05, + "loss": 0.0897, + "step": 35920 + }, + { + "epoch": 0.8420448085702398, + "grad_norm": 0.2645460069179535, + "learning_rate": 1.2066467690215832e-05, + "loss": 0.0306, + "step": 35921 + }, + { + "epoch": 0.8420682501450447, + "grad_norm": 0.4643079340457916, + "learning_rate": 1.2062960141611845e-05, + "loss": 0.0606, + "step": 35922 + }, + { + "epoch": 0.8420916917198498, + "grad_norm": 0.45951953530311584, + "learning_rate": 1.2059453070156702e-05, + "loss": 0.0464, + "step": 35923 + }, + { + "epoch": 0.8421151332946547, + "grad_norm": 0.2150055468082428, + "learning_rate": 1.2055946475869417e-05, + "loss": 0.049, + "step": 35924 + }, + { + "epoch": 0.8421385748694598, + "grad_norm": 0.4471692442893982, + "learning_rate": 1.2052440358769058e-05, + "loss": 0.0643, + "step": 35925 + }, + { + "epoch": 0.8421620164442647, + "grad_norm": 0.23758994042873383, + "learning_rate": 1.2048934718874638e-05, + "loss": 0.0436, + "step": 35926 + }, + { + "epoch": 0.8421854580190697, + "grad_norm": 0.30584821105003357, + "learning_rate": 1.2045429556205168e-05, + "loss": 0.0642, + "step": 35927 + }, + { + "epoch": 0.8422088995938747, + "grad_norm": 0.3389282822608948, + "learning_rate": 1.2041924870779675e-05, + "loss": 0.0848, + "step": 35928 + }, + { + "epoch": 0.8422323411686797, + "grad_norm": 0.1936274766921997, + "learning_rate": 1.2038420662617167e-05, + "loss": 0.0158, + "step": 35929 + }, + { + "epoch": 0.8422557827434847, + "grad_norm": 0.2991924285888672, + "learning_rate": 1.203491693173664e-05, + "loss": 0.0606, + "step": 35930 + }, + { + "epoch": 0.8422792243182897, + "grad_norm": 0.08485500514507294, + "learning_rate": 1.2031413678157122e-05, + "loss": 0.0132, + "step": 35931 + }, + { + "epoch": 0.8423026658930947, + "grad_norm": 0.17909133434295654, + "learning_rate": 1.2027910901897654e-05, + "loss": 0.0334, + "step": 35932 + }, + { + "epoch": 0.8423261074678997, + "grad_norm": 0.24576279520988464, + "learning_rate": 1.2024408602977222e-05, + "loss": 0.0204, + "step": 35933 + }, + { + "epoch": 0.8423495490427046, + "grad_norm": 0.3905908167362213, + "learning_rate": 1.202090678141481e-05, + "loss": 0.0961, + "step": 35934 + }, + { + "epoch": 0.8423729906175097, + "grad_norm": 0.5751295685768127, + "learning_rate": 1.2017405437229434e-05, + "loss": 0.103, + "step": 35935 + }, + { + "epoch": 0.8423964321923146, + "grad_norm": 0.3561325967311859, + "learning_rate": 1.2013904570440093e-05, + "loss": 0.0283, + "step": 35936 + }, + { + "epoch": 0.8424198737671197, + "grad_norm": 0.25517353415489197, + "learning_rate": 1.201040418106576e-05, + "loss": 0.0318, + "step": 35937 + }, + { + "epoch": 0.8424433153419246, + "grad_norm": 0.6005575060844421, + "learning_rate": 1.2006904269125464e-05, + "loss": 0.7103, + "step": 35938 + }, + { + "epoch": 0.8424667569167297, + "grad_norm": 0.21673361957073212, + "learning_rate": 1.2003404834638187e-05, + "loss": 0.0359, + "step": 35939 + }, + { + "epoch": 0.8424901984915346, + "grad_norm": 0.49815142154693604, + "learning_rate": 1.1999905877622907e-05, + "loss": 0.4805, + "step": 35940 + }, + { + "epoch": 0.8425136400663397, + "grad_norm": 0.16908946633338928, + "learning_rate": 1.1996407398098607e-05, + "loss": 0.0398, + "step": 35941 + }, + { + "epoch": 0.8425370816411446, + "grad_norm": 0.4657023847103119, + "learning_rate": 1.1992909396084263e-05, + "loss": 0.6276, + "step": 35942 + }, + { + "epoch": 0.8425605232159497, + "grad_norm": 0.4396754205226898, + "learning_rate": 1.1989411871598877e-05, + "loss": 0.085, + "step": 35943 + }, + { + "epoch": 0.8425839647907546, + "grad_norm": 0.5718942284584045, + "learning_rate": 1.1985914824661404e-05, + "loss": 0.0765, + "step": 35944 + }, + { + "epoch": 0.8426074063655596, + "grad_norm": 0.08611026406288147, + "learning_rate": 1.198241825529085e-05, + "loss": 0.0101, + "step": 35945 + }, + { + "epoch": 0.8426308479403646, + "grad_norm": 0.3298960328102112, + "learning_rate": 1.1978922163506168e-05, + "loss": 0.0553, + "step": 35946 + }, + { + "epoch": 0.8426542895151696, + "grad_norm": 0.18905360996723175, + "learning_rate": 1.197542654932633e-05, + "loss": 0.0243, + "step": 35947 + }, + { + "epoch": 0.8426777310899747, + "grad_norm": 0.253715842962265, + "learning_rate": 1.1971931412770298e-05, + "loss": 0.0588, + "step": 35948 + }, + { + "epoch": 0.8427011726647796, + "grad_norm": 0.6439100503921509, + "learning_rate": 1.1968436753857049e-05, + "loss": 0.0724, + "step": 35949 + }, + { + "epoch": 0.8427246142395847, + "grad_norm": 0.6041244864463806, + "learning_rate": 1.196494257260552e-05, + "loss": 0.5695, + "step": 35950 + }, + { + "epoch": 0.8427480558143896, + "grad_norm": 0.07862502336502075, + "learning_rate": 1.1961448869034698e-05, + "loss": 0.0147, + "step": 35951 + }, + { + "epoch": 0.8427714973891947, + "grad_norm": 0.27601882815361023, + "learning_rate": 1.1957955643163531e-05, + "loss": 0.0669, + "step": 35952 + }, + { + "epoch": 0.8427949389639996, + "grad_norm": 0.1138949766755104, + "learning_rate": 1.1954462895010975e-05, + "loss": 0.018, + "step": 35953 + }, + { + "epoch": 0.8428183805388046, + "grad_norm": 0.1228618249297142, + "learning_rate": 1.195097062459598e-05, + "loss": 0.0178, + "step": 35954 + }, + { + "epoch": 0.8428418221136096, + "grad_norm": 0.3604393005371094, + "learning_rate": 1.1947478831937475e-05, + "loss": 0.0748, + "step": 35955 + }, + { + "epoch": 0.8428652636884146, + "grad_norm": 0.29749563336372375, + "learning_rate": 1.1943987517054445e-05, + "loss": 0.0455, + "step": 35956 + }, + { + "epoch": 0.8428887052632196, + "grad_norm": 0.6012477278709412, + "learning_rate": 1.1940496679965796e-05, + "loss": 0.0248, + "step": 35957 + }, + { + "epoch": 0.8429121468380246, + "grad_norm": 0.7249293327331543, + "learning_rate": 1.1937006320690503e-05, + "loss": 0.0953, + "step": 35958 + }, + { + "epoch": 0.8429355884128296, + "grad_norm": 0.5244027972221375, + "learning_rate": 1.1933516439247494e-05, + "loss": 0.0657, + "step": 35959 + }, + { + "epoch": 0.8429590299876346, + "grad_norm": 0.3045022487640381, + "learning_rate": 1.19300270356557e-05, + "loss": 0.0523, + "step": 35960 + }, + { + "epoch": 0.8429824715624396, + "grad_norm": 0.44135645031929016, + "learning_rate": 1.192653810993406e-05, + "loss": 0.0987, + "step": 35961 + }, + { + "epoch": 0.8430059131372446, + "grad_norm": 0.3241385519504547, + "learning_rate": 1.192304966210147e-05, + "loss": 0.0476, + "step": 35962 + }, + { + "epoch": 0.8430293547120495, + "grad_norm": 0.1877770870923996, + "learning_rate": 1.1919561692176917e-05, + "loss": 0.0242, + "step": 35963 + }, + { + "epoch": 0.8430527962868546, + "grad_norm": 0.6323461532592773, + "learning_rate": 1.1916074200179306e-05, + "loss": 0.0485, + "step": 35964 + }, + { + "epoch": 0.8430762378616595, + "grad_norm": 0.6648895740509033, + "learning_rate": 1.1912587186127545e-05, + "loss": 0.1235, + "step": 35965 + }, + { + "epoch": 0.8430996794364646, + "grad_norm": 0.47178879380226135, + "learning_rate": 1.190910065004056e-05, + "loss": 0.0571, + "step": 35966 + }, + { + "epoch": 0.8431231210112695, + "grad_norm": 0.4170088469982147, + "learning_rate": 1.1905614591937252e-05, + "loss": 0.0905, + "step": 35967 + }, + { + "epoch": 0.8431465625860746, + "grad_norm": 0.3576771914958954, + "learning_rate": 1.190212901183657e-05, + "loss": 0.047, + "step": 35968 + }, + { + "epoch": 0.8431700041608795, + "grad_norm": 0.6047565340995789, + "learning_rate": 1.1898643909757412e-05, + "loss": 0.5688, + "step": 35969 + }, + { + "epoch": 0.8431934457356846, + "grad_norm": 0.3932250738143921, + "learning_rate": 1.1895159285718693e-05, + "loss": 0.0691, + "step": 35970 + }, + { + "epoch": 0.8432168873104895, + "grad_norm": 0.10065370053052902, + "learning_rate": 1.1891675139739321e-05, + "loss": 0.0185, + "step": 35971 + }, + { + "epoch": 0.8432403288852945, + "grad_norm": 0.13453033566474915, + "learning_rate": 1.1888191471838205e-05, + "loss": 0.0153, + "step": 35972 + }, + { + "epoch": 0.8432637704600995, + "grad_norm": 0.2478523999452591, + "learning_rate": 1.188470828203423e-05, + "loss": 0.0328, + "step": 35973 + }, + { + "epoch": 0.8432872120349045, + "grad_norm": 0.6490843296051025, + "learning_rate": 1.188122557034631e-05, + "loss": 0.1062, + "step": 35974 + }, + { + "epoch": 0.8433106536097095, + "grad_norm": 0.531507670879364, + "learning_rate": 1.1877743336793313e-05, + "loss": 0.1448, + "step": 35975 + }, + { + "epoch": 0.8433340951845145, + "grad_norm": 0.35487377643585205, + "learning_rate": 1.1874261581394187e-05, + "loss": 0.6944, + "step": 35976 + }, + { + "epoch": 0.8433575367593195, + "grad_norm": 0.19890980422496796, + "learning_rate": 1.1870780304167783e-05, + "loss": 0.012, + "step": 35977 + }, + { + "epoch": 0.8433809783341245, + "grad_norm": 0.37461212277412415, + "learning_rate": 1.1867299505133012e-05, + "loss": 0.0627, + "step": 35978 + }, + { + "epoch": 0.8434044199089294, + "grad_norm": 0.3310256898403168, + "learning_rate": 1.1863819184308733e-05, + "loss": 0.0381, + "step": 35979 + }, + { + "epoch": 0.8434278614837345, + "grad_norm": 0.09761874377727509, + "learning_rate": 1.1860339341713866e-05, + "loss": 0.0096, + "step": 35980 + }, + { + "epoch": 0.8434513030585394, + "grad_norm": 0.6777492165565491, + "learning_rate": 1.1856859977367273e-05, + "loss": 0.1138, + "step": 35981 + }, + { + "epoch": 0.8434747446333445, + "grad_norm": 0.5463589429855347, + "learning_rate": 1.185338109128782e-05, + "loss": 0.0444, + "step": 35982 + }, + { + "epoch": 0.8434981862081494, + "grad_norm": 0.3571357727050781, + "learning_rate": 1.1849902683494418e-05, + "loss": 0.0513, + "step": 35983 + }, + { + "epoch": 0.8435216277829545, + "grad_norm": 0.20873941481113434, + "learning_rate": 1.1846424754005925e-05, + "loss": 0.0169, + "step": 35984 + }, + { + "epoch": 0.8435450693577594, + "grad_norm": 0.320333868265152, + "learning_rate": 1.1842947302841213e-05, + "loss": 0.0482, + "step": 35985 + }, + { + "epoch": 0.8435685109325645, + "grad_norm": 0.2462896704673767, + "learning_rate": 1.1839470330019143e-05, + "loss": 0.0475, + "step": 35986 + }, + { + "epoch": 0.8435919525073694, + "grad_norm": 0.6023612022399902, + "learning_rate": 1.1835993835558589e-05, + "loss": 0.1216, + "step": 35987 + }, + { + "epoch": 0.8436153940821745, + "grad_norm": 0.6249653100967407, + "learning_rate": 1.1832517819478383e-05, + "loss": 0.5509, + "step": 35988 + }, + { + "epoch": 0.8436388356569794, + "grad_norm": 0.18936343491077423, + "learning_rate": 1.1829042281797443e-05, + "loss": 0.0308, + "step": 35989 + }, + { + "epoch": 0.8436622772317844, + "grad_norm": 0.1505691409111023, + "learning_rate": 1.1825567222534595e-05, + "loss": 0.0106, + "step": 35990 + }, + { + "epoch": 0.8436857188065894, + "grad_norm": 0.3857853412628174, + "learning_rate": 1.1822092641708683e-05, + "loss": 0.0766, + "step": 35991 + }, + { + "epoch": 0.8437091603813944, + "grad_norm": 0.45001232624053955, + "learning_rate": 1.1818618539338588e-05, + "loss": 0.07, + "step": 35992 + }, + { + "epoch": 0.8437326019561994, + "grad_norm": 0.3511058986186981, + "learning_rate": 1.1815144915443155e-05, + "loss": 0.0785, + "step": 35993 + }, + { + "epoch": 0.8437560435310044, + "grad_norm": 0.3737092614173889, + "learning_rate": 1.18116717700412e-05, + "loss": 0.0577, + "step": 35994 + }, + { + "epoch": 0.8437794851058094, + "grad_norm": 0.4837082326412201, + "learning_rate": 1.1808199103151619e-05, + "loss": 0.0299, + "step": 35995 + }, + { + "epoch": 0.8438029266806144, + "grad_norm": 0.35079503059387207, + "learning_rate": 1.1804726914793218e-05, + "loss": 0.0607, + "step": 35996 + }, + { + "epoch": 0.8438263682554193, + "grad_norm": 0.3983781337738037, + "learning_rate": 1.1801255204984862e-05, + "loss": 0.0612, + "step": 35997 + }, + { + "epoch": 0.8438498098302244, + "grad_norm": 0.3774247169494629, + "learning_rate": 1.1797783973745369e-05, + "loss": 0.0779, + "step": 35998 + }, + { + "epoch": 0.8438732514050294, + "grad_norm": 0.2524937093257904, + "learning_rate": 1.1794313221093579e-05, + "loss": 0.0374, + "step": 35999 + }, + { + "epoch": 0.8438966929798344, + "grad_norm": 0.5901714563369751, + "learning_rate": 1.1790842947048308e-05, + "loss": 0.0996, + "step": 36000 + }, + { + "epoch": 0.8439201345546394, + "grad_norm": 0.20460040867328644, + "learning_rate": 1.1787373151628423e-05, + "loss": 0.0482, + "step": 36001 + }, + { + "epoch": 0.8439435761294444, + "grad_norm": 0.23604586720466614, + "learning_rate": 1.1783903834852738e-05, + "loss": 0.0264, + "step": 36002 + }, + { + "epoch": 0.8439670177042494, + "grad_norm": 0.5401679277420044, + "learning_rate": 1.1780434996740042e-05, + "loss": 0.0551, + "step": 36003 + }, + { + "epoch": 0.8439904592790544, + "grad_norm": 0.365152508020401, + "learning_rate": 1.1776966637309205e-05, + "loss": 0.0491, + "step": 36004 + }, + { + "epoch": 0.8440139008538594, + "grad_norm": 0.45685189962387085, + "learning_rate": 1.1773498756579037e-05, + "loss": 0.0511, + "step": 36005 + }, + { + "epoch": 0.8440373424286644, + "grad_norm": 0.3939683735370636, + "learning_rate": 1.1770031354568345e-05, + "loss": 0.1099, + "step": 36006 + }, + { + "epoch": 0.8440607840034694, + "grad_norm": 0.1569557636976242, + "learning_rate": 1.1766564431295912e-05, + "loss": 0.0277, + "step": 36007 + }, + { + "epoch": 0.8440842255782743, + "grad_norm": 0.2981017529964447, + "learning_rate": 1.1763097986780613e-05, + "loss": 0.0552, + "step": 36008 + }, + { + "epoch": 0.8441076671530794, + "grad_norm": 0.365875780582428, + "learning_rate": 1.1759632021041223e-05, + "loss": 0.0468, + "step": 36009 + }, + { + "epoch": 0.8441311087278843, + "grad_norm": 0.09670310467481613, + "learning_rate": 1.1756166534096547e-05, + "loss": 0.0182, + "step": 36010 + }, + { + "epoch": 0.8441545503026894, + "grad_norm": 0.3243992328643799, + "learning_rate": 1.1752701525965393e-05, + "loss": 0.0621, + "step": 36011 + }, + { + "epoch": 0.8441779918774943, + "grad_norm": 0.34384194016456604, + "learning_rate": 1.1749236996666568e-05, + "loss": 0.3725, + "step": 36012 + }, + { + "epoch": 0.8442014334522994, + "grad_norm": 0.4457535147666931, + "learning_rate": 1.1745772946218835e-05, + "loss": 0.072, + "step": 36013 + }, + { + "epoch": 0.8442248750271043, + "grad_norm": 0.39126160740852356, + "learning_rate": 1.1742309374641047e-05, + "loss": 0.0474, + "step": 36014 + }, + { + "epoch": 0.8442483166019094, + "grad_norm": 0.4799928069114685, + "learning_rate": 1.1738846281951943e-05, + "loss": 0.092, + "step": 36015 + }, + { + "epoch": 0.8442717581767143, + "grad_norm": 0.4033815562725067, + "learning_rate": 1.1735383668170363e-05, + "loss": 0.0703, + "step": 36016 + }, + { + "epoch": 0.8442951997515193, + "grad_norm": 0.681081235408783, + "learning_rate": 1.173192153331507e-05, + "loss": 0.6143, + "step": 36017 + }, + { + "epoch": 0.8443186413263243, + "grad_norm": 0.4423580467700958, + "learning_rate": 1.1728459877404863e-05, + "loss": 0.0593, + "step": 36018 + }, + { + "epoch": 0.8443420829011293, + "grad_norm": 0.4738890826702118, + "learning_rate": 1.1724998700458512e-05, + "loss": 0.0738, + "step": 36019 + }, + { + "epoch": 0.8443655244759343, + "grad_norm": 0.5818467140197754, + "learning_rate": 1.172153800249477e-05, + "loss": 0.638, + "step": 36020 + }, + { + "epoch": 0.8443889660507393, + "grad_norm": 0.33532842993736267, + "learning_rate": 1.1718077783532477e-05, + "loss": 0.0899, + "step": 36021 + }, + { + "epoch": 0.8444124076255443, + "grad_norm": 0.48568710684776306, + "learning_rate": 1.1714618043590376e-05, + "loss": 0.0722, + "step": 36022 + }, + { + "epoch": 0.8444358492003493, + "grad_norm": 0.4688463509082794, + "learning_rate": 1.1711158782687226e-05, + "loss": 0.0398, + "step": 36023 + }, + { + "epoch": 0.8444592907751542, + "grad_norm": 0.0870361402630806, + "learning_rate": 1.1707700000841825e-05, + "loss": 0.0155, + "step": 36024 + }, + { + "epoch": 0.8444827323499593, + "grad_norm": 0.4965610206127167, + "learning_rate": 1.1704241698072916e-05, + "loss": 0.0735, + "step": 36025 + }, + { + "epoch": 0.8445061739247642, + "grad_norm": 0.6282817721366882, + "learning_rate": 1.1700783874399269e-05, + "loss": 0.0879, + "step": 36026 + }, + { + "epoch": 0.8445296154995693, + "grad_norm": 0.2176327109336853, + "learning_rate": 1.1697326529839658e-05, + "loss": 0.0343, + "step": 36027 + }, + { + "epoch": 0.8445530570743742, + "grad_norm": 0.2636815309524536, + "learning_rate": 1.1693869664412815e-05, + "loss": 0.0375, + "step": 36028 + }, + { + "epoch": 0.8445764986491793, + "grad_norm": 0.07917768508195877, + "learning_rate": 1.1690413278137547e-05, + "loss": 0.0046, + "step": 36029 + }, + { + "epoch": 0.8445999402239842, + "grad_norm": 0.11811742186546326, + "learning_rate": 1.168695737103258e-05, + "loss": 0.0215, + "step": 36030 + }, + { + "epoch": 0.8446233817987893, + "grad_norm": 0.062218036502599716, + "learning_rate": 1.1683501943116658e-05, + "loss": 0.0053, + "step": 36031 + }, + { + "epoch": 0.8446468233735942, + "grad_norm": 0.45157986879348755, + "learning_rate": 1.1680046994408522e-05, + "loss": 0.7609, + "step": 36032 + }, + { + "epoch": 0.8446702649483993, + "grad_norm": 0.15114068984985352, + "learning_rate": 1.1676592524926955e-05, + "loss": 0.0294, + "step": 36033 + }, + { + "epoch": 0.8446937065232042, + "grad_norm": 0.3641945719718933, + "learning_rate": 1.1673138534690675e-05, + "loss": 0.0645, + "step": 36034 + }, + { + "epoch": 0.8447171480980092, + "grad_norm": 0.19499075412750244, + "learning_rate": 1.1669685023718435e-05, + "loss": 0.0135, + "step": 36035 + }, + { + "epoch": 0.8447405896728142, + "grad_norm": 0.13323667645454407, + "learning_rate": 1.1666231992028964e-05, + "loss": 0.0211, + "step": 36036 + }, + { + "epoch": 0.8447640312476192, + "grad_norm": 0.38354840874671936, + "learning_rate": 1.1662779439641003e-05, + "loss": 0.0573, + "step": 36037 + }, + { + "epoch": 0.8447874728224242, + "grad_norm": 0.48415854573249817, + "learning_rate": 1.165932736657327e-05, + "loss": 0.0355, + "step": 36038 + }, + { + "epoch": 0.8448109143972292, + "grad_norm": 0.4470669627189636, + "learning_rate": 1.1655875772844516e-05, + "loss": 0.0588, + "step": 36039 + }, + { + "epoch": 0.8448343559720342, + "grad_norm": 0.4096679389476776, + "learning_rate": 1.165242465847346e-05, + "loss": 0.0422, + "step": 36040 + }, + { + "epoch": 0.8448577975468392, + "grad_norm": 0.34863606095314026, + "learning_rate": 1.1648974023478842e-05, + "loss": 0.0426, + "step": 36041 + }, + { + "epoch": 0.8448812391216441, + "grad_norm": 0.5175008773803711, + "learning_rate": 1.1645523867879382e-05, + "loss": 0.0866, + "step": 36042 + }, + { + "epoch": 0.8449046806964492, + "grad_norm": 0.13628806173801422, + "learning_rate": 1.1642074191693785e-05, + "loss": 0.018, + "step": 36043 + }, + { + "epoch": 0.8449281222712541, + "grad_norm": 0.49960628151893616, + "learning_rate": 1.1638624994940783e-05, + "loss": 0.1194, + "step": 36044 + }, + { + "epoch": 0.8449515638460592, + "grad_norm": 0.20191429555416107, + "learning_rate": 1.163517627763907e-05, + "loss": 0.0598, + "step": 36045 + }, + { + "epoch": 0.8449750054208641, + "grad_norm": 0.39411047101020813, + "learning_rate": 1.1631728039807388e-05, + "loss": 0.0604, + "step": 36046 + }, + { + "epoch": 0.8449984469956692, + "grad_norm": 0.29139214754104614, + "learning_rate": 1.1628280281464443e-05, + "loss": 0.0227, + "step": 36047 + }, + { + "epoch": 0.8450218885704741, + "grad_norm": 0.15704113245010376, + "learning_rate": 1.1624833002628932e-05, + "loss": 0.0254, + "step": 36048 + }, + { + "epoch": 0.8450453301452792, + "grad_norm": 0.13117453455924988, + "learning_rate": 1.1621386203319551e-05, + "loss": 0.0096, + "step": 36049 + }, + { + "epoch": 0.8450687717200841, + "grad_norm": 0.6024000644683838, + "learning_rate": 1.1617939883555029e-05, + "loss": 0.0858, + "step": 36050 + }, + { + "epoch": 0.8450922132948891, + "grad_norm": 0.4508453905582428, + "learning_rate": 1.1614494043354018e-05, + "loss": 0.0545, + "step": 36051 + }, + { + "epoch": 0.8451156548696942, + "grad_norm": 0.47565966844558716, + "learning_rate": 1.1611048682735248e-05, + "loss": 0.074, + "step": 36052 + }, + { + "epoch": 0.8451390964444991, + "grad_norm": 0.5745630860328674, + "learning_rate": 1.1607603801717437e-05, + "loss": 0.0864, + "step": 36053 + }, + { + "epoch": 0.8451625380193042, + "grad_norm": 0.3634313941001892, + "learning_rate": 1.160415940031926e-05, + "loss": 0.0284, + "step": 36054 + }, + { + "epoch": 0.8451859795941091, + "grad_norm": 0.3479490876197815, + "learning_rate": 1.1600715478559387e-05, + "loss": 0.0755, + "step": 36055 + }, + { + "epoch": 0.8452094211689142, + "grad_norm": 0.22524632513523102, + "learning_rate": 1.159727203645653e-05, + "loss": 0.0241, + "step": 36056 + }, + { + "epoch": 0.8452328627437191, + "grad_norm": 0.8666447401046753, + "learning_rate": 1.159382907402935e-05, + "loss": 0.0768, + "step": 36057 + }, + { + "epoch": 0.8452563043185242, + "grad_norm": 0.3300997316837311, + "learning_rate": 1.1590386591296521e-05, + "loss": 0.0372, + "step": 36058 + }, + { + "epoch": 0.8452797458933291, + "grad_norm": 0.4555940628051758, + "learning_rate": 1.1586944588276772e-05, + "loss": 0.0636, + "step": 36059 + }, + { + "epoch": 0.8453031874681342, + "grad_norm": 0.20533889532089233, + "learning_rate": 1.1583503064988732e-05, + "loss": 0.0222, + "step": 36060 + }, + { + "epoch": 0.8453266290429391, + "grad_norm": 0.44192415475845337, + "learning_rate": 1.15800620214511e-05, + "loss": 0.0605, + "step": 36061 + }, + { + "epoch": 0.8453500706177441, + "grad_norm": 0.20871485769748688, + "learning_rate": 1.1576621457682535e-05, + "loss": 0.0402, + "step": 36062 + }, + { + "epoch": 0.8453735121925491, + "grad_norm": 0.33249491453170776, + "learning_rate": 1.1573181373701703e-05, + "loss": 0.0713, + "step": 36063 + }, + { + "epoch": 0.8453969537673541, + "grad_norm": 0.20327407121658325, + "learning_rate": 1.1569741769527265e-05, + "loss": 0.0292, + "step": 36064 + }, + { + "epoch": 0.8454203953421591, + "grad_norm": 0.43140271306037903, + "learning_rate": 1.1566302645177884e-05, + "loss": 0.0563, + "step": 36065 + }, + { + "epoch": 0.8454438369169641, + "grad_norm": 0.3783288300037384, + "learning_rate": 1.1562864000672258e-05, + "loss": 0.0453, + "step": 36066 + }, + { + "epoch": 0.8454672784917691, + "grad_norm": 0.5790485739707947, + "learning_rate": 1.1559425836029013e-05, + "loss": 0.6179, + "step": 36067 + }, + { + "epoch": 0.8454907200665741, + "grad_norm": 0.246117502450943, + "learning_rate": 1.1555988151266816e-05, + "loss": 0.0466, + "step": 36068 + }, + { + "epoch": 0.845514161641379, + "grad_norm": 0.5033441781997681, + "learning_rate": 1.1552550946404317e-05, + "loss": 0.6015, + "step": 36069 + }, + { + "epoch": 0.8455376032161841, + "grad_norm": 0.5142913460731506, + "learning_rate": 1.1549114221460133e-05, + "loss": 0.0505, + "step": 36070 + }, + { + "epoch": 0.845561044790989, + "grad_norm": 0.09076918661594391, + "learning_rate": 1.1545677976452974e-05, + "loss": 0.0107, + "step": 36071 + }, + { + "epoch": 0.8455844863657941, + "grad_norm": 0.7430880069732666, + "learning_rate": 1.1542242211401444e-05, + "loss": 0.167, + "step": 36072 + }, + { + "epoch": 0.845607927940599, + "grad_norm": 0.4704490602016449, + "learning_rate": 1.1538806926324198e-05, + "loss": 0.5593, + "step": 36073 + }, + { + "epoch": 0.8456313695154041, + "grad_norm": 0.6435251235961914, + "learning_rate": 1.1535372121239874e-05, + "loss": 0.1361, + "step": 36074 + }, + { + "epoch": 0.845654811090209, + "grad_norm": 0.08806730806827545, + "learning_rate": 1.1531937796167114e-05, + "loss": 0.0124, + "step": 36075 + }, + { + "epoch": 0.8456782526650141, + "grad_norm": 0.6810750961303711, + "learning_rate": 1.1528503951124514e-05, + "loss": 0.1445, + "step": 36076 + }, + { + "epoch": 0.845701694239819, + "grad_norm": 0.24767637252807617, + "learning_rate": 1.1525070586130749e-05, + "loss": 0.0467, + "step": 36077 + }, + { + "epoch": 0.845725135814624, + "grad_norm": 0.3545711636543274, + "learning_rate": 1.1521637701204458e-05, + "loss": 0.0246, + "step": 36078 + }, + { + "epoch": 0.845748577389429, + "grad_norm": 0.44151487946510315, + "learning_rate": 1.1518205296364238e-05, + "loss": 0.087, + "step": 36079 + }, + { + "epoch": 0.845772018964234, + "grad_norm": 0.266524076461792, + "learning_rate": 1.151477337162874e-05, + "loss": 0.0369, + "step": 36080 + }, + { + "epoch": 0.845795460539039, + "grad_norm": 0.4192885458469391, + "learning_rate": 1.151134192701656e-05, + "loss": 0.0542, + "step": 36081 + }, + { + "epoch": 0.845818902113844, + "grad_norm": 0.2932094931602478, + "learning_rate": 1.1507910962546331e-05, + "loss": 0.0302, + "step": 36082 + }, + { + "epoch": 0.845842343688649, + "grad_norm": 0.2545245587825775, + "learning_rate": 1.1504480478236635e-05, + "loss": 0.04, + "step": 36083 + }, + { + "epoch": 0.845865785263454, + "grad_norm": 0.33462825417518616, + "learning_rate": 1.1501050474106146e-05, + "loss": 0.0569, + "step": 36084 + }, + { + "epoch": 0.845889226838259, + "grad_norm": 0.07261139154434204, + "learning_rate": 1.1497620950173448e-05, + "loss": 0.0166, + "step": 36085 + }, + { + "epoch": 0.845912668413064, + "grad_norm": 0.301684707403183, + "learning_rate": 1.1494191906457141e-05, + "loss": 0.034, + "step": 36086 + }, + { + "epoch": 0.845936109987869, + "grad_norm": 0.2570466995239258, + "learning_rate": 1.1490763342975842e-05, + "loss": 0.0265, + "step": 36087 + }, + { + "epoch": 0.845959551562674, + "grad_norm": 0.20114746689796448, + "learning_rate": 1.1487335259748122e-05, + "loss": 0.0308, + "step": 36088 + }, + { + "epoch": 0.8459829931374789, + "grad_norm": 0.3502994477748871, + "learning_rate": 1.1483907656792637e-05, + "loss": 0.0619, + "step": 36089 + }, + { + "epoch": 0.846006434712284, + "grad_norm": 0.5928285717964172, + "learning_rate": 1.1480480534127935e-05, + "loss": 0.5661, + "step": 36090 + }, + { + "epoch": 0.8460298762870889, + "grad_norm": 0.5917401909828186, + "learning_rate": 1.1477053891772649e-05, + "loss": 0.0539, + "step": 36091 + }, + { + "epoch": 0.846053317861894, + "grad_norm": 0.269376665353775, + "learning_rate": 1.1473627729745373e-05, + "loss": 0.0265, + "step": 36092 + }, + { + "epoch": 0.8460767594366989, + "grad_norm": 0.6954563856124878, + "learning_rate": 1.147020204806467e-05, + "loss": 0.1131, + "step": 36093 + }, + { + "epoch": 0.846100201011504, + "grad_norm": 0.4226309657096863, + "learning_rate": 1.1466776846749138e-05, + "loss": 0.1, + "step": 36094 + }, + { + "epoch": 0.8461236425863089, + "grad_norm": 0.2997116446495056, + "learning_rate": 1.1463352125817372e-05, + "loss": 0.0304, + "step": 36095 + }, + { + "epoch": 0.846147084161114, + "grad_norm": 0.555623471736908, + "learning_rate": 1.1459927885287924e-05, + "loss": 0.1194, + "step": 36096 + }, + { + "epoch": 0.8461705257359189, + "grad_norm": 0.5473924875259399, + "learning_rate": 1.1456504125179412e-05, + "loss": 0.6249, + "step": 36097 + }, + { + "epoch": 0.8461939673107239, + "grad_norm": 0.2602459788322449, + "learning_rate": 1.1453080845510411e-05, + "loss": 0.0327, + "step": 36098 + }, + { + "epoch": 0.8462174088855289, + "grad_norm": 0.2189694344997406, + "learning_rate": 1.1449658046299471e-05, + "loss": 0.0275, + "step": 36099 + }, + { + "epoch": 0.8462408504603339, + "grad_norm": 0.353045254945755, + "learning_rate": 1.1446235727565158e-05, + "loss": 0.0557, + "step": 36100 + }, + { + "epoch": 0.8462642920351389, + "grad_norm": 0.5578975081443787, + "learning_rate": 1.1442813889326087e-05, + "loss": 0.0661, + "step": 36101 + }, + { + "epoch": 0.8462877336099439, + "grad_norm": 0.29338541626930237, + "learning_rate": 1.143939253160079e-05, + "loss": 0.0542, + "step": 36102 + }, + { + "epoch": 0.846311175184749, + "grad_norm": 0.3767949044704437, + "learning_rate": 1.1435971654407817e-05, + "loss": 0.0494, + "step": 36103 + }, + { + "epoch": 0.8463346167595539, + "grad_norm": 0.3744231164455414, + "learning_rate": 1.1432551257765777e-05, + "loss": 0.3381, + "step": 36104 + }, + { + "epoch": 0.846358058334359, + "grad_norm": 0.5099907517433167, + "learning_rate": 1.1429131341693211e-05, + "loss": 0.0478, + "step": 36105 + }, + { + "epoch": 0.8463814999091639, + "grad_norm": 0.5954089164733887, + "learning_rate": 1.1425711906208658e-05, + "loss": 0.1581, + "step": 36106 + }, + { + "epoch": 0.8464049414839689, + "grad_norm": 0.3886389136314392, + "learning_rate": 1.1422292951330683e-05, + "loss": 0.0531, + "step": 36107 + }, + { + "epoch": 0.8464283830587739, + "grad_norm": 0.5742383599281311, + "learning_rate": 1.1418874477077823e-05, + "loss": 0.0575, + "step": 36108 + }, + { + "epoch": 0.8464518246335789, + "grad_norm": 1.0656068325042725, + "learning_rate": 1.1415456483468657e-05, + "loss": 0.0595, + "step": 36109 + }, + { + "epoch": 0.8464752662083839, + "grad_norm": 0.23794618248939514, + "learning_rate": 1.141203897052172e-05, + "loss": 0.047, + "step": 36110 + }, + { + "epoch": 0.8464987077831889, + "grad_norm": 0.3480725884437561, + "learning_rate": 1.1408621938255548e-05, + "loss": 0.1053, + "step": 36111 + }, + { + "epoch": 0.8465221493579939, + "grad_norm": 0.7629507184028625, + "learning_rate": 1.1405205386688678e-05, + "loss": 0.1362, + "step": 36112 + }, + { + "epoch": 0.8465455909327989, + "grad_norm": 0.5492976903915405, + "learning_rate": 1.140178931583964e-05, + "loss": 0.0743, + "step": 36113 + }, + { + "epoch": 0.8465690325076038, + "grad_norm": 0.2307790219783783, + "learning_rate": 1.1398373725727008e-05, + "loss": 0.0335, + "step": 36114 + }, + { + "epoch": 0.8465924740824089, + "grad_norm": 0.5134386420249939, + "learning_rate": 1.1394958616369267e-05, + "loss": 0.4402, + "step": 36115 + }, + { + "epoch": 0.8466159156572138, + "grad_norm": 0.13884663581848145, + "learning_rate": 1.1391543987784992e-05, + "loss": 0.0219, + "step": 36116 + }, + { + "epoch": 0.8466393572320189, + "grad_norm": 0.3968223035335541, + "learning_rate": 1.138812983999269e-05, + "loss": 0.0667, + "step": 36117 + }, + { + "epoch": 0.8466627988068238, + "grad_norm": 0.10593615472316742, + "learning_rate": 1.138471617301089e-05, + "loss": 0.0159, + "step": 36118 + }, + { + "epoch": 0.8466862403816289, + "grad_norm": 0.40780386328697205, + "learning_rate": 1.1381302986858112e-05, + "loss": 0.0516, + "step": 36119 + }, + { + "epoch": 0.8467096819564338, + "grad_norm": 0.1139666736125946, + "learning_rate": 1.1377890281552872e-05, + "loss": 0.0125, + "step": 36120 + }, + { + "epoch": 0.8467331235312389, + "grad_norm": 0.21547800302505493, + "learning_rate": 1.1374478057113669e-05, + "loss": 0.0582, + "step": 36121 + }, + { + "epoch": 0.8467565651060438, + "grad_norm": 0.40425074100494385, + "learning_rate": 1.1371066313559065e-05, + "loss": 0.0792, + "step": 36122 + }, + { + "epoch": 0.8467800066808489, + "grad_norm": 0.1391851007938385, + "learning_rate": 1.1367655050907544e-05, + "loss": 0.0144, + "step": 36123 + }, + { + "epoch": 0.8468034482556538, + "grad_norm": 0.3940099775791168, + "learning_rate": 1.1364244269177615e-05, + "loss": 0.0508, + "step": 36124 + }, + { + "epoch": 0.8468268898304588, + "grad_norm": 0.4961681365966797, + "learning_rate": 1.1360833968387774e-05, + "loss": 0.0865, + "step": 36125 + }, + { + "epoch": 0.8468503314052638, + "grad_norm": 0.3233782649040222, + "learning_rate": 1.1357424148556561e-05, + "loss": 0.0337, + "step": 36126 + }, + { + "epoch": 0.8468737729800688, + "grad_norm": 0.251390278339386, + "learning_rate": 1.1354014809702462e-05, + "loss": 0.0547, + "step": 36127 + }, + { + "epoch": 0.8468972145548738, + "grad_norm": 0.5280258059501648, + "learning_rate": 1.135060595184394e-05, + "loss": 0.0617, + "step": 36128 + }, + { + "epoch": 0.8469206561296788, + "grad_norm": 0.0885893777012825, + "learning_rate": 1.1347197574999557e-05, + "loss": 0.0122, + "step": 36129 + }, + { + "epoch": 0.8469440977044838, + "grad_norm": 0.41352808475494385, + "learning_rate": 1.1343789679187756e-05, + "loss": 0.1122, + "step": 36130 + }, + { + "epoch": 0.8469675392792888, + "grad_norm": 0.10889876633882523, + "learning_rate": 1.1340382264427064e-05, + "loss": 0.0116, + "step": 36131 + }, + { + "epoch": 0.8469909808540937, + "grad_norm": 0.5306865572929382, + "learning_rate": 1.1336975330735932e-05, + "loss": 0.1119, + "step": 36132 + }, + { + "epoch": 0.8470144224288988, + "grad_norm": 0.3418016731739044, + "learning_rate": 1.1333568878132872e-05, + "loss": 0.056, + "step": 36133 + }, + { + "epoch": 0.8470378640037037, + "grad_norm": 0.2915590703487396, + "learning_rate": 1.1330162906636343e-05, + "loss": 0.0588, + "step": 36134 + }, + { + "epoch": 0.8470613055785088, + "grad_norm": 0.3465489149093628, + "learning_rate": 1.1326757416264866e-05, + "loss": 0.064, + "step": 36135 + }, + { + "epoch": 0.8470847471533137, + "grad_norm": 0.20341899991035461, + "learning_rate": 1.1323352407036891e-05, + "loss": 0.028, + "step": 36136 + }, + { + "epoch": 0.8471081887281188, + "grad_norm": 0.3768683075904846, + "learning_rate": 1.1319947878970872e-05, + "loss": 0.0574, + "step": 36137 + }, + { + "epoch": 0.8471316303029237, + "grad_norm": 0.20012561976909637, + "learning_rate": 1.1316543832085335e-05, + "loss": 0.0541, + "step": 36138 + }, + { + "epoch": 0.8471550718777288, + "grad_norm": 0.2159336805343628, + "learning_rate": 1.1313140266398736e-05, + "loss": 0.0477, + "step": 36139 + }, + { + "epoch": 0.8471785134525337, + "grad_norm": 0.24586312472820282, + "learning_rate": 1.1309737181929525e-05, + "loss": 0.0341, + "step": 36140 + }, + { + "epoch": 0.8472019550273387, + "grad_norm": 0.8012711405754089, + "learning_rate": 1.1306334578696142e-05, + "loss": 0.0967, + "step": 36141 + }, + { + "epoch": 0.8472253966021437, + "grad_norm": 0.3373144865036011, + "learning_rate": 1.130293245671712e-05, + "loss": 0.0287, + "step": 36142 + }, + { + "epoch": 0.8472488381769487, + "grad_norm": 0.29402798414230347, + "learning_rate": 1.1299530816010861e-05, + "loss": 0.0333, + "step": 36143 + }, + { + "epoch": 0.8472722797517537, + "grad_norm": 0.26671478152275085, + "learning_rate": 1.1296129656595856e-05, + "loss": 0.0132, + "step": 36144 + }, + { + "epoch": 0.8472957213265587, + "grad_norm": 0.8347368836402893, + "learning_rate": 1.129272897849054e-05, + "loss": 0.1095, + "step": 36145 + }, + { + "epoch": 0.8473191629013637, + "grad_norm": 0.09515148401260376, + "learning_rate": 1.128932878171336e-05, + "loss": 0.0061, + "step": 36146 + }, + { + "epoch": 0.8473426044761687, + "grad_norm": 0.30885839462280273, + "learning_rate": 1.1285929066282785e-05, + "loss": 0.0339, + "step": 36147 + }, + { + "epoch": 0.8473660460509737, + "grad_norm": 0.25230714678764343, + "learning_rate": 1.1282529832217259e-05, + "loss": 0.0357, + "step": 36148 + }, + { + "epoch": 0.8473894876257787, + "grad_norm": 0.1108599603176117, + "learning_rate": 1.12791310795352e-05, + "loss": 0.0135, + "step": 36149 + }, + { + "epoch": 0.8474129292005836, + "grad_norm": 0.2204163819551468, + "learning_rate": 1.127573280825508e-05, + "loss": 0.0345, + "step": 36150 + }, + { + "epoch": 0.8474363707753887, + "grad_norm": 0.336765855550766, + "learning_rate": 1.1272335018395341e-05, + "loss": 0.029, + "step": 36151 + }, + { + "epoch": 0.8474598123501936, + "grad_norm": 0.16219183802604675, + "learning_rate": 1.1268937709974403e-05, + "loss": 0.0266, + "step": 36152 + }, + { + "epoch": 0.8474832539249987, + "grad_norm": 0.4009976089000702, + "learning_rate": 1.1265540883010684e-05, + "loss": 0.0329, + "step": 36153 + }, + { + "epoch": 0.8475066954998037, + "grad_norm": 0.4078163206577301, + "learning_rate": 1.1262144537522657e-05, + "loss": 0.0432, + "step": 36154 + }, + { + "epoch": 0.8475301370746087, + "grad_norm": 0.4704589545726776, + "learning_rate": 1.125874867352873e-05, + "loss": 0.0341, + "step": 36155 + }, + { + "epoch": 0.8475535786494137, + "grad_norm": 0.34194663166999817, + "learning_rate": 1.1255353291047321e-05, + "loss": 0.0323, + "step": 36156 + }, + { + "epoch": 0.8475770202242187, + "grad_norm": 0.3705286681652069, + "learning_rate": 1.1251958390096874e-05, + "loss": 0.0769, + "step": 36157 + }, + { + "epoch": 0.8476004617990237, + "grad_norm": 0.22549447417259216, + "learning_rate": 1.124856397069578e-05, + "loss": 0.0217, + "step": 36158 + }, + { + "epoch": 0.8476239033738286, + "grad_norm": 0.23632971942424774, + "learning_rate": 1.1245170032862463e-05, + "loss": 0.0489, + "step": 36159 + }, + { + "epoch": 0.8476473449486337, + "grad_norm": 0.32940706610679626, + "learning_rate": 1.1241776576615359e-05, + "loss": 0.0203, + "step": 36160 + }, + { + "epoch": 0.8476707865234386, + "grad_norm": 0.41623738408088684, + "learning_rate": 1.1238383601972857e-05, + "loss": 0.0616, + "step": 36161 + }, + { + "epoch": 0.8476942280982437, + "grad_norm": 0.3550839126110077, + "learning_rate": 1.1234991108953408e-05, + "loss": 0.0552, + "step": 36162 + }, + { + "epoch": 0.8477176696730486, + "grad_norm": 0.412513792514801, + "learning_rate": 1.1231599097575385e-05, + "loss": 0.0698, + "step": 36163 + }, + { + "epoch": 0.8477411112478537, + "grad_norm": 0.20773689448833466, + "learning_rate": 1.1228207567857208e-05, + "loss": 0.0251, + "step": 36164 + }, + { + "epoch": 0.8477645528226586, + "grad_norm": 0.5983902812004089, + "learning_rate": 1.1224816519817272e-05, + "loss": 0.0846, + "step": 36165 + }, + { + "epoch": 0.8477879943974637, + "grad_norm": 0.37301862239837646, + "learning_rate": 1.1221425953473952e-05, + "loss": 0.0302, + "step": 36166 + }, + { + "epoch": 0.8478114359722686, + "grad_norm": 0.5525056719779968, + "learning_rate": 1.121803586884569e-05, + "loss": 0.127, + "step": 36167 + }, + { + "epoch": 0.8478348775470737, + "grad_norm": 0.4092031419277191, + "learning_rate": 1.1214646265950868e-05, + "loss": 0.0512, + "step": 36168 + }, + { + "epoch": 0.8478583191218786, + "grad_norm": 0.8876424431800842, + "learning_rate": 1.1211257144807863e-05, + "loss": 0.2137, + "step": 36169 + }, + { + "epoch": 0.8478817606966836, + "grad_norm": 0.30220693349838257, + "learning_rate": 1.1207868505435082e-05, + "loss": 0.0629, + "step": 36170 + }, + { + "epoch": 0.8479052022714886, + "grad_norm": 0.430726557970047, + "learning_rate": 1.1204480347850898e-05, + "loss": 0.0656, + "step": 36171 + }, + { + "epoch": 0.8479286438462936, + "grad_norm": 0.5109672546386719, + "learning_rate": 1.1201092672073688e-05, + "loss": 0.0769, + "step": 36172 + }, + { + "epoch": 0.8479520854210986, + "grad_norm": 0.37374934554100037, + "learning_rate": 1.1197705478121833e-05, + "loss": 0.0402, + "step": 36173 + }, + { + "epoch": 0.8479755269959036, + "grad_norm": 0.13406847417354584, + "learning_rate": 1.1194318766013756e-05, + "loss": 0.021, + "step": 36174 + }, + { + "epoch": 0.8479989685707086, + "grad_norm": 0.554389476776123, + "learning_rate": 1.1190932535767795e-05, + "loss": 0.0835, + "step": 36175 + }, + { + "epoch": 0.8480224101455136, + "grad_norm": 0.3884612023830414, + "learning_rate": 1.1187546787402336e-05, + "loss": 0.0574, + "step": 36176 + }, + { + "epoch": 0.8480458517203185, + "grad_norm": 0.30624860525131226, + "learning_rate": 1.1184161520935742e-05, + "loss": 0.0544, + "step": 36177 + }, + { + "epoch": 0.8480692932951236, + "grad_norm": 0.1919955313205719, + "learning_rate": 1.118077673638639e-05, + "loss": 0.0446, + "step": 36178 + }, + { + "epoch": 0.8480927348699285, + "grad_norm": 0.5636441111564636, + "learning_rate": 1.1177392433772615e-05, + "loss": 0.0895, + "step": 36179 + }, + { + "epoch": 0.8481161764447336, + "grad_norm": 0.12348403036594391, + "learning_rate": 1.1174008613112829e-05, + "loss": 0.0218, + "step": 36180 + }, + { + "epoch": 0.8481396180195385, + "grad_norm": 0.5539320111274719, + "learning_rate": 1.1170625274425373e-05, + "loss": 0.1137, + "step": 36181 + }, + { + "epoch": 0.8481630595943436, + "grad_norm": 0.39986151456832886, + "learning_rate": 1.1167242417728608e-05, + "loss": 0.0503, + "step": 36182 + }, + { + "epoch": 0.8481865011691485, + "grad_norm": 0.46331965923309326, + "learning_rate": 1.1163860043040863e-05, + "loss": 0.556, + "step": 36183 + }, + { + "epoch": 0.8482099427439536, + "grad_norm": 0.4325988292694092, + "learning_rate": 1.1160478150380504e-05, + "loss": 0.0375, + "step": 36184 + }, + { + "epoch": 0.8482333843187585, + "grad_norm": 0.21746420860290527, + "learning_rate": 1.1157096739765916e-05, + "loss": 0.0295, + "step": 36185 + }, + { + "epoch": 0.8482568258935635, + "grad_norm": 0.6223337650299072, + "learning_rate": 1.1153715811215382e-05, + "loss": 0.4359, + "step": 36186 + }, + { + "epoch": 0.8482802674683685, + "grad_norm": 0.1359628587961197, + "learning_rate": 1.1150335364747311e-05, + "loss": 0.0142, + "step": 36187 + }, + { + "epoch": 0.8483037090431735, + "grad_norm": 0.5603695511817932, + "learning_rate": 1.1146955400380021e-05, + "loss": 0.0813, + "step": 36188 + }, + { + "epoch": 0.8483271506179785, + "grad_norm": 0.4280756711959839, + "learning_rate": 1.1143575918131854e-05, + "loss": 0.0566, + "step": 36189 + }, + { + "epoch": 0.8483505921927835, + "grad_norm": 0.44084855914115906, + "learning_rate": 1.1140196918021128e-05, + "loss": 0.0796, + "step": 36190 + }, + { + "epoch": 0.8483740337675885, + "grad_norm": 0.45121854543685913, + "learning_rate": 1.1136818400066184e-05, + "loss": 0.0493, + "step": 36191 + }, + { + "epoch": 0.8483974753423935, + "grad_norm": 0.6575531363487244, + "learning_rate": 1.1133440364285374e-05, + "loss": 0.1231, + "step": 36192 + }, + { + "epoch": 0.8484209169171985, + "grad_norm": 0.4196426570415497, + "learning_rate": 1.1130062810697028e-05, + "loss": 0.0611, + "step": 36193 + }, + { + "epoch": 0.8484443584920035, + "grad_norm": 0.24557538330554962, + "learning_rate": 1.1126685739319454e-05, + "loss": 0.0181, + "step": 36194 + }, + { + "epoch": 0.8484678000668084, + "grad_norm": 0.49968913197517395, + "learning_rate": 1.112330915017098e-05, + "loss": 0.0841, + "step": 36195 + }, + { + "epoch": 0.8484912416416135, + "grad_norm": 0.5023306608200073, + "learning_rate": 1.1119933043269926e-05, + "loss": 0.109, + "step": 36196 + }, + { + "epoch": 0.8485146832164184, + "grad_norm": 0.22899268567562103, + "learning_rate": 1.11165574186346e-05, + "loss": 0.0189, + "step": 36197 + }, + { + "epoch": 0.8485381247912235, + "grad_norm": 0.5135529637336731, + "learning_rate": 1.111318227628333e-05, + "loss": 0.0396, + "step": 36198 + }, + { + "epoch": 0.8485615663660284, + "grad_norm": 0.29077836871147156, + "learning_rate": 1.1109807616234447e-05, + "loss": 0.0331, + "step": 36199 + }, + { + "epoch": 0.8485850079408335, + "grad_norm": 0.4967471957206726, + "learning_rate": 1.1106433438506258e-05, + "loss": 0.0565, + "step": 36200 + }, + { + "epoch": 0.8486084495156384, + "grad_norm": 0.4692879319190979, + "learning_rate": 1.110305974311705e-05, + "loss": 0.0307, + "step": 36201 + }, + { + "epoch": 0.8486318910904435, + "grad_norm": 0.21031522750854492, + "learning_rate": 1.1099686530085151e-05, + "loss": 0.0317, + "step": 36202 + }, + { + "epoch": 0.8486553326652484, + "grad_norm": 0.02984040416777134, + "learning_rate": 1.1096313799428848e-05, + "loss": 0.0042, + "step": 36203 + }, + { + "epoch": 0.8486787742400534, + "grad_norm": 0.9425045847892761, + "learning_rate": 1.1092941551166425e-05, + "loss": 0.151, + "step": 36204 + }, + { + "epoch": 0.8487022158148585, + "grad_norm": 0.6590411067008972, + "learning_rate": 1.1089569785316212e-05, + "loss": 0.4624, + "step": 36205 + }, + { + "epoch": 0.8487256573896634, + "grad_norm": 0.4172409474849701, + "learning_rate": 1.1086198501896505e-05, + "loss": 0.1218, + "step": 36206 + }, + { + "epoch": 0.8487490989644685, + "grad_norm": 0.2733609676361084, + "learning_rate": 1.108282770092558e-05, + "loss": 0.0537, + "step": 36207 + }, + { + "epoch": 0.8487725405392734, + "grad_norm": 0.33343496918678284, + "learning_rate": 1.1079457382421742e-05, + "loss": 0.0738, + "step": 36208 + }, + { + "epoch": 0.8487959821140785, + "grad_norm": 0.18878421187400818, + "learning_rate": 1.1076087546403258e-05, + "loss": 0.0416, + "step": 36209 + }, + { + "epoch": 0.8488194236888834, + "grad_norm": 0.05857657268643379, + "learning_rate": 1.1072718192888409e-05, + "loss": 0.0096, + "step": 36210 + }, + { + "epoch": 0.8488428652636885, + "grad_norm": 0.45800551772117615, + "learning_rate": 1.1069349321895483e-05, + "loss": 0.0813, + "step": 36211 + }, + { + "epoch": 0.8488663068384934, + "grad_norm": 0.09284529834985733, + "learning_rate": 1.1065980933442799e-05, + "loss": 0.0123, + "step": 36212 + }, + { + "epoch": 0.8488897484132985, + "grad_norm": 0.44482293725013733, + "learning_rate": 1.1062613027548597e-05, + "loss": 0.0741, + "step": 36213 + }, + { + "epoch": 0.8489131899881034, + "grad_norm": 0.3465767502784729, + "learning_rate": 1.105924560423115e-05, + "loss": 0.0506, + "step": 36214 + }, + { + "epoch": 0.8489366315629084, + "grad_norm": 0.6244933009147644, + "learning_rate": 1.1055878663508745e-05, + "loss": 0.1058, + "step": 36215 + }, + { + "epoch": 0.8489600731377134, + "grad_norm": 0.44034698605537415, + "learning_rate": 1.1052512205399646e-05, + "loss": 0.048, + "step": 36216 + }, + { + "epoch": 0.8489835147125184, + "grad_norm": 0.23775675892829895, + "learning_rate": 1.104914622992209e-05, + "loss": 0.0342, + "step": 36217 + }, + { + "epoch": 0.8490069562873234, + "grad_norm": 0.3348162770271301, + "learning_rate": 1.1045780737094392e-05, + "loss": 0.4559, + "step": 36218 + }, + { + "epoch": 0.8490303978621284, + "grad_norm": 0.2861323952674866, + "learning_rate": 1.1042415726934796e-05, + "loss": 0.05, + "step": 36219 + }, + { + "epoch": 0.8490538394369334, + "grad_norm": 0.41973841190338135, + "learning_rate": 1.1039051199461536e-05, + "loss": 0.0544, + "step": 36220 + }, + { + "epoch": 0.8490772810117384, + "grad_norm": 0.610167920589447, + "learning_rate": 1.1035687154692898e-05, + "loss": 0.152, + "step": 36221 + }, + { + "epoch": 0.8491007225865433, + "grad_norm": 0.35695236921310425, + "learning_rate": 1.1032323592647099e-05, + "loss": 0.5086, + "step": 36222 + }, + { + "epoch": 0.8491241641613484, + "grad_norm": 0.2298620641231537, + "learning_rate": 1.1028960513342435e-05, + "loss": 0.0377, + "step": 36223 + }, + { + "epoch": 0.8491476057361533, + "grad_norm": 0.5590099692344666, + "learning_rate": 1.1025597916797115e-05, + "loss": 0.4418, + "step": 36224 + }, + { + "epoch": 0.8491710473109584, + "grad_norm": 0.2942159175872803, + "learning_rate": 1.1022235803029423e-05, + "loss": 0.0321, + "step": 36225 + }, + { + "epoch": 0.8491944888857633, + "grad_norm": 0.2761819362640381, + "learning_rate": 1.101887417205758e-05, + "loss": 0.2599, + "step": 36226 + }, + { + "epoch": 0.8492179304605684, + "grad_norm": 0.7346005439758301, + "learning_rate": 1.1015513023899826e-05, + "loss": 0.1235, + "step": 36227 + }, + { + "epoch": 0.8492413720353733, + "grad_norm": 0.679509699344635, + "learning_rate": 1.1012152358574401e-05, + "loss": 0.1454, + "step": 36228 + }, + { + "epoch": 0.8492648136101784, + "grad_norm": 0.16436083614826202, + "learning_rate": 1.1008792176099514e-05, + "loss": 0.0248, + "step": 36229 + }, + { + "epoch": 0.8492882551849833, + "grad_norm": 0.0973222479224205, + "learning_rate": 1.1005432476493449e-05, + "loss": 0.0163, + "step": 36230 + }, + { + "epoch": 0.8493116967597883, + "grad_norm": 0.2917250096797943, + "learning_rate": 1.1002073259774414e-05, + "loss": 0.0371, + "step": 36231 + }, + { + "epoch": 0.8493351383345933, + "grad_norm": 0.5411692261695862, + "learning_rate": 1.099871452596063e-05, + "loss": 0.0762, + "step": 36232 + }, + { + "epoch": 0.8493585799093983, + "grad_norm": 0.39216452836990356, + "learning_rate": 1.0995356275070312e-05, + "loss": 0.0362, + "step": 36233 + }, + { + "epoch": 0.8493820214842033, + "grad_norm": 0.12331221997737885, + "learning_rate": 1.0991998507121682e-05, + "loss": 0.0257, + "step": 36234 + }, + { + "epoch": 0.8494054630590083, + "grad_norm": 0.595710813999176, + "learning_rate": 1.098864122213299e-05, + "loss": 0.0781, + "step": 36235 + }, + { + "epoch": 0.8494289046338133, + "grad_norm": 0.2010914832353592, + "learning_rate": 1.0985284420122422e-05, + "loss": 0.0316, + "step": 36236 + }, + { + "epoch": 0.8494523462086183, + "grad_norm": 0.17263418436050415, + "learning_rate": 1.0981928101108208e-05, + "loss": 0.047, + "step": 36237 + }, + { + "epoch": 0.8494757877834233, + "grad_norm": 0.7562089562416077, + "learning_rate": 1.0978572265108567e-05, + "loss": 0.0793, + "step": 36238 + }, + { + "epoch": 0.8494992293582283, + "grad_norm": 0.3653669059276581, + "learning_rate": 1.0975216912141695e-05, + "loss": 0.0346, + "step": 36239 + }, + { + "epoch": 0.8495226709330332, + "grad_norm": 0.17451590299606323, + "learning_rate": 1.0971862042225789e-05, + "loss": 0.0116, + "step": 36240 + }, + { + "epoch": 0.8495461125078383, + "grad_norm": 0.20931574702262878, + "learning_rate": 1.0968507655379067e-05, + "loss": 0.0308, + "step": 36241 + }, + { + "epoch": 0.8495695540826432, + "grad_norm": 0.4947563409805298, + "learning_rate": 1.0965153751619717e-05, + "loss": 0.0988, + "step": 36242 + }, + { + "epoch": 0.8495929956574483, + "grad_norm": 0.4427349269390106, + "learning_rate": 1.0961800330965955e-05, + "loss": 0.0545, + "step": 36243 + }, + { + "epoch": 0.8496164372322532, + "grad_norm": 0.5395512580871582, + "learning_rate": 1.0958447393435967e-05, + "loss": 0.0555, + "step": 36244 + }, + { + "epoch": 0.8496398788070583, + "grad_norm": 0.38774967193603516, + "learning_rate": 1.095509493904795e-05, + "loss": 0.0478, + "step": 36245 + }, + { + "epoch": 0.8496633203818632, + "grad_norm": 0.05267457664012909, + "learning_rate": 1.0951742967820078e-05, + "loss": 0.0067, + "step": 36246 + }, + { + "epoch": 0.8496867619566683, + "grad_norm": 0.42132076621055603, + "learning_rate": 1.094839147977057e-05, + "loss": 0.0813, + "step": 36247 + }, + { + "epoch": 0.8497102035314732, + "grad_norm": 0.2818702161312103, + "learning_rate": 1.094504047491759e-05, + "loss": 0.0446, + "step": 36248 + }, + { + "epoch": 0.8497336451062782, + "grad_norm": 0.501757800579071, + "learning_rate": 1.0941689953279299e-05, + "loss": 0.5865, + "step": 36249 + }, + { + "epoch": 0.8497570866810832, + "grad_norm": 0.43632015585899353, + "learning_rate": 1.093833991487393e-05, + "loss": 0.0714, + "step": 36250 + }, + { + "epoch": 0.8497805282558882, + "grad_norm": 0.3605601191520691, + "learning_rate": 1.0934990359719633e-05, + "loss": 0.0524, + "step": 36251 + }, + { + "epoch": 0.8498039698306932, + "grad_norm": 0.21871809661388397, + "learning_rate": 1.0931641287834582e-05, + "loss": 0.0231, + "step": 36252 + }, + { + "epoch": 0.8498274114054982, + "grad_norm": 0.10862986743450165, + "learning_rate": 1.0928292699236941e-05, + "loss": 0.0116, + "step": 36253 + }, + { + "epoch": 0.8498508529803032, + "grad_norm": 0.6213966608047485, + "learning_rate": 1.0924944593944896e-05, + "loss": 0.5908, + "step": 36254 + }, + { + "epoch": 0.8498742945551082, + "grad_norm": 0.7350158095359802, + "learning_rate": 1.0921596971976588e-05, + "loss": 0.0821, + "step": 36255 + }, + { + "epoch": 0.8498977361299133, + "grad_norm": 0.28605917096138, + "learning_rate": 1.0918249833350225e-05, + "loss": 0.0653, + "step": 36256 + }, + { + "epoch": 0.8499211777047182, + "grad_norm": 0.5501291751861572, + "learning_rate": 1.0914903178083935e-05, + "loss": 0.1011, + "step": 36257 + }, + { + "epoch": 0.8499446192795233, + "grad_norm": 0.42693230509757996, + "learning_rate": 1.0911557006195883e-05, + "loss": 0.0721, + "step": 36258 + }, + { + "epoch": 0.8499680608543282, + "grad_norm": 0.5479793548583984, + "learning_rate": 1.0908211317704208e-05, + "loss": 0.0951, + "step": 36259 + }, + { + "epoch": 0.8499915024291332, + "grad_norm": 0.5660821795463562, + "learning_rate": 1.090486611262711e-05, + "loss": 0.072, + "step": 36260 + }, + { + "epoch": 0.8500149440039382, + "grad_norm": 0.6387656331062317, + "learning_rate": 1.0901521390982705e-05, + "loss": 0.101, + "step": 36261 + }, + { + "epoch": 0.8500383855787432, + "grad_norm": 0.04160541296005249, + "learning_rate": 1.0898177152789135e-05, + "loss": 0.0028, + "step": 36262 + }, + { + "epoch": 0.8500618271535482, + "grad_norm": 0.3236271142959595, + "learning_rate": 1.0894833398064585e-05, + "loss": 0.0246, + "step": 36263 + }, + { + "epoch": 0.8500852687283532, + "grad_norm": 0.11950702965259552, + "learning_rate": 1.0891490126827165e-05, + "loss": 0.021, + "step": 36264 + }, + { + "epoch": 0.8501087103031582, + "grad_norm": 0.6258472204208374, + "learning_rate": 1.0888147339095034e-05, + "loss": 0.1097, + "step": 36265 + }, + { + "epoch": 0.8501321518779632, + "grad_norm": 0.17060324549674988, + "learning_rate": 1.0884805034886314e-05, + "loss": 0.0266, + "step": 36266 + }, + { + "epoch": 0.8501555934527681, + "grad_norm": 0.1219298467040062, + "learning_rate": 1.0881463214219135e-05, + "loss": 0.0162, + "step": 36267 + }, + { + "epoch": 0.8501790350275732, + "grad_norm": 0.3948633670806885, + "learning_rate": 1.0878121877111657e-05, + "loss": 0.0511, + "step": 36268 + }, + { + "epoch": 0.8502024766023781, + "grad_norm": 0.4547373950481415, + "learning_rate": 1.0874781023581993e-05, + "loss": 0.068, + "step": 36269 + }, + { + "epoch": 0.8502259181771832, + "grad_norm": 0.4923335611820221, + "learning_rate": 1.087144065364828e-05, + "loss": 0.0623, + "step": 36270 + }, + { + "epoch": 0.8502493597519881, + "grad_norm": 0.5042901635169983, + "learning_rate": 1.0868100767328616e-05, + "loss": 0.1206, + "step": 36271 + }, + { + "epoch": 0.8502728013267932, + "grad_norm": 0.5865179300308228, + "learning_rate": 1.0864761364641151e-05, + "loss": 0.4988, + "step": 36272 + }, + { + "epoch": 0.8502962429015981, + "grad_norm": 0.34786075353622437, + "learning_rate": 1.0861422445604009e-05, + "loss": 0.055, + "step": 36273 + }, + { + "epoch": 0.8503196844764032, + "grad_norm": 0.233620747923851, + "learning_rate": 1.085808401023527e-05, + "loss": 0.0448, + "step": 36274 + }, + { + "epoch": 0.8503431260512081, + "grad_norm": 0.4336976408958435, + "learning_rate": 1.08547460585531e-05, + "loss": 0.0587, + "step": 36275 + }, + { + "epoch": 0.8503665676260131, + "grad_norm": 0.3229091167449951, + "learning_rate": 1.0851408590575585e-05, + "loss": 0.0549, + "step": 36276 + }, + { + "epoch": 0.8503900092008181, + "grad_norm": 0.2710692286491394, + "learning_rate": 1.0848071606320832e-05, + "loss": 0.0307, + "step": 36277 + }, + { + "epoch": 0.8504134507756231, + "grad_norm": 0.28030532598495483, + "learning_rate": 1.0844735105806947e-05, + "loss": 0.0427, + "step": 36278 + }, + { + "epoch": 0.8504368923504281, + "grad_norm": 0.5405828356742859, + "learning_rate": 1.084139908905203e-05, + "loss": 0.6028, + "step": 36279 + }, + { + "epoch": 0.8504603339252331, + "grad_norm": 0.1962834894657135, + "learning_rate": 1.0838063556074175e-05, + "loss": 0.0208, + "step": 36280 + }, + { + "epoch": 0.8504837755000381, + "grad_norm": 0.30451154708862305, + "learning_rate": 1.0834728506891512e-05, + "loss": 0.0426, + "step": 36281 + }, + { + "epoch": 0.8505072170748431, + "grad_norm": 0.33075129985809326, + "learning_rate": 1.0831393941522117e-05, + "loss": 0.0459, + "step": 36282 + }, + { + "epoch": 0.850530658649648, + "grad_norm": 0.4304684102535248, + "learning_rate": 1.0828059859984074e-05, + "loss": 0.0465, + "step": 36283 + }, + { + "epoch": 0.8505541002244531, + "grad_norm": 0.3322048783302307, + "learning_rate": 1.0824726262295503e-05, + "loss": 0.048, + "step": 36284 + }, + { + "epoch": 0.850577541799258, + "grad_norm": 0.5005885362625122, + "learning_rate": 1.0821393148474468e-05, + "loss": 0.1112, + "step": 36285 + }, + { + "epoch": 0.8506009833740631, + "grad_norm": 0.23242373764514923, + "learning_rate": 1.0818060518539064e-05, + "loss": 0.027, + "step": 36286 + }, + { + "epoch": 0.850624424948868, + "grad_norm": 0.1503632515668869, + "learning_rate": 1.0814728372507355e-05, + "loss": 0.0273, + "step": 36287 + }, + { + "epoch": 0.8506478665236731, + "grad_norm": 0.3765100836753845, + "learning_rate": 1.081139671039746e-05, + "loss": 0.078, + "step": 36288 + }, + { + "epoch": 0.850671308098478, + "grad_norm": 0.5886933207511902, + "learning_rate": 1.080806553222743e-05, + "loss": 0.1284, + "step": 36289 + }, + { + "epoch": 0.8506947496732831, + "grad_norm": 0.5748564004898071, + "learning_rate": 1.080473483801534e-05, + "loss": 0.5113, + "step": 36290 + }, + { + "epoch": 0.850718191248088, + "grad_norm": 0.18895867466926575, + "learning_rate": 1.0801404627779278e-05, + "loss": 0.0342, + "step": 36291 + }, + { + "epoch": 0.8507416328228931, + "grad_norm": 0.49633675813674927, + "learning_rate": 1.0798074901537291e-05, + "loss": 0.1205, + "step": 36292 + }, + { + "epoch": 0.850765074397698, + "grad_norm": 0.1896645426750183, + "learning_rate": 1.0794745659307437e-05, + "loss": 0.0261, + "step": 36293 + }, + { + "epoch": 0.850788515972503, + "grad_norm": 0.5061383843421936, + "learning_rate": 1.079141690110783e-05, + "loss": 0.0733, + "step": 36294 + }, + { + "epoch": 0.850811957547308, + "grad_norm": 0.5392611622810364, + "learning_rate": 1.0788088626956482e-05, + "loss": 0.0765, + "step": 36295 + }, + { + "epoch": 0.850835399122113, + "grad_norm": 0.5223490595817566, + "learning_rate": 1.0784760836871489e-05, + "loss": 0.5843, + "step": 36296 + }, + { + "epoch": 0.850858840696918, + "grad_norm": 0.14186318218708038, + "learning_rate": 1.0781433530870888e-05, + "loss": 0.0175, + "step": 36297 + }, + { + "epoch": 0.850882282271723, + "grad_norm": 0.3517704904079437, + "learning_rate": 1.0778106708972746e-05, + "loss": 0.0363, + "step": 36298 + }, + { + "epoch": 0.850905723846528, + "grad_norm": 0.3478816747665405, + "learning_rate": 1.0774780371195104e-05, + "loss": 0.041, + "step": 36299 + }, + { + "epoch": 0.850929165421333, + "grad_norm": 0.710889458656311, + "learning_rate": 1.077145451755599e-05, + "loss": 0.1284, + "step": 36300 + }, + { + "epoch": 0.850952606996138, + "grad_norm": 0.06577969342470169, + "learning_rate": 1.076812914807349e-05, + "loss": 0.0038, + "step": 36301 + }, + { + "epoch": 0.850976048570943, + "grad_norm": 0.23902149498462677, + "learning_rate": 1.0764804262765638e-05, + "loss": 0.0365, + "step": 36302 + }, + { + "epoch": 0.8509994901457479, + "grad_norm": 0.40896129608154297, + "learning_rate": 1.076147986165047e-05, + "loss": 0.03, + "step": 36303 + }, + { + "epoch": 0.851022931720553, + "grad_norm": 0.08475944399833679, + "learning_rate": 1.0758155944746007e-05, + "loss": 0.014, + "step": 36304 + }, + { + "epoch": 0.8510463732953579, + "grad_norm": 0.5341225266456604, + "learning_rate": 1.075483251207029e-05, + "loss": 0.0789, + "step": 36305 + }, + { + "epoch": 0.851069814870163, + "grad_norm": 0.6187232136726379, + "learning_rate": 1.0751509563641381e-05, + "loss": 0.1077, + "step": 36306 + }, + { + "epoch": 0.851093256444968, + "grad_norm": 0.11545019596815109, + "learning_rate": 1.074818709947727e-05, + "loss": 0.0245, + "step": 36307 + }, + { + "epoch": 0.851116698019773, + "grad_norm": 0.3032612204551697, + "learning_rate": 1.0744865119596026e-05, + "loss": 0.0273, + "step": 36308 + }, + { + "epoch": 0.851140139594578, + "grad_norm": 0.15249283611774445, + "learning_rate": 1.0741543624015648e-05, + "loss": 0.0311, + "step": 36309 + }, + { + "epoch": 0.851163581169383, + "grad_norm": 0.43775686621665955, + "learning_rate": 1.0738222612754168e-05, + "loss": 0.0477, + "step": 36310 + }, + { + "epoch": 0.851187022744188, + "grad_norm": 0.6647657752037048, + "learning_rate": 1.0734902085829601e-05, + "loss": 0.0979, + "step": 36311 + }, + { + "epoch": 0.8512104643189929, + "grad_norm": 0.5299220681190491, + "learning_rate": 1.0731582043259936e-05, + "loss": 0.1121, + "step": 36312 + }, + { + "epoch": 0.851233905893798, + "grad_norm": 0.2308550328016281, + "learning_rate": 1.0728262485063246e-05, + "loss": 0.0388, + "step": 36313 + }, + { + "epoch": 0.8512573474686029, + "grad_norm": 0.4373174011707306, + "learning_rate": 1.0724943411257515e-05, + "loss": 0.0609, + "step": 36314 + }, + { + "epoch": 0.851280789043408, + "grad_norm": 0.4879259467124939, + "learning_rate": 1.0721624821860754e-05, + "loss": 0.1026, + "step": 36315 + }, + { + "epoch": 0.8513042306182129, + "grad_norm": 0.31909772753715515, + "learning_rate": 1.0718306716890958e-05, + "loss": 0.0529, + "step": 36316 + }, + { + "epoch": 0.851327672193018, + "grad_norm": 0.3920249938964844, + "learning_rate": 1.0714989096366145e-05, + "loss": 0.0665, + "step": 36317 + }, + { + "epoch": 0.8513511137678229, + "grad_norm": 0.7107418775558472, + "learning_rate": 1.0711671960304292e-05, + "loss": 0.1, + "step": 36318 + }, + { + "epoch": 0.851374555342628, + "grad_norm": 0.3693946599960327, + "learning_rate": 1.0708355308723417e-05, + "loss": 0.0149, + "step": 36319 + }, + { + "epoch": 0.8513979969174329, + "grad_norm": 0.42633163928985596, + "learning_rate": 1.0705039141641537e-05, + "loss": 0.0878, + "step": 36320 + }, + { + "epoch": 0.851421438492238, + "grad_norm": 0.23023106157779694, + "learning_rate": 1.0701723459076629e-05, + "loss": 0.0176, + "step": 36321 + }, + { + "epoch": 0.8514448800670429, + "grad_norm": 0.2898084819316864, + "learning_rate": 1.0698408261046677e-05, + "loss": 0.0312, + "step": 36322 + }, + { + "epoch": 0.8514683216418479, + "grad_norm": 0.425950825214386, + "learning_rate": 1.0695093547569668e-05, + "loss": 0.0664, + "step": 36323 + }, + { + "epoch": 0.8514917632166529, + "grad_norm": 0.3342149555683136, + "learning_rate": 1.0691779318663597e-05, + "loss": 0.026, + "step": 36324 + }, + { + "epoch": 0.8515152047914579, + "grad_norm": 0.7102892994880676, + "learning_rate": 1.068846557434643e-05, + "loss": 0.1145, + "step": 36325 + }, + { + "epoch": 0.8515386463662629, + "grad_norm": 0.49799075722694397, + "learning_rate": 1.0685152314636171e-05, + "loss": 0.1139, + "step": 36326 + }, + { + "epoch": 0.8515620879410679, + "grad_norm": 0.38029247522354126, + "learning_rate": 1.0681839539550787e-05, + "loss": 0.2734, + "step": 36327 + }, + { + "epoch": 0.8515855295158729, + "grad_norm": 0.35560429096221924, + "learning_rate": 1.067852724910825e-05, + "loss": 0.043, + "step": 36328 + }, + { + "epoch": 0.8516089710906779, + "grad_norm": 0.38446369767189026, + "learning_rate": 1.0675215443326536e-05, + "loss": 0.6023, + "step": 36329 + }, + { + "epoch": 0.8516324126654828, + "grad_norm": 0.41669636964797974, + "learning_rate": 1.067190412222362e-05, + "loss": 0.3191, + "step": 36330 + }, + { + "epoch": 0.8516558542402879, + "grad_norm": 0.23573651909828186, + "learning_rate": 1.066859328581744e-05, + "loss": 0.0553, + "step": 36331 + }, + { + "epoch": 0.8516792958150928, + "grad_norm": 0.5052312612533569, + "learning_rate": 1.0665282934125998e-05, + "loss": 0.6282, + "step": 36332 + }, + { + "epoch": 0.8517027373898979, + "grad_norm": 0.1371699869632721, + "learning_rate": 1.0661973067167241e-05, + "loss": 0.0224, + "step": 36333 + }, + { + "epoch": 0.8517261789647028, + "grad_norm": 0.4950830042362213, + "learning_rate": 1.0658663684959146e-05, + "loss": 0.0609, + "step": 36334 + }, + { + "epoch": 0.8517496205395079, + "grad_norm": 0.31926217675209045, + "learning_rate": 1.0655354787519645e-05, + "loss": 0.0449, + "step": 36335 + }, + { + "epoch": 0.8517730621143128, + "grad_norm": 0.4117577373981476, + "learning_rate": 1.065204637486671e-05, + "loss": 0.0419, + "step": 36336 + }, + { + "epoch": 0.8517965036891179, + "grad_norm": 0.42688798904418945, + "learning_rate": 1.0648738447018259e-05, + "loss": 0.0594, + "step": 36337 + }, + { + "epoch": 0.8518199452639228, + "grad_norm": 0.31621426343917847, + "learning_rate": 1.064543100399228e-05, + "loss": 0.0537, + "step": 36338 + }, + { + "epoch": 0.8518433868387278, + "grad_norm": 0.5013882517814636, + "learning_rate": 1.0642124045806712e-05, + "loss": 0.7161, + "step": 36339 + }, + { + "epoch": 0.8518668284135328, + "grad_norm": 0.45824840664863586, + "learning_rate": 1.0638817572479486e-05, + "loss": 0.0975, + "step": 36340 + }, + { + "epoch": 0.8518902699883378, + "grad_norm": 0.31606340408325195, + "learning_rate": 1.0635511584028557e-05, + "loss": 0.0774, + "step": 36341 + }, + { + "epoch": 0.8519137115631428, + "grad_norm": 0.256533682346344, + "learning_rate": 1.0632206080471851e-05, + "loss": 0.0175, + "step": 36342 + }, + { + "epoch": 0.8519371531379478, + "grad_norm": 0.028367305174469948, + "learning_rate": 1.062890106182728e-05, + "loss": 0.0018, + "step": 36343 + }, + { + "epoch": 0.8519605947127528, + "grad_norm": 0.4555727243423462, + "learning_rate": 1.0625596528112813e-05, + "loss": 0.6653, + "step": 36344 + }, + { + "epoch": 0.8519840362875578, + "grad_norm": 0.32249775528907776, + "learning_rate": 1.0622292479346385e-05, + "loss": 0.0344, + "step": 36345 + }, + { + "epoch": 0.8520074778623627, + "grad_norm": 0.3045037090778351, + "learning_rate": 1.0618988915545924e-05, + "loss": 0.0246, + "step": 36346 + }, + { + "epoch": 0.8520309194371678, + "grad_norm": 0.41966503858566284, + "learning_rate": 1.0615685836729328e-05, + "loss": 0.4748, + "step": 36347 + }, + { + "epoch": 0.8520543610119727, + "grad_norm": 0.39342889189720154, + "learning_rate": 1.0612383242914547e-05, + "loss": 0.0534, + "step": 36348 + }, + { + "epoch": 0.8520778025867778, + "grad_norm": 0.46829789876937866, + "learning_rate": 1.0609081134119481e-05, + "loss": 0.0794, + "step": 36349 + }, + { + "epoch": 0.8521012441615827, + "grad_norm": 0.11064239591360092, + "learning_rate": 1.0605779510362035e-05, + "loss": 0.0104, + "step": 36350 + }, + { + "epoch": 0.8521246857363878, + "grad_norm": 0.5047574639320374, + "learning_rate": 1.0602478371660163e-05, + "loss": 0.0704, + "step": 36351 + }, + { + "epoch": 0.8521481273111927, + "grad_norm": 0.13410139083862305, + "learning_rate": 1.059917771803175e-05, + "loss": 0.0256, + "step": 36352 + }, + { + "epoch": 0.8521715688859978, + "grad_norm": 0.3474133610725403, + "learning_rate": 1.0595877549494727e-05, + "loss": 0.0686, + "step": 36353 + }, + { + "epoch": 0.8521950104608027, + "grad_norm": 0.320654034614563, + "learning_rate": 1.0592577866066977e-05, + "loss": 0.0753, + "step": 36354 + }, + { + "epoch": 0.8522184520356078, + "grad_norm": 0.4189552962779999, + "learning_rate": 1.0589278667766412e-05, + "loss": 0.0669, + "step": 36355 + }, + { + "epoch": 0.8522418936104127, + "grad_norm": 0.10787367075681686, + "learning_rate": 1.0585979954610915e-05, + "loss": 0.0173, + "step": 36356 + }, + { + "epoch": 0.8522653351852177, + "grad_norm": 0.7528020143508911, + "learning_rate": 1.0582681726618404e-05, + "loss": 0.1135, + "step": 36357 + }, + { + "epoch": 0.8522887767600228, + "grad_norm": 0.24813592433929443, + "learning_rate": 1.05793839838068e-05, + "loss": 0.0342, + "step": 36358 + }, + { + "epoch": 0.8523122183348277, + "grad_norm": 0.25258690118789673, + "learning_rate": 1.0576086726193979e-05, + "loss": 0.0367, + "step": 36359 + }, + { + "epoch": 0.8523356599096328, + "grad_norm": 0.4244902729988098, + "learning_rate": 1.0572789953797813e-05, + "loss": 0.0624, + "step": 36360 + }, + { + "epoch": 0.8523591014844377, + "grad_norm": 0.34008392691612244, + "learning_rate": 1.056949366663621e-05, + "loss": 0.0648, + "step": 36361 + }, + { + "epoch": 0.8523825430592428, + "grad_norm": 0.7981470227241516, + "learning_rate": 1.0566197864727046e-05, + "loss": 0.1118, + "step": 36362 + }, + { + "epoch": 0.8524059846340477, + "grad_norm": 0.4008544087409973, + "learning_rate": 1.0562902548088194e-05, + "loss": 0.0414, + "step": 36363 + }, + { + "epoch": 0.8524294262088528, + "grad_norm": 0.563178300857544, + "learning_rate": 1.0559607716737563e-05, + "loss": 0.1095, + "step": 36364 + }, + { + "epoch": 0.8524528677836577, + "grad_norm": 0.3982944190502167, + "learning_rate": 1.0556313370693017e-05, + "loss": 0.0863, + "step": 36365 + }, + { + "epoch": 0.8524763093584627, + "grad_norm": 0.1928262561559677, + "learning_rate": 1.0553019509972428e-05, + "loss": 0.0298, + "step": 36366 + }, + { + "epoch": 0.8524997509332677, + "grad_norm": 0.3422229290008545, + "learning_rate": 1.0549726134593663e-05, + "loss": 0.0458, + "step": 36367 + }, + { + "epoch": 0.8525231925080727, + "grad_norm": 0.6167683005332947, + "learning_rate": 1.0546433244574583e-05, + "loss": 0.0783, + "step": 36368 + }, + { + "epoch": 0.8525466340828777, + "grad_norm": 0.3730602264404297, + "learning_rate": 1.0543140839933097e-05, + "loss": 0.0308, + "step": 36369 + }, + { + "epoch": 0.8525700756576827, + "grad_norm": 0.46293097734451294, + "learning_rate": 1.0539848920687024e-05, + "loss": 0.0628, + "step": 36370 + }, + { + "epoch": 0.8525935172324877, + "grad_norm": 0.3308444619178772, + "learning_rate": 1.053655748685426e-05, + "loss": 0.3624, + "step": 36371 + }, + { + "epoch": 0.8526169588072927, + "grad_norm": 0.28180205821990967, + "learning_rate": 1.0533266538452658e-05, + "loss": 0.0503, + "step": 36372 + }, + { + "epoch": 0.8526404003820977, + "grad_norm": 0.28957924246788025, + "learning_rate": 1.0529976075500058e-05, + "loss": 0.0505, + "step": 36373 + }, + { + "epoch": 0.8526638419569027, + "grad_norm": 0.33003562688827515, + "learning_rate": 1.0526686098014338e-05, + "loss": 0.0473, + "step": 36374 + }, + { + "epoch": 0.8526872835317076, + "grad_norm": 0.24878205358982086, + "learning_rate": 1.0523396606013302e-05, + "loss": 0.033, + "step": 36375 + }, + { + "epoch": 0.8527107251065127, + "grad_norm": 0.3556749224662781, + "learning_rate": 1.052010759951485e-05, + "loss": 0.0376, + "step": 36376 + }, + { + "epoch": 0.8527341666813176, + "grad_norm": 0.19569523632526398, + "learning_rate": 1.0516819078536821e-05, + "loss": 0.0177, + "step": 36377 + }, + { + "epoch": 0.8527576082561227, + "grad_norm": 0.1813388168811798, + "learning_rate": 1.0513531043097047e-05, + "loss": 0.0339, + "step": 36378 + }, + { + "epoch": 0.8527810498309276, + "grad_norm": 0.4542868435382843, + "learning_rate": 1.0510243493213356e-05, + "loss": 0.0667, + "step": 36379 + }, + { + "epoch": 0.8528044914057327, + "grad_norm": 0.3256160020828247, + "learning_rate": 1.0506956428903592e-05, + "loss": 0.0408, + "step": 36380 + }, + { + "epoch": 0.8528279329805376, + "grad_norm": 0.04040994495153427, + "learning_rate": 1.0503669850185615e-05, + "loss": 0.0053, + "step": 36381 + }, + { + "epoch": 0.8528513745553427, + "grad_norm": 0.3890116512775421, + "learning_rate": 1.0500383757077225e-05, + "loss": 0.0672, + "step": 36382 + }, + { + "epoch": 0.8528748161301476, + "grad_norm": 0.45410940051078796, + "learning_rate": 1.0497098149596286e-05, + "loss": 0.0556, + "step": 36383 + }, + { + "epoch": 0.8528982577049526, + "grad_norm": 0.41216841340065, + "learning_rate": 1.0493813027760612e-05, + "loss": 0.0765, + "step": 36384 + }, + { + "epoch": 0.8529216992797576, + "grad_norm": 0.13173329830169678, + "learning_rate": 1.0490528391588028e-05, + "loss": 0.0182, + "step": 36385 + }, + { + "epoch": 0.8529451408545626, + "grad_norm": 0.39517077803611755, + "learning_rate": 1.048724424109635e-05, + "loss": 0.0587, + "step": 36386 + }, + { + "epoch": 0.8529685824293676, + "grad_norm": 0.33386343717575073, + "learning_rate": 1.0483960576303398e-05, + "loss": 0.3447, + "step": 36387 + }, + { + "epoch": 0.8529920240041726, + "grad_norm": 0.6760401129722595, + "learning_rate": 1.048067739722698e-05, + "loss": 0.597, + "step": 36388 + }, + { + "epoch": 0.8530154655789776, + "grad_norm": 0.47006893157958984, + "learning_rate": 1.0477394703884936e-05, + "loss": 0.0583, + "step": 36389 + }, + { + "epoch": 0.8530389071537826, + "grad_norm": 0.7809101939201355, + "learning_rate": 1.0474112496295064e-05, + "loss": 0.1891, + "step": 36390 + }, + { + "epoch": 0.8530623487285875, + "grad_norm": 0.4929344952106476, + "learning_rate": 1.0470830774475182e-05, + "loss": 0.0939, + "step": 36391 + }, + { + "epoch": 0.8530857903033926, + "grad_norm": 0.8050488829612732, + "learning_rate": 1.0467549538443066e-05, + "loss": 0.0725, + "step": 36392 + }, + { + "epoch": 0.8531092318781975, + "grad_norm": 0.42332974076271057, + "learning_rate": 1.0464268788216568e-05, + "loss": 0.0622, + "step": 36393 + }, + { + "epoch": 0.8531326734530026, + "grad_norm": 0.17618107795715332, + "learning_rate": 1.0460988523813464e-05, + "loss": 0.0161, + "step": 36394 + }, + { + "epoch": 0.8531561150278075, + "grad_norm": 0.11233220994472504, + "learning_rate": 1.0457708745251527e-05, + "loss": 0.0141, + "step": 36395 + }, + { + "epoch": 0.8531795566026126, + "grad_norm": 0.32904374599456787, + "learning_rate": 1.0454429452548608e-05, + "loss": 0.0278, + "step": 36396 + }, + { + "epoch": 0.8532029981774175, + "grad_norm": 0.23532670736312866, + "learning_rate": 1.0451150645722463e-05, + "loss": 0.0353, + "step": 36397 + }, + { + "epoch": 0.8532264397522226, + "grad_norm": 0.26438435912132263, + "learning_rate": 1.0447872324790898e-05, + "loss": 0.0361, + "step": 36398 + }, + { + "epoch": 0.8532498813270275, + "grad_norm": 0.08899998664855957, + "learning_rate": 1.0444594489771698e-05, + "loss": 0.0147, + "step": 36399 + }, + { + "epoch": 0.8532733229018326, + "grad_norm": 0.10740305483341217, + "learning_rate": 1.044131714068265e-05, + "loss": 0.0113, + "step": 36400 + }, + { + "epoch": 0.8532967644766375, + "grad_norm": 0.3306109607219696, + "learning_rate": 1.0438040277541506e-05, + "loss": 0.0438, + "step": 36401 + }, + { + "epoch": 0.8533202060514425, + "grad_norm": 0.4699128270149231, + "learning_rate": 1.0434763900366096e-05, + "loss": 0.0829, + "step": 36402 + }, + { + "epoch": 0.8533436476262475, + "grad_norm": 0.2701602578163147, + "learning_rate": 1.0431488009174174e-05, + "loss": 0.0438, + "step": 36403 + }, + { + "epoch": 0.8533670892010525, + "grad_norm": 0.18386155366897583, + "learning_rate": 1.04282126039835e-05, + "loss": 0.0264, + "step": 36404 + }, + { + "epoch": 0.8533905307758575, + "grad_norm": 0.38415059447288513, + "learning_rate": 1.0424937684811876e-05, + "loss": 0.0621, + "step": 36405 + }, + { + "epoch": 0.8534139723506625, + "grad_norm": 0.43322983384132385, + "learning_rate": 1.0421663251677071e-05, + "loss": 0.0832, + "step": 36406 + }, + { + "epoch": 0.8534374139254675, + "grad_norm": 0.5533907413482666, + "learning_rate": 1.0418389304596832e-05, + "loss": 0.1435, + "step": 36407 + }, + { + "epoch": 0.8534608555002725, + "grad_norm": 0.2995971143245697, + "learning_rate": 1.0415115843588907e-05, + "loss": 0.0597, + "step": 36408 + }, + { + "epoch": 0.8534842970750776, + "grad_norm": 0.44945019483566284, + "learning_rate": 1.0411842868671106e-05, + "loss": 0.5423, + "step": 36409 + }, + { + "epoch": 0.8535077386498825, + "grad_norm": 0.4740440249443054, + "learning_rate": 1.0408570379861172e-05, + "loss": 0.4409, + "step": 36410 + }, + { + "epoch": 0.8535311802246875, + "grad_norm": 0.12487596273422241, + "learning_rate": 1.0405298377176842e-05, + "loss": 0.023, + "step": 36411 + }, + { + "epoch": 0.8535546217994925, + "grad_norm": 0.3562884032726288, + "learning_rate": 1.0402026860635894e-05, + "loss": 0.3031, + "step": 36412 + }, + { + "epoch": 0.8535780633742975, + "grad_norm": 0.5131003856658936, + "learning_rate": 1.0398755830256035e-05, + "loss": 0.0457, + "step": 36413 + }, + { + "epoch": 0.8536015049491025, + "grad_norm": 0.26754510402679443, + "learning_rate": 1.0395485286055074e-05, + "loss": 0.337, + "step": 36414 + }, + { + "epoch": 0.8536249465239075, + "grad_norm": 0.24558724462985992, + "learning_rate": 1.0392215228050727e-05, + "loss": 0.0491, + "step": 36415 + }, + { + "epoch": 0.8536483880987125, + "grad_norm": 0.24560970067977905, + "learning_rate": 1.038894565626074e-05, + "loss": 0.0326, + "step": 36416 + }, + { + "epoch": 0.8536718296735175, + "grad_norm": 0.2141740620136261, + "learning_rate": 1.038567657070284e-05, + "loss": 0.0238, + "step": 36417 + }, + { + "epoch": 0.8536952712483225, + "grad_norm": 0.19187502562999725, + "learning_rate": 1.0382407971394792e-05, + "loss": 0.027, + "step": 36418 + }, + { + "epoch": 0.8537187128231275, + "grad_norm": 0.37980249524116516, + "learning_rate": 1.0379139858354326e-05, + "loss": 0.0497, + "step": 36419 + }, + { + "epoch": 0.8537421543979324, + "grad_norm": 0.3975622057914734, + "learning_rate": 1.0375872231599137e-05, + "loss": 0.0634, + "step": 36420 + }, + { + "epoch": 0.8537655959727375, + "grad_norm": 0.4253259301185608, + "learning_rate": 1.0372605091147003e-05, + "loss": 0.0503, + "step": 36421 + }, + { + "epoch": 0.8537890375475424, + "grad_norm": 0.505409836769104, + "learning_rate": 1.0369338437015641e-05, + "loss": 0.6723, + "step": 36422 + }, + { + "epoch": 0.8538124791223475, + "grad_norm": 0.4121423065662384, + "learning_rate": 1.036607226922277e-05, + "loss": 0.0488, + "step": 36423 + }, + { + "epoch": 0.8538359206971524, + "grad_norm": 0.4102713465690613, + "learning_rate": 1.0362806587786111e-05, + "loss": 0.0241, + "step": 36424 + }, + { + "epoch": 0.8538593622719575, + "grad_norm": 0.1550186425447464, + "learning_rate": 1.0359541392723382e-05, + "loss": 0.0287, + "step": 36425 + }, + { + "epoch": 0.8538828038467624, + "grad_norm": 0.53496253490448, + "learning_rate": 1.0356276684052279e-05, + "loss": 0.5407, + "step": 36426 + }, + { + "epoch": 0.8539062454215675, + "grad_norm": 0.7654925584793091, + "learning_rate": 1.0353012461790557e-05, + "loss": 0.1661, + "step": 36427 + }, + { + "epoch": 0.8539296869963724, + "grad_norm": 0.6742231845855713, + "learning_rate": 1.0349748725955921e-05, + "loss": 0.412, + "step": 36428 + }, + { + "epoch": 0.8539531285711774, + "grad_norm": 0.55040043592453, + "learning_rate": 1.0346485476566038e-05, + "loss": 0.0913, + "step": 36429 + }, + { + "epoch": 0.8539765701459824, + "grad_norm": 0.46371254324913025, + "learning_rate": 1.0343222713638668e-05, + "loss": 0.0648, + "step": 36430 + }, + { + "epoch": 0.8540000117207874, + "grad_norm": 0.13196122646331787, + "learning_rate": 1.0339960437191499e-05, + "loss": 0.0268, + "step": 36431 + }, + { + "epoch": 0.8540234532955924, + "grad_norm": 0.23086614906787872, + "learning_rate": 1.0336698647242227e-05, + "loss": 0.0245, + "step": 36432 + }, + { + "epoch": 0.8540468948703974, + "grad_norm": 0.5700145363807678, + "learning_rate": 1.0333437343808516e-05, + "loss": 0.0577, + "step": 36433 + }, + { + "epoch": 0.8540703364452024, + "grad_norm": 0.358313649892807, + "learning_rate": 1.0330176526908132e-05, + "loss": 0.0765, + "step": 36434 + }, + { + "epoch": 0.8540937780200074, + "grad_norm": 0.46854060888290405, + "learning_rate": 1.0326916196558711e-05, + "loss": 0.095, + "step": 36435 + }, + { + "epoch": 0.8541172195948123, + "grad_norm": 0.13454122841358185, + "learning_rate": 1.0323656352777978e-05, + "loss": 0.012, + "step": 36436 + }, + { + "epoch": 0.8541406611696174, + "grad_norm": 0.5772005319595337, + "learning_rate": 1.0320396995583603e-05, + "loss": 0.9406, + "step": 36437 + }, + { + "epoch": 0.8541641027444223, + "grad_norm": 0.4980190098285675, + "learning_rate": 1.0317138124993286e-05, + "loss": 0.6787, + "step": 36438 + }, + { + "epoch": 0.8541875443192274, + "grad_norm": 0.5529283881187439, + "learning_rate": 1.0313879741024669e-05, + "loss": 0.0739, + "step": 36439 + }, + { + "epoch": 0.8542109858940323, + "grad_norm": 0.09121263772249222, + "learning_rate": 1.0310621843695478e-05, + "loss": 0.0138, + "step": 36440 + }, + { + "epoch": 0.8542344274688374, + "grad_norm": 0.5441542267799377, + "learning_rate": 1.0307364433023358e-05, + "loss": 0.0469, + "step": 36441 + }, + { + "epoch": 0.8542578690436423, + "grad_norm": 0.18813878297805786, + "learning_rate": 1.0304107509026028e-05, + "loss": 0.0318, + "step": 36442 + }, + { + "epoch": 0.8542813106184474, + "grad_norm": 0.11657264083623886, + "learning_rate": 1.030085107172113e-05, + "loss": 0.0196, + "step": 36443 + }, + { + "epoch": 0.8543047521932523, + "grad_norm": 0.4648337960243225, + "learning_rate": 1.0297595121126336e-05, + "loss": 0.0603, + "step": 36444 + }, + { + "epoch": 0.8543281937680574, + "grad_norm": 0.23084941506385803, + "learning_rate": 1.029433965725931e-05, + "loss": 0.0437, + "step": 36445 + }, + { + "epoch": 0.8543516353428623, + "grad_norm": 0.49761101603507996, + "learning_rate": 1.0291084680137697e-05, + "loss": 0.0976, + "step": 36446 + }, + { + "epoch": 0.8543750769176673, + "grad_norm": 0.30888548493385315, + "learning_rate": 1.0287830189779201e-05, + "loss": 0.0339, + "step": 36447 + }, + { + "epoch": 0.8543985184924723, + "grad_norm": 0.6354297995567322, + "learning_rate": 1.0284576186201466e-05, + "loss": 0.5368, + "step": 36448 + }, + { + "epoch": 0.8544219600672773, + "grad_norm": 0.2777194678783417, + "learning_rate": 1.0281322669422144e-05, + "loss": 0.0354, + "step": 36449 + }, + { + "epoch": 0.8544454016420823, + "grad_norm": 0.9662895202636719, + "learning_rate": 1.0278069639458876e-05, + "loss": 0.1427, + "step": 36450 + }, + { + "epoch": 0.8544688432168873, + "grad_norm": 0.5942108035087585, + "learning_rate": 1.0274817096329315e-05, + "loss": 0.1146, + "step": 36451 + }, + { + "epoch": 0.8544922847916923, + "grad_norm": 0.23510374128818512, + "learning_rate": 1.0271565040051146e-05, + "loss": 0.0527, + "step": 36452 + }, + { + "epoch": 0.8545157263664973, + "grad_norm": 0.5069189667701721, + "learning_rate": 1.0268313470641955e-05, + "loss": 0.0679, + "step": 36453 + }, + { + "epoch": 0.8545391679413022, + "grad_norm": 0.22963756322860718, + "learning_rate": 1.026506238811945e-05, + "loss": 0.0506, + "step": 36454 + }, + { + "epoch": 0.8545626095161073, + "grad_norm": 0.5079128742218018, + "learning_rate": 1.0261811792501231e-05, + "loss": 0.0549, + "step": 36455 + }, + { + "epoch": 0.8545860510909122, + "grad_norm": 0.39001256227493286, + "learning_rate": 1.0258561683804945e-05, + "loss": 0.0771, + "step": 36456 + }, + { + "epoch": 0.8546094926657173, + "grad_norm": 0.6398895382881165, + "learning_rate": 1.0255312062048227e-05, + "loss": 0.6779, + "step": 36457 + }, + { + "epoch": 0.8546329342405222, + "grad_norm": 0.17275771498680115, + "learning_rate": 1.0252062927248684e-05, + "loss": 0.0293, + "step": 36458 + }, + { + "epoch": 0.8546563758153273, + "grad_norm": 0.4896237552165985, + "learning_rate": 1.0248814279424002e-05, + "loss": 0.0916, + "step": 36459 + }, + { + "epoch": 0.8546798173901323, + "grad_norm": 0.43043333292007446, + "learning_rate": 1.0245566118591776e-05, + "loss": 0.0751, + "step": 36460 + }, + { + "epoch": 0.8547032589649373, + "grad_norm": 0.18237806856632233, + "learning_rate": 1.0242318444769627e-05, + "loss": 0.0218, + "step": 36461 + }, + { + "epoch": 0.8547267005397423, + "grad_norm": 0.1870604008436203, + "learning_rate": 1.0239071257975186e-05, + "loss": 0.0378, + "step": 36462 + }, + { + "epoch": 0.8547501421145473, + "grad_norm": 0.35286736488342285, + "learning_rate": 1.023582455822606e-05, + "loss": 0.0921, + "step": 36463 + }, + { + "epoch": 0.8547735836893523, + "grad_norm": 0.8308972716331482, + "learning_rate": 1.0232578345539856e-05, + "loss": 0.1427, + "step": 36464 + }, + { + "epoch": 0.8547970252641572, + "grad_norm": 0.3422626554965973, + "learning_rate": 1.0229332619934218e-05, + "loss": 0.0449, + "step": 36465 + }, + { + "epoch": 0.8548204668389623, + "grad_norm": 0.140036940574646, + "learning_rate": 1.0226087381426752e-05, + "loss": 0.0248, + "step": 36466 + }, + { + "epoch": 0.8548439084137672, + "grad_norm": 0.74634850025177, + "learning_rate": 1.0222842630035068e-05, + "loss": 0.136, + "step": 36467 + }, + { + "epoch": 0.8548673499885723, + "grad_norm": 0.3984244465827942, + "learning_rate": 1.021959836577676e-05, + "loss": 0.0654, + "step": 36468 + }, + { + "epoch": 0.8548907915633772, + "grad_norm": 0.38316160440444946, + "learning_rate": 1.021635458866943e-05, + "loss": 0.0289, + "step": 36469 + }, + { + "epoch": 0.8549142331381823, + "grad_norm": 0.42761439085006714, + "learning_rate": 1.0213111298730693e-05, + "loss": 0.0561, + "step": 36470 + }, + { + "epoch": 0.8549376747129872, + "grad_norm": 0.23043696582317352, + "learning_rate": 1.0209868495978114e-05, + "loss": 0.0419, + "step": 36471 + }, + { + "epoch": 0.8549611162877923, + "grad_norm": 0.39967915415763855, + "learning_rate": 1.0206626180429347e-05, + "loss": 0.053, + "step": 36472 + }, + { + "epoch": 0.8549845578625972, + "grad_norm": 0.34238389134407043, + "learning_rate": 1.020338435210193e-05, + "loss": 0.0357, + "step": 36473 + }, + { + "epoch": 0.8550079994374022, + "grad_norm": 0.40501832962036133, + "learning_rate": 1.0200143011013485e-05, + "loss": 0.0886, + "step": 36474 + }, + { + "epoch": 0.8550314410122072, + "grad_norm": 0.4439077079296112, + "learning_rate": 1.0196902157181588e-05, + "loss": 0.0329, + "step": 36475 + }, + { + "epoch": 0.8550548825870122, + "grad_norm": 0.5342549681663513, + "learning_rate": 1.0193661790623831e-05, + "loss": 0.4657, + "step": 36476 + }, + { + "epoch": 0.8550783241618172, + "grad_norm": 0.36739301681518555, + "learning_rate": 1.0190421911357762e-05, + "loss": 0.0691, + "step": 36477 + }, + { + "epoch": 0.8551017657366222, + "grad_norm": 0.3929573893547058, + "learning_rate": 1.0187182519400995e-05, + "loss": 0.0769, + "step": 36478 + }, + { + "epoch": 0.8551252073114272, + "grad_norm": 0.47353270649909973, + "learning_rate": 1.018394361477112e-05, + "loss": 0.0727, + "step": 36479 + }, + { + "epoch": 0.8551486488862322, + "grad_norm": 0.21610885858535767, + "learning_rate": 1.0180705197485685e-05, + "loss": 0.0261, + "step": 36480 + }, + { + "epoch": 0.8551720904610371, + "grad_norm": 0.13346412777900696, + "learning_rate": 1.017746726756228e-05, + "loss": 0.0182, + "step": 36481 + }, + { + "epoch": 0.8551955320358422, + "grad_norm": 0.494833379983902, + "learning_rate": 1.0174229825018455e-05, + "loss": 0.0439, + "step": 36482 + }, + { + "epoch": 0.8552189736106471, + "grad_norm": 0.1202179417014122, + "learning_rate": 1.0170992869871787e-05, + "loss": 0.0255, + "step": 36483 + }, + { + "epoch": 0.8552424151854522, + "grad_norm": 0.4841194152832031, + "learning_rate": 1.0167756402139816e-05, + "loss": 0.104, + "step": 36484 + }, + { + "epoch": 0.8552658567602571, + "grad_norm": 0.20821143686771393, + "learning_rate": 1.016452042184014e-05, + "loss": 0.0458, + "step": 36485 + }, + { + "epoch": 0.8552892983350622, + "grad_norm": 0.10687298327684402, + "learning_rate": 1.0161284928990312e-05, + "loss": 0.0136, + "step": 36486 + }, + { + "epoch": 0.8553127399098671, + "grad_norm": 0.33218592405319214, + "learning_rate": 1.015804992360787e-05, + "loss": 0.0743, + "step": 36487 + }, + { + "epoch": 0.8553361814846722, + "grad_norm": 0.3523058593273163, + "learning_rate": 1.0154815405710383e-05, + "loss": 0.071, + "step": 36488 + }, + { + "epoch": 0.8553596230594771, + "grad_norm": 0.3911038339138031, + "learning_rate": 1.0151581375315367e-05, + "loss": 0.0654, + "step": 36489 + }, + { + "epoch": 0.8553830646342822, + "grad_norm": 0.07056517899036407, + "learning_rate": 1.0148347832440419e-05, + "loss": 0.0066, + "step": 36490 + }, + { + "epoch": 0.8554065062090871, + "grad_norm": 0.20328448712825775, + "learning_rate": 1.014511477710305e-05, + "loss": 0.034, + "step": 36491 + }, + { + "epoch": 0.8554299477838921, + "grad_norm": 0.14477013051509857, + "learning_rate": 1.014188220932083e-05, + "loss": 0.0233, + "step": 36492 + }, + { + "epoch": 0.8554533893586971, + "grad_norm": 0.5602903962135315, + "learning_rate": 1.0138650129111272e-05, + "loss": 0.0972, + "step": 36493 + }, + { + "epoch": 0.8554768309335021, + "grad_norm": 0.6516924500465393, + "learning_rate": 1.0135418536491936e-05, + "loss": 0.0844, + "step": 36494 + }, + { + "epoch": 0.8555002725083071, + "grad_norm": 0.5067784786224365, + "learning_rate": 1.0132187431480344e-05, + "loss": 0.0878, + "step": 36495 + }, + { + "epoch": 0.8555237140831121, + "grad_norm": 0.47116678953170776, + "learning_rate": 1.0128956814094015e-05, + "loss": 0.6331, + "step": 36496 + }, + { + "epoch": 0.8555471556579171, + "grad_norm": 0.16508999466896057, + "learning_rate": 1.01257266843505e-05, + "loss": 0.0361, + "step": 36497 + }, + { + "epoch": 0.8555705972327221, + "grad_norm": 0.5330265760421753, + "learning_rate": 1.0122497042267332e-05, + "loss": 0.0789, + "step": 36498 + }, + { + "epoch": 0.855594038807527, + "grad_norm": 0.05436963215470314, + "learning_rate": 1.0119267887862017e-05, + "loss": 0.0077, + "step": 36499 + }, + { + "epoch": 0.8556174803823321, + "grad_norm": 0.24738344550132751, + "learning_rate": 1.0116039221152086e-05, + "loss": 0.0264, + "step": 36500 + }, + { + "epoch": 0.855640921957137, + "grad_norm": 0.15665794909000397, + "learning_rate": 1.0112811042155045e-05, + "loss": 0.0348, + "step": 36501 + }, + { + "epoch": 0.8556643635319421, + "grad_norm": 0.08058223873376846, + "learning_rate": 1.0109583350888407e-05, + "loss": 0.0093, + "step": 36502 + }, + { + "epoch": 0.855687805106747, + "grad_norm": 0.5142582654953003, + "learning_rate": 1.0106356147369689e-05, + "loss": 0.0987, + "step": 36503 + }, + { + "epoch": 0.8557112466815521, + "grad_norm": 0.24827447533607483, + "learning_rate": 1.0103129431616442e-05, + "loss": 0.0337, + "step": 36504 + }, + { + "epoch": 0.855734688256357, + "grad_norm": 0.2992965877056122, + "learning_rate": 1.0099903203646132e-05, + "loss": 0.046, + "step": 36505 + }, + { + "epoch": 0.8557581298311621, + "grad_norm": 0.09737880527973175, + "learning_rate": 1.0096677463476278e-05, + "loss": 0.008, + "step": 36506 + }, + { + "epoch": 0.855781571405967, + "grad_norm": 0.34426775574684143, + "learning_rate": 1.0093452211124376e-05, + "loss": 0.0537, + "step": 36507 + }, + { + "epoch": 0.855805012980772, + "grad_norm": 0.5058679580688477, + "learning_rate": 1.0090227446607936e-05, + "loss": 0.0851, + "step": 36508 + }, + { + "epoch": 0.855828454555577, + "grad_norm": 0.5214501619338989, + "learning_rate": 1.0087003169944432e-05, + "loss": 0.0487, + "step": 36509 + }, + { + "epoch": 0.855851896130382, + "grad_norm": 0.301652729511261, + "learning_rate": 1.0083779381151393e-05, + "loss": 0.0586, + "step": 36510 + }, + { + "epoch": 0.8558753377051871, + "grad_norm": 0.3849904239177704, + "learning_rate": 1.0080556080246295e-05, + "loss": 0.0522, + "step": 36511 + }, + { + "epoch": 0.855898779279992, + "grad_norm": 0.32709595561027527, + "learning_rate": 1.0077333267246624e-05, + "loss": 0.0347, + "step": 36512 + }, + { + "epoch": 0.8559222208547971, + "grad_norm": 0.4620124399662018, + "learning_rate": 1.0074110942169878e-05, + "loss": 0.0807, + "step": 36513 + }, + { + "epoch": 0.855945662429602, + "grad_norm": 0.0758671835064888, + "learning_rate": 1.007088910503351e-05, + "loss": 0.0156, + "step": 36514 + }, + { + "epoch": 0.8559691040044071, + "grad_norm": 0.08376531302928925, + "learning_rate": 1.0067667755855048e-05, + "loss": 0.0044, + "step": 36515 + }, + { + "epoch": 0.855992545579212, + "grad_norm": 0.19140684604644775, + "learning_rate": 1.0064446894651936e-05, + "loss": 0.0286, + "step": 36516 + }, + { + "epoch": 0.856015987154017, + "grad_norm": 0.4589374363422394, + "learning_rate": 1.006122652144168e-05, + "loss": 0.1073, + "step": 36517 + }, + { + "epoch": 0.856039428728822, + "grad_norm": 0.22582510113716125, + "learning_rate": 1.0058006636241735e-05, + "loss": 0.0676, + "step": 36518 + }, + { + "epoch": 0.856062870303627, + "grad_norm": 0.6506443023681641, + "learning_rate": 1.0054787239069574e-05, + "loss": 0.1168, + "step": 36519 + }, + { + "epoch": 0.856086311878432, + "grad_norm": 0.24864724278450012, + "learning_rate": 1.0051568329942662e-05, + "loss": 0.0627, + "step": 36520 + }, + { + "epoch": 0.856109753453237, + "grad_norm": 0.39852213859558105, + "learning_rate": 1.0048349908878484e-05, + "loss": 0.0423, + "step": 36521 + }, + { + "epoch": 0.856133195028042, + "grad_norm": 0.5896488428115845, + "learning_rate": 1.0045131975894451e-05, + "loss": 0.0991, + "step": 36522 + }, + { + "epoch": 0.856156636602847, + "grad_norm": 0.3392546474933624, + "learning_rate": 1.004191453100809e-05, + "loss": 0.0396, + "step": 36523 + }, + { + "epoch": 0.856180078177652, + "grad_norm": 0.2055804282426834, + "learning_rate": 1.0038697574236833e-05, + "loss": 0.0459, + "step": 36524 + }, + { + "epoch": 0.856203519752457, + "grad_norm": 0.09520453214645386, + "learning_rate": 1.0035481105598133e-05, + "loss": 0.011, + "step": 36525 + }, + { + "epoch": 0.856226961327262, + "grad_norm": 0.336199015378952, + "learning_rate": 1.0032265125109419e-05, + "loss": 0.078, + "step": 36526 + }, + { + "epoch": 0.856250402902067, + "grad_norm": 0.289618581533432, + "learning_rate": 1.0029049632788178e-05, + "loss": 0.0514, + "step": 36527 + }, + { + "epoch": 0.8562738444768719, + "grad_norm": 0.312608927488327, + "learning_rate": 1.0025834628651854e-05, + "loss": 0.3672, + "step": 36528 + }, + { + "epoch": 0.856297286051677, + "grad_norm": 0.38262227177619934, + "learning_rate": 1.0022620112717862e-05, + "loss": 0.062, + "step": 36529 + }, + { + "epoch": 0.8563207276264819, + "grad_norm": 0.28001269698143005, + "learning_rate": 1.0019406085003669e-05, + "loss": 0.039, + "step": 36530 + }, + { + "epoch": 0.856344169201287, + "grad_norm": 0.23563870787620544, + "learning_rate": 1.0016192545526725e-05, + "loss": 0.0294, + "step": 36531 + }, + { + "epoch": 0.8563676107760919, + "grad_norm": 0.4613690972328186, + "learning_rate": 1.0012979494304441e-05, + "loss": 0.0855, + "step": 36532 + }, + { + "epoch": 0.856391052350897, + "grad_norm": 0.3168259859085083, + "learning_rate": 1.0009766931354258e-05, + "loss": 0.0539, + "step": 36533 + }, + { + "epoch": 0.8564144939257019, + "grad_norm": 0.3159045875072479, + "learning_rate": 1.0006554856693594e-05, + "loss": 0.04, + "step": 36534 + }, + { + "epoch": 0.856437935500507, + "grad_norm": 0.5514063239097595, + "learning_rate": 1.0003343270339915e-05, + "loss": 0.5262, + "step": 36535 + }, + { + "epoch": 0.8564613770753119, + "grad_norm": 0.8666208982467651, + "learning_rate": 1.0000132172310629e-05, + "loss": 0.0899, + "step": 36536 + }, + { + "epoch": 0.8564848186501169, + "grad_norm": 0.3732958734035492, + "learning_rate": 9.996921562623151e-06, + "loss": 0.0215, + "step": 36537 + }, + { + "epoch": 0.8565082602249219, + "grad_norm": 0.9202884435653687, + "learning_rate": 9.993711441294884e-06, + "loss": 0.1567, + "step": 36538 + }, + { + "epoch": 0.8565317017997269, + "grad_norm": 0.7441366314888, + "learning_rate": 9.9905018083433e-06, + "loss": 0.2015, + "step": 36539 + }, + { + "epoch": 0.8565551433745319, + "grad_norm": 0.3407461643218994, + "learning_rate": 9.987292663785775e-06, + "loss": 0.0437, + "step": 36540 + }, + { + "epoch": 0.8565785849493369, + "grad_norm": 0.4044312834739685, + "learning_rate": 9.984084007639716e-06, + "loss": 0.0582, + "step": 36541 + }, + { + "epoch": 0.8566020265241419, + "grad_norm": 0.332502156496048, + "learning_rate": 9.980875839922566e-06, + "loss": 0.0621, + "step": 36542 + }, + { + "epoch": 0.8566254680989469, + "grad_norm": 0.4519656002521515, + "learning_rate": 9.97766816065171e-06, + "loss": 0.0744, + "step": 36543 + }, + { + "epoch": 0.8566489096737518, + "grad_norm": 0.24656663835048676, + "learning_rate": 9.974460969844567e-06, + "loss": 0.0404, + "step": 36544 + }, + { + "epoch": 0.8566723512485569, + "grad_norm": 0.16429558396339417, + "learning_rate": 9.971254267518526e-06, + "loss": 0.0147, + "step": 36545 + }, + { + "epoch": 0.8566957928233618, + "grad_norm": 0.33366939425468445, + "learning_rate": 9.968048053690992e-06, + "loss": 0.0275, + "step": 36546 + }, + { + "epoch": 0.8567192343981669, + "grad_norm": 0.4375510811805725, + "learning_rate": 9.964842328379332e-06, + "loss": 0.0785, + "step": 36547 + }, + { + "epoch": 0.8567426759729718, + "grad_norm": 0.34178870916366577, + "learning_rate": 9.961637091600995e-06, + "loss": 0.0524, + "step": 36548 + }, + { + "epoch": 0.8567661175477769, + "grad_norm": 0.4198300242424011, + "learning_rate": 9.958432343373359e-06, + "loss": 0.3522, + "step": 36549 + }, + { + "epoch": 0.8567895591225818, + "grad_norm": 0.30640897154808044, + "learning_rate": 9.955228083713763e-06, + "loss": 0.0433, + "step": 36550 + }, + { + "epoch": 0.8568130006973869, + "grad_norm": 0.49205973744392395, + "learning_rate": 9.952024312639662e-06, + "loss": 0.1059, + "step": 36551 + }, + { + "epoch": 0.8568364422721918, + "grad_norm": 0.5151247978210449, + "learning_rate": 9.948821030168398e-06, + "loss": 0.1276, + "step": 36552 + }, + { + "epoch": 0.8568598838469969, + "grad_norm": 0.2722657322883606, + "learning_rate": 9.945618236317377e-06, + "loss": 0.035, + "step": 36553 + }, + { + "epoch": 0.8568833254218018, + "grad_norm": 0.4745585024356842, + "learning_rate": 9.942415931103933e-06, + "loss": 0.0542, + "step": 36554 + }, + { + "epoch": 0.8569067669966068, + "grad_norm": 0.5858834385871887, + "learning_rate": 9.939214114545503e-06, + "loss": 0.5544, + "step": 36555 + }, + { + "epoch": 0.8569302085714118, + "grad_norm": 0.4473501741886139, + "learning_rate": 9.936012786659421e-06, + "loss": 0.0933, + "step": 36556 + }, + { + "epoch": 0.8569536501462168, + "grad_norm": 0.3992219865322113, + "learning_rate": 9.932811947463061e-06, + "loss": 0.074, + "step": 36557 + }, + { + "epoch": 0.8569770917210218, + "grad_norm": 0.4649699032306671, + "learning_rate": 9.929611596973809e-06, + "loss": 0.058, + "step": 36558 + }, + { + "epoch": 0.8570005332958268, + "grad_norm": 0.1765083372592926, + "learning_rate": 9.926411735209006e-06, + "loss": 0.0117, + "step": 36559 + }, + { + "epoch": 0.8570239748706318, + "grad_norm": 0.29373472929000854, + "learning_rate": 9.923212362186007e-06, + "loss": 0.0275, + "step": 36560 + }, + { + "epoch": 0.8570474164454368, + "grad_norm": 0.5871338844299316, + "learning_rate": 9.920013477922219e-06, + "loss": 0.1157, + "step": 36561 + }, + { + "epoch": 0.8570708580202419, + "grad_norm": 0.09493423253297806, + "learning_rate": 9.91681508243495e-06, + "loss": 0.0179, + "step": 36562 + }, + { + "epoch": 0.8570942995950468, + "grad_norm": 0.6969418525695801, + "learning_rate": 9.913617175741585e-06, + "loss": 0.1419, + "step": 36563 + }, + { + "epoch": 0.8571177411698518, + "grad_norm": 0.45712757110595703, + "learning_rate": 9.910419757859468e-06, + "loss": 0.068, + "step": 36564 + }, + { + "epoch": 0.8571411827446568, + "grad_norm": 0.1323109120130539, + "learning_rate": 9.907222828805951e-06, + "loss": 0.0192, + "step": 36565 + }, + { + "epoch": 0.8571646243194618, + "grad_norm": 0.5087748169898987, + "learning_rate": 9.904026388598375e-06, + "loss": 0.1748, + "step": 36566 + }, + { + "epoch": 0.8571880658942668, + "grad_norm": 0.48364493250846863, + "learning_rate": 9.900830437254071e-06, + "loss": 0.0959, + "step": 36567 + }, + { + "epoch": 0.8572115074690718, + "grad_norm": 0.18369914591312408, + "learning_rate": 9.897634974790404e-06, + "loss": 0.1101, + "step": 36568 + }, + { + "epoch": 0.8572349490438768, + "grad_norm": 0.33186426758766174, + "learning_rate": 9.894440001224714e-06, + "loss": 0.0522, + "step": 36569 + }, + { + "epoch": 0.8572583906186818, + "grad_norm": 0.4600661098957062, + "learning_rate": 9.891245516574321e-06, + "loss": 0.051, + "step": 36570 + }, + { + "epoch": 0.8572818321934867, + "grad_norm": 0.5437189936637878, + "learning_rate": 9.888051520856557e-06, + "loss": 0.4999, + "step": 36571 + }, + { + "epoch": 0.8573052737682918, + "grad_norm": 0.09740687161684036, + "learning_rate": 9.884858014088749e-06, + "loss": 0.0158, + "step": 36572 + }, + { + "epoch": 0.8573287153430967, + "grad_norm": 0.40411078929901123, + "learning_rate": 9.881664996288254e-06, + "loss": 0.0899, + "step": 36573 + }, + { + "epoch": 0.8573521569179018, + "grad_norm": 0.15973573923110962, + "learning_rate": 9.878472467472366e-06, + "loss": 0.0203, + "step": 36574 + }, + { + "epoch": 0.8573755984927067, + "grad_norm": 0.09364401549100876, + "learning_rate": 9.875280427658418e-06, + "loss": 0.0095, + "step": 36575 + }, + { + "epoch": 0.8573990400675118, + "grad_norm": 0.22655074298381805, + "learning_rate": 9.872088876863739e-06, + "loss": 0.0369, + "step": 36576 + }, + { + "epoch": 0.8574224816423167, + "grad_norm": 0.471762090921402, + "learning_rate": 9.86889781510565e-06, + "loss": 0.0943, + "step": 36577 + }, + { + "epoch": 0.8574459232171218, + "grad_norm": 0.3775453269481659, + "learning_rate": 9.865707242401445e-06, + "loss": 0.0565, + "step": 36578 + }, + { + "epoch": 0.8574693647919267, + "grad_norm": 0.4505917727947235, + "learning_rate": 9.862517158768436e-06, + "loss": 0.0553, + "step": 36579 + }, + { + "epoch": 0.8574928063667318, + "grad_norm": 0.23813249170780182, + "learning_rate": 9.859327564223963e-06, + "loss": 0.0137, + "step": 36580 + }, + { + "epoch": 0.8575162479415367, + "grad_norm": 0.8930478096008301, + "learning_rate": 9.856138458785314e-06, + "loss": 0.0822, + "step": 36581 + }, + { + "epoch": 0.8575396895163417, + "grad_norm": 1.8791629076004028, + "learning_rate": 9.852949842469783e-06, + "loss": 0.0671, + "step": 36582 + }, + { + "epoch": 0.8575631310911467, + "grad_norm": 1.3032187223434448, + "learning_rate": 9.849761715294692e-06, + "loss": 0.105, + "step": 36583 + }, + { + "epoch": 0.8575865726659517, + "grad_norm": 0.5106277465820312, + "learning_rate": 9.846574077277315e-06, + "loss": 0.625, + "step": 36584 + }, + { + "epoch": 0.8576100142407567, + "grad_norm": 0.4243975281715393, + "learning_rate": 9.84338692843495e-06, + "loss": 0.0984, + "step": 36585 + }, + { + "epoch": 0.8576334558155617, + "grad_norm": 0.4430501163005829, + "learning_rate": 9.840200268784928e-06, + "loss": 0.033, + "step": 36586 + }, + { + "epoch": 0.8576568973903667, + "grad_norm": 0.07251568138599396, + "learning_rate": 9.837014098344488e-06, + "loss": 0.0088, + "step": 36587 + }, + { + "epoch": 0.8576803389651717, + "grad_norm": 0.6683980822563171, + "learning_rate": 9.833828417130964e-06, + "loss": 0.0699, + "step": 36588 + }, + { + "epoch": 0.8577037805399766, + "grad_norm": 0.28474920988082886, + "learning_rate": 9.830643225161628e-06, + "loss": 0.034, + "step": 36589 + }, + { + "epoch": 0.8577272221147817, + "grad_norm": 0.5181770324707031, + "learning_rate": 9.827458522453748e-06, + "loss": 0.1047, + "step": 36590 + }, + { + "epoch": 0.8577506636895866, + "grad_norm": 0.2907697558403015, + "learning_rate": 9.824274309024628e-06, + "loss": 0.046, + "step": 36591 + }, + { + "epoch": 0.8577741052643917, + "grad_norm": 0.3820562958717346, + "learning_rate": 9.821090584891501e-06, + "loss": 0.5488, + "step": 36592 + }, + { + "epoch": 0.8577975468391966, + "grad_norm": 0.596641480922699, + "learning_rate": 9.817907350071687e-06, + "loss": 0.1024, + "step": 36593 + }, + { + "epoch": 0.8578209884140017, + "grad_norm": 0.5950325727462769, + "learning_rate": 9.814724604582448e-06, + "loss": 0.5777, + "step": 36594 + }, + { + "epoch": 0.8578444299888066, + "grad_norm": 0.3957415521144867, + "learning_rate": 9.811542348441049e-06, + "loss": 0.0559, + "step": 36595 + }, + { + "epoch": 0.8578678715636117, + "grad_norm": 0.26668983697891235, + "learning_rate": 9.808360581664767e-06, + "loss": 0.0396, + "step": 36596 + }, + { + "epoch": 0.8578913131384166, + "grad_norm": 0.2944563031196594, + "learning_rate": 9.80517930427084e-06, + "loss": 0.0257, + "step": 36597 + }, + { + "epoch": 0.8579147547132217, + "grad_norm": 0.201071098446846, + "learning_rate": 9.801998516276534e-06, + "loss": 0.0294, + "step": 36598 + }, + { + "epoch": 0.8579381962880266, + "grad_norm": 0.5055160522460938, + "learning_rate": 9.798818217699124e-06, + "loss": 0.0561, + "step": 36599 + }, + { + "epoch": 0.8579616378628316, + "grad_norm": 0.6378568410873413, + "learning_rate": 9.795638408555885e-06, + "loss": 0.0977, + "step": 36600 + }, + { + "epoch": 0.8579850794376366, + "grad_norm": 0.2640118896961212, + "learning_rate": 9.792459088864037e-06, + "loss": 0.0262, + "step": 36601 + }, + { + "epoch": 0.8580085210124416, + "grad_norm": 0.48325803875923157, + "learning_rate": 9.789280258640853e-06, + "loss": 0.5482, + "step": 36602 + }, + { + "epoch": 0.8580319625872466, + "grad_norm": 0.7108412981033325, + "learning_rate": 9.786101917903556e-06, + "loss": 0.072, + "step": 36603 + }, + { + "epoch": 0.8580554041620516, + "grad_norm": 0.08550592511892319, + "learning_rate": 9.78292406666942e-06, + "loss": 0.0148, + "step": 36604 + }, + { + "epoch": 0.8580788457368566, + "grad_norm": 0.39494606852531433, + "learning_rate": 9.77974670495565e-06, + "loss": 0.069, + "step": 36605 + }, + { + "epoch": 0.8581022873116616, + "grad_norm": 0.2849303185939789, + "learning_rate": 9.776569832779537e-06, + "loss": 0.2951, + "step": 36606 + }, + { + "epoch": 0.8581257288864665, + "grad_norm": 0.6445016860961914, + "learning_rate": 9.773393450158285e-06, + "loss": 0.0954, + "step": 36607 + }, + { + "epoch": 0.8581491704612716, + "grad_norm": 0.7203782200813293, + "learning_rate": 9.770217557109141e-06, + "loss": 0.1243, + "step": 36608 + }, + { + "epoch": 0.8581726120360765, + "grad_norm": 0.33736181259155273, + "learning_rate": 9.76704215364933e-06, + "loss": 0.0601, + "step": 36609 + }, + { + "epoch": 0.8581960536108816, + "grad_norm": 0.1769220381975174, + "learning_rate": 9.763867239796065e-06, + "loss": 0.0341, + "step": 36610 + }, + { + "epoch": 0.8582194951856865, + "grad_norm": 0.22962017357349396, + "learning_rate": 9.760692815566608e-06, + "loss": 0.0344, + "step": 36611 + }, + { + "epoch": 0.8582429367604916, + "grad_norm": 0.2829749286174774, + "learning_rate": 9.757518880978167e-06, + "loss": 0.0285, + "step": 36612 + }, + { + "epoch": 0.8582663783352966, + "grad_norm": 0.40441882610321045, + "learning_rate": 9.754345436047984e-06, + "loss": 0.0508, + "step": 36613 + }, + { + "epoch": 0.8582898199101016, + "grad_norm": 0.4926089644432068, + "learning_rate": 9.751172480793247e-06, + "loss": 0.088, + "step": 36614 + }, + { + "epoch": 0.8583132614849066, + "grad_norm": 0.12389392405748367, + "learning_rate": 9.748000015231196e-06, + "loss": 0.0278, + "step": 36615 + }, + { + "epoch": 0.8583367030597115, + "grad_norm": 0.11309293657541275, + "learning_rate": 9.744828039379028e-06, + "loss": 0.0065, + "step": 36616 + }, + { + "epoch": 0.8583601446345166, + "grad_norm": 0.2391335517168045, + "learning_rate": 9.741656553253942e-06, + "loss": 0.0581, + "step": 36617 + }, + { + "epoch": 0.8583835862093215, + "grad_norm": 0.22332331538200378, + "learning_rate": 9.738485556873189e-06, + "loss": 0.0475, + "step": 36618 + }, + { + "epoch": 0.8584070277841266, + "grad_norm": 0.08819755911827087, + "learning_rate": 9.735315050253946e-06, + "loss": 0.011, + "step": 36619 + }, + { + "epoch": 0.8584304693589315, + "grad_norm": 0.5151594281196594, + "learning_rate": 9.73214503341342e-06, + "loss": 0.1438, + "step": 36620 + }, + { + "epoch": 0.8584539109337366, + "grad_norm": 0.2970782518386841, + "learning_rate": 9.72897550636882e-06, + "loss": 0.0294, + "step": 36621 + }, + { + "epoch": 0.8584773525085415, + "grad_norm": 0.39519909024238586, + "learning_rate": 9.725806469137333e-06, + "loss": 0.0603, + "step": 36622 + }, + { + "epoch": 0.8585007940833466, + "grad_norm": 0.5774315595626831, + "learning_rate": 9.722637921736144e-06, + "loss": 0.0923, + "step": 36623 + }, + { + "epoch": 0.8585242356581515, + "grad_norm": 0.45140838623046875, + "learning_rate": 9.719469864182452e-06, + "loss": 0.4324, + "step": 36624 + }, + { + "epoch": 0.8585476772329566, + "grad_norm": 0.3089063763618469, + "learning_rate": 9.716302296493484e-06, + "loss": 0.0346, + "step": 36625 + }, + { + "epoch": 0.8585711188077615, + "grad_norm": 0.25761017203330994, + "learning_rate": 9.713135218686386e-06, + "loss": 0.0576, + "step": 36626 + }, + { + "epoch": 0.8585945603825665, + "grad_norm": 0.3485805094242096, + "learning_rate": 9.709968630778366e-06, + "loss": 0.0495, + "step": 36627 + }, + { + "epoch": 0.8586180019573715, + "grad_norm": 0.16899721324443817, + "learning_rate": 9.706802532786596e-06, + "loss": 0.0241, + "step": 36628 + }, + { + "epoch": 0.8586414435321765, + "grad_norm": 1.067073941230774, + "learning_rate": 9.703636924728244e-06, + "loss": 0.1442, + "step": 36629 + }, + { + "epoch": 0.8586648851069815, + "grad_norm": 0.15377068519592285, + "learning_rate": 9.70047180662048e-06, + "loss": 0.0241, + "step": 36630 + }, + { + "epoch": 0.8586883266817865, + "grad_norm": 0.5187035202980042, + "learning_rate": 9.69730717848052e-06, + "loss": 0.1379, + "step": 36631 + }, + { + "epoch": 0.8587117682565915, + "grad_norm": 0.695580780506134, + "learning_rate": 9.69414304032551e-06, + "loss": 0.104, + "step": 36632 + }, + { + "epoch": 0.8587352098313965, + "grad_norm": 0.40011411905288696, + "learning_rate": 9.690979392172616e-06, + "loss": 0.0554, + "step": 36633 + }, + { + "epoch": 0.8587586514062014, + "grad_norm": 0.4935731887817383, + "learning_rate": 9.687816234039004e-06, + "loss": 0.0665, + "step": 36634 + }, + { + "epoch": 0.8587820929810065, + "grad_norm": 0.9116895794868469, + "learning_rate": 9.684653565941825e-06, + "loss": 0.1371, + "step": 36635 + }, + { + "epoch": 0.8588055345558114, + "grad_norm": 0.28586870431900024, + "learning_rate": 9.681491387898278e-06, + "loss": 0.0433, + "step": 36636 + }, + { + "epoch": 0.8588289761306165, + "grad_norm": 0.5686224102973938, + "learning_rate": 9.67832969992547e-06, + "loss": 0.1091, + "step": 36637 + }, + { + "epoch": 0.8588524177054214, + "grad_norm": 0.604544460773468, + "learning_rate": 9.67516850204061e-06, + "loss": 0.5007, + "step": 36638 + }, + { + "epoch": 0.8588758592802265, + "grad_norm": 0.1385018676519394, + "learning_rate": 9.672007794260829e-06, + "loss": 0.0152, + "step": 36639 + }, + { + "epoch": 0.8588993008550314, + "grad_norm": 0.42541807889938354, + "learning_rate": 9.668847576603269e-06, + "loss": 0.044, + "step": 36640 + }, + { + "epoch": 0.8589227424298365, + "grad_norm": 0.6541110277175903, + "learning_rate": 9.665687849085082e-06, + "loss": 0.0985, + "step": 36641 + }, + { + "epoch": 0.8589461840046414, + "grad_norm": 0.3635087311267853, + "learning_rate": 9.662528611723387e-06, + "loss": 0.0601, + "step": 36642 + }, + { + "epoch": 0.8589696255794464, + "grad_norm": 0.298947274684906, + "learning_rate": 9.659369864535383e-06, + "loss": 0.0664, + "step": 36643 + }, + { + "epoch": 0.8589930671542514, + "grad_norm": 0.4675978124141693, + "learning_rate": 9.656211607538169e-06, + "loss": 0.0912, + "step": 36644 + }, + { + "epoch": 0.8590165087290564, + "grad_norm": 0.4875580370426178, + "learning_rate": 9.653053840748893e-06, + "loss": 0.0604, + "step": 36645 + }, + { + "epoch": 0.8590399503038614, + "grad_norm": 0.44334539771080017, + "learning_rate": 9.64989656418469e-06, + "loss": 0.1021, + "step": 36646 + }, + { + "epoch": 0.8590633918786664, + "grad_norm": 0.704437792301178, + "learning_rate": 9.64673977786269e-06, + "loss": 0.1287, + "step": 36647 + }, + { + "epoch": 0.8590868334534714, + "grad_norm": 0.47981247305870056, + "learning_rate": 9.6435834818e-06, + "loss": 0.1344, + "step": 36648 + }, + { + "epoch": 0.8591102750282764, + "grad_norm": 0.124305360019207, + "learning_rate": 9.640427676013764e-06, + "loss": 0.0191, + "step": 36649 + }, + { + "epoch": 0.8591337166030814, + "grad_norm": 0.30928856134414673, + "learning_rate": 9.637272360521133e-06, + "loss": 0.0383, + "step": 36650 + }, + { + "epoch": 0.8591571581778864, + "grad_norm": 0.39161059260368347, + "learning_rate": 9.634117535339193e-06, + "loss": 0.0804, + "step": 36651 + }, + { + "epoch": 0.8591805997526913, + "grad_norm": 0.2799147367477417, + "learning_rate": 9.630963200485077e-06, + "loss": 0.0521, + "step": 36652 + }, + { + "epoch": 0.8592040413274964, + "grad_norm": 0.10116279125213623, + "learning_rate": 9.627809355975904e-06, + "loss": 0.0188, + "step": 36653 + }, + { + "epoch": 0.8592274829023013, + "grad_norm": 0.5950043797492981, + "learning_rate": 9.624656001828769e-06, + "loss": 0.0837, + "step": 36654 + }, + { + "epoch": 0.8592509244771064, + "grad_norm": 0.15122145414352417, + "learning_rate": 9.621503138060772e-06, + "loss": 0.0279, + "step": 36655 + }, + { + "epoch": 0.8592743660519113, + "grad_norm": 0.28905925154685974, + "learning_rate": 9.618350764689065e-06, + "loss": 0.0589, + "step": 36656 + }, + { + "epoch": 0.8592978076267164, + "grad_norm": 0.5718055963516235, + "learning_rate": 9.615198881730726e-06, + "loss": 0.0533, + "step": 36657 + }, + { + "epoch": 0.8593212492015213, + "grad_norm": 0.44328048825263977, + "learning_rate": 9.612047489202858e-06, + "loss": 0.0994, + "step": 36658 + }, + { + "epoch": 0.8593446907763264, + "grad_norm": 0.07058220356702805, + "learning_rate": 9.608896587122562e-06, + "loss": 0.0109, + "step": 36659 + }, + { + "epoch": 0.8593681323511313, + "grad_norm": 0.5405561923980713, + "learning_rate": 9.605746175506913e-06, + "loss": 0.5025, + "step": 36660 + }, + { + "epoch": 0.8593915739259363, + "grad_norm": 0.5252174735069275, + "learning_rate": 9.602596254373041e-06, + "loss": 0.6308, + "step": 36661 + }, + { + "epoch": 0.8594150155007413, + "grad_norm": 0.48160111904144287, + "learning_rate": 9.599446823738022e-06, + "loss": 0.053, + "step": 36662 + }, + { + "epoch": 0.8594384570755463, + "grad_norm": 0.05637592449784279, + "learning_rate": 9.596297883618955e-06, + "loss": 0.0042, + "step": 36663 + }, + { + "epoch": 0.8594618986503514, + "grad_norm": 0.32156452536582947, + "learning_rate": 9.593149434032912e-06, + "loss": 0.0459, + "step": 36664 + }, + { + "epoch": 0.8594853402251563, + "grad_norm": 0.08780469745397568, + "learning_rate": 9.59000147499699e-06, + "loss": 0.0092, + "step": 36665 + }, + { + "epoch": 0.8595087817999614, + "grad_norm": 0.14923560619354248, + "learning_rate": 9.586854006528256e-06, + "loss": 0.022, + "step": 36666 + }, + { + "epoch": 0.8595322233747663, + "grad_norm": 0.5033846497535706, + "learning_rate": 9.583707028643807e-06, + "loss": 0.0519, + "step": 36667 + }, + { + "epoch": 0.8595556649495714, + "grad_norm": 0.09505002200603485, + "learning_rate": 9.580560541360673e-06, + "loss": 0.0099, + "step": 36668 + }, + { + "epoch": 0.8595791065243763, + "grad_norm": 0.5067272782325745, + "learning_rate": 9.577414544695984e-06, + "loss": 0.654, + "step": 36669 + }, + { + "epoch": 0.8596025480991814, + "grad_norm": 0.1479412466287613, + "learning_rate": 9.574269038666784e-06, + "loss": 0.0199, + "step": 36670 + }, + { + "epoch": 0.8596259896739863, + "grad_norm": 0.2685847580432892, + "learning_rate": 9.571124023290146e-06, + "loss": 0.0401, + "step": 36671 + }, + { + "epoch": 0.8596494312487913, + "grad_norm": 0.10613568872213364, + "learning_rate": 9.567979498583113e-06, + "loss": 0.0189, + "step": 36672 + }, + { + "epoch": 0.8596728728235963, + "grad_norm": 0.5923414826393127, + "learning_rate": 9.564835464562783e-06, + "loss": 0.4917, + "step": 36673 + }, + { + "epoch": 0.8596963143984013, + "grad_norm": 0.26524025201797485, + "learning_rate": 9.561691921246207e-06, + "loss": 0.0318, + "step": 36674 + }, + { + "epoch": 0.8597197559732063, + "grad_norm": 0.2836417555809021, + "learning_rate": 9.558548868650407e-06, + "loss": 0.0355, + "step": 36675 + }, + { + "epoch": 0.8597431975480113, + "grad_norm": 0.15506085753440857, + "learning_rate": 9.55540630679248e-06, + "loss": 0.0332, + "step": 36676 + }, + { + "epoch": 0.8597666391228163, + "grad_norm": 0.18693557381629944, + "learning_rate": 9.552264235689467e-06, + "loss": 0.0295, + "step": 36677 + }, + { + "epoch": 0.8597900806976213, + "grad_norm": 0.16983479261398315, + "learning_rate": 9.5491226553584e-06, + "loss": 0.0188, + "step": 36678 + }, + { + "epoch": 0.8598135222724262, + "grad_norm": 1.1514548063278198, + "learning_rate": 9.545981565816353e-06, + "loss": 0.0734, + "step": 36679 + }, + { + "epoch": 0.8598369638472313, + "grad_norm": 0.24939113855361938, + "learning_rate": 9.542840967080324e-06, + "loss": 0.0176, + "step": 36680 + }, + { + "epoch": 0.8598604054220362, + "grad_norm": 0.2850465774536133, + "learning_rate": 9.53970085916741e-06, + "loss": 0.0527, + "step": 36681 + }, + { + "epoch": 0.8598838469968413, + "grad_norm": 0.6541739106178284, + "learning_rate": 9.536561242094611e-06, + "loss": 0.1262, + "step": 36682 + }, + { + "epoch": 0.8599072885716462, + "grad_norm": 0.44674253463745117, + "learning_rate": 9.533422115878987e-06, + "loss": 0.722, + "step": 36683 + }, + { + "epoch": 0.8599307301464513, + "grad_norm": 0.348565012216568, + "learning_rate": 9.530283480537528e-06, + "loss": 0.032, + "step": 36684 + }, + { + "epoch": 0.8599541717212562, + "grad_norm": 0.12974251806735992, + "learning_rate": 9.527145336087317e-06, + "loss": 0.0135, + "step": 36685 + }, + { + "epoch": 0.8599776132960613, + "grad_norm": 0.3828097879886627, + "learning_rate": 9.524007682545366e-06, + "loss": 0.0883, + "step": 36686 + }, + { + "epoch": 0.8600010548708662, + "grad_norm": 0.4659944474697113, + "learning_rate": 9.520870519928671e-06, + "loss": 0.1297, + "step": 36687 + }, + { + "epoch": 0.8600244964456712, + "grad_norm": 0.10037092119455338, + "learning_rate": 9.517733848254295e-06, + "loss": 0.0181, + "step": 36688 + }, + { + "epoch": 0.8600479380204762, + "grad_norm": 0.08608115464448929, + "learning_rate": 9.514597667539238e-06, + "loss": 0.0088, + "step": 36689 + }, + { + "epoch": 0.8600713795952812, + "grad_norm": 0.09841412305831909, + "learning_rate": 9.511461977800518e-06, + "loss": 0.0165, + "step": 36690 + }, + { + "epoch": 0.8600948211700862, + "grad_norm": 0.5402186512947083, + "learning_rate": 9.508326779055154e-06, + "loss": 0.1226, + "step": 36691 + }, + { + "epoch": 0.8601182627448912, + "grad_norm": 0.38434454798698425, + "learning_rate": 9.505192071320145e-06, + "loss": 0.0712, + "step": 36692 + }, + { + "epoch": 0.8601417043196962, + "grad_norm": 0.25246626138687134, + "learning_rate": 9.50205785461249e-06, + "loss": 0.2151, + "step": 36693 + }, + { + "epoch": 0.8601651458945012, + "grad_norm": 0.5591349601745605, + "learning_rate": 9.498924128949239e-06, + "loss": 0.0127, + "step": 36694 + }, + { + "epoch": 0.8601885874693062, + "grad_norm": 0.7977109551429749, + "learning_rate": 9.495790894347367e-06, + "loss": 0.1757, + "step": 36695 + }, + { + "epoch": 0.8602120290441112, + "grad_norm": 0.5607653856277466, + "learning_rate": 9.492658150823864e-06, + "loss": 0.0566, + "step": 36696 + }, + { + "epoch": 0.8602354706189161, + "grad_norm": 0.6121835708618164, + "learning_rate": 9.489525898395758e-06, + "loss": 0.6443, + "step": 36697 + }, + { + "epoch": 0.8602589121937212, + "grad_norm": 0.8938284516334534, + "learning_rate": 9.486394137080034e-06, + "loss": 0.6086, + "step": 36698 + }, + { + "epoch": 0.8602823537685261, + "grad_norm": 0.24595433473587036, + "learning_rate": 9.483262866893672e-06, + "loss": 0.0289, + "step": 36699 + }, + { + "epoch": 0.8603057953433312, + "grad_norm": 0.580985963344574, + "learning_rate": 9.480132087853654e-06, + "loss": 0.6621, + "step": 36700 + }, + { + "epoch": 0.8603292369181361, + "grad_norm": 0.5099486112594604, + "learning_rate": 9.477001799977003e-06, + "loss": 0.0802, + "step": 36701 + }, + { + "epoch": 0.8603526784929412, + "grad_norm": 0.3932153880596161, + "learning_rate": 9.473872003280693e-06, + "loss": 0.0811, + "step": 36702 + }, + { + "epoch": 0.8603761200677461, + "grad_norm": 0.22489811480045319, + "learning_rate": 9.470742697781687e-06, + "loss": 0.0276, + "step": 36703 + }, + { + "epoch": 0.8603995616425512, + "grad_norm": 0.6433992981910706, + "learning_rate": 9.467613883496983e-06, + "loss": 0.1342, + "step": 36704 + }, + { + "epoch": 0.8604230032173561, + "grad_norm": 0.409595251083374, + "learning_rate": 9.464485560443559e-06, + "loss": 0.077, + "step": 36705 + }, + { + "epoch": 0.8604464447921611, + "grad_norm": 0.39239513874053955, + "learning_rate": 9.461357728638354e-06, + "loss": 0.073, + "step": 36706 + }, + { + "epoch": 0.8604698863669661, + "grad_norm": 0.4273958206176758, + "learning_rate": 9.458230388098388e-06, + "loss": 0.1186, + "step": 36707 + }, + { + "epoch": 0.8604933279417711, + "grad_norm": 0.4123478829860687, + "learning_rate": 9.455103538840592e-06, + "loss": 0.0616, + "step": 36708 + }, + { + "epoch": 0.8605167695165761, + "grad_norm": 0.12102935463190079, + "learning_rate": 9.451977180881966e-06, + "loss": 0.0191, + "step": 36709 + }, + { + "epoch": 0.8605402110913811, + "grad_norm": 0.09453587234020233, + "learning_rate": 9.44885131423946e-06, + "loss": 0.0107, + "step": 36710 + }, + { + "epoch": 0.8605636526661861, + "grad_norm": 0.27940723299980164, + "learning_rate": 9.445725938930039e-06, + "loss": 0.0388, + "step": 36711 + }, + { + "epoch": 0.8605870942409911, + "grad_norm": 0.1628401279449463, + "learning_rate": 9.442601054970645e-06, + "loss": 0.0242, + "step": 36712 + }, + { + "epoch": 0.860610535815796, + "grad_norm": 0.14837031066417694, + "learning_rate": 9.439476662378233e-06, + "loss": 0.024, + "step": 36713 + }, + { + "epoch": 0.8606339773906011, + "grad_norm": 0.16469581425189972, + "learning_rate": 9.436352761169775e-06, + "loss": 0.0164, + "step": 36714 + }, + { + "epoch": 0.8606574189654062, + "grad_norm": 0.24854595959186554, + "learning_rate": 9.433229351362227e-06, + "loss": 0.2032, + "step": 36715 + }, + { + "epoch": 0.8606808605402111, + "grad_norm": 0.07112547010183334, + "learning_rate": 9.430106432972519e-06, + "loss": 0.0091, + "step": 36716 + }, + { + "epoch": 0.8607043021150161, + "grad_norm": 0.2434111386537552, + "learning_rate": 9.426984006017592e-06, + "loss": 0.0458, + "step": 36717 + }, + { + "epoch": 0.8607277436898211, + "grad_norm": 0.3198886215686798, + "learning_rate": 9.423862070514377e-06, + "loss": 0.038, + "step": 36718 + }, + { + "epoch": 0.8607511852646261, + "grad_norm": 0.25026369094848633, + "learning_rate": 9.420740626479863e-06, + "loss": 0.0259, + "step": 36719 + }, + { + "epoch": 0.8607746268394311, + "grad_norm": 0.39030522108078003, + "learning_rate": 9.417619673930932e-06, + "loss": 0.0619, + "step": 36720 + }, + { + "epoch": 0.8607980684142361, + "grad_norm": 0.541375994682312, + "learning_rate": 9.414499212884564e-06, + "loss": 0.4662, + "step": 36721 + }, + { + "epoch": 0.8608215099890411, + "grad_norm": 0.30422595143318176, + "learning_rate": 9.411379243357677e-06, + "loss": 0.0434, + "step": 36722 + }, + { + "epoch": 0.8608449515638461, + "grad_norm": 0.5263041853904724, + "learning_rate": 9.408259765367178e-06, + "loss": 0.0586, + "step": 36723 + }, + { + "epoch": 0.860868393138651, + "grad_norm": 0.38858070969581604, + "learning_rate": 9.405140778930021e-06, + "loss": 0.0747, + "step": 36724 + }, + { + "epoch": 0.8608918347134561, + "grad_norm": 0.21484321355819702, + "learning_rate": 9.402022284063095e-06, + "loss": 0.0504, + "step": 36725 + }, + { + "epoch": 0.860915276288261, + "grad_norm": 0.3406102955341339, + "learning_rate": 9.39890428078336e-06, + "loss": 0.0735, + "step": 36726 + }, + { + "epoch": 0.8609387178630661, + "grad_norm": 0.296462744474411, + "learning_rate": 9.395786769107728e-06, + "loss": 0.0595, + "step": 36727 + }, + { + "epoch": 0.860962159437871, + "grad_norm": 0.8570088148117065, + "learning_rate": 9.392669749053096e-06, + "loss": 0.1482, + "step": 36728 + }, + { + "epoch": 0.8609856010126761, + "grad_norm": 0.46043840050697327, + "learning_rate": 9.389553220636383e-06, + "loss": 0.0955, + "step": 36729 + }, + { + "epoch": 0.861009042587481, + "grad_norm": 0.4168214499950409, + "learning_rate": 9.386437183874508e-06, + "loss": 0.0638, + "step": 36730 + }, + { + "epoch": 0.8610324841622861, + "grad_norm": 0.1565273255109787, + "learning_rate": 9.383321638784349e-06, + "loss": 0.0205, + "step": 36731 + }, + { + "epoch": 0.861055925737091, + "grad_norm": 0.02668742835521698, + "learning_rate": 9.380206585382856e-06, + "loss": 0.0016, + "step": 36732 + }, + { + "epoch": 0.861079367311896, + "grad_norm": 0.3450513184070587, + "learning_rate": 9.377092023686884e-06, + "loss": 0.0511, + "step": 36733 + }, + { + "epoch": 0.861102808886701, + "grad_norm": 0.22274066507816315, + "learning_rate": 9.373977953713387e-06, + "loss": 0.0443, + "step": 36734 + }, + { + "epoch": 0.861126250461506, + "grad_norm": 0.45814386010169983, + "learning_rate": 9.370864375479238e-06, + "loss": 0.4748, + "step": 36735 + }, + { + "epoch": 0.861149692036311, + "grad_norm": 0.42802268266677856, + "learning_rate": 9.367751289001314e-06, + "loss": 0.0658, + "step": 36736 + }, + { + "epoch": 0.861173133611116, + "grad_norm": 0.40591374039649963, + "learning_rate": 9.364638694296534e-06, + "loss": 0.0389, + "step": 36737 + }, + { + "epoch": 0.861196575185921, + "grad_norm": 0.3137516677379608, + "learning_rate": 9.361526591381742e-06, + "loss": 0.0515, + "step": 36738 + }, + { + "epoch": 0.861220016760726, + "grad_norm": 0.11620920896530151, + "learning_rate": 9.358414980273878e-06, + "loss": 0.0115, + "step": 36739 + }, + { + "epoch": 0.861243458335531, + "grad_norm": 0.236031636595726, + "learning_rate": 9.355303860989805e-06, + "loss": 0.0398, + "step": 36740 + }, + { + "epoch": 0.861266899910336, + "grad_norm": 0.10440658777952194, + "learning_rate": 9.352193233546402e-06, + "loss": 0.0101, + "step": 36741 + }, + { + "epoch": 0.8612903414851409, + "grad_norm": 0.39144304394721985, + "learning_rate": 9.349083097960543e-06, + "loss": 0.1105, + "step": 36742 + }, + { + "epoch": 0.861313783059946, + "grad_norm": 0.621060848236084, + "learning_rate": 9.345973454249114e-06, + "loss": 0.0973, + "step": 36743 + }, + { + "epoch": 0.8613372246347509, + "grad_norm": 0.3149369955062866, + "learning_rate": 9.342864302428955e-06, + "loss": 0.2009, + "step": 36744 + }, + { + "epoch": 0.861360666209556, + "grad_norm": 0.20152392983436584, + "learning_rate": 9.339755642516968e-06, + "loss": 0.0415, + "step": 36745 + }, + { + "epoch": 0.8613841077843609, + "grad_norm": 0.5115137100219727, + "learning_rate": 9.336647474530036e-06, + "loss": 0.1259, + "step": 36746 + }, + { + "epoch": 0.861407549359166, + "grad_norm": 0.4151934087276459, + "learning_rate": 9.333539798485003e-06, + "loss": 0.0427, + "step": 36747 + }, + { + "epoch": 0.8614309909339709, + "grad_norm": 0.7365441918373108, + "learning_rate": 9.330432614398732e-06, + "loss": 0.1875, + "step": 36748 + }, + { + "epoch": 0.861454432508776, + "grad_norm": 0.4403044879436493, + "learning_rate": 9.327325922288089e-06, + "loss": 0.0865, + "step": 36749 + }, + { + "epoch": 0.8614778740835809, + "grad_norm": 0.5682581663131714, + "learning_rate": 9.324219722169913e-06, + "loss": 0.126, + "step": 36750 + }, + { + "epoch": 0.861501315658386, + "grad_norm": 0.13645654916763306, + "learning_rate": 9.321114014061062e-06, + "loss": 0.0261, + "step": 36751 + }, + { + "epoch": 0.8615247572331909, + "grad_norm": 0.1980675756931305, + "learning_rate": 9.318008797978406e-06, + "loss": 0.0599, + "step": 36752 + }, + { + "epoch": 0.8615481988079959, + "grad_norm": 0.5341518521308899, + "learning_rate": 9.31490407393879e-06, + "loss": 0.4818, + "step": 36753 + }, + { + "epoch": 0.8615716403828009, + "grad_norm": 0.6801672577857971, + "learning_rate": 9.311799841959057e-06, + "loss": 0.0963, + "step": 36754 + }, + { + "epoch": 0.8615950819576059, + "grad_norm": 0.14394676685333252, + "learning_rate": 9.308696102056047e-06, + "loss": 0.015, + "step": 36755 + }, + { + "epoch": 0.8616185235324109, + "grad_norm": 0.26554644107818604, + "learning_rate": 9.30559285424657e-06, + "loss": 0.0441, + "step": 36756 + }, + { + "epoch": 0.8616419651072159, + "grad_norm": 0.39381957054138184, + "learning_rate": 9.302490098547534e-06, + "loss": 0.0558, + "step": 36757 + }, + { + "epoch": 0.8616654066820209, + "grad_norm": 0.5053015351295471, + "learning_rate": 9.299387834975703e-06, + "loss": 0.1097, + "step": 36758 + }, + { + "epoch": 0.8616888482568259, + "grad_norm": 0.03273015841841698, + "learning_rate": 9.296286063547977e-06, + "loss": 0.0016, + "step": 36759 + }, + { + "epoch": 0.8617122898316308, + "grad_norm": 0.2404208481311798, + "learning_rate": 9.293184784281139e-06, + "loss": 0.0505, + "step": 36760 + }, + { + "epoch": 0.8617357314064359, + "grad_norm": 0.17624203860759735, + "learning_rate": 9.290083997192034e-06, + "loss": 0.0374, + "step": 36761 + }, + { + "epoch": 0.8617591729812408, + "grad_norm": 0.3406442701816559, + "learning_rate": 9.286983702297491e-06, + "loss": 0.061, + "step": 36762 + }, + { + "epoch": 0.8617826145560459, + "grad_norm": 1.5166573524475098, + "learning_rate": 9.283883899614299e-06, + "loss": 0.1376, + "step": 36763 + }, + { + "epoch": 0.8618060561308508, + "grad_norm": 0.4804295599460602, + "learning_rate": 9.28078458915933e-06, + "loss": 0.1071, + "step": 36764 + }, + { + "epoch": 0.8618294977056559, + "grad_norm": 0.18876402080059052, + "learning_rate": 9.277685770949373e-06, + "loss": 0.0331, + "step": 36765 + }, + { + "epoch": 0.8618529392804609, + "grad_norm": 0.4355129301548004, + "learning_rate": 9.274587445001238e-06, + "loss": 0.0387, + "step": 36766 + }, + { + "epoch": 0.8618763808552659, + "grad_norm": 0.11978697776794434, + "learning_rate": 9.271489611331741e-06, + "loss": 0.0137, + "step": 36767 + }, + { + "epoch": 0.8618998224300709, + "grad_norm": 0.34945017099380493, + "learning_rate": 9.268392269957704e-06, + "loss": 0.036, + "step": 36768 + }, + { + "epoch": 0.8619232640048758, + "grad_norm": 0.19837673008441925, + "learning_rate": 9.265295420895891e-06, + "loss": 0.0244, + "step": 36769 + }, + { + "epoch": 0.8619467055796809, + "grad_norm": 0.31637370586395264, + "learning_rate": 9.262199064163147e-06, + "loss": 0.0333, + "step": 36770 + }, + { + "epoch": 0.8619701471544858, + "grad_norm": 0.33594605326652527, + "learning_rate": 9.259103199776275e-06, + "loss": 0.0316, + "step": 36771 + }, + { + "epoch": 0.8619935887292909, + "grad_norm": 0.3219623863697052, + "learning_rate": 9.256007827752055e-06, + "loss": 0.0681, + "step": 36772 + }, + { + "epoch": 0.8620170303040958, + "grad_norm": 0.3135455250740051, + "learning_rate": 9.252912948107295e-06, + "loss": 0.0532, + "step": 36773 + }, + { + "epoch": 0.8620404718789009, + "grad_norm": 0.19555449485778809, + "learning_rate": 9.24981856085878e-06, + "loss": 0.0275, + "step": 36774 + }, + { + "epoch": 0.8620639134537058, + "grad_norm": 0.17814214527606964, + "learning_rate": 9.246724666023299e-06, + "loss": 0.0203, + "step": 36775 + }, + { + "epoch": 0.8620873550285109, + "grad_norm": 0.12581807374954224, + "learning_rate": 9.243631263617625e-06, + "loss": 0.0179, + "step": 36776 + }, + { + "epoch": 0.8621107966033158, + "grad_norm": 0.362400621175766, + "learning_rate": 9.24053835365858e-06, + "loss": 0.0741, + "step": 36777 + }, + { + "epoch": 0.8621342381781208, + "grad_norm": 0.40173909068107605, + "learning_rate": 9.237445936162925e-06, + "loss": 0.0472, + "step": 36778 + }, + { + "epoch": 0.8621576797529258, + "grad_norm": 2.0805859565734863, + "learning_rate": 9.23435401114744e-06, + "loss": 0.131, + "step": 36779 + }, + { + "epoch": 0.8621811213277308, + "grad_norm": 0.21945108473300934, + "learning_rate": 9.231262578628908e-06, + "loss": 0.032, + "step": 36780 + }, + { + "epoch": 0.8622045629025358, + "grad_norm": 0.3112441599369049, + "learning_rate": 9.228171638624062e-06, + "loss": 0.045, + "step": 36781 + }, + { + "epoch": 0.8622280044773408, + "grad_norm": 0.6554080247879028, + "learning_rate": 9.225081191149743e-06, + "loss": 0.1732, + "step": 36782 + }, + { + "epoch": 0.8622514460521458, + "grad_norm": 0.21942445635795593, + "learning_rate": 9.22199123622266e-06, + "loss": 0.0311, + "step": 36783 + }, + { + "epoch": 0.8622748876269508, + "grad_norm": 0.45002302527427673, + "learning_rate": 9.218901773859623e-06, + "loss": 0.0569, + "step": 36784 + }, + { + "epoch": 0.8622983292017558, + "grad_norm": 0.32584208250045776, + "learning_rate": 9.215812804077384e-06, + "loss": 0.0673, + "step": 36785 + }, + { + "epoch": 0.8623217707765608, + "grad_norm": 0.5094361901283264, + "learning_rate": 9.212724326892696e-06, + "loss": 0.0848, + "step": 36786 + }, + { + "epoch": 0.8623452123513657, + "grad_norm": 0.13999156653881073, + "learning_rate": 9.209636342322314e-06, + "loss": 0.0247, + "step": 36787 + }, + { + "epoch": 0.8623686539261708, + "grad_norm": 0.34420496225357056, + "learning_rate": 9.206548850383002e-06, + "loss": 0.0607, + "step": 36788 + }, + { + "epoch": 0.8623920955009757, + "grad_norm": 0.48621129989624023, + "learning_rate": 9.203461851091488e-06, + "loss": 0.0928, + "step": 36789 + }, + { + "epoch": 0.8624155370757808, + "grad_norm": 0.49845096468925476, + "learning_rate": 9.200375344464563e-06, + "loss": 0.0579, + "step": 36790 + }, + { + "epoch": 0.8624389786505857, + "grad_norm": 0.5245177745819092, + "learning_rate": 9.197289330518955e-06, + "loss": 0.1155, + "step": 36791 + }, + { + "epoch": 0.8624624202253908, + "grad_norm": 0.1240864023566246, + "learning_rate": 9.194203809271406e-06, + "loss": 0.0056, + "step": 36792 + }, + { + "epoch": 0.8624858618001957, + "grad_norm": 0.31057217717170715, + "learning_rate": 9.19111878073864e-06, + "loss": 0.0524, + "step": 36793 + }, + { + "epoch": 0.8625093033750008, + "grad_norm": 0.33168357610702515, + "learning_rate": 9.188034244937438e-06, + "loss": 0.0739, + "step": 36794 + }, + { + "epoch": 0.8625327449498057, + "grad_norm": 0.30272072553634644, + "learning_rate": 9.184950201884512e-06, + "loss": 0.0433, + "step": 36795 + }, + { + "epoch": 0.8625561865246107, + "grad_norm": 0.29802805185317993, + "learning_rate": 9.181866651596593e-06, + "loss": 0.0348, + "step": 36796 + }, + { + "epoch": 0.8625796280994157, + "grad_norm": 0.32680219411849976, + "learning_rate": 9.178783594090423e-06, + "loss": 0.0285, + "step": 36797 + }, + { + "epoch": 0.8626030696742207, + "grad_norm": 0.5482035875320435, + "learning_rate": 9.175701029382733e-06, + "loss": 0.0816, + "step": 36798 + }, + { + "epoch": 0.8626265112490257, + "grad_norm": 1.330161690711975, + "learning_rate": 9.172618957490242e-06, + "loss": 0.1527, + "step": 36799 + }, + { + "epoch": 0.8626499528238307, + "grad_norm": 0.328043133020401, + "learning_rate": 9.169537378429683e-06, + "loss": 0.0504, + "step": 36800 + }, + { + "epoch": 0.8626733943986357, + "grad_norm": 0.5177689790725708, + "learning_rate": 9.16645629221774e-06, + "loss": 0.0487, + "step": 36801 + }, + { + "epoch": 0.8626968359734407, + "grad_norm": 0.35286450386047363, + "learning_rate": 9.16337569887118e-06, + "loss": 0.0747, + "step": 36802 + }, + { + "epoch": 0.8627202775482457, + "grad_norm": 0.23795627057552338, + "learning_rate": 9.1602955984067e-06, + "loss": 0.0309, + "step": 36803 + }, + { + "epoch": 0.8627437191230507, + "grad_norm": 0.06677281111478806, + "learning_rate": 9.157215990841005e-06, + "loss": 0.0052, + "step": 36804 + }, + { + "epoch": 0.8627671606978556, + "grad_norm": 0.1597820222377777, + "learning_rate": 9.15413687619081e-06, + "loss": 0.0274, + "step": 36805 + }, + { + "epoch": 0.8627906022726607, + "grad_norm": 0.22697007656097412, + "learning_rate": 9.15105825447281e-06, + "loss": 0.0335, + "step": 36806 + }, + { + "epoch": 0.8628140438474656, + "grad_norm": 0.436405211687088, + "learning_rate": 9.147980125703736e-06, + "loss": 0.0547, + "step": 36807 + }, + { + "epoch": 0.8628374854222707, + "grad_norm": 0.10011259466409683, + "learning_rate": 9.144902489900265e-06, + "loss": 0.0087, + "step": 36808 + }, + { + "epoch": 0.8628609269970756, + "grad_norm": 0.3437802493572235, + "learning_rate": 9.141825347079114e-06, + "loss": 0.0336, + "step": 36809 + }, + { + "epoch": 0.8628843685718807, + "grad_norm": 0.46237701177597046, + "learning_rate": 9.138748697256982e-06, + "loss": 0.1079, + "step": 36810 + }, + { + "epoch": 0.8629078101466856, + "grad_norm": 0.2823537290096283, + "learning_rate": 9.135672540450558e-06, + "loss": 0.0348, + "step": 36811 + }, + { + "epoch": 0.8629312517214907, + "grad_norm": 0.4971712827682495, + "learning_rate": 9.132596876676524e-06, + "loss": 0.0586, + "step": 36812 + }, + { + "epoch": 0.8629546932962956, + "grad_norm": 0.4406847059726715, + "learning_rate": 9.129521705951572e-06, + "loss": 0.0662, + "step": 36813 + }, + { + "epoch": 0.8629781348711006, + "grad_norm": 0.3727649748325348, + "learning_rate": 9.126447028292372e-06, + "loss": 0.0628, + "step": 36814 + }, + { + "epoch": 0.8630015764459056, + "grad_norm": 0.24081555008888245, + "learning_rate": 9.12337284371565e-06, + "loss": 0.0285, + "step": 36815 + }, + { + "epoch": 0.8630250180207106, + "grad_norm": 0.3711349070072174, + "learning_rate": 9.120299152238054e-06, + "loss": 0.1453, + "step": 36816 + }, + { + "epoch": 0.8630484595955156, + "grad_norm": 0.6176302433013916, + "learning_rate": 9.117225953876274e-06, + "loss": 0.0791, + "step": 36817 + }, + { + "epoch": 0.8630719011703206, + "grad_norm": 0.3516792953014374, + "learning_rate": 9.114153248646962e-06, + "loss": 0.0671, + "step": 36818 + }, + { + "epoch": 0.8630953427451257, + "grad_norm": 0.47353771328926086, + "learning_rate": 9.111081036566837e-06, + "loss": 0.1045, + "step": 36819 + }, + { + "epoch": 0.8631187843199306, + "grad_norm": 0.43422049283981323, + "learning_rate": 9.108009317652534e-06, + "loss": 0.0432, + "step": 36820 + }, + { + "epoch": 0.8631422258947357, + "grad_norm": 0.2889321446418762, + "learning_rate": 9.104938091920712e-06, + "loss": 0.0347, + "step": 36821 + }, + { + "epoch": 0.8631656674695406, + "grad_norm": 0.14444176852703094, + "learning_rate": 9.101867359388061e-06, + "loss": 0.0131, + "step": 36822 + }, + { + "epoch": 0.8631891090443456, + "grad_norm": 0.639817476272583, + "learning_rate": 9.098797120071245e-06, + "loss": 0.662, + "step": 36823 + }, + { + "epoch": 0.8632125506191506, + "grad_norm": 0.07095538079738617, + "learning_rate": 9.095727373986907e-06, + "loss": 0.0084, + "step": 36824 + }, + { + "epoch": 0.8632359921939556, + "grad_norm": 0.4444730579853058, + "learning_rate": 9.092658121151709e-06, + "loss": 0.0589, + "step": 36825 + }, + { + "epoch": 0.8632594337687606, + "grad_norm": 0.6569893956184387, + "learning_rate": 9.089589361582307e-06, + "loss": 0.079, + "step": 36826 + }, + { + "epoch": 0.8632828753435656, + "grad_norm": 0.2426876723766327, + "learning_rate": 9.086521095295331e-06, + "loss": 0.0459, + "step": 36827 + }, + { + "epoch": 0.8633063169183706, + "grad_norm": 0.5399582386016846, + "learning_rate": 9.083453322307466e-06, + "loss": 0.1388, + "step": 36828 + }, + { + "epoch": 0.8633297584931756, + "grad_norm": 0.36891740560531616, + "learning_rate": 9.080386042635336e-06, + "loss": 0.0657, + "step": 36829 + }, + { + "epoch": 0.8633532000679806, + "grad_norm": 0.23626500368118286, + "learning_rate": 9.077319256295569e-06, + "loss": 0.0209, + "step": 36830 + }, + { + "epoch": 0.8633766416427856, + "grad_norm": 0.4938238263130188, + "learning_rate": 9.07425296330484e-06, + "loss": 0.1032, + "step": 36831 + }, + { + "epoch": 0.8634000832175905, + "grad_norm": 0.4922615885734558, + "learning_rate": 9.071187163679773e-06, + "loss": 0.0921, + "step": 36832 + }, + { + "epoch": 0.8634235247923956, + "grad_norm": 0.31855547428131104, + "learning_rate": 9.068121857437007e-06, + "loss": 0.0357, + "step": 36833 + }, + { + "epoch": 0.8634469663672005, + "grad_norm": 0.5933419466018677, + "learning_rate": 9.06505704459315e-06, + "loss": 0.6226, + "step": 36834 + }, + { + "epoch": 0.8634704079420056, + "grad_norm": 0.41042280197143555, + "learning_rate": 9.061992725164858e-06, + "loss": 0.0622, + "step": 36835 + }, + { + "epoch": 0.8634938495168105, + "grad_norm": 0.21964497864246368, + "learning_rate": 9.05892889916876e-06, + "loss": 0.0131, + "step": 36836 + }, + { + "epoch": 0.8635172910916156, + "grad_norm": 0.41698333621025085, + "learning_rate": 9.055865566621469e-06, + "loss": 0.0566, + "step": 36837 + }, + { + "epoch": 0.8635407326664205, + "grad_norm": 0.2600995600223541, + "learning_rate": 9.05280272753961e-06, + "loss": 0.0374, + "step": 36838 + }, + { + "epoch": 0.8635641742412256, + "grad_norm": 0.20933914184570312, + "learning_rate": 9.049740381939787e-06, + "loss": 0.0163, + "step": 36839 + }, + { + "epoch": 0.8635876158160305, + "grad_norm": 0.31473004817962646, + "learning_rate": 9.046678529838637e-06, + "loss": 0.2873, + "step": 36840 + }, + { + "epoch": 0.8636110573908355, + "grad_norm": 0.3428831398487091, + "learning_rate": 9.043617171252771e-06, + "loss": 0.0727, + "step": 36841 + }, + { + "epoch": 0.8636344989656405, + "grad_norm": 0.5700575709342957, + "learning_rate": 9.040556306198777e-06, + "loss": 0.5752, + "step": 36842 + }, + { + "epoch": 0.8636579405404455, + "grad_norm": 0.11562550812959671, + "learning_rate": 9.037495934693296e-06, + "loss": 0.0149, + "step": 36843 + }, + { + "epoch": 0.8636813821152505, + "grad_norm": 0.3503316044807434, + "learning_rate": 9.034436056752927e-06, + "loss": 0.0553, + "step": 36844 + }, + { + "epoch": 0.8637048236900555, + "grad_norm": 0.43418532609939575, + "learning_rate": 9.031376672394265e-06, + "loss": 0.1106, + "step": 36845 + }, + { + "epoch": 0.8637282652648605, + "grad_norm": 0.44697144627571106, + "learning_rate": 9.028317781633889e-06, + "loss": 0.0752, + "step": 36846 + }, + { + "epoch": 0.8637517068396655, + "grad_norm": 0.2770187556743622, + "learning_rate": 9.025259384488439e-06, + "loss": 0.0434, + "step": 36847 + }, + { + "epoch": 0.8637751484144705, + "grad_norm": 0.38315054774284363, + "learning_rate": 9.02220148097449e-06, + "loss": 0.0532, + "step": 36848 + }, + { + "epoch": 0.8637985899892755, + "grad_norm": 0.5831729173660278, + "learning_rate": 9.019144071108631e-06, + "loss": 0.0742, + "step": 36849 + }, + { + "epoch": 0.8638220315640804, + "grad_norm": 0.7589969038963318, + "learning_rate": 9.01608715490746e-06, + "loss": 0.1672, + "step": 36850 + }, + { + "epoch": 0.8638454731388855, + "grad_norm": 0.1671656370162964, + "learning_rate": 9.013030732387562e-06, + "loss": 0.028, + "step": 36851 + }, + { + "epoch": 0.8638689147136904, + "grad_norm": 0.4796138107776642, + "learning_rate": 9.00997480356549e-06, + "loss": 0.0906, + "step": 36852 + }, + { + "epoch": 0.8638923562884955, + "grad_norm": 0.4017831087112427, + "learning_rate": 9.006919368457867e-06, + "loss": 0.0798, + "step": 36853 + }, + { + "epoch": 0.8639157978633004, + "grad_norm": 0.7945322394371033, + "learning_rate": 9.003864427081254e-06, + "loss": 1.0745, + "step": 36854 + }, + { + "epoch": 0.8639392394381055, + "grad_norm": 0.4098627269268036, + "learning_rate": 9.00080997945224e-06, + "loss": 0.0564, + "step": 36855 + }, + { + "epoch": 0.8639626810129104, + "grad_norm": 0.2107032984495163, + "learning_rate": 8.99775602558739e-06, + "loss": 0.0448, + "step": 36856 + }, + { + "epoch": 0.8639861225877155, + "grad_norm": 0.37738704681396484, + "learning_rate": 8.994702565503278e-06, + "loss": 0.0649, + "step": 36857 + }, + { + "epoch": 0.8640095641625204, + "grad_norm": 0.38928309082984924, + "learning_rate": 8.991649599216467e-06, + "loss": 0.0573, + "step": 36858 + }, + { + "epoch": 0.8640330057373254, + "grad_norm": 0.24947960674762726, + "learning_rate": 8.988597126743502e-06, + "loss": 0.0162, + "step": 36859 + }, + { + "epoch": 0.8640564473121304, + "grad_norm": 0.4575570523738861, + "learning_rate": 8.985545148100994e-06, + "loss": 0.5964, + "step": 36860 + }, + { + "epoch": 0.8640798888869354, + "grad_norm": 0.8615202307701111, + "learning_rate": 8.982493663305468e-06, + "loss": 0.175, + "step": 36861 + }, + { + "epoch": 0.8641033304617404, + "grad_norm": 0.6320362687110901, + "learning_rate": 8.979442672373506e-06, + "loss": 0.0698, + "step": 36862 + }, + { + "epoch": 0.8641267720365454, + "grad_norm": 0.5484130382537842, + "learning_rate": 8.976392175321635e-06, + "loss": 0.1197, + "step": 36863 + }, + { + "epoch": 0.8641502136113504, + "grad_norm": 0.3597138524055481, + "learning_rate": 8.973342172166422e-06, + "loss": 0.0589, + "step": 36864 + }, + { + "epoch": 0.8641736551861554, + "grad_norm": 0.3758346438407898, + "learning_rate": 8.970292662924396e-06, + "loss": 0.0624, + "step": 36865 + }, + { + "epoch": 0.8641970967609603, + "grad_norm": 0.21626079082489014, + "learning_rate": 8.967243647612122e-06, + "loss": 0.0257, + "step": 36866 + }, + { + "epoch": 0.8642205383357654, + "grad_norm": 0.5126163363456726, + "learning_rate": 8.964195126246155e-06, + "loss": 0.0977, + "step": 36867 + }, + { + "epoch": 0.8642439799105703, + "grad_norm": 0.37096869945526123, + "learning_rate": 8.961147098843038e-06, + "loss": 0.059, + "step": 36868 + }, + { + "epoch": 0.8642674214853754, + "grad_norm": 0.19067049026489258, + "learning_rate": 8.958099565419287e-06, + "loss": 0.0316, + "step": 36869 + }, + { + "epoch": 0.8642908630601804, + "grad_norm": 0.4368435740470886, + "learning_rate": 8.955052525991448e-06, + "loss": 0.0516, + "step": 36870 + }, + { + "epoch": 0.8643143046349854, + "grad_norm": 0.09925699234008789, + "learning_rate": 8.95200598057605e-06, + "loss": 0.0152, + "step": 36871 + }, + { + "epoch": 0.8643377462097904, + "grad_norm": 0.206989586353302, + "learning_rate": 8.948959929189615e-06, + "loss": 0.034, + "step": 36872 + }, + { + "epoch": 0.8643611877845954, + "grad_norm": 0.2886110544204712, + "learning_rate": 8.945914371848707e-06, + "loss": 0.0558, + "step": 36873 + }, + { + "epoch": 0.8643846293594004, + "grad_norm": 0.2856823801994324, + "learning_rate": 8.942869308569824e-06, + "loss": 0.0469, + "step": 36874 + }, + { + "epoch": 0.8644080709342054, + "grad_norm": 0.46196046471595764, + "learning_rate": 8.939824739369496e-06, + "loss": 0.5069, + "step": 36875 + }, + { + "epoch": 0.8644315125090104, + "grad_norm": 0.6441255211830139, + "learning_rate": 8.936780664264233e-06, + "loss": 0.128, + "step": 36876 + }, + { + "epoch": 0.8644549540838153, + "grad_norm": 0.5221302509307861, + "learning_rate": 8.933737083270544e-06, + "loss": 0.0856, + "step": 36877 + }, + { + "epoch": 0.8644783956586204, + "grad_norm": 0.4127650260925293, + "learning_rate": 8.930693996404971e-06, + "loss": 0.5835, + "step": 36878 + }, + { + "epoch": 0.8645018372334253, + "grad_norm": 0.12329726666212082, + "learning_rate": 8.927651403684e-06, + "loss": 0.0287, + "step": 36879 + }, + { + "epoch": 0.8645252788082304, + "grad_norm": 0.5389674305915833, + "learning_rate": 8.924609305124165e-06, + "loss": 0.8379, + "step": 36880 + }, + { + "epoch": 0.8645487203830353, + "grad_norm": 0.18726873397827148, + "learning_rate": 8.921567700741962e-06, + "loss": 0.0498, + "step": 36881 + }, + { + "epoch": 0.8645721619578404, + "grad_norm": 0.12613818049430847, + "learning_rate": 8.918526590553899e-06, + "loss": 0.0156, + "step": 36882 + }, + { + "epoch": 0.8645956035326453, + "grad_norm": 0.2867598831653595, + "learning_rate": 8.915485974576477e-06, + "loss": 0.0308, + "step": 36883 + }, + { + "epoch": 0.8646190451074504, + "grad_norm": 0.237571120262146, + "learning_rate": 8.912445852826157e-06, + "loss": 0.0144, + "step": 36884 + }, + { + "epoch": 0.8646424866822553, + "grad_norm": 0.2763838469982147, + "learning_rate": 8.909406225319495e-06, + "loss": 0.0263, + "step": 36885 + }, + { + "epoch": 0.8646659282570603, + "grad_norm": 0.26288700103759766, + "learning_rate": 8.906367092072954e-06, + "loss": 0.0626, + "step": 36886 + }, + { + "epoch": 0.8646893698318653, + "grad_norm": 0.13921524584293365, + "learning_rate": 8.903328453103021e-06, + "loss": 0.0127, + "step": 36887 + }, + { + "epoch": 0.8647128114066703, + "grad_norm": 0.4645269513130188, + "learning_rate": 8.900290308426195e-06, + "loss": 0.065, + "step": 36888 + }, + { + "epoch": 0.8647362529814753, + "grad_norm": 0.16043545305728912, + "learning_rate": 8.897252658058952e-06, + "loss": 0.0242, + "step": 36889 + }, + { + "epoch": 0.8647596945562803, + "grad_norm": 0.2441142201423645, + "learning_rate": 8.894215502017766e-06, + "loss": 0.0468, + "step": 36890 + }, + { + "epoch": 0.8647831361310853, + "grad_norm": 0.602753758430481, + "learning_rate": 8.891178840319126e-06, + "loss": 0.0677, + "step": 36891 + }, + { + "epoch": 0.8648065777058903, + "grad_norm": 0.38491612672805786, + "learning_rate": 8.888142672979527e-06, + "loss": 0.0708, + "step": 36892 + }, + { + "epoch": 0.8648300192806953, + "grad_norm": 0.5116010904312134, + "learning_rate": 8.885107000015424e-06, + "loss": 0.0425, + "step": 36893 + }, + { + "epoch": 0.8648534608555003, + "grad_norm": 0.4363682270050049, + "learning_rate": 8.882071821443294e-06, + "loss": 0.0278, + "step": 36894 + }, + { + "epoch": 0.8648769024303052, + "grad_norm": 0.18148639798164368, + "learning_rate": 8.879037137279611e-06, + "loss": 0.029, + "step": 36895 + }, + { + "epoch": 0.8649003440051103, + "grad_norm": 0.40531396865844727, + "learning_rate": 8.87600294754083e-06, + "loss": 0.0694, + "step": 36896 + }, + { + "epoch": 0.8649237855799152, + "grad_norm": 0.511433482170105, + "learning_rate": 8.872969252243401e-06, + "loss": 0.1757, + "step": 36897 + }, + { + "epoch": 0.8649472271547203, + "grad_norm": 0.41917598247528076, + "learning_rate": 8.869936051403815e-06, + "loss": 0.0747, + "step": 36898 + }, + { + "epoch": 0.8649706687295252, + "grad_norm": 0.17813998460769653, + "learning_rate": 8.866903345038523e-06, + "loss": 0.0337, + "step": 36899 + }, + { + "epoch": 0.8649941103043303, + "grad_norm": 0.08587305247783661, + "learning_rate": 8.863871133163982e-06, + "loss": 0.0163, + "step": 36900 + }, + { + "epoch": 0.8650175518791352, + "grad_norm": 0.15870049595832825, + "learning_rate": 8.86083941579664e-06, + "loss": 0.0222, + "step": 36901 + }, + { + "epoch": 0.8650409934539403, + "grad_norm": 0.5801240801811218, + "learning_rate": 8.857808192952932e-06, + "loss": 0.0826, + "step": 36902 + }, + { + "epoch": 0.8650644350287452, + "grad_norm": 0.08663590997457504, + "learning_rate": 8.85477746464931e-06, + "loss": 0.0125, + "step": 36903 + }, + { + "epoch": 0.8650878766035502, + "grad_norm": 0.2694694995880127, + "learning_rate": 8.85174723090223e-06, + "loss": 0.0463, + "step": 36904 + }, + { + "epoch": 0.8651113181783552, + "grad_norm": 0.46735095977783203, + "learning_rate": 8.848717491728154e-06, + "loss": 0.6004, + "step": 36905 + }, + { + "epoch": 0.8651347597531602, + "grad_norm": 0.5520772933959961, + "learning_rate": 8.845688247143492e-06, + "loss": 0.1934, + "step": 36906 + }, + { + "epoch": 0.8651582013279652, + "grad_norm": 0.3831876516342163, + "learning_rate": 8.842659497164684e-06, + "loss": 0.0688, + "step": 36907 + }, + { + "epoch": 0.8651816429027702, + "grad_norm": 0.39394623041152954, + "learning_rate": 8.839631241808178e-06, + "loss": 0.0657, + "step": 36908 + }, + { + "epoch": 0.8652050844775752, + "grad_norm": 0.23072494566440582, + "learning_rate": 8.8366034810904e-06, + "loss": 0.0516, + "step": 36909 + }, + { + "epoch": 0.8652285260523802, + "grad_norm": 0.3115532696247101, + "learning_rate": 8.83357621502775e-06, + "loss": 0.0538, + "step": 36910 + }, + { + "epoch": 0.8652519676271851, + "grad_norm": 0.5280576944351196, + "learning_rate": 8.830549443636693e-06, + "loss": 0.1052, + "step": 36911 + }, + { + "epoch": 0.8652754092019902, + "grad_norm": 0.42289817333221436, + "learning_rate": 8.827523166933637e-06, + "loss": 0.0355, + "step": 36912 + }, + { + "epoch": 0.8652988507767951, + "grad_norm": 0.522426187992096, + "learning_rate": 8.824497384935016e-06, + "loss": 0.0889, + "step": 36913 + }, + { + "epoch": 0.8653222923516002, + "grad_norm": 0.5047671794891357, + "learning_rate": 8.821472097657224e-06, + "loss": 0.0948, + "step": 36914 + }, + { + "epoch": 0.8653457339264051, + "grad_norm": 0.32491064071655273, + "learning_rate": 8.818447305116672e-06, + "loss": 0.0319, + "step": 36915 + }, + { + "epoch": 0.8653691755012102, + "grad_norm": 0.45762601494789124, + "learning_rate": 8.815423007329793e-06, + "loss": 0.0572, + "step": 36916 + }, + { + "epoch": 0.8653926170760151, + "grad_norm": 0.7403847575187683, + "learning_rate": 8.812399204313015e-06, + "loss": 0.1089, + "step": 36917 + }, + { + "epoch": 0.8654160586508202, + "grad_norm": 0.4108617305755615, + "learning_rate": 8.809375896082717e-06, + "loss": 0.0651, + "step": 36918 + }, + { + "epoch": 0.8654395002256251, + "grad_norm": 0.5535673499107361, + "learning_rate": 8.806353082655305e-06, + "loss": 0.1075, + "step": 36919 + }, + { + "epoch": 0.8654629418004302, + "grad_norm": 0.2867233455181122, + "learning_rate": 8.803330764047191e-06, + "loss": 0.0384, + "step": 36920 + }, + { + "epoch": 0.8654863833752352, + "grad_norm": 0.4552595019340515, + "learning_rate": 8.80030894027477e-06, + "loss": 0.0409, + "step": 36921 + }, + { + "epoch": 0.8655098249500401, + "grad_norm": 0.24735093116760254, + "learning_rate": 8.797287611354421e-06, + "loss": 0.0609, + "step": 36922 + }, + { + "epoch": 0.8655332665248452, + "grad_norm": 0.7278627157211304, + "learning_rate": 8.794266777302563e-06, + "loss": 0.1363, + "step": 36923 + }, + { + "epoch": 0.8655567080996501, + "grad_norm": 0.41325730085372925, + "learning_rate": 8.79124643813558e-06, + "loss": 0.049, + "step": 36924 + }, + { + "epoch": 0.8655801496744552, + "grad_norm": 0.19456170499324799, + "learning_rate": 8.788226593869864e-06, + "loss": 0.0166, + "step": 36925 + }, + { + "epoch": 0.8656035912492601, + "grad_norm": 0.1950424164533615, + "learning_rate": 8.785207244521797e-06, + "loss": 0.0264, + "step": 36926 + }, + { + "epoch": 0.8656270328240652, + "grad_norm": 0.3962370455265045, + "learning_rate": 8.782188390107737e-06, + "loss": 0.0507, + "step": 36927 + }, + { + "epoch": 0.8656504743988701, + "grad_norm": 0.5391606092453003, + "learning_rate": 8.779170030644102e-06, + "loss": 0.1001, + "step": 36928 + }, + { + "epoch": 0.8656739159736752, + "grad_norm": 0.32941219210624695, + "learning_rate": 8.776152166147245e-06, + "loss": 0.2546, + "step": 36929 + }, + { + "epoch": 0.8656973575484801, + "grad_norm": 0.1645997017621994, + "learning_rate": 8.773134796633564e-06, + "loss": 0.0259, + "step": 36930 + }, + { + "epoch": 0.8657207991232851, + "grad_norm": 0.3831172585487366, + "learning_rate": 8.770117922119414e-06, + "loss": 0.0422, + "step": 36931 + }, + { + "epoch": 0.8657442406980901, + "grad_norm": 0.41855305433273315, + "learning_rate": 8.76710154262118e-06, + "loss": 0.0721, + "step": 36932 + }, + { + "epoch": 0.8657676822728951, + "grad_norm": 0.5283374786376953, + "learning_rate": 8.764085658155209e-06, + "loss": 0.0217, + "step": 36933 + }, + { + "epoch": 0.8657911238477001, + "grad_norm": 0.44810396432876587, + "learning_rate": 8.761070268737882e-06, + "loss": 0.13, + "step": 36934 + }, + { + "epoch": 0.8658145654225051, + "grad_norm": 0.2349095195531845, + "learning_rate": 8.758055374385532e-06, + "loss": 0.0504, + "step": 36935 + }, + { + "epoch": 0.8658380069973101, + "grad_norm": 0.08909893035888672, + "learning_rate": 8.755040975114559e-06, + "loss": 0.0142, + "step": 36936 + }, + { + "epoch": 0.8658614485721151, + "grad_norm": 0.15265394747257233, + "learning_rate": 8.752027070941294e-06, + "loss": 0.0294, + "step": 36937 + }, + { + "epoch": 0.86588489014692, + "grad_norm": 0.29554009437561035, + "learning_rate": 8.7490136618821e-06, + "loss": 0.0426, + "step": 36938 + }, + { + "epoch": 0.8659083317217251, + "grad_norm": 0.45783451199531555, + "learning_rate": 8.746000747953309e-06, + "loss": 0.0737, + "step": 36939 + }, + { + "epoch": 0.86593177329653, + "grad_norm": 0.16960518062114716, + "learning_rate": 8.742988329171297e-06, + "loss": 0.0184, + "step": 36940 + }, + { + "epoch": 0.8659552148713351, + "grad_norm": 0.6250306367874146, + "learning_rate": 8.739976405552398e-06, + "loss": 0.0695, + "step": 36941 + }, + { + "epoch": 0.86597865644614, + "grad_norm": 0.4443645179271698, + "learning_rate": 8.736964977112926e-06, + "loss": 0.0903, + "step": 36942 + }, + { + "epoch": 0.8660020980209451, + "grad_norm": 0.5790375471115112, + "learning_rate": 8.733954043869285e-06, + "loss": 0.1291, + "step": 36943 + }, + { + "epoch": 0.86602553959575, + "grad_norm": 0.5902940630912781, + "learning_rate": 8.73094360583776e-06, + "loss": 0.0851, + "step": 36944 + }, + { + "epoch": 0.8660489811705551, + "grad_norm": 0.5354527235031128, + "learning_rate": 8.727933663034704e-06, + "loss": 0.0611, + "step": 36945 + }, + { + "epoch": 0.86607242274536, + "grad_norm": 0.6693971157073975, + "learning_rate": 8.724924215476448e-06, + "loss": 0.1334, + "step": 36946 + }, + { + "epoch": 0.866095864320165, + "grad_norm": 0.42652738094329834, + "learning_rate": 8.721915263179325e-06, + "loss": 0.1248, + "step": 36947 + }, + { + "epoch": 0.86611930589497, + "grad_norm": 0.510890781879425, + "learning_rate": 8.718906806159632e-06, + "loss": 0.1076, + "step": 36948 + }, + { + "epoch": 0.866142747469775, + "grad_norm": 0.5678269267082214, + "learning_rate": 8.715898844433734e-06, + "loss": 0.1147, + "step": 36949 + }, + { + "epoch": 0.86616618904458, + "grad_norm": 0.21483856439590454, + "learning_rate": 8.71289137801794e-06, + "loss": 0.0229, + "step": 36950 + }, + { + "epoch": 0.866189630619385, + "grad_norm": 0.5753536224365234, + "learning_rate": 8.709884406928549e-06, + "loss": 0.0537, + "step": 36951 + }, + { + "epoch": 0.86621307219419, + "grad_norm": 0.27005258202552795, + "learning_rate": 8.7068779311819e-06, + "loss": 0.0424, + "step": 36952 + }, + { + "epoch": 0.866236513768995, + "grad_norm": 0.5810830593109131, + "learning_rate": 8.703871950794306e-06, + "loss": 0.4972, + "step": 36953 + }, + { + "epoch": 0.8662599553438, + "grad_norm": 0.5544264316558838, + "learning_rate": 8.700866465782042e-06, + "loss": 0.0613, + "step": 36954 + }, + { + "epoch": 0.866283396918605, + "grad_norm": 0.17039383947849274, + "learning_rate": 8.697861476161472e-06, + "loss": 0.0474, + "step": 36955 + }, + { + "epoch": 0.86630683849341, + "grad_norm": 0.32092708349227905, + "learning_rate": 8.694856981948873e-06, + "loss": 0.31, + "step": 36956 + }, + { + "epoch": 0.866330280068215, + "grad_norm": 0.38841530680656433, + "learning_rate": 8.691852983160542e-06, + "loss": 0.0448, + "step": 36957 + }, + { + "epoch": 0.8663537216430199, + "grad_norm": 0.11227263510227203, + "learning_rate": 8.688849479812788e-06, + "loss": 0.0238, + "step": 36958 + }, + { + "epoch": 0.866377163217825, + "grad_norm": 0.2845226228237152, + "learning_rate": 8.6858464719219e-06, + "loss": 0.0303, + "step": 36959 + }, + { + "epoch": 0.8664006047926299, + "grad_norm": 0.3349775969982147, + "learning_rate": 8.68284395950416e-06, + "loss": 0.0802, + "step": 36960 + }, + { + "epoch": 0.866424046367435, + "grad_norm": 0.18322955071926117, + "learning_rate": 8.679841942575895e-06, + "loss": 0.0273, + "step": 36961 + }, + { + "epoch": 0.8664474879422399, + "grad_norm": 0.2095373123884201, + "learning_rate": 8.676840421153376e-06, + "loss": 0.0461, + "step": 36962 + }, + { + "epoch": 0.866470929517045, + "grad_norm": 0.1965428739786148, + "learning_rate": 8.673839395252892e-06, + "loss": 0.0413, + "step": 36963 + }, + { + "epoch": 0.8664943710918499, + "grad_norm": 0.5291723608970642, + "learning_rate": 8.670838864890696e-06, + "loss": 0.1082, + "step": 36964 + }, + { + "epoch": 0.866517812666655, + "grad_norm": 0.4431471526622772, + "learning_rate": 8.66783883008312e-06, + "loss": 0.0954, + "step": 36965 + }, + { + "epoch": 0.8665412542414599, + "grad_norm": 0.12199081480503082, + "learning_rate": 8.664839290846427e-06, + "loss": 0.0205, + "step": 36966 + }, + { + "epoch": 0.8665646958162649, + "grad_norm": 0.20107844471931458, + "learning_rate": 8.66184024719685e-06, + "loss": 0.0397, + "step": 36967 + }, + { + "epoch": 0.8665881373910699, + "grad_norm": 0.36522072553634644, + "learning_rate": 8.658841699150722e-06, + "loss": 0.0355, + "step": 36968 + }, + { + "epoch": 0.8666115789658749, + "grad_norm": 0.6119146943092346, + "learning_rate": 8.655843646724294e-06, + "loss": 0.0789, + "step": 36969 + }, + { + "epoch": 0.8666350205406799, + "grad_norm": 0.42817509174346924, + "learning_rate": 8.652846089933818e-06, + "loss": 0.0644, + "step": 36970 + }, + { + "epoch": 0.8666584621154849, + "grad_norm": 0.3353953957557678, + "learning_rate": 8.649849028795565e-06, + "loss": 0.0356, + "step": 36971 + }, + { + "epoch": 0.86668190369029, + "grad_norm": 0.5009434819221497, + "learning_rate": 8.646852463325805e-06, + "loss": 0.0877, + "step": 36972 + }, + { + "epoch": 0.8667053452650949, + "grad_norm": 0.12065409123897552, + "learning_rate": 8.643856393540772e-06, + "loss": 0.017, + "step": 36973 + }, + { + "epoch": 0.8667287868399, + "grad_norm": 0.07363012433052063, + "learning_rate": 8.640860819456764e-06, + "loss": 0.0102, + "step": 36974 + }, + { + "epoch": 0.8667522284147049, + "grad_norm": 0.4123449921607971, + "learning_rate": 8.637865741090012e-06, + "loss": 0.0433, + "step": 36975 + }, + { + "epoch": 0.86677566998951, + "grad_norm": 0.4766250252723694, + "learning_rate": 8.634871158456747e-06, + "loss": 0.1076, + "step": 36976 + }, + { + "epoch": 0.8667991115643149, + "grad_norm": 0.12382844090461731, + "learning_rate": 8.63187707157327e-06, + "loss": 0.0182, + "step": 36977 + }, + { + "epoch": 0.8668225531391199, + "grad_norm": 0.6599447131156921, + "learning_rate": 8.628883480455785e-06, + "loss": 0.0777, + "step": 36978 + }, + { + "epoch": 0.8668459947139249, + "grad_norm": 0.5963013172149658, + "learning_rate": 8.625890385120549e-06, + "loss": 0.1818, + "step": 36979 + }, + { + "epoch": 0.8668694362887299, + "grad_norm": 0.4903249740600586, + "learning_rate": 8.622897785583784e-06, + "loss": 0.094, + "step": 36980 + }, + { + "epoch": 0.8668928778635349, + "grad_norm": 0.39926454424858093, + "learning_rate": 8.619905681861762e-06, + "loss": 0.0708, + "step": 36981 + }, + { + "epoch": 0.8669163194383399, + "grad_norm": 0.5652291774749756, + "learning_rate": 8.616914073970705e-06, + "loss": 0.1184, + "step": 36982 + }, + { + "epoch": 0.8669397610131448, + "grad_norm": 0.9031699299812317, + "learning_rate": 8.613922961926845e-06, + "loss": 0.15, + "step": 36983 + }, + { + "epoch": 0.8669632025879499, + "grad_norm": 0.0909004956483841, + "learning_rate": 8.610932345746403e-06, + "loss": 0.0207, + "step": 36984 + }, + { + "epoch": 0.8669866441627548, + "grad_norm": 0.6892582774162292, + "learning_rate": 8.607942225445597e-06, + "loss": 0.1183, + "step": 36985 + }, + { + "epoch": 0.8670100857375599, + "grad_norm": 0.45128437876701355, + "learning_rate": 8.604952601040683e-06, + "loss": 0.0641, + "step": 36986 + }, + { + "epoch": 0.8670335273123648, + "grad_norm": 0.9726346135139465, + "learning_rate": 8.601963472547859e-06, + "loss": 0.1922, + "step": 36987 + }, + { + "epoch": 0.8670569688871699, + "grad_norm": 0.5042247176170349, + "learning_rate": 8.598974839983342e-06, + "loss": 0.4035, + "step": 36988 + }, + { + "epoch": 0.8670804104619748, + "grad_norm": 0.362070232629776, + "learning_rate": 8.595986703363368e-06, + "loss": 0.0528, + "step": 36989 + }, + { + "epoch": 0.8671038520367799, + "grad_norm": 0.19806566834449768, + "learning_rate": 8.592999062704143e-06, + "loss": 0.0299, + "step": 36990 + }, + { + "epoch": 0.8671272936115848, + "grad_norm": 0.3230719268321991, + "learning_rate": 8.590011918021878e-06, + "loss": 0.0299, + "step": 36991 + }, + { + "epoch": 0.8671507351863899, + "grad_norm": 0.8811330199241638, + "learning_rate": 8.58702526933276e-06, + "loss": 0.0711, + "step": 36992 + }, + { + "epoch": 0.8671741767611948, + "grad_norm": 0.29207006096839905, + "learning_rate": 8.584039116653031e-06, + "loss": 0.0405, + "step": 36993 + }, + { + "epoch": 0.8671976183359998, + "grad_norm": 0.14878171682357788, + "learning_rate": 8.581053459998879e-06, + "loss": 0.0143, + "step": 36994 + }, + { + "epoch": 0.8672210599108048, + "grad_norm": 0.19480054080486298, + "learning_rate": 8.5780682993865e-06, + "loss": 0.0339, + "step": 36995 + }, + { + "epoch": 0.8672445014856098, + "grad_norm": 0.3174760639667511, + "learning_rate": 8.575083634832092e-06, + "loss": 0.0559, + "step": 36996 + }, + { + "epoch": 0.8672679430604148, + "grad_norm": 0.10687986016273499, + "learning_rate": 8.572099466351846e-06, + "loss": 0.0161, + "step": 36997 + }, + { + "epoch": 0.8672913846352198, + "grad_norm": 0.12110532075166702, + "learning_rate": 8.569115793961946e-06, + "loss": 0.0117, + "step": 36998 + }, + { + "epoch": 0.8673148262100248, + "grad_norm": 0.5125367045402527, + "learning_rate": 8.566132617678612e-06, + "loss": 0.0714, + "step": 36999 + }, + { + "epoch": 0.8673382677848298, + "grad_norm": 0.5095008015632629, + "learning_rate": 8.563149937517989e-06, + "loss": 0.0595, + "step": 37000 + }, + { + "epoch": 0.8673617093596347, + "grad_norm": 0.2991850972175598, + "learning_rate": 8.560167753496318e-06, + "loss": 0.0241, + "step": 37001 + }, + { + "epoch": 0.8673851509344398, + "grad_norm": 0.33771663904190063, + "learning_rate": 8.55718606562973e-06, + "loss": 0.0485, + "step": 37002 + }, + { + "epoch": 0.8674085925092447, + "grad_norm": 0.370790034532547, + "learning_rate": 8.554204873934435e-06, + "loss": 0.0546, + "step": 37003 + }, + { + "epoch": 0.8674320340840498, + "grad_norm": 0.3886294662952423, + "learning_rate": 8.551224178426587e-06, + "loss": 0.069, + "step": 37004 + }, + { + "epoch": 0.8674554756588547, + "grad_norm": 0.2314312905073166, + "learning_rate": 8.548243979122362e-06, + "loss": 0.0242, + "step": 37005 + }, + { + "epoch": 0.8674789172336598, + "grad_norm": 0.4221636950969696, + "learning_rate": 8.545264276037945e-06, + "loss": 0.0774, + "step": 37006 + }, + { + "epoch": 0.8675023588084647, + "grad_norm": 0.3555540144443512, + "learning_rate": 8.542285069189493e-06, + "loss": 0.0538, + "step": 37007 + }, + { + "epoch": 0.8675258003832698, + "grad_norm": 0.3288043439388275, + "learning_rate": 8.53930635859318e-06, + "loss": 0.054, + "step": 37008 + }, + { + "epoch": 0.8675492419580747, + "grad_norm": 0.15805259346961975, + "learning_rate": 8.53632814426516e-06, + "loss": 0.027, + "step": 37009 + }, + { + "epoch": 0.8675726835328798, + "grad_norm": 0.19123852252960205, + "learning_rate": 8.533350426221597e-06, + "loss": 0.029, + "step": 37010 + }, + { + "epoch": 0.8675961251076847, + "grad_norm": 0.137241929769516, + "learning_rate": 8.530373204478626e-06, + "loss": 0.0183, + "step": 37011 + }, + { + "epoch": 0.8676195666824897, + "grad_norm": 0.35465550422668457, + "learning_rate": 8.527396479052419e-06, + "loss": 0.0498, + "step": 37012 + }, + { + "epoch": 0.8676430082572947, + "grad_norm": 0.39966830611228943, + "learning_rate": 8.524420249959152e-06, + "loss": 0.0754, + "step": 37013 + }, + { + "epoch": 0.8676664498320997, + "grad_norm": 0.328896164894104, + "learning_rate": 8.52144451721496e-06, + "loss": 0.0516, + "step": 37014 + }, + { + "epoch": 0.8676898914069047, + "grad_norm": 0.34492525458335876, + "learning_rate": 8.518469280835973e-06, + "loss": 0.0358, + "step": 37015 + }, + { + "epoch": 0.8677133329817097, + "grad_norm": 0.2335067093372345, + "learning_rate": 8.515494540838342e-06, + "loss": 0.0369, + "step": 37016 + }, + { + "epoch": 0.8677367745565147, + "grad_norm": 0.42571771144866943, + "learning_rate": 8.512520297238213e-06, + "loss": 0.0503, + "step": 37017 + }, + { + "epoch": 0.8677602161313197, + "grad_norm": 0.4343152344226837, + "learning_rate": 8.509546550051706e-06, + "loss": 0.059, + "step": 37018 + }, + { + "epoch": 0.8677836577061246, + "grad_norm": 0.8606798648834229, + "learning_rate": 8.506573299294996e-06, + "loss": 0.1584, + "step": 37019 + }, + { + "epoch": 0.8678070992809297, + "grad_norm": 0.5163941979408264, + "learning_rate": 8.50360054498418e-06, + "loss": 0.0594, + "step": 37020 + }, + { + "epoch": 0.8678305408557346, + "grad_norm": 0.23115651309490204, + "learning_rate": 8.500628287135404e-06, + "loss": 0.0531, + "step": 37021 + }, + { + "epoch": 0.8678539824305397, + "grad_norm": 0.37262699007987976, + "learning_rate": 8.497656525764797e-06, + "loss": 0.0438, + "step": 37022 + }, + { + "epoch": 0.8678774240053447, + "grad_norm": 0.17849482595920563, + "learning_rate": 8.494685260888458e-06, + "loss": 0.0226, + "step": 37023 + }, + { + "epoch": 0.8679008655801497, + "grad_norm": 0.707560658454895, + "learning_rate": 8.49171449252254e-06, + "loss": 0.1056, + "step": 37024 + }, + { + "epoch": 0.8679243071549547, + "grad_norm": 0.19732119143009186, + "learning_rate": 8.488744220683143e-06, + "loss": 0.034, + "step": 37025 + }, + { + "epoch": 0.8679477487297597, + "grad_norm": 0.2492022067308426, + "learning_rate": 8.485774445386407e-06, + "loss": 0.0329, + "step": 37026 + }, + { + "epoch": 0.8679711903045647, + "grad_norm": 0.47961312532424927, + "learning_rate": 8.482805166648433e-06, + "loss": 0.0839, + "step": 37027 + }, + { + "epoch": 0.8679946318793696, + "grad_norm": 0.34050452709198, + "learning_rate": 8.479836384485329e-06, + "loss": 0.0524, + "step": 37028 + }, + { + "epoch": 0.8680180734541747, + "grad_norm": 0.2723890244960785, + "learning_rate": 8.476868098913205e-06, + "loss": 0.036, + "step": 37029 + }, + { + "epoch": 0.8680415150289796, + "grad_norm": 0.5204792022705078, + "learning_rate": 8.473900309948158e-06, + "loss": 0.1024, + "step": 37030 + }, + { + "epoch": 0.8680649566037847, + "grad_norm": 0.33935418725013733, + "learning_rate": 8.470933017606309e-06, + "loss": 0.0717, + "step": 37031 + }, + { + "epoch": 0.8680883981785896, + "grad_norm": 0.5285207629203796, + "learning_rate": 8.467966221903756e-06, + "loss": 0.1474, + "step": 37032 + }, + { + "epoch": 0.8681118397533947, + "grad_norm": 0.32866883277893066, + "learning_rate": 8.464999922856597e-06, + "loss": 0.0523, + "step": 37033 + }, + { + "epoch": 0.8681352813281996, + "grad_norm": 0.19626033306121826, + "learning_rate": 8.46203412048091e-06, + "loss": 0.0346, + "step": 37034 + }, + { + "epoch": 0.8681587229030047, + "grad_norm": 0.8358954787254333, + "learning_rate": 8.459068814792814e-06, + "loss": 0.4729, + "step": 37035 + }, + { + "epoch": 0.8681821644778096, + "grad_norm": 0.23346057534217834, + "learning_rate": 8.456104005808352e-06, + "loss": 0.0433, + "step": 37036 + }, + { + "epoch": 0.8682056060526147, + "grad_norm": 0.5119355320930481, + "learning_rate": 8.453139693543654e-06, + "loss": 0.0806, + "step": 37037 + }, + { + "epoch": 0.8682290476274196, + "grad_norm": 0.6893850564956665, + "learning_rate": 8.45017587801481e-06, + "loss": 0.101, + "step": 37038 + }, + { + "epoch": 0.8682524892022246, + "grad_norm": 0.15211059153079987, + "learning_rate": 8.447212559237883e-06, + "loss": 0.0161, + "step": 37039 + }, + { + "epoch": 0.8682759307770296, + "grad_norm": 0.2048008292913437, + "learning_rate": 8.44424973722896e-06, + "loss": 0.0418, + "step": 37040 + }, + { + "epoch": 0.8682993723518346, + "grad_norm": 0.5403864979743958, + "learning_rate": 8.441287412004106e-06, + "loss": 0.0911, + "step": 37041 + }, + { + "epoch": 0.8683228139266396, + "grad_norm": 0.2946442663669586, + "learning_rate": 8.438325583579409e-06, + "loss": 0.0203, + "step": 37042 + }, + { + "epoch": 0.8683462555014446, + "grad_norm": 0.4136546552181244, + "learning_rate": 8.435364251970911e-06, + "loss": 0.0381, + "step": 37043 + }, + { + "epoch": 0.8683696970762496, + "grad_norm": 0.7809600830078125, + "learning_rate": 8.432403417194712e-06, + "loss": 0.0863, + "step": 37044 + }, + { + "epoch": 0.8683931386510546, + "grad_norm": 0.10617855191230774, + "learning_rate": 8.429443079266875e-06, + "loss": 0.0155, + "step": 37045 + }, + { + "epoch": 0.8684165802258595, + "grad_norm": 0.1838960498571396, + "learning_rate": 8.426483238203452e-06, + "loss": 0.0398, + "step": 37046 + }, + { + "epoch": 0.8684400218006646, + "grad_norm": 0.06955751031637192, + "learning_rate": 8.423523894020502e-06, + "loss": 0.0095, + "step": 37047 + }, + { + "epoch": 0.8684634633754695, + "grad_norm": 0.3840530216693878, + "learning_rate": 8.420565046734097e-06, + "loss": 0.0378, + "step": 37048 + }, + { + "epoch": 0.8684869049502746, + "grad_norm": 0.4439796805381775, + "learning_rate": 8.417606696360247e-06, + "loss": 0.0796, + "step": 37049 + }, + { + "epoch": 0.8685103465250795, + "grad_norm": 0.3895701766014099, + "learning_rate": 8.414648842915051e-06, + "loss": 0.0366, + "step": 37050 + }, + { + "epoch": 0.8685337880998846, + "grad_norm": 0.2703852653503418, + "learning_rate": 8.411691486414563e-06, + "loss": 0.0365, + "step": 37051 + }, + { + "epoch": 0.8685572296746895, + "grad_norm": 0.6652196049690247, + "learning_rate": 8.408734626874804e-06, + "loss": 0.1034, + "step": 37052 + }, + { + "epoch": 0.8685806712494946, + "grad_norm": 0.08617579936981201, + "learning_rate": 8.405778264311837e-06, + "loss": 0.01, + "step": 37053 + }, + { + "epoch": 0.8686041128242995, + "grad_norm": 0.5141445994377136, + "learning_rate": 8.402822398741684e-06, + "loss": 0.0681, + "step": 37054 + }, + { + "epoch": 0.8686275543991046, + "grad_norm": 0.44341573119163513, + "learning_rate": 8.399867030180398e-06, + "loss": 0.0845, + "step": 37055 + }, + { + "epoch": 0.8686509959739095, + "grad_norm": 0.5231159925460815, + "learning_rate": 8.396912158643977e-06, + "loss": 0.0601, + "step": 37056 + }, + { + "epoch": 0.8686744375487145, + "grad_norm": 0.26038873195648193, + "learning_rate": 8.39395778414852e-06, + "loss": 0.0486, + "step": 37057 + }, + { + "epoch": 0.8686978791235195, + "grad_norm": 0.20186908543109894, + "learning_rate": 8.391003906710016e-06, + "loss": 0.0371, + "step": 37058 + }, + { + "epoch": 0.8687213206983245, + "grad_norm": 0.33163660764694214, + "learning_rate": 8.388050526344505e-06, + "loss": 0.052, + "step": 37059 + }, + { + "epoch": 0.8687447622731295, + "grad_norm": 0.14566412568092346, + "learning_rate": 8.385097643068007e-06, + "loss": 0.024, + "step": 37060 + }, + { + "epoch": 0.8687682038479345, + "grad_norm": 0.323215126991272, + "learning_rate": 8.382145256896523e-06, + "loss": 0.0656, + "step": 37061 + }, + { + "epoch": 0.8687916454227395, + "grad_norm": 0.18144367635250092, + "learning_rate": 8.379193367846116e-06, + "loss": 0.0578, + "step": 37062 + }, + { + "epoch": 0.8688150869975445, + "grad_norm": 0.5237029790878296, + "learning_rate": 8.376241975932764e-06, + "loss": 0.1003, + "step": 37063 + }, + { + "epoch": 0.8688385285723494, + "grad_norm": 0.4044927954673767, + "learning_rate": 8.373291081172518e-06, + "loss": 0.042, + "step": 37064 + }, + { + "epoch": 0.8688619701471545, + "grad_norm": 0.5838087797164917, + "learning_rate": 8.370340683581368e-06, + "loss": 0.1403, + "step": 37065 + }, + { + "epoch": 0.8688854117219594, + "grad_norm": 0.4735978841781616, + "learning_rate": 8.367390783175333e-06, + "loss": 0.0593, + "step": 37066 + }, + { + "epoch": 0.8689088532967645, + "grad_norm": 0.41721704602241516, + "learning_rate": 8.364441379970411e-06, + "loss": 0.0488, + "step": 37067 + }, + { + "epoch": 0.8689322948715694, + "grad_norm": 0.14050689339637756, + "learning_rate": 8.361492473982579e-06, + "loss": 0.0239, + "step": 37068 + }, + { + "epoch": 0.8689557364463745, + "grad_norm": 0.2984437048435211, + "learning_rate": 8.358544065227891e-06, + "loss": 0.043, + "step": 37069 + }, + { + "epoch": 0.8689791780211794, + "grad_norm": 0.1419447958469391, + "learning_rate": 8.355596153722323e-06, + "loss": 0.0169, + "step": 37070 + }, + { + "epoch": 0.8690026195959845, + "grad_norm": 0.2045673131942749, + "learning_rate": 8.35264873948186e-06, + "loss": 0.0237, + "step": 37071 + }, + { + "epoch": 0.8690260611707894, + "grad_norm": 0.6016363501548767, + "learning_rate": 8.349701822522504e-06, + "loss": 0.0992, + "step": 37072 + }, + { + "epoch": 0.8690495027455944, + "grad_norm": 0.3734024167060852, + "learning_rate": 8.346755402860218e-06, + "loss": 0.0775, + "step": 37073 + }, + { + "epoch": 0.8690729443203995, + "grad_norm": 0.41344916820526123, + "learning_rate": 8.343809480511045e-06, + "loss": 0.0457, + "step": 37074 + }, + { + "epoch": 0.8690963858952044, + "grad_norm": 0.4584848880767822, + "learning_rate": 8.340864055490916e-06, + "loss": 0.0678, + "step": 37075 + }, + { + "epoch": 0.8691198274700095, + "grad_norm": 0.8274841904640198, + "learning_rate": 8.337919127815853e-06, + "loss": 0.0566, + "step": 37076 + }, + { + "epoch": 0.8691432690448144, + "grad_norm": 0.18933585286140442, + "learning_rate": 8.334974697501818e-06, + "loss": 0.0346, + "step": 37077 + }, + { + "epoch": 0.8691667106196195, + "grad_norm": 0.38732418417930603, + "learning_rate": 8.33203076456478e-06, + "loss": 0.0601, + "step": 37078 + }, + { + "epoch": 0.8691901521944244, + "grad_norm": 0.1013772115111351, + "learning_rate": 8.329087329020735e-06, + "loss": 0.0268, + "step": 37079 + }, + { + "epoch": 0.8692135937692295, + "grad_norm": 0.16147027909755707, + "learning_rate": 8.326144390885626e-06, + "loss": 0.0276, + "step": 37080 + }, + { + "epoch": 0.8692370353440344, + "grad_norm": 0.47195497155189514, + "learning_rate": 8.32320195017543e-06, + "loss": 0.1271, + "step": 37081 + }, + { + "epoch": 0.8692604769188395, + "grad_norm": 0.21765738725662231, + "learning_rate": 8.320260006906122e-06, + "loss": 0.0186, + "step": 37082 + }, + { + "epoch": 0.8692839184936444, + "grad_norm": 0.19709065556526184, + "learning_rate": 8.31731856109368e-06, + "loss": 0.0316, + "step": 37083 + }, + { + "epoch": 0.8693073600684494, + "grad_norm": 0.08960296213626862, + "learning_rate": 8.314377612754032e-06, + "loss": 0.0193, + "step": 37084 + }, + { + "epoch": 0.8693308016432544, + "grad_norm": 0.5472486615180969, + "learning_rate": 8.311437161903124e-06, + "loss": 0.5178, + "step": 37085 + }, + { + "epoch": 0.8693542432180594, + "grad_norm": 0.21484169363975525, + "learning_rate": 8.308497208556964e-06, + "loss": 0.0442, + "step": 37086 + }, + { + "epoch": 0.8693776847928644, + "grad_norm": 0.3202495276927948, + "learning_rate": 8.305557752731485e-06, + "loss": 0.0469, + "step": 37087 + }, + { + "epoch": 0.8694011263676694, + "grad_norm": 0.04962499812245369, + "learning_rate": 8.302618794442596e-06, + "loss": 0.0057, + "step": 37088 + }, + { + "epoch": 0.8694245679424744, + "grad_norm": 0.4086644947528839, + "learning_rate": 8.299680333706295e-06, + "loss": 0.0764, + "step": 37089 + }, + { + "epoch": 0.8694480095172794, + "grad_norm": 0.5881868004798889, + "learning_rate": 8.296742370538513e-06, + "loss": 0.0969, + "step": 37090 + }, + { + "epoch": 0.8694714510920843, + "grad_norm": 0.5739198923110962, + "learning_rate": 8.293804904955182e-06, + "loss": 0.068, + "step": 37091 + }, + { + "epoch": 0.8694948926668894, + "grad_norm": 0.1425773799419403, + "learning_rate": 8.290867936972246e-06, + "loss": 0.021, + "step": 37092 + }, + { + "epoch": 0.8695183342416943, + "grad_norm": 0.3976025879383087, + "learning_rate": 8.287931466605636e-06, + "loss": 0.0674, + "step": 37093 + }, + { + "epoch": 0.8695417758164994, + "grad_norm": 0.07811901718378067, + "learning_rate": 8.284995493871273e-06, + "loss": 0.0109, + "step": 37094 + }, + { + "epoch": 0.8695652173913043, + "grad_norm": 0.1345428228378296, + "learning_rate": 8.282060018785121e-06, + "loss": 0.0188, + "step": 37095 + }, + { + "epoch": 0.8695886589661094, + "grad_norm": 0.24913202226161957, + "learning_rate": 8.279125041363089e-06, + "loss": 0.0155, + "step": 37096 + }, + { + "epoch": 0.8696121005409143, + "grad_norm": 0.11416174471378326, + "learning_rate": 8.276190561621089e-06, + "loss": 0.026, + "step": 37097 + }, + { + "epoch": 0.8696355421157194, + "grad_norm": 0.523600161075592, + "learning_rate": 8.273256579575073e-06, + "loss": 0.5652, + "step": 37098 + }, + { + "epoch": 0.8696589836905243, + "grad_norm": 0.26391294598579407, + "learning_rate": 8.27032309524094e-06, + "loss": 0.041, + "step": 37099 + }, + { + "epoch": 0.8696824252653294, + "grad_norm": 0.4544416666030884, + "learning_rate": 8.26739010863462e-06, + "loss": 0.0896, + "step": 37100 + }, + { + "epoch": 0.8697058668401343, + "grad_norm": 0.590575098991394, + "learning_rate": 8.264457619772004e-06, + "loss": 0.0641, + "step": 37101 + }, + { + "epoch": 0.8697293084149393, + "grad_norm": 0.48569729924201965, + "learning_rate": 8.261525628669043e-06, + "loss": 0.0489, + "step": 37102 + }, + { + "epoch": 0.8697527499897443, + "grad_norm": 0.3261564075946808, + "learning_rate": 8.258594135341612e-06, + "loss": 0.0275, + "step": 37103 + }, + { + "epoch": 0.8697761915645493, + "grad_norm": 0.16369511187076569, + "learning_rate": 8.255663139805636e-06, + "loss": 0.0269, + "step": 37104 + }, + { + "epoch": 0.8697996331393543, + "grad_norm": 0.29459235072135925, + "learning_rate": 8.25273264207701e-06, + "loss": 0.0343, + "step": 37105 + }, + { + "epoch": 0.8698230747141593, + "grad_norm": 0.46420058608055115, + "learning_rate": 8.24980264217161e-06, + "loss": 0.1, + "step": 37106 + }, + { + "epoch": 0.8698465162889643, + "grad_norm": 0.6726618409156799, + "learning_rate": 8.246873140105394e-06, + "loss": 0.1333, + "step": 37107 + }, + { + "epoch": 0.8698699578637693, + "grad_norm": 0.4153188467025757, + "learning_rate": 8.24394413589421e-06, + "loss": 0.093, + "step": 37108 + }, + { + "epoch": 0.8698933994385742, + "grad_norm": 0.46584537625312805, + "learning_rate": 8.241015629553972e-06, + "loss": 0.049, + "step": 37109 + }, + { + "epoch": 0.8699168410133793, + "grad_norm": 0.09427589178085327, + "learning_rate": 8.238087621100543e-06, + "loss": 0.0169, + "step": 37110 + }, + { + "epoch": 0.8699402825881842, + "grad_norm": 0.36390405893325806, + "learning_rate": 8.235160110549856e-06, + "loss": 0.0616, + "step": 37111 + }, + { + "epoch": 0.8699637241629893, + "grad_norm": 0.5871526598930359, + "learning_rate": 8.232233097917763e-06, + "loss": 0.1229, + "step": 37112 + }, + { + "epoch": 0.8699871657377942, + "grad_norm": 0.4818679690361023, + "learning_rate": 8.22930658322013e-06, + "loss": 0.5539, + "step": 37113 + }, + { + "epoch": 0.8700106073125993, + "grad_norm": 0.3313566744327545, + "learning_rate": 8.22638056647289e-06, + "loss": 0.0495, + "step": 37114 + }, + { + "epoch": 0.8700340488874042, + "grad_norm": 0.31892359256744385, + "learning_rate": 8.223455047691886e-06, + "loss": 0.0466, + "step": 37115 + }, + { + "epoch": 0.8700574904622093, + "grad_norm": 0.23132501542568207, + "learning_rate": 8.220530026892992e-06, + "loss": 0.0363, + "step": 37116 + }, + { + "epoch": 0.8700809320370142, + "grad_norm": 0.48661941289901733, + "learning_rate": 8.217605504092085e-06, + "loss": 0.0689, + "step": 37117 + }, + { + "epoch": 0.8701043736118192, + "grad_norm": 0.656548261642456, + "learning_rate": 8.21468147930503e-06, + "loss": 0.0817, + "step": 37118 + }, + { + "epoch": 0.8701278151866242, + "grad_norm": 0.6502102613449097, + "learning_rate": 8.211757952547683e-06, + "loss": 0.569, + "step": 37119 + }, + { + "epoch": 0.8701512567614292, + "grad_norm": 0.682375431060791, + "learning_rate": 8.208834923835928e-06, + "loss": 0.1554, + "step": 37120 + }, + { + "epoch": 0.8701746983362342, + "grad_norm": 0.15968824923038483, + "learning_rate": 8.205912393185611e-06, + "loss": 0.0385, + "step": 37121 + }, + { + "epoch": 0.8701981399110392, + "grad_norm": 0.4957405924797058, + "learning_rate": 8.202990360612583e-06, + "loss": 0.1196, + "step": 37122 + }, + { + "epoch": 0.8702215814858442, + "grad_norm": 0.47132059931755066, + "learning_rate": 8.200068826132733e-06, + "loss": 0.3741, + "step": 37123 + }, + { + "epoch": 0.8702450230606492, + "grad_norm": 0.19011831283569336, + "learning_rate": 8.197147789761884e-06, + "loss": 0.0281, + "step": 37124 + }, + { + "epoch": 0.8702684646354543, + "grad_norm": 0.45591163635253906, + "learning_rate": 8.194227251515896e-06, + "loss": 0.1086, + "step": 37125 + }, + { + "epoch": 0.8702919062102592, + "grad_norm": 0.7835047245025635, + "learning_rate": 8.191307211410592e-06, + "loss": 0.117, + "step": 37126 + }, + { + "epoch": 0.8703153477850643, + "grad_norm": 0.09776873886585236, + "learning_rate": 8.18838766946186e-06, + "loss": 0.0159, + "step": 37127 + }, + { + "epoch": 0.8703387893598692, + "grad_norm": 0.1437552273273468, + "learning_rate": 8.18546862568551e-06, + "loss": 0.0159, + "step": 37128 + }, + { + "epoch": 0.8703622309346742, + "grad_norm": 0.27198484539985657, + "learning_rate": 8.182550080097396e-06, + "loss": 0.0457, + "step": 37129 + }, + { + "epoch": 0.8703856725094792, + "grad_norm": 0.30146345496177673, + "learning_rate": 8.179632032713335e-06, + "loss": 0.0843, + "step": 37130 + }, + { + "epoch": 0.8704091140842842, + "grad_norm": 0.3713899254798889, + "learning_rate": 8.176714483549186e-06, + "loss": 0.0467, + "step": 37131 + }, + { + "epoch": 0.8704325556590892, + "grad_norm": 0.1999761462211609, + "learning_rate": 8.173797432620744e-06, + "loss": 0.0403, + "step": 37132 + }, + { + "epoch": 0.8704559972338942, + "grad_norm": 1.4219536781311035, + "learning_rate": 8.170880879943876e-06, + "loss": 0.0863, + "step": 37133 + }, + { + "epoch": 0.8704794388086992, + "grad_norm": 0.180633082985878, + "learning_rate": 8.167964825534369e-06, + "loss": 0.0435, + "step": 37134 + }, + { + "epoch": 0.8705028803835042, + "grad_norm": 0.6383451223373413, + "learning_rate": 8.165049269408087e-06, + "loss": 0.1696, + "step": 37135 + }, + { + "epoch": 0.8705263219583091, + "grad_norm": 1.0202044248580933, + "learning_rate": 8.16213421158083e-06, + "loss": 0.0383, + "step": 37136 + }, + { + "epoch": 0.8705497635331142, + "grad_norm": 0.37149110436439514, + "learning_rate": 8.159219652068418e-06, + "loss": 0.0373, + "step": 37137 + }, + { + "epoch": 0.8705732051079191, + "grad_norm": 0.5344756245613098, + "learning_rate": 8.15630559088666e-06, + "loss": 0.0775, + "step": 37138 + }, + { + "epoch": 0.8705966466827242, + "grad_norm": 0.574984610080719, + "learning_rate": 8.153392028051355e-06, + "loss": 0.1766, + "step": 37139 + }, + { + "epoch": 0.8706200882575291, + "grad_norm": 0.18492621183395386, + "learning_rate": 8.150478963578346e-06, + "loss": 0.0259, + "step": 37140 + }, + { + "epoch": 0.8706435298323342, + "grad_norm": 0.3818351626396179, + "learning_rate": 8.147566397483409e-06, + "loss": 0.0774, + "step": 37141 + }, + { + "epoch": 0.8706669714071391, + "grad_norm": 0.46039044857025146, + "learning_rate": 8.144654329782375e-06, + "loss": 0.0902, + "step": 37142 + }, + { + "epoch": 0.8706904129819442, + "grad_norm": 0.4795841872692108, + "learning_rate": 8.141742760491022e-06, + "loss": 0.6018, + "step": 37143 + }, + { + "epoch": 0.8707138545567491, + "grad_norm": 0.36751988530158997, + "learning_rate": 8.138831689625137e-06, + "loss": 0.0739, + "step": 37144 + }, + { + "epoch": 0.8707372961315542, + "grad_norm": 0.33031773567199707, + "learning_rate": 8.135921117200551e-06, + "loss": 0.0741, + "step": 37145 + }, + { + "epoch": 0.8707607377063591, + "grad_norm": 0.3385672867298126, + "learning_rate": 8.133011043233019e-06, + "loss": 0.0589, + "step": 37146 + }, + { + "epoch": 0.8707841792811641, + "grad_norm": 0.35173362493515015, + "learning_rate": 8.130101467738383e-06, + "loss": 0.0257, + "step": 37147 + }, + { + "epoch": 0.8708076208559691, + "grad_norm": 0.35922515392303467, + "learning_rate": 8.127192390732386e-06, + "loss": 0.0147, + "step": 37148 + }, + { + "epoch": 0.8708310624307741, + "grad_norm": 0.39166635274887085, + "learning_rate": 8.124283812230838e-06, + "loss": 0.0809, + "step": 37149 + }, + { + "epoch": 0.8708545040055791, + "grad_norm": 0.11114644259214401, + "learning_rate": 8.121375732249503e-06, + "loss": 0.0255, + "step": 37150 + }, + { + "epoch": 0.8708779455803841, + "grad_norm": 0.5506593585014343, + "learning_rate": 8.118468150804148e-06, + "loss": 0.0812, + "step": 37151 + }, + { + "epoch": 0.870901387155189, + "grad_norm": 0.5226848125457764, + "learning_rate": 8.115561067910583e-06, + "loss": 0.1389, + "step": 37152 + }, + { + "epoch": 0.8709248287299941, + "grad_norm": 0.5192579030990601, + "learning_rate": 8.11265448358457e-06, + "loss": 0.4858, + "step": 37153 + }, + { + "epoch": 0.870948270304799, + "grad_norm": 0.19184346497058868, + "learning_rate": 8.109748397841888e-06, + "loss": 0.0259, + "step": 37154 + }, + { + "epoch": 0.8709717118796041, + "grad_norm": 0.1414073258638382, + "learning_rate": 8.106842810698278e-06, + "loss": 0.0256, + "step": 37155 + }, + { + "epoch": 0.870995153454409, + "grad_norm": 0.1799813210964203, + "learning_rate": 8.10393772216953e-06, + "loss": 0.0267, + "step": 37156 + }, + { + "epoch": 0.8710185950292141, + "grad_norm": 0.37696781754493713, + "learning_rate": 8.101033132271385e-06, + "loss": 0.0355, + "step": 37157 + }, + { + "epoch": 0.871042036604019, + "grad_norm": 0.3356599509716034, + "learning_rate": 8.09812904101962e-06, + "loss": 0.0594, + "step": 37158 + }, + { + "epoch": 0.8710654781788241, + "grad_norm": 0.7915602922439575, + "learning_rate": 8.09522544843e-06, + "loss": 0.6263, + "step": 37159 + }, + { + "epoch": 0.871088919753629, + "grad_norm": 0.3667466342449188, + "learning_rate": 8.092322354518278e-06, + "loss": 0.0678, + "step": 37160 + }, + { + "epoch": 0.8711123613284341, + "grad_norm": 0.2350068986415863, + "learning_rate": 8.089419759300198e-06, + "loss": 0.0349, + "step": 37161 + }, + { + "epoch": 0.871135802903239, + "grad_norm": 0.3570092022418976, + "learning_rate": 8.086517662791503e-06, + "loss": 0.0465, + "step": 37162 + }, + { + "epoch": 0.871159244478044, + "grad_norm": 0.408785343170166, + "learning_rate": 8.08361606500796e-06, + "loss": 0.0877, + "step": 37163 + }, + { + "epoch": 0.871182686052849, + "grad_norm": 0.17986328899860382, + "learning_rate": 8.080714965965274e-06, + "loss": 0.016, + "step": 37164 + }, + { + "epoch": 0.871206127627654, + "grad_norm": 0.623643696308136, + "learning_rate": 8.077814365679237e-06, + "loss": 0.6699, + "step": 37165 + }, + { + "epoch": 0.871229569202459, + "grad_norm": 0.4569076597690582, + "learning_rate": 8.074914264165556e-06, + "loss": 0.0735, + "step": 37166 + }, + { + "epoch": 0.871253010777264, + "grad_norm": 0.3970988094806671, + "learning_rate": 8.072014661439987e-06, + "loss": 0.0323, + "step": 37167 + }, + { + "epoch": 0.871276452352069, + "grad_norm": 0.37830063700675964, + "learning_rate": 8.069115557518236e-06, + "loss": 0.0555, + "step": 37168 + }, + { + "epoch": 0.871299893926874, + "grad_norm": 0.5872248411178589, + "learning_rate": 8.066216952416061e-06, + "loss": 0.509, + "step": 37169 + }, + { + "epoch": 0.871323335501679, + "grad_norm": 0.13572630286216736, + "learning_rate": 8.063318846149159e-06, + "loss": 0.0164, + "step": 37170 + }, + { + "epoch": 0.871346777076484, + "grad_norm": 0.5049492716789246, + "learning_rate": 8.060421238733262e-06, + "loss": 0.0539, + "step": 37171 + }, + { + "epoch": 0.8713702186512889, + "grad_norm": 0.18651187419891357, + "learning_rate": 8.057524130184135e-06, + "loss": 0.0308, + "step": 37172 + }, + { + "epoch": 0.871393660226094, + "grad_norm": 0.4629312753677368, + "learning_rate": 8.054627520517467e-06, + "loss": 0.4956, + "step": 37173 + }, + { + "epoch": 0.8714171018008989, + "grad_norm": 0.38842615485191345, + "learning_rate": 8.051731409748964e-06, + "loss": 0.0653, + "step": 37174 + }, + { + "epoch": 0.871440543375704, + "grad_norm": 0.43402165174484253, + "learning_rate": 8.048835797894361e-06, + "loss": 0.0594, + "step": 37175 + }, + { + "epoch": 0.871463984950509, + "grad_norm": 0.2780461013317108, + "learning_rate": 8.045940684969356e-06, + "loss": 0.0376, + "step": 37176 + }, + { + "epoch": 0.871487426525314, + "grad_norm": 0.3454720079898834, + "learning_rate": 8.043046070989647e-06, + "loss": 0.0718, + "step": 37177 + }, + { + "epoch": 0.871510868100119, + "grad_norm": 0.08216257393360138, + "learning_rate": 8.040151955970977e-06, + "loss": 0.0104, + "step": 37178 + }, + { + "epoch": 0.871534309674924, + "grad_norm": 0.5938512682914734, + "learning_rate": 8.037258339929021e-06, + "loss": 0.5337, + "step": 37179 + }, + { + "epoch": 0.871557751249729, + "grad_norm": 0.4407544732093811, + "learning_rate": 8.034365222879491e-06, + "loss": 0.0652, + "step": 37180 + }, + { + "epoch": 0.871581192824534, + "grad_norm": 0.3825191557407379, + "learning_rate": 8.031472604838087e-06, + "loss": 0.0823, + "step": 37181 + }, + { + "epoch": 0.871604634399339, + "grad_norm": 0.7300187349319458, + "learning_rate": 8.02858048582047e-06, + "loss": 0.1279, + "step": 37182 + }, + { + "epoch": 0.8716280759741439, + "grad_norm": 0.3043626844882965, + "learning_rate": 8.025688865842385e-06, + "loss": 0.0739, + "step": 37183 + }, + { + "epoch": 0.871651517548949, + "grad_norm": 0.5078951120376587, + "learning_rate": 8.022797744919475e-06, + "loss": 0.0974, + "step": 37184 + }, + { + "epoch": 0.8716749591237539, + "grad_norm": 0.2938925623893738, + "learning_rate": 8.019907123067482e-06, + "loss": 0.0623, + "step": 37185 + }, + { + "epoch": 0.871698400698559, + "grad_norm": 0.5330824255943298, + "learning_rate": 8.017017000302041e-06, + "loss": 0.114, + "step": 37186 + }, + { + "epoch": 0.8717218422733639, + "grad_norm": 0.44619104266166687, + "learning_rate": 8.014127376638869e-06, + "loss": 0.099, + "step": 37187 + }, + { + "epoch": 0.871745283848169, + "grad_norm": 0.21250203251838684, + "learning_rate": 8.011238252093611e-06, + "loss": 0.023, + "step": 37188 + }, + { + "epoch": 0.8717687254229739, + "grad_norm": 0.18316024541854858, + "learning_rate": 8.008349626681954e-06, + "loss": 0.0322, + "step": 37189 + }, + { + "epoch": 0.871792166997779, + "grad_norm": 0.08241678029298782, + "learning_rate": 8.005461500419598e-06, + "loss": 0.0098, + "step": 37190 + }, + { + "epoch": 0.8718156085725839, + "grad_norm": 0.22379620373249054, + "learning_rate": 8.002573873322195e-06, + "loss": 0.0376, + "step": 37191 + }, + { + "epoch": 0.8718390501473889, + "grad_norm": 0.0980898067355156, + "learning_rate": 7.9996867454054e-06, + "loss": 0.0134, + "step": 37192 + }, + { + "epoch": 0.8718624917221939, + "grad_norm": 0.5853103399276733, + "learning_rate": 7.9968001166849e-06, + "loss": 0.6277, + "step": 37193 + }, + { + "epoch": 0.8718859332969989, + "grad_norm": 0.32139700651168823, + "learning_rate": 7.99391398717635e-06, + "loss": 0.049, + "step": 37194 + }, + { + "epoch": 0.8719093748718039, + "grad_norm": 0.7569466829299927, + "learning_rate": 7.991028356895391e-06, + "loss": 0.525, + "step": 37195 + }, + { + "epoch": 0.8719328164466089, + "grad_norm": 0.2918246388435364, + "learning_rate": 7.988143225857703e-06, + "loss": 0.0452, + "step": 37196 + }, + { + "epoch": 0.8719562580214139, + "grad_norm": 0.5834864377975464, + "learning_rate": 7.985258594078948e-06, + "loss": 0.1136, + "step": 37197 + }, + { + "epoch": 0.8719796995962189, + "grad_norm": 0.33107101917266846, + "learning_rate": 7.98237446157477e-06, + "loss": 0.0287, + "step": 37198 + }, + { + "epoch": 0.8720031411710238, + "grad_norm": 0.19551850855350494, + "learning_rate": 7.979490828360826e-06, + "loss": 0.0245, + "step": 37199 + }, + { + "epoch": 0.8720265827458289, + "grad_norm": 0.4134233295917511, + "learning_rate": 7.976607694452742e-06, + "loss": 0.0853, + "step": 37200 + }, + { + "epoch": 0.8720500243206338, + "grad_norm": 0.14653009176254272, + "learning_rate": 7.973725059866166e-06, + "loss": 0.0268, + "step": 37201 + }, + { + "epoch": 0.8720734658954389, + "grad_norm": 0.44783857464790344, + "learning_rate": 7.970842924616739e-06, + "loss": 0.0446, + "step": 37202 + }, + { + "epoch": 0.8720969074702438, + "grad_norm": 0.3529803454875946, + "learning_rate": 7.967961288720117e-06, + "loss": 0.0647, + "step": 37203 + }, + { + "epoch": 0.8721203490450489, + "grad_norm": 0.4354826509952545, + "learning_rate": 7.965080152191929e-06, + "loss": 0.0376, + "step": 37204 + }, + { + "epoch": 0.8721437906198538, + "grad_norm": 0.5219693779945374, + "learning_rate": 7.962199515047796e-06, + "loss": 0.3808, + "step": 37205 + }, + { + "epoch": 0.8721672321946589, + "grad_norm": 0.18864379823207855, + "learning_rate": 7.959319377303364e-06, + "loss": 0.018, + "step": 37206 + }, + { + "epoch": 0.8721906737694638, + "grad_norm": 0.5081382393836975, + "learning_rate": 7.95643973897423e-06, + "loss": 0.6746, + "step": 37207 + }, + { + "epoch": 0.8722141153442688, + "grad_norm": 0.4773108959197998, + "learning_rate": 7.953560600076072e-06, + "loss": 0.1079, + "step": 37208 + }, + { + "epoch": 0.8722375569190738, + "grad_norm": 0.40991559624671936, + "learning_rate": 7.950681960624451e-06, + "loss": 0.0598, + "step": 37209 + }, + { + "epoch": 0.8722609984938788, + "grad_norm": 0.38027194142341614, + "learning_rate": 7.947803820635035e-06, + "loss": 0.069, + "step": 37210 + }, + { + "epoch": 0.8722844400686838, + "grad_norm": 0.46969178318977356, + "learning_rate": 7.944926180123435e-06, + "loss": 0.1329, + "step": 37211 + }, + { + "epoch": 0.8723078816434888, + "grad_norm": 0.22772984206676483, + "learning_rate": 7.942049039105248e-06, + "loss": 0.0409, + "step": 37212 + }, + { + "epoch": 0.8723313232182938, + "grad_norm": 0.3995530605316162, + "learning_rate": 7.939172397596085e-06, + "loss": 0.0562, + "step": 37213 + }, + { + "epoch": 0.8723547647930988, + "grad_norm": 0.10993510484695435, + "learning_rate": 7.936296255611575e-06, + "loss": 0.0145, + "step": 37214 + }, + { + "epoch": 0.8723782063679038, + "grad_norm": 0.6552298665046692, + "learning_rate": 7.933420613167287e-06, + "loss": 0.1147, + "step": 37215 + }, + { + "epoch": 0.8724016479427088, + "grad_norm": 0.09969601780176163, + "learning_rate": 7.93054547027886e-06, + "loss": 0.0164, + "step": 37216 + }, + { + "epoch": 0.8724250895175137, + "grad_norm": 0.30195945501327515, + "learning_rate": 7.927670826961887e-06, + "loss": 0.0167, + "step": 37217 + }, + { + "epoch": 0.8724485310923188, + "grad_norm": 0.6262759566307068, + "learning_rate": 7.92479668323195e-06, + "loss": 0.6423, + "step": 37218 + }, + { + "epoch": 0.8724719726671237, + "grad_norm": 0.6068996787071228, + "learning_rate": 7.92192303910465e-06, + "loss": 0.0529, + "step": 37219 + }, + { + "epoch": 0.8724954142419288, + "grad_norm": 0.31988272070884705, + "learning_rate": 7.919049894595598e-06, + "loss": 0.0536, + "step": 37220 + }, + { + "epoch": 0.8725188558167337, + "grad_norm": 0.359139621257782, + "learning_rate": 7.916177249720357e-06, + "loss": 0.0535, + "step": 37221 + }, + { + "epoch": 0.8725422973915388, + "grad_norm": 0.40494564175605774, + "learning_rate": 7.91330510449454e-06, + "loss": 0.0573, + "step": 37222 + }, + { + "epoch": 0.8725657389663437, + "grad_norm": 0.5331653356552124, + "learning_rate": 7.91043345893372e-06, + "loss": 0.0722, + "step": 37223 + }, + { + "epoch": 0.8725891805411488, + "grad_norm": 0.5340524315834045, + "learning_rate": 7.907562313053473e-06, + "loss": 0.6046, + "step": 37224 + }, + { + "epoch": 0.8726126221159537, + "grad_norm": 0.31400227546691895, + "learning_rate": 7.904691666869391e-06, + "loss": 0.047, + "step": 37225 + }, + { + "epoch": 0.8726360636907587, + "grad_norm": 0.3169269263744354, + "learning_rate": 7.901821520397034e-06, + "loss": 0.0657, + "step": 37226 + }, + { + "epoch": 0.8726595052655638, + "grad_norm": 0.191063791513443, + "learning_rate": 7.898951873651972e-06, + "loss": 0.0126, + "step": 37227 + }, + { + "epoch": 0.8726829468403687, + "grad_norm": 0.5959331393241882, + "learning_rate": 7.8960827266498e-06, + "loss": 0.5503, + "step": 37228 + }, + { + "epoch": 0.8727063884151738, + "grad_norm": 0.12909600138664246, + "learning_rate": 7.893214079406075e-06, + "loss": 0.0093, + "step": 37229 + }, + { + "epoch": 0.8727298299899787, + "grad_norm": 0.1110740676522255, + "learning_rate": 7.89034593193635e-06, + "loss": 0.0087, + "step": 37230 + }, + { + "epoch": 0.8727532715647838, + "grad_norm": 0.37524518370628357, + "learning_rate": 7.887478284256188e-06, + "loss": 0.0485, + "step": 37231 + }, + { + "epoch": 0.8727767131395887, + "grad_norm": 0.2976346015930176, + "learning_rate": 7.884611136381182e-06, + "loss": 0.0699, + "step": 37232 + }, + { + "epoch": 0.8728001547143938, + "grad_norm": 0.5084398984909058, + "learning_rate": 7.881744488326859e-06, + "loss": 0.0411, + "step": 37233 + }, + { + "epoch": 0.8728235962891987, + "grad_norm": 0.6936661601066589, + "learning_rate": 7.878878340108765e-06, + "loss": 0.1107, + "step": 37234 + }, + { + "epoch": 0.8728470378640037, + "grad_norm": 0.09059912711381912, + "learning_rate": 7.876012691742485e-06, + "loss": 0.0084, + "step": 37235 + }, + { + "epoch": 0.8728704794388087, + "grad_norm": 0.32909178733825684, + "learning_rate": 7.873147543243554e-06, + "loss": 0.059, + "step": 37236 + }, + { + "epoch": 0.8728939210136137, + "grad_norm": 0.34709516167640686, + "learning_rate": 7.870282894627512e-06, + "loss": 0.0555, + "step": 37237 + }, + { + "epoch": 0.8729173625884187, + "grad_norm": 0.8754006028175354, + "learning_rate": 7.867418745909905e-06, + "loss": 0.1707, + "step": 37238 + }, + { + "epoch": 0.8729408041632237, + "grad_norm": 0.3327626585960388, + "learning_rate": 7.864555097106274e-06, + "loss": 0.0476, + "step": 37239 + }, + { + "epoch": 0.8729642457380287, + "grad_norm": 0.6038399934768677, + "learning_rate": 7.861691948232141e-06, + "loss": 0.087, + "step": 37240 + }, + { + "epoch": 0.8729876873128337, + "grad_norm": 0.45165467262268066, + "learning_rate": 7.858829299303072e-06, + "loss": 0.0676, + "step": 37241 + }, + { + "epoch": 0.8730111288876387, + "grad_norm": 0.8159343600273132, + "learning_rate": 7.855967150334598e-06, + "loss": 0.6386, + "step": 37242 + }, + { + "epoch": 0.8730345704624437, + "grad_norm": 0.20401765406131744, + "learning_rate": 7.853105501342206e-06, + "loss": 0.0218, + "step": 37243 + }, + { + "epoch": 0.8730580120372486, + "grad_norm": 0.9224815368652344, + "learning_rate": 7.850244352341484e-06, + "loss": 0.1305, + "step": 37244 + }, + { + "epoch": 0.8730814536120537, + "grad_norm": 0.3474165201187134, + "learning_rate": 7.84738370334791e-06, + "loss": 0.0635, + "step": 37245 + }, + { + "epoch": 0.8731048951868586, + "grad_norm": 0.20281505584716797, + "learning_rate": 7.844523554377037e-06, + "loss": 0.056, + "step": 37246 + }, + { + "epoch": 0.8731283367616637, + "grad_norm": 0.5758538246154785, + "learning_rate": 7.841663905444351e-06, + "loss": 0.1238, + "step": 37247 + }, + { + "epoch": 0.8731517783364686, + "grad_norm": 0.29222479462623596, + "learning_rate": 7.838804756565398e-06, + "loss": 0.0699, + "step": 37248 + }, + { + "epoch": 0.8731752199112737, + "grad_norm": 0.44869479537010193, + "learning_rate": 7.835946107755698e-06, + "loss": 0.079, + "step": 37249 + }, + { + "epoch": 0.8731986614860786, + "grad_norm": 0.09473913908004761, + "learning_rate": 7.833087959030727e-06, + "loss": 0.0183, + "step": 37250 + }, + { + "epoch": 0.8732221030608837, + "grad_norm": 0.29553067684173584, + "learning_rate": 7.830230310406029e-06, + "loss": 0.037, + "step": 37251 + }, + { + "epoch": 0.8732455446356886, + "grad_norm": 0.16849733889102936, + "learning_rate": 7.827373161897078e-06, + "loss": 0.0407, + "step": 37252 + }, + { + "epoch": 0.8732689862104936, + "grad_norm": 0.6764296293258667, + "learning_rate": 7.824516513519387e-06, + "loss": 0.6658, + "step": 37253 + }, + { + "epoch": 0.8732924277852986, + "grad_norm": 0.30706366896629333, + "learning_rate": 7.821660365288463e-06, + "loss": 0.0597, + "step": 37254 + }, + { + "epoch": 0.8733158693601036, + "grad_norm": 0.36452436447143555, + "learning_rate": 7.818804717219796e-06, + "loss": 0.0616, + "step": 37255 + }, + { + "epoch": 0.8733393109349086, + "grad_norm": 0.7735652327537537, + "learning_rate": 7.815949569328906e-06, + "loss": 0.0547, + "step": 37256 + }, + { + "epoch": 0.8733627525097136, + "grad_norm": 0.16769468784332275, + "learning_rate": 7.81309492163126e-06, + "loss": 0.0465, + "step": 37257 + }, + { + "epoch": 0.8733861940845186, + "grad_norm": 0.38132721185684204, + "learning_rate": 7.810240774142352e-06, + "loss": 0.045, + "step": 37258 + }, + { + "epoch": 0.8734096356593236, + "grad_norm": 0.3436175584793091, + "learning_rate": 7.807387126877652e-06, + "loss": 0.0248, + "step": 37259 + }, + { + "epoch": 0.8734330772341286, + "grad_norm": 0.25590866804122925, + "learning_rate": 7.80453397985269e-06, + "loss": 0.0183, + "step": 37260 + }, + { + "epoch": 0.8734565188089336, + "grad_norm": 0.33630406856536865, + "learning_rate": 7.801681333082911e-06, + "loss": 0.0754, + "step": 37261 + }, + { + "epoch": 0.8734799603837385, + "grad_norm": 0.30624574422836304, + "learning_rate": 7.798829186583801e-06, + "loss": 0.0619, + "step": 37262 + }, + { + "epoch": 0.8735034019585436, + "grad_norm": 0.23644982278347015, + "learning_rate": 7.795977540370835e-06, + "loss": 0.0241, + "step": 37263 + }, + { + "epoch": 0.8735268435333485, + "grad_norm": 0.19949372112751007, + "learning_rate": 7.793126394459482e-06, + "loss": 0.0224, + "step": 37264 + }, + { + "epoch": 0.8735502851081536, + "grad_norm": 0.3135392665863037, + "learning_rate": 7.790275748865205e-06, + "loss": 0.0409, + "step": 37265 + }, + { + "epoch": 0.8735737266829585, + "grad_norm": 0.2670188844203949, + "learning_rate": 7.787425603603504e-06, + "loss": 0.0268, + "step": 37266 + }, + { + "epoch": 0.8735971682577636, + "grad_norm": 0.4081428050994873, + "learning_rate": 7.784575958689833e-06, + "loss": 0.0864, + "step": 37267 + }, + { + "epoch": 0.8736206098325685, + "grad_norm": 0.32689353823661804, + "learning_rate": 7.781726814139612e-06, + "loss": 0.0554, + "step": 37268 + }, + { + "epoch": 0.8736440514073736, + "grad_norm": 0.6560879945755005, + "learning_rate": 7.778878169968363e-06, + "loss": 0.148, + "step": 37269 + }, + { + "epoch": 0.8736674929821785, + "grad_norm": 0.7480723261833191, + "learning_rate": 7.776030026191516e-06, + "loss": 0.1508, + "step": 37270 + }, + { + "epoch": 0.8736909345569835, + "grad_norm": 0.08943469077348709, + "learning_rate": 7.773182382824517e-06, + "loss": 0.0106, + "step": 37271 + }, + { + "epoch": 0.8737143761317885, + "grad_norm": 0.4447992146015167, + "learning_rate": 7.770335239882809e-06, + "loss": 0.1127, + "step": 37272 + }, + { + "epoch": 0.8737378177065935, + "grad_norm": 0.11080743372440338, + "learning_rate": 7.767488597381867e-06, + "loss": 0.0235, + "step": 37273 + }, + { + "epoch": 0.8737612592813985, + "grad_norm": 0.5489100813865662, + "learning_rate": 7.764642455337123e-06, + "loss": 0.1493, + "step": 37274 + }, + { + "epoch": 0.8737847008562035, + "grad_norm": 0.6867482662200928, + "learning_rate": 7.761796813764032e-06, + "loss": 0.0861, + "step": 37275 + }, + { + "epoch": 0.8738081424310085, + "grad_norm": 0.628300666809082, + "learning_rate": 7.758951672678016e-06, + "loss": 0.0452, + "step": 37276 + }, + { + "epoch": 0.8738315840058135, + "grad_norm": 0.8446623682975769, + "learning_rate": 7.756107032094528e-06, + "loss": 0.1813, + "step": 37277 + }, + { + "epoch": 0.8738550255806186, + "grad_norm": 0.5315043926239014, + "learning_rate": 7.753262892028979e-06, + "loss": 0.1224, + "step": 37278 + }, + { + "epoch": 0.8738784671554235, + "grad_norm": 0.714056670665741, + "learning_rate": 7.750419252496844e-06, + "loss": 0.1214, + "step": 37279 + }, + { + "epoch": 0.8739019087302285, + "grad_norm": 0.48051634430885315, + "learning_rate": 7.747576113513499e-06, + "loss": 0.0647, + "step": 37280 + }, + { + "epoch": 0.8739253503050335, + "grad_norm": 0.14900264143943787, + "learning_rate": 7.744733475094423e-06, + "loss": 0.0206, + "step": 37281 + }, + { + "epoch": 0.8739487918798385, + "grad_norm": 0.4213999807834625, + "learning_rate": 7.741891337255014e-06, + "loss": 0.0667, + "step": 37282 + }, + { + "epoch": 0.8739722334546435, + "grad_norm": 0.40600481629371643, + "learning_rate": 7.739049700010703e-06, + "loss": 0.0651, + "step": 37283 + }, + { + "epoch": 0.8739956750294485, + "grad_norm": 0.5690408945083618, + "learning_rate": 7.73620856337689e-06, + "loss": 0.0973, + "step": 37284 + }, + { + "epoch": 0.8740191166042535, + "grad_norm": 0.6619420051574707, + "learning_rate": 7.733367927368995e-06, + "loss": 0.5067, + "step": 37285 + }, + { + "epoch": 0.8740425581790585, + "grad_norm": 0.49764782190322876, + "learning_rate": 7.730527792002462e-06, + "loss": 0.7527, + "step": 37286 + }, + { + "epoch": 0.8740659997538635, + "grad_norm": 0.2759178876876831, + "learning_rate": 7.727688157292668e-06, + "loss": 0.0337, + "step": 37287 + }, + { + "epoch": 0.8740894413286685, + "grad_norm": 0.29664379358291626, + "learning_rate": 7.724849023255042e-06, + "loss": 0.0424, + "step": 37288 + }, + { + "epoch": 0.8741128829034734, + "grad_norm": 0.2118992954492569, + "learning_rate": 7.722010389904977e-06, + "loss": 0.017, + "step": 37289 + }, + { + "epoch": 0.8741363244782785, + "grad_norm": 0.11637160181999207, + "learning_rate": 7.719172257257878e-06, + "loss": 0.0135, + "step": 37290 + }, + { + "epoch": 0.8741597660530834, + "grad_norm": 0.5145365595817566, + "learning_rate": 7.716334625329125e-06, + "loss": 0.663, + "step": 37291 + }, + { + "epoch": 0.8741832076278885, + "grad_norm": 0.44320148229599, + "learning_rate": 7.713497494134136e-06, + "loss": 0.1229, + "step": 37292 + }, + { + "epoch": 0.8742066492026934, + "grad_norm": 0.1974249929189682, + "learning_rate": 7.710660863688323e-06, + "loss": 0.0077, + "step": 37293 + }, + { + "epoch": 0.8742300907774985, + "grad_norm": 0.26887837052345276, + "learning_rate": 7.707824734007053e-06, + "loss": 0.0137, + "step": 37294 + }, + { + "epoch": 0.8742535323523034, + "grad_norm": 0.4115435779094696, + "learning_rate": 7.704989105105731e-06, + "loss": 0.0629, + "step": 37295 + }, + { + "epoch": 0.8742769739271085, + "grad_norm": 0.09626534581184387, + "learning_rate": 7.702153976999726e-06, + "loss": 0.0097, + "step": 37296 + }, + { + "epoch": 0.8743004155019134, + "grad_norm": 0.28036653995513916, + "learning_rate": 7.699319349704415e-06, + "loss": 0.0305, + "step": 37297 + }, + { + "epoch": 0.8743238570767184, + "grad_norm": 0.8034406900405884, + "learning_rate": 7.696485223235205e-06, + "loss": 0.1412, + "step": 37298 + }, + { + "epoch": 0.8743472986515234, + "grad_norm": 0.47959545254707336, + "learning_rate": 7.693651597607477e-06, + "loss": 0.0463, + "step": 37299 + }, + { + "epoch": 0.8743707402263284, + "grad_norm": 0.2912718653678894, + "learning_rate": 7.69081847283657e-06, + "loss": 0.0302, + "step": 37300 + }, + { + "epoch": 0.8743941818011334, + "grad_norm": 0.23279431462287903, + "learning_rate": 7.687985848937896e-06, + "loss": 0.049, + "step": 37301 + }, + { + "epoch": 0.8744176233759384, + "grad_norm": 0.5141311883926392, + "learning_rate": 7.685153725926797e-06, + "loss": 0.1059, + "step": 37302 + }, + { + "epoch": 0.8744410649507434, + "grad_norm": 0.42052677273750305, + "learning_rate": 7.682322103818629e-06, + "loss": 0.0834, + "step": 37303 + }, + { + "epoch": 0.8744645065255484, + "grad_norm": 0.41148704290390015, + "learning_rate": 7.67949098262878e-06, + "loss": 0.05, + "step": 37304 + }, + { + "epoch": 0.8744879481003534, + "grad_norm": 0.47914981842041016, + "learning_rate": 7.676660362372622e-06, + "loss": 0.7137, + "step": 37305 + }, + { + "epoch": 0.8745113896751584, + "grad_norm": 0.5294004082679749, + "learning_rate": 7.673830243065506e-06, + "loss": 0.0607, + "step": 37306 + }, + { + "epoch": 0.8745348312499633, + "grad_norm": 0.44768524169921875, + "learning_rate": 7.67100062472279e-06, + "loss": 0.1013, + "step": 37307 + }, + { + "epoch": 0.8745582728247684, + "grad_norm": 0.526576578617096, + "learning_rate": 7.668171507359811e-06, + "loss": 0.0801, + "step": 37308 + }, + { + "epoch": 0.8745817143995733, + "grad_norm": 0.5044174790382385, + "learning_rate": 7.665342890991934e-06, + "loss": 0.1286, + "step": 37309 + }, + { + "epoch": 0.8746051559743784, + "grad_norm": 0.21548645198345184, + "learning_rate": 7.662514775634477e-06, + "loss": 0.048, + "step": 37310 + }, + { + "epoch": 0.8746285975491833, + "grad_norm": 0.7542541027069092, + "learning_rate": 7.659687161302842e-06, + "loss": 0.1168, + "step": 37311 + }, + { + "epoch": 0.8746520391239884, + "grad_norm": 0.12053544819355011, + "learning_rate": 7.656860048012326e-06, + "loss": 0.0257, + "step": 37312 + }, + { + "epoch": 0.8746754806987933, + "grad_norm": 0.5464521646499634, + "learning_rate": 7.654033435778296e-06, + "loss": 0.0885, + "step": 37313 + }, + { + "epoch": 0.8746989222735984, + "grad_norm": 0.45273226499557495, + "learning_rate": 7.651207324616071e-06, + "loss": 0.0924, + "step": 37314 + }, + { + "epoch": 0.8747223638484033, + "grad_norm": 0.3685905635356903, + "learning_rate": 7.648381714540998e-06, + "loss": 0.0452, + "step": 37315 + }, + { + "epoch": 0.8747458054232083, + "grad_norm": 0.6040007472038269, + "learning_rate": 7.645556605568372e-06, + "loss": 0.1009, + "step": 37316 + }, + { + "epoch": 0.8747692469980133, + "grad_norm": 0.23989243805408478, + "learning_rate": 7.64273199771357e-06, + "loss": 0.0212, + "step": 37317 + }, + { + "epoch": 0.8747926885728183, + "grad_norm": 0.465762197971344, + "learning_rate": 7.639907890991904e-06, + "loss": 0.1048, + "step": 37318 + }, + { + "epoch": 0.8748161301476233, + "grad_norm": 0.34363019466400146, + "learning_rate": 7.637084285418705e-06, + "loss": 0.3345, + "step": 37319 + }, + { + "epoch": 0.8748395717224283, + "grad_norm": 0.10302615910768509, + "learning_rate": 7.634261181009273e-06, + "loss": 0.0196, + "step": 37320 + }, + { + "epoch": 0.8748630132972333, + "grad_norm": 0.6065829396247864, + "learning_rate": 7.63143857777895e-06, + "loss": 0.4741, + "step": 37321 + }, + { + "epoch": 0.8748864548720383, + "grad_norm": 0.23204222321510315, + "learning_rate": 7.628616475743022e-06, + "loss": 0.0409, + "step": 37322 + }, + { + "epoch": 0.8749098964468432, + "grad_norm": 0.31590452790260315, + "learning_rate": 7.625794874916814e-06, + "loss": 0.045, + "step": 37323 + }, + { + "epoch": 0.8749333380216483, + "grad_norm": 0.44424208998680115, + "learning_rate": 7.622973775315656e-06, + "loss": 0.0766, + "step": 37324 + }, + { + "epoch": 0.8749567795964532, + "grad_norm": 0.23827165365219116, + "learning_rate": 7.620153176954836e-06, + "loss": 0.0393, + "step": 37325 + }, + { + "epoch": 0.8749802211712583, + "grad_norm": 0.6026433110237122, + "learning_rate": 7.617333079849665e-06, + "loss": 0.144, + "step": 37326 + }, + { + "epoch": 0.8750036627460632, + "grad_norm": 0.409898579120636, + "learning_rate": 7.61451348401544e-06, + "loss": 0.063, + "step": 37327 + }, + { + "epoch": 0.8750271043208683, + "grad_norm": 0.3119218349456787, + "learning_rate": 7.611694389467439e-06, + "loss": 0.0402, + "step": 37328 + }, + { + "epoch": 0.8750505458956733, + "grad_norm": 0.6038531064987183, + "learning_rate": 7.6088757962210154e-06, + "loss": 0.063, + "step": 37329 + }, + { + "epoch": 0.8750739874704783, + "grad_norm": 0.40081965923309326, + "learning_rate": 7.606057704291403e-06, + "loss": 0.051, + "step": 37330 + }, + { + "epoch": 0.8750974290452833, + "grad_norm": 0.09736903011798859, + "learning_rate": 7.603240113693932e-06, + "loss": 0.0184, + "step": 37331 + }, + { + "epoch": 0.8751208706200883, + "grad_norm": 0.13243411481380463, + "learning_rate": 7.6004230244438815e-06, + "loss": 0.0196, + "step": 37332 + }, + { + "epoch": 0.8751443121948933, + "grad_norm": 0.2829208970069885, + "learning_rate": 7.597606436556537e-06, + "loss": 0.0744, + "step": 37333 + }, + { + "epoch": 0.8751677537696982, + "grad_norm": 0.06344567984342575, + "learning_rate": 7.594790350047176e-06, + "loss": 0.0128, + "step": 37334 + }, + { + "epoch": 0.8751911953445033, + "grad_norm": 0.44996151328086853, + "learning_rate": 7.591974764931064e-06, + "loss": 0.0862, + "step": 37335 + }, + { + "epoch": 0.8752146369193082, + "grad_norm": 0.45902419090270996, + "learning_rate": 7.589159681223512e-06, + "loss": 0.4889, + "step": 37336 + }, + { + "epoch": 0.8752380784941133, + "grad_norm": 0.4430762827396393, + "learning_rate": 7.586345098939784e-06, + "loss": 0.051, + "step": 37337 + }, + { + "epoch": 0.8752615200689182, + "grad_norm": 0.638731837272644, + "learning_rate": 7.583531018095136e-06, + "loss": 0.0791, + "step": 37338 + }, + { + "epoch": 0.8752849616437233, + "grad_norm": 0.1643541157245636, + "learning_rate": 7.580717438704854e-06, + "loss": 0.0208, + "step": 37339 + }, + { + "epoch": 0.8753084032185282, + "grad_norm": 0.5531128644943237, + "learning_rate": 7.577904360784194e-06, + "loss": 0.4751, + "step": 37340 + }, + { + "epoch": 0.8753318447933333, + "grad_norm": 0.23887944221496582, + "learning_rate": 7.57509178434841e-06, + "loss": 0.0411, + "step": 37341 + }, + { + "epoch": 0.8753552863681382, + "grad_norm": 0.4069421887397766, + "learning_rate": 7.572279709412777e-06, + "loss": 0.0381, + "step": 37342 + }, + { + "epoch": 0.8753787279429432, + "grad_norm": 0.6329189538955688, + "learning_rate": 7.569468135992564e-06, + "loss": 0.0762, + "step": 37343 + }, + { + "epoch": 0.8754021695177482, + "grad_norm": 0.18338140845298767, + "learning_rate": 7.566657064103022e-06, + "loss": 0.0118, + "step": 37344 + }, + { + "epoch": 0.8754256110925532, + "grad_norm": 0.13409002125263214, + "learning_rate": 7.563846493759397e-06, + "loss": 0.0186, + "step": 37345 + }, + { + "epoch": 0.8754490526673582, + "grad_norm": 0.19308218359947205, + "learning_rate": 7.561036424976931e-06, + "loss": 0.0347, + "step": 37346 + }, + { + "epoch": 0.8754724942421632, + "grad_norm": 0.4250355660915375, + "learning_rate": 7.558226857770889e-06, + "loss": 0.0292, + "step": 37347 + }, + { + "epoch": 0.8754959358169682, + "grad_norm": 0.45322874188423157, + "learning_rate": 7.555417792156494e-06, + "loss": 0.0615, + "step": 37348 + }, + { + "epoch": 0.8755193773917732, + "grad_norm": 0.3363255560398102, + "learning_rate": 7.55260922814901e-06, + "loss": 0.0518, + "step": 37349 + }, + { + "epoch": 0.8755428189665782, + "grad_norm": 0.12468019127845764, + "learning_rate": 7.5498011657636705e-06, + "loss": 0.0261, + "step": 37350 + }, + { + "epoch": 0.8755662605413832, + "grad_norm": 0.07409362494945526, + "learning_rate": 7.546993605015706e-06, + "loss": 0.0048, + "step": 37351 + }, + { + "epoch": 0.8755897021161881, + "grad_norm": 0.09717828780412674, + "learning_rate": 7.544186545920362e-06, + "loss": 0.0191, + "step": 37352 + }, + { + "epoch": 0.8756131436909932, + "grad_norm": 0.14903727173805237, + "learning_rate": 7.541379988492836e-06, + "loss": 0.017, + "step": 37353 + }, + { + "epoch": 0.8756365852657981, + "grad_norm": 0.4010317921638489, + "learning_rate": 7.538573932748405e-06, + "loss": 0.0416, + "step": 37354 + }, + { + "epoch": 0.8756600268406032, + "grad_norm": 0.35433030128479004, + "learning_rate": 7.535768378702257e-06, + "loss": 0.0306, + "step": 37355 + }, + { + "epoch": 0.8756834684154081, + "grad_norm": 0.26704734563827515, + "learning_rate": 7.532963326369636e-06, + "loss": 0.0483, + "step": 37356 + }, + { + "epoch": 0.8757069099902132, + "grad_norm": 0.5716302990913391, + "learning_rate": 7.530158775765772e-06, + "loss": 0.1545, + "step": 37357 + }, + { + "epoch": 0.8757303515650181, + "grad_norm": 0.2283434420824051, + "learning_rate": 7.527354726905856e-06, + "loss": 0.03, + "step": 37358 + }, + { + "epoch": 0.8757537931398232, + "grad_norm": 0.6049038767814636, + "learning_rate": 7.524551179805106e-06, + "loss": 0.0936, + "step": 37359 + }, + { + "epoch": 0.8757772347146281, + "grad_norm": 0.8108927607536316, + "learning_rate": 7.521748134478756e-06, + "loss": 0.1211, + "step": 37360 + }, + { + "epoch": 0.8758006762894331, + "grad_norm": 0.29830601811408997, + "learning_rate": 7.518945590941972e-06, + "loss": 0.0417, + "step": 37361 + }, + { + "epoch": 0.8758241178642381, + "grad_norm": 0.12950918078422546, + "learning_rate": 7.5161435492100195e-06, + "loss": 0.0168, + "step": 37362 + }, + { + "epoch": 0.8758475594390431, + "grad_norm": 0.295406311750412, + "learning_rate": 7.513342009298063e-06, + "loss": 0.0311, + "step": 37363 + }, + { + "epoch": 0.8758710010138481, + "grad_norm": 0.6522355079650879, + "learning_rate": 7.510540971221314e-06, + "loss": 0.1105, + "step": 37364 + }, + { + "epoch": 0.8758944425886531, + "grad_norm": 0.6004290580749512, + "learning_rate": 7.507740434994959e-06, + "loss": 0.0785, + "step": 37365 + }, + { + "epoch": 0.8759178841634581, + "grad_norm": 0.3829300105571747, + "learning_rate": 7.504940400634208e-06, + "loss": 0.2355, + "step": 37366 + }, + { + "epoch": 0.8759413257382631, + "grad_norm": 0.2207176834344864, + "learning_rate": 7.502140868154262e-06, + "loss": 0.0298, + "step": 37367 + }, + { + "epoch": 0.875964767313068, + "grad_norm": 0.2621096670627594, + "learning_rate": 7.499341837570284e-06, + "loss": 0.0264, + "step": 37368 + }, + { + "epoch": 0.8759882088878731, + "grad_norm": 0.11318796873092651, + "learning_rate": 7.496543308897497e-06, + "loss": 0.0225, + "step": 37369 + }, + { + "epoch": 0.876011650462678, + "grad_norm": 0.510751485824585, + "learning_rate": 7.4937452821510654e-06, + "loss": 0.6555, + "step": 37370 + }, + { + "epoch": 0.8760350920374831, + "grad_norm": 0.43255817890167236, + "learning_rate": 7.490947757346167e-06, + "loss": 0.0833, + "step": 37371 + }, + { + "epoch": 0.876058533612288, + "grad_norm": 0.6368259787559509, + "learning_rate": 7.488150734497989e-06, + "loss": 0.076, + "step": 37372 + }, + { + "epoch": 0.8760819751870931, + "grad_norm": 0.4228776693344116, + "learning_rate": 7.485354213621698e-06, + "loss": 0.0966, + "step": 37373 + }, + { + "epoch": 0.876105416761898, + "grad_norm": 0.2934544086456299, + "learning_rate": 7.482558194732481e-06, + "loss": 0.0626, + "step": 37374 + }, + { + "epoch": 0.8761288583367031, + "grad_norm": 0.12319740653038025, + "learning_rate": 7.479762677845514e-06, + "loss": 0.0145, + "step": 37375 + }, + { + "epoch": 0.876152299911508, + "grad_norm": 0.5640546083450317, + "learning_rate": 7.4769676629759534e-06, + "loss": 0.1465, + "step": 37376 + }, + { + "epoch": 0.876175741486313, + "grad_norm": 0.4012366235256195, + "learning_rate": 7.474173150138941e-06, + "loss": 0.0742, + "step": 37377 + }, + { + "epoch": 0.876199183061118, + "grad_norm": 0.169613316655159, + "learning_rate": 7.471379139349699e-06, + "loss": 0.022, + "step": 37378 + }, + { + "epoch": 0.876222624635923, + "grad_norm": 0.15611466765403748, + "learning_rate": 7.468585630623348e-06, + "loss": 0.0322, + "step": 37379 + }, + { + "epoch": 0.8762460662107281, + "grad_norm": 0.3940616846084595, + "learning_rate": 7.465792623975032e-06, + "loss": 0.1026, + "step": 37380 + }, + { + "epoch": 0.876269507785533, + "grad_norm": 0.4951286315917969, + "learning_rate": 7.463000119419949e-06, + "loss": 0.574, + "step": 37381 + }, + { + "epoch": 0.8762929493603381, + "grad_norm": 0.12320384383201599, + "learning_rate": 7.460208116973233e-06, + "loss": 0.0106, + "step": 37382 + }, + { + "epoch": 0.876316390935143, + "grad_norm": 0.4681801497936249, + "learning_rate": 7.457416616650026e-06, + "loss": 0.0805, + "step": 37383 + }, + { + "epoch": 0.8763398325099481, + "grad_norm": 0.04397869110107422, + "learning_rate": 7.454625618465483e-06, + "loss": 0.0019, + "step": 37384 + }, + { + "epoch": 0.876363274084753, + "grad_norm": 0.5780584216117859, + "learning_rate": 7.451835122434736e-06, + "loss": 0.1136, + "step": 37385 + }, + { + "epoch": 0.8763867156595581, + "grad_norm": 0.4528134763240814, + "learning_rate": 7.449045128572918e-06, + "loss": 0.3799, + "step": 37386 + }, + { + "epoch": 0.876410157234363, + "grad_norm": 0.6150633096694946, + "learning_rate": 7.446255636895205e-06, + "loss": 0.0954, + "step": 37387 + }, + { + "epoch": 0.876433598809168, + "grad_norm": 0.38798201084136963, + "learning_rate": 7.443466647416708e-06, + "loss": 0.0523, + "step": 37388 + }, + { + "epoch": 0.876457040383973, + "grad_norm": 0.47941404581069946, + "learning_rate": 7.440678160152547e-06, + "loss": 0.4121, + "step": 37389 + }, + { + "epoch": 0.876480481958778, + "grad_norm": 0.11577805131673813, + "learning_rate": 7.43789017511789e-06, + "loss": 0.021, + "step": 37390 + }, + { + "epoch": 0.876503923533583, + "grad_norm": 0.7405924201011658, + "learning_rate": 7.435102692327833e-06, + "loss": 0.7009, + "step": 37391 + }, + { + "epoch": 0.876527365108388, + "grad_norm": 0.08929760754108429, + "learning_rate": 7.432315711797522e-06, + "loss": 0.0083, + "step": 37392 + }, + { + "epoch": 0.876550806683193, + "grad_norm": 0.49737808108329773, + "learning_rate": 7.429529233542043e-06, + "loss": 0.0762, + "step": 37393 + }, + { + "epoch": 0.876574248257998, + "grad_norm": 0.11541755497455597, + "learning_rate": 7.426743257576563e-06, + "loss": 0.0121, + "step": 37394 + }, + { + "epoch": 0.876597689832803, + "grad_norm": 0.40026673674583435, + "learning_rate": 7.42395778391618e-06, + "loss": 0.0427, + "step": 37395 + }, + { + "epoch": 0.876621131407608, + "grad_norm": 0.36875227093696594, + "learning_rate": 7.421172812576005e-06, + "loss": 0.0502, + "step": 37396 + }, + { + "epoch": 0.8766445729824129, + "grad_norm": 0.2255251556634903, + "learning_rate": 7.418388343571147e-06, + "loss": 0.0085, + "step": 37397 + }, + { + "epoch": 0.876668014557218, + "grad_norm": 0.11465268582105637, + "learning_rate": 7.415604376916718e-06, + "loss": 0.0297, + "step": 37398 + }, + { + "epoch": 0.8766914561320229, + "grad_norm": 0.39392828941345215, + "learning_rate": 7.412820912627805e-06, + "loss": 0.0351, + "step": 37399 + }, + { + "epoch": 0.876714897706828, + "grad_norm": 0.15611013770103455, + "learning_rate": 7.410037950719539e-06, + "loss": 0.0296, + "step": 37400 + }, + { + "epoch": 0.8767383392816329, + "grad_norm": 0.10116343945264816, + "learning_rate": 7.407255491206999e-06, + "loss": 0.0132, + "step": 37401 + }, + { + "epoch": 0.876761780856438, + "grad_norm": 0.16053837537765503, + "learning_rate": 7.404473534105317e-06, + "loss": 0.0271, + "step": 37402 + }, + { + "epoch": 0.8767852224312429, + "grad_norm": 0.10803687572479248, + "learning_rate": 7.401692079429556e-06, + "loss": 0.012, + "step": 37403 + }, + { + "epoch": 0.876808664006048, + "grad_norm": 0.28403711318969727, + "learning_rate": 7.39891112719483e-06, + "loss": 0.0324, + "step": 37404 + }, + { + "epoch": 0.8768321055808529, + "grad_norm": 0.2791517972946167, + "learning_rate": 7.396130677416202e-06, + "loss": 0.032, + "step": 37405 + }, + { + "epoch": 0.8768555471556579, + "grad_norm": 0.1609157919883728, + "learning_rate": 7.393350730108762e-06, + "loss": 0.0209, + "step": 37406 + }, + { + "epoch": 0.8768789887304629, + "grad_norm": 0.11873659491539001, + "learning_rate": 7.390571285287618e-06, + "loss": 0.0163, + "step": 37407 + }, + { + "epoch": 0.8769024303052679, + "grad_norm": 0.3206269145011902, + "learning_rate": 7.387792342967847e-06, + "loss": 0.0481, + "step": 37408 + }, + { + "epoch": 0.8769258718800729, + "grad_norm": 0.2639482617378235, + "learning_rate": 7.385013903164506e-06, + "loss": 0.0134, + "step": 37409 + }, + { + "epoch": 0.8769493134548779, + "grad_norm": 0.35975977778434753, + "learning_rate": 7.382235965892692e-06, + "loss": 0.0775, + "step": 37410 + }, + { + "epoch": 0.8769727550296829, + "grad_norm": 0.3267992436885834, + "learning_rate": 7.379458531167449e-06, + "loss": 0.0607, + "step": 37411 + }, + { + "epoch": 0.8769961966044879, + "grad_norm": 0.07631774991750717, + "learning_rate": 7.376681599003899e-06, + "loss": 0.0154, + "step": 37412 + }, + { + "epoch": 0.8770196381792928, + "grad_norm": 0.2978112995624542, + "learning_rate": 7.3739051694170505e-06, + "loss": 0.0718, + "step": 37413 + }, + { + "epoch": 0.8770430797540979, + "grad_norm": 0.17313094437122345, + "learning_rate": 7.371129242422003e-06, + "loss": 0.0262, + "step": 37414 + }, + { + "epoch": 0.8770665213289028, + "grad_norm": 0.2569606900215149, + "learning_rate": 7.368353818033835e-06, + "loss": 0.0372, + "step": 37415 + }, + { + "epoch": 0.8770899629037079, + "grad_norm": 0.4237094819545746, + "learning_rate": 7.365578896267567e-06, + "loss": 0.0452, + "step": 37416 + }, + { + "epoch": 0.8771134044785128, + "grad_norm": 0.09465526789426804, + "learning_rate": 7.362804477138274e-06, + "loss": 0.0092, + "step": 37417 + }, + { + "epoch": 0.8771368460533179, + "grad_norm": 0.08872642368078232, + "learning_rate": 7.360030560661002e-06, + "loss": 0.0152, + "step": 37418 + }, + { + "epoch": 0.8771602876281228, + "grad_norm": 0.1953800618648529, + "learning_rate": 7.357257146850816e-06, + "loss": 0.007, + "step": 37419 + }, + { + "epoch": 0.8771837292029279, + "grad_norm": 0.0978388860821724, + "learning_rate": 7.354484235722769e-06, + "loss": 0.0225, + "step": 37420 + }, + { + "epoch": 0.8772071707777328, + "grad_norm": 0.8555507063865662, + "learning_rate": 7.351711827291885e-06, + "loss": 0.7641, + "step": 37421 + }, + { + "epoch": 0.8772306123525379, + "grad_norm": 0.19077564775943756, + "learning_rate": 7.348939921573217e-06, + "loss": 0.0277, + "step": 37422 + }, + { + "epoch": 0.8772540539273428, + "grad_norm": 0.6407315731048584, + "learning_rate": 7.346168518581809e-06, + "loss": 0.1314, + "step": 37423 + }, + { + "epoch": 0.8772774955021478, + "grad_norm": 0.39731064438819885, + "learning_rate": 7.343397618332681e-06, + "loss": 0.0482, + "step": 37424 + }, + { + "epoch": 0.8773009370769528, + "grad_norm": 0.5610589981079102, + "learning_rate": 7.340627220840901e-06, + "loss": 0.1133, + "step": 37425 + }, + { + "epoch": 0.8773243786517578, + "grad_norm": 0.4756630063056946, + "learning_rate": 7.337857326121466e-06, + "loss": 0.061, + "step": 37426 + }, + { + "epoch": 0.8773478202265628, + "grad_norm": 0.11937865614891052, + "learning_rate": 7.335087934189433e-06, + "loss": 0.0187, + "step": 37427 + }, + { + "epoch": 0.8773712618013678, + "grad_norm": 0.48822396993637085, + "learning_rate": 7.332319045059821e-06, + "loss": 0.0915, + "step": 37428 + }, + { + "epoch": 0.8773947033761728, + "grad_norm": 0.19313892722129822, + "learning_rate": 7.329550658747653e-06, + "loss": 0.0364, + "step": 37429 + }, + { + "epoch": 0.8774181449509778, + "grad_norm": 0.26320672035217285, + "learning_rate": 7.3267827752679485e-06, + "loss": 0.0368, + "step": 37430 + }, + { + "epoch": 0.8774415865257829, + "grad_norm": 0.3171290457248688, + "learning_rate": 7.324015394635708e-06, + "loss": 0.0364, + "step": 37431 + }, + { + "epoch": 0.8774650281005878, + "grad_norm": 0.426444947719574, + "learning_rate": 7.321248516865986e-06, + "loss": 0.1084, + "step": 37432 + }, + { + "epoch": 0.8774884696753928, + "grad_norm": 0.6586542725563049, + "learning_rate": 7.318482141973781e-06, + "loss": 0.0528, + "step": 37433 + }, + { + "epoch": 0.8775119112501978, + "grad_norm": 0.19816739857196808, + "learning_rate": 7.3157162699740825e-06, + "loss": 0.0156, + "step": 37434 + }, + { + "epoch": 0.8775353528250028, + "grad_norm": 0.3092435598373413, + "learning_rate": 7.312950900881921e-06, + "loss": 0.0351, + "step": 37435 + }, + { + "epoch": 0.8775587943998078, + "grad_norm": 0.6025130152702332, + "learning_rate": 7.310186034712296e-06, + "loss": 0.0826, + "step": 37436 + }, + { + "epoch": 0.8775822359746128, + "grad_norm": 0.16037707030773163, + "learning_rate": 7.307421671480197e-06, + "loss": 0.0271, + "step": 37437 + }, + { + "epoch": 0.8776056775494178, + "grad_norm": 0.5171896815299988, + "learning_rate": 7.3046578112006215e-06, + "loss": 0.5867, + "step": 37438 + }, + { + "epoch": 0.8776291191242228, + "grad_norm": 0.31576210260391235, + "learning_rate": 7.301894453888602e-06, + "loss": 0.0578, + "step": 37439 + }, + { + "epoch": 0.8776525606990278, + "grad_norm": 0.2002900391817093, + "learning_rate": 7.299131599559106e-06, + "loss": 0.0116, + "step": 37440 + }, + { + "epoch": 0.8776760022738328, + "grad_norm": 0.32611674070358276, + "learning_rate": 7.296369248227131e-06, + "loss": 0.0377, + "step": 37441 + }, + { + "epoch": 0.8776994438486377, + "grad_norm": 0.26159802079200745, + "learning_rate": 7.293607399907665e-06, + "loss": 0.075, + "step": 37442 + }, + { + "epoch": 0.8777228854234428, + "grad_norm": 0.4822947382926941, + "learning_rate": 7.2908460546156965e-06, + "loss": 0.0519, + "step": 37443 + }, + { + "epoch": 0.8777463269982477, + "grad_norm": 0.3280450403690338, + "learning_rate": 7.28808521236618e-06, + "loss": 0.049, + "step": 37444 + }, + { + "epoch": 0.8777697685730528, + "grad_norm": 0.376212477684021, + "learning_rate": 7.2853248731741375e-06, + "loss": 0.0379, + "step": 37445 + }, + { + "epoch": 0.8777932101478577, + "grad_norm": 0.3288024365901947, + "learning_rate": 7.282565037054534e-06, + "loss": 0.0284, + "step": 37446 + }, + { + "epoch": 0.8778166517226628, + "grad_norm": 0.723456084728241, + "learning_rate": 7.279805704022324e-06, + "loss": 0.1788, + "step": 37447 + }, + { + "epoch": 0.8778400932974677, + "grad_norm": 0.0681433156132698, + "learning_rate": 7.277046874092508e-06, + "loss": 0.0106, + "step": 37448 + }, + { + "epoch": 0.8778635348722728, + "grad_norm": 0.06323355436325073, + "learning_rate": 7.274288547280028e-06, + "loss": 0.0156, + "step": 37449 + }, + { + "epoch": 0.8778869764470777, + "grad_norm": 0.7436367869377136, + "learning_rate": 7.271530723599884e-06, + "loss": 0.093, + "step": 37450 + }, + { + "epoch": 0.8779104180218827, + "grad_norm": 0.600135326385498, + "learning_rate": 7.268773403066998e-06, + "loss": 0.0696, + "step": 37451 + }, + { + "epoch": 0.8779338595966877, + "grad_norm": 0.35719946026802063, + "learning_rate": 7.266016585696367e-06, + "loss": 0.0319, + "step": 37452 + }, + { + "epoch": 0.8779573011714927, + "grad_norm": 0.3490016460418701, + "learning_rate": 7.263260271502948e-06, + "loss": 0.0459, + "step": 37453 + }, + { + "epoch": 0.8779807427462977, + "grad_norm": 0.15464946627616882, + "learning_rate": 7.260504460501682e-06, + "loss": 0.013, + "step": 37454 + }, + { + "epoch": 0.8780041843211027, + "grad_norm": 0.49549543857574463, + "learning_rate": 7.257749152707527e-06, + "loss": 0.1106, + "step": 37455 + }, + { + "epoch": 0.8780276258959077, + "grad_norm": 0.35411298274993896, + "learning_rate": 7.254994348135413e-06, + "loss": 0.043, + "step": 37456 + }, + { + "epoch": 0.8780510674707127, + "grad_norm": 0.5901895761489868, + "learning_rate": 7.25224004680033e-06, + "loss": 0.1514, + "step": 37457 + }, + { + "epoch": 0.8780745090455176, + "grad_norm": 0.5445123314857483, + "learning_rate": 7.249486248717196e-06, + "loss": 0.1083, + "step": 37458 + }, + { + "epoch": 0.8780979506203227, + "grad_norm": 0.33458828926086426, + "learning_rate": 7.2467329539009585e-06, + "loss": 0.0469, + "step": 37459 + }, + { + "epoch": 0.8781213921951276, + "grad_norm": 0.757025420665741, + "learning_rate": 7.24398016236656e-06, + "loss": 0.1031, + "step": 37460 + }, + { + "epoch": 0.8781448337699327, + "grad_norm": 0.741464376449585, + "learning_rate": 7.241227874128931e-06, + "loss": 0.1446, + "step": 37461 + }, + { + "epoch": 0.8781682753447376, + "grad_norm": 0.24223928153514862, + "learning_rate": 7.238476089202983e-06, + "loss": 0.0376, + "step": 37462 + }, + { + "epoch": 0.8781917169195427, + "grad_norm": 0.6465704441070557, + "learning_rate": 7.235724807603683e-06, + "loss": 0.1167, + "step": 37463 + }, + { + "epoch": 0.8782151584943476, + "grad_norm": 0.6123731136322021, + "learning_rate": 7.232974029345963e-06, + "loss": 0.4972, + "step": 37464 + }, + { + "epoch": 0.8782386000691527, + "grad_norm": 0.41827893257141113, + "learning_rate": 7.2302237544447205e-06, + "loss": 0.039, + "step": 37465 + }, + { + "epoch": 0.8782620416439576, + "grad_norm": 0.18710261583328247, + "learning_rate": 7.227473982914912e-06, + "loss": 0.0321, + "step": 37466 + }, + { + "epoch": 0.8782854832187627, + "grad_norm": 0.5906187295913696, + "learning_rate": 7.224724714771424e-06, + "loss": 0.1064, + "step": 37467 + }, + { + "epoch": 0.8783089247935676, + "grad_norm": 0.22472424805164337, + "learning_rate": 7.221975950029192e-06, + "loss": 0.0189, + "step": 37468 + }, + { + "epoch": 0.8783323663683726, + "grad_norm": 0.2701544761657715, + "learning_rate": 7.219227688703123e-06, + "loss": 0.3613, + "step": 37469 + }, + { + "epoch": 0.8783558079431776, + "grad_norm": 0.16720302402973175, + "learning_rate": 7.216479930808129e-06, + "loss": 0.0303, + "step": 37470 + }, + { + "epoch": 0.8783792495179826, + "grad_norm": 0.34360581636428833, + "learning_rate": 7.2137326763591416e-06, + "loss": 0.0696, + "step": 37471 + }, + { + "epoch": 0.8784026910927876, + "grad_norm": 0.4176456332206726, + "learning_rate": 7.210985925371039e-06, + "loss": 0.0734, + "step": 37472 + }, + { + "epoch": 0.8784261326675926, + "grad_norm": 0.2641662061214447, + "learning_rate": 7.208239677858741e-06, + "loss": 0.0545, + "step": 37473 + }, + { + "epoch": 0.8784495742423976, + "grad_norm": 0.45662087202072144, + "learning_rate": 7.205493933837127e-06, + "loss": 0.0724, + "step": 37474 + }, + { + "epoch": 0.8784730158172026, + "grad_norm": 0.1784692406654358, + "learning_rate": 7.202748693321126e-06, + "loss": 0.0267, + "step": 37475 + }, + { + "epoch": 0.8784964573920075, + "grad_norm": 0.29904717206954956, + "learning_rate": 7.200003956325607e-06, + "loss": 0.0559, + "step": 37476 + }, + { + "epoch": 0.8785198989668126, + "grad_norm": 0.27849215269088745, + "learning_rate": 7.1972597228654795e-06, + "loss": 0.0469, + "step": 37477 + }, + { + "epoch": 0.8785433405416175, + "grad_norm": 0.21836240589618683, + "learning_rate": 7.194515992955642e-06, + "loss": 0.0348, + "step": 37478 + }, + { + "epoch": 0.8785667821164226, + "grad_norm": 0.3738550841808319, + "learning_rate": 7.191772766610972e-06, + "loss": 0.0795, + "step": 37479 + }, + { + "epoch": 0.8785902236912275, + "grad_norm": 0.6657761335372925, + "learning_rate": 7.189030043846345e-06, + "loss": 0.1332, + "step": 37480 + }, + { + "epoch": 0.8786136652660326, + "grad_norm": 0.31320980191230774, + "learning_rate": 7.186287824676663e-06, + "loss": 0.0398, + "step": 37481 + }, + { + "epoch": 0.8786371068408376, + "grad_norm": 0.3736247420310974, + "learning_rate": 7.183546109116768e-06, + "loss": 0.0574, + "step": 37482 + }, + { + "epoch": 0.8786605484156426, + "grad_norm": 0.527034342288971, + "learning_rate": 7.18080489718157e-06, + "loss": 0.0389, + "step": 37483 + }, + { + "epoch": 0.8786839899904476, + "grad_norm": 0.12051139771938324, + "learning_rate": 7.1780641888859465e-06, + "loss": 0.0114, + "step": 37484 + }, + { + "epoch": 0.8787074315652526, + "grad_norm": 0.5784034729003906, + "learning_rate": 7.175323984244753e-06, + "loss": 0.4837, + "step": 37485 + }, + { + "epoch": 0.8787308731400576, + "grad_norm": 0.39317744970321655, + "learning_rate": 7.172584283272843e-06, + "loss": 0.0707, + "step": 37486 + }, + { + "epoch": 0.8787543147148625, + "grad_norm": 0.25332164764404297, + "learning_rate": 7.169845085985117e-06, + "loss": 0.0249, + "step": 37487 + }, + { + "epoch": 0.8787777562896676, + "grad_norm": 0.7439438700675964, + "learning_rate": 7.167106392396427e-06, + "loss": 0.1033, + "step": 37488 + }, + { + "epoch": 0.8788011978644725, + "grad_norm": 0.1997561752796173, + "learning_rate": 7.16436820252161e-06, + "loss": 0.0445, + "step": 37489 + }, + { + "epoch": 0.8788246394392776, + "grad_norm": 0.21685466170310974, + "learning_rate": 7.16163051637555e-06, + "loss": 0.0126, + "step": 37490 + }, + { + "epoch": 0.8788480810140825, + "grad_norm": 0.2756163477897644, + "learning_rate": 7.158893333973105e-06, + "loss": 0.0234, + "step": 37491 + }, + { + "epoch": 0.8788715225888876, + "grad_norm": 0.3709331154823303, + "learning_rate": 7.156156655329105e-06, + "loss": 0.0519, + "step": 37492 + }, + { + "epoch": 0.8788949641636925, + "grad_norm": 0.53831547498703, + "learning_rate": 7.153420480458417e-06, + "loss": 0.1054, + "step": 37493 + }, + { + "epoch": 0.8789184057384976, + "grad_norm": 0.37145060300827026, + "learning_rate": 7.150684809375863e-06, + "loss": 0.0456, + "step": 37494 + }, + { + "epoch": 0.8789418473133025, + "grad_norm": 0.36266806721687317, + "learning_rate": 7.147949642096308e-06, + "loss": 0.0325, + "step": 37495 + }, + { + "epoch": 0.8789652888881075, + "grad_norm": 0.5369294881820679, + "learning_rate": 7.145214978634607e-06, + "loss": 0.1419, + "step": 37496 + }, + { + "epoch": 0.8789887304629125, + "grad_norm": 0.2619832754135132, + "learning_rate": 7.142480819005571e-06, + "loss": 0.0534, + "step": 37497 + }, + { + "epoch": 0.8790121720377175, + "grad_norm": 0.45668429136276245, + "learning_rate": 7.1397471632240424e-06, + "loss": 0.0753, + "step": 37498 + }, + { + "epoch": 0.8790356136125225, + "grad_norm": 0.4076296091079712, + "learning_rate": 7.1370140113048435e-06, + "loss": 0.0702, + "step": 37499 + }, + { + "epoch": 0.8790590551873275, + "grad_norm": 0.35785427689552307, + "learning_rate": 7.13428136326284e-06, + "loss": 0.0202, + "step": 37500 + }, + { + "epoch": 0.8790824967621325, + "grad_norm": 0.2896097004413605, + "learning_rate": 7.13154921911281e-06, + "loss": 0.0687, + "step": 37501 + }, + { + "epoch": 0.8791059383369375, + "grad_norm": 0.4133777320384979, + "learning_rate": 7.128817578869629e-06, + "loss": 0.0561, + "step": 37502 + }, + { + "epoch": 0.8791293799117424, + "grad_norm": 0.3507389724254608, + "learning_rate": 7.126086442548096e-06, + "loss": 0.3734, + "step": 37503 + }, + { + "epoch": 0.8791528214865475, + "grad_norm": 0.407160222530365, + "learning_rate": 7.123355810163024e-06, + "loss": 0.0636, + "step": 37504 + }, + { + "epoch": 0.8791762630613524, + "grad_norm": 0.4494641125202179, + "learning_rate": 7.120625681729231e-06, + "loss": 0.0485, + "step": 37505 + }, + { + "epoch": 0.8791997046361575, + "grad_norm": 0.10249408334493637, + "learning_rate": 7.117896057261541e-06, + "loss": 0.0253, + "step": 37506 + }, + { + "epoch": 0.8792231462109624, + "grad_norm": 0.3984028995037079, + "learning_rate": 7.115166936774742e-06, + "loss": 0.049, + "step": 37507 + }, + { + "epoch": 0.8792465877857675, + "grad_norm": 0.8829701542854309, + "learning_rate": 7.112438320283676e-06, + "loss": 0.1588, + "step": 37508 + }, + { + "epoch": 0.8792700293605724, + "grad_norm": 0.4109954833984375, + "learning_rate": 7.109710207803133e-06, + "loss": 0.0767, + "step": 37509 + }, + { + "epoch": 0.8792934709353775, + "grad_norm": 0.2698534429073334, + "learning_rate": 7.106982599347911e-06, + "loss": 0.0266, + "step": 37510 + }, + { + "epoch": 0.8793169125101824, + "grad_norm": 0.0700150802731514, + "learning_rate": 7.104255494932799e-06, + "loss": 0.0067, + "step": 37511 + }, + { + "epoch": 0.8793403540849875, + "grad_norm": 0.2675282657146454, + "learning_rate": 7.101528894572629e-06, + "loss": 0.0559, + "step": 37512 + }, + { + "epoch": 0.8793637956597924, + "grad_norm": 0.506241500377655, + "learning_rate": 7.098802798282167e-06, + "loss": 0.5761, + "step": 37513 + }, + { + "epoch": 0.8793872372345974, + "grad_norm": 0.25520411133766174, + "learning_rate": 7.096077206076202e-06, + "loss": 0.0412, + "step": 37514 + }, + { + "epoch": 0.8794106788094024, + "grad_norm": 0.6438788771629333, + "learning_rate": 7.093352117969543e-06, + "loss": 0.0584, + "step": 37515 + }, + { + "epoch": 0.8794341203842074, + "grad_norm": 0.5205272436141968, + "learning_rate": 7.090627533976979e-06, + "loss": 0.0596, + "step": 37516 + }, + { + "epoch": 0.8794575619590124, + "grad_norm": 0.48434555530548096, + "learning_rate": 7.087903454113276e-06, + "loss": 0.0852, + "step": 37517 + }, + { + "epoch": 0.8794810035338174, + "grad_norm": 0.0998835414648056, + "learning_rate": 7.085179878393222e-06, + "loss": 0.0205, + "step": 37518 + }, + { + "epoch": 0.8795044451086224, + "grad_norm": 0.6349037289619446, + "learning_rate": 7.082456806831595e-06, + "loss": 0.071, + "step": 37519 + }, + { + "epoch": 0.8795278866834274, + "grad_norm": 0.632887065410614, + "learning_rate": 7.079734239443148e-06, + "loss": 0.128, + "step": 37520 + }, + { + "epoch": 0.8795513282582323, + "grad_norm": 0.16778166592121124, + "learning_rate": 7.077012176242703e-06, + "loss": 0.0172, + "step": 37521 + }, + { + "epoch": 0.8795747698330374, + "grad_norm": 0.48819682002067566, + "learning_rate": 7.0742906172449945e-06, + "loss": 0.0622, + "step": 37522 + }, + { + "epoch": 0.8795982114078423, + "grad_norm": 0.4640512764453888, + "learning_rate": 7.071569562464786e-06, + "loss": 0.5825, + "step": 37523 + }, + { + "epoch": 0.8796216529826474, + "grad_norm": 0.28660449385643005, + "learning_rate": 7.068849011916867e-06, + "loss": 0.0368, + "step": 37524 + }, + { + "epoch": 0.8796450945574523, + "grad_norm": 0.6414266228675842, + "learning_rate": 7.066128965615993e-06, + "loss": 0.115, + "step": 37525 + }, + { + "epoch": 0.8796685361322574, + "grad_norm": 0.3439674973487854, + "learning_rate": 7.063409423576906e-06, + "loss": 0.0546, + "step": 37526 + }, + { + "epoch": 0.8796919777070623, + "grad_norm": 0.4242708086967468, + "learning_rate": 7.060690385814361e-06, + "loss": 0.0476, + "step": 37527 + }, + { + "epoch": 0.8797154192818674, + "grad_norm": 0.3284188508987427, + "learning_rate": 7.057971852343137e-06, + "loss": 0.0447, + "step": 37528 + }, + { + "epoch": 0.8797388608566723, + "grad_norm": 0.12428060173988342, + "learning_rate": 7.055253823177976e-06, + "loss": 0.0152, + "step": 37529 + }, + { + "epoch": 0.8797623024314773, + "grad_norm": 0.37422657012939453, + "learning_rate": 7.0525362983336225e-06, + "loss": 0.0352, + "step": 37530 + }, + { + "epoch": 0.8797857440062823, + "grad_norm": 0.38275760412216187, + "learning_rate": 7.0498192778248095e-06, + "loss": 0.0626, + "step": 37531 + }, + { + "epoch": 0.8798091855810873, + "grad_norm": 0.3894263803958893, + "learning_rate": 7.0471027616662795e-06, + "loss": 0.0584, + "step": 37532 + }, + { + "epoch": 0.8798326271558924, + "grad_norm": 0.3248386085033417, + "learning_rate": 7.0443867498728115e-06, + "loss": 0.0681, + "step": 37533 + }, + { + "epoch": 0.8798560687306973, + "grad_norm": 0.6103174686431885, + "learning_rate": 7.041671242459102e-06, + "loss": 0.5993, + "step": 37534 + }, + { + "epoch": 0.8798795103055024, + "grad_norm": 0.40886008739471436, + "learning_rate": 7.038956239439886e-06, + "loss": 0.0799, + "step": 37535 + }, + { + "epoch": 0.8799029518803073, + "grad_norm": 0.5394263863563538, + "learning_rate": 7.0362417408299185e-06, + "loss": 0.0922, + "step": 37536 + }, + { + "epoch": 0.8799263934551124, + "grad_norm": 0.3590346574783325, + "learning_rate": 7.03352774664392e-06, + "loss": 0.0323, + "step": 37537 + }, + { + "epoch": 0.8799498350299173, + "grad_norm": 0.503240704536438, + "learning_rate": 7.030814256896623e-06, + "loss": 0.1216, + "step": 37538 + }, + { + "epoch": 0.8799732766047224, + "grad_norm": 0.5347021818161011, + "learning_rate": 7.028101271602716e-06, + "loss": 0.1108, + "step": 37539 + }, + { + "epoch": 0.8799967181795273, + "grad_norm": 0.5284743905067444, + "learning_rate": 7.0253887907769654e-06, + "loss": 0.093, + "step": 37540 + }, + { + "epoch": 0.8800201597543323, + "grad_norm": 0.5204463601112366, + "learning_rate": 7.022676814434059e-06, + "loss": 0.7136, + "step": 37541 + }, + { + "epoch": 0.8800436013291373, + "grad_norm": 0.39616459608078003, + "learning_rate": 7.019965342588742e-06, + "loss": 0.052, + "step": 37542 + }, + { + "epoch": 0.8800670429039423, + "grad_norm": 0.7171606421470642, + "learning_rate": 7.017254375255689e-06, + "loss": 0.0943, + "step": 37543 + }, + { + "epoch": 0.8800904844787473, + "grad_norm": 0.2606904208660126, + "learning_rate": 7.0145439124496334e-06, + "loss": 0.0641, + "step": 37544 + }, + { + "epoch": 0.8801139260535523, + "grad_norm": 0.5023125410079956, + "learning_rate": 7.011833954185265e-06, + "loss": 0.1248, + "step": 37545 + }, + { + "epoch": 0.8801373676283573, + "grad_norm": 0.2616751790046692, + "learning_rate": 7.009124500477316e-06, + "loss": 0.0204, + "step": 37546 + }, + { + "epoch": 0.8801608092031623, + "grad_norm": 0.1183057352900505, + "learning_rate": 7.006415551340462e-06, + "loss": 0.0171, + "step": 37547 + }, + { + "epoch": 0.8801842507779672, + "grad_norm": 0.26429283618927, + "learning_rate": 7.003707106789414e-06, + "loss": 0.0304, + "step": 37548 + }, + { + "epoch": 0.8802076923527723, + "grad_norm": 0.08446517586708069, + "learning_rate": 7.000999166838873e-06, + "loss": 0.0111, + "step": 37549 + }, + { + "epoch": 0.8802311339275772, + "grad_norm": 0.258362740278244, + "learning_rate": 6.998291731503526e-06, + "loss": 0.0382, + "step": 37550 + }, + { + "epoch": 0.8802545755023823, + "grad_norm": 0.4663912355899811, + "learning_rate": 6.99558480079806e-06, + "loss": 0.0784, + "step": 37551 + }, + { + "epoch": 0.8802780170771872, + "grad_norm": 0.09224725514650345, + "learning_rate": 6.992878374737155e-06, + "loss": 0.0149, + "step": 37552 + }, + { + "epoch": 0.8803014586519923, + "grad_norm": 0.3363931477069855, + "learning_rate": 6.990172453335519e-06, + "loss": 0.3917, + "step": 37553 + }, + { + "epoch": 0.8803249002267972, + "grad_norm": 0.5715828537940979, + "learning_rate": 6.987467036607831e-06, + "loss": 0.1267, + "step": 37554 + }, + { + "epoch": 0.8803483418016023, + "grad_norm": 0.5068187713623047, + "learning_rate": 6.984762124568744e-06, + "loss": 0.0844, + "step": 37555 + }, + { + "epoch": 0.8803717833764072, + "grad_norm": 0.13264812529087067, + "learning_rate": 6.982057717232971e-06, + "loss": 0.0177, + "step": 37556 + }, + { + "epoch": 0.8803952249512123, + "grad_norm": 0.31769850850105286, + "learning_rate": 6.979353814615164e-06, + "loss": 0.0388, + "step": 37557 + }, + { + "epoch": 0.8804186665260172, + "grad_norm": 0.415435791015625, + "learning_rate": 6.97665041672998e-06, + "loss": 0.0489, + "step": 37558 + }, + { + "epoch": 0.8804421081008222, + "grad_norm": 0.2532854974269867, + "learning_rate": 6.973947523592106e-06, + "loss": 0.0488, + "step": 37559 + }, + { + "epoch": 0.8804655496756272, + "grad_norm": 0.6018964052200317, + "learning_rate": 6.97124513521622e-06, + "loss": 0.525, + "step": 37560 + }, + { + "epoch": 0.8804889912504322, + "grad_norm": 0.2727065682411194, + "learning_rate": 6.968543251616977e-06, + "loss": 0.0506, + "step": 37561 + }, + { + "epoch": 0.8805124328252372, + "grad_norm": 0.1692141890525818, + "learning_rate": 6.965841872809042e-06, + "loss": 0.0193, + "step": 37562 + }, + { + "epoch": 0.8805358744000422, + "grad_norm": 0.23310653865337372, + "learning_rate": 6.96314099880705e-06, + "loss": 0.0396, + "step": 37563 + }, + { + "epoch": 0.8805593159748472, + "grad_norm": 0.2534494996070862, + "learning_rate": 6.9604406296256645e-06, + "loss": 0.0377, + "step": 37564 + }, + { + "epoch": 0.8805827575496522, + "grad_norm": 0.11456375569105148, + "learning_rate": 6.957740765279563e-06, + "loss": 0.017, + "step": 37565 + }, + { + "epoch": 0.8806061991244571, + "grad_norm": 0.3551406264305115, + "learning_rate": 6.955041405783369e-06, + "loss": 0.07, + "step": 37566 + }, + { + "epoch": 0.8806296406992622, + "grad_norm": 0.3576451241970062, + "learning_rate": 6.952342551151747e-06, + "loss": 0.0319, + "step": 37567 + }, + { + "epoch": 0.8806530822740671, + "grad_norm": 0.5043984651565552, + "learning_rate": 6.949644201399319e-06, + "loss": 0.0838, + "step": 37568 + }, + { + "epoch": 0.8806765238488722, + "grad_norm": 0.3884619474411011, + "learning_rate": 6.94694635654074e-06, + "loss": 0.0532, + "step": 37569 + }, + { + "epoch": 0.8806999654236771, + "grad_norm": 0.19796165823936462, + "learning_rate": 6.944249016590631e-06, + "loss": 0.0304, + "step": 37570 + }, + { + "epoch": 0.8807234069984822, + "grad_norm": 0.1525653898715973, + "learning_rate": 6.94155218156366e-06, + "loss": 0.0336, + "step": 37571 + }, + { + "epoch": 0.8807468485732871, + "grad_norm": 0.37698882818222046, + "learning_rate": 6.938855851474435e-06, + "loss": 0.0414, + "step": 37572 + }, + { + "epoch": 0.8807702901480922, + "grad_norm": 0.261355996131897, + "learning_rate": 6.936160026337601e-06, + "loss": 0.039, + "step": 37573 + }, + { + "epoch": 0.8807937317228971, + "grad_norm": 0.27649641036987305, + "learning_rate": 6.9334647061677806e-06, + "loss": 0.0498, + "step": 37574 + }, + { + "epoch": 0.8808171732977021, + "grad_norm": 0.2949145436286926, + "learning_rate": 6.930769890979605e-06, + "loss": 0.0464, + "step": 37575 + }, + { + "epoch": 0.8808406148725071, + "grad_norm": 0.34018293023109436, + "learning_rate": 6.928075580787685e-06, + "loss": 0.0355, + "step": 37576 + }, + { + "epoch": 0.8808640564473121, + "grad_norm": 0.10777325183153152, + "learning_rate": 6.925381775606621e-06, + "loss": 0.0123, + "step": 37577 + }, + { + "epoch": 0.8808874980221171, + "grad_norm": 0.6330381035804749, + "learning_rate": 6.922688475451078e-06, + "loss": 0.1333, + "step": 37578 + }, + { + "epoch": 0.8809109395969221, + "grad_norm": 0.3602442145347595, + "learning_rate": 6.919995680335634e-06, + "loss": 0.0867, + "step": 37579 + }, + { + "epoch": 0.8809343811717271, + "grad_norm": 0.32639262080192566, + "learning_rate": 6.917303390274921e-06, + "loss": 0.0493, + "step": 37580 + }, + { + "epoch": 0.8809578227465321, + "grad_norm": 0.610849142074585, + "learning_rate": 6.914611605283539e-06, + "loss": 0.1137, + "step": 37581 + }, + { + "epoch": 0.880981264321337, + "grad_norm": 0.47356879711151123, + "learning_rate": 6.911920325376076e-06, + "loss": 0.0441, + "step": 37582 + }, + { + "epoch": 0.8810047058961421, + "grad_norm": 0.2916248142719269, + "learning_rate": 6.9092295505671535e-06, + "loss": 0.0432, + "step": 37583 + }, + { + "epoch": 0.881028147470947, + "grad_norm": 0.3226188123226166, + "learning_rate": 6.906539280871361e-06, + "loss": 0.0578, + "step": 37584 + }, + { + "epoch": 0.8810515890457521, + "grad_norm": 0.5021461844444275, + "learning_rate": 6.903849516303318e-06, + "loss": 0.0741, + "step": 37585 + }, + { + "epoch": 0.8810750306205571, + "grad_norm": 0.15688085556030273, + "learning_rate": 6.901160256877603e-06, + "loss": 0.0232, + "step": 37586 + }, + { + "epoch": 0.8810984721953621, + "grad_norm": 0.5916434526443481, + "learning_rate": 6.898471502608816e-06, + "loss": 0.1342, + "step": 37587 + }, + { + "epoch": 0.8811219137701671, + "grad_norm": 0.3323461711406708, + "learning_rate": 6.8957832535115345e-06, + "loss": 0.044, + "step": 37588 + }, + { + "epoch": 0.8811453553449721, + "grad_norm": 0.33886098861694336, + "learning_rate": 6.893095509600356e-06, + "loss": 0.0718, + "step": 37589 + }, + { + "epoch": 0.8811687969197771, + "grad_norm": 0.05231528729200363, + "learning_rate": 6.890408270889837e-06, + "loss": 0.0036, + "step": 37590 + }, + { + "epoch": 0.8811922384945821, + "grad_norm": 0.16352052986621857, + "learning_rate": 6.887721537394598e-06, + "loss": 0.0354, + "step": 37591 + }, + { + "epoch": 0.8812156800693871, + "grad_norm": 0.1880168318748474, + "learning_rate": 6.885035309129207e-06, + "loss": 0.0227, + "step": 37592 + }, + { + "epoch": 0.881239121644192, + "grad_norm": 0.46395599842071533, + "learning_rate": 6.882349586108227e-06, + "loss": 0.1212, + "step": 37593 + }, + { + "epoch": 0.8812625632189971, + "grad_norm": 0.5702337622642517, + "learning_rate": 6.879664368346239e-06, + "loss": 0.6816, + "step": 37594 + }, + { + "epoch": 0.881286004793802, + "grad_norm": 0.10587790608406067, + "learning_rate": 6.876979655857807e-06, + "loss": 0.0158, + "step": 37595 + }, + { + "epoch": 0.8813094463686071, + "grad_norm": 0.32716962695121765, + "learning_rate": 6.8742954486574864e-06, + "loss": 0.0593, + "step": 37596 + }, + { + "epoch": 0.881332887943412, + "grad_norm": 0.4094085097312927, + "learning_rate": 6.871611746759865e-06, + "loss": 0.0575, + "step": 37597 + }, + { + "epoch": 0.8813563295182171, + "grad_norm": 0.09573297202587128, + "learning_rate": 6.86892855017951e-06, + "loss": 0.0178, + "step": 37598 + }, + { + "epoch": 0.881379771093022, + "grad_norm": 0.19149836897850037, + "learning_rate": 6.866245858930964e-06, + "loss": 0.0383, + "step": 37599 + }, + { + "epoch": 0.8814032126678271, + "grad_norm": 0.537799060344696, + "learning_rate": 6.8635636730287835e-06, + "loss": 0.0889, + "step": 37600 + }, + { + "epoch": 0.881426654242632, + "grad_norm": 0.41615861654281616, + "learning_rate": 6.860881992487522e-06, + "loss": 0.0601, + "step": 37601 + }, + { + "epoch": 0.881450095817437, + "grad_norm": 0.509280800819397, + "learning_rate": 6.858200817321725e-06, + "loss": 0.0572, + "step": 37602 + }, + { + "epoch": 0.881473537392242, + "grad_norm": 0.8155250549316406, + "learning_rate": 6.855520147545969e-06, + "loss": 0.179, + "step": 37603 + }, + { + "epoch": 0.881496978967047, + "grad_norm": 0.5091949105262756, + "learning_rate": 6.852839983174775e-06, + "loss": 0.0952, + "step": 37604 + }, + { + "epoch": 0.881520420541852, + "grad_norm": 0.3238702416419983, + "learning_rate": 6.850160324222687e-06, + "loss": 0.0385, + "step": 37605 + }, + { + "epoch": 0.881543862116657, + "grad_norm": 0.5616970062255859, + "learning_rate": 6.8474811707042505e-06, + "loss": 0.0487, + "step": 37606 + }, + { + "epoch": 0.881567303691462, + "grad_norm": 0.42523348331451416, + "learning_rate": 6.844802522634008e-06, + "loss": 0.0361, + "step": 37607 + }, + { + "epoch": 0.881590745266267, + "grad_norm": 0.14788146317005157, + "learning_rate": 6.84212438002646e-06, + "loss": 0.026, + "step": 37608 + }, + { + "epoch": 0.881614186841072, + "grad_norm": 0.3319612741470337, + "learning_rate": 6.8394467428961715e-06, + "loss": 0.0541, + "step": 37609 + }, + { + "epoch": 0.881637628415877, + "grad_norm": 0.8254494667053223, + "learning_rate": 6.836769611257676e-06, + "loss": 0.1738, + "step": 37610 + }, + { + "epoch": 0.881661069990682, + "grad_norm": 0.4768473207950592, + "learning_rate": 6.834092985125484e-06, + "loss": 0.4432, + "step": 37611 + }, + { + "epoch": 0.881684511565487, + "grad_norm": 0.17247521877288818, + "learning_rate": 6.831416864514128e-06, + "loss": 0.0365, + "step": 37612 + }, + { + "epoch": 0.8817079531402919, + "grad_norm": 0.4166991412639618, + "learning_rate": 6.828741249438131e-06, + "loss": 0.0858, + "step": 37613 + }, + { + "epoch": 0.881731394715097, + "grad_norm": 0.5789377093315125, + "learning_rate": 6.826066139911991e-06, + "loss": 0.0885, + "step": 37614 + }, + { + "epoch": 0.8817548362899019, + "grad_norm": 0.36648741364479065, + "learning_rate": 6.82339153595023e-06, + "loss": 0.0301, + "step": 37615 + }, + { + "epoch": 0.881778277864707, + "grad_norm": 0.5270841121673584, + "learning_rate": 6.820717437567381e-06, + "loss": 0.1118, + "step": 37616 + }, + { + "epoch": 0.8818017194395119, + "grad_norm": 0.5079622864723206, + "learning_rate": 6.818043844777933e-06, + "loss": 0.0761, + "step": 37617 + }, + { + "epoch": 0.881825161014317, + "grad_norm": 0.40685904026031494, + "learning_rate": 6.815370757596407e-06, + "loss": 0.0766, + "step": 37618 + }, + { + "epoch": 0.8818486025891219, + "grad_norm": 0.5376818180084229, + "learning_rate": 6.812698176037291e-06, + "loss": 0.0685, + "step": 37619 + }, + { + "epoch": 0.881872044163927, + "grad_norm": 0.6055702567100525, + "learning_rate": 6.810026100115086e-06, + "loss": 0.0891, + "step": 37620 + }, + { + "epoch": 0.8818954857387319, + "grad_norm": 0.358107328414917, + "learning_rate": 6.807354529844312e-06, + "loss": 0.0282, + "step": 37621 + }, + { + "epoch": 0.8819189273135369, + "grad_norm": 0.4591411352157593, + "learning_rate": 6.804683465239437e-06, + "loss": 0.0447, + "step": 37622 + }, + { + "epoch": 0.8819423688883419, + "grad_norm": 0.2310512214899063, + "learning_rate": 6.8020129063149805e-06, + "loss": 0.0255, + "step": 37623 + }, + { + "epoch": 0.8819658104631469, + "grad_norm": 0.34591197967529297, + "learning_rate": 6.799342853085433e-06, + "loss": 0.0544, + "step": 37624 + }, + { + "epoch": 0.8819892520379519, + "grad_norm": 0.35722410678863525, + "learning_rate": 6.7966733055652595e-06, + "loss": 0.0665, + "step": 37625 + }, + { + "epoch": 0.8820126936127569, + "grad_norm": 0.27710866928100586, + "learning_rate": 6.794004263768972e-06, + "loss": 0.0326, + "step": 37626 + }, + { + "epoch": 0.8820361351875619, + "grad_norm": 0.4346937835216522, + "learning_rate": 6.791335727711023e-06, + "loss": 0.1015, + "step": 37627 + }, + { + "epoch": 0.8820595767623669, + "grad_norm": 0.3873552680015564, + "learning_rate": 6.788667697405904e-06, + "loss": 0.0458, + "step": 37628 + }, + { + "epoch": 0.8820830183371718, + "grad_norm": 0.36285313963890076, + "learning_rate": 6.786000172868101e-06, + "loss": 0.065, + "step": 37629 + }, + { + "epoch": 0.8821064599119769, + "grad_norm": 0.4679648280143738, + "learning_rate": 6.783333154112093e-06, + "loss": 0.0781, + "step": 37630 + }, + { + "epoch": 0.8821299014867818, + "grad_norm": 0.3255597651004791, + "learning_rate": 6.7806666411523335e-06, + "loss": 0.0755, + "step": 37631 + }, + { + "epoch": 0.8821533430615869, + "grad_norm": 0.31678473949432373, + "learning_rate": 6.7780006340032786e-06, + "loss": 0.0514, + "step": 37632 + }, + { + "epoch": 0.8821767846363918, + "grad_norm": 0.3077729046344757, + "learning_rate": 6.775335132679439e-06, + "loss": 0.0829, + "step": 37633 + }, + { + "epoch": 0.8822002262111969, + "grad_norm": 0.5678882002830505, + "learning_rate": 6.772670137195247e-06, + "loss": 0.0858, + "step": 37634 + }, + { + "epoch": 0.8822236677860018, + "grad_norm": 0.7935620546340942, + "learning_rate": 6.770005647565147e-06, + "loss": 0.0968, + "step": 37635 + }, + { + "epoch": 0.8822471093608069, + "grad_norm": 0.4425513446331024, + "learning_rate": 6.767341663803639e-06, + "loss": 0.0959, + "step": 37636 + }, + { + "epoch": 0.8822705509356119, + "grad_norm": 0.12749512493610382, + "learning_rate": 6.764678185925155e-06, + "loss": 0.0148, + "step": 37637 + }, + { + "epoch": 0.8822939925104168, + "grad_norm": 0.33462682366371155, + "learning_rate": 6.762015213944151e-06, + "loss": 0.0493, + "step": 37638 + }, + { + "epoch": 0.8823174340852219, + "grad_norm": 0.4705902636051178, + "learning_rate": 6.7593527478750696e-06, + "loss": 0.0436, + "step": 37639 + }, + { + "epoch": 0.8823408756600268, + "grad_norm": 0.38159647583961487, + "learning_rate": 6.756690787732356e-06, + "loss": 0.0476, + "step": 37640 + }, + { + "epoch": 0.8823643172348319, + "grad_norm": 0.35347625613212585, + "learning_rate": 6.754029333530465e-06, + "loss": 0.0347, + "step": 37641 + }, + { + "epoch": 0.8823877588096368, + "grad_norm": 0.14875981211662292, + "learning_rate": 6.75136838528384e-06, + "loss": 0.0157, + "step": 37642 + }, + { + "epoch": 0.8824112003844419, + "grad_norm": 0.4888702929019928, + "learning_rate": 6.748707943006904e-06, + "loss": 0.0944, + "step": 37643 + }, + { + "epoch": 0.8824346419592468, + "grad_norm": 0.5739205479621887, + "learning_rate": 6.74604800671409e-06, + "loss": 0.087, + "step": 37644 + }, + { + "epoch": 0.8824580835340519, + "grad_norm": 0.3572051525115967, + "learning_rate": 6.743388576419851e-06, + "loss": 0.4598, + "step": 37645 + }, + { + "epoch": 0.8824815251088568, + "grad_norm": 0.21998143196105957, + "learning_rate": 6.740729652138622e-06, + "loss": 0.0442, + "step": 37646 + }, + { + "epoch": 0.8825049666836619, + "grad_norm": 0.23006854951381683, + "learning_rate": 6.73807123388479e-06, + "loss": 0.0316, + "step": 37647 + }, + { + "epoch": 0.8825284082584668, + "grad_norm": 0.38922399282455444, + "learning_rate": 6.735413321672823e-06, + "loss": 0.0729, + "step": 37648 + }, + { + "epoch": 0.8825518498332718, + "grad_norm": 0.3807355761528015, + "learning_rate": 6.73275591551713e-06, + "loss": 0.0683, + "step": 37649 + }, + { + "epoch": 0.8825752914080768, + "grad_norm": 0.6198509335517883, + "learning_rate": 6.730099015432123e-06, + "loss": 0.6035, + "step": 37650 + }, + { + "epoch": 0.8825987329828818, + "grad_norm": 0.3704826235771179, + "learning_rate": 6.727442621432212e-06, + "loss": 0.0592, + "step": 37651 + }, + { + "epoch": 0.8826221745576868, + "grad_norm": 0.17041875422000885, + "learning_rate": 6.7247867335318296e-06, + "loss": 0.0174, + "step": 37652 + }, + { + "epoch": 0.8826456161324918, + "grad_norm": 0.15346205234527588, + "learning_rate": 6.722131351745365e-06, + "loss": 0.0101, + "step": 37653 + }, + { + "epoch": 0.8826690577072968, + "grad_norm": 0.34536492824554443, + "learning_rate": 6.71947647608725e-06, + "loss": 0.0501, + "step": 37654 + }, + { + "epoch": 0.8826924992821018, + "grad_norm": 0.16687732934951782, + "learning_rate": 6.716822106571874e-06, + "loss": 0.0293, + "step": 37655 + }, + { + "epoch": 0.8827159408569067, + "grad_norm": 0.5756327509880066, + "learning_rate": 6.714168243213648e-06, + "loss": 0.1125, + "step": 37656 + }, + { + "epoch": 0.8827393824317118, + "grad_norm": 0.35199692845344543, + "learning_rate": 6.711514886026948e-06, + "loss": 0.1042, + "step": 37657 + }, + { + "epoch": 0.8827628240065167, + "grad_norm": 0.6353052258491516, + "learning_rate": 6.70886203502622e-06, + "loss": 0.0447, + "step": 37658 + }, + { + "epoch": 0.8827862655813218, + "grad_norm": 0.37009426951408386, + "learning_rate": 6.706209690225818e-06, + "loss": 0.04, + "step": 37659 + }, + { + "epoch": 0.8828097071561267, + "grad_norm": 0.5511947274208069, + "learning_rate": 6.70355785164013e-06, + "loss": 0.0972, + "step": 37660 + }, + { + "epoch": 0.8828331487309318, + "grad_norm": 0.4631982147693634, + "learning_rate": 6.700906519283578e-06, + "loss": 0.0367, + "step": 37661 + }, + { + "epoch": 0.8828565903057367, + "grad_norm": 0.3895622789859772, + "learning_rate": 6.69825569317053e-06, + "loss": 0.0551, + "step": 37662 + }, + { + "epoch": 0.8828800318805418, + "grad_norm": 0.21837545931339264, + "learning_rate": 6.695605373315372e-06, + "loss": 0.0407, + "step": 37663 + }, + { + "epoch": 0.8829034734553467, + "grad_norm": 0.07993350178003311, + "learning_rate": 6.692955559732483e-06, + "loss": 0.007, + "step": 37664 + }, + { + "epoch": 0.8829269150301517, + "grad_norm": 0.26445433497428894, + "learning_rate": 6.69030625243624e-06, + "loss": 0.0371, + "step": 37665 + }, + { + "epoch": 0.8829503566049567, + "grad_norm": 0.5619495511054993, + "learning_rate": 6.687657451440999e-06, + "loss": 0.663, + "step": 37666 + }, + { + "epoch": 0.8829737981797617, + "grad_norm": 0.22371616959571838, + "learning_rate": 6.6850091567611795e-06, + "loss": 0.0237, + "step": 37667 + }, + { + "epoch": 0.8829972397545667, + "grad_norm": 0.3933281898498535, + "learning_rate": 6.682361368411116e-06, + "loss": 0.0323, + "step": 37668 + }, + { + "epoch": 0.8830206813293717, + "grad_norm": 0.3856077790260315, + "learning_rate": 6.679714086405164e-06, + "loss": 0.0248, + "step": 37669 + }, + { + "epoch": 0.8830441229041767, + "grad_norm": 0.2683209180831909, + "learning_rate": 6.677067310757734e-06, + "loss": 0.0571, + "step": 37670 + }, + { + "epoch": 0.8830675644789817, + "grad_norm": 0.5727317333221436, + "learning_rate": 6.674421041483148e-06, + "loss": 0.116, + "step": 37671 + }, + { + "epoch": 0.8830910060537867, + "grad_norm": 0.3385602533817291, + "learning_rate": 6.671775278595782e-06, + "loss": 0.0818, + "step": 37672 + }, + { + "epoch": 0.8831144476285917, + "grad_norm": 0.6234508156776428, + "learning_rate": 6.669130022109981e-06, + "loss": 0.1268, + "step": 37673 + }, + { + "epoch": 0.8831378892033966, + "grad_norm": 0.4016469717025757, + "learning_rate": 6.6664852720401125e-06, + "loss": 0.0492, + "step": 37674 + }, + { + "epoch": 0.8831613307782017, + "grad_norm": 0.432687371969223, + "learning_rate": 6.663841028400519e-06, + "loss": 0.072, + "step": 37675 + }, + { + "epoch": 0.8831847723530066, + "grad_norm": 0.3586123287677765, + "learning_rate": 6.6611972912055455e-06, + "loss": 0.0427, + "step": 37676 + }, + { + "epoch": 0.8832082139278117, + "grad_norm": 0.32626819610595703, + "learning_rate": 6.658554060469546e-06, + "loss": 0.0773, + "step": 37677 + }, + { + "epoch": 0.8832316555026166, + "grad_norm": 0.5026081204414368, + "learning_rate": 6.6559113362068324e-06, + "loss": 0.0415, + "step": 37678 + }, + { + "epoch": 0.8832550970774217, + "grad_norm": 0.5671998858451843, + "learning_rate": 6.653269118431793e-06, + "loss": 0.0749, + "step": 37679 + }, + { + "epoch": 0.8832785386522266, + "grad_norm": 0.44621026515960693, + "learning_rate": 6.6506274071587385e-06, + "loss": 0.4845, + "step": 37680 + }, + { + "epoch": 0.8833019802270317, + "grad_norm": 0.14812518656253815, + "learning_rate": 6.64798620240199e-06, + "loss": 0.0139, + "step": 37681 + }, + { + "epoch": 0.8833254218018366, + "grad_norm": 0.4865974187850952, + "learning_rate": 6.645345504175904e-06, + "loss": 0.0655, + "step": 37682 + }, + { + "epoch": 0.8833488633766416, + "grad_norm": 0.23341195285320282, + "learning_rate": 6.642705312494802e-06, + "loss": 0.0314, + "step": 37683 + }, + { + "epoch": 0.8833723049514466, + "grad_norm": 0.24496859312057495, + "learning_rate": 6.640065627373005e-06, + "loss": 0.0416, + "step": 37684 + }, + { + "epoch": 0.8833957465262516, + "grad_norm": 0.7470492124557495, + "learning_rate": 6.637426448824824e-06, + "loss": 0.1739, + "step": 37685 + }, + { + "epoch": 0.8834191881010566, + "grad_norm": 0.5624990463256836, + "learning_rate": 6.634787776864604e-06, + "loss": 0.1218, + "step": 37686 + }, + { + "epoch": 0.8834426296758616, + "grad_norm": 0.460431307554245, + "learning_rate": 6.632149611506644e-06, + "loss": 0.0557, + "step": 37687 + }, + { + "epoch": 0.8834660712506667, + "grad_norm": 0.4310734272003174, + "learning_rate": 6.629511952765277e-06, + "loss": 0.0409, + "step": 37688 + }, + { + "epoch": 0.8834895128254716, + "grad_norm": 0.7024415135383606, + "learning_rate": 6.626874800654803e-06, + "loss": 0.1608, + "step": 37689 + }, + { + "epoch": 0.8835129544002767, + "grad_norm": 0.23379428684711456, + "learning_rate": 6.6242381551895216e-06, + "loss": 0.0387, + "step": 37690 + }, + { + "epoch": 0.8835363959750816, + "grad_norm": 0.11974494159221649, + "learning_rate": 6.621602016383743e-06, + "loss": 0.015, + "step": 37691 + }, + { + "epoch": 0.8835598375498867, + "grad_norm": 0.37726646661758423, + "learning_rate": 6.61896638425179e-06, + "loss": 0.0415, + "step": 37692 + }, + { + "epoch": 0.8835832791246916, + "grad_norm": 0.484082967042923, + "learning_rate": 6.616331258807939e-06, + "loss": 0.0848, + "step": 37693 + }, + { + "epoch": 0.8836067206994966, + "grad_norm": 0.33838802576065063, + "learning_rate": 6.613696640066514e-06, + "loss": 0.0601, + "step": 37694 + }, + { + "epoch": 0.8836301622743016, + "grad_norm": 0.2152526080608368, + "learning_rate": 6.611062528041789e-06, + "loss": 0.0412, + "step": 37695 + }, + { + "epoch": 0.8836536038491066, + "grad_norm": 0.4872557818889618, + "learning_rate": 6.608428922748078e-06, + "loss": 0.0572, + "step": 37696 + }, + { + "epoch": 0.8836770454239116, + "grad_norm": 0.2560877799987793, + "learning_rate": 6.605795824199645e-06, + "loss": 0.0271, + "step": 37697 + }, + { + "epoch": 0.8837004869987166, + "grad_norm": 0.4009300470352173, + "learning_rate": 6.603163232410781e-06, + "loss": 0.1037, + "step": 37698 + }, + { + "epoch": 0.8837239285735216, + "grad_norm": 0.5737979412078857, + "learning_rate": 6.600531147395794e-06, + "loss": 0.6426, + "step": 37699 + }, + { + "epoch": 0.8837473701483266, + "grad_norm": 0.28936246037483215, + "learning_rate": 6.597899569168953e-06, + "loss": 0.0442, + "step": 37700 + }, + { + "epoch": 0.8837708117231315, + "grad_norm": 0.14521074295043945, + "learning_rate": 6.595268497744533e-06, + "loss": 0.0203, + "step": 37701 + }, + { + "epoch": 0.8837942532979366, + "grad_norm": 0.12988077104091644, + "learning_rate": 6.592637933136814e-06, + "loss": 0.0118, + "step": 37702 + }, + { + "epoch": 0.8838176948727415, + "grad_norm": 0.3115506172180176, + "learning_rate": 6.5900078753600715e-06, + "loss": 0.0584, + "step": 37703 + }, + { + "epoch": 0.8838411364475466, + "grad_norm": 0.513746976852417, + "learning_rate": 6.58737832442855e-06, + "loss": 0.0544, + "step": 37704 + }, + { + "epoch": 0.8838645780223515, + "grad_norm": 0.6915988326072693, + "learning_rate": 6.584749280356539e-06, + "loss": 0.1249, + "step": 37705 + }, + { + "epoch": 0.8838880195971566, + "grad_norm": 0.41655269265174866, + "learning_rate": 6.582120743158326e-06, + "loss": 0.0457, + "step": 37706 + }, + { + "epoch": 0.8839114611719615, + "grad_norm": 0.17537572979927063, + "learning_rate": 6.579492712848156e-06, + "loss": 0.0256, + "step": 37707 + }, + { + "epoch": 0.8839349027467666, + "grad_norm": 0.6494714021682739, + "learning_rate": 6.576865189440274e-06, + "loss": 0.1098, + "step": 37708 + }, + { + "epoch": 0.8839583443215715, + "grad_norm": 0.4072454571723938, + "learning_rate": 6.574238172948954e-06, + "loss": 0.0349, + "step": 37709 + }, + { + "epoch": 0.8839817858963765, + "grad_norm": 0.4164850413799286, + "learning_rate": 6.571611663388444e-06, + "loss": 0.0968, + "step": 37710 + }, + { + "epoch": 0.8840052274711815, + "grad_norm": 0.8412789702415466, + "learning_rate": 6.568985660772986e-06, + "loss": 0.1311, + "step": 37711 + }, + { + "epoch": 0.8840286690459865, + "grad_norm": 0.35244131088256836, + "learning_rate": 6.566360165116847e-06, + "loss": 0.0746, + "step": 37712 + }, + { + "epoch": 0.8840521106207915, + "grad_norm": 0.3470543622970581, + "learning_rate": 6.563735176434271e-06, + "loss": 0.0589, + "step": 37713 + }, + { + "epoch": 0.8840755521955965, + "grad_norm": 0.09868025034666061, + "learning_rate": 6.561110694739492e-06, + "loss": 0.0171, + "step": 37714 + }, + { + "epoch": 0.8840989937704015, + "grad_norm": 0.31268271803855896, + "learning_rate": 6.558486720046753e-06, + "loss": 0.053, + "step": 37715 + }, + { + "epoch": 0.8841224353452065, + "grad_norm": 0.6053199172019958, + "learning_rate": 6.555863252370276e-06, + "loss": 0.5264, + "step": 37716 + }, + { + "epoch": 0.8841458769200115, + "grad_norm": 0.7520334124565125, + "learning_rate": 6.5532402917243275e-06, + "loss": 0.0811, + "step": 37717 + }, + { + "epoch": 0.8841693184948165, + "grad_norm": 0.2720091640949249, + "learning_rate": 6.550617838123119e-06, + "loss": 0.0338, + "step": 37718 + }, + { + "epoch": 0.8841927600696214, + "grad_norm": 0.475040078163147, + "learning_rate": 6.547995891580894e-06, + "loss": 0.551, + "step": 37719 + }, + { + "epoch": 0.8842162016444265, + "grad_norm": 0.12211108952760696, + "learning_rate": 6.5453744521118635e-06, + "loss": 0.0295, + "step": 37720 + }, + { + "epoch": 0.8842396432192314, + "grad_norm": 0.23581622540950775, + "learning_rate": 6.542753519730271e-06, + "loss": 0.024, + "step": 37721 + }, + { + "epoch": 0.8842630847940365, + "grad_norm": 0.2840246260166168, + "learning_rate": 6.540133094450318e-06, + "loss": 0.074, + "step": 37722 + }, + { + "epoch": 0.8842865263688414, + "grad_norm": 0.10924121737480164, + "learning_rate": 6.537513176286214e-06, + "loss": 0.0171, + "step": 37723 + }, + { + "epoch": 0.8843099679436465, + "grad_norm": 0.6863338947296143, + "learning_rate": 6.534893765252214e-06, + "loss": 0.1033, + "step": 37724 + }, + { + "epoch": 0.8843334095184514, + "grad_norm": 0.3909788727760315, + "learning_rate": 6.532274861362508e-06, + "loss": 0.0537, + "step": 37725 + }, + { + "epoch": 0.8843568510932565, + "grad_norm": 0.3304767906665802, + "learning_rate": 6.529656464631295e-06, + "loss": 0.0529, + "step": 37726 + }, + { + "epoch": 0.8843802926680614, + "grad_norm": 0.18261033296585083, + "learning_rate": 6.527038575072808e-06, + "loss": 0.0213, + "step": 37727 + }, + { + "epoch": 0.8844037342428664, + "grad_norm": 0.2085266411304474, + "learning_rate": 6.524421192701235e-06, + "loss": 0.038, + "step": 37728 + }, + { + "epoch": 0.8844271758176714, + "grad_norm": 0.496845543384552, + "learning_rate": 6.521804317530755e-06, + "loss": 0.0547, + "step": 37729 + }, + { + "epoch": 0.8844506173924764, + "grad_norm": 0.24587756395339966, + "learning_rate": 6.519187949575612e-06, + "loss": 0.0544, + "step": 37730 + }, + { + "epoch": 0.8844740589672814, + "grad_norm": 0.33608904480934143, + "learning_rate": 6.5165720888499814e-06, + "loss": 0.0424, + "step": 37731 + }, + { + "epoch": 0.8844975005420864, + "grad_norm": 0.3901631236076355, + "learning_rate": 6.5139567353680656e-06, + "loss": 0.0682, + "step": 37732 + }, + { + "epoch": 0.8845209421168914, + "grad_norm": 0.1702558994293213, + "learning_rate": 6.511341889144051e-06, + "loss": 0.0222, + "step": 37733 + }, + { + "epoch": 0.8845443836916964, + "grad_norm": 0.059705931693315506, + "learning_rate": 6.5087275501921285e-06, + "loss": 0.0043, + "step": 37734 + }, + { + "epoch": 0.8845678252665014, + "grad_norm": 0.27656587958335876, + "learning_rate": 6.5061137185264634e-06, + "loss": 0.0313, + "step": 37735 + }, + { + "epoch": 0.8845912668413064, + "grad_norm": 0.22177061438560486, + "learning_rate": 6.503500394161255e-06, + "loss": 0.0141, + "step": 37736 + }, + { + "epoch": 0.8846147084161113, + "grad_norm": 0.2813240885734558, + "learning_rate": 6.500887577110693e-06, + "loss": 0.0441, + "step": 37737 + }, + { + "epoch": 0.8846381499909164, + "grad_norm": 0.5073275566101074, + "learning_rate": 6.4982752673889426e-06, + "loss": 0.0471, + "step": 37738 + }, + { + "epoch": 0.8846615915657214, + "grad_norm": 0.18934497237205505, + "learning_rate": 6.4956634650101714e-06, + "loss": 0.0344, + "step": 37739 + }, + { + "epoch": 0.8846850331405264, + "grad_norm": 0.4297195374965668, + "learning_rate": 6.493052169988567e-06, + "loss": 0.0836, + "step": 37740 + }, + { + "epoch": 0.8847084747153314, + "grad_norm": 0.481006383895874, + "learning_rate": 6.490441382338297e-06, + "loss": 0.0789, + "step": 37741 + }, + { + "epoch": 0.8847319162901364, + "grad_norm": 0.6858729720115662, + "learning_rate": 6.487831102073494e-06, + "loss": 0.4441, + "step": 37742 + }, + { + "epoch": 0.8847553578649414, + "grad_norm": 0.6735512614250183, + "learning_rate": 6.485221329208357e-06, + "loss": 0.0631, + "step": 37743 + }, + { + "epoch": 0.8847787994397464, + "grad_norm": 0.2818768322467804, + "learning_rate": 6.482612063757054e-06, + "loss": 0.0377, + "step": 37744 + }, + { + "epoch": 0.8848022410145514, + "grad_norm": 0.2375541627407074, + "learning_rate": 6.480003305733728e-06, + "loss": 0.0448, + "step": 37745 + }, + { + "epoch": 0.8848256825893563, + "grad_norm": 0.33060765266418457, + "learning_rate": 6.477395055152524e-06, + "loss": 0.0572, + "step": 37746 + }, + { + "epoch": 0.8848491241641614, + "grad_norm": 0.068604975938797, + "learning_rate": 6.474787312027608e-06, + "loss": 0.0056, + "step": 37747 + }, + { + "epoch": 0.8848725657389663, + "grad_norm": 0.08296869695186615, + "learning_rate": 6.472180076373136e-06, + "loss": 0.0148, + "step": 37748 + }, + { + "epoch": 0.8848960073137714, + "grad_norm": 0.28810447454452515, + "learning_rate": 6.469573348203217e-06, + "loss": 0.0252, + "step": 37749 + }, + { + "epoch": 0.8849194488885763, + "grad_norm": 0.5326131582260132, + "learning_rate": 6.466967127532042e-06, + "loss": 0.0595, + "step": 37750 + }, + { + "epoch": 0.8849428904633814, + "grad_norm": 0.37273842096328735, + "learning_rate": 6.464361414373732e-06, + "loss": 0.0241, + "step": 37751 + }, + { + "epoch": 0.8849663320381863, + "grad_norm": 0.5426522493362427, + "learning_rate": 6.461756208742431e-06, + "loss": 0.0667, + "step": 37752 + }, + { + "epoch": 0.8849897736129914, + "grad_norm": 0.5647262334823608, + "learning_rate": 6.4591515106522725e-06, + "loss": 0.1374, + "step": 37753 + }, + { + "epoch": 0.8850132151877963, + "grad_norm": 0.36329376697540283, + "learning_rate": 6.456547320117368e-06, + "loss": 0.0475, + "step": 37754 + }, + { + "epoch": 0.8850366567626013, + "grad_norm": 0.4617184102535248, + "learning_rate": 6.453943637151882e-06, + "loss": 0.0749, + "step": 37755 + }, + { + "epoch": 0.8850600983374063, + "grad_norm": 0.10460992902517319, + "learning_rate": 6.451340461769917e-06, + "loss": 0.01, + "step": 37756 + }, + { + "epoch": 0.8850835399122113, + "grad_norm": 0.5800383687019348, + "learning_rate": 6.448737793985626e-06, + "loss": 0.1187, + "step": 37757 + }, + { + "epoch": 0.8851069814870163, + "grad_norm": 0.12761826813220978, + "learning_rate": 6.446135633813111e-06, + "loss": 0.0267, + "step": 37758 + }, + { + "epoch": 0.8851304230618213, + "grad_norm": 0.5768454074859619, + "learning_rate": 6.4435339812664915e-06, + "loss": 0.437, + "step": 37759 + }, + { + "epoch": 0.8851538646366263, + "grad_norm": 0.10047203302383423, + "learning_rate": 6.440932836359892e-06, + "loss": 0.0136, + "step": 37760 + }, + { + "epoch": 0.8851773062114313, + "grad_norm": 0.5185726284980774, + "learning_rate": 6.4383321991074e-06, + "loss": 0.0701, + "step": 37761 + }, + { + "epoch": 0.8852007477862363, + "grad_norm": 0.3973225951194763, + "learning_rate": 6.43573206952317e-06, + "loss": 0.0801, + "step": 37762 + }, + { + "epoch": 0.8852241893610413, + "grad_norm": 0.35355913639068604, + "learning_rate": 6.433132447621282e-06, + "loss": 0.0533, + "step": 37763 + }, + { + "epoch": 0.8852476309358462, + "grad_norm": 0.12534642219543457, + "learning_rate": 6.430533333415856e-06, + "loss": 0.0181, + "step": 37764 + }, + { + "epoch": 0.8852710725106513, + "grad_norm": 0.5456512570381165, + "learning_rate": 6.427934726920981e-06, + "loss": 0.0859, + "step": 37765 + }, + { + "epoch": 0.8852945140854562, + "grad_norm": 0.4475696086883545, + "learning_rate": 6.425336628150758e-06, + "loss": 0.0913, + "step": 37766 + }, + { + "epoch": 0.8853179556602613, + "grad_norm": 0.39025843143463135, + "learning_rate": 6.4227390371192965e-06, + "loss": 0.0751, + "step": 37767 + }, + { + "epoch": 0.8853413972350662, + "grad_norm": 0.33926835656166077, + "learning_rate": 6.420141953840664e-06, + "loss": 0.0401, + "step": 37768 + }, + { + "epoch": 0.8853648388098713, + "grad_norm": 0.1870785653591156, + "learning_rate": 6.417545378328993e-06, + "loss": 0.0195, + "step": 37769 + }, + { + "epoch": 0.8853882803846762, + "grad_norm": 0.16733601689338684, + "learning_rate": 6.4149493105983506e-06, + "loss": 0.0383, + "step": 37770 + }, + { + "epoch": 0.8854117219594813, + "grad_norm": 0.3241589069366455, + "learning_rate": 6.412353750662836e-06, + "loss": 0.0252, + "step": 37771 + }, + { + "epoch": 0.8854351635342862, + "grad_norm": 0.31148406863212585, + "learning_rate": 6.4097586985365055e-06, + "loss": 0.0589, + "step": 37772 + }, + { + "epoch": 0.8854586051090912, + "grad_norm": 0.3202403485774994, + "learning_rate": 6.407164154233469e-06, + "loss": 0.0618, + "step": 37773 + }, + { + "epoch": 0.8854820466838962, + "grad_norm": 0.31854447722435, + "learning_rate": 6.404570117767761e-06, + "loss": 0.0274, + "step": 37774 + }, + { + "epoch": 0.8855054882587012, + "grad_norm": 0.46128010749816895, + "learning_rate": 6.401976589153513e-06, + "loss": 0.0668, + "step": 37775 + }, + { + "epoch": 0.8855289298335062, + "grad_norm": 0.5581173896789551, + "learning_rate": 6.3993835684047706e-06, + "loss": 0.0766, + "step": 37776 + }, + { + "epoch": 0.8855523714083112, + "grad_norm": 0.12330840528011322, + "learning_rate": 6.396791055535601e-06, + "loss": 0.0122, + "step": 37777 + }, + { + "epoch": 0.8855758129831162, + "grad_norm": 0.32906192541122437, + "learning_rate": 6.394199050560068e-06, + "loss": 0.0597, + "step": 37778 + }, + { + "epoch": 0.8855992545579212, + "grad_norm": 0.22445538640022278, + "learning_rate": 6.391607553492251e-06, + "loss": 0.0364, + "step": 37779 + }, + { + "epoch": 0.8856226961327262, + "grad_norm": 0.37171077728271484, + "learning_rate": 6.389016564346207e-06, + "loss": 0.0219, + "step": 37780 + }, + { + "epoch": 0.8856461377075312, + "grad_norm": 0.48217836022377014, + "learning_rate": 6.386426083135977e-06, + "loss": 0.0615, + "step": 37781 + }, + { + "epoch": 0.8856695792823361, + "grad_norm": 0.327399343252182, + "learning_rate": 6.383836109875641e-06, + "loss": 0.0547, + "step": 37782 + }, + { + "epoch": 0.8856930208571412, + "grad_norm": 0.5145132541656494, + "learning_rate": 6.381246644579242e-06, + "loss": 0.0761, + "step": 37783 + }, + { + "epoch": 0.8857164624319461, + "grad_norm": 0.6799772381782532, + "learning_rate": 6.378657687260836e-06, + "loss": 0.0674, + "step": 37784 + }, + { + "epoch": 0.8857399040067512, + "grad_norm": 0.424393892288208, + "learning_rate": 6.376069237934468e-06, + "loss": 0.1205, + "step": 37785 + }, + { + "epoch": 0.8857633455815561, + "grad_norm": 0.17462848126888275, + "learning_rate": 6.37348129661417e-06, + "loss": 0.0291, + "step": 37786 + }, + { + "epoch": 0.8857867871563612, + "grad_norm": 0.13667075335979462, + "learning_rate": 6.370893863313987e-06, + "loss": 0.0206, + "step": 37787 + }, + { + "epoch": 0.8858102287311661, + "grad_norm": 0.43186450004577637, + "learning_rate": 6.368306938047985e-06, + "loss": 0.0363, + "step": 37788 + }, + { + "epoch": 0.8858336703059712, + "grad_norm": 0.30448585748672485, + "learning_rate": 6.365720520830176e-06, + "loss": 0.0468, + "step": 37789 + }, + { + "epoch": 0.8858571118807762, + "grad_norm": 0.4278252124786377, + "learning_rate": 6.363134611674593e-06, + "loss": 0.039, + "step": 37790 + }, + { + "epoch": 0.8858805534555811, + "grad_norm": 0.3961513936519623, + "learning_rate": 6.3605492105952794e-06, + "loss": 0.4121, + "step": 37791 + }, + { + "epoch": 0.8859039950303862, + "grad_norm": 0.56386399269104, + "learning_rate": 6.357964317606269e-06, + "loss": 0.1179, + "step": 37792 + }, + { + "epoch": 0.8859274366051911, + "grad_norm": 0.6799221038818359, + "learning_rate": 6.355379932721561e-06, + "loss": 0.1119, + "step": 37793 + }, + { + "epoch": 0.8859508781799962, + "grad_norm": 0.08235321938991547, + "learning_rate": 6.35279605595519e-06, + "loss": 0.0044, + "step": 37794 + }, + { + "epoch": 0.8859743197548011, + "grad_norm": 0.3490946888923645, + "learning_rate": 6.3502126873212e-06, + "loss": 0.0471, + "step": 37795 + }, + { + "epoch": 0.8859977613296062, + "grad_norm": 0.3703860640525818, + "learning_rate": 6.34762982683359e-06, + "loss": 0.261, + "step": 37796 + }, + { + "epoch": 0.8860212029044111, + "grad_norm": 0.4700314402580261, + "learning_rate": 6.34504747450636e-06, + "loss": 0.0663, + "step": 37797 + }, + { + "epoch": 0.8860446444792162, + "grad_norm": 0.5129050612449646, + "learning_rate": 6.342465630353556e-06, + "loss": 0.427, + "step": 37798 + }, + { + "epoch": 0.8860680860540211, + "grad_norm": 0.348479300737381, + "learning_rate": 6.339884294389131e-06, + "loss": 0.0293, + "step": 37799 + }, + { + "epoch": 0.8860915276288261, + "grad_norm": 0.3734692633152008, + "learning_rate": 6.337303466627153e-06, + "loss": 0.0682, + "step": 37800 + }, + { + "epoch": 0.8861149692036311, + "grad_norm": 0.4530983865261078, + "learning_rate": 6.3347231470816005e-06, + "loss": 0.0895, + "step": 37801 + }, + { + "epoch": 0.8861384107784361, + "grad_norm": 0.22115574777126312, + "learning_rate": 6.332143335766472e-06, + "loss": 0.1933, + "step": 37802 + }, + { + "epoch": 0.8861618523532411, + "grad_norm": 0.8421698808670044, + "learning_rate": 6.3295640326957566e-06, + "loss": 0.149, + "step": 37803 + }, + { + "epoch": 0.8861852939280461, + "grad_norm": 0.6078940033912659, + "learning_rate": 6.326985237883465e-06, + "loss": 0.1205, + "step": 37804 + }, + { + "epoch": 0.8862087355028511, + "grad_norm": 0.5431757569313049, + "learning_rate": 6.3244069513435975e-06, + "loss": 0.0626, + "step": 37805 + }, + { + "epoch": 0.8862321770776561, + "grad_norm": 0.12060307711362839, + "learning_rate": 6.32182917309011e-06, + "loss": 0.0099, + "step": 37806 + }, + { + "epoch": 0.886255618652461, + "grad_norm": 0.3881871700286865, + "learning_rate": 6.319251903137025e-06, + "loss": 0.0769, + "step": 37807 + }, + { + "epoch": 0.8862790602272661, + "grad_norm": 0.3664894700050354, + "learning_rate": 6.316675141498318e-06, + "loss": 0.0636, + "step": 37808 + }, + { + "epoch": 0.886302501802071, + "grad_norm": 0.5035122036933899, + "learning_rate": 6.314098888187969e-06, + "loss": 0.6181, + "step": 37809 + }, + { + "epoch": 0.8863259433768761, + "grad_norm": 0.26788511872291565, + "learning_rate": 6.311523143219944e-06, + "loss": 0.0633, + "step": 37810 + }, + { + "epoch": 0.886349384951681, + "grad_norm": 0.1292405128479004, + "learning_rate": 6.308947906608231e-06, + "loss": 0.031, + "step": 37811 + }, + { + "epoch": 0.8863728265264861, + "grad_norm": 0.230117067694664, + "learning_rate": 6.306373178366787e-06, + "loss": 0.1481, + "step": 37812 + }, + { + "epoch": 0.886396268101291, + "grad_norm": 0.4493148624897003, + "learning_rate": 6.30379895850961e-06, + "loss": 0.0535, + "step": 37813 + }, + { + "epoch": 0.8864197096760961, + "grad_norm": 0.3844805955886841, + "learning_rate": 6.301225247050668e-06, + "loss": 0.6208, + "step": 37814 + }, + { + "epoch": 0.886443151250901, + "grad_norm": 0.24790358543395996, + "learning_rate": 6.298652044003883e-06, + "loss": 0.0187, + "step": 37815 + }, + { + "epoch": 0.8864665928257061, + "grad_norm": 0.6264965534210205, + "learning_rate": 6.2960793493832656e-06, + "loss": 0.4841, + "step": 37816 + }, + { + "epoch": 0.886490034400511, + "grad_norm": 0.3948349952697754, + "learning_rate": 6.29350716320275e-06, + "loss": 0.0563, + "step": 37817 + }, + { + "epoch": 0.886513475975316, + "grad_norm": 0.3606700897216797, + "learning_rate": 6.290935485476312e-06, + "loss": 0.0305, + "step": 37818 + }, + { + "epoch": 0.886536917550121, + "grad_norm": 0.08941090106964111, + "learning_rate": 6.288364316217865e-06, + "loss": 0.0091, + "step": 37819 + }, + { + "epoch": 0.886560359124926, + "grad_norm": 0.11978945881128311, + "learning_rate": 6.285793655441408e-06, + "loss": 0.014, + "step": 37820 + }, + { + "epoch": 0.886583800699731, + "grad_norm": 0.2691701054573059, + "learning_rate": 6.283223503160873e-06, + "loss": 0.0326, + "step": 37821 + }, + { + "epoch": 0.886607242274536, + "grad_norm": 0.1550562083721161, + "learning_rate": 6.280653859390206e-06, + "loss": 0.0269, + "step": 37822 + }, + { + "epoch": 0.886630683849341, + "grad_norm": 0.4895227551460266, + "learning_rate": 6.27808472414334e-06, + "loss": 0.0618, + "step": 37823 + }, + { + "epoch": 0.886654125424146, + "grad_norm": 0.6800141334533691, + "learning_rate": 6.27551609743422e-06, + "loss": 0.1086, + "step": 37824 + }, + { + "epoch": 0.886677566998951, + "grad_norm": 0.5958423614501953, + "learning_rate": 6.2729479792767665e-06, + "loss": 0.0749, + "step": 37825 + }, + { + "epoch": 0.886701008573756, + "grad_norm": 0.17244020104408264, + "learning_rate": 6.270380369684959e-06, + "loss": 0.0288, + "step": 37826 + }, + { + "epoch": 0.8867244501485609, + "grad_norm": 0.3401133716106415, + "learning_rate": 6.2678132686726846e-06, + "loss": 0.0652, + "step": 37827 + }, + { + "epoch": 0.886747891723366, + "grad_norm": 0.3741130232810974, + "learning_rate": 6.2652466762539e-06, + "loss": 0.2294, + "step": 37828 + }, + { + "epoch": 0.8867713332981709, + "grad_norm": 0.5136774778366089, + "learning_rate": 6.2626805924425265e-06, + "loss": 0.0404, + "step": 37829 + }, + { + "epoch": 0.886794774872976, + "grad_norm": 0.3323603570461273, + "learning_rate": 6.260115017252488e-06, + "loss": 0.0459, + "step": 37830 + }, + { + "epoch": 0.8868182164477809, + "grad_norm": 0.29092511534690857, + "learning_rate": 6.257549950697694e-06, + "loss": 0.0533, + "step": 37831 + }, + { + "epoch": 0.886841658022586, + "grad_norm": 0.4423607289791107, + "learning_rate": 6.254985392792057e-06, + "loss": 0.0594, + "step": 37832 + }, + { + "epoch": 0.8868650995973909, + "grad_norm": 0.4070608913898468, + "learning_rate": 6.252421343549519e-06, + "loss": 0.0685, + "step": 37833 + }, + { + "epoch": 0.886888541172196, + "grad_norm": 0.2502429783344269, + "learning_rate": 6.249857802983983e-06, + "loss": 0.0369, + "step": 37834 + }, + { + "epoch": 0.8869119827470009, + "grad_norm": 0.3809599280357361, + "learning_rate": 6.247294771109347e-06, + "loss": 0.0743, + "step": 37835 + }, + { + "epoch": 0.8869354243218059, + "grad_norm": 0.07850697636604309, + "learning_rate": 6.2447322479395335e-06, + "loss": 0.0145, + "step": 37836 + }, + { + "epoch": 0.8869588658966109, + "grad_norm": 0.2943565845489502, + "learning_rate": 6.24217023348842e-06, + "loss": 0.0448, + "step": 37837 + }, + { + "epoch": 0.8869823074714159, + "grad_norm": 0.3540504574775696, + "learning_rate": 6.2396087277699415e-06, + "loss": 0.0331, + "step": 37838 + }, + { + "epoch": 0.8870057490462209, + "grad_norm": 0.32370495796203613, + "learning_rate": 6.2370477307979735e-06, + "loss": 0.0239, + "step": 37839 + }, + { + "epoch": 0.8870291906210259, + "grad_norm": 0.12448851019144058, + "learning_rate": 6.234487242586428e-06, + "loss": 0.0271, + "step": 37840 + }, + { + "epoch": 0.887052632195831, + "grad_norm": 0.7467740774154663, + "learning_rate": 6.2319272631491954e-06, + "loss": 0.0773, + "step": 37841 + }, + { + "epoch": 0.8870760737706359, + "grad_norm": 0.3137747645378113, + "learning_rate": 6.229367792500163e-06, + "loss": 0.0434, + "step": 37842 + }, + { + "epoch": 0.887099515345441, + "grad_norm": 0.22154638171195984, + "learning_rate": 6.2268088306532076e-06, + "loss": 0.0416, + "step": 37843 + }, + { + "epoch": 0.8871229569202459, + "grad_norm": 0.48397910594940186, + "learning_rate": 6.224250377622221e-06, + "loss": 0.4561, + "step": 37844 + }, + { + "epoch": 0.887146398495051, + "grad_norm": 0.44638100266456604, + "learning_rate": 6.2216924334211e-06, + "loss": 0.0812, + "step": 37845 + }, + { + "epoch": 0.8871698400698559, + "grad_norm": 0.4031457304954529, + "learning_rate": 6.219134998063713e-06, + "loss": 0.0545, + "step": 37846 + }, + { + "epoch": 0.8871932816446609, + "grad_norm": 0.5166687369346619, + "learning_rate": 6.216578071563939e-06, + "loss": 0.1044, + "step": 37847 + }, + { + "epoch": 0.8872167232194659, + "grad_norm": 0.4765651524066925, + "learning_rate": 6.2140216539356426e-06, + "loss": 0.1337, + "step": 37848 + }, + { + "epoch": 0.8872401647942709, + "grad_norm": 0.9108617901802063, + "learning_rate": 6.211465745192702e-06, + "loss": 0.1208, + "step": 37849 + }, + { + "epoch": 0.8872636063690759, + "grad_norm": 0.08921753615140915, + "learning_rate": 6.208910345348973e-06, + "loss": 0.0077, + "step": 37850 + }, + { + "epoch": 0.8872870479438809, + "grad_norm": 0.42275190353393555, + "learning_rate": 6.2063554544183334e-06, + "loss": 0.0804, + "step": 37851 + }, + { + "epoch": 0.8873104895186859, + "grad_norm": 0.809004545211792, + "learning_rate": 6.203801072414661e-06, + "loss": 0.1334, + "step": 37852 + }, + { + "epoch": 0.8873339310934909, + "grad_norm": 0.3275797367095947, + "learning_rate": 6.2012471993518116e-06, + "loss": 0.0807, + "step": 37853 + }, + { + "epoch": 0.8873573726682958, + "grad_norm": 0.7012537121772766, + "learning_rate": 6.198693835243619e-06, + "loss": 1.005, + "step": 37854 + }, + { + "epoch": 0.8873808142431009, + "grad_norm": 0.788865327835083, + "learning_rate": 6.19614098010396e-06, + "loss": 0.7911, + "step": 37855 + }, + { + "epoch": 0.8874042558179058, + "grad_norm": 0.35379329323768616, + "learning_rate": 6.193588633946668e-06, + "loss": 0.0634, + "step": 37856 + }, + { + "epoch": 0.8874276973927109, + "grad_norm": 0.2628863453865051, + "learning_rate": 6.191036796785587e-06, + "loss": 0.0762, + "step": 37857 + }, + { + "epoch": 0.8874511389675158, + "grad_norm": 0.5362815856933594, + "learning_rate": 6.1884854686345974e-06, + "loss": 0.06, + "step": 37858 + }, + { + "epoch": 0.8874745805423209, + "grad_norm": 0.418692409992218, + "learning_rate": 6.18593464950753e-06, + "loss": 0.055, + "step": 37859 + }, + { + "epoch": 0.8874980221171258, + "grad_norm": 0.42175304889678955, + "learning_rate": 6.183384339418208e-06, + "loss": 0.0924, + "step": 37860 + }, + { + "epoch": 0.8875214636919309, + "grad_norm": 0.3317807912826538, + "learning_rate": 6.180834538380486e-06, + "loss": 0.0429, + "step": 37861 + }, + { + "epoch": 0.8875449052667358, + "grad_norm": 0.13655522465705872, + "learning_rate": 6.178285246408199e-06, + "loss": 0.0345, + "step": 37862 + }, + { + "epoch": 0.8875683468415408, + "grad_norm": 0.7016710042953491, + "learning_rate": 6.1757364635151475e-06, + "loss": 0.1303, + "step": 37863 + }, + { + "epoch": 0.8875917884163458, + "grad_norm": 0.33832603693008423, + "learning_rate": 6.1731881897151955e-06, + "loss": 0.0674, + "step": 37864 + }, + { + "epoch": 0.8876152299911508, + "grad_norm": 0.433366596698761, + "learning_rate": 6.170640425022178e-06, + "loss": 0.0835, + "step": 37865 + }, + { + "epoch": 0.8876386715659558, + "grad_norm": 0.5200762152671814, + "learning_rate": 6.168093169449895e-06, + "loss": 0.5574, + "step": 37866 + }, + { + "epoch": 0.8876621131407608, + "grad_norm": 0.5596558451652527, + "learning_rate": 6.16554642301218e-06, + "loss": 0.0618, + "step": 37867 + }, + { + "epoch": 0.8876855547155658, + "grad_norm": 0.529262125492096, + "learning_rate": 6.163000185722856e-06, + "loss": 0.1116, + "step": 37868 + }, + { + "epoch": 0.8877089962903708, + "grad_norm": 0.3659062683582306, + "learning_rate": 6.160454457595733e-06, + "loss": 0.0476, + "step": 37869 + }, + { + "epoch": 0.8877324378651757, + "grad_norm": 0.2475089728832245, + "learning_rate": 6.1579092386445995e-06, + "loss": 0.0238, + "step": 37870 + }, + { + "epoch": 0.8877558794399808, + "grad_norm": 0.6902294754981995, + "learning_rate": 6.155364528883301e-06, + "loss": 0.429, + "step": 37871 + }, + { + "epoch": 0.8877793210147857, + "grad_norm": 0.6909551024436951, + "learning_rate": 6.1528203283256366e-06, + "loss": 0.0764, + "step": 37872 + }, + { + "epoch": 0.8878027625895908, + "grad_norm": 0.4824979603290558, + "learning_rate": 6.150276636985419e-06, + "loss": 0.0857, + "step": 37873 + }, + { + "epoch": 0.8878262041643957, + "grad_norm": 0.38276001811027527, + "learning_rate": 6.147733454876425e-06, + "loss": 0.0291, + "step": 37874 + }, + { + "epoch": 0.8878496457392008, + "grad_norm": 0.1947711557149887, + "learning_rate": 6.1451907820124555e-06, + "loss": 0.025, + "step": 37875 + }, + { + "epoch": 0.8878730873140057, + "grad_norm": 0.2244095504283905, + "learning_rate": 6.142648618407332e-06, + "loss": 0.024, + "step": 37876 + }, + { + "epoch": 0.8878965288888108, + "grad_norm": 0.28320497274398804, + "learning_rate": 6.140106964074832e-06, + "loss": 0.041, + "step": 37877 + }, + { + "epoch": 0.8879199704636157, + "grad_norm": 0.449974924325943, + "learning_rate": 6.137565819028768e-06, + "loss": 0.0867, + "step": 37878 + }, + { + "epoch": 0.8879434120384208, + "grad_norm": 0.5476368069648743, + "learning_rate": 6.135025183282906e-06, + "loss": 0.1258, + "step": 37879 + }, + { + "epoch": 0.8879668536132257, + "grad_norm": 0.42278826236724854, + "learning_rate": 6.132485056851034e-06, + "loss": 0.1137, + "step": 37880 + }, + { + "epoch": 0.8879902951880307, + "grad_norm": 0.8859908580780029, + "learning_rate": 6.129945439746931e-06, + "loss": 0.1405, + "step": 37881 + }, + { + "epoch": 0.8880137367628357, + "grad_norm": 0.5794281363487244, + "learning_rate": 6.127406331984375e-06, + "loss": 0.0917, + "step": 37882 + }, + { + "epoch": 0.8880371783376407, + "grad_norm": 0.24641694128513336, + "learning_rate": 6.124867733577167e-06, + "loss": 0.031, + "step": 37883 + }, + { + "epoch": 0.8880606199124457, + "grad_norm": 0.32509687542915344, + "learning_rate": 6.122329644539071e-06, + "loss": 0.2245, + "step": 37884 + }, + { + "epoch": 0.8880840614872507, + "grad_norm": 0.3329191505908966, + "learning_rate": 6.119792064883844e-06, + "loss": 0.073, + "step": 37885 + }, + { + "epoch": 0.8881075030620557, + "grad_norm": 0.28934717178344727, + "learning_rate": 6.117254994625265e-06, + "loss": 0.0658, + "step": 37886 + }, + { + "epoch": 0.8881309446368607, + "grad_norm": 0.1792593151330948, + "learning_rate": 6.114718433777112e-06, + "loss": 0.0319, + "step": 37887 + }, + { + "epoch": 0.8881543862116656, + "grad_norm": 0.45470350980758667, + "learning_rate": 6.112182382353116e-06, + "loss": 0.1494, + "step": 37888 + }, + { + "epoch": 0.8881778277864707, + "grad_norm": 0.5273029804229736, + "learning_rate": 6.109646840367056e-06, + "loss": 0.0831, + "step": 37889 + }, + { + "epoch": 0.8882012693612756, + "grad_norm": 0.2041756510734558, + "learning_rate": 6.107111807832699e-06, + "loss": 0.0264, + "step": 37890 + }, + { + "epoch": 0.8882247109360807, + "grad_norm": 0.16074292361736298, + "learning_rate": 6.104577284763813e-06, + "loss": 0.025, + "step": 37891 + }, + { + "epoch": 0.8882481525108857, + "grad_norm": 0.2212716042995453, + "learning_rate": 6.102043271174118e-06, + "loss": 0.0261, + "step": 37892 + }, + { + "epoch": 0.8882715940856907, + "grad_norm": 0.453945517539978, + "learning_rate": 6.0995097670773716e-06, + "loss": 0.4501, + "step": 37893 + }, + { + "epoch": 0.8882950356604957, + "grad_norm": 0.3282754421234131, + "learning_rate": 6.09697677248734e-06, + "loss": 0.2586, + "step": 37894 + }, + { + "epoch": 0.8883184772353007, + "grad_norm": 0.27361631393432617, + "learning_rate": 6.094444287417722e-06, + "loss": 0.0324, + "step": 37895 + }, + { + "epoch": 0.8883419188101057, + "grad_norm": 0.271918922662735, + "learning_rate": 6.09191231188232e-06, + "loss": 0.0432, + "step": 37896 + }, + { + "epoch": 0.8883653603849107, + "grad_norm": 0.5634030103683472, + "learning_rate": 6.089380845894832e-06, + "loss": 0.0861, + "step": 37897 + }, + { + "epoch": 0.8883888019597157, + "grad_norm": 0.35546836256980896, + "learning_rate": 6.086849889469004e-06, + "loss": 0.0262, + "step": 37898 + }, + { + "epoch": 0.8884122435345206, + "grad_norm": 0.4521459937095642, + "learning_rate": 6.08431944261858e-06, + "loss": 0.071, + "step": 37899 + }, + { + "epoch": 0.8884356851093257, + "grad_norm": 0.3133261203765869, + "learning_rate": 6.081789505357249e-06, + "loss": 0.0637, + "step": 37900 + }, + { + "epoch": 0.8884591266841306, + "grad_norm": 0.2747097909450531, + "learning_rate": 6.07926007769879e-06, + "loss": 0.2465, + "step": 37901 + }, + { + "epoch": 0.8884825682589357, + "grad_norm": 0.5705540180206299, + "learning_rate": 6.076731159656901e-06, + "loss": 0.1012, + "step": 37902 + }, + { + "epoch": 0.8885060098337406, + "grad_norm": 0.598789393901825, + "learning_rate": 6.074202751245317e-06, + "loss": 0.1946, + "step": 37903 + }, + { + "epoch": 0.8885294514085457, + "grad_norm": 0.3166505992412567, + "learning_rate": 6.071674852477749e-06, + "loss": 0.0562, + "step": 37904 + }, + { + "epoch": 0.8885528929833506, + "grad_norm": 0.19776949286460876, + "learning_rate": 6.069147463367908e-06, + "loss": 0.0308, + "step": 37905 + }, + { + "epoch": 0.8885763345581557, + "grad_norm": 0.3618939518928528, + "learning_rate": 6.0666205839295275e-06, + "loss": 0.062, + "step": 37906 + }, + { + "epoch": 0.8885997761329606, + "grad_norm": 0.23860472440719604, + "learning_rate": 6.064094214176275e-06, + "loss": 0.0381, + "step": 37907 + }, + { + "epoch": 0.8886232177077656, + "grad_norm": 0.2971441149711609, + "learning_rate": 6.061568354121916e-06, + "loss": 0.049, + "step": 37908 + }, + { + "epoch": 0.8886466592825706, + "grad_norm": 0.41483259201049805, + "learning_rate": 6.059043003780129e-06, + "loss": 0.0414, + "step": 37909 + }, + { + "epoch": 0.8886701008573756, + "grad_norm": 0.23346742987632751, + "learning_rate": 6.056518163164604e-06, + "loss": 0.0167, + "step": 37910 + }, + { + "epoch": 0.8886935424321806, + "grad_norm": 0.2684250473976135, + "learning_rate": 6.053993832289062e-06, + "loss": 0.0477, + "step": 37911 + }, + { + "epoch": 0.8887169840069856, + "grad_norm": 0.2566772699356079, + "learning_rate": 6.051470011167171e-06, + "loss": 0.0239, + "step": 37912 + }, + { + "epoch": 0.8887404255817906, + "grad_norm": 0.490939199924469, + "learning_rate": 6.048946699812663e-06, + "loss": 0.1145, + "step": 37913 + }, + { + "epoch": 0.8887638671565956, + "grad_norm": 0.257688045501709, + "learning_rate": 6.046423898239207e-06, + "loss": 0.0415, + "step": 37914 + }, + { + "epoch": 0.8887873087314005, + "grad_norm": 0.6234440207481384, + "learning_rate": 6.0439016064605004e-06, + "loss": 0.1472, + "step": 37915 + }, + { + "epoch": 0.8888107503062056, + "grad_norm": 0.33953621983528137, + "learning_rate": 6.041379824490223e-06, + "loss": 0.0221, + "step": 37916 + }, + { + "epoch": 0.8888341918810105, + "grad_norm": 0.2564409673213959, + "learning_rate": 6.0388585523420636e-06, + "loss": 0.053, + "step": 37917 + }, + { + "epoch": 0.8888576334558156, + "grad_norm": 0.6018890142440796, + "learning_rate": 6.0363377900297006e-06, + "loss": 0.0521, + "step": 37918 + }, + { + "epoch": 0.8888810750306205, + "grad_norm": 0.3156359791755676, + "learning_rate": 6.033817537566811e-06, + "loss": 0.0318, + "step": 37919 + }, + { + "epoch": 0.8889045166054256, + "grad_norm": 0.27764129638671875, + "learning_rate": 6.03129779496705e-06, + "loss": 0.0364, + "step": 37920 + }, + { + "epoch": 0.8889279581802305, + "grad_norm": 0.7306680083274841, + "learning_rate": 6.028778562244131e-06, + "loss": 0.1179, + "step": 37921 + }, + { + "epoch": 0.8889513997550356, + "grad_norm": 0.1619509607553482, + "learning_rate": 6.026259839411708e-06, + "loss": 0.0196, + "step": 37922 + }, + { + "epoch": 0.8889748413298405, + "grad_norm": 0.3082537055015564, + "learning_rate": 6.023741626483426e-06, + "loss": 0.0503, + "step": 37923 + }, + { + "epoch": 0.8889982829046456, + "grad_norm": 0.6847975254058838, + "learning_rate": 6.0212239234729516e-06, + "loss": 0.4095, + "step": 37924 + }, + { + "epoch": 0.8890217244794505, + "grad_norm": 0.2123987376689911, + "learning_rate": 6.018706730393986e-06, + "loss": 0.0202, + "step": 37925 + }, + { + "epoch": 0.8890451660542555, + "grad_norm": 0.09379307925701141, + "learning_rate": 6.016190047260151e-06, + "loss": 0.0131, + "step": 37926 + }, + { + "epoch": 0.8890686076290605, + "grad_norm": 0.20138157904148102, + "learning_rate": 6.013673874085102e-06, + "loss": 0.017, + "step": 37927 + }, + { + "epoch": 0.8890920492038655, + "grad_norm": 0.4400530755519867, + "learning_rate": 6.011158210882506e-06, + "loss": 0.0555, + "step": 37928 + }, + { + "epoch": 0.8891154907786705, + "grad_norm": 0.5390361547470093, + "learning_rate": 6.008643057666019e-06, + "loss": 0.0618, + "step": 37929 + }, + { + "epoch": 0.8891389323534755, + "grad_norm": 0.5773357152938843, + "learning_rate": 6.006128414449275e-06, + "loss": 0.0993, + "step": 37930 + }, + { + "epoch": 0.8891623739282805, + "grad_norm": 0.29550400376319885, + "learning_rate": 6.003614281245917e-06, + "loss": 0.0201, + "step": 37931 + }, + { + "epoch": 0.8891858155030855, + "grad_norm": 0.1979360282421112, + "learning_rate": 6.001100658069592e-06, + "loss": 0.0526, + "step": 37932 + }, + { + "epoch": 0.8892092570778904, + "grad_norm": 0.1301048994064331, + "learning_rate": 5.99858754493392e-06, + "loss": 0.0149, + "step": 37933 + }, + { + "epoch": 0.8892326986526955, + "grad_norm": 0.5050537586212158, + "learning_rate": 5.99607494185257e-06, + "loss": 0.1216, + "step": 37934 + }, + { + "epoch": 0.8892561402275004, + "grad_norm": 0.4484752118587494, + "learning_rate": 5.993562848839174e-06, + "loss": 0.0879, + "step": 37935 + }, + { + "epoch": 0.8892795818023055, + "grad_norm": 0.5043268203735352, + "learning_rate": 5.991051265907321e-06, + "loss": 0.5722, + "step": 37936 + }, + { + "epoch": 0.8893030233771104, + "grad_norm": 0.22516213357448578, + "learning_rate": 5.98854019307068e-06, + "loss": 0.0437, + "step": 37937 + }, + { + "epoch": 0.8893264649519155, + "grad_norm": 0.36841633915901184, + "learning_rate": 5.9860296303428705e-06, + "loss": 0.0565, + "step": 37938 + }, + { + "epoch": 0.8893499065267204, + "grad_norm": 0.0993686094880104, + "learning_rate": 5.983519577737506e-06, + "loss": 0.0142, + "step": 37939 + }, + { + "epoch": 0.8893733481015255, + "grad_norm": 0.10423488914966583, + "learning_rate": 5.981010035268198e-06, + "loss": 0.0112, + "step": 37940 + }, + { + "epoch": 0.8893967896763304, + "grad_norm": 0.1398201882839203, + "learning_rate": 5.9785010029485774e-06, + "loss": 0.0145, + "step": 37941 + }, + { + "epoch": 0.8894202312511355, + "grad_norm": 0.3532240688800812, + "learning_rate": 5.975992480792259e-06, + "loss": 0.0667, + "step": 37942 + }, + { + "epoch": 0.8894436728259405, + "grad_norm": 0.3467988669872284, + "learning_rate": 5.973484468812851e-06, + "loss": 0.0594, + "step": 37943 + }, + { + "epoch": 0.8894671144007454, + "grad_norm": 0.23977097868919373, + "learning_rate": 5.970976967023956e-06, + "loss": 0.0377, + "step": 37944 + }, + { + "epoch": 0.8894905559755505, + "grad_norm": 0.1919536292552948, + "learning_rate": 5.968469975439173e-06, + "loss": 0.0394, + "step": 37945 + }, + { + "epoch": 0.8895139975503554, + "grad_norm": 0.5379719734191895, + "learning_rate": 5.965963494072135e-06, + "loss": 0.0881, + "step": 37946 + }, + { + "epoch": 0.8895374391251605, + "grad_norm": 0.28956928849220276, + "learning_rate": 5.963457522936411e-06, + "loss": 0.0295, + "step": 37947 + }, + { + "epoch": 0.8895608806999654, + "grad_norm": 0.6773368120193481, + "learning_rate": 5.96095206204561e-06, + "loss": 0.1107, + "step": 37948 + }, + { + "epoch": 0.8895843222747705, + "grad_norm": 0.2740088999271393, + "learning_rate": 5.958447111413334e-06, + "loss": 0.0209, + "step": 37949 + }, + { + "epoch": 0.8896077638495754, + "grad_norm": 0.11470021307468414, + "learning_rate": 5.955942671053161e-06, + "loss": 0.0141, + "step": 37950 + }, + { + "epoch": 0.8896312054243805, + "grad_norm": 0.55184006690979, + "learning_rate": 5.953438740978701e-06, + "loss": 0.1436, + "step": 37951 + }, + { + "epoch": 0.8896546469991854, + "grad_norm": 0.24823975563049316, + "learning_rate": 5.9509353212034994e-06, + "loss": 0.0447, + "step": 37952 + }, + { + "epoch": 0.8896780885739904, + "grad_norm": 0.4834316074848175, + "learning_rate": 5.948432411741189e-06, + "loss": 0.521, + "step": 37953 + }, + { + "epoch": 0.8897015301487954, + "grad_norm": 0.10573648661375046, + "learning_rate": 5.945930012605327e-06, + "loss": 0.0103, + "step": 37954 + }, + { + "epoch": 0.8897249717236004, + "grad_norm": 0.35611265897750854, + "learning_rate": 5.943428123809503e-06, + "loss": 0.0809, + "step": 37955 + }, + { + "epoch": 0.8897484132984054, + "grad_norm": 0.12671823799610138, + "learning_rate": 5.94092674536727e-06, + "loss": 0.023, + "step": 37956 + }, + { + "epoch": 0.8897718548732104, + "grad_norm": 0.28108417987823486, + "learning_rate": 5.93842587729222e-06, + "loss": 0.045, + "step": 37957 + }, + { + "epoch": 0.8897952964480154, + "grad_norm": 0.3644540309906006, + "learning_rate": 5.935925519597896e-06, + "loss": 0.0413, + "step": 37958 + }, + { + "epoch": 0.8898187380228204, + "grad_norm": 0.39817890524864197, + "learning_rate": 5.933425672297899e-06, + "loss": 0.0745, + "step": 37959 + }, + { + "epoch": 0.8898421795976253, + "grad_norm": 0.37136995792388916, + "learning_rate": 5.9309263354057844e-06, + "loss": 0.0892, + "step": 37960 + }, + { + "epoch": 0.8898656211724304, + "grad_norm": 0.5696057081222534, + "learning_rate": 5.928427508935097e-06, + "loss": 0.6614, + "step": 37961 + }, + { + "epoch": 0.8898890627472353, + "grad_norm": 0.529430627822876, + "learning_rate": 5.9259291928994155e-06, + "loss": 0.082, + "step": 37962 + }, + { + "epoch": 0.8899125043220404, + "grad_norm": 0.18521785736083984, + "learning_rate": 5.923431387312284e-06, + "loss": 0.03, + "step": 37963 + }, + { + "epoch": 0.8899359458968453, + "grad_norm": 0.33018773794174194, + "learning_rate": 5.920934092187269e-06, + "loss": 0.0463, + "step": 37964 + }, + { + "epoch": 0.8899593874716504, + "grad_norm": 0.39512884616851807, + "learning_rate": 5.918437307537883e-06, + "loss": 0.0371, + "step": 37965 + }, + { + "epoch": 0.8899828290464553, + "grad_norm": 0.2816406786441803, + "learning_rate": 5.9159410333777255e-06, + "loss": 0.0293, + "step": 37966 + }, + { + "epoch": 0.8900062706212604, + "grad_norm": 0.36059463024139404, + "learning_rate": 5.913445269720319e-06, + "loss": 0.0648, + "step": 37967 + }, + { + "epoch": 0.8900297121960653, + "grad_norm": 0.1604011207818985, + "learning_rate": 5.9109500165791975e-06, + "loss": 0.013, + "step": 37968 + }, + { + "epoch": 0.8900531537708704, + "grad_norm": 0.1110324040055275, + "learning_rate": 5.9084552739679165e-06, + "loss": 0.0123, + "step": 37969 + }, + { + "epoch": 0.8900765953456753, + "grad_norm": 0.19520938396453857, + "learning_rate": 5.905961041899988e-06, + "loss": 0.0327, + "step": 37970 + }, + { + "epoch": 0.8901000369204803, + "grad_norm": 0.3469855487346649, + "learning_rate": 5.903467320388955e-06, + "loss": 0.0835, + "step": 37971 + }, + { + "epoch": 0.8901234784952853, + "grad_norm": 0.6037992238998413, + "learning_rate": 5.900974109448376e-06, + "loss": 0.6927, + "step": 37972 + }, + { + "epoch": 0.8901469200700903, + "grad_norm": 0.3690468966960907, + "learning_rate": 5.898481409091727e-06, + "loss": 0.069, + "step": 37973 + }, + { + "epoch": 0.8901703616448953, + "grad_norm": 0.4597511291503906, + "learning_rate": 5.895989219332587e-06, + "loss": 0.1047, + "step": 37974 + }, + { + "epoch": 0.8901938032197003, + "grad_norm": 0.3916483223438263, + "learning_rate": 5.893497540184456e-06, + "loss": 0.0703, + "step": 37975 + }, + { + "epoch": 0.8902172447945053, + "grad_norm": 0.45455092191696167, + "learning_rate": 5.891006371660856e-06, + "loss": 0.0328, + "step": 37976 + }, + { + "epoch": 0.8902406863693103, + "grad_norm": 0.13827911019325256, + "learning_rate": 5.888515713775289e-06, + "loss": 0.0172, + "step": 37977 + }, + { + "epoch": 0.8902641279441152, + "grad_norm": 0.24251419305801392, + "learning_rate": 5.886025566541276e-06, + "loss": 0.0463, + "step": 37978 + }, + { + "epoch": 0.8902875695189203, + "grad_norm": 0.3335946798324585, + "learning_rate": 5.8835359299723505e-06, + "loss": 0.4692, + "step": 37979 + }, + { + "epoch": 0.8903110110937252, + "grad_norm": 0.1764824539422989, + "learning_rate": 5.881046804082002e-06, + "loss": 0.0272, + "step": 37980 + }, + { + "epoch": 0.8903344526685303, + "grad_norm": 0.2077282965183258, + "learning_rate": 5.878558188883743e-06, + "loss": 0.0259, + "step": 37981 + }, + { + "epoch": 0.8903578942433352, + "grad_norm": 0.3281104564666748, + "learning_rate": 5.876070084391072e-06, + "loss": 0.0529, + "step": 37982 + }, + { + "epoch": 0.8903813358181403, + "grad_norm": 0.21481485664844513, + "learning_rate": 5.87358249061748e-06, + "loss": 0.0284, + "step": 37983 + }, + { + "epoch": 0.8904047773929452, + "grad_norm": 0.3640354871749878, + "learning_rate": 5.871095407576488e-06, + "loss": 0.0683, + "step": 37984 + }, + { + "epoch": 0.8904282189677503, + "grad_norm": 0.45371994376182556, + "learning_rate": 5.868608835281564e-06, + "loss": 0.6867, + "step": 37985 + }, + { + "epoch": 0.8904516605425552, + "grad_norm": 0.15411749482154846, + "learning_rate": 5.86612277374623e-06, + "loss": 0.0142, + "step": 37986 + }, + { + "epoch": 0.8904751021173603, + "grad_norm": 0.585813045501709, + "learning_rate": 5.863637222983953e-06, + "loss": 0.1403, + "step": 37987 + }, + { + "epoch": 0.8904985436921652, + "grad_norm": 0.15610621869564056, + "learning_rate": 5.861152183008245e-06, + "loss": 0.0173, + "step": 37988 + }, + { + "epoch": 0.8905219852669702, + "grad_norm": 0.1789720505475998, + "learning_rate": 5.858667653832561e-06, + "loss": 0.0253, + "step": 37989 + }, + { + "epoch": 0.8905454268417752, + "grad_norm": 0.3867132365703583, + "learning_rate": 5.856183635470369e-06, + "loss": 0.0537, + "step": 37990 + }, + { + "epoch": 0.8905688684165802, + "grad_norm": 0.46358510851860046, + "learning_rate": 5.853700127935202e-06, + "loss": 0.1124, + "step": 37991 + }, + { + "epoch": 0.8905923099913852, + "grad_norm": 0.408864289522171, + "learning_rate": 5.851217131240505e-06, + "loss": 0.0937, + "step": 37992 + }, + { + "epoch": 0.8906157515661902, + "grad_norm": 0.22286289930343628, + "learning_rate": 5.848734645399745e-06, + "loss": 0.026, + "step": 37993 + }, + { + "epoch": 0.8906391931409953, + "grad_norm": 0.6453284025192261, + "learning_rate": 5.8462526704264e-06, + "loss": 0.7729, + "step": 37994 + }, + { + "epoch": 0.8906626347158002, + "grad_norm": 0.3221381902694702, + "learning_rate": 5.843771206333937e-06, + "loss": 0.0474, + "step": 37995 + }, + { + "epoch": 0.8906860762906053, + "grad_norm": 0.44121643900871277, + "learning_rate": 5.8412902531358e-06, + "loss": 0.0571, + "step": 37996 + }, + { + "epoch": 0.8907095178654102, + "grad_norm": 0.07703094184398651, + "learning_rate": 5.838809810845469e-06, + "loss": 0.0092, + "step": 37997 + }, + { + "epoch": 0.8907329594402152, + "grad_norm": 0.4145999848842621, + "learning_rate": 5.836329879476421e-06, + "loss": 0.0617, + "step": 37998 + }, + { + "epoch": 0.8907564010150202, + "grad_norm": 0.7361862659454346, + "learning_rate": 5.833850459042089e-06, + "loss": 0.1807, + "step": 37999 + }, + { + "epoch": 0.8907798425898252, + "grad_norm": 0.13702426850795746, + "learning_rate": 5.831371549555941e-06, + "loss": 0.0088, + "step": 38000 + }, + { + "epoch": 0.8908032841646302, + "grad_norm": 0.2492615133523941, + "learning_rate": 5.828893151031412e-06, + "loss": 0.0516, + "step": 38001 + }, + { + "epoch": 0.8908267257394352, + "grad_norm": 0.32008999586105347, + "learning_rate": 5.826415263481966e-06, + "loss": 0.0499, + "step": 38002 + }, + { + "epoch": 0.8908501673142402, + "grad_norm": 0.797057569026947, + "learning_rate": 5.823937886921016e-06, + "loss": 0.1298, + "step": 38003 + }, + { + "epoch": 0.8908736088890452, + "grad_norm": 0.11996299028396606, + "learning_rate": 5.8214610213620406e-06, + "loss": 0.0159, + "step": 38004 + }, + { + "epoch": 0.8908970504638501, + "grad_norm": 0.40535926818847656, + "learning_rate": 5.8189846668184724e-06, + "loss": 0.0414, + "step": 38005 + }, + { + "epoch": 0.8909204920386552, + "grad_norm": 0.5328738689422607, + "learning_rate": 5.816508823303735e-06, + "loss": 0.0751, + "step": 38006 + }, + { + "epoch": 0.8909439336134601, + "grad_norm": 0.22932276129722595, + "learning_rate": 5.814033490831272e-06, + "loss": 0.0336, + "step": 38007 + }, + { + "epoch": 0.8909673751882652, + "grad_norm": 0.26783987879753113, + "learning_rate": 5.811558669414519e-06, + "loss": 0.0277, + "step": 38008 + }, + { + "epoch": 0.8909908167630701, + "grad_norm": 0.466910183429718, + "learning_rate": 5.809084359066874e-06, + "loss": 0.0655, + "step": 38009 + }, + { + "epoch": 0.8910142583378752, + "grad_norm": 0.32975876331329346, + "learning_rate": 5.806610559801784e-06, + "loss": 0.052, + "step": 38010 + }, + { + "epoch": 0.8910376999126801, + "grad_norm": 0.29405656456947327, + "learning_rate": 5.804137271632692e-06, + "loss": 0.0264, + "step": 38011 + }, + { + "epoch": 0.8910611414874852, + "grad_norm": 0.3763252794742584, + "learning_rate": 5.801664494573e-06, + "loss": 0.0552, + "step": 38012 + }, + { + "epoch": 0.8910845830622901, + "grad_norm": 0.1359838843345642, + "learning_rate": 5.799192228636119e-06, + "loss": 0.0112, + "step": 38013 + }, + { + "epoch": 0.8911080246370952, + "grad_norm": 0.19998914003372192, + "learning_rate": 5.796720473835471e-06, + "loss": 0.0412, + "step": 38014 + }, + { + "epoch": 0.8911314662119001, + "grad_norm": 0.5830320119857788, + "learning_rate": 5.794249230184467e-06, + "loss": 0.5044, + "step": 38015 + }, + { + "epoch": 0.8911549077867051, + "grad_norm": 0.40107980370521545, + "learning_rate": 5.79177849769651e-06, + "loss": 0.0235, + "step": 38016 + }, + { + "epoch": 0.8911783493615101, + "grad_norm": 0.7262521982192993, + "learning_rate": 5.789308276385019e-06, + "loss": 0.1292, + "step": 38017 + }, + { + "epoch": 0.8912017909363151, + "grad_norm": 0.13955974578857422, + "learning_rate": 5.786838566263386e-06, + "loss": 0.0111, + "step": 38018 + }, + { + "epoch": 0.8912252325111201, + "grad_norm": 0.6429092288017273, + "learning_rate": 5.7843693673450215e-06, + "loss": 0.5846, + "step": 38019 + }, + { + "epoch": 0.8912486740859251, + "grad_norm": 0.309400349855423, + "learning_rate": 5.781900679643326e-06, + "loss": 0.2115, + "step": 38020 + }, + { + "epoch": 0.8912721156607301, + "grad_norm": 0.45769691467285156, + "learning_rate": 5.779432503171655e-06, + "loss": 0.0842, + "step": 38021 + }, + { + "epoch": 0.8912955572355351, + "grad_norm": 0.37130844593048096, + "learning_rate": 5.776964837943466e-06, + "loss": 0.0616, + "step": 38022 + }, + { + "epoch": 0.89131899881034, + "grad_norm": 0.1625405102968216, + "learning_rate": 5.77449768397208e-06, + "loss": 0.0203, + "step": 38023 + }, + { + "epoch": 0.8913424403851451, + "grad_norm": 0.35939377546310425, + "learning_rate": 5.772031041270953e-06, + "loss": 0.0506, + "step": 38024 + }, + { + "epoch": 0.89136588195995, + "grad_norm": 0.42281436920166016, + "learning_rate": 5.76956490985342e-06, + "loss": 0.0512, + "step": 38025 + }, + { + "epoch": 0.8913893235347551, + "grad_norm": 0.13938573002815247, + "learning_rate": 5.7670992897328805e-06, + "loss": 0.0157, + "step": 38026 + }, + { + "epoch": 0.89141276510956, + "grad_norm": 0.38596606254577637, + "learning_rate": 5.7646341809227125e-06, + "loss": 0.0574, + "step": 38027 + }, + { + "epoch": 0.8914362066843651, + "grad_norm": 0.32543012499809265, + "learning_rate": 5.762169583436283e-06, + "loss": 0.0543, + "step": 38028 + }, + { + "epoch": 0.89145964825917, + "grad_norm": 0.5409165620803833, + "learning_rate": 5.759705497286971e-06, + "loss": 0.1149, + "step": 38029 + }, + { + "epoch": 0.8914830898339751, + "grad_norm": 0.36630523204803467, + "learning_rate": 5.757241922488154e-06, + "loss": 0.2768, + "step": 38030 + }, + { + "epoch": 0.89150653140878, + "grad_norm": 0.466304749250412, + "learning_rate": 5.7547788590531894e-06, + "loss": 0.0636, + "step": 38031 + }, + { + "epoch": 0.891529972983585, + "grad_norm": 0.36059412360191345, + "learning_rate": 5.752316306995453e-06, + "loss": 0.0253, + "step": 38032 + }, + { + "epoch": 0.89155341455839, + "grad_norm": 0.41185885667800903, + "learning_rate": 5.749854266328292e-06, + "loss": 0.1529, + "step": 38033 + }, + { + "epoch": 0.891576856133195, + "grad_norm": 0.401381254196167, + "learning_rate": 5.747392737065049e-06, + "loss": 0.0718, + "step": 38034 + }, + { + "epoch": 0.891600297708, + "grad_norm": 0.504185140132904, + "learning_rate": 5.744931719219116e-06, + "loss": 0.0993, + "step": 38035 + }, + { + "epoch": 0.891623739282805, + "grad_norm": 0.36495810747146606, + "learning_rate": 5.742471212803846e-06, + "loss": 0.0923, + "step": 38036 + }, + { + "epoch": 0.89164718085761, + "grad_norm": 0.30969688296318054, + "learning_rate": 5.7400112178325745e-06, + "loss": 0.0726, + "step": 38037 + }, + { + "epoch": 0.891670622432415, + "grad_norm": 0.10049311071634293, + "learning_rate": 5.7375517343186466e-06, + "loss": 0.0225, + "step": 38038 + }, + { + "epoch": 0.89169406400722, + "grad_norm": 0.30635377764701843, + "learning_rate": 5.735092762275418e-06, + "loss": 0.0504, + "step": 38039 + }, + { + "epoch": 0.891717505582025, + "grad_norm": 0.40382009744644165, + "learning_rate": 5.732634301716233e-06, + "loss": 0.0702, + "step": 38040 + }, + { + "epoch": 0.8917409471568299, + "grad_norm": 0.4716818928718567, + "learning_rate": 5.730176352654404e-06, + "loss": 0.0908, + "step": 38041 + }, + { + "epoch": 0.891764388731635, + "grad_norm": 0.30948084592819214, + "learning_rate": 5.727718915103309e-06, + "loss": 0.0688, + "step": 38042 + }, + { + "epoch": 0.8917878303064399, + "grad_norm": 0.1826014220714569, + "learning_rate": 5.7252619890762585e-06, + "loss": 0.0354, + "step": 38043 + }, + { + "epoch": 0.891811271881245, + "grad_norm": 0.22722190618515015, + "learning_rate": 5.722805574586587e-06, + "loss": 0.2774, + "step": 38044 + }, + { + "epoch": 0.89183471345605, + "grad_norm": 0.14848439395427704, + "learning_rate": 5.720349671647618e-06, + "loss": 0.0131, + "step": 38045 + }, + { + "epoch": 0.891858155030855, + "grad_norm": 0.0857929140329361, + "learning_rate": 5.717894280272673e-06, + "loss": 0.0186, + "step": 38046 + }, + { + "epoch": 0.89188159660566, + "grad_norm": 0.36179476976394653, + "learning_rate": 5.7154394004750975e-06, + "loss": 0.0309, + "step": 38047 + }, + { + "epoch": 0.891905038180465, + "grad_norm": 0.09485313296318054, + "learning_rate": 5.712985032268192e-06, + "loss": 0.0113, + "step": 38048 + }, + { + "epoch": 0.89192847975527, + "grad_norm": 0.4132022261619568, + "learning_rate": 5.710531175665279e-06, + "loss": 0.0878, + "step": 38049 + }, + { + "epoch": 0.891951921330075, + "grad_norm": 0.45987430214881897, + "learning_rate": 5.708077830679692e-06, + "loss": 0.1054, + "step": 38050 + }, + { + "epoch": 0.89197536290488, + "grad_norm": 0.6581549644470215, + "learning_rate": 5.705624997324721e-06, + "loss": 0.0899, + "step": 38051 + }, + { + "epoch": 0.8919988044796849, + "grad_norm": 0.48993515968322754, + "learning_rate": 5.703172675613677e-06, + "loss": 0.057, + "step": 38052 + }, + { + "epoch": 0.89202224605449, + "grad_norm": 0.1187913566827774, + "learning_rate": 5.700720865559872e-06, + "loss": 0.0104, + "step": 38053 + }, + { + "epoch": 0.8920456876292949, + "grad_norm": 0.38874953985214233, + "learning_rate": 5.698269567176584e-06, + "loss": 0.0546, + "step": 38054 + }, + { + "epoch": 0.8920691292041, + "grad_norm": 0.3940654993057251, + "learning_rate": 5.695818780477158e-06, + "loss": 0.0556, + "step": 38055 + }, + { + "epoch": 0.8920925707789049, + "grad_norm": 0.37241098284721375, + "learning_rate": 5.693368505474872e-06, + "loss": 0.0925, + "step": 38056 + }, + { + "epoch": 0.89211601235371, + "grad_norm": 0.5335540771484375, + "learning_rate": 5.6909187421830265e-06, + "loss": 0.0649, + "step": 38057 + }, + { + "epoch": 0.8921394539285149, + "grad_norm": 0.34337273240089417, + "learning_rate": 5.688469490614878e-06, + "loss": 0.0597, + "step": 38058 + }, + { + "epoch": 0.89216289550332, + "grad_norm": 0.18230123817920685, + "learning_rate": 5.686020750783772e-06, + "loss": 0.0196, + "step": 38059 + }, + { + "epoch": 0.8921863370781249, + "grad_norm": 0.5111965537071228, + "learning_rate": 5.683572522702962e-06, + "loss": 0.0609, + "step": 38060 + }, + { + "epoch": 0.8922097786529299, + "grad_norm": 0.5039265155792236, + "learning_rate": 5.68112480638573e-06, + "loss": 0.0709, + "step": 38061 + }, + { + "epoch": 0.8922332202277349, + "grad_norm": 0.6303063631057739, + "learning_rate": 5.678677601845384e-06, + "loss": 0.1075, + "step": 38062 + }, + { + "epoch": 0.8922566618025399, + "grad_norm": 0.38277578353881836, + "learning_rate": 5.676230909095181e-06, + "loss": 0.0503, + "step": 38063 + }, + { + "epoch": 0.8922801033773449, + "grad_norm": 0.297027587890625, + "learning_rate": 5.673784728148401e-06, + "loss": 0.0635, + "step": 38064 + }, + { + "epoch": 0.8923035449521499, + "grad_norm": 0.5541319847106934, + "learning_rate": 5.671339059018321e-06, + "loss": 0.6762, + "step": 38065 + }, + { + "epoch": 0.8923269865269549, + "grad_norm": 0.3281194269657135, + "learning_rate": 5.668893901718186e-06, + "loss": 0.0553, + "step": 38066 + }, + { + "epoch": 0.8923504281017599, + "grad_norm": 0.46128973364830017, + "learning_rate": 5.666449256261308e-06, + "loss": 0.0301, + "step": 38067 + }, + { + "epoch": 0.8923738696765648, + "grad_norm": 0.4076770842075348, + "learning_rate": 5.66400512266092e-06, + "loss": 0.0622, + "step": 38068 + }, + { + "epoch": 0.8923973112513699, + "grad_norm": 0.256387323141098, + "learning_rate": 5.661561500930301e-06, + "loss": 0.0362, + "step": 38069 + }, + { + "epoch": 0.8924207528261748, + "grad_norm": 0.2897562086582184, + "learning_rate": 5.659118391082685e-06, + "loss": 0.0372, + "step": 38070 + }, + { + "epoch": 0.8924441944009799, + "grad_norm": 0.1810396909713745, + "learning_rate": 5.6566757931313496e-06, + "loss": 0.0283, + "step": 38071 + }, + { + "epoch": 0.8924676359757848, + "grad_norm": 0.1795211136341095, + "learning_rate": 5.654233707089562e-06, + "loss": 0.0351, + "step": 38072 + }, + { + "epoch": 0.8924910775505899, + "grad_norm": 0.3114618957042694, + "learning_rate": 5.651792132970524e-06, + "loss": 0.0352, + "step": 38073 + }, + { + "epoch": 0.8925145191253948, + "grad_norm": 0.5140883326530457, + "learning_rate": 5.649351070787534e-06, + "loss": 0.5959, + "step": 38074 + }, + { + "epoch": 0.8925379607001999, + "grad_norm": 0.1952601671218872, + "learning_rate": 5.6469105205538275e-06, + "loss": 0.0239, + "step": 38075 + }, + { + "epoch": 0.8925614022750048, + "grad_norm": 0.4666445553302765, + "learning_rate": 5.644470482282626e-06, + "loss": 0.0718, + "step": 38076 + }, + { + "epoch": 0.8925848438498099, + "grad_norm": 0.224638432264328, + "learning_rate": 5.642030955987187e-06, + "loss": 0.0527, + "step": 38077 + }, + { + "epoch": 0.8926082854246148, + "grad_norm": 0.48987776041030884, + "learning_rate": 5.639591941680744e-06, + "loss": 0.0374, + "step": 38078 + }, + { + "epoch": 0.8926317269994198, + "grad_norm": 0.43157288432121277, + "learning_rate": 5.637153439376508e-06, + "loss": 0.0786, + "step": 38079 + }, + { + "epoch": 0.8926551685742248, + "grad_norm": 0.1220681443810463, + "learning_rate": 5.634715449087746e-06, + "loss": 0.0237, + "step": 38080 + }, + { + "epoch": 0.8926786101490298, + "grad_norm": 0.6278554797172546, + "learning_rate": 5.63227797082767e-06, + "loss": 0.0898, + "step": 38081 + }, + { + "epoch": 0.8927020517238348, + "grad_norm": 0.34523338079452515, + "learning_rate": 5.629841004609493e-06, + "loss": 0.049, + "step": 38082 + }, + { + "epoch": 0.8927254932986398, + "grad_norm": 0.27183857560157776, + "learning_rate": 5.627404550446469e-06, + "loss": 0.0304, + "step": 38083 + }, + { + "epoch": 0.8927489348734448, + "grad_norm": 0.4887891113758087, + "learning_rate": 5.624968608351799e-06, + "loss": 0.5595, + "step": 38084 + }, + { + "epoch": 0.8927723764482498, + "grad_norm": 0.5220147967338562, + "learning_rate": 5.622533178338707e-06, + "loss": 0.6278, + "step": 38085 + }, + { + "epoch": 0.8927958180230547, + "grad_norm": 0.4089975357055664, + "learning_rate": 5.620098260420381e-06, + "loss": 0.0788, + "step": 38086 + }, + { + "epoch": 0.8928192595978598, + "grad_norm": 0.42820465564727783, + "learning_rate": 5.617663854610078e-06, + "loss": 0.0783, + "step": 38087 + }, + { + "epoch": 0.8928427011726647, + "grad_norm": 1.0243486166000366, + "learning_rate": 5.615229960920987e-06, + "loss": 0.1569, + "step": 38088 + }, + { + "epoch": 0.8928661427474698, + "grad_norm": 0.26493561267852783, + "learning_rate": 5.612796579366297e-06, + "loss": 0.0155, + "step": 38089 + }, + { + "epoch": 0.8928895843222747, + "grad_norm": 0.18470916152000427, + "learning_rate": 5.610363709959243e-06, + "loss": 0.0239, + "step": 38090 + }, + { + "epoch": 0.8929130258970798, + "grad_norm": 0.23091799020767212, + "learning_rate": 5.607931352713003e-06, + "loss": 0.0469, + "step": 38091 + }, + { + "epoch": 0.8929364674718847, + "grad_norm": 0.3982636332511902, + "learning_rate": 5.605499507640766e-06, + "loss": 0.0621, + "step": 38092 + }, + { + "epoch": 0.8929599090466898, + "grad_norm": 0.5412545800209045, + "learning_rate": 5.603068174755755e-06, + "loss": 0.5454, + "step": 38093 + }, + { + "epoch": 0.8929833506214947, + "grad_norm": 0.11578110605478287, + "learning_rate": 5.600637354071147e-06, + "loss": 0.0168, + "step": 38094 + }, + { + "epoch": 0.8930067921962997, + "grad_norm": 0.534846842288971, + "learning_rate": 5.598207045600135e-06, + "loss": 0.0922, + "step": 38095 + }, + { + "epoch": 0.8930302337711048, + "grad_norm": 0.05875658988952637, + "learning_rate": 5.5957772493559154e-06, + "loss": 0.0071, + "step": 38096 + }, + { + "epoch": 0.8930536753459097, + "grad_norm": 0.35494667291641235, + "learning_rate": 5.593347965351658e-06, + "loss": 0.2896, + "step": 38097 + }, + { + "epoch": 0.8930771169207148, + "grad_norm": 0.5804619789123535, + "learning_rate": 5.590919193600552e-06, + "loss": 0.1084, + "step": 38098 + }, + { + "epoch": 0.8931005584955197, + "grad_norm": 0.2347451150417328, + "learning_rate": 5.588490934115764e-06, + "loss": 0.0301, + "step": 38099 + }, + { + "epoch": 0.8931240000703248, + "grad_norm": 0.25372710824012756, + "learning_rate": 5.586063186910484e-06, + "loss": 0.0229, + "step": 38100 + }, + { + "epoch": 0.8931474416451297, + "grad_norm": 0.43670400977134705, + "learning_rate": 5.583635951997878e-06, + "loss": 0.0744, + "step": 38101 + }, + { + "epoch": 0.8931708832199348, + "grad_norm": 0.3727593421936035, + "learning_rate": 5.581209229391126e-06, + "loss": 0.0546, + "step": 38102 + }, + { + "epoch": 0.8931943247947397, + "grad_norm": 0.101983942091465, + "learning_rate": 5.578783019103384e-06, + "loss": 0.0084, + "step": 38103 + }, + { + "epoch": 0.8932177663695448, + "grad_norm": 0.26428529620170593, + "learning_rate": 5.5763573211478074e-06, + "loss": 0.0375, + "step": 38104 + }, + { + "epoch": 0.8932412079443497, + "grad_norm": 0.6370849609375, + "learning_rate": 5.573932135537585e-06, + "loss": 0.4264, + "step": 38105 + }, + { + "epoch": 0.8932646495191547, + "grad_norm": 0.39154741168022156, + "learning_rate": 5.571507462285852e-06, + "loss": 0.0578, + "step": 38106 + }, + { + "epoch": 0.8932880910939597, + "grad_norm": 0.21351221203804016, + "learning_rate": 5.5690833014057864e-06, + "loss": 0.0261, + "step": 38107 + }, + { + "epoch": 0.8933115326687647, + "grad_norm": 0.3149591386318207, + "learning_rate": 5.566659652910522e-06, + "loss": 0.0448, + "step": 38108 + }, + { + "epoch": 0.8933349742435697, + "grad_norm": 0.455294132232666, + "learning_rate": 5.564236516813226e-06, + "loss": 0.0992, + "step": 38109 + }, + { + "epoch": 0.8933584158183747, + "grad_norm": 0.7970960140228271, + "learning_rate": 5.561813893127032e-06, + "loss": 0.2081, + "step": 38110 + }, + { + "epoch": 0.8933818573931797, + "grad_norm": 0.43891826272010803, + "learning_rate": 5.559391781865075e-06, + "loss": 0.0562, + "step": 38111 + }, + { + "epoch": 0.8934052989679847, + "grad_norm": 0.314525842666626, + "learning_rate": 5.556970183040533e-06, + "loss": 0.0537, + "step": 38112 + }, + { + "epoch": 0.8934287405427896, + "grad_norm": 0.59517902135849, + "learning_rate": 5.554549096666528e-06, + "loss": 0.0927, + "step": 38113 + }, + { + "epoch": 0.8934521821175947, + "grad_norm": 0.35572314262390137, + "learning_rate": 5.552128522756195e-06, + "loss": 0.0621, + "step": 38114 + }, + { + "epoch": 0.8934756236923996, + "grad_norm": 0.5017609000205994, + "learning_rate": 5.549708461322667e-06, + "loss": 0.0923, + "step": 38115 + }, + { + "epoch": 0.8934990652672047, + "grad_norm": 0.02914826199412346, + "learning_rate": 5.547288912379078e-06, + "loss": 0.0022, + "step": 38116 + }, + { + "epoch": 0.8935225068420096, + "grad_norm": 0.24813984334468842, + "learning_rate": 5.544869875938541e-06, + "loss": 0.0596, + "step": 38117 + }, + { + "epoch": 0.8935459484168147, + "grad_norm": 0.2599330544471741, + "learning_rate": 5.542451352014199e-06, + "loss": 0.2207, + "step": 38118 + }, + { + "epoch": 0.8935693899916196, + "grad_norm": 0.28283601999282837, + "learning_rate": 5.540033340619177e-06, + "loss": 0.0489, + "step": 38119 + }, + { + "epoch": 0.8935928315664247, + "grad_norm": 0.4915640354156494, + "learning_rate": 5.5376158417665855e-06, + "loss": 0.0677, + "step": 38120 + }, + { + "epoch": 0.8936162731412296, + "grad_norm": 0.32097724080085754, + "learning_rate": 5.535198855469559e-06, + "loss": 0.0499, + "step": 38121 + }, + { + "epoch": 0.8936397147160346, + "grad_norm": 0.15306486189365387, + "learning_rate": 5.532782381741197e-06, + "loss": 0.0119, + "step": 38122 + }, + { + "epoch": 0.8936631562908396, + "grad_norm": 0.31269893050193787, + "learning_rate": 5.530366420594612e-06, + "loss": 0.0665, + "step": 38123 + }, + { + "epoch": 0.8936865978656446, + "grad_norm": 0.6063216328620911, + "learning_rate": 5.5279509720429055e-06, + "loss": 0.1181, + "step": 38124 + }, + { + "epoch": 0.8937100394404496, + "grad_norm": 0.3327799141407013, + "learning_rate": 5.525536036099199e-06, + "loss": 0.0322, + "step": 38125 + }, + { + "epoch": 0.8937334810152546, + "grad_norm": 0.32793909311294556, + "learning_rate": 5.523121612776594e-06, + "loss": 0.0559, + "step": 38126 + }, + { + "epoch": 0.8937569225900596, + "grad_norm": 0.3289411962032318, + "learning_rate": 5.52070770208819e-06, + "loss": 0.0428, + "step": 38127 + }, + { + "epoch": 0.8937803641648646, + "grad_norm": 0.24952921271324158, + "learning_rate": 5.518294304047078e-06, + "loss": 0.0401, + "step": 38128 + }, + { + "epoch": 0.8938038057396696, + "grad_norm": 0.7661489248275757, + "learning_rate": 5.515881418666369e-06, + "loss": 0.1211, + "step": 38129 + }, + { + "epoch": 0.8938272473144746, + "grad_norm": 0.11725866049528122, + "learning_rate": 5.513469045959118e-06, + "loss": 0.0162, + "step": 38130 + }, + { + "epoch": 0.8938506888892795, + "grad_norm": 0.2632697820663452, + "learning_rate": 5.511057185938439e-06, + "loss": 0.0261, + "step": 38131 + }, + { + "epoch": 0.8938741304640846, + "grad_norm": 0.5191553831100464, + "learning_rate": 5.5086458386174435e-06, + "loss": 0.0975, + "step": 38132 + }, + { + "epoch": 0.8938975720388895, + "grad_norm": 0.12764692306518555, + "learning_rate": 5.506235004009186e-06, + "loss": 0.0185, + "step": 38133 + }, + { + "epoch": 0.8939210136136946, + "grad_norm": 0.6079715490341187, + "learning_rate": 5.503824682126757e-06, + "loss": 0.5009, + "step": 38134 + }, + { + "epoch": 0.8939444551884995, + "grad_norm": 0.22469958662986755, + "learning_rate": 5.501414872983235e-06, + "loss": 0.0111, + "step": 38135 + }, + { + "epoch": 0.8939678967633046, + "grad_norm": 0.2520405650138855, + "learning_rate": 5.499005576591687e-06, + "loss": 0.0318, + "step": 38136 + }, + { + "epoch": 0.8939913383381095, + "grad_norm": 0.23580174148082733, + "learning_rate": 5.496596792965181e-06, + "loss": 0.0251, + "step": 38137 + }, + { + "epoch": 0.8940147799129146, + "grad_norm": 0.37065568566322327, + "learning_rate": 5.494188522116816e-06, + "loss": 0.0428, + "step": 38138 + }, + { + "epoch": 0.8940382214877195, + "grad_norm": 1.131326675415039, + "learning_rate": 5.491780764059639e-06, + "loss": 0.1432, + "step": 38139 + }, + { + "epoch": 0.8940616630625245, + "grad_norm": 0.5256208181381226, + "learning_rate": 5.489373518806718e-06, + "loss": 0.1027, + "step": 38140 + }, + { + "epoch": 0.8940851046373295, + "grad_norm": 0.05376942828297615, + "learning_rate": 5.486966786371106e-06, + "loss": 0.0072, + "step": 38141 + }, + { + "epoch": 0.8941085462121345, + "grad_norm": 0.15136389434337616, + "learning_rate": 5.484560566765862e-06, + "loss": 0.023, + "step": 38142 + }, + { + "epoch": 0.8941319877869395, + "grad_norm": 0.32668522000312805, + "learning_rate": 5.482154860004052e-06, + "loss": 0.0773, + "step": 38143 + }, + { + "epoch": 0.8941554293617445, + "grad_norm": 0.37088340520858765, + "learning_rate": 5.4797496660987325e-06, + "loss": 0.0709, + "step": 38144 + }, + { + "epoch": 0.8941788709365495, + "grad_norm": 0.25453582406044006, + "learning_rate": 5.47734498506296e-06, + "loss": 0.0483, + "step": 38145 + }, + { + "epoch": 0.8942023125113545, + "grad_norm": 0.16979607939720154, + "learning_rate": 5.474940816909768e-06, + "loss": 0.0169, + "step": 38146 + }, + { + "epoch": 0.8942257540861596, + "grad_norm": 0.5008288621902466, + "learning_rate": 5.472537161652203e-06, + "loss": 0.6034, + "step": 38147 + }, + { + "epoch": 0.8942491956609645, + "grad_norm": 0.3513907492160797, + "learning_rate": 5.470134019303308e-06, + "loss": 0.0765, + "step": 38148 + }, + { + "epoch": 0.8942726372357696, + "grad_norm": 0.5674881935119629, + "learning_rate": 5.467731389876107e-06, + "loss": 0.0795, + "step": 38149 + }, + { + "epoch": 0.8942960788105745, + "grad_norm": 0.09733645617961884, + "learning_rate": 5.4653292733836794e-06, + "loss": 0.0176, + "step": 38150 + }, + { + "epoch": 0.8943195203853795, + "grad_norm": 0.3168669044971466, + "learning_rate": 5.462927669839024e-06, + "loss": 0.0318, + "step": 38151 + }, + { + "epoch": 0.8943429619601845, + "grad_norm": 0.39502018690109253, + "learning_rate": 5.460526579255176e-06, + "loss": 0.0943, + "step": 38152 + }, + { + "epoch": 0.8943664035349895, + "grad_norm": 0.4762381911277771, + "learning_rate": 5.458126001645181e-06, + "loss": 0.0902, + "step": 38153 + }, + { + "epoch": 0.8943898451097945, + "grad_norm": 0.4365522563457489, + "learning_rate": 5.45572593702205e-06, + "loss": 0.0663, + "step": 38154 + }, + { + "epoch": 0.8944132866845995, + "grad_norm": 0.11396954953670502, + "learning_rate": 5.4533263853987825e-06, + "loss": 0.0148, + "step": 38155 + }, + { + "epoch": 0.8944367282594045, + "grad_norm": 0.3148055374622345, + "learning_rate": 5.4509273467884366e-06, + "loss": 0.0522, + "step": 38156 + }, + { + "epoch": 0.8944601698342095, + "grad_norm": 0.12027812749147415, + "learning_rate": 5.448528821204036e-06, + "loss": 0.0212, + "step": 38157 + }, + { + "epoch": 0.8944836114090144, + "grad_norm": 0.6498952507972717, + "learning_rate": 5.4461308086585675e-06, + "loss": 0.7013, + "step": 38158 + }, + { + "epoch": 0.8945070529838195, + "grad_norm": 0.3417576849460602, + "learning_rate": 5.443733309165044e-06, + "loss": 0.0704, + "step": 38159 + }, + { + "epoch": 0.8945304945586244, + "grad_norm": 0.41950586438179016, + "learning_rate": 5.441336322736501e-06, + "loss": 0.0534, + "step": 38160 + }, + { + "epoch": 0.8945539361334295, + "grad_norm": 0.659680962562561, + "learning_rate": 5.4389398493859156e-06, + "loss": 0.0483, + "step": 38161 + }, + { + "epoch": 0.8945773777082344, + "grad_norm": 0.42568567395210266, + "learning_rate": 5.436543889126289e-06, + "loss": 0.0818, + "step": 38162 + }, + { + "epoch": 0.8946008192830395, + "grad_norm": 0.6729206442832947, + "learning_rate": 5.434148441970643e-06, + "loss": 0.1006, + "step": 38163 + }, + { + "epoch": 0.8946242608578444, + "grad_norm": 0.41526493430137634, + "learning_rate": 5.4317535079319694e-06, + "loss": 0.0411, + "step": 38164 + }, + { + "epoch": 0.8946477024326495, + "grad_norm": 0.6283143758773804, + "learning_rate": 5.429359087023267e-06, + "loss": 0.1116, + "step": 38165 + }, + { + "epoch": 0.8946711440074544, + "grad_norm": 0.4247898757457733, + "learning_rate": 5.426965179257515e-06, + "loss": 0.5153, + "step": 38166 + }, + { + "epoch": 0.8946945855822594, + "grad_norm": 0.42054715752601624, + "learning_rate": 5.424571784647692e-06, + "loss": 0.0877, + "step": 38167 + }, + { + "epoch": 0.8947180271570644, + "grad_norm": 0.5939892530441284, + "learning_rate": 5.422178903206821e-06, + "loss": 0.518, + "step": 38168 + }, + { + "epoch": 0.8947414687318694, + "grad_norm": 0.45655328035354614, + "learning_rate": 5.419786534947846e-06, + "loss": 0.0683, + "step": 38169 + }, + { + "epoch": 0.8947649103066744, + "grad_norm": 0.33004412055015564, + "learning_rate": 5.4173946798837915e-06, + "loss": 0.0435, + "step": 38170 + }, + { + "epoch": 0.8947883518814794, + "grad_norm": 0.12807878851890564, + "learning_rate": 5.415003338027613e-06, + "loss": 0.0348, + "step": 38171 + }, + { + "epoch": 0.8948117934562844, + "grad_norm": 0.11695300042629242, + "learning_rate": 5.412612509392279e-06, + "loss": 0.0216, + "step": 38172 + }, + { + "epoch": 0.8948352350310894, + "grad_norm": 0.06147585064172745, + "learning_rate": 5.410222193990777e-06, + "loss": 0.0083, + "step": 38173 + }, + { + "epoch": 0.8948586766058944, + "grad_norm": 0.2361348569393158, + "learning_rate": 5.407832391836065e-06, + "loss": 0.0212, + "step": 38174 + }, + { + "epoch": 0.8948821181806994, + "grad_norm": 0.6766130328178406, + "learning_rate": 5.405443102941099e-06, + "loss": 0.5539, + "step": 38175 + }, + { + "epoch": 0.8949055597555043, + "grad_norm": 0.536090075969696, + "learning_rate": 5.403054327318868e-06, + "loss": 0.1009, + "step": 38176 + }, + { + "epoch": 0.8949290013303094, + "grad_norm": 0.1680833250284195, + "learning_rate": 5.400666064982329e-06, + "loss": 0.0319, + "step": 38177 + }, + { + "epoch": 0.8949524429051143, + "grad_norm": 0.08820222318172455, + "learning_rate": 5.398278315944438e-06, + "loss": 0.0141, + "step": 38178 + }, + { + "epoch": 0.8949758844799194, + "grad_norm": 0.8894732594490051, + "learning_rate": 5.39589108021813e-06, + "loss": 0.1295, + "step": 38179 + }, + { + "epoch": 0.8949993260547243, + "grad_norm": 0.4330958127975464, + "learning_rate": 5.393504357816392e-06, + "loss": 0.0681, + "step": 38180 + }, + { + "epoch": 0.8950227676295294, + "grad_norm": 0.5006887316703796, + "learning_rate": 5.39111814875215e-06, + "loss": 0.0863, + "step": 38181 + }, + { + "epoch": 0.8950462092043343, + "grad_norm": 0.2822452187538147, + "learning_rate": 5.38873245303837e-06, + "loss": 0.0753, + "step": 38182 + }, + { + "epoch": 0.8950696507791394, + "grad_norm": 0.13370440900325775, + "learning_rate": 5.386347270687997e-06, + "loss": 0.0196, + "step": 38183 + }, + { + "epoch": 0.8950930923539443, + "grad_norm": 0.2358957976102829, + "learning_rate": 5.3839626017139545e-06, + "loss": 0.0271, + "step": 38184 + }, + { + "epoch": 0.8951165339287493, + "grad_norm": 0.47442805767059326, + "learning_rate": 5.381578446129198e-06, + "loss": 0.1365, + "step": 38185 + }, + { + "epoch": 0.8951399755035543, + "grad_norm": 0.3665647804737091, + "learning_rate": 5.3791948039466525e-06, + "loss": 0.0365, + "step": 38186 + }, + { + "epoch": 0.8951634170783593, + "grad_norm": 0.518194317817688, + "learning_rate": 5.376811675179239e-06, + "loss": 0.0241, + "step": 38187 + }, + { + "epoch": 0.8951868586531643, + "grad_norm": 0.24685893952846527, + "learning_rate": 5.374429059839925e-06, + "loss": 0.0334, + "step": 38188 + }, + { + "epoch": 0.8952103002279693, + "grad_norm": 0.3573038876056671, + "learning_rate": 5.372046957941623e-06, + "loss": 0.0318, + "step": 38189 + }, + { + "epoch": 0.8952337418027743, + "grad_norm": 0.2523064911365509, + "learning_rate": 5.369665369497245e-06, + "loss": 0.0291, + "step": 38190 + }, + { + "epoch": 0.8952571833775793, + "grad_norm": 0.1829608976840973, + "learning_rate": 5.367284294519736e-06, + "loss": 0.0231, + "step": 38191 + }, + { + "epoch": 0.8952806249523843, + "grad_norm": 0.4665127694606781, + "learning_rate": 5.364903733021975e-06, + "loss": 0.0871, + "step": 38192 + }, + { + "epoch": 0.8953040665271893, + "grad_norm": 0.34288573265075684, + "learning_rate": 5.362523685016929e-06, + "loss": 0.4082, + "step": 38193 + }, + { + "epoch": 0.8953275081019942, + "grad_norm": 0.11570106446743011, + "learning_rate": 5.360144150517476e-06, + "loss": 0.0167, + "step": 38194 + }, + { + "epoch": 0.8953509496767993, + "grad_norm": 0.2918853163719177, + "learning_rate": 5.357765129536563e-06, + "loss": 0.0605, + "step": 38195 + }, + { + "epoch": 0.8953743912516042, + "grad_norm": 0.23028023540973663, + "learning_rate": 5.355386622087066e-06, + "loss": 0.0421, + "step": 38196 + }, + { + "epoch": 0.8953978328264093, + "grad_norm": 0.19631652534008026, + "learning_rate": 5.35300862818191e-06, + "loss": 0.0116, + "step": 38197 + }, + { + "epoch": 0.8954212744012143, + "grad_norm": 1.0018953084945679, + "learning_rate": 5.350631147833984e-06, + "loss": 0.1909, + "step": 38198 + }, + { + "epoch": 0.8954447159760193, + "grad_norm": 0.2033129632472992, + "learning_rate": 5.3482541810562004e-06, + "loss": 0.0458, + "step": 38199 + }, + { + "epoch": 0.8954681575508243, + "grad_norm": 0.11916209012269974, + "learning_rate": 5.345877727861437e-06, + "loss": 0.0139, + "step": 38200 + }, + { + "epoch": 0.8954915991256293, + "grad_norm": 0.07730481028556824, + "learning_rate": 5.343501788262606e-06, + "loss": 0.0115, + "step": 38201 + }, + { + "epoch": 0.8955150407004343, + "grad_norm": 0.6156221628189087, + "learning_rate": 5.341126362272608e-06, + "loss": 0.1686, + "step": 38202 + }, + { + "epoch": 0.8955384822752392, + "grad_norm": 0.5783931612968445, + "learning_rate": 5.338751449904322e-06, + "loss": 0.6237, + "step": 38203 + }, + { + "epoch": 0.8955619238500443, + "grad_norm": 0.24285954236984253, + "learning_rate": 5.3363770511706044e-06, + "loss": 0.031, + "step": 38204 + }, + { + "epoch": 0.8955853654248492, + "grad_norm": 0.24989189207553864, + "learning_rate": 5.334003166084389e-06, + "loss": 0.0393, + "step": 38205 + }, + { + "epoch": 0.8956088069996543, + "grad_norm": 0.39514657855033875, + "learning_rate": 5.331629794658533e-06, + "loss": 0.0637, + "step": 38206 + }, + { + "epoch": 0.8956322485744592, + "grad_norm": 0.24587923288345337, + "learning_rate": 5.329256936905902e-06, + "loss": 0.0456, + "step": 38207 + }, + { + "epoch": 0.8956556901492643, + "grad_norm": 0.10368599742650986, + "learning_rate": 5.326884592839398e-06, + "loss": 0.014, + "step": 38208 + }, + { + "epoch": 0.8956791317240692, + "grad_norm": 0.1525108516216278, + "learning_rate": 5.32451276247189e-06, + "loss": 0.032, + "step": 38209 + }, + { + "epoch": 0.8957025732988743, + "grad_norm": 0.17404095828533173, + "learning_rate": 5.32214144581622e-06, + "loss": 0.0349, + "step": 38210 + }, + { + "epoch": 0.8957260148736792, + "grad_norm": 0.26863324642181396, + "learning_rate": 5.31977064288528e-06, + "loss": 0.0258, + "step": 38211 + }, + { + "epoch": 0.8957494564484842, + "grad_norm": 0.6093875169754028, + "learning_rate": 5.317400353691915e-06, + "loss": 0.0793, + "step": 38212 + }, + { + "epoch": 0.8957728980232892, + "grad_norm": 0.5634408593177795, + "learning_rate": 5.315030578249014e-06, + "loss": 0.1234, + "step": 38213 + }, + { + "epoch": 0.8957963395980942, + "grad_norm": 0.17599380016326904, + "learning_rate": 5.312661316569423e-06, + "loss": 0.0273, + "step": 38214 + }, + { + "epoch": 0.8958197811728992, + "grad_norm": 0.45621296763420105, + "learning_rate": 5.310292568665986e-06, + "loss": 0.0696, + "step": 38215 + }, + { + "epoch": 0.8958432227477042, + "grad_norm": 0.5072683095932007, + "learning_rate": 5.307924334551551e-06, + "loss": 0.0992, + "step": 38216 + }, + { + "epoch": 0.8958666643225092, + "grad_norm": 0.9131737947463989, + "learning_rate": 5.305556614238994e-06, + "loss": 0.1838, + "step": 38217 + }, + { + "epoch": 0.8958901058973142, + "grad_norm": 0.2629393935203552, + "learning_rate": 5.303189407741149e-06, + "loss": 0.0436, + "step": 38218 + }, + { + "epoch": 0.8959135474721192, + "grad_norm": 0.5556791424751282, + "learning_rate": 5.300822715070852e-06, + "loss": 0.1145, + "step": 38219 + }, + { + "epoch": 0.8959369890469242, + "grad_norm": 0.5851923227310181, + "learning_rate": 5.298456536240959e-06, + "loss": 0.091, + "step": 38220 + }, + { + "epoch": 0.8959604306217291, + "grad_norm": 0.36488577723503113, + "learning_rate": 5.2960908712643145e-06, + "loss": 0.0767, + "step": 38221 + }, + { + "epoch": 0.8959838721965342, + "grad_norm": 0.26041823625564575, + "learning_rate": 5.2937257201537305e-06, + "loss": 0.0554, + "step": 38222 + }, + { + "epoch": 0.8960073137713391, + "grad_norm": 0.42625564336776733, + "learning_rate": 5.291361082922064e-06, + "loss": 0.5439, + "step": 38223 + }, + { + "epoch": 0.8960307553461442, + "grad_norm": 0.4109346568584442, + "learning_rate": 5.288996959582138e-06, + "loss": 0.2183, + "step": 38224 + }, + { + "epoch": 0.8960541969209491, + "grad_norm": 0.42469286918640137, + "learning_rate": 5.286633350146752e-06, + "loss": 0.0444, + "step": 38225 + }, + { + "epoch": 0.8960776384957542, + "grad_norm": 0.5587431192398071, + "learning_rate": 5.284270254628776e-06, + "loss": 0.6464, + "step": 38226 + }, + { + "epoch": 0.8961010800705591, + "grad_norm": 0.37132808566093445, + "learning_rate": 5.28190767304102e-06, + "loss": 0.0476, + "step": 38227 + }, + { + "epoch": 0.8961245216453642, + "grad_norm": 0.36254146695137024, + "learning_rate": 5.279545605396274e-06, + "loss": 0.0426, + "step": 38228 + }, + { + "epoch": 0.8961479632201691, + "grad_norm": 0.2491067349910736, + "learning_rate": 5.277184051707395e-06, + "loss": 0.0452, + "step": 38229 + }, + { + "epoch": 0.8961714047949741, + "grad_norm": 0.1918036788702011, + "learning_rate": 5.274823011987173e-06, + "loss": 0.0425, + "step": 38230 + }, + { + "epoch": 0.8961948463697791, + "grad_norm": 0.35119253396987915, + "learning_rate": 5.272462486248431e-06, + "loss": 0.0357, + "step": 38231 + }, + { + "epoch": 0.8962182879445841, + "grad_norm": 0.3831726908683777, + "learning_rate": 5.2701024745039575e-06, + "loss": 0.0526, + "step": 38232 + }, + { + "epoch": 0.8962417295193891, + "grad_norm": 0.14427849650382996, + "learning_rate": 5.267742976766576e-06, + "loss": 0.0264, + "step": 38233 + }, + { + "epoch": 0.8962651710941941, + "grad_norm": 0.49838384985923767, + "learning_rate": 5.265383993049089e-06, + "loss": 0.4683, + "step": 38234 + }, + { + "epoch": 0.8962886126689991, + "grad_norm": 0.468548059463501, + "learning_rate": 5.263025523364296e-06, + "loss": 0.0916, + "step": 38235 + }, + { + "epoch": 0.8963120542438041, + "grad_norm": 0.14696814119815826, + "learning_rate": 5.260667567724986e-06, + "loss": 0.0155, + "step": 38236 + }, + { + "epoch": 0.896335495818609, + "grad_norm": 0.4155886173248291, + "learning_rate": 5.2583101261439615e-06, + "loss": 0.0584, + "step": 38237 + }, + { + "epoch": 0.8963589373934141, + "grad_norm": 0.08150853216648102, + "learning_rate": 5.255953198634001e-06, + "loss": 0.0106, + "step": 38238 + }, + { + "epoch": 0.896382378968219, + "grad_norm": 0.3961971700191498, + "learning_rate": 5.253596785207904e-06, + "loss": 0.0625, + "step": 38239 + }, + { + "epoch": 0.8964058205430241, + "grad_norm": 0.24344754219055176, + "learning_rate": 5.25124088587845e-06, + "loss": 0.0154, + "step": 38240 + }, + { + "epoch": 0.896429262117829, + "grad_norm": 0.24592097103595734, + "learning_rate": 5.24888550065844e-06, + "loss": 0.0396, + "step": 38241 + }, + { + "epoch": 0.8964527036926341, + "grad_norm": 0.8582802414894104, + "learning_rate": 5.246530629560642e-06, + "loss": 0.0536, + "step": 38242 + }, + { + "epoch": 0.896476145267439, + "grad_norm": 0.4168851971626282, + "learning_rate": 5.244176272597822e-06, + "loss": 0.0696, + "step": 38243 + }, + { + "epoch": 0.8964995868422441, + "grad_norm": 0.0817231610417366, + "learning_rate": 5.241822429782772e-06, + "loss": 0.0119, + "step": 38244 + }, + { + "epoch": 0.896523028417049, + "grad_norm": 0.3230926990509033, + "learning_rate": 5.239469101128247e-06, + "loss": 0.0349, + "step": 38245 + }, + { + "epoch": 0.8965464699918541, + "grad_norm": 0.2912740409374237, + "learning_rate": 5.237116286647048e-06, + "loss": 0.0529, + "step": 38246 + }, + { + "epoch": 0.896569911566659, + "grad_norm": 0.4081452190876007, + "learning_rate": 5.234763986351909e-06, + "loss": 0.0522, + "step": 38247 + }, + { + "epoch": 0.896593353141464, + "grad_norm": 0.5291147232055664, + "learning_rate": 5.232412200255609e-06, + "loss": 0.0932, + "step": 38248 + }, + { + "epoch": 0.8966167947162691, + "grad_norm": 0.4035654067993164, + "learning_rate": 5.230060928370905e-06, + "loss": 0.5162, + "step": 38249 + }, + { + "epoch": 0.896640236291074, + "grad_norm": 0.9842692017555237, + "learning_rate": 5.227710170710542e-06, + "loss": 0.1446, + "step": 38250 + }, + { + "epoch": 0.8966636778658791, + "grad_norm": 0.7672808766365051, + "learning_rate": 5.225359927287299e-06, + "loss": 0.1226, + "step": 38251 + }, + { + "epoch": 0.896687119440684, + "grad_norm": 0.5046867728233337, + "learning_rate": 5.223010198113909e-06, + "loss": 0.0578, + "step": 38252 + }, + { + "epoch": 0.8967105610154891, + "grad_norm": 0.46257174015045166, + "learning_rate": 5.220660983203141e-06, + "loss": 0.0585, + "step": 38253 + }, + { + "epoch": 0.896734002590294, + "grad_norm": 0.3170796036720276, + "learning_rate": 5.218312282567739e-06, + "loss": 0.0411, + "step": 38254 + }, + { + "epoch": 0.8967574441650991, + "grad_norm": 0.2116033434867859, + "learning_rate": 5.215964096220427e-06, + "loss": 0.0302, + "step": 38255 + }, + { + "epoch": 0.896780885739904, + "grad_norm": 0.6171924471855164, + "learning_rate": 5.213616424173962e-06, + "loss": 0.5281, + "step": 38256 + }, + { + "epoch": 0.896804327314709, + "grad_norm": 0.15399527549743652, + "learning_rate": 5.211269266441077e-06, + "loss": 0.0093, + "step": 38257 + }, + { + "epoch": 0.896827768889514, + "grad_norm": 0.44398385286331177, + "learning_rate": 5.208922623034507e-06, + "loss": 0.1163, + "step": 38258 + }, + { + "epoch": 0.896851210464319, + "grad_norm": 0.3645056188106537, + "learning_rate": 5.206576493966997e-06, + "loss": 0.0453, + "step": 38259 + }, + { + "epoch": 0.896874652039124, + "grad_norm": 0.2056359052658081, + "learning_rate": 5.204230879251271e-06, + "loss": 0.0236, + "step": 38260 + }, + { + "epoch": 0.896898093613929, + "grad_norm": 0.41872304677963257, + "learning_rate": 5.20188577890005e-06, + "loss": 0.065, + "step": 38261 + }, + { + "epoch": 0.896921535188734, + "grad_norm": 0.4458984136581421, + "learning_rate": 5.1995411929260606e-06, + "loss": 0.0599, + "step": 38262 + }, + { + "epoch": 0.896944976763539, + "grad_norm": 0.422911673784256, + "learning_rate": 5.1971971213420124e-06, + "loss": 0.094, + "step": 38263 + }, + { + "epoch": 0.896968418338344, + "grad_norm": 0.18034732341766357, + "learning_rate": 5.194853564160662e-06, + "loss": 0.0278, + "step": 38264 + }, + { + "epoch": 0.896991859913149, + "grad_norm": 0.28928956389427185, + "learning_rate": 5.192510521394678e-06, + "loss": 0.058, + "step": 38265 + }, + { + "epoch": 0.8970153014879539, + "grad_norm": 0.431493878364563, + "learning_rate": 5.190167993056816e-06, + "loss": 0.0823, + "step": 38266 + }, + { + "epoch": 0.897038743062759, + "grad_norm": 0.528792679309845, + "learning_rate": 5.187825979159766e-06, + "loss": 0.6184, + "step": 38267 + }, + { + "epoch": 0.8970621846375639, + "grad_norm": 0.4685317277908325, + "learning_rate": 5.185484479716241e-06, + "loss": 0.0857, + "step": 38268 + }, + { + "epoch": 0.897085626212369, + "grad_norm": 0.36941397190093994, + "learning_rate": 5.1831434947389405e-06, + "loss": 0.09, + "step": 38269 + }, + { + "epoch": 0.8971090677871739, + "grad_norm": 0.5032102465629578, + "learning_rate": 5.180803024240554e-06, + "loss": 0.083, + "step": 38270 + }, + { + "epoch": 0.897132509361979, + "grad_norm": 0.5964880585670471, + "learning_rate": 5.178463068233807e-06, + "loss": 0.0791, + "step": 38271 + }, + { + "epoch": 0.8971559509367839, + "grad_norm": 0.36574599146842957, + "learning_rate": 5.176123626731389e-06, + "loss": 0.0405, + "step": 38272 + }, + { + "epoch": 0.897179392511589, + "grad_norm": 0.08668634295463562, + "learning_rate": 5.173784699745998e-06, + "loss": 0.0082, + "step": 38273 + }, + { + "epoch": 0.8972028340863939, + "grad_norm": 0.4685659408569336, + "learning_rate": 5.1714462872903045e-06, + "loss": 0.1071, + "step": 38274 + }, + { + "epoch": 0.897226275661199, + "grad_norm": 0.3555941879749298, + "learning_rate": 5.1691083893770195e-06, + "loss": 0.0451, + "step": 38275 + }, + { + "epoch": 0.8972497172360039, + "grad_norm": 0.6621829867362976, + "learning_rate": 5.1667710060188e-06, + "loss": 0.0902, + "step": 38276 + }, + { + "epoch": 0.8972731588108089, + "grad_norm": 0.690384566783905, + "learning_rate": 5.164434137228347e-06, + "loss": 0.0896, + "step": 38277 + }, + { + "epoch": 0.8972966003856139, + "grad_norm": 0.35614803433418274, + "learning_rate": 5.1620977830183605e-06, + "loss": 0.0568, + "step": 38278 + }, + { + "epoch": 0.8973200419604189, + "grad_norm": 0.5852177143096924, + "learning_rate": 5.159761943401498e-06, + "loss": 0.1451, + "step": 38279 + }, + { + "epoch": 0.8973434835352239, + "grad_norm": 0.34309858083724976, + "learning_rate": 5.157426618390426e-06, + "loss": 0.0426, + "step": 38280 + }, + { + "epoch": 0.8973669251100289, + "grad_norm": 0.5931301116943359, + "learning_rate": 5.155091807997825e-06, + "loss": 0.0398, + "step": 38281 + }, + { + "epoch": 0.8973903666848339, + "grad_norm": 0.3569474220275879, + "learning_rate": 5.152757512236372e-06, + "loss": 0.0507, + "step": 38282 + }, + { + "epoch": 0.8974138082596389, + "grad_norm": 0.3123340308666229, + "learning_rate": 5.150423731118703e-06, + "loss": 0.0516, + "step": 38283 + }, + { + "epoch": 0.8974372498344438, + "grad_norm": 0.11388206481933594, + "learning_rate": 5.148090464657518e-06, + "loss": 0.0222, + "step": 38284 + }, + { + "epoch": 0.8974606914092489, + "grad_norm": 0.5694579482078552, + "learning_rate": 5.1457577128654625e-06, + "loss": 0.117, + "step": 38285 + }, + { + "epoch": 0.8974841329840538, + "grad_norm": 0.32282429933547974, + "learning_rate": 5.1434254757551815e-06, + "loss": 0.0368, + "step": 38286 + }, + { + "epoch": 0.8975075745588589, + "grad_norm": 0.4775063991546631, + "learning_rate": 5.141093753339355e-06, + "loss": 0.05, + "step": 38287 + }, + { + "epoch": 0.8975310161336638, + "grad_norm": 0.5877529978752136, + "learning_rate": 5.138762545630604e-06, + "loss": 0.0935, + "step": 38288 + }, + { + "epoch": 0.8975544577084689, + "grad_norm": 0.4360826313495636, + "learning_rate": 5.136431852641599e-06, + "loss": 0.0865, + "step": 38289 + }, + { + "epoch": 0.8975778992832738, + "grad_norm": 0.3554633557796478, + "learning_rate": 5.134101674384973e-06, + "loss": 0.0898, + "step": 38290 + }, + { + "epoch": 0.8976013408580789, + "grad_norm": 0.28109410405158997, + "learning_rate": 5.131772010873403e-06, + "loss": 0.0515, + "step": 38291 + }, + { + "epoch": 0.8976247824328838, + "grad_norm": 0.35585638880729675, + "learning_rate": 5.129442862119494e-06, + "loss": 0.0549, + "step": 38292 + }, + { + "epoch": 0.8976482240076888, + "grad_norm": 0.375067800283432, + "learning_rate": 5.127114228135899e-06, + "loss": 0.0243, + "step": 38293 + }, + { + "epoch": 0.8976716655824938, + "grad_norm": 0.39866527915000916, + "learning_rate": 5.124786108935254e-06, + "loss": 0.0346, + "step": 38294 + }, + { + "epoch": 0.8976951071572988, + "grad_norm": 0.1496557891368866, + "learning_rate": 5.122458504530159e-06, + "loss": 0.031, + "step": 38295 + }, + { + "epoch": 0.8977185487321038, + "grad_norm": 0.5804451704025269, + "learning_rate": 5.120131414933304e-06, + "loss": 0.1326, + "step": 38296 + }, + { + "epoch": 0.8977419903069088, + "grad_norm": 0.5750083327293396, + "learning_rate": 5.11780484015727e-06, + "loss": 0.0517, + "step": 38297 + }, + { + "epoch": 0.8977654318817138, + "grad_norm": 0.29892513155937195, + "learning_rate": 5.1154787802147e-06, + "loss": 0.0304, + "step": 38298 + }, + { + "epoch": 0.8977888734565188, + "grad_norm": 0.5204654335975647, + "learning_rate": 5.113153235118217e-06, + "loss": 0.0876, + "step": 38299 + }, + { + "epoch": 0.8978123150313237, + "grad_norm": 0.4034561514854431, + "learning_rate": 5.110828204880424e-06, + "loss": 0.0319, + "step": 38300 + }, + { + "epoch": 0.8978357566061288, + "grad_norm": 0.24063913524150848, + "learning_rate": 5.108503689513933e-06, + "loss": 0.0452, + "step": 38301 + }, + { + "epoch": 0.8978591981809338, + "grad_norm": 0.09557948261499405, + "learning_rate": 5.1061796890313655e-06, + "loss": 0.0171, + "step": 38302 + }, + { + "epoch": 0.8978826397557388, + "grad_norm": 0.5762243866920471, + "learning_rate": 5.103856203445356e-06, + "loss": 0.6782, + "step": 38303 + }, + { + "epoch": 0.8979060813305438, + "grad_norm": 0.16706642508506775, + "learning_rate": 5.101533232768485e-06, + "loss": 0.0234, + "step": 38304 + }, + { + "epoch": 0.8979295229053488, + "grad_norm": 0.46551254391670227, + "learning_rate": 5.0992107770133635e-06, + "loss": 0.6779, + "step": 38305 + }, + { + "epoch": 0.8979529644801538, + "grad_norm": 0.5133293271064758, + "learning_rate": 5.0968888361926034e-06, + "loss": 0.5669, + "step": 38306 + }, + { + "epoch": 0.8979764060549588, + "grad_norm": 0.33491843938827515, + "learning_rate": 5.094567410318785e-06, + "loss": 0.0511, + "step": 38307 + }, + { + "epoch": 0.8979998476297638, + "grad_norm": 0.2845783531665802, + "learning_rate": 5.092246499404496e-06, + "loss": 0.0404, + "step": 38308 + }, + { + "epoch": 0.8980232892045688, + "grad_norm": 0.31089359521865845, + "learning_rate": 5.089926103462361e-06, + "loss": 0.0503, + "step": 38309 + }, + { + "epoch": 0.8980467307793738, + "grad_norm": 0.2749592661857605, + "learning_rate": 5.087606222504959e-06, + "loss": 0.0479, + "step": 38310 + }, + { + "epoch": 0.8980701723541787, + "grad_norm": 0.4374435245990753, + "learning_rate": 5.085286856544868e-06, + "loss": 0.0619, + "step": 38311 + }, + { + "epoch": 0.8980936139289838, + "grad_norm": 0.4219180643558502, + "learning_rate": 5.082968005594679e-06, + "loss": 0.1034, + "step": 38312 + }, + { + "epoch": 0.8981170555037887, + "grad_norm": 0.42110925912857056, + "learning_rate": 5.0806496696669704e-06, + "loss": 0.0628, + "step": 38313 + }, + { + "epoch": 0.8981404970785938, + "grad_norm": 0.3671770989894867, + "learning_rate": 5.078331848774332e-06, + "loss": 0.0493, + "step": 38314 + }, + { + "epoch": 0.8981639386533987, + "grad_norm": 0.39692291617393494, + "learning_rate": 5.0760145429293194e-06, + "loss": 0.0712, + "step": 38315 + }, + { + "epoch": 0.8981873802282038, + "grad_norm": 0.36360639333724976, + "learning_rate": 5.073697752144535e-06, + "loss": 0.0588, + "step": 38316 + }, + { + "epoch": 0.8982108218030087, + "grad_norm": 0.2530742287635803, + "learning_rate": 5.071381476432546e-06, + "loss": 0.0363, + "step": 38317 + }, + { + "epoch": 0.8982342633778138, + "grad_norm": 0.36780208349227905, + "learning_rate": 5.069065715805898e-06, + "loss": 0.3522, + "step": 38318 + }, + { + "epoch": 0.8982577049526187, + "grad_norm": 0.6865718364715576, + "learning_rate": 5.066750470277171e-06, + "loss": 0.1108, + "step": 38319 + }, + { + "epoch": 0.8982811465274237, + "grad_norm": 0.43019434809684753, + "learning_rate": 5.064435739858931e-06, + "loss": 0.0599, + "step": 38320 + }, + { + "epoch": 0.8983045881022287, + "grad_norm": 0.11449752002954483, + "learning_rate": 5.062121524563712e-06, + "loss": 0.0085, + "step": 38321 + }, + { + "epoch": 0.8983280296770337, + "grad_norm": 0.43254411220550537, + "learning_rate": 5.059807824404106e-06, + "loss": 0.0829, + "step": 38322 + }, + { + "epoch": 0.8983514712518387, + "grad_norm": 0.7146916389465332, + "learning_rate": 5.057494639392657e-06, + "loss": 0.1596, + "step": 38323 + }, + { + "epoch": 0.8983749128266437, + "grad_norm": 0.6129162311553955, + "learning_rate": 5.055181969541911e-06, + "loss": 0.0943, + "step": 38324 + }, + { + "epoch": 0.8983983544014487, + "grad_norm": 0.561939001083374, + "learning_rate": 5.0528698148643915e-06, + "loss": 0.0652, + "step": 38325 + }, + { + "epoch": 0.8984217959762537, + "grad_norm": 0.4348638653755188, + "learning_rate": 5.0505581753726996e-06, + "loss": 0.0472, + "step": 38326 + }, + { + "epoch": 0.8984452375510587, + "grad_norm": 0.17641349136829376, + "learning_rate": 5.0482470510793355e-06, + "loss": 0.0243, + "step": 38327 + }, + { + "epoch": 0.8984686791258637, + "grad_norm": 0.6396041512489319, + "learning_rate": 5.045936441996846e-06, + "loss": 0.03, + "step": 38328 + }, + { + "epoch": 0.8984921207006686, + "grad_norm": 0.4833908975124359, + "learning_rate": 5.043626348137787e-06, + "loss": 0.1251, + "step": 38329 + }, + { + "epoch": 0.8985155622754737, + "grad_norm": 0.4023808538913727, + "learning_rate": 5.0413167695146924e-06, + "loss": 0.0344, + "step": 38330 + }, + { + "epoch": 0.8985390038502786, + "grad_norm": 0.15903623402118683, + "learning_rate": 5.039007706140064e-06, + "loss": 0.0259, + "step": 38331 + }, + { + "epoch": 0.8985624454250837, + "grad_norm": 0.11446358263492584, + "learning_rate": 5.036699158026459e-06, + "loss": 0.0207, + "step": 38332 + }, + { + "epoch": 0.8985858869998886, + "grad_norm": 0.5044768452644348, + "learning_rate": 5.034391125186388e-06, + "loss": 0.5163, + "step": 38333 + }, + { + "epoch": 0.8986093285746937, + "grad_norm": 0.10629582405090332, + "learning_rate": 5.032083607632387e-06, + "loss": 0.0171, + "step": 38334 + }, + { + "epoch": 0.8986327701494986, + "grad_norm": 0.17690907418727875, + "learning_rate": 5.029776605376968e-06, + "loss": 0.0275, + "step": 38335 + }, + { + "epoch": 0.8986562117243037, + "grad_norm": 1.257982611656189, + "learning_rate": 5.027470118432653e-06, + "loss": 1.4772, + "step": 38336 + }, + { + "epoch": 0.8986796532991086, + "grad_norm": 0.2350306212902069, + "learning_rate": 5.0251641468119445e-06, + "loss": 0.0322, + "step": 38337 + }, + { + "epoch": 0.8987030948739136, + "grad_norm": 0.2856980264186859, + "learning_rate": 5.022858690527377e-06, + "loss": 0.0456, + "step": 38338 + }, + { + "epoch": 0.8987265364487186, + "grad_norm": 0.12647025287151337, + "learning_rate": 5.02055374959145e-06, + "loss": 0.0234, + "step": 38339 + }, + { + "epoch": 0.8987499780235236, + "grad_norm": 0.3861047625541687, + "learning_rate": 5.018249324016655e-06, + "loss": 0.0177, + "step": 38340 + }, + { + "epoch": 0.8987734195983286, + "grad_norm": 0.4030616879463196, + "learning_rate": 5.015945413815526e-06, + "loss": 0.0324, + "step": 38341 + }, + { + "epoch": 0.8987968611731336, + "grad_norm": 0.17090889811515808, + "learning_rate": 5.013642019000553e-06, + "loss": 0.0239, + "step": 38342 + }, + { + "epoch": 0.8988203027479386, + "grad_norm": 0.4036068618297577, + "learning_rate": 5.011339139584214e-06, + "loss": 0.0805, + "step": 38343 + }, + { + "epoch": 0.8988437443227436, + "grad_norm": 0.345730185508728, + "learning_rate": 5.009036775579035e-06, + "loss": 0.0489, + "step": 38344 + }, + { + "epoch": 0.8988671858975485, + "grad_norm": 0.6061595678329468, + "learning_rate": 5.006734926997481e-06, + "loss": 0.0952, + "step": 38345 + }, + { + "epoch": 0.8988906274723536, + "grad_norm": 0.6667006611824036, + "learning_rate": 5.004433593852053e-06, + "loss": 0.1026, + "step": 38346 + }, + { + "epoch": 0.8989140690471585, + "grad_norm": 0.41592738032341003, + "learning_rate": 5.002132776155244e-06, + "loss": 0.0974, + "step": 38347 + }, + { + "epoch": 0.8989375106219636, + "grad_norm": 0.5047606825828552, + "learning_rate": 4.99983247391953e-06, + "loss": 0.0543, + "step": 38348 + }, + { + "epoch": 0.8989609521967685, + "grad_norm": 0.30969786643981934, + "learning_rate": 4.997532687157402e-06, + "loss": 0.0473, + "step": 38349 + }, + { + "epoch": 0.8989843937715736, + "grad_norm": 0.08151501417160034, + "learning_rate": 4.995233415881317e-06, + "loss": 0.0081, + "step": 38350 + }, + { + "epoch": 0.8990078353463785, + "grad_norm": 0.4786906838417053, + "learning_rate": 4.992934660103788e-06, + "loss": 0.0791, + "step": 38351 + }, + { + "epoch": 0.8990312769211836, + "grad_norm": 0.3132191300392151, + "learning_rate": 4.9906364198372575e-06, + "loss": 0.0354, + "step": 38352 + }, + { + "epoch": 0.8990547184959886, + "grad_norm": 0.5181288719177246, + "learning_rate": 4.9883386950941964e-06, + "loss": 0.6607, + "step": 38353 + }, + { + "epoch": 0.8990781600707936, + "grad_norm": 0.3373252749443054, + "learning_rate": 4.986041485887094e-06, + "loss": 0.0621, + "step": 38354 + }, + { + "epoch": 0.8991016016455986, + "grad_norm": 0.4582329988479614, + "learning_rate": 4.983744792228407e-06, + "loss": 0.0702, + "step": 38355 + }, + { + "epoch": 0.8991250432204035, + "grad_norm": 0.47852370142936707, + "learning_rate": 4.981448614130579e-06, + "loss": 0.0325, + "step": 38356 + }, + { + "epoch": 0.8991484847952086, + "grad_norm": 0.12344824522733688, + "learning_rate": 4.979152951606092e-06, + "loss": 0.0134, + "step": 38357 + }, + { + "epoch": 0.8991719263700135, + "grad_norm": 0.12432760745286942, + "learning_rate": 4.9768578046673895e-06, + "loss": 0.0308, + "step": 38358 + }, + { + "epoch": 0.8991953679448186, + "grad_norm": 0.7491111755371094, + "learning_rate": 4.974563173326918e-06, + "loss": 0.1171, + "step": 38359 + }, + { + "epoch": 0.8992188095196235, + "grad_norm": 0.36523255705833435, + "learning_rate": 4.972269057597156e-06, + "loss": 0.0697, + "step": 38360 + }, + { + "epoch": 0.8992422510944286, + "grad_norm": 0.5224372148513794, + "learning_rate": 4.969975457490528e-06, + "loss": 0.6365, + "step": 38361 + }, + { + "epoch": 0.8992656926692335, + "grad_norm": 0.38356778025627136, + "learning_rate": 4.967682373019467e-06, + "loss": 0.0487, + "step": 38362 + }, + { + "epoch": 0.8992891342440386, + "grad_norm": 0.4621927738189697, + "learning_rate": 4.965389804196452e-06, + "loss": 0.0635, + "step": 38363 + }, + { + "epoch": 0.8993125758188435, + "grad_norm": 0.18070879578590393, + "learning_rate": 4.9630977510339075e-06, + "loss": 0.0389, + "step": 38364 + }, + { + "epoch": 0.8993360173936485, + "grad_norm": 0.5490111708641052, + "learning_rate": 4.9608062135442665e-06, + "loss": 0.0923, + "step": 38365 + }, + { + "epoch": 0.8993594589684535, + "grad_norm": 0.3239811956882477, + "learning_rate": 4.958515191739943e-06, + "loss": 0.0269, + "step": 38366 + }, + { + "epoch": 0.8993829005432585, + "grad_norm": 0.10415419936180115, + "learning_rate": 4.956224685633415e-06, + "loss": 0.0212, + "step": 38367 + }, + { + "epoch": 0.8994063421180635, + "grad_norm": 0.3182961642742157, + "learning_rate": 4.953934695237072e-06, + "loss": 0.0405, + "step": 38368 + }, + { + "epoch": 0.8994297836928685, + "grad_norm": 0.17284803092479706, + "learning_rate": 4.95164522056335e-06, + "loss": 0.0178, + "step": 38369 + }, + { + "epoch": 0.8994532252676735, + "grad_norm": 0.8748748302459717, + "learning_rate": 4.949356261624683e-06, + "loss": 0.1512, + "step": 38370 + }, + { + "epoch": 0.8994766668424785, + "grad_norm": 0.39664241671562195, + "learning_rate": 4.947067818433459e-06, + "loss": 0.0419, + "step": 38371 + }, + { + "epoch": 0.8995001084172835, + "grad_norm": 0.2092640995979309, + "learning_rate": 4.944779891002138e-06, + "loss": 0.0253, + "step": 38372 + }, + { + "epoch": 0.8995235499920885, + "grad_norm": 0.4806910455226898, + "learning_rate": 4.9424924793431085e-06, + "loss": 0.0575, + "step": 38373 + }, + { + "epoch": 0.8995469915668934, + "grad_norm": 0.3825593590736389, + "learning_rate": 4.940205583468782e-06, + "loss": 0.0475, + "step": 38374 + }, + { + "epoch": 0.8995704331416985, + "grad_norm": 0.11333636939525604, + "learning_rate": 4.937919203391583e-06, + "loss": 0.0153, + "step": 38375 + }, + { + "epoch": 0.8995938747165034, + "grad_norm": 0.180143341422081, + "learning_rate": 4.935633339123902e-06, + "loss": 0.0386, + "step": 38376 + }, + { + "epoch": 0.8996173162913085, + "grad_norm": 0.12043248862028122, + "learning_rate": 4.93334799067815e-06, + "loss": 0.014, + "step": 38377 + }, + { + "epoch": 0.8996407578661134, + "grad_norm": 0.3246876895427704, + "learning_rate": 4.931063158066718e-06, + "loss": 0.0647, + "step": 38378 + }, + { + "epoch": 0.8996641994409185, + "grad_norm": 0.32043537497520447, + "learning_rate": 4.928778841302017e-06, + "loss": 0.0393, + "step": 38379 + }, + { + "epoch": 0.8996876410157234, + "grad_norm": 0.4502348303794861, + "learning_rate": 4.926495040396439e-06, + "loss": 0.0881, + "step": 38380 + }, + { + "epoch": 0.8997110825905285, + "grad_norm": 0.42853349447250366, + "learning_rate": 4.924211755362373e-06, + "loss": 0.0528, + "step": 38381 + }, + { + "epoch": 0.8997345241653334, + "grad_norm": 0.49062615633010864, + "learning_rate": 4.921928986212199e-06, + "loss": 0.0922, + "step": 38382 + }, + { + "epoch": 0.8997579657401384, + "grad_norm": 0.20331142842769623, + "learning_rate": 4.919646732958316e-06, + "loss": 0.0251, + "step": 38383 + }, + { + "epoch": 0.8997814073149434, + "grad_norm": 0.1765558272600174, + "learning_rate": 4.917364995613094e-06, + "loss": 0.0249, + "step": 38384 + }, + { + "epoch": 0.8998048488897484, + "grad_norm": 0.139740452170372, + "learning_rate": 4.915083774188934e-06, + "loss": 0.0162, + "step": 38385 + }, + { + "epoch": 0.8998282904645534, + "grad_norm": 0.3201640546321869, + "learning_rate": 4.912803068698191e-06, + "loss": 0.0564, + "step": 38386 + }, + { + "epoch": 0.8998517320393584, + "grad_norm": 0.7342008948326111, + "learning_rate": 4.9105228791532566e-06, + "loss": 0.103, + "step": 38387 + }, + { + "epoch": 0.8998751736141634, + "grad_norm": 0.41833794116973877, + "learning_rate": 4.90824320556651e-06, + "loss": 0.0779, + "step": 38388 + }, + { + "epoch": 0.8998986151889684, + "grad_norm": 0.11560893803834915, + "learning_rate": 4.905964047950306e-06, + "loss": 0.0136, + "step": 38389 + }, + { + "epoch": 0.8999220567637733, + "grad_norm": 0.07082528620958328, + "learning_rate": 4.903685406317016e-06, + "loss": 0.0066, + "step": 38390 + }, + { + "epoch": 0.8999454983385784, + "grad_norm": 0.43768370151519775, + "learning_rate": 4.901407280678982e-06, + "loss": 0.1362, + "step": 38391 + }, + { + "epoch": 0.8999689399133833, + "grad_norm": 0.5069825053215027, + "learning_rate": 4.899129671048608e-06, + "loss": 0.0926, + "step": 38392 + }, + { + "epoch": 0.8999923814881884, + "grad_norm": 0.2869303822517395, + "learning_rate": 4.896852577438227e-06, + "loss": 0.0339, + "step": 38393 + }, + { + "epoch": 0.9000158230629933, + "grad_norm": 0.3720737099647522, + "learning_rate": 4.894575999860196e-06, + "loss": 0.0655, + "step": 38394 + }, + { + "epoch": 0.9000392646377984, + "grad_norm": 0.5482960343360901, + "learning_rate": 4.892299938326872e-06, + "loss": 0.0898, + "step": 38395 + }, + { + "epoch": 0.9000627062126033, + "grad_norm": 0.30739614367485046, + "learning_rate": 4.890024392850601e-06, + "loss": 0.0646, + "step": 38396 + }, + { + "epoch": 0.9000861477874084, + "grad_norm": 0.48289647698402405, + "learning_rate": 4.887749363443728e-06, + "loss": 0.0875, + "step": 38397 + }, + { + "epoch": 0.9001095893622133, + "grad_norm": 0.3067740201950073, + "learning_rate": 4.885474850118587e-06, + "loss": 0.0383, + "step": 38398 + }, + { + "epoch": 0.9001330309370184, + "grad_norm": 0.1326090395450592, + "learning_rate": 4.8832008528875595e-06, + "loss": 0.0072, + "step": 38399 + }, + { + "epoch": 0.9001564725118233, + "grad_norm": 0.327455997467041, + "learning_rate": 4.880927371762955e-06, + "loss": 0.3203, + "step": 38400 + }, + { + "epoch": 0.9001799140866283, + "grad_norm": 0.43155843019485474, + "learning_rate": 4.878654406757111e-06, + "loss": 0.1057, + "step": 38401 + }, + { + "epoch": 0.9002033556614333, + "grad_norm": 0.48190838098526, + "learning_rate": 4.8763819578823586e-06, + "loss": 0.0655, + "step": 38402 + }, + { + "epoch": 0.9002267972362383, + "grad_norm": 0.5170167684555054, + "learning_rate": 4.874110025151035e-06, + "loss": 0.1252, + "step": 38403 + }, + { + "epoch": 0.9002502388110434, + "grad_norm": 0.4768918752670288, + "learning_rate": 4.871838608575452e-06, + "loss": 0.0879, + "step": 38404 + }, + { + "epoch": 0.9002736803858483, + "grad_norm": 0.20476478338241577, + "learning_rate": 4.869567708167966e-06, + "loss": 0.0229, + "step": 38405 + }, + { + "epoch": 0.9002971219606534, + "grad_norm": 0.2760773301124573, + "learning_rate": 4.867297323940878e-06, + "loss": 0.0341, + "step": 38406 + }, + { + "epoch": 0.9003205635354583, + "grad_norm": 0.13393215835094452, + "learning_rate": 4.865027455906502e-06, + "loss": 0.0199, + "step": 38407 + }, + { + "epoch": 0.9003440051102634, + "grad_norm": 0.3276747167110443, + "learning_rate": 4.86275810407717e-06, + "loss": 0.0484, + "step": 38408 + }, + { + "epoch": 0.9003674466850683, + "grad_norm": 0.09384670108556747, + "learning_rate": 4.8604892684651735e-06, + "loss": 0.0143, + "step": 38409 + }, + { + "epoch": 0.9003908882598733, + "grad_norm": 0.22425223886966705, + "learning_rate": 4.8582209490828474e-06, + "loss": 0.0371, + "step": 38410 + }, + { + "epoch": 0.9004143298346783, + "grad_norm": 0.4821225106716156, + "learning_rate": 4.855953145942482e-06, + "loss": 0.1015, + "step": 38411 + }, + { + "epoch": 0.9004377714094833, + "grad_norm": 0.2277127504348755, + "learning_rate": 4.853685859056401e-06, + "loss": 0.0291, + "step": 38412 + }, + { + "epoch": 0.9004612129842883, + "grad_norm": 0.3146370053291321, + "learning_rate": 4.8514190884368924e-06, + "loss": 0.0316, + "step": 38413 + }, + { + "epoch": 0.9004846545590933, + "grad_norm": 0.10762358456850052, + "learning_rate": 4.84915283409626e-06, + "loss": 0.0174, + "step": 38414 + }, + { + "epoch": 0.9005080961338983, + "grad_norm": 0.40343570709228516, + "learning_rate": 4.8468870960468036e-06, + "loss": 0.0457, + "step": 38415 + }, + { + "epoch": 0.9005315377087033, + "grad_norm": 0.391704797744751, + "learning_rate": 4.844621874300792e-06, + "loss": 0.4526, + "step": 38416 + }, + { + "epoch": 0.9005549792835083, + "grad_norm": 0.3339865207672119, + "learning_rate": 4.842357168870559e-06, + "loss": 0.047, + "step": 38417 + }, + { + "epoch": 0.9005784208583133, + "grad_norm": 0.08786240965127945, + "learning_rate": 4.840092979768374e-06, + "loss": 0.0118, + "step": 38418 + }, + { + "epoch": 0.9006018624331182, + "grad_norm": 0.6139530539512634, + "learning_rate": 4.837829307006514e-06, + "loss": 0.1309, + "step": 38419 + }, + { + "epoch": 0.9006253040079233, + "grad_norm": 0.45741090178489685, + "learning_rate": 4.83556615059727e-06, + "loss": 0.0712, + "step": 38420 + }, + { + "epoch": 0.9006487455827282, + "grad_norm": 0.37448155879974365, + "learning_rate": 4.833303510552911e-06, + "loss": 0.0464, + "step": 38421 + }, + { + "epoch": 0.9006721871575333, + "grad_norm": 0.2885880172252655, + "learning_rate": 4.831041386885726e-06, + "loss": 0.0273, + "step": 38422 + }, + { + "epoch": 0.9006956287323382, + "grad_norm": 0.2793804109096527, + "learning_rate": 4.828779779607972e-06, + "loss": 0.0262, + "step": 38423 + }, + { + "epoch": 0.9007190703071433, + "grad_norm": 0.4635496139526367, + "learning_rate": 4.826518688731951e-06, + "loss": 0.0931, + "step": 38424 + }, + { + "epoch": 0.9007425118819482, + "grad_norm": 0.4631579518318176, + "learning_rate": 4.82425811426992e-06, + "loss": 0.0743, + "step": 38425 + }, + { + "epoch": 0.9007659534567533, + "grad_norm": 0.32488560676574707, + "learning_rate": 4.821998056234145e-06, + "loss": 0.0531, + "step": 38426 + }, + { + "epoch": 0.9007893950315582, + "grad_norm": 0.5401944518089294, + "learning_rate": 4.8197385146368734e-06, + "loss": 0.0948, + "step": 38427 + }, + { + "epoch": 0.9008128366063632, + "grad_norm": 0.7332643866539001, + "learning_rate": 4.817479489490384e-06, + "loss": 0.1338, + "step": 38428 + }, + { + "epoch": 0.9008362781811682, + "grad_norm": 0.35045358538627625, + "learning_rate": 4.815220980806912e-06, + "loss": 0.0338, + "step": 38429 + }, + { + "epoch": 0.9008597197559732, + "grad_norm": 0.45125338435173035, + "learning_rate": 4.812962988598746e-06, + "loss": 0.0875, + "step": 38430 + }, + { + "epoch": 0.9008831613307782, + "grad_norm": 0.12264131009578705, + "learning_rate": 4.810705512878111e-06, + "loss": 0.0248, + "step": 38431 + }, + { + "epoch": 0.9009066029055832, + "grad_norm": 0.48288193345069885, + "learning_rate": 4.8084485536572745e-06, + "loss": 0.1047, + "step": 38432 + }, + { + "epoch": 0.9009300444803882, + "grad_norm": 0.33356741070747375, + "learning_rate": 4.8061921109484594e-06, + "loss": 0.0281, + "step": 38433 + }, + { + "epoch": 0.9009534860551932, + "grad_norm": 0.3103336691856384, + "learning_rate": 4.8039361847639355e-06, + "loss": 0.0493, + "step": 38434 + }, + { + "epoch": 0.9009769276299981, + "grad_norm": 0.2980944514274597, + "learning_rate": 4.8016807751159024e-06, + "loss": 0.0691, + "step": 38435 + }, + { + "epoch": 0.9010003692048032, + "grad_norm": 0.42774415016174316, + "learning_rate": 4.799425882016639e-06, + "loss": 0.0483, + "step": 38436 + }, + { + "epoch": 0.9010238107796081, + "grad_norm": 0.622599720954895, + "learning_rate": 4.79717150547837e-06, + "loss": 0.6182, + "step": 38437 + }, + { + "epoch": 0.9010472523544132, + "grad_norm": 0.4841398298740387, + "learning_rate": 4.794917645513342e-06, + "loss": 0.0929, + "step": 38438 + }, + { + "epoch": 0.9010706939292181, + "grad_norm": 0.4485158920288086, + "learning_rate": 4.7926643021337535e-06, + "loss": 0.0552, + "step": 38439 + }, + { + "epoch": 0.9010941355040232, + "grad_norm": 0.26357993483543396, + "learning_rate": 4.7904114753518525e-06, + "loss": 0.0491, + "step": 38440 + }, + { + "epoch": 0.9011175770788281, + "grad_norm": 0.13671983778476715, + "learning_rate": 4.788159165179851e-06, + "loss": 0.0137, + "step": 38441 + }, + { + "epoch": 0.9011410186536332, + "grad_norm": 0.15020380914211273, + "learning_rate": 4.78590737162996e-06, + "loss": 0.0205, + "step": 38442 + }, + { + "epoch": 0.9011644602284381, + "grad_norm": 0.507911741733551, + "learning_rate": 4.783656094714428e-06, + "loss": 0.1079, + "step": 38443 + }, + { + "epoch": 0.9011879018032432, + "grad_norm": 0.08327357470989227, + "learning_rate": 4.781405334445454e-06, + "loss": 0.0103, + "step": 38444 + }, + { + "epoch": 0.9012113433780481, + "grad_norm": 0.4884870648384094, + "learning_rate": 4.779155090835264e-06, + "loss": 0.0783, + "step": 38445 + }, + { + "epoch": 0.9012347849528531, + "grad_norm": 0.0806087851524353, + "learning_rate": 4.776905363896045e-06, + "loss": 0.0131, + "step": 38446 + }, + { + "epoch": 0.9012582265276581, + "grad_norm": 0.32803216576576233, + "learning_rate": 4.77465615364e-06, + "loss": 0.0597, + "step": 38447 + }, + { + "epoch": 0.9012816681024631, + "grad_norm": 0.4424560070037842, + "learning_rate": 4.772407460079353e-06, + "loss": 0.0585, + "step": 38448 + }, + { + "epoch": 0.9013051096772681, + "grad_norm": 0.6972913146018982, + "learning_rate": 4.7701592832263165e-06, + "loss": 0.1573, + "step": 38449 + }, + { + "epoch": 0.9013285512520731, + "grad_norm": 0.20856420695781708, + "learning_rate": 4.767911623093069e-06, + "loss": 0.0211, + "step": 38450 + }, + { + "epoch": 0.9013519928268781, + "grad_norm": 0.4698921740055084, + "learning_rate": 4.765664479691811e-06, + "loss": 0.6068, + "step": 38451 + }, + { + "epoch": 0.9013754344016831, + "grad_norm": 0.26566553115844727, + "learning_rate": 4.763417853034735e-06, + "loss": 0.0691, + "step": 38452 + }, + { + "epoch": 0.901398875976488, + "grad_norm": 0.3713912069797516, + "learning_rate": 4.761171743134029e-06, + "loss": 0.0351, + "step": 38453 + }, + { + "epoch": 0.9014223175512931, + "grad_norm": 0.5727226138114929, + "learning_rate": 4.758926150001874e-06, + "loss": 0.1019, + "step": 38454 + }, + { + "epoch": 0.9014457591260981, + "grad_norm": 0.3728539049625397, + "learning_rate": 4.7566810736504705e-06, + "loss": 0.0603, + "step": 38455 + }, + { + "epoch": 0.9014692007009031, + "grad_norm": 0.573147714138031, + "learning_rate": 4.754436514091998e-06, + "loss": 0.1226, + "step": 38456 + }, + { + "epoch": 0.9014926422757081, + "grad_norm": 0.23347152769565582, + "learning_rate": 4.752192471338635e-06, + "loss": 0.0364, + "step": 38457 + }, + { + "epoch": 0.9015160838505131, + "grad_norm": 0.1885523498058319, + "learning_rate": 4.749948945402549e-06, + "loss": 0.0205, + "step": 38458 + }, + { + "epoch": 0.9015395254253181, + "grad_norm": 0.4892050623893738, + "learning_rate": 4.74770593629591e-06, + "loss": 0.0952, + "step": 38459 + }, + { + "epoch": 0.9015629670001231, + "grad_norm": 0.7401622533798218, + "learning_rate": 4.745463444030906e-06, + "loss": 0.1497, + "step": 38460 + }, + { + "epoch": 0.9015864085749281, + "grad_norm": 0.1937878578901291, + "learning_rate": 4.743221468619685e-06, + "loss": 0.039, + "step": 38461 + }, + { + "epoch": 0.901609850149733, + "grad_norm": 0.7892956733703613, + "learning_rate": 4.740980010074425e-06, + "loss": 0.1181, + "step": 38462 + }, + { + "epoch": 0.9016332917245381, + "grad_norm": 0.668076753616333, + "learning_rate": 4.738739068407305e-06, + "loss": 0.2185, + "step": 38463 + }, + { + "epoch": 0.901656733299343, + "grad_norm": 0.27522775530815125, + "learning_rate": 4.736498643630449e-06, + "loss": 0.0279, + "step": 38464 + }, + { + "epoch": 0.9016801748741481, + "grad_norm": 0.5797110199928284, + "learning_rate": 4.734258735756037e-06, + "loss": 0.0639, + "step": 38465 + }, + { + "epoch": 0.901703616448953, + "grad_norm": 0.45037388801574707, + "learning_rate": 4.732019344796212e-06, + "loss": 0.5201, + "step": 38466 + }, + { + "epoch": 0.9017270580237581, + "grad_norm": 0.44872042536735535, + "learning_rate": 4.729780470763123e-06, + "loss": 0.1132, + "step": 38467 + }, + { + "epoch": 0.901750499598563, + "grad_norm": 0.6104639172554016, + "learning_rate": 4.7275421136689255e-06, + "loss": 0.0811, + "step": 38468 + }, + { + "epoch": 0.9017739411733681, + "grad_norm": 0.18482156097888947, + "learning_rate": 4.725304273525777e-06, + "loss": 0.0245, + "step": 38469 + }, + { + "epoch": 0.901797382748173, + "grad_norm": 0.43887677788734436, + "learning_rate": 4.723066950345801e-06, + "loss": 0.0444, + "step": 38470 + }, + { + "epoch": 0.901820824322978, + "grad_norm": 0.42510727047920227, + "learning_rate": 4.72083014414112e-06, + "loss": 0.0566, + "step": 38471 + }, + { + "epoch": 0.901844265897783, + "grad_norm": 0.08073867857456207, + "learning_rate": 4.718593854923925e-06, + "loss": 0.0106, + "step": 38472 + }, + { + "epoch": 0.901867707472588, + "grad_norm": 0.42917904257774353, + "learning_rate": 4.716358082706307e-06, + "loss": 0.0515, + "step": 38473 + }, + { + "epoch": 0.901891149047393, + "grad_norm": 0.19196335971355438, + "learning_rate": 4.714122827500389e-06, + "loss": 0.0445, + "step": 38474 + }, + { + "epoch": 0.901914590622198, + "grad_norm": 0.4145778715610504, + "learning_rate": 4.711888089318351e-06, + "loss": 0.0312, + "step": 38475 + }, + { + "epoch": 0.901938032197003, + "grad_norm": 0.2443750947713852, + "learning_rate": 4.70965386817227e-06, + "loss": 0.0248, + "step": 38476 + }, + { + "epoch": 0.901961473771808, + "grad_norm": 0.2418278604745865, + "learning_rate": 4.707420164074294e-06, + "loss": 0.0511, + "step": 38477 + }, + { + "epoch": 0.901984915346613, + "grad_norm": 0.5939562320709229, + "learning_rate": 4.705186977036546e-06, + "loss": 0.0367, + "step": 38478 + }, + { + "epoch": 0.902008356921418, + "grad_norm": 0.40353652834892273, + "learning_rate": 4.702954307071117e-06, + "loss": 0.0466, + "step": 38479 + }, + { + "epoch": 0.902031798496223, + "grad_norm": 0.1591213345527649, + "learning_rate": 4.7007221541901295e-06, + "loss": 0.0217, + "step": 38480 + }, + { + "epoch": 0.902055240071028, + "grad_norm": 0.330392062664032, + "learning_rate": 4.698490518405718e-06, + "loss": 0.0466, + "step": 38481 + }, + { + "epoch": 0.9020786816458329, + "grad_norm": 0.5764310359954834, + "learning_rate": 4.696259399729974e-06, + "loss": 0.1068, + "step": 38482 + }, + { + "epoch": 0.902102123220638, + "grad_norm": 0.48337215185165405, + "learning_rate": 4.694028798174999e-06, + "loss": 0.1422, + "step": 38483 + }, + { + "epoch": 0.9021255647954429, + "grad_norm": 0.8297198414802551, + "learning_rate": 4.691798713752915e-06, + "loss": 0.1633, + "step": 38484 + }, + { + "epoch": 0.902149006370248, + "grad_norm": 0.3759828507900238, + "learning_rate": 4.689569146475803e-06, + "loss": 0.0618, + "step": 38485 + }, + { + "epoch": 0.9021724479450529, + "grad_norm": 0.6711859107017517, + "learning_rate": 4.687340096355764e-06, + "loss": 0.1429, + "step": 38486 + }, + { + "epoch": 0.902195889519858, + "grad_norm": 0.402344673871994, + "learning_rate": 4.685111563404909e-06, + "loss": 0.2866, + "step": 38487 + }, + { + "epoch": 0.9022193310946629, + "grad_norm": 0.7911205291748047, + "learning_rate": 4.68288354763532e-06, + "loss": 0.1456, + "step": 38488 + }, + { + "epoch": 0.902242772669468, + "grad_norm": 0.17107227444648743, + "learning_rate": 4.680656049059073e-06, + "loss": 0.0257, + "step": 38489 + }, + { + "epoch": 0.9022662142442729, + "grad_norm": 0.25462254881858826, + "learning_rate": 4.678429067688283e-06, + "loss": 0.0395, + "step": 38490 + }, + { + "epoch": 0.9022896558190779, + "grad_norm": 0.5590956807136536, + "learning_rate": 4.676202603535007e-06, + "loss": 0.0757, + "step": 38491 + }, + { + "epoch": 0.9023130973938829, + "grad_norm": 0.14239907264709473, + "learning_rate": 4.673976656611323e-06, + "loss": 0.0168, + "step": 38492 + }, + { + "epoch": 0.9023365389686879, + "grad_norm": 0.07431141287088394, + "learning_rate": 4.671751226929344e-06, + "loss": 0.0091, + "step": 38493 + }, + { + "epoch": 0.9023599805434929, + "grad_norm": 1.0571101903915405, + "learning_rate": 4.669526314501116e-06, + "loss": 0.1058, + "step": 38494 + }, + { + "epoch": 0.9023834221182979, + "grad_norm": 0.44995203614234924, + "learning_rate": 4.667301919338718e-06, + "loss": 0.0476, + "step": 38495 + }, + { + "epoch": 0.9024068636931029, + "grad_norm": 0.4214942455291748, + "learning_rate": 4.665078041454207e-06, + "loss": 0.0749, + "step": 38496 + }, + { + "epoch": 0.9024303052679079, + "grad_norm": 0.39416053891181946, + "learning_rate": 4.662854680859685e-06, + "loss": 0.0528, + "step": 38497 + }, + { + "epoch": 0.9024537468427128, + "grad_norm": 0.4812600612640381, + "learning_rate": 4.660631837567186e-06, + "loss": 0.1177, + "step": 38498 + }, + { + "epoch": 0.9024771884175179, + "grad_norm": 0.4045383930206299, + "learning_rate": 4.658409511588768e-06, + "loss": 0.0626, + "step": 38499 + }, + { + "epoch": 0.9025006299923228, + "grad_norm": 0.507561206817627, + "learning_rate": 4.656187702936509e-06, + "loss": 0.0663, + "step": 38500 + }, + { + "epoch": 0.9025240715671279, + "grad_norm": 0.4292411804199219, + "learning_rate": 4.653966411622457e-06, + "loss": 0.4829, + "step": 38501 + }, + { + "epoch": 0.9025475131419328, + "grad_norm": 0.24340829253196716, + "learning_rate": 4.6517456376586775e-06, + "loss": 0.0298, + "step": 38502 + }, + { + "epoch": 0.9025709547167379, + "grad_norm": 0.14467233419418335, + "learning_rate": 4.649525381057196e-06, + "loss": 0.0354, + "step": 38503 + }, + { + "epoch": 0.9025943962915428, + "grad_norm": 0.6763561964035034, + "learning_rate": 4.6473056418300686e-06, + "loss": 0.0861, + "step": 38504 + }, + { + "epoch": 0.9026178378663479, + "grad_norm": 0.2740972638130188, + "learning_rate": 4.645086419989331e-06, + "loss": 0.0144, + "step": 38505 + }, + { + "epoch": 0.9026412794411529, + "grad_norm": 0.16828173398971558, + "learning_rate": 4.64286771554705e-06, + "loss": 0.0356, + "step": 38506 + }, + { + "epoch": 0.9026647210159578, + "grad_norm": 0.332216739654541, + "learning_rate": 4.64064952851525e-06, + "loss": 0.4093, + "step": 38507 + }, + { + "epoch": 0.9026881625907629, + "grad_norm": 0.18898777663707733, + "learning_rate": 4.6384318589059565e-06, + "loss": 0.0282, + "step": 38508 + }, + { + "epoch": 0.9027116041655678, + "grad_norm": 0.37886178493499756, + "learning_rate": 4.636214706731224e-06, + "loss": 0.0652, + "step": 38509 + }, + { + "epoch": 0.9027350457403729, + "grad_norm": 0.4554482400417328, + "learning_rate": 4.633998072003076e-06, + "loss": 0.0897, + "step": 38510 + }, + { + "epoch": 0.9027584873151778, + "grad_norm": 0.35286664962768555, + "learning_rate": 4.631781954733527e-06, + "loss": 0.0836, + "step": 38511 + }, + { + "epoch": 0.9027819288899829, + "grad_norm": 0.3621836304664612, + "learning_rate": 4.6295663549346e-06, + "loss": 0.0759, + "step": 38512 + }, + { + "epoch": 0.9028053704647878, + "grad_norm": 0.2927343249320984, + "learning_rate": 4.627351272618341e-06, + "loss": 0.079, + "step": 38513 + }, + { + "epoch": 0.9028288120395929, + "grad_norm": 0.34120744466781616, + "learning_rate": 4.625136707796763e-06, + "loss": 0.0832, + "step": 38514 + }, + { + "epoch": 0.9028522536143978, + "grad_norm": 0.3944378197193146, + "learning_rate": 4.622922660481865e-06, + "loss": 0.0401, + "step": 38515 + }, + { + "epoch": 0.9028756951892029, + "grad_norm": 0.5249764323234558, + "learning_rate": 4.620709130685674e-06, + "loss": 0.1152, + "step": 38516 + }, + { + "epoch": 0.9028991367640078, + "grad_norm": 0.16877324879169464, + "learning_rate": 4.618496118420202e-06, + "loss": 0.0161, + "step": 38517 + }, + { + "epoch": 0.9029225783388128, + "grad_norm": 0.03702666983008385, + "learning_rate": 4.616283623697437e-06, + "loss": 0.0034, + "step": 38518 + }, + { + "epoch": 0.9029460199136178, + "grad_norm": 0.2839142978191376, + "learning_rate": 4.614071646529405e-06, + "loss": 0.0422, + "step": 38519 + }, + { + "epoch": 0.9029694614884228, + "grad_norm": 0.1570798009634018, + "learning_rate": 4.611860186928097e-06, + "loss": 0.0353, + "step": 38520 + }, + { + "epoch": 0.9029929030632278, + "grad_norm": 0.26713013648986816, + "learning_rate": 4.6096492449055225e-06, + "loss": 0.0239, + "step": 38521 + }, + { + "epoch": 0.9030163446380328, + "grad_norm": 0.6264806389808655, + "learning_rate": 4.607438820473686e-06, + "loss": 0.0994, + "step": 38522 + }, + { + "epoch": 0.9030397862128378, + "grad_norm": 0.44851046800613403, + "learning_rate": 4.6052289136445545e-06, + "loss": 0.5645, + "step": 38523 + }, + { + "epoch": 0.9030632277876428, + "grad_norm": 0.45485809445381165, + "learning_rate": 4.60301952443013e-06, + "loss": 0.0879, + "step": 38524 + }, + { + "epoch": 0.9030866693624477, + "grad_norm": 0.48009592294692993, + "learning_rate": 4.600810652842413e-06, + "loss": 0.0662, + "step": 38525 + }, + { + "epoch": 0.9031101109372528, + "grad_norm": 0.3942950963973999, + "learning_rate": 4.598602298893384e-06, + "loss": 0.0811, + "step": 38526 + }, + { + "epoch": 0.9031335525120577, + "grad_norm": 1.044960856437683, + "learning_rate": 4.59639446259501e-06, + "loss": 0.1705, + "step": 38527 + }, + { + "epoch": 0.9031569940868628, + "grad_norm": 0.3742747902870178, + "learning_rate": 4.594187143959294e-06, + "loss": 0.0608, + "step": 38528 + }, + { + "epoch": 0.9031804356616677, + "grad_norm": 0.24593432247638702, + "learning_rate": 4.591980342998203e-06, + "loss": 0.0232, + "step": 38529 + }, + { + "epoch": 0.9032038772364728, + "grad_norm": 0.14845451712608337, + "learning_rate": 4.5897740597236835e-06, + "loss": 0.0304, + "step": 38530 + }, + { + "epoch": 0.9032273188112777, + "grad_norm": 0.0627160370349884, + "learning_rate": 4.587568294147759e-06, + "loss": 0.0061, + "step": 38531 + }, + { + "epoch": 0.9032507603860828, + "grad_norm": 0.32865995168685913, + "learning_rate": 4.5853630462823425e-06, + "loss": 0.0442, + "step": 38532 + }, + { + "epoch": 0.9032742019608877, + "grad_norm": 0.17929887771606445, + "learning_rate": 4.583158316139458e-06, + "loss": 0.0248, + "step": 38533 + }, + { + "epoch": 0.9032976435356928, + "grad_norm": 0.3755156695842743, + "learning_rate": 4.5809541037310275e-06, + "loss": 0.041, + "step": 38534 + }, + { + "epoch": 0.9033210851104977, + "grad_norm": 0.3769298493862152, + "learning_rate": 4.5787504090690324e-06, + "loss": 0.0498, + "step": 38535 + }, + { + "epoch": 0.9033445266853027, + "grad_norm": 0.288194477558136, + "learning_rate": 4.576547232165418e-06, + "loss": 0.0661, + "step": 38536 + }, + { + "epoch": 0.9033679682601077, + "grad_norm": 0.0798935517668724, + "learning_rate": 4.57434457303213e-06, + "loss": 0.0084, + "step": 38537 + }, + { + "epoch": 0.9033914098349127, + "grad_norm": 0.82259202003479, + "learning_rate": 4.5721424316811366e-06, + "loss": 0.0992, + "step": 38538 + }, + { + "epoch": 0.9034148514097177, + "grad_norm": 0.25854218006134033, + "learning_rate": 4.569940808124384e-06, + "loss": 0.0611, + "step": 38539 + }, + { + "epoch": 0.9034382929845227, + "grad_norm": 0.3871343731880188, + "learning_rate": 4.567739702373819e-06, + "loss": 0.0513, + "step": 38540 + }, + { + "epoch": 0.9034617345593277, + "grad_norm": 0.1604612022638321, + "learning_rate": 4.565539114441386e-06, + "loss": 0.0147, + "step": 38541 + }, + { + "epoch": 0.9034851761341327, + "grad_norm": 0.16309724748134613, + "learning_rate": 4.563339044339021e-06, + "loss": 0.0262, + "step": 38542 + }, + { + "epoch": 0.9035086177089376, + "grad_norm": 0.5602695941925049, + "learning_rate": 4.561139492078637e-06, + "loss": 0.0939, + "step": 38543 + }, + { + "epoch": 0.9035320592837427, + "grad_norm": 0.7063530683517456, + "learning_rate": 4.5589404576722005e-06, + "loss": 0.1024, + "step": 38544 + }, + { + "epoch": 0.9035555008585476, + "grad_norm": 0.15486641228199005, + "learning_rate": 4.556741941131648e-06, + "loss": 0.017, + "step": 38545 + }, + { + "epoch": 0.9035789424333527, + "grad_norm": 0.8768227696418762, + "learning_rate": 4.554543942468903e-06, + "loss": 0.1022, + "step": 38546 + }, + { + "epoch": 0.9036023840081576, + "grad_norm": 0.14566335082054138, + "learning_rate": 4.552346461695878e-06, + "loss": 0.0162, + "step": 38547 + }, + { + "epoch": 0.9036258255829627, + "grad_norm": 0.3422774374485016, + "learning_rate": 4.5501494988245184e-06, + "loss": 0.0318, + "step": 38548 + }, + { + "epoch": 0.9036492671577676, + "grad_norm": 0.4360879957675934, + "learning_rate": 4.547953053866716e-06, + "loss": 0.0522, + "step": 38549 + }, + { + "epoch": 0.9036727087325727, + "grad_norm": 0.5354763269424438, + "learning_rate": 4.545757126834405e-06, + "loss": 0.0947, + "step": 38550 + }, + { + "epoch": 0.9036961503073776, + "grad_norm": 0.14416080713272095, + "learning_rate": 4.543561717739497e-06, + "loss": 0.0298, + "step": 38551 + }, + { + "epoch": 0.9037195918821826, + "grad_norm": 0.6190761923789978, + "learning_rate": 4.541366826593918e-06, + "loss": 0.0741, + "step": 38552 + }, + { + "epoch": 0.9037430334569876, + "grad_norm": 0.08529195934534073, + "learning_rate": 4.539172453409568e-06, + "loss": 0.0176, + "step": 38553 + }, + { + "epoch": 0.9037664750317926, + "grad_norm": 0.3391062617301941, + "learning_rate": 4.536978598198349e-06, + "loss": 0.0289, + "step": 38554 + }, + { + "epoch": 0.9037899166065976, + "grad_norm": 0.08646838366985321, + "learning_rate": 4.534785260972152e-06, + "loss": 0.0075, + "step": 38555 + }, + { + "epoch": 0.9038133581814026, + "grad_norm": 0.3449406623840332, + "learning_rate": 4.532592441742911e-06, + "loss": 0.0378, + "step": 38556 + }, + { + "epoch": 0.9038367997562077, + "grad_norm": 0.3172959089279175, + "learning_rate": 4.530400140522495e-06, + "loss": 0.0248, + "step": 38557 + }, + { + "epoch": 0.9038602413310126, + "grad_norm": 0.5409082174301147, + "learning_rate": 4.528208357322816e-06, + "loss": 0.1381, + "step": 38558 + }, + { + "epoch": 0.9038836829058177, + "grad_norm": 0.1864226758480072, + "learning_rate": 4.526017092155776e-06, + "loss": 0.0303, + "step": 38559 + }, + { + "epoch": 0.9039071244806226, + "grad_norm": 0.5558201670646667, + "learning_rate": 4.523826345033244e-06, + "loss": 0.1182, + "step": 38560 + }, + { + "epoch": 0.9039305660554277, + "grad_norm": 0.4636998176574707, + "learning_rate": 4.521636115967121e-06, + "loss": 0.0841, + "step": 38561 + }, + { + "epoch": 0.9039540076302326, + "grad_norm": 0.3689987361431122, + "learning_rate": 4.519446404969263e-06, + "loss": 0.0384, + "step": 38562 + }, + { + "epoch": 0.9039774492050376, + "grad_norm": 0.32529255747795105, + "learning_rate": 4.5172572120515955e-06, + "loss": 0.0697, + "step": 38563 + }, + { + "epoch": 0.9040008907798426, + "grad_norm": 0.34510862827301025, + "learning_rate": 4.515068537225975e-06, + "loss": 0.0295, + "step": 38564 + }, + { + "epoch": 0.9040243323546476, + "grad_norm": 0.32947978377342224, + "learning_rate": 4.51288038050427e-06, + "loss": 0.0593, + "step": 38565 + }, + { + "epoch": 0.9040477739294526, + "grad_norm": 0.320588618516922, + "learning_rate": 4.5106927418983704e-06, + "loss": 0.0466, + "step": 38566 + }, + { + "epoch": 0.9040712155042576, + "grad_norm": 0.6501911282539368, + "learning_rate": 4.5085056214201335e-06, + "loss": 0.0911, + "step": 38567 + }, + { + "epoch": 0.9040946570790626, + "grad_norm": 0.7202446460723877, + "learning_rate": 4.506319019081407e-06, + "loss": 0.6416, + "step": 38568 + }, + { + "epoch": 0.9041180986538676, + "grad_norm": 0.41176170110702515, + "learning_rate": 4.50413293489409e-06, + "loss": 0.0472, + "step": 38569 + }, + { + "epoch": 0.9041415402286725, + "grad_norm": 0.4570220410823822, + "learning_rate": 4.501947368870041e-06, + "loss": 0.1024, + "step": 38570 + }, + { + "epoch": 0.9041649818034776, + "grad_norm": 0.12058086693286896, + "learning_rate": 4.499762321021117e-06, + "loss": 0.019, + "step": 38571 + }, + { + "epoch": 0.9041884233782825, + "grad_norm": 0.1464802324771881, + "learning_rate": 4.497577791359164e-06, + "loss": 0.0221, + "step": 38572 + }, + { + "epoch": 0.9042118649530876, + "grad_norm": 0.4408285915851593, + "learning_rate": 4.49539377989604e-06, + "loss": 0.0673, + "step": 38573 + }, + { + "epoch": 0.9042353065278925, + "grad_norm": 0.21550972759723663, + "learning_rate": 4.49321028664359e-06, + "loss": 0.0381, + "step": 38574 + }, + { + "epoch": 0.9042587481026976, + "grad_norm": 0.16531409323215485, + "learning_rate": 4.4910273116136495e-06, + "loss": 0.0268, + "step": 38575 + }, + { + "epoch": 0.9042821896775025, + "grad_norm": 0.07850316166877747, + "learning_rate": 4.488844854818108e-06, + "loss": 0.01, + "step": 38576 + }, + { + "epoch": 0.9043056312523076, + "grad_norm": 0.5883169770240784, + "learning_rate": 4.486662916268769e-06, + "loss": 0.4969, + "step": 38577 + }, + { + "epoch": 0.9043290728271125, + "grad_norm": 0.45794761180877686, + "learning_rate": 4.484481495977488e-06, + "loss": 0.0798, + "step": 38578 + }, + { + "epoch": 0.9043525144019176, + "grad_norm": 0.23851455748081207, + "learning_rate": 4.48230059395609e-06, + "loss": 0.0364, + "step": 38579 + }, + { + "epoch": 0.9043759559767225, + "grad_norm": 0.7820460796356201, + "learning_rate": 4.480120210216421e-06, + "loss": 0.1353, + "step": 38580 + }, + { + "epoch": 0.9043993975515275, + "grad_norm": 0.5725988745689392, + "learning_rate": 4.477940344770293e-06, + "loss": 0.1239, + "step": 38581 + }, + { + "epoch": 0.9044228391263325, + "grad_norm": 0.14660704135894775, + "learning_rate": 4.475760997629541e-06, + "loss": 0.0311, + "step": 38582 + }, + { + "epoch": 0.9044462807011375, + "grad_norm": 0.267844021320343, + "learning_rate": 4.473582168806012e-06, + "loss": 0.0437, + "step": 38583 + }, + { + "epoch": 0.9044697222759425, + "grad_norm": 0.37737351655960083, + "learning_rate": 4.471403858311507e-06, + "loss": 0.4107, + "step": 38584 + }, + { + "epoch": 0.9044931638507475, + "grad_norm": 0.3113589286804199, + "learning_rate": 4.469226066157861e-06, + "loss": 0.0515, + "step": 38585 + }, + { + "epoch": 0.9045166054255525, + "grad_norm": 0.698296844959259, + "learning_rate": 4.467048792356876e-06, + "loss": 0.0632, + "step": 38586 + }, + { + "epoch": 0.9045400470003575, + "grad_norm": 0.3441912531852722, + "learning_rate": 4.464872036920365e-06, + "loss": 0.0762, + "step": 38587 + }, + { + "epoch": 0.9045634885751624, + "grad_norm": 0.7481638789176941, + "learning_rate": 4.462695799860139e-06, + "loss": 0.0776, + "step": 38588 + }, + { + "epoch": 0.9045869301499675, + "grad_norm": 0.33414241671562195, + "learning_rate": 4.460520081188014e-06, + "loss": 0.0391, + "step": 38589 + }, + { + "epoch": 0.9046103717247724, + "grad_norm": 1.1200910806655884, + "learning_rate": 4.458344880915799e-06, + "loss": 0.199, + "step": 38590 + }, + { + "epoch": 0.9046338132995775, + "grad_norm": 0.6881405115127563, + "learning_rate": 4.456170199055298e-06, + "loss": 0.0562, + "step": 38591 + }, + { + "epoch": 0.9046572548743824, + "grad_norm": 0.22324949502944946, + "learning_rate": 4.453996035618302e-06, + "loss": 0.0315, + "step": 38592 + }, + { + "epoch": 0.9046806964491875, + "grad_norm": 0.23081974685192108, + "learning_rate": 4.451822390616589e-06, + "loss": 0.0365, + "step": 38593 + }, + { + "epoch": 0.9047041380239924, + "grad_norm": 0.1324283480644226, + "learning_rate": 4.449649264061995e-06, + "loss": 0.0242, + "step": 38594 + }, + { + "epoch": 0.9047275795987975, + "grad_norm": 0.3065894544124603, + "learning_rate": 4.4474766559662764e-06, + "loss": 0.0268, + "step": 38595 + }, + { + "epoch": 0.9047510211736024, + "grad_norm": 0.3310220241546631, + "learning_rate": 4.445304566341246e-06, + "loss": 0.047, + "step": 38596 + }, + { + "epoch": 0.9047744627484074, + "grad_norm": 0.42287078499794006, + "learning_rate": 4.4431329951986735e-06, + "loss": 0.0995, + "step": 38597 + }, + { + "epoch": 0.9047979043232124, + "grad_norm": 0.33988919854164124, + "learning_rate": 4.440961942550359e-06, + "loss": 0.0876, + "step": 38598 + }, + { + "epoch": 0.9048213458980174, + "grad_norm": 0.4020940661430359, + "learning_rate": 4.438791408408072e-06, + "loss": 0.0461, + "step": 38599 + }, + { + "epoch": 0.9048447874728224, + "grad_norm": 0.41928187012672424, + "learning_rate": 4.436621392783569e-06, + "loss": 0.0825, + "step": 38600 + }, + { + "epoch": 0.9048682290476274, + "grad_norm": 0.5738756060600281, + "learning_rate": 4.4344518956886625e-06, + "loss": 0.1016, + "step": 38601 + }, + { + "epoch": 0.9048916706224324, + "grad_norm": 0.19254133105278015, + "learning_rate": 4.432282917135111e-06, + "loss": 0.0249, + "step": 38602 + }, + { + "epoch": 0.9049151121972374, + "grad_norm": 0.04818791523575783, + "learning_rate": 4.430114457134671e-06, + "loss": 0.0053, + "step": 38603 + }, + { + "epoch": 0.9049385537720424, + "grad_norm": 0.11432024836540222, + "learning_rate": 4.427946515699121e-06, + "loss": 0.0203, + "step": 38604 + }, + { + "epoch": 0.9049619953468474, + "grad_norm": 0.19808615744113922, + "learning_rate": 4.42577909284021e-06, + "loss": 0.0209, + "step": 38605 + }, + { + "epoch": 0.9049854369216523, + "grad_norm": 0.1168549582362175, + "learning_rate": 4.423612188569725e-06, + "loss": 0.0252, + "step": 38606 + }, + { + "epoch": 0.9050088784964574, + "grad_norm": 0.7015200257301331, + "learning_rate": 4.421445802899394e-06, + "loss": 0.0904, + "step": 38607 + }, + { + "epoch": 0.9050323200712624, + "grad_norm": 0.1954779028892517, + "learning_rate": 4.419279935840992e-06, + "loss": 0.0364, + "step": 38608 + }, + { + "epoch": 0.9050557616460674, + "grad_norm": 0.17855490744113922, + "learning_rate": 4.4171145874062684e-06, + "loss": 0.0142, + "step": 38609 + }, + { + "epoch": 0.9050792032208724, + "grad_norm": 0.43321678042411804, + "learning_rate": 4.4149497576069675e-06, + "loss": 0.0575, + "step": 38610 + }, + { + "epoch": 0.9051026447956774, + "grad_norm": 0.33878225088119507, + "learning_rate": 4.412785446454837e-06, + "loss": 0.0653, + "step": 38611 + }, + { + "epoch": 0.9051260863704824, + "grad_norm": 0.28484782576560974, + "learning_rate": 4.4106216539616215e-06, + "loss": 0.0313, + "step": 38612 + }, + { + "epoch": 0.9051495279452874, + "grad_norm": 0.45166829228401184, + "learning_rate": 4.408458380139047e-06, + "loss": 0.06, + "step": 38613 + }, + { + "epoch": 0.9051729695200924, + "grad_norm": 0.2833903431892395, + "learning_rate": 4.4062956249988685e-06, + "loss": 0.054, + "step": 38614 + }, + { + "epoch": 0.9051964110948973, + "grad_norm": 0.5431506633758545, + "learning_rate": 4.404133388552833e-06, + "loss": 0.0531, + "step": 38615 + }, + { + "epoch": 0.9052198526697024, + "grad_norm": 0.18471169471740723, + "learning_rate": 4.401971670812655e-06, + "loss": 0.0261, + "step": 38616 + }, + { + "epoch": 0.9052432942445073, + "grad_norm": 0.2918342351913452, + "learning_rate": 4.399810471790045e-06, + "loss": 0.0401, + "step": 38617 + }, + { + "epoch": 0.9052667358193124, + "grad_norm": 0.48907333612442017, + "learning_rate": 4.397649791496772e-06, + "loss": 0.1276, + "step": 38618 + }, + { + "epoch": 0.9052901773941173, + "grad_norm": 0.12422800809144974, + "learning_rate": 4.395489629944538e-06, + "loss": 0.0184, + "step": 38619 + }, + { + "epoch": 0.9053136189689224, + "grad_norm": 0.49243542551994324, + "learning_rate": 4.3933299871450455e-06, + "loss": 0.1137, + "step": 38620 + }, + { + "epoch": 0.9053370605437273, + "grad_norm": 0.09970088303089142, + "learning_rate": 4.39117086311005e-06, + "loss": 0.0073, + "step": 38621 + }, + { + "epoch": 0.9053605021185324, + "grad_norm": 0.20933328568935394, + "learning_rate": 4.389012257851255e-06, + "loss": 0.0171, + "step": 38622 + }, + { + "epoch": 0.9053839436933373, + "grad_norm": 0.43636003136634827, + "learning_rate": 4.386854171380361e-06, + "loss": 0.072, + "step": 38623 + }, + { + "epoch": 0.9054073852681424, + "grad_norm": 0.37597858905792236, + "learning_rate": 4.384696603709093e-06, + "loss": 0.0584, + "step": 38624 + }, + { + "epoch": 0.9054308268429473, + "grad_norm": 0.1784360110759735, + "learning_rate": 4.382539554849141e-06, + "loss": 0.0221, + "step": 38625 + }, + { + "epoch": 0.9054542684177523, + "grad_norm": 0.5487973093986511, + "learning_rate": 4.380383024812207e-06, + "loss": 0.0932, + "step": 38626 + }, + { + "epoch": 0.9054777099925573, + "grad_norm": 0.18849757313728333, + "learning_rate": 4.378227013610015e-06, + "loss": 0.0253, + "step": 38627 + }, + { + "epoch": 0.9055011515673623, + "grad_norm": 0.25968390703201294, + "learning_rate": 4.3760715212542546e-06, + "loss": 0.044, + "step": 38628 + }, + { + "epoch": 0.9055245931421673, + "grad_norm": 0.3387696444988251, + "learning_rate": 4.373916547756607e-06, + "loss": 0.0654, + "step": 38629 + }, + { + "epoch": 0.9055480347169723, + "grad_norm": 0.4378979206085205, + "learning_rate": 4.3717620931287954e-06, + "loss": 0.0767, + "step": 38630 + }, + { + "epoch": 0.9055714762917773, + "grad_norm": 0.3620951473712921, + "learning_rate": 4.369608157382476e-06, + "loss": 0.0383, + "step": 38631 + }, + { + "epoch": 0.9055949178665823, + "grad_norm": 0.33209002017974854, + "learning_rate": 4.367454740529364e-06, + "loss": 0.4061, + "step": 38632 + }, + { + "epoch": 0.9056183594413872, + "grad_norm": 0.5320050120353699, + "learning_rate": 4.3653018425811155e-06, + "loss": 0.0687, + "step": 38633 + }, + { + "epoch": 0.9056418010161923, + "grad_norm": 0.09328141808509827, + "learning_rate": 4.363149463549443e-06, + "loss": 0.0096, + "step": 38634 + }, + { + "epoch": 0.9056652425909972, + "grad_norm": 0.45260995626449585, + "learning_rate": 4.3609976034460045e-06, + "loss": 0.1122, + "step": 38635 + }, + { + "epoch": 0.9056886841658023, + "grad_norm": 0.2960692048072815, + "learning_rate": 4.35884626228249e-06, + "loss": 0.0488, + "step": 38636 + }, + { + "epoch": 0.9057121257406072, + "grad_norm": 0.21747040748596191, + "learning_rate": 4.3566954400705575e-06, + "loss": 0.0195, + "step": 38637 + }, + { + "epoch": 0.9057355673154123, + "grad_norm": 0.28583309054374695, + "learning_rate": 4.354545136821875e-06, + "loss": 0.0219, + "step": 38638 + }, + { + "epoch": 0.9057590088902172, + "grad_norm": 0.6158350110054016, + "learning_rate": 4.352395352548133e-06, + "loss": 0.4172, + "step": 38639 + }, + { + "epoch": 0.9057824504650223, + "grad_norm": 0.45707157254219055, + "learning_rate": 4.350246087260979e-06, + "loss": 0.0429, + "step": 38640 + }, + { + "epoch": 0.9058058920398272, + "grad_norm": 0.5733203887939453, + "learning_rate": 4.34809734097208e-06, + "loss": 0.1182, + "step": 38641 + }, + { + "epoch": 0.9058293336146322, + "grad_norm": 0.08485713601112366, + "learning_rate": 4.345949113693093e-06, + "loss": 0.0083, + "step": 38642 + }, + { + "epoch": 0.9058527751894372, + "grad_norm": 0.16583293676376343, + "learning_rate": 4.343801405435677e-06, + "loss": 0.0236, + "step": 38643 + }, + { + "epoch": 0.9058762167642422, + "grad_norm": 0.28128644824028015, + "learning_rate": 4.341654216211488e-06, + "loss": 0.0438, + "step": 38644 + }, + { + "epoch": 0.9058996583390472, + "grad_norm": 0.30114129185676575, + "learning_rate": 4.3395075460321624e-06, + "loss": 0.0362, + "step": 38645 + }, + { + "epoch": 0.9059230999138522, + "grad_norm": 0.29694893956184387, + "learning_rate": 4.3373613949093675e-06, + "loss": 0.0828, + "step": 38646 + }, + { + "epoch": 0.9059465414886572, + "grad_norm": 0.09661302715539932, + "learning_rate": 4.335215762854738e-06, + "loss": 0.0171, + "step": 38647 + }, + { + "epoch": 0.9059699830634622, + "grad_norm": 0.3759040832519531, + "learning_rate": 4.333070649879922e-06, + "loss": 0.0578, + "step": 38648 + }, + { + "epoch": 0.9059934246382672, + "grad_norm": 0.34381887316703796, + "learning_rate": 4.330926055996553e-06, + "loss": 0.0861, + "step": 38649 + }, + { + "epoch": 0.9060168662130722, + "grad_norm": 0.35462021827697754, + "learning_rate": 4.3287819812162675e-06, + "loss": 0.0377, + "step": 38650 + }, + { + "epoch": 0.9060403077878771, + "grad_norm": 0.5682254433631897, + "learning_rate": 4.3266384255506885e-06, + "loss": 0.0769, + "step": 38651 + }, + { + "epoch": 0.9060637493626822, + "grad_norm": 0.15718014538288116, + "learning_rate": 4.324495389011474e-06, + "loss": 0.0224, + "step": 38652 + }, + { + "epoch": 0.9060871909374871, + "grad_norm": 0.2639773190021515, + "learning_rate": 4.322352871610236e-06, + "loss": 0.0618, + "step": 38653 + }, + { + "epoch": 0.9061106325122922, + "grad_norm": 0.33945971727371216, + "learning_rate": 4.320210873358588e-06, + "loss": 0.0399, + "step": 38654 + }, + { + "epoch": 0.9061340740870971, + "grad_norm": 0.4627796411514282, + "learning_rate": 4.3180693942681875e-06, + "loss": 0.0569, + "step": 38655 + }, + { + "epoch": 0.9061575156619022, + "grad_norm": 0.13950522243976593, + "learning_rate": 4.315928434350624e-06, + "loss": 0.0288, + "step": 38656 + }, + { + "epoch": 0.9061809572367071, + "grad_norm": 0.14066770672798157, + "learning_rate": 4.313787993617524e-06, + "loss": 0.0161, + "step": 38657 + }, + { + "epoch": 0.9062043988115122, + "grad_norm": 0.352657675743103, + "learning_rate": 4.311648072080488e-06, + "loss": 0.0346, + "step": 38658 + }, + { + "epoch": 0.9062278403863172, + "grad_norm": 0.6332957744598389, + "learning_rate": 4.309508669751161e-06, + "loss": 0.1107, + "step": 38659 + }, + { + "epoch": 0.9062512819611221, + "grad_norm": 0.48818162083625793, + "learning_rate": 4.307369786641124e-06, + "loss": 0.0732, + "step": 38660 + }, + { + "epoch": 0.9062747235359272, + "grad_norm": 0.3751709759235382, + "learning_rate": 4.3052314227619796e-06, + "loss": 0.0635, + "step": 38661 + }, + { + "epoch": 0.9062981651107321, + "grad_norm": 0.3462831974029541, + "learning_rate": 4.3030935781253504e-06, + "loss": 0.4286, + "step": 38662 + }, + { + "epoch": 0.9063216066855372, + "grad_norm": 0.2593202292919159, + "learning_rate": 4.300956252742827e-06, + "loss": 0.0416, + "step": 38663 + }, + { + "epoch": 0.9063450482603421, + "grad_norm": 0.18445242941379547, + "learning_rate": 4.29881944662599e-06, + "loss": 0.0396, + "step": 38664 + }, + { + "epoch": 0.9063684898351472, + "grad_norm": 0.2855711281299591, + "learning_rate": 4.296683159786463e-06, + "loss": 0.058, + "step": 38665 + }, + { + "epoch": 0.9063919314099521, + "grad_norm": 0.0779290720820427, + "learning_rate": 4.294547392235815e-06, + "loss": 0.0141, + "step": 38666 + }, + { + "epoch": 0.9064153729847572, + "grad_norm": 0.2047181874513626, + "learning_rate": 4.292412143985658e-06, + "loss": 0.0202, + "step": 38667 + }, + { + "epoch": 0.9064388145595621, + "grad_norm": 0.29821521043777466, + "learning_rate": 4.29027741504755e-06, + "loss": 0.0524, + "step": 38668 + }, + { + "epoch": 0.9064622561343672, + "grad_norm": 0.2897945046424866, + "learning_rate": 4.288143205433104e-06, + "loss": 0.0765, + "step": 38669 + }, + { + "epoch": 0.9064856977091721, + "grad_norm": 0.18758432567119598, + "learning_rate": 4.286009515153877e-06, + "loss": 0.0394, + "step": 38670 + }, + { + "epoch": 0.9065091392839771, + "grad_norm": 0.6453201174736023, + "learning_rate": 4.283876344221438e-06, + "loss": 0.1085, + "step": 38671 + }, + { + "epoch": 0.9065325808587821, + "grad_norm": 0.8137009739875793, + "learning_rate": 4.281743692647389e-06, + "loss": 0.1951, + "step": 38672 + }, + { + "epoch": 0.9065560224335871, + "grad_norm": 0.35510051250457764, + "learning_rate": 4.279611560443297e-06, + "loss": 0.0644, + "step": 38673 + }, + { + "epoch": 0.9065794640083921, + "grad_norm": 0.9284756183624268, + "learning_rate": 4.277479947620721e-06, + "loss": 0.148, + "step": 38674 + }, + { + "epoch": 0.9066029055831971, + "grad_norm": 0.09371993690729141, + "learning_rate": 4.27534885419123e-06, + "loss": 0.0111, + "step": 38675 + }, + { + "epoch": 0.906626347158002, + "grad_norm": 0.2302350550889969, + "learning_rate": 4.27321828016638e-06, + "loss": 0.03, + "step": 38676 + }, + { + "epoch": 0.9066497887328071, + "grad_norm": 0.13953527808189392, + "learning_rate": 4.271088225557751e-06, + "loss": 0.0168, + "step": 38677 + }, + { + "epoch": 0.906673230307612, + "grad_norm": 0.11651880294084549, + "learning_rate": 4.268958690376868e-06, + "loss": 0.0123, + "step": 38678 + }, + { + "epoch": 0.9066966718824171, + "grad_norm": 0.30695271492004395, + "learning_rate": 4.26682967463532e-06, + "loss": 0.0597, + "step": 38679 + }, + { + "epoch": 0.906720113457222, + "grad_norm": 0.33238181471824646, + "learning_rate": 4.264701178344654e-06, + "loss": 0.0833, + "step": 38680 + }, + { + "epoch": 0.9067435550320271, + "grad_norm": 0.26888391375541687, + "learning_rate": 4.2625732015164065e-06, + "loss": 0.0589, + "step": 38681 + }, + { + "epoch": 0.906766996606832, + "grad_norm": 1.4550710916519165, + "learning_rate": 4.260445744162134e-06, + "loss": 0.1143, + "step": 38682 + }, + { + "epoch": 0.9067904381816371, + "grad_norm": 0.46846237778663635, + "learning_rate": 4.258318806293349e-06, + "loss": 0.0183, + "step": 38683 + }, + { + "epoch": 0.906813879756442, + "grad_norm": 0.5680191516876221, + "learning_rate": 4.256192387921643e-06, + "loss": 0.5793, + "step": 38684 + }, + { + "epoch": 0.9068373213312471, + "grad_norm": 0.31245630979537964, + "learning_rate": 4.254066489058528e-06, + "loss": 0.0447, + "step": 38685 + }, + { + "epoch": 0.906860762906052, + "grad_norm": 0.5877832174301147, + "learning_rate": 4.25194110971554e-06, + "loss": 0.0557, + "step": 38686 + }, + { + "epoch": 0.906884204480857, + "grad_norm": 0.4275391101837158, + "learning_rate": 4.249816249904204e-06, + "loss": 0.0797, + "step": 38687 + }, + { + "epoch": 0.906907646055662, + "grad_norm": 0.4734005928039551, + "learning_rate": 4.247691909636064e-06, + "loss": 0.0684, + "step": 38688 + }, + { + "epoch": 0.906931087630467, + "grad_norm": 0.39115476608276367, + "learning_rate": 4.245568088922625e-06, + "loss": 0.0818, + "step": 38689 + }, + { + "epoch": 0.906954529205272, + "grad_norm": 0.28752750158309937, + "learning_rate": 4.2434447877754305e-06, + "loss": 0.0608, + "step": 38690 + }, + { + "epoch": 0.906977970780077, + "grad_norm": 0.5521062016487122, + "learning_rate": 4.241322006206005e-06, + "loss": 0.0494, + "step": 38691 + }, + { + "epoch": 0.907001412354882, + "grad_norm": 0.48651400208473206, + "learning_rate": 4.239199744225864e-06, + "loss": 0.5405, + "step": 38692 + }, + { + "epoch": 0.907024853929687, + "grad_norm": 0.7964061498641968, + "learning_rate": 4.237078001846517e-06, + "loss": 0.1289, + "step": 38693 + }, + { + "epoch": 0.907048295504492, + "grad_norm": 0.5307581424713135, + "learning_rate": 4.234956779079469e-06, + "loss": 0.0869, + "step": 38694 + }, + { + "epoch": 0.907071737079297, + "grad_norm": 0.3021865785121918, + "learning_rate": 4.232836075936242e-06, + "loss": 0.0182, + "step": 38695 + }, + { + "epoch": 0.9070951786541019, + "grad_norm": 0.33703821897506714, + "learning_rate": 4.230715892428327e-06, + "loss": 0.0346, + "step": 38696 + }, + { + "epoch": 0.907118620228907, + "grad_norm": 0.4824385344982147, + "learning_rate": 4.228596228567261e-06, + "loss": 0.0733, + "step": 38697 + }, + { + "epoch": 0.9071420618037119, + "grad_norm": 0.4539291262626648, + "learning_rate": 4.226477084364522e-06, + "loss": 0.1178, + "step": 38698 + }, + { + "epoch": 0.907165503378517, + "grad_norm": 0.6818481087684631, + "learning_rate": 4.224358459831601e-06, + "loss": 0.5969, + "step": 38699 + }, + { + "epoch": 0.9071889449533219, + "grad_norm": 0.46066179871559143, + "learning_rate": 4.222240354980012e-06, + "loss": 0.0915, + "step": 38700 + }, + { + "epoch": 0.907212386528127, + "grad_norm": 0.2834857702255249, + "learning_rate": 4.220122769821233e-06, + "loss": 0.0465, + "step": 38701 + }, + { + "epoch": 0.9072358281029319, + "grad_norm": 0.4142194390296936, + "learning_rate": 4.218005704366756e-06, + "loss": 0.0736, + "step": 38702 + }, + { + "epoch": 0.907259269677737, + "grad_norm": 0.4294672906398773, + "learning_rate": 4.215889158628073e-06, + "loss": 0.5355, + "step": 38703 + }, + { + "epoch": 0.9072827112525419, + "grad_norm": 0.13005870580673218, + "learning_rate": 4.213773132616683e-06, + "loss": 0.0186, + "step": 38704 + }, + { + "epoch": 0.907306152827347, + "grad_norm": 0.15964075922966003, + "learning_rate": 4.211657626344045e-06, + "loss": 0.0155, + "step": 38705 + }, + { + "epoch": 0.9073295944021519, + "grad_norm": 0.29990893602371216, + "learning_rate": 4.209542639821651e-06, + "loss": 0.0282, + "step": 38706 + }, + { + "epoch": 0.9073530359769569, + "grad_norm": 0.5380011200904846, + "learning_rate": 4.207428173060968e-06, + "loss": 0.4859, + "step": 38707 + }, + { + "epoch": 0.9073764775517619, + "grad_norm": 0.3977324962615967, + "learning_rate": 4.205314226073476e-06, + "loss": 0.057, + "step": 38708 + }, + { + "epoch": 0.9073999191265669, + "grad_norm": 0.34026435017585754, + "learning_rate": 4.2032007988706325e-06, + "loss": 0.0285, + "step": 38709 + }, + { + "epoch": 0.907423360701372, + "grad_norm": 0.14662829041481018, + "learning_rate": 4.201087891463928e-06, + "loss": 0.0291, + "step": 38710 + }, + { + "epoch": 0.9074468022761769, + "grad_norm": 0.29432395100593567, + "learning_rate": 4.19897550386481e-06, + "loss": 0.3227, + "step": 38711 + }, + { + "epoch": 0.907470243850982, + "grad_norm": 0.8744010329246521, + "learning_rate": 4.196863636084747e-06, + "loss": 0.073, + "step": 38712 + }, + { + "epoch": 0.9074936854257869, + "grad_norm": 0.4518586993217468, + "learning_rate": 4.1947522881351956e-06, + "loss": 0.0648, + "step": 38713 + }, + { + "epoch": 0.907517127000592, + "grad_norm": 0.23285117745399475, + "learning_rate": 4.192641460027602e-06, + "loss": 0.0387, + "step": 38714 + }, + { + "epoch": 0.9075405685753969, + "grad_norm": 0.20563863217830658, + "learning_rate": 4.190531151773436e-06, + "loss": 0.0377, + "step": 38715 + }, + { + "epoch": 0.9075640101502019, + "grad_norm": 0.14006677269935608, + "learning_rate": 4.188421363384132e-06, + "loss": 0.018, + "step": 38716 + }, + { + "epoch": 0.9075874517250069, + "grad_norm": 0.2553512752056122, + "learning_rate": 4.186312094871159e-06, + "loss": 0.0421, + "step": 38717 + }, + { + "epoch": 0.9076108932998119, + "grad_norm": 0.46011239290237427, + "learning_rate": 4.184203346245952e-06, + "loss": 0.6232, + "step": 38718 + }, + { + "epoch": 0.9076343348746169, + "grad_norm": 0.12908045947551727, + "learning_rate": 4.182095117519957e-06, + "loss": 0.0215, + "step": 38719 + }, + { + "epoch": 0.9076577764494219, + "grad_norm": 0.13979654014110565, + "learning_rate": 4.1799874087046e-06, + "loss": 0.0214, + "step": 38720 + }, + { + "epoch": 0.9076812180242269, + "grad_norm": 0.511476993560791, + "learning_rate": 4.177880219811314e-06, + "loss": 0.0853, + "step": 38721 + }, + { + "epoch": 0.9077046595990319, + "grad_norm": 0.22216226160526276, + "learning_rate": 4.1757735508515585e-06, + "loss": 0.0136, + "step": 38722 + }, + { + "epoch": 0.9077281011738368, + "grad_norm": 0.28884533047676086, + "learning_rate": 4.173667401836756e-06, + "loss": 0.0657, + "step": 38723 + }, + { + "epoch": 0.9077515427486419, + "grad_norm": 0.48919349908828735, + "learning_rate": 4.171561772778321e-06, + "loss": 0.1309, + "step": 38724 + }, + { + "epoch": 0.9077749843234468, + "grad_norm": 0.18220682442188263, + "learning_rate": 4.169456663687688e-06, + "loss": 0.0378, + "step": 38725 + }, + { + "epoch": 0.9077984258982519, + "grad_norm": 0.22778207063674927, + "learning_rate": 4.16735207457628e-06, + "loss": 0.0163, + "step": 38726 + }, + { + "epoch": 0.9078218674730568, + "grad_norm": 0.19551685452461243, + "learning_rate": 4.165248005455502e-06, + "loss": 0.0251, + "step": 38727 + }, + { + "epoch": 0.9078453090478619, + "grad_norm": 0.11381158232688904, + "learning_rate": 4.163144456336776e-06, + "loss": 0.018, + "step": 38728 + }, + { + "epoch": 0.9078687506226668, + "grad_norm": 0.1709708422422409, + "learning_rate": 4.161041427231538e-06, + "loss": 0.0183, + "step": 38729 + }, + { + "epoch": 0.9078921921974719, + "grad_norm": 0.1769058108329773, + "learning_rate": 4.1589389181511895e-06, + "loss": 0.0253, + "step": 38730 + }, + { + "epoch": 0.9079156337722768, + "grad_norm": 0.35155728459358215, + "learning_rate": 4.156836929107133e-06, + "loss": 0.3405, + "step": 38731 + }, + { + "epoch": 0.9079390753470818, + "grad_norm": 0.4093841314315796, + "learning_rate": 4.15473546011077e-06, + "loss": 0.0346, + "step": 38732 + }, + { + "epoch": 0.9079625169218868, + "grad_norm": 0.42810362577438354, + "learning_rate": 4.152634511173503e-06, + "loss": 0.0555, + "step": 38733 + }, + { + "epoch": 0.9079859584966918, + "grad_norm": 0.48896756768226624, + "learning_rate": 4.1505340823067344e-06, + "loss": 0.0694, + "step": 38734 + }, + { + "epoch": 0.9080094000714968, + "grad_norm": 0.9313531517982483, + "learning_rate": 4.148434173521865e-06, + "loss": 0.1217, + "step": 38735 + }, + { + "epoch": 0.9080328416463018, + "grad_norm": 0.20834672451019287, + "learning_rate": 4.146334784830297e-06, + "loss": 0.0329, + "step": 38736 + }, + { + "epoch": 0.9080562832211068, + "grad_norm": 0.5926608443260193, + "learning_rate": 4.1442359162433995e-06, + "loss": 0.1274, + "step": 38737 + }, + { + "epoch": 0.9080797247959118, + "grad_norm": 0.5009017586708069, + "learning_rate": 4.142137567772586e-06, + "loss": 0.1002, + "step": 38738 + }, + { + "epoch": 0.9081031663707168, + "grad_norm": 0.7165907621383667, + "learning_rate": 4.140039739429202e-06, + "loss": 0.0977, + "step": 38739 + }, + { + "epoch": 0.9081266079455218, + "grad_norm": 0.140475332736969, + "learning_rate": 4.137942431224684e-06, + "loss": 0.0117, + "step": 38740 + }, + { + "epoch": 0.9081500495203267, + "grad_norm": 0.48606815934181213, + "learning_rate": 4.135845643170366e-06, + "loss": 0.0617, + "step": 38741 + }, + { + "epoch": 0.9081734910951318, + "grad_norm": 0.11797057092189789, + "learning_rate": 4.1337493752776625e-06, + "loss": 0.0136, + "step": 38742 + }, + { + "epoch": 0.9081969326699367, + "grad_norm": 0.23269404470920563, + "learning_rate": 4.13165362755793e-06, + "loss": 0.022, + "step": 38743 + }, + { + "epoch": 0.9082203742447418, + "grad_norm": 0.49775058031082153, + "learning_rate": 4.129558400022537e-06, + "loss": 0.5204, + "step": 38744 + }, + { + "epoch": 0.9082438158195467, + "grad_norm": 0.13059478998184204, + "learning_rate": 4.127463692682853e-06, + "loss": 0.0107, + "step": 38745 + }, + { + "epoch": 0.9082672573943518, + "grad_norm": 0.7465593814849854, + "learning_rate": 4.125369505550258e-06, + "loss": 0.8275, + "step": 38746 + }, + { + "epoch": 0.9082906989691567, + "grad_norm": 0.4769078493118286, + "learning_rate": 4.123275838636076e-06, + "loss": 0.0436, + "step": 38747 + }, + { + "epoch": 0.9083141405439618, + "grad_norm": 0.3262280821800232, + "learning_rate": 4.121182691951719e-06, + "loss": 0.0668, + "step": 38748 + }, + { + "epoch": 0.9083375821187667, + "grad_norm": 0.12315578013658524, + "learning_rate": 4.119090065508513e-06, + "loss": 0.0094, + "step": 38749 + }, + { + "epoch": 0.9083610236935717, + "grad_norm": 0.7193379998207092, + "learning_rate": 4.116997959317825e-06, + "loss": 0.4551, + "step": 38750 + }, + { + "epoch": 0.9083844652683767, + "grad_norm": 0.28755271434783936, + "learning_rate": 4.1149063733909924e-06, + "loss": 0.0241, + "step": 38751 + }, + { + "epoch": 0.9084079068431817, + "grad_norm": 0.14966031908988953, + "learning_rate": 4.112815307739381e-06, + "loss": 0.0257, + "step": 38752 + }, + { + "epoch": 0.9084313484179867, + "grad_norm": 0.4857092797756195, + "learning_rate": 4.110724762374329e-06, + "loss": 0.0549, + "step": 38753 + }, + { + "epoch": 0.9084547899927917, + "grad_norm": 0.612790048122406, + "learning_rate": 4.108634737307171e-06, + "loss": 0.0583, + "step": 38754 + }, + { + "epoch": 0.9084782315675967, + "grad_norm": 0.41518157720565796, + "learning_rate": 4.106545232549275e-06, + "loss": 0.0587, + "step": 38755 + }, + { + "epoch": 0.9085016731424017, + "grad_norm": 0.4963463544845581, + "learning_rate": 4.104456248111954e-06, + "loss": 0.058, + "step": 38756 + }, + { + "epoch": 0.9085251147172066, + "grad_norm": 0.20728479325771332, + "learning_rate": 4.102367784006555e-06, + "loss": 0.0177, + "step": 38757 + }, + { + "epoch": 0.9085485562920117, + "grad_norm": 0.372102290391922, + "learning_rate": 4.100279840244403e-06, + "loss": 0.0466, + "step": 38758 + }, + { + "epoch": 0.9085719978668166, + "grad_norm": 0.2607003450393677, + "learning_rate": 4.0981924168368105e-06, + "loss": 0.0182, + "step": 38759 + }, + { + "epoch": 0.9085954394416217, + "grad_norm": 0.336748868227005, + "learning_rate": 4.096105513795134e-06, + "loss": 0.0388, + "step": 38760 + }, + { + "epoch": 0.9086188810164267, + "grad_norm": 0.2574702501296997, + "learning_rate": 4.0940191311307e-06, + "loss": 0.0377, + "step": 38761 + }, + { + "epoch": 0.9086423225912317, + "grad_norm": 0.1999209225177765, + "learning_rate": 4.091933268854797e-06, + "loss": 0.0102, + "step": 38762 + }, + { + "epoch": 0.9086657641660367, + "grad_norm": 0.5811483263969421, + "learning_rate": 4.0898479269787625e-06, + "loss": 0.045, + "step": 38763 + }, + { + "epoch": 0.9086892057408417, + "grad_norm": 0.09958247095346451, + "learning_rate": 4.08776310551392e-06, + "loss": 0.0134, + "step": 38764 + }, + { + "epoch": 0.9087126473156467, + "grad_norm": 0.24462637305259705, + "learning_rate": 4.08567880447156e-06, + "loss": 0.0409, + "step": 38765 + }, + { + "epoch": 0.9087360888904517, + "grad_norm": 0.7324296236038208, + "learning_rate": 4.0835950238630076e-06, + "loss": 0.1526, + "step": 38766 + }, + { + "epoch": 0.9087595304652567, + "grad_norm": 0.35773855447769165, + "learning_rate": 4.081511763699564e-06, + "loss": 0.0463, + "step": 38767 + }, + { + "epoch": 0.9087829720400616, + "grad_norm": 0.4738655090332031, + "learning_rate": 4.079429023992542e-06, + "loss": 0.0758, + "step": 38768 + }, + { + "epoch": 0.9088064136148667, + "grad_norm": 0.5938723087310791, + "learning_rate": 4.077346804753235e-06, + "loss": 0.1168, + "step": 38769 + }, + { + "epoch": 0.9088298551896716, + "grad_norm": 0.26802101731300354, + "learning_rate": 4.0752651059929425e-06, + "loss": 0.0197, + "step": 38770 + }, + { + "epoch": 0.9088532967644767, + "grad_norm": 0.46743524074554443, + "learning_rate": 4.073183927722956e-06, + "loss": 0.0783, + "step": 38771 + }, + { + "epoch": 0.9088767383392816, + "grad_norm": 0.18278178572654724, + "learning_rate": 4.071103269954557e-06, + "loss": 0.0291, + "step": 38772 + }, + { + "epoch": 0.9089001799140867, + "grad_norm": 0.44093090295791626, + "learning_rate": 4.069023132699057e-06, + "loss": 0.0848, + "step": 38773 + }, + { + "epoch": 0.9089236214888916, + "grad_norm": 0.35692209005355835, + "learning_rate": 4.066943515967747e-06, + "loss": 0.0377, + "step": 38774 + }, + { + "epoch": 0.9089470630636967, + "grad_norm": 0.4436832666397095, + "learning_rate": 4.0648644197718745e-06, + "loss": 0.0777, + "step": 38775 + }, + { + "epoch": 0.9089705046385016, + "grad_norm": 0.61277174949646, + "learning_rate": 4.062785844122763e-06, + "loss": 0.1126, + "step": 38776 + }, + { + "epoch": 0.9089939462133066, + "grad_norm": 0.49405956268310547, + "learning_rate": 4.06070778903167e-06, + "loss": 0.4463, + "step": 38777 + }, + { + "epoch": 0.9090173877881116, + "grad_norm": 0.6130736470222473, + "learning_rate": 4.058630254509888e-06, + "loss": 0.0928, + "step": 38778 + }, + { + "epoch": 0.9090408293629166, + "grad_norm": 0.16143916547298431, + "learning_rate": 4.0565532405686505e-06, + "loss": 0.0149, + "step": 38779 + }, + { + "epoch": 0.9090642709377216, + "grad_norm": 0.5091866850852966, + "learning_rate": 4.05447674721926e-06, + "loss": 0.1016, + "step": 38780 + }, + { + "epoch": 0.9090877125125266, + "grad_norm": 0.5598315000534058, + "learning_rate": 4.052400774472987e-06, + "loss": 0.1268, + "step": 38781 + }, + { + "epoch": 0.9091111540873316, + "grad_norm": 0.17575664818286896, + "learning_rate": 4.050325322341087e-06, + "loss": 0.1145, + "step": 38782 + }, + { + "epoch": 0.9091345956621366, + "grad_norm": 0.4306204319000244, + "learning_rate": 4.0482503908348065e-06, + "loss": 0.0794, + "step": 38783 + }, + { + "epoch": 0.9091580372369416, + "grad_norm": 0.37427210807800293, + "learning_rate": 4.0461759799654275e-06, + "loss": 0.0574, + "step": 38784 + }, + { + "epoch": 0.9091814788117466, + "grad_norm": 0.5715280175209045, + "learning_rate": 4.044102089744184e-06, + "loss": 0.1014, + "step": 38785 + }, + { + "epoch": 0.9092049203865515, + "grad_norm": 0.2651415169239044, + "learning_rate": 4.042028720182356e-06, + "loss": 0.035, + "step": 38786 + }, + { + "epoch": 0.9092283619613566, + "grad_norm": 0.29684582352638245, + "learning_rate": 4.0399558712911565e-06, + "loss": 0.0543, + "step": 38787 + }, + { + "epoch": 0.9092518035361615, + "grad_norm": 0.36333099007606506, + "learning_rate": 4.037883543081867e-06, + "loss": 0.0641, + "step": 38788 + }, + { + "epoch": 0.9092752451109666, + "grad_norm": 0.4804827868938446, + "learning_rate": 4.03581173556572e-06, + "loss": 0.0619, + "step": 38789 + }, + { + "epoch": 0.9092986866857715, + "grad_norm": 0.5699872374534607, + "learning_rate": 4.0337404487539646e-06, + "loss": 0.0489, + "step": 38790 + }, + { + "epoch": 0.9093221282605766, + "grad_norm": 0.18489249050617218, + "learning_rate": 4.031669682657812e-06, + "loss": 0.0264, + "step": 38791 + }, + { + "epoch": 0.9093455698353815, + "grad_norm": 0.36378952860832214, + "learning_rate": 4.029599437288534e-06, + "loss": 0.0739, + "step": 38792 + }, + { + "epoch": 0.9093690114101866, + "grad_norm": 0.15380245447158813, + "learning_rate": 4.027529712657352e-06, + "loss": 0.02, + "step": 38793 + }, + { + "epoch": 0.9093924529849915, + "grad_norm": 0.6437494158744812, + "learning_rate": 4.025460508775492e-06, + "loss": 0.0915, + "step": 38794 + }, + { + "epoch": 0.9094158945597965, + "grad_norm": 0.29669681191444397, + "learning_rate": 4.023391825654177e-06, + "loss": 0.0233, + "step": 38795 + }, + { + "epoch": 0.9094393361346015, + "grad_norm": 0.1905946582555771, + "learning_rate": 4.021323663304633e-06, + "loss": 0.0214, + "step": 38796 + }, + { + "epoch": 0.9094627777094065, + "grad_norm": 0.27969637513160706, + "learning_rate": 4.019256021738072e-06, + "loss": 0.0432, + "step": 38797 + }, + { + "epoch": 0.9094862192842115, + "grad_norm": 0.33515745401382446, + "learning_rate": 4.017188900965752e-06, + "loss": 0.0342, + "step": 38798 + }, + { + "epoch": 0.9095096608590165, + "grad_norm": 0.8092637658119202, + "learning_rate": 4.0151223009988415e-06, + "loss": 0.0794, + "step": 38799 + }, + { + "epoch": 0.9095331024338215, + "grad_norm": 0.15644046664237976, + "learning_rate": 4.013056221848588e-06, + "loss": 0.0207, + "step": 38800 + }, + { + "epoch": 0.9095565440086265, + "grad_norm": 0.2899602949619293, + "learning_rate": 4.010990663526193e-06, + "loss": 0.0555, + "step": 38801 + }, + { + "epoch": 0.9095799855834314, + "grad_norm": 0.6012691259384155, + "learning_rate": 4.0089256260428584e-06, + "loss": 0.1111, + "step": 38802 + }, + { + "epoch": 0.9096034271582365, + "grad_norm": 0.22402922809123993, + "learning_rate": 4.006861109409799e-06, + "loss": 0.026, + "step": 38803 + }, + { + "epoch": 0.9096268687330414, + "grad_norm": 0.1422688066959381, + "learning_rate": 4.004797113638192e-06, + "loss": 0.0241, + "step": 38804 + }, + { + "epoch": 0.9096503103078465, + "grad_norm": 0.5267189145088196, + "learning_rate": 4.002733638739275e-06, + "loss": 0.0749, + "step": 38805 + }, + { + "epoch": 0.9096737518826514, + "grad_norm": 0.47298890352249146, + "learning_rate": 4.000670684724217e-06, + "loss": 0.1052, + "step": 38806 + }, + { + "epoch": 0.9096971934574565, + "grad_norm": 0.38883715867996216, + "learning_rate": 3.998608251604219e-06, + "loss": 0.0461, + "step": 38807 + }, + { + "epoch": 0.9097206350322614, + "grad_norm": 0.5337674617767334, + "learning_rate": 3.996546339390472e-06, + "loss": 0.0726, + "step": 38808 + }, + { + "epoch": 0.9097440766070665, + "grad_norm": 0.48620349168777466, + "learning_rate": 3.9944849480941685e-06, + "loss": 0.5682, + "step": 38809 + }, + { + "epoch": 0.9097675181818714, + "grad_norm": 0.367996484041214, + "learning_rate": 3.992424077726476e-06, + "loss": 0.0599, + "step": 38810 + }, + { + "epoch": 0.9097909597566765, + "grad_norm": 0.12863698601722717, + "learning_rate": 3.9903637282986076e-06, + "loss": 0.013, + "step": 38811 + }, + { + "epoch": 0.9098144013314815, + "grad_norm": 0.2380373477935791, + "learning_rate": 3.988303899821699e-06, + "loss": 0.0291, + "step": 38812 + }, + { + "epoch": 0.9098378429062864, + "grad_norm": 0.2837294638156891, + "learning_rate": 3.986244592306976e-06, + "loss": 0.0384, + "step": 38813 + }, + { + "epoch": 0.9098612844810915, + "grad_norm": 0.7895844578742981, + "learning_rate": 3.984185805765583e-06, + "loss": 0.1341, + "step": 38814 + }, + { + "epoch": 0.9098847260558964, + "grad_norm": 0.47568443417549133, + "learning_rate": 3.982127540208702e-06, + "loss": 0.0855, + "step": 38815 + }, + { + "epoch": 0.9099081676307015, + "grad_norm": 0.25965192914009094, + "learning_rate": 3.980069795647501e-06, + "loss": 0.0316, + "step": 38816 + }, + { + "epoch": 0.9099316092055064, + "grad_norm": 0.6627116203308105, + "learning_rate": 3.978012572093115e-06, + "loss": 0.1436, + "step": 38817 + }, + { + "epoch": 0.9099550507803115, + "grad_norm": 0.49119532108306885, + "learning_rate": 3.975955869556758e-06, + "loss": 0.0762, + "step": 38818 + }, + { + "epoch": 0.9099784923551164, + "grad_norm": 0.34271085262298584, + "learning_rate": 3.973899688049554e-06, + "loss": 0.0666, + "step": 38819 + }, + { + "epoch": 0.9100019339299215, + "grad_norm": 0.19608961045742035, + "learning_rate": 3.971844027582683e-06, + "loss": 0.0578, + "step": 38820 + }, + { + "epoch": 0.9100253755047264, + "grad_norm": 0.13644760847091675, + "learning_rate": 3.9697888881672806e-06, + "loss": 0.0319, + "step": 38821 + }, + { + "epoch": 0.9100488170795314, + "grad_norm": 0.8369969129562378, + "learning_rate": 3.967734269814494e-06, + "loss": 0.1277, + "step": 38822 + }, + { + "epoch": 0.9100722586543364, + "grad_norm": 0.6930314898490906, + "learning_rate": 3.9656801725354795e-06, + "loss": 0.1385, + "step": 38823 + }, + { + "epoch": 0.9100957002291414, + "grad_norm": 0.4388563334941864, + "learning_rate": 3.963626596341386e-06, + "loss": 0.5481, + "step": 38824 + }, + { + "epoch": 0.9101191418039464, + "grad_norm": 0.37609267234802246, + "learning_rate": 3.961573541243369e-06, + "loss": 0.4139, + "step": 38825 + }, + { + "epoch": 0.9101425833787514, + "grad_norm": 0.5888903737068176, + "learning_rate": 3.9595210072525425e-06, + "loss": 0.0857, + "step": 38826 + }, + { + "epoch": 0.9101660249535564, + "grad_norm": 0.2545942962169647, + "learning_rate": 3.957468994380064e-06, + "loss": 0.0444, + "step": 38827 + }, + { + "epoch": 0.9101894665283614, + "grad_norm": 0.17601150274276733, + "learning_rate": 3.95541750263706e-06, + "loss": 0.0222, + "step": 38828 + }, + { + "epoch": 0.9102129081031664, + "grad_norm": 0.3706717789173126, + "learning_rate": 3.953366532034641e-06, + "loss": 0.06, + "step": 38829 + }, + { + "epoch": 0.9102363496779714, + "grad_norm": 0.3799205422401428, + "learning_rate": 3.951316082583978e-06, + "loss": 0.0916, + "step": 38830 + }, + { + "epoch": 0.9102597912527763, + "grad_norm": 0.13344226777553558, + "learning_rate": 3.949266154296172e-06, + "loss": 0.018, + "step": 38831 + }, + { + "epoch": 0.9102832328275814, + "grad_norm": 0.24402619898319244, + "learning_rate": 3.9472167471823475e-06, + "loss": 0.0553, + "step": 38832 + }, + { + "epoch": 0.9103066744023863, + "grad_norm": 0.4295792579650879, + "learning_rate": 3.945167861253618e-06, + "loss": 0.0889, + "step": 38833 + }, + { + "epoch": 0.9103301159771914, + "grad_norm": 0.2595107853412628, + "learning_rate": 3.943119496521119e-06, + "loss": 0.0308, + "step": 38834 + }, + { + "epoch": 0.9103535575519963, + "grad_norm": 0.1986107975244522, + "learning_rate": 3.9410716529959425e-06, + "loss": 0.0383, + "step": 38835 + }, + { + "epoch": 0.9103769991268014, + "grad_norm": 0.4511110484600067, + "learning_rate": 3.939024330689201e-06, + "loss": 0.0997, + "step": 38836 + }, + { + "epoch": 0.9104004407016063, + "grad_norm": 0.522576630115509, + "learning_rate": 3.9369775296120405e-06, + "loss": 0.4005, + "step": 38837 + }, + { + "epoch": 0.9104238822764114, + "grad_norm": 0.5953429341316223, + "learning_rate": 3.934931249775531e-06, + "loss": 0.609, + "step": 38838 + }, + { + "epoch": 0.9104473238512163, + "grad_norm": 0.508956253528595, + "learning_rate": 3.932885491190785e-06, + "loss": 0.0936, + "step": 38839 + }, + { + "epoch": 0.9104707654260213, + "grad_norm": 0.6343119740486145, + "learning_rate": 3.930840253868906e-06, + "loss": 0.0802, + "step": 38840 + }, + { + "epoch": 0.9104942070008263, + "grad_norm": 0.4256681799888611, + "learning_rate": 3.928795537820995e-06, + "loss": 0.0694, + "step": 38841 + }, + { + "epoch": 0.9105176485756313, + "grad_norm": 0.26967549324035645, + "learning_rate": 3.926751343058121e-06, + "loss": 0.0348, + "step": 38842 + }, + { + "epoch": 0.9105410901504363, + "grad_norm": 0.1919751614332199, + "learning_rate": 3.92470766959141e-06, + "loss": 0.0114, + "step": 38843 + }, + { + "epoch": 0.9105645317252413, + "grad_norm": 0.49138835072517395, + "learning_rate": 3.92266451743194e-06, + "loss": 0.071, + "step": 38844 + }, + { + "epoch": 0.9105879733000463, + "grad_norm": 0.4122154712677002, + "learning_rate": 3.920621886590792e-06, + "loss": 0.0656, + "step": 38845 + }, + { + "epoch": 0.9106114148748513, + "grad_norm": 0.09630041569471359, + "learning_rate": 3.918579777079046e-06, + "loss": 0.0151, + "step": 38846 + }, + { + "epoch": 0.9106348564496562, + "grad_norm": 0.11021064966917038, + "learning_rate": 3.916538188907792e-06, + "loss": 0.0215, + "step": 38847 + }, + { + "epoch": 0.9106582980244613, + "grad_norm": 0.2164839506149292, + "learning_rate": 3.914497122088101e-06, + "loss": 0.0249, + "step": 38848 + }, + { + "epoch": 0.9106817395992662, + "grad_norm": 0.39267614483833313, + "learning_rate": 3.91245657663103e-06, + "loss": 0.0742, + "step": 38849 + }, + { + "epoch": 0.9107051811740713, + "grad_norm": 0.20794963836669922, + "learning_rate": 3.910416552547702e-06, + "loss": 0.0322, + "step": 38850 + }, + { + "epoch": 0.9107286227488762, + "grad_norm": 0.30214643478393555, + "learning_rate": 3.9083770498491435e-06, + "loss": 0.0259, + "step": 38851 + }, + { + "epoch": 0.9107520643236813, + "grad_norm": 0.4184999465942383, + "learning_rate": 3.9063380685464335e-06, + "loss": 0.0571, + "step": 38852 + }, + { + "epoch": 0.9107755058984862, + "grad_norm": 0.4318957030773163, + "learning_rate": 3.904299608650641e-06, + "loss": 0.0741, + "step": 38853 + }, + { + "epoch": 0.9107989474732913, + "grad_norm": 0.16283921897411346, + "learning_rate": 3.902261670172824e-06, + "loss": 0.0263, + "step": 38854 + }, + { + "epoch": 0.9108223890480962, + "grad_norm": 0.4358248710632324, + "learning_rate": 3.9002242531240185e-06, + "loss": 0.0653, + "step": 38855 + }, + { + "epoch": 0.9108458306229013, + "grad_norm": 0.49741071462631226, + "learning_rate": 3.898187357515315e-06, + "loss": 0.0839, + "step": 38856 + }, + { + "epoch": 0.9108692721977062, + "grad_norm": 0.2610912621021271, + "learning_rate": 3.896150983357738e-06, + "loss": 0.0361, + "step": 38857 + }, + { + "epoch": 0.9108927137725112, + "grad_norm": 0.52618807554245, + "learning_rate": 3.894115130662357e-06, + "loss": 0.068, + "step": 38858 + }, + { + "epoch": 0.9109161553473162, + "grad_norm": 0.33132457733154297, + "learning_rate": 3.892079799440207e-06, + "loss": 0.0662, + "step": 38859 + }, + { + "epoch": 0.9109395969221212, + "grad_norm": 0.3180379569530487, + "learning_rate": 3.890044989702324e-06, + "loss": 0.053, + "step": 38860 + }, + { + "epoch": 0.9109630384969262, + "grad_norm": 0.31387409567832947, + "learning_rate": 3.888010701459777e-06, + "loss": 0.0614, + "step": 38861 + }, + { + "epoch": 0.9109864800717312, + "grad_norm": 0.47999969124794006, + "learning_rate": 3.885976934723568e-06, + "loss": 0.0702, + "step": 38862 + }, + { + "epoch": 0.9110099216465363, + "grad_norm": 0.7354022264480591, + "learning_rate": 3.883943689504754e-06, + "loss": 0.125, + "step": 38863 + }, + { + "epoch": 0.9110333632213412, + "grad_norm": 0.10068074613809586, + "learning_rate": 3.881910965814373e-06, + "loss": 0.013, + "step": 38864 + }, + { + "epoch": 0.9110568047961463, + "grad_norm": 0.618039608001709, + "learning_rate": 3.879878763663447e-06, + "loss": 0.4755, + "step": 38865 + }, + { + "epoch": 0.9110802463709512, + "grad_norm": 0.4992727041244507, + "learning_rate": 3.877847083063002e-06, + "loss": 0.5162, + "step": 38866 + }, + { + "epoch": 0.9111036879457562, + "grad_norm": 0.46455326676368713, + "learning_rate": 3.875815924024051e-06, + "loss": 0.0958, + "step": 38867 + }, + { + "epoch": 0.9111271295205612, + "grad_norm": 0.13459862768650055, + "learning_rate": 3.87378528655763e-06, + "loss": 0.0172, + "step": 38868 + }, + { + "epoch": 0.9111505710953662, + "grad_norm": 0.39206627011299133, + "learning_rate": 3.871755170674751e-06, + "loss": 0.0816, + "step": 38869 + }, + { + "epoch": 0.9111740126701712, + "grad_norm": 0.08585355430841446, + "learning_rate": 3.86972557638644e-06, + "loss": 0.0094, + "step": 38870 + }, + { + "epoch": 0.9111974542449762, + "grad_norm": 0.6024913191795349, + "learning_rate": 3.8676965037037e-06, + "loss": 0.7105, + "step": 38871 + }, + { + "epoch": 0.9112208958197812, + "grad_norm": 0.2742942273616791, + "learning_rate": 3.86566795263753e-06, + "loss": 0.0461, + "step": 38872 + }, + { + "epoch": 0.9112443373945862, + "grad_norm": 0.3666864037513733, + "learning_rate": 3.8636399231989585e-06, + "loss": 0.0527, + "step": 38873 + }, + { + "epoch": 0.9112677789693912, + "grad_norm": 0.2819744348526001, + "learning_rate": 3.861612415398963e-06, + "loss": 0.2713, + "step": 38874 + }, + { + "epoch": 0.9112912205441962, + "grad_norm": 0.4846237301826477, + "learning_rate": 3.85958542924858e-06, + "loss": 0.0761, + "step": 38875 + }, + { + "epoch": 0.9113146621190011, + "grad_norm": 0.521207869052887, + "learning_rate": 3.85755896475879e-06, + "loss": 0.0704, + "step": 38876 + }, + { + "epoch": 0.9113381036938062, + "grad_norm": 0.4828602969646454, + "learning_rate": 3.855533021940583e-06, + "loss": 0.0667, + "step": 38877 + }, + { + "epoch": 0.9113615452686111, + "grad_norm": 0.45000481605529785, + "learning_rate": 3.853507600804962e-06, + "loss": 0.5806, + "step": 38878 + }, + { + "epoch": 0.9113849868434162, + "grad_norm": 0.8058789372444153, + "learning_rate": 3.851482701362918e-06, + "loss": 0.172, + "step": 38879 + }, + { + "epoch": 0.9114084284182211, + "grad_norm": 0.18365682661533356, + "learning_rate": 3.84945832362541e-06, + "loss": 0.036, + "step": 38880 + }, + { + "epoch": 0.9114318699930262, + "grad_norm": 0.299925297498703, + "learning_rate": 3.847434467603461e-06, + "loss": 0.0226, + "step": 38881 + }, + { + "epoch": 0.9114553115678311, + "grad_norm": 0.407265305519104, + "learning_rate": 3.84541113330803e-06, + "loss": 0.0792, + "step": 38882 + }, + { + "epoch": 0.9114787531426362, + "grad_norm": 0.4753277599811554, + "learning_rate": 3.843388320750107e-06, + "loss": 0.0802, + "step": 38883 + }, + { + "epoch": 0.9115021947174411, + "grad_norm": 0.14092624187469482, + "learning_rate": 3.841366029940663e-06, + "loss": 0.0213, + "step": 38884 + }, + { + "epoch": 0.9115256362922461, + "grad_norm": 0.05366359278559685, + "learning_rate": 3.839344260890665e-06, + "loss": 0.0069, + "step": 38885 + }, + { + "epoch": 0.9115490778670511, + "grad_norm": 0.35748812556266785, + "learning_rate": 3.837323013611094e-06, + "loss": 0.0862, + "step": 38886 + }, + { + "epoch": 0.9115725194418561, + "grad_norm": 0.3723500370979309, + "learning_rate": 3.835302288112896e-06, + "loss": 0.058, + "step": 38887 + }, + { + "epoch": 0.9115959610166611, + "grad_norm": 0.4845596253871918, + "learning_rate": 3.833282084407064e-06, + "loss": 0.5246, + "step": 38888 + }, + { + "epoch": 0.9116194025914661, + "grad_norm": 0.21460460126399994, + "learning_rate": 3.831262402504554e-06, + "loss": 0.0399, + "step": 38889 + }, + { + "epoch": 0.9116428441662711, + "grad_norm": 0.4345172643661499, + "learning_rate": 3.829243242416314e-06, + "loss": 0.0762, + "step": 38890 + }, + { + "epoch": 0.9116662857410761, + "grad_norm": 0.29891300201416016, + "learning_rate": 3.827224604153312e-06, + "loss": 0.054, + "step": 38891 + }, + { + "epoch": 0.911689727315881, + "grad_norm": 0.44912514090538025, + "learning_rate": 3.825206487726485e-06, + "loss": 0.0892, + "step": 38892 + }, + { + "epoch": 0.9117131688906861, + "grad_norm": 0.577626645565033, + "learning_rate": 3.8231888931467785e-06, + "loss": 0.0578, + "step": 38893 + }, + { + "epoch": 0.911736610465491, + "grad_norm": 0.16671140491962433, + "learning_rate": 3.821171820425173e-06, + "loss": 0.0372, + "step": 38894 + }, + { + "epoch": 0.9117600520402961, + "grad_norm": 0.3360001742839813, + "learning_rate": 3.819155269572594e-06, + "loss": 0.0421, + "step": 38895 + }, + { + "epoch": 0.911783493615101, + "grad_norm": 0.37304240465164185, + "learning_rate": 3.817139240599976e-06, + "loss": 0.0575, + "step": 38896 + }, + { + "epoch": 0.9118069351899061, + "grad_norm": 0.5780459046363831, + "learning_rate": 3.815123733518255e-06, + "loss": 0.1018, + "step": 38897 + }, + { + "epoch": 0.911830376764711, + "grad_norm": 0.14049813151359558, + "learning_rate": 3.81310874833839e-06, + "loss": 0.0121, + "step": 38898 + }, + { + "epoch": 0.9118538183395161, + "grad_norm": 0.3158322274684906, + "learning_rate": 3.811094285071293e-06, + "loss": 0.0383, + "step": 38899 + }, + { + "epoch": 0.911877259914321, + "grad_norm": 0.3583672046661377, + "learning_rate": 3.8090803437279e-06, + "loss": 0.0327, + "step": 38900 + }, + { + "epoch": 0.911900701489126, + "grad_norm": 0.136040598154068, + "learning_rate": 3.807066924319158e-06, + "loss": 0.0141, + "step": 38901 + }, + { + "epoch": 0.911924143063931, + "grad_norm": 0.43555891513824463, + "learning_rate": 3.8050540268559697e-06, + "loss": 0.0949, + "step": 38902 + }, + { + "epoch": 0.911947584638736, + "grad_norm": 0.3653567135334015, + "learning_rate": 3.8030416513492595e-06, + "loss": 0.0384, + "step": 38903 + }, + { + "epoch": 0.911971026213541, + "grad_norm": 0.1863693743944168, + "learning_rate": 3.8010297978099516e-06, + "loss": 0.0268, + "step": 38904 + }, + { + "epoch": 0.911994467788346, + "grad_norm": 0.28031986951828003, + "learning_rate": 3.79901846624896e-06, + "loss": 0.0385, + "step": 38905 + }, + { + "epoch": 0.912017909363151, + "grad_norm": 0.40802451968193054, + "learning_rate": 3.7970076566771982e-06, + "loss": 0.0346, + "step": 38906 + }, + { + "epoch": 0.912041350937956, + "grad_norm": 0.2681375741958618, + "learning_rate": 3.7949973691055794e-06, + "loss": 0.027, + "step": 38907 + }, + { + "epoch": 0.912064792512761, + "grad_norm": 0.0662945806980133, + "learning_rate": 3.7929876035450174e-06, + "loss": 0.0066, + "step": 38908 + }, + { + "epoch": 0.912088234087566, + "grad_norm": 0.43641892075538635, + "learning_rate": 3.7909783600063918e-06, + "loss": 0.0432, + "step": 38909 + }, + { + "epoch": 0.912111675662371, + "grad_norm": 0.12739428877830505, + "learning_rate": 3.788969638500639e-06, + "loss": 0.0212, + "step": 38910 + }, + { + "epoch": 0.912135117237176, + "grad_norm": 0.2880420982837677, + "learning_rate": 3.7869614390386386e-06, + "loss": 0.0456, + "step": 38911 + }, + { + "epoch": 0.9121585588119809, + "grad_norm": 0.45971032977104187, + "learning_rate": 3.7849537616312713e-06, + "loss": 0.0817, + "step": 38912 + }, + { + "epoch": 0.912182000386786, + "grad_norm": 0.1178613230586052, + "learning_rate": 3.7829466062894723e-06, + "loss": 0.025, + "step": 38913 + }, + { + "epoch": 0.912205441961591, + "grad_norm": 0.2385839819908142, + "learning_rate": 3.7809399730241114e-06, + "loss": 0.0513, + "step": 38914 + }, + { + "epoch": 0.912228883536396, + "grad_norm": 0.38071560859680176, + "learning_rate": 3.778933861846068e-06, + "loss": 0.0731, + "step": 38915 + }, + { + "epoch": 0.912252325111201, + "grad_norm": 0.24003644287586212, + "learning_rate": 3.776928272766245e-06, + "loss": 0.0449, + "step": 38916 + }, + { + "epoch": 0.912275766686006, + "grad_norm": 0.38103699684143066, + "learning_rate": 3.7749232057955107e-06, + "loss": 0.0471, + "step": 38917 + }, + { + "epoch": 0.912299208260811, + "grad_norm": 0.7506433129310608, + "learning_rate": 3.7729186609447355e-06, + "loss": 0.1697, + "step": 38918 + }, + { + "epoch": 0.912322649835616, + "grad_norm": 0.437431275844574, + "learning_rate": 3.7709146382248205e-06, + "loss": 0.0824, + "step": 38919 + }, + { + "epoch": 0.912346091410421, + "grad_norm": 0.5986083149909973, + "learning_rate": 3.7689111376466246e-06, + "loss": 0.0947, + "step": 38920 + }, + { + "epoch": 0.9123695329852259, + "grad_norm": 0.4225867986679077, + "learning_rate": 3.7669081592210167e-06, + "loss": 0.061, + "step": 38921 + }, + { + "epoch": 0.912392974560031, + "grad_norm": 0.5666364431381226, + "learning_rate": 3.7649057029588763e-06, + "loss": 0.0831, + "step": 38922 + }, + { + "epoch": 0.9124164161348359, + "grad_norm": 0.11727698147296906, + "learning_rate": 3.762903768871073e-06, + "loss": 0.0358, + "step": 38923 + }, + { + "epoch": 0.912439857709641, + "grad_norm": 0.07127140462398529, + "learning_rate": 3.760902356968454e-06, + "loss": 0.0169, + "step": 38924 + }, + { + "epoch": 0.9124632992844459, + "grad_norm": 0.14699093997478485, + "learning_rate": 3.7589014672618772e-06, + "loss": 0.0258, + "step": 38925 + }, + { + "epoch": 0.912486740859251, + "grad_norm": 0.10423999279737473, + "learning_rate": 3.7569010997622113e-06, + "loss": 0.0182, + "step": 38926 + }, + { + "epoch": 0.9125101824340559, + "grad_norm": 0.22085998952388763, + "learning_rate": 3.754901254480314e-06, + "loss": 0.019, + "step": 38927 + }, + { + "epoch": 0.912533624008861, + "grad_norm": 0.36399903893470764, + "learning_rate": 3.752901931427022e-06, + "loss": 0.0499, + "step": 38928 + }, + { + "epoch": 0.9125570655836659, + "grad_norm": 0.47643816471099854, + "learning_rate": 3.7509031306131926e-06, + "loss": 0.7211, + "step": 38929 + }, + { + "epoch": 0.9125805071584709, + "grad_norm": 0.34495314955711365, + "learning_rate": 3.7489048520496726e-06, + "loss": 0.0606, + "step": 38930 + }, + { + "epoch": 0.9126039487332759, + "grad_norm": 0.28536084294319153, + "learning_rate": 3.7469070957472873e-06, + "loss": 0.0426, + "step": 38931 + }, + { + "epoch": 0.9126273903080809, + "grad_norm": 0.340573251247406, + "learning_rate": 3.7449098617169054e-06, + "loss": 0.053, + "step": 38932 + }, + { + "epoch": 0.9126508318828859, + "grad_norm": 0.25926467776298523, + "learning_rate": 3.742913149969329e-06, + "loss": 0.0322, + "step": 38933 + }, + { + "epoch": 0.9126742734576909, + "grad_norm": 0.3356390595436096, + "learning_rate": 3.740916960515428e-06, + "loss": 0.3349, + "step": 38934 + }, + { + "epoch": 0.9126977150324959, + "grad_norm": 0.6071176528930664, + "learning_rate": 3.7389212933660266e-06, + "loss": 0.1002, + "step": 38935 + }, + { + "epoch": 0.9127211566073009, + "grad_norm": 0.2752193808555603, + "learning_rate": 3.7369261485319275e-06, + "loss": 0.0468, + "step": 38936 + }, + { + "epoch": 0.9127445981821058, + "grad_norm": 0.11558547616004944, + "learning_rate": 3.7349315260239882e-06, + "loss": 0.0097, + "step": 38937 + }, + { + "epoch": 0.9127680397569109, + "grad_norm": 0.5252336263656616, + "learning_rate": 3.7329374258530003e-06, + "loss": 0.0695, + "step": 38938 + }, + { + "epoch": 0.9127914813317158, + "grad_norm": 0.65348219871521, + "learning_rate": 3.730943848029811e-06, + "loss": 0.5454, + "step": 38939 + }, + { + "epoch": 0.9128149229065209, + "grad_norm": 0.6108977794647217, + "learning_rate": 3.7289507925652336e-06, + "loss": 0.0775, + "step": 38940 + }, + { + "epoch": 0.9128383644813258, + "grad_norm": 0.35611024498939514, + "learning_rate": 3.72695825947007e-06, + "loss": 0.0506, + "step": 38941 + }, + { + "epoch": 0.9128618060561309, + "grad_norm": 0.342617928981781, + "learning_rate": 3.7249662487551353e-06, + "loss": 0.0381, + "step": 38942 + }, + { + "epoch": 0.9128852476309358, + "grad_norm": 0.45351675152778625, + "learning_rate": 3.7229747604312415e-06, + "loss": 0.0727, + "step": 38943 + }, + { + "epoch": 0.9129086892057409, + "grad_norm": 0.7368939518928528, + "learning_rate": 3.720983794509192e-06, + "loss": 0.0958, + "step": 38944 + }, + { + "epoch": 0.9129321307805458, + "grad_norm": 0.5419182777404785, + "learning_rate": 3.7189933509997887e-06, + "loss": 0.1078, + "step": 38945 + }, + { + "epoch": 0.9129555723553509, + "grad_norm": 0.4593704044818878, + "learning_rate": 3.717003429913846e-06, + "loss": 0.0624, + "step": 38946 + }, + { + "epoch": 0.9129790139301558, + "grad_norm": 0.4790915250778198, + "learning_rate": 3.715014031262143e-06, + "loss": 0.0756, + "step": 38947 + }, + { + "epoch": 0.9130024555049608, + "grad_norm": 0.42523568868637085, + "learning_rate": 3.713025155055483e-06, + "loss": 0.4265, + "step": 38948 + }, + { + "epoch": 0.9130258970797658, + "grad_norm": 0.5239539742469788, + "learning_rate": 3.711036801304657e-06, + "loss": 0.4596, + "step": 38949 + }, + { + "epoch": 0.9130493386545708, + "grad_norm": 0.35830771923065186, + "learning_rate": 3.7090489700204456e-06, + "loss": 0.0424, + "step": 38950 + }, + { + "epoch": 0.9130727802293758, + "grad_norm": 0.3130730986595154, + "learning_rate": 3.707061661213651e-06, + "loss": 0.0318, + "step": 38951 + }, + { + "epoch": 0.9130962218041808, + "grad_norm": 0.12739011645317078, + "learning_rate": 3.7050748748950424e-06, + "loss": 0.0088, + "step": 38952 + }, + { + "epoch": 0.9131196633789858, + "grad_norm": 0.7143094539642334, + "learning_rate": 3.7030886110754114e-06, + "loss": 0.0828, + "step": 38953 + }, + { + "epoch": 0.9131431049537908, + "grad_norm": 0.5665123462677002, + "learning_rate": 3.701102869765527e-06, + "loss": 0.64, + "step": 38954 + }, + { + "epoch": 0.9131665465285957, + "grad_norm": 0.365979939699173, + "learning_rate": 3.699117650976169e-06, + "loss": 0.1095, + "step": 38955 + }, + { + "epoch": 0.9131899881034008, + "grad_norm": 0.4334649443626404, + "learning_rate": 3.6971329547180854e-06, + "loss": 0.0925, + "step": 38956 + }, + { + "epoch": 0.9132134296782057, + "grad_norm": 0.40690878033638, + "learning_rate": 3.695148781002078e-06, + "loss": 0.0281, + "step": 38957 + }, + { + "epoch": 0.9132368712530108, + "grad_norm": 0.4215516448020935, + "learning_rate": 3.693165129838905e-06, + "loss": 0.0877, + "step": 38958 + }, + { + "epoch": 0.9132603128278157, + "grad_norm": 0.12426381558179855, + "learning_rate": 3.691182001239324e-06, + "loss": 0.0205, + "step": 38959 + }, + { + "epoch": 0.9132837544026208, + "grad_norm": 0.18863244354724884, + "learning_rate": 3.6891993952141047e-06, + "loss": 0.0377, + "step": 38960 + }, + { + "epoch": 0.9133071959774257, + "grad_norm": 0.5145124793052673, + "learning_rate": 3.687217311773994e-06, + "loss": 0.3376, + "step": 38961 + }, + { + "epoch": 0.9133306375522308, + "grad_norm": 0.18132388591766357, + "learning_rate": 3.6852357509297497e-06, + "loss": 0.0195, + "step": 38962 + }, + { + "epoch": 0.9133540791270357, + "grad_norm": 0.24538244307041168, + "learning_rate": 3.683254712692108e-06, + "loss": 0.1774, + "step": 38963 + }, + { + "epoch": 0.9133775207018408, + "grad_norm": 0.48367369174957275, + "learning_rate": 3.68127419707186e-06, + "loss": 0.547, + "step": 38964 + }, + { + "epoch": 0.9134009622766458, + "grad_norm": 0.6595854163169861, + "learning_rate": 3.6792942040797196e-06, + "loss": 0.0832, + "step": 38965 + }, + { + "epoch": 0.9134244038514507, + "grad_norm": 0.5034907460212708, + "learning_rate": 3.6773147337264336e-06, + "loss": 0.1287, + "step": 38966 + }, + { + "epoch": 0.9134478454262558, + "grad_norm": 0.5760002136230469, + "learning_rate": 3.675335786022749e-06, + "loss": 0.572, + "step": 38967 + }, + { + "epoch": 0.9134712870010607, + "grad_norm": 0.6113753914833069, + "learning_rate": 3.6733573609794015e-06, + "loss": 0.0941, + "step": 38968 + }, + { + "epoch": 0.9134947285758658, + "grad_norm": 0.8265548348426819, + "learning_rate": 3.671379458607105e-06, + "loss": 0.1679, + "step": 38969 + }, + { + "epoch": 0.9135181701506707, + "grad_norm": 0.07180798798799515, + "learning_rate": 3.6694020789166172e-06, + "loss": 0.0076, + "step": 38970 + }, + { + "epoch": 0.9135416117254758, + "grad_norm": 0.17009253799915314, + "learning_rate": 3.6674252219186857e-06, + "loss": 0.0341, + "step": 38971 + }, + { + "epoch": 0.9135650533002807, + "grad_norm": 0.5088191628456116, + "learning_rate": 3.665448887624001e-06, + "loss": 0.0385, + "step": 38972 + }, + { + "epoch": 0.9135884948750858, + "grad_norm": 0.14781421422958374, + "learning_rate": 3.6634730760433002e-06, + "loss": 0.0185, + "step": 38973 + }, + { + "epoch": 0.9136119364498907, + "grad_norm": 0.4244741201400757, + "learning_rate": 3.6614977871873066e-06, + "loss": 0.0596, + "step": 38974 + }, + { + "epoch": 0.9136353780246957, + "grad_norm": 0.5235979557037354, + "learning_rate": 3.6595230210667354e-06, + "loss": 0.0927, + "step": 38975 + }, + { + "epoch": 0.9136588195995007, + "grad_norm": 0.406703919172287, + "learning_rate": 3.657548777692288e-06, + "loss": 0.6847, + "step": 38976 + }, + { + "epoch": 0.9136822611743057, + "grad_norm": 0.38325226306915283, + "learning_rate": 3.6555750570747005e-06, + "loss": 0.0933, + "step": 38977 + }, + { + "epoch": 0.9137057027491107, + "grad_norm": 0.6456860303878784, + "learning_rate": 3.6536018592246646e-06, + "loss": 0.1155, + "step": 38978 + }, + { + "epoch": 0.9137291443239157, + "grad_norm": 0.33663874864578247, + "learning_rate": 3.651629184152905e-06, + "loss": 0.0646, + "step": 38979 + }, + { + "epoch": 0.9137525858987207, + "grad_norm": 0.510405421257019, + "learning_rate": 3.649657031870113e-06, + "loss": 0.0908, + "step": 38980 + }, + { + "epoch": 0.9137760274735257, + "grad_norm": 0.32307755947113037, + "learning_rate": 3.64768540238698e-06, + "loss": 0.0312, + "step": 38981 + }, + { + "epoch": 0.9137994690483306, + "grad_norm": 0.29639723896980286, + "learning_rate": 3.64571429571422e-06, + "loss": 0.0384, + "step": 38982 + }, + { + "epoch": 0.9138229106231357, + "grad_norm": 0.30577749013900757, + "learning_rate": 3.643743711862524e-06, + "loss": 0.0487, + "step": 38983 + }, + { + "epoch": 0.9138463521979406, + "grad_norm": 0.6818931102752686, + "learning_rate": 3.641773650842584e-06, + "loss": 0.5146, + "step": 38984 + }, + { + "epoch": 0.9138697937727457, + "grad_norm": 0.30470186471939087, + "learning_rate": 3.6398041126650906e-06, + "loss": 0.2908, + "step": 38985 + }, + { + "epoch": 0.9138932353475506, + "grad_norm": 0.23847748339176178, + "learning_rate": 3.637835097340736e-06, + "loss": 0.0203, + "step": 38986 + }, + { + "epoch": 0.9139166769223557, + "grad_norm": 0.9082601070404053, + "learning_rate": 3.6358666048801894e-06, + "loss": 0.1906, + "step": 38987 + }, + { + "epoch": 0.9139401184971606, + "grad_norm": 0.24281086027622223, + "learning_rate": 3.633898635294142e-06, + "loss": 0.0288, + "step": 38988 + }, + { + "epoch": 0.9139635600719657, + "grad_norm": 0.35823163390159607, + "learning_rate": 3.6319311885932627e-06, + "loss": 0.0356, + "step": 38989 + }, + { + "epoch": 0.9139870016467706, + "grad_norm": 0.10586635023355484, + "learning_rate": 3.6299642647882547e-06, + "loss": 0.0093, + "step": 38990 + }, + { + "epoch": 0.9140104432215757, + "grad_norm": 0.584827184677124, + "learning_rate": 3.6279978638897537e-06, + "loss": 0.101, + "step": 38991 + }, + { + "epoch": 0.9140338847963806, + "grad_norm": 0.4539087116718292, + "learning_rate": 3.6260319859084513e-06, + "loss": 0.0675, + "step": 38992 + }, + { + "epoch": 0.9140573263711856, + "grad_norm": 0.4214378595352173, + "learning_rate": 3.6240666308550163e-06, + "loss": 0.0494, + "step": 38993 + }, + { + "epoch": 0.9140807679459906, + "grad_norm": 0.49765241146087646, + "learning_rate": 3.6221017987400853e-06, + "loss": 0.0779, + "step": 38994 + }, + { + "epoch": 0.9141042095207956, + "grad_norm": 0.26669883728027344, + "learning_rate": 3.6201374895743377e-06, + "loss": 0.0199, + "step": 38995 + }, + { + "epoch": 0.9141276510956006, + "grad_norm": 0.05691368505358696, + "learning_rate": 3.618173703368455e-06, + "loss": 0.006, + "step": 38996 + }, + { + "epoch": 0.9141510926704056, + "grad_norm": 0.2689782679080963, + "learning_rate": 3.6162104401330722e-06, + "loss": 0.0336, + "step": 38997 + }, + { + "epoch": 0.9141745342452106, + "grad_norm": 0.19597484171390533, + "learning_rate": 3.6142476998788365e-06, + "loss": 0.044, + "step": 38998 + }, + { + "epoch": 0.9141979758200156, + "grad_norm": 0.2885647118091583, + "learning_rate": 3.6122854826163954e-06, + "loss": 0.0302, + "step": 38999 + }, + { + "epoch": 0.9142214173948205, + "grad_norm": 0.2716212570667267, + "learning_rate": 3.6103237883564178e-06, + "loss": 0.0419, + "step": 39000 + }, + { + "epoch": 0.9142448589696256, + "grad_norm": 0.26311200857162476, + "learning_rate": 3.6083626171095174e-06, + "loss": 0.0371, + "step": 39001 + }, + { + "epoch": 0.9142683005444305, + "grad_norm": 0.525230884552002, + "learning_rate": 3.6064019688863525e-06, + "loss": 0.0533, + "step": 39002 + }, + { + "epoch": 0.9142917421192356, + "grad_norm": 0.5821173191070557, + "learning_rate": 3.60444184369757e-06, + "loss": 0.1159, + "step": 39003 + }, + { + "epoch": 0.9143151836940405, + "grad_norm": 0.1997520625591278, + "learning_rate": 3.6024822415537953e-06, + "loss": 0.0252, + "step": 39004 + }, + { + "epoch": 0.9143386252688456, + "grad_norm": 0.4777009189128876, + "learning_rate": 3.600523162465663e-06, + "loss": 0.0755, + "step": 39005 + }, + { + "epoch": 0.9143620668436505, + "grad_norm": 0.7055186033248901, + "learning_rate": 3.5985646064437994e-06, + "loss": 0.1122, + "step": 39006 + }, + { + "epoch": 0.9143855084184556, + "grad_norm": 0.10962063819169998, + "learning_rate": 3.5966065734988398e-06, + "loss": 0.0091, + "step": 39007 + }, + { + "epoch": 0.9144089499932605, + "grad_norm": 0.3979590833187103, + "learning_rate": 3.5946490636413865e-06, + "loss": 0.077, + "step": 39008 + }, + { + "epoch": 0.9144323915680656, + "grad_norm": 0.33067721128463745, + "learning_rate": 3.5926920768820984e-06, + "loss": 0.0288, + "step": 39009 + }, + { + "epoch": 0.9144558331428705, + "grad_norm": 0.3465020954608917, + "learning_rate": 3.590735613231566e-06, + "loss": 0.0247, + "step": 39010 + }, + { + "epoch": 0.9144792747176755, + "grad_norm": 0.06597158312797546, + "learning_rate": 3.588779672700415e-06, + "loss": 0.0056, + "step": 39011 + }, + { + "epoch": 0.9145027162924805, + "grad_norm": 0.5509863495826721, + "learning_rate": 3.5868242552992594e-06, + "loss": 0.0867, + "step": 39012 + }, + { + "epoch": 0.9145261578672855, + "grad_norm": 0.42228373885154724, + "learning_rate": 3.584869361038701e-06, + "loss": 0.0843, + "step": 39013 + }, + { + "epoch": 0.9145495994420905, + "grad_norm": 0.5023787021636963, + "learning_rate": 3.582914989929342e-06, + "loss": 0.6886, + "step": 39014 + }, + { + "epoch": 0.9145730410168955, + "grad_norm": 0.6271995306015015, + "learning_rate": 3.580961141981809e-06, + "loss": 0.1066, + "step": 39015 + }, + { + "epoch": 0.9145964825917006, + "grad_norm": 0.3888055682182312, + "learning_rate": 3.579007817206692e-06, + "loss": 0.0746, + "step": 39016 + }, + { + "epoch": 0.9146199241665055, + "grad_norm": 0.4052712917327881, + "learning_rate": 3.5770550156145944e-06, + "loss": 0.073, + "step": 39017 + }, + { + "epoch": 0.9146433657413106, + "grad_norm": 0.06192455440759659, + "learning_rate": 3.5751027372160963e-06, + "loss": 0.0052, + "step": 39018 + }, + { + "epoch": 0.9146668073161155, + "grad_norm": 0.5617801547050476, + "learning_rate": 3.5731509820218113e-06, + "loss": 0.1416, + "step": 39019 + }, + { + "epoch": 0.9146902488909205, + "grad_norm": 0.8750585913658142, + "learning_rate": 3.57119975004232e-06, + "loss": 0.1692, + "step": 39020 + }, + { + "epoch": 0.9147136904657255, + "grad_norm": 0.07725238800048828, + "learning_rate": 3.569249041288192e-06, + "loss": 0.0052, + "step": 39021 + }, + { + "epoch": 0.9147371320405305, + "grad_norm": 0.8495526909828186, + "learning_rate": 3.5672988557700515e-06, + "loss": 0.1405, + "step": 39022 + }, + { + "epoch": 0.9147605736153355, + "grad_norm": 0.5409951210021973, + "learning_rate": 3.5653491934984573e-06, + "loss": 0.095, + "step": 39023 + }, + { + "epoch": 0.9147840151901405, + "grad_norm": 0.48028725385665894, + "learning_rate": 3.5634000544839896e-06, + "loss": 0.0663, + "step": 39024 + }, + { + "epoch": 0.9148074567649455, + "grad_norm": 0.5445708632469177, + "learning_rate": 3.5614514387372176e-06, + "loss": 0.5199, + "step": 39025 + }, + { + "epoch": 0.9148308983397505, + "grad_norm": 0.42965003848075867, + "learning_rate": 3.559503346268711e-06, + "loss": 0.0829, + "step": 39026 + }, + { + "epoch": 0.9148543399145554, + "grad_norm": 0.2727417051792145, + "learning_rate": 3.5575557770890723e-06, + "loss": 0.0484, + "step": 39027 + }, + { + "epoch": 0.9148777814893605, + "grad_norm": 0.6254030466079712, + "learning_rate": 3.555608731208837e-06, + "loss": 0.1291, + "step": 39028 + }, + { + "epoch": 0.9149012230641654, + "grad_norm": 0.5571695566177368, + "learning_rate": 3.5536622086385863e-06, + "loss": 0.5873, + "step": 39029 + }, + { + "epoch": 0.9149246646389705, + "grad_norm": 0.11541536450386047, + "learning_rate": 3.5517162093888665e-06, + "loss": 0.0112, + "step": 39030 + }, + { + "epoch": 0.9149481062137754, + "grad_norm": 0.2929888963699341, + "learning_rate": 3.549770733470259e-06, + "loss": 0.0417, + "step": 39031 + }, + { + "epoch": 0.9149715477885805, + "grad_norm": 0.6106635332107544, + "learning_rate": 3.5478257808932992e-06, + "loss": 0.0755, + "step": 39032 + }, + { + "epoch": 0.9149949893633854, + "grad_norm": 0.5224538445472717, + "learning_rate": 3.5458813516685453e-06, + "loss": 0.0793, + "step": 39033 + }, + { + "epoch": 0.9150184309381905, + "grad_norm": 0.14018131792545319, + "learning_rate": 3.5439374458065556e-06, + "loss": 0.0259, + "step": 39034 + }, + { + "epoch": 0.9150418725129954, + "grad_norm": 0.08982640504837036, + "learning_rate": 3.5419940633178774e-06, + "loss": 0.013, + "step": 39035 + }, + { + "epoch": 0.9150653140878005, + "grad_norm": 0.3060513436794281, + "learning_rate": 3.5400512042130574e-06, + "loss": 0.0623, + "step": 39036 + }, + { + "epoch": 0.9150887556626054, + "grad_norm": 0.311331182718277, + "learning_rate": 3.538108868502621e-06, + "loss": 0.0277, + "step": 39037 + }, + { + "epoch": 0.9151121972374104, + "grad_norm": 0.22268952429294586, + "learning_rate": 3.5361670561971262e-06, + "loss": 0.0341, + "step": 39038 + }, + { + "epoch": 0.9151356388122154, + "grad_norm": 0.3196316659450531, + "learning_rate": 3.5342257673070867e-06, + "loss": 0.0342, + "step": 39039 + }, + { + "epoch": 0.9151590803870204, + "grad_norm": 0.15023019909858704, + "learning_rate": 3.532285001843061e-06, + "loss": 0.025, + "step": 39040 + }, + { + "epoch": 0.9151825219618254, + "grad_norm": 0.4925919473171234, + "learning_rate": 3.5303447598155624e-06, + "loss": 0.1153, + "step": 39041 + }, + { + "epoch": 0.9152059635366304, + "grad_norm": 0.13554050028324127, + "learning_rate": 3.528405041235139e-06, + "loss": 0.0172, + "step": 39042 + }, + { + "epoch": 0.9152294051114354, + "grad_norm": 0.5581582188606262, + "learning_rate": 3.5264658461122814e-06, + "loss": 0.8407, + "step": 39043 + }, + { + "epoch": 0.9152528466862404, + "grad_norm": 0.6514267921447754, + "learning_rate": 3.524527174457548e-06, + "loss": 0.0866, + "step": 39044 + }, + { + "epoch": 0.9152762882610453, + "grad_norm": 0.2441840022802353, + "learning_rate": 3.522589026281442e-06, + "loss": 0.1527, + "step": 39045 + }, + { + "epoch": 0.9152997298358504, + "grad_norm": 0.18540675938129425, + "learning_rate": 3.5206514015944767e-06, + "loss": 0.0321, + "step": 39046 + }, + { + "epoch": 0.9153231714106553, + "grad_norm": 0.14637380838394165, + "learning_rate": 3.5187143004071777e-06, + "loss": 0.0135, + "step": 39047 + }, + { + "epoch": 0.9153466129854604, + "grad_norm": 0.2001224011182785, + "learning_rate": 3.516777722730036e-06, + "loss": 0.014, + "step": 39048 + }, + { + "epoch": 0.9153700545602653, + "grad_norm": 0.47598737478256226, + "learning_rate": 3.5148416685735876e-06, + "loss": 0.0838, + "step": 39049 + }, + { + "epoch": 0.9153934961350704, + "grad_norm": 0.41759514808654785, + "learning_rate": 3.512906137948313e-06, + "loss": 0.0145, + "step": 39050 + }, + { + "epoch": 0.9154169377098753, + "grad_norm": 0.501275360584259, + "learning_rate": 3.510971130864715e-06, + "loss": 0.0708, + "step": 39051 + }, + { + "epoch": 0.9154403792846804, + "grad_norm": 0.2820700407028198, + "learning_rate": 3.509036647333297e-06, + "loss": 0.041, + "step": 39052 + }, + { + "epoch": 0.9154638208594853, + "grad_norm": 0.10894431173801422, + "learning_rate": 3.507102687364572e-06, + "loss": 0.0221, + "step": 39053 + }, + { + "epoch": 0.9154872624342903, + "grad_norm": 0.282062828540802, + "learning_rate": 3.5051692509690205e-06, + "loss": 0.0346, + "step": 39054 + }, + { + "epoch": 0.9155107040090953, + "grad_norm": 0.28438565135002136, + "learning_rate": 3.503236338157123e-06, + "loss": 0.0414, + "step": 39055 + }, + { + "epoch": 0.9155341455839003, + "grad_norm": 0.46601763367652893, + "learning_rate": 3.501303948939394e-06, + "loss": 0.0918, + "step": 39056 + }, + { + "epoch": 0.9155575871587053, + "grad_norm": 0.559024453163147, + "learning_rate": 3.499372083326291e-06, + "loss": 0.1394, + "step": 39057 + }, + { + "epoch": 0.9155810287335103, + "grad_norm": 0.1860072761774063, + "learning_rate": 3.4974407413283172e-06, + "loss": 0.0268, + "step": 39058 + }, + { + "epoch": 0.9156044703083153, + "grad_norm": 0.37902912497520447, + "learning_rate": 3.495509922955931e-06, + "loss": 0.0462, + "step": 39059 + }, + { + "epoch": 0.9156279118831203, + "grad_norm": 0.4169555902481079, + "learning_rate": 3.493579628219634e-06, + "loss": 0.1184, + "step": 39060 + }, + { + "epoch": 0.9156513534579253, + "grad_norm": 0.31727463006973267, + "learning_rate": 3.4916498571298862e-06, + "loss": 0.0439, + "step": 39061 + }, + { + "epoch": 0.9156747950327303, + "grad_norm": 0.29996103048324585, + "learning_rate": 3.489720609697156e-06, + "loss": 0.0333, + "step": 39062 + }, + { + "epoch": 0.9156982366075352, + "grad_norm": 0.37287387251853943, + "learning_rate": 3.4877918859319124e-06, + "loss": 0.0702, + "step": 39063 + }, + { + "epoch": 0.9157216781823403, + "grad_norm": 0.12438113242387772, + "learning_rate": 3.485863685844626e-06, + "loss": 0.0135, + "step": 39064 + }, + { + "epoch": 0.9157451197571452, + "grad_norm": 0.23721647262573242, + "learning_rate": 3.483936009445754e-06, + "loss": 0.0243, + "step": 39065 + }, + { + "epoch": 0.9157685613319503, + "grad_norm": 0.340559720993042, + "learning_rate": 3.4820088567457666e-06, + "loss": 0.0348, + "step": 39066 + }, + { + "epoch": 0.9157920029067552, + "grad_norm": 0.456252783536911, + "learning_rate": 3.4800822277550993e-06, + "loss": 0.1578, + "step": 39067 + }, + { + "epoch": 0.9158154444815603, + "grad_norm": 0.2848847508430481, + "learning_rate": 3.478156122484233e-06, + "loss": 0.0396, + "step": 39068 + }, + { + "epoch": 0.9158388860563653, + "grad_norm": 0.4368343949317932, + "learning_rate": 3.4762305409436037e-06, + "loss": 0.1124, + "step": 39069 + }, + { + "epoch": 0.9158623276311703, + "grad_norm": 0.3703644871711731, + "learning_rate": 3.4743054831436695e-06, + "loss": 0.03, + "step": 39070 + }, + { + "epoch": 0.9158857692059753, + "grad_norm": 0.14313508570194244, + "learning_rate": 3.4723809490948446e-06, + "loss": 0.0134, + "step": 39071 + }, + { + "epoch": 0.9159092107807802, + "grad_norm": 0.07910292595624924, + "learning_rate": 3.470456938807609e-06, + "loss": 0.0115, + "step": 39072 + }, + { + "epoch": 0.9159326523555853, + "grad_norm": 0.3633187711238861, + "learning_rate": 3.4685334522923997e-06, + "loss": 0.0752, + "step": 39073 + }, + { + "epoch": 0.9159560939303902, + "grad_norm": 0.9079835414886475, + "learning_rate": 3.46661048955963e-06, + "loss": 0.1014, + "step": 39074 + }, + { + "epoch": 0.9159795355051953, + "grad_norm": 0.4087633490562439, + "learning_rate": 3.4646880506197466e-06, + "loss": 0.065, + "step": 39075 + }, + { + "epoch": 0.9160029770800002, + "grad_norm": 0.5065134167671204, + "learning_rate": 3.4627661354831863e-06, + "loss": 0.0422, + "step": 39076 + }, + { + "epoch": 0.9160264186548053, + "grad_norm": 0.7297402620315552, + "learning_rate": 3.460844744160352e-06, + "loss": 0.0701, + "step": 39077 + }, + { + "epoch": 0.9160498602296102, + "grad_norm": 0.7830748558044434, + "learning_rate": 3.4589238766617015e-06, + "loss": 0.0934, + "step": 39078 + }, + { + "epoch": 0.9160733018044153, + "grad_norm": 0.14266365766525269, + "learning_rate": 3.457003532997638e-06, + "loss": 0.0225, + "step": 39079 + }, + { + "epoch": 0.9160967433792202, + "grad_norm": 0.29228729009628296, + "learning_rate": 3.455083713178597e-06, + "loss": 0.0439, + "step": 39080 + }, + { + "epoch": 0.9161201849540253, + "grad_norm": 0.668376624584198, + "learning_rate": 3.4531644172149823e-06, + "loss": 0.6952, + "step": 39081 + }, + { + "epoch": 0.9161436265288302, + "grad_norm": 0.384965181350708, + "learning_rate": 3.4512456451172183e-06, + "loss": 0.0595, + "step": 39082 + }, + { + "epoch": 0.9161670681036352, + "grad_norm": 0.25571656227111816, + "learning_rate": 3.4493273968957073e-06, + "loss": 0.0535, + "step": 39083 + }, + { + "epoch": 0.9161905096784402, + "grad_norm": 0.0919007733464241, + "learning_rate": 3.447409672560853e-06, + "loss": 0.0195, + "step": 39084 + }, + { + "epoch": 0.9162139512532452, + "grad_norm": 0.34437960386276245, + "learning_rate": 3.4454924721230795e-06, + "loss": 0.039, + "step": 39085 + }, + { + "epoch": 0.9162373928280502, + "grad_norm": 0.7273131012916565, + "learning_rate": 3.4435757955927684e-06, + "loss": 0.097, + "step": 39086 + }, + { + "epoch": 0.9162608344028552, + "grad_norm": 0.2307533621788025, + "learning_rate": 3.441659642980344e-06, + "loss": 0.0337, + "step": 39087 + }, + { + "epoch": 0.9162842759776602, + "grad_norm": 0.3591004014015198, + "learning_rate": 3.4397440142961757e-06, + "loss": 0.4199, + "step": 39088 + }, + { + "epoch": 0.9163077175524652, + "grad_norm": 0.35365432500839233, + "learning_rate": 3.437828909550689e-06, + "loss": 0.0644, + "step": 39089 + }, + { + "epoch": 0.9163311591272701, + "grad_norm": 0.11084137111902237, + "learning_rate": 3.4359143287542306e-06, + "loss": 0.0062, + "step": 39090 + }, + { + "epoch": 0.9163546007020752, + "grad_norm": 0.6352010369300842, + "learning_rate": 3.434000271917226e-06, + "loss": 0.1033, + "step": 39091 + }, + { + "epoch": 0.9163780422768801, + "grad_norm": 0.5493470430374146, + "learning_rate": 3.4320867390500554e-06, + "loss": 0.143, + "step": 39092 + }, + { + "epoch": 0.9164014838516852, + "grad_norm": 0.32324081659317017, + "learning_rate": 3.4301737301631e-06, + "loss": 0.0469, + "step": 39093 + }, + { + "epoch": 0.9164249254264901, + "grad_norm": 0.3287024199962616, + "learning_rate": 3.4282612452667393e-06, + "loss": 0.0689, + "step": 39094 + }, + { + "epoch": 0.9164483670012952, + "grad_norm": 0.4694918394088745, + "learning_rate": 3.426349284371344e-06, + "loss": 0.1193, + "step": 39095 + }, + { + "epoch": 0.9164718085761001, + "grad_norm": 0.6603008508682251, + "learning_rate": 3.424437847487294e-06, + "loss": 0.1006, + "step": 39096 + }, + { + "epoch": 0.9164952501509052, + "grad_norm": 0.26941630244255066, + "learning_rate": 3.4225269346249476e-06, + "loss": 0.0405, + "step": 39097 + }, + { + "epoch": 0.9165186917257101, + "grad_norm": 0.48202136158943176, + "learning_rate": 3.420616545794708e-06, + "loss": 0.072, + "step": 39098 + }, + { + "epoch": 0.9165421333005151, + "grad_norm": 0.1669824868440628, + "learning_rate": 3.4187066810069003e-06, + "loss": 0.0332, + "step": 39099 + }, + { + "epoch": 0.9165655748753201, + "grad_norm": 0.22329223155975342, + "learning_rate": 3.4167973402719155e-06, + "loss": 0.0354, + "step": 39100 + }, + { + "epoch": 0.9165890164501251, + "grad_norm": 0.5534812808036804, + "learning_rate": 3.4148885236001128e-06, + "loss": 0.08, + "step": 39101 + }, + { + "epoch": 0.9166124580249301, + "grad_norm": 0.4451132118701935, + "learning_rate": 3.4129802310018166e-06, + "loss": 0.0974, + "step": 39102 + }, + { + "epoch": 0.9166358995997351, + "grad_norm": 1.0079251527786255, + "learning_rate": 3.411072462487408e-06, + "loss": 0.1042, + "step": 39103 + }, + { + "epoch": 0.9166593411745401, + "grad_norm": 0.2854044437408447, + "learning_rate": 3.4091652180672452e-06, + "loss": 0.0415, + "step": 39104 + }, + { + "epoch": 0.9166827827493451, + "grad_norm": 0.3382579982280731, + "learning_rate": 3.4072584977516753e-06, + "loss": 0.0592, + "step": 39105 + }, + { + "epoch": 0.91670622432415, + "grad_norm": 0.8852648735046387, + "learning_rate": 3.405352301551035e-06, + "loss": 0.0441, + "step": 39106 + }, + { + "epoch": 0.9167296658989551, + "grad_norm": 0.1296425312757492, + "learning_rate": 3.4034466294756593e-06, + "loss": 0.0201, + "step": 39107 + }, + { + "epoch": 0.91675310747376, + "grad_norm": 0.37929391860961914, + "learning_rate": 3.401541481535908e-06, + "loss": 0.0407, + "step": 39108 + }, + { + "epoch": 0.9167765490485651, + "grad_norm": 0.42596402764320374, + "learning_rate": 3.3996368577420947e-06, + "loss": 0.0631, + "step": 39109 + }, + { + "epoch": 0.91679999062337, + "grad_norm": 0.6323529481887817, + "learning_rate": 3.397732758104577e-06, + "loss": 0.1276, + "step": 39110 + }, + { + "epoch": 0.9168234321981751, + "grad_norm": 0.22806039452552795, + "learning_rate": 3.395829182633681e-06, + "loss": 0.0469, + "step": 39111 + }, + { + "epoch": 0.91684687377298, + "grad_norm": 0.2017417699098587, + "learning_rate": 3.3939261313397308e-06, + "loss": 0.0225, + "step": 39112 + }, + { + "epoch": 0.9168703153477851, + "grad_norm": 0.4497678875923157, + "learning_rate": 3.392023604233063e-06, + "loss": 0.0811, + "step": 39113 + }, + { + "epoch": 0.91689375692259, + "grad_norm": 0.38181284070014954, + "learning_rate": 3.390121601323981e-06, + "loss": 0.0545, + "step": 39114 + }, + { + "epoch": 0.9169171984973951, + "grad_norm": 0.13287755846977234, + "learning_rate": 3.3882201226228093e-06, + "loss": 0.0177, + "step": 39115 + }, + { + "epoch": 0.9169406400722, + "grad_norm": 0.1595577895641327, + "learning_rate": 3.3863191681398622e-06, + "loss": 0.0259, + "step": 39116 + }, + { + "epoch": 0.916964081647005, + "grad_norm": 0.19360752403736115, + "learning_rate": 3.3844187378854864e-06, + "loss": 0.0377, + "step": 39117 + }, + { + "epoch": 0.91698752322181, + "grad_norm": 0.3173507750034332, + "learning_rate": 3.382518831869963e-06, + "loss": 0.0581, + "step": 39118 + }, + { + "epoch": 0.917010964796615, + "grad_norm": 0.058737121522426605, + "learning_rate": 3.380619450103606e-06, + "loss": 0.0064, + "step": 39119 + }, + { + "epoch": 0.9170344063714201, + "grad_norm": 0.13542477786540985, + "learning_rate": 3.3787205925967403e-06, + "loss": 0.0076, + "step": 39120 + }, + { + "epoch": 0.917057847946225, + "grad_norm": 0.6153483986854553, + "learning_rate": 3.3768222593596353e-06, + "loss": 0.4221, + "step": 39121 + }, + { + "epoch": 0.9170812895210301, + "grad_norm": 0.3506695330142975, + "learning_rate": 3.374924450402617e-06, + "loss": 0.028, + "step": 39122 + }, + { + "epoch": 0.917104731095835, + "grad_norm": 0.5195209383964539, + "learning_rate": 3.373027165735976e-06, + "loss": 0.079, + "step": 39123 + }, + { + "epoch": 0.9171281726706401, + "grad_norm": 0.30425018072128296, + "learning_rate": 3.371130405370004e-06, + "loss": 0.028, + "step": 39124 + }, + { + "epoch": 0.917151614245445, + "grad_norm": 0.6105926036834717, + "learning_rate": 3.3692341693149942e-06, + "loss": 0.6733, + "step": 39125 + }, + { + "epoch": 0.91717505582025, + "grad_norm": 0.1998719573020935, + "learning_rate": 3.3673384575812484e-06, + "loss": 0.0302, + "step": 39126 + }, + { + "epoch": 0.917198497395055, + "grad_norm": 0.3941783308982849, + "learning_rate": 3.365443270179025e-06, + "loss": 0.035, + "step": 39127 + }, + { + "epoch": 0.91722193896986, + "grad_norm": 0.1605040729045868, + "learning_rate": 3.3635486071186273e-06, + "loss": 0.0117, + "step": 39128 + }, + { + "epoch": 0.917245380544665, + "grad_norm": 0.4656812250614166, + "learning_rate": 3.3616544684103247e-06, + "loss": 0.625, + "step": 39129 + }, + { + "epoch": 0.91726882211947, + "grad_norm": 0.14681120216846466, + "learning_rate": 3.359760854064409e-06, + "loss": 0.0245, + "step": 39130 + }, + { + "epoch": 0.917292263694275, + "grad_norm": 0.505124032497406, + "learning_rate": 3.3578677640911603e-06, + "loss": 0.0924, + "step": 39131 + }, + { + "epoch": 0.91731570526908, + "grad_norm": 0.385762095451355, + "learning_rate": 3.355975198500827e-06, + "loss": 0.3732, + "step": 39132 + }, + { + "epoch": 0.917339146843885, + "grad_norm": 0.5923713445663452, + "learning_rate": 3.3540831573037003e-06, + "loss": 0.0706, + "step": 39133 + }, + { + "epoch": 0.91736258841869, + "grad_norm": 0.3175552487373352, + "learning_rate": 3.3521916405100163e-06, + "loss": 0.1668, + "step": 39134 + }, + { + "epoch": 0.917386029993495, + "grad_norm": 0.2383379191160202, + "learning_rate": 3.350300648130067e-06, + "loss": 0.0253, + "step": 39135 + }, + { + "epoch": 0.9174094715683, + "grad_norm": 0.2683844268321991, + "learning_rate": 3.3484101801741107e-06, + "loss": 0.0189, + "step": 39136 + }, + { + "epoch": 0.9174329131431049, + "grad_norm": 0.34303879737854004, + "learning_rate": 3.346520236652395e-06, + "loss": 0.0497, + "step": 39137 + }, + { + "epoch": 0.91745635471791, + "grad_norm": 0.07661563158035278, + "learning_rate": 3.3446308175751896e-06, + "loss": 0.0135, + "step": 39138 + }, + { + "epoch": 0.9174797962927149, + "grad_norm": 0.5991618037223816, + "learning_rate": 3.342741922952719e-06, + "loss": 0.0871, + "step": 39139 + }, + { + "epoch": 0.91750323786752, + "grad_norm": 0.5350205302238464, + "learning_rate": 3.340853552795242e-06, + "loss": 0.1116, + "step": 39140 + }, + { + "epoch": 0.9175266794423249, + "grad_norm": 0.5005561709403992, + "learning_rate": 3.3389657071130174e-06, + "loss": 0.0396, + "step": 39141 + }, + { + "epoch": 0.91755012101713, + "grad_norm": 0.15324419736862183, + "learning_rate": 3.3370783859162923e-06, + "loss": 0.0194, + "step": 39142 + }, + { + "epoch": 0.9175735625919349, + "grad_norm": 0.9911743402481079, + "learning_rate": 3.3351915892152918e-06, + "loss": 0.188, + "step": 39143 + }, + { + "epoch": 0.91759700416674, + "grad_norm": 0.4336603581905365, + "learning_rate": 3.333305317020263e-06, + "loss": 0.0577, + "step": 39144 + }, + { + "epoch": 0.9176204457415449, + "grad_norm": 0.0509505495429039, + "learning_rate": 3.331419569341432e-06, + "loss": 0.0036, + "step": 39145 + }, + { + "epoch": 0.9176438873163499, + "grad_norm": 0.357806533575058, + "learning_rate": 3.329534346189045e-06, + "loss": 0.0662, + "step": 39146 + }, + { + "epoch": 0.9176673288911549, + "grad_norm": 0.25429457426071167, + "learning_rate": 3.3276496475733053e-06, + "loss": 0.0197, + "step": 39147 + }, + { + "epoch": 0.9176907704659599, + "grad_norm": 0.40106236934661865, + "learning_rate": 3.325765473504472e-06, + "loss": 0.081, + "step": 39148 + }, + { + "epoch": 0.9177142120407649, + "grad_norm": 0.4857618510723114, + "learning_rate": 3.323881823992758e-06, + "loss": 0.678, + "step": 39149 + }, + { + "epoch": 0.9177376536155699, + "grad_norm": 0.25357586145401, + "learning_rate": 3.3219986990483674e-06, + "loss": 0.0279, + "step": 39150 + }, + { + "epoch": 0.9177610951903749, + "grad_norm": 0.5228712558746338, + "learning_rate": 3.320116098681536e-06, + "loss": 0.1159, + "step": 39151 + }, + { + "epoch": 0.9177845367651799, + "grad_norm": 0.4491972327232361, + "learning_rate": 3.318234022902467e-06, + "loss": 0.0866, + "step": 39152 + }, + { + "epoch": 0.9178079783399848, + "grad_norm": 0.1984671801328659, + "learning_rate": 3.3163524717213847e-06, + "loss": 0.0242, + "step": 39153 + }, + { + "epoch": 0.9178314199147899, + "grad_norm": 0.3755369782447815, + "learning_rate": 3.3144714451484813e-06, + "loss": 0.0556, + "step": 39154 + }, + { + "epoch": 0.9178548614895948, + "grad_norm": 0.5562315583229065, + "learning_rate": 3.312590943193983e-06, + "loss": 0.0438, + "step": 39155 + }, + { + "epoch": 0.9178783030643999, + "grad_norm": 0.15271908044815063, + "learning_rate": 3.3107109658680912e-06, + "loss": 0.0192, + "step": 39156 + }, + { + "epoch": 0.9179017446392048, + "grad_norm": 0.9697504043579102, + "learning_rate": 3.308831513180999e-06, + "loss": 0.1192, + "step": 39157 + }, + { + "epoch": 0.9179251862140099, + "grad_norm": 0.27210816740989685, + "learning_rate": 3.3069525851429085e-06, + "loss": 0.0268, + "step": 39158 + }, + { + "epoch": 0.9179486277888148, + "grad_norm": 0.10615811496973038, + "learning_rate": 3.3050741817640006e-06, + "loss": 0.0151, + "step": 39159 + }, + { + "epoch": 0.9179720693636199, + "grad_norm": 0.5162174701690674, + "learning_rate": 3.303196303054479e-06, + "loss": 0.0974, + "step": 39160 + }, + { + "epoch": 0.9179955109384248, + "grad_norm": 0.26767903566360474, + "learning_rate": 3.3013189490245344e-06, + "loss": 0.0542, + "step": 39161 + }, + { + "epoch": 0.9180189525132298, + "grad_norm": 0.20496952533721924, + "learning_rate": 3.2994421196843595e-06, + "loss": 0.0365, + "step": 39162 + }, + { + "epoch": 0.9180423940880348, + "grad_norm": 0.09372169524431229, + "learning_rate": 3.297565815044135e-06, + "loss": 0.0109, + "step": 39163 + }, + { + "epoch": 0.9180658356628398, + "grad_norm": 0.4733653664588928, + "learning_rate": 3.2956900351140187e-06, + "loss": 0.2531, + "step": 39164 + }, + { + "epoch": 0.9180892772376448, + "grad_norm": 0.19323278963565826, + "learning_rate": 3.293814779904214e-06, + "loss": 0.0154, + "step": 39165 + }, + { + "epoch": 0.9181127188124498, + "grad_norm": 0.4941387474536896, + "learning_rate": 3.291940049424902e-06, + "loss": 0.0662, + "step": 39166 + }, + { + "epoch": 0.9181361603872548, + "grad_norm": 0.2396833449602127, + "learning_rate": 3.2900658436862186e-06, + "loss": 0.0201, + "step": 39167 + }, + { + "epoch": 0.9181596019620598, + "grad_norm": 0.5087657570838928, + "learning_rate": 3.288192162698378e-06, + "loss": 0.0595, + "step": 39168 + }, + { + "epoch": 0.9181830435368648, + "grad_norm": 0.6055035591125488, + "learning_rate": 3.286319006471528e-06, + "loss": 0.6181, + "step": 39169 + }, + { + "epoch": 0.9182064851116698, + "grad_norm": 0.336921364068985, + "learning_rate": 3.284446375015826e-06, + "loss": 0.0579, + "step": 39170 + }, + { + "epoch": 0.9182299266864749, + "grad_norm": 0.428554892539978, + "learning_rate": 3.2825742683414317e-06, + "loss": 0.463, + "step": 39171 + }, + { + "epoch": 0.9182533682612798, + "grad_norm": 0.08325812220573425, + "learning_rate": 3.280702686458503e-06, + "loss": 0.0189, + "step": 39172 + }, + { + "epoch": 0.9182768098360848, + "grad_norm": 0.22303198277950287, + "learning_rate": 3.27883162937721e-06, + "loss": 0.0377, + "step": 39173 + }, + { + "epoch": 0.9183002514108898, + "grad_norm": 0.26892486214637756, + "learning_rate": 3.2769610971077e-06, + "loss": 0.0523, + "step": 39174 + }, + { + "epoch": 0.9183236929856948, + "grad_norm": 0.3205837309360504, + "learning_rate": 3.2750910896601207e-06, + "loss": 0.0614, + "step": 39175 + }, + { + "epoch": 0.9183471345604998, + "grad_norm": 0.0833398625254631, + "learning_rate": 3.2732216070445963e-06, + "loss": 0.0091, + "step": 39176 + }, + { + "epoch": 0.9183705761353048, + "grad_norm": 0.47651347517967224, + "learning_rate": 3.2713526492713087e-06, + "loss": 0.0483, + "step": 39177 + }, + { + "epoch": 0.9183940177101098, + "grad_norm": 0.5166176557540894, + "learning_rate": 3.2694842163503825e-06, + "loss": 0.7192, + "step": 39178 + }, + { + "epoch": 0.9184174592849148, + "grad_norm": 0.1632796972990036, + "learning_rate": 3.267616308291943e-06, + "loss": 0.0199, + "step": 39179 + }, + { + "epoch": 0.9184409008597197, + "grad_norm": 0.5307711958885193, + "learning_rate": 3.265748925106138e-06, + "loss": 0.0958, + "step": 39180 + }, + { + "epoch": 0.9184643424345248, + "grad_norm": 0.22318759560585022, + "learning_rate": 3.2638820668031145e-06, + "loss": 0.0367, + "step": 39181 + }, + { + "epoch": 0.9184877840093297, + "grad_norm": 0.530354380607605, + "learning_rate": 3.262015733392976e-06, + "loss": 0.1183, + "step": 39182 + }, + { + "epoch": 0.9185112255841348, + "grad_norm": 0.3759760558605194, + "learning_rate": 3.260149924885869e-06, + "loss": 0.0436, + "step": 39183 + }, + { + "epoch": 0.9185346671589397, + "grad_norm": 0.3211272358894348, + "learning_rate": 3.2582846412918976e-06, + "loss": 0.044, + "step": 39184 + }, + { + "epoch": 0.9185581087337448, + "grad_norm": 0.43136370182037354, + "learning_rate": 3.2564198826211976e-06, + "loss": 0.0905, + "step": 39185 + }, + { + "epoch": 0.9185815503085497, + "grad_norm": 0.43664804100990295, + "learning_rate": 3.2545556488838837e-06, + "loss": 0.0795, + "step": 39186 + }, + { + "epoch": 0.9186049918833548, + "grad_norm": 0.5951036810874939, + "learning_rate": 3.252691940090069e-06, + "loss": 0.1173, + "step": 39187 + }, + { + "epoch": 0.9186284334581597, + "grad_norm": 0.6432479619979858, + "learning_rate": 3.2508287562498684e-06, + "loss": 0.1431, + "step": 39188 + }, + { + "epoch": 0.9186518750329647, + "grad_norm": 0.2637730538845062, + "learning_rate": 3.248966097373396e-06, + "loss": 0.0377, + "step": 39189 + }, + { + "epoch": 0.9186753166077697, + "grad_norm": 0.5468705892562866, + "learning_rate": 3.2471039634707545e-06, + "loss": 0.0421, + "step": 39190 + }, + { + "epoch": 0.9186987581825747, + "grad_norm": 0.30419760942459106, + "learning_rate": 3.245242354552036e-06, + "loss": 0.0622, + "step": 39191 + }, + { + "epoch": 0.9187221997573797, + "grad_norm": 0.7997422814369202, + "learning_rate": 3.2433812706273547e-06, + "loss": 0.0835, + "step": 39192 + }, + { + "epoch": 0.9187456413321847, + "grad_norm": 0.4137301743030548, + "learning_rate": 3.2415207117068135e-06, + "loss": 0.0757, + "step": 39193 + }, + { + "epoch": 0.9187690829069897, + "grad_norm": 0.11139591783285141, + "learning_rate": 3.2396606778005046e-06, + "loss": 0.0236, + "step": 39194 + }, + { + "epoch": 0.9187925244817947, + "grad_norm": 0.5117552876472473, + "learning_rate": 3.2378011689185193e-06, + "loss": 0.1311, + "step": 39195 + }, + { + "epoch": 0.9188159660565997, + "grad_norm": 0.2631508409976959, + "learning_rate": 3.2359421850709503e-06, + "loss": 0.0315, + "step": 39196 + }, + { + "epoch": 0.9188394076314047, + "grad_norm": 0.5862762928009033, + "learning_rate": 3.234083726267867e-06, + "loss": 0.1434, + "step": 39197 + }, + { + "epoch": 0.9188628492062096, + "grad_norm": 0.2707007825374603, + "learning_rate": 3.232225792519361e-06, + "loss": 0.029, + "step": 39198 + }, + { + "epoch": 0.9188862907810147, + "grad_norm": 0.292743057012558, + "learning_rate": 3.230368383835536e-06, + "loss": 0.0489, + "step": 39199 + }, + { + "epoch": 0.9189097323558196, + "grad_norm": 0.3826669752597809, + "learning_rate": 3.228511500226439e-06, + "loss": 0.0675, + "step": 39200 + }, + { + "epoch": 0.9189331739306247, + "grad_norm": 0.37132060527801514, + "learning_rate": 3.2266551417021617e-06, + "loss": 0.0834, + "step": 39201 + }, + { + "epoch": 0.9189566155054296, + "grad_norm": 0.19276542961597443, + "learning_rate": 3.2247993082727747e-06, + "loss": 0.0425, + "step": 39202 + }, + { + "epoch": 0.9189800570802347, + "grad_norm": 0.3748624920845032, + "learning_rate": 3.222943999948358e-06, + "loss": 0.0505, + "step": 39203 + }, + { + "epoch": 0.9190034986550396, + "grad_norm": 0.3675926923751831, + "learning_rate": 3.2210892167389594e-06, + "loss": 0.0619, + "step": 39204 + }, + { + "epoch": 0.9190269402298447, + "grad_norm": 0.5097628831863403, + "learning_rate": 3.219234958654649e-06, + "loss": 0.1205, + "step": 39205 + }, + { + "epoch": 0.9190503818046496, + "grad_norm": 0.3886793255805969, + "learning_rate": 3.2173812257054957e-06, + "loss": 0.0838, + "step": 39206 + }, + { + "epoch": 0.9190738233794546, + "grad_norm": 0.6250425577163696, + "learning_rate": 3.215528017901559e-06, + "loss": 0.125, + "step": 39207 + }, + { + "epoch": 0.9190972649542596, + "grad_norm": 0.23830142617225647, + "learning_rate": 3.2136753352528857e-06, + "loss": 0.0536, + "step": 39208 + }, + { + "epoch": 0.9191207065290646, + "grad_norm": 0.15831626951694489, + "learning_rate": 3.211823177769524e-06, + "loss": 0.0281, + "step": 39209 + }, + { + "epoch": 0.9191441481038696, + "grad_norm": 0.07036016136407852, + "learning_rate": 3.209971545461532e-06, + "loss": 0.0145, + "step": 39210 + }, + { + "epoch": 0.9191675896786746, + "grad_norm": 0.2007395476102829, + "learning_rate": 3.2081204383389572e-06, + "loss": 0.0373, + "step": 39211 + }, + { + "epoch": 0.9191910312534796, + "grad_norm": 0.50757896900177, + "learning_rate": 3.2062698564118478e-06, + "loss": 0.4234, + "step": 39212 + }, + { + "epoch": 0.9192144728282846, + "grad_norm": 0.4782693386077881, + "learning_rate": 3.204419799690228e-06, + "loss": 0.1133, + "step": 39213 + }, + { + "epoch": 0.9192379144030896, + "grad_norm": 0.5338391661643982, + "learning_rate": 3.2025702681841574e-06, + "loss": 0.4743, + "step": 39214 + }, + { + "epoch": 0.9192613559778946, + "grad_norm": 0.5918217301368713, + "learning_rate": 3.200721261903661e-06, + "loss": 0.1366, + "step": 39215 + }, + { + "epoch": 0.9192847975526995, + "grad_norm": 0.2325650304555893, + "learning_rate": 3.198872780858786e-06, + "loss": 0.0227, + "step": 39216 + }, + { + "epoch": 0.9193082391275046, + "grad_norm": 0.1735965758562088, + "learning_rate": 3.1970248250595246e-06, + "loss": 0.0244, + "step": 39217 + }, + { + "epoch": 0.9193316807023095, + "grad_norm": 0.11016301065683365, + "learning_rate": 3.1951773945159357e-06, + "loss": 0.0128, + "step": 39218 + }, + { + "epoch": 0.9193551222771146, + "grad_norm": 0.2500292956829071, + "learning_rate": 3.1933304892380446e-06, + "loss": 0.0351, + "step": 39219 + }, + { + "epoch": 0.9193785638519195, + "grad_norm": 0.1588941514492035, + "learning_rate": 3.191484109235865e-06, + "loss": 0.0195, + "step": 39220 + }, + { + "epoch": 0.9194020054267246, + "grad_norm": 0.38177159428596497, + "learning_rate": 3.1896382545194116e-06, + "loss": 0.079, + "step": 39221 + }, + { + "epoch": 0.9194254470015296, + "grad_norm": 0.5675920248031616, + "learning_rate": 3.1877929250987093e-06, + "loss": 0.0737, + "step": 39222 + }, + { + "epoch": 0.9194488885763346, + "grad_norm": 0.27609434723854065, + "learning_rate": 3.1859481209837393e-06, + "loss": 0.0551, + "step": 39223 + }, + { + "epoch": 0.9194723301511396, + "grad_norm": 0.19291962683200836, + "learning_rate": 3.1841038421845603e-06, + "loss": 0.0333, + "step": 39224 + }, + { + "epoch": 0.9194957717259445, + "grad_norm": 0.4058024287223816, + "learning_rate": 3.182260088711142e-06, + "loss": 0.0547, + "step": 39225 + }, + { + "epoch": 0.9195192133007496, + "grad_norm": 0.5389703512191772, + "learning_rate": 3.1804168605735097e-06, + "loss": 0.0803, + "step": 39226 + }, + { + "epoch": 0.9195426548755545, + "grad_norm": 0.4336630702018738, + "learning_rate": 3.1785741577816663e-06, + "loss": 0.0925, + "step": 39227 + }, + { + "epoch": 0.9195660964503596, + "grad_norm": 0.29140299558639526, + "learning_rate": 3.1767319803455932e-06, + "loss": 0.0669, + "step": 39228 + }, + { + "epoch": 0.9195895380251645, + "grad_norm": 0.4907107949256897, + "learning_rate": 3.174890328275304e-06, + "loss": 0.041, + "step": 39229 + }, + { + "epoch": 0.9196129795999696, + "grad_norm": 0.34124407172203064, + "learning_rate": 3.173049201580769e-06, + "loss": 0.0515, + "step": 39230 + }, + { + "epoch": 0.9196364211747745, + "grad_norm": 0.16905900835990906, + "learning_rate": 3.171208600271991e-06, + "loss": 0.0312, + "step": 39231 + }, + { + "epoch": 0.9196598627495796, + "grad_norm": 0.5952872633934021, + "learning_rate": 3.1693685243589733e-06, + "loss": 0.1198, + "step": 39232 + }, + { + "epoch": 0.9196833043243845, + "grad_norm": 0.4012174606323242, + "learning_rate": 3.1675289738516744e-06, + "loss": 0.0489, + "step": 39233 + }, + { + "epoch": 0.9197067458991895, + "grad_norm": 0.1514482945203781, + "learning_rate": 3.1656899487600867e-06, + "loss": 0.0302, + "step": 39234 + }, + { + "epoch": 0.9197301874739945, + "grad_norm": 0.2881390452384949, + "learning_rate": 3.1638514490941908e-06, + "loss": 0.0547, + "step": 39235 + }, + { + "epoch": 0.9197536290487995, + "grad_norm": 0.37479761242866516, + "learning_rate": 3.162013474863945e-06, + "loss": 0.0488, + "step": 39236 + }, + { + "epoch": 0.9197770706236045, + "grad_norm": 0.365538090467453, + "learning_rate": 3.1601760260793423e-06, + "loss": 0.0723, + "step": 39237 + }, + { + "epoch": 0.9198005121984095, + "grad_norm": 0.15364007651805878, + "learning_rate": 3.158339102750363e-06, + "loss": 0.0179, + "step": 39238 + }, + { + "epoch": 0.9198239537732145, + "grad_norm": 0.33763793110847473, + "learning_rate": 3.1565027048869543e-06, + "loss": 0.0856, + "step": 39239 + }, + { + "epoch": 0.9198473953480195, + "grad_norm": 0.31597915291786194, + "learning_rate": 3.1546668324990756e-06, + "loss": 0.0439, + "step": 39240 + }, + { + "epoch": 0.9198708369228245, + "grad_norm": 0.3206937313079834, + "learning_rate": 3.152831485596708e-06, + "loss": 0.0231, + "step": 39241 + }, + { + "epoch": 0.9198942784976295, + "grad_norm": 0.4531386196613312, + "learning_rate": 3.1509966641897982e-06, + "loss": 0.0909, + "step": 39242 + }, + { + "epoch": 0.9199177200724344, + "grad_norm": 0.17244741320610046, + "learning_rate": 3.149162368288294e-06, + "loss": 0.012, + "step": 39243 + }, + { + "epoch": 0.9199411616472395, + "grad_norm": 0.15609151124954224, + "learning_rate": 3.147328597902166e-06, + "loss": 0.0212, + "step": 39244 + }, + { + "epoch": 0.9199646032220444, + "grad_norm": 0.29389095306396484, + "learning_rate": 3.1454953530413502e-06, + "loss": 0.0498, + "step": 39245 + }, + { + "epoch": 0.9199880447968495, + "grad_norm": 0.39357703924179077, + "learning_rate": 3.1436626337158046e-06, + "loss": 0.0412, + "step": 39246 + }, + { + "epoch": 0.9200114863716544, + "grad_norm": 0.5776379704475403, + "learning_rate": 3.141830439935478e-06, + "loss": 0.1009, + "step": 39247 + }, + { + "epoch": 0.9200349279464595, + "grad_norm": 0.05549630522727966, + "learning_rate": 3.139998771710284e-06, + "loss": 0.007, + "step": 39248 + }, + { + "epoch": 0.9200583695212644, + "grad_norm": 0.364096999168396, + "learning_rate": 3.1381676290501927e-06, + "loss": 0.0996, + "step": 39249 + }, + { + "epoch": 0.9200818110960695, + "grad_norm": 0.4900308847427368, + "learning_rate": 3.136337011965107e-06, + "loss": 0.5052, + "step": 39250 + }, + { + "epoch": 0.9201052526708744, + "grad_norm": 0.4605095684528351, + "learning_rate": 3.134506920465008e-06, + "loss": 0.0172, + "step": 39251 + }, + { + "epoch": 0.9201286942456794, + "grad_norm": 0.3757382035255432, + "learning_rate": 3.1326773545597875e-06, + "loss": 0.0451, + "step": 39252 + }, + { + "epoch": 0.9201521358204844, + "grad_norm": 0.6501836180686951, + "learning_rate": 3.1308483142593825e-06, + "loss": 0.0682, + "step": 39253 + }, + { + "epoch": 0.9201755773952894, + "grad_norm": 0.21143032610416412, + "learning_rate": 3.1290197995737182e-06, + "loss": 0.0307, + "step": 39254 + }, + { + "epoch": 0.9201990189700944, + "grad_norm": 0.27488481998443604, + "learning_rate": 3.1271918105127087e-06, + "loss": 0.0416, + "step": 39255 + }, + { + "epoch": 0.9202224605448994, + "grad_norm": 0.2185484915971756, + "learning_rate": 3.1253643470862903e-06, + "loss": 0.0333, + "step": 39256 + }, + { + "epoch": 0.9202459021197044, + "grad_norm": 0.5490623116493225, + "learning_rate": 3.1235374093043668e-06, + "loss": 0.0893, + "step": 39257 + }, + { + "epoch": 0.9202693436945094, + "grad_norm": 0.4078299403190613, + "learning_rate": 3.121710997176863e-06, + "loss": 0.0679, + "step": 39258 + }, + { + "epoch": 0.9202927852693144, + "grad_norm": 0.371309369802475, + "learning_rate": 3.119885110713672e-06, + "loss": 0.0537, + "step": 39259 + }, + { + "epoch": 0.9203162268441194, + "grad_norm": 0.1875101774930954, + "learning_rate": 3.1180597499247066e-06, + "loss": 0.0264, + "step": 39260 + }, + { + "epoch": 0.9203396684189243, + "grad_norm": 0.35953789949417114, + "learning_rate": 3.11623491481986e-06, + "loss": 0.0723, + "step": 39261 + }, + { + "epoch": 0.9203631099937294, + "grad_norm": 0.41330254077911377, + "learning_rate": 3.1144106054090572e-06, + "loss": 0.0535, + "step": 39262 + }, + { + "epoch": 0.9203865515685343, + "grad_norm": 0.1596290022134781, + "learning_rate": 3.1125868217021903e-06, + "loss": 0.0301, + "step": 39263 + }, + { + "epoch": 0.9204099931433394, + "grad_norm": 0.12913894653320312, + "learning_rate": 3.1107635637091517e-06, + "loss": 0.0177, + "step": 39264 + }, + { + "epoch": 0.9204334347181443, + "grad_norm": 0.7730231285095215, + "learning_rate": 3.108940831439833e-06, + "loss": 0.1328, + "step": 39265 + }, + { + "epoch": 0.9204568762929494, + "grad_norm": 0.409515380859375, + "learning_rate": 3.107118624904126e-06, + "loss": 0.0414, + "step": 39266 + }, + { + "epoch": 0.9204803178677543, + "grad_norm": 0.18914254009723663, + "learning_rate": 3.105296944111924e-06, + "loss": 0.0429, + "step": 39267 + }, + { + "epoch": 0.9205037594425594, + "grad_norm": 0.1443643420934677, + "learning_rate": 3.1034757890730847e-06, + "loss": 0.0174, + "step": 39268 + }, + { + "epoch": 0.9205272010173643, + "grad_norm": 0.33245840668678284, + "learning_rate": 3.101655159797534e-06, + "loss": 0.0759, + "step": 39269 + }, + { + "epoch": 0.9205506425921693, + "grad_norm": 0.12262387573719025, + "learning_rate": 3.0998350562951085e-06, + "loss": 0.0114, + "step": 39270 + }, + { + "epoch": 0.9205740841669743, + "grad_norm": 0.14118672907352448, + "learning_rate": 3.098015478575722e-06, + "loss": 0.0224, + "step": 39271 + }, + { + "epoch": 0.9205975257417793, + "grad_norm": 0.36532333493232727, + "learning_rate": 3.0961964266492114e-06, + "loss": 0.0723, + "step": 39272 + }, + { + "epoch": 0.9206209673165844, + "grad_norm": 0.27268657088279724, + "learning_rate": 3.0943779005254805e-06, + "loss": 0.0533, + "step": 39273 + }, + { + "epoch": 0.9206444088913893, + "grad_norm": 0.5818398594856262, + "learning_rate": 3.0925599002143536e-06, + "loss": 0.1236, + "step": 39274 + }, + { + "epoch": 0.9206678504661944, + "grad_norm": 0.5421985387802124, + "learning_rate": 3.0907424257257344e-06, + "loss": 0.057, + "step": 39275 + }, + { + "epoch": 0.9206912920409993, + "grad_norm": 0.5218101739883423, + "learning_rate": 3.088925477069471e-06, + "loss": 0.1108, + "step": 39276 + }, + { + "epoch": 0.9207147336158044, + "grad_norm": 0.6926719546318054, + "learning_rate": 3.0871090542554326e-06, + "loss": 0.0815, + "step": 39277 + }, + { + "epoch": 0.9207381751906093, + "grad_norm": 0.8460837006568909, + "learning_rate": 3.0852931572934562e-06, + "loss": 0.1082, + "step": 39278 + }, + { + "epoch": 0.9207616167654143, + "grad_norm": 0.1541641652584076, + "learning_rate": 3.083477786193412e-06, + "loss": 0.0233, + "step": 39279 + }, + { + "epoch": 0.9207850583402193, + "grad_norm": 0.24266639351844788, + "learning_rate": 3.0816629409651355e-06, + "loss": 0.0366, + "step": 39280 + }, + { + "epoch": 0.9208084999150243, + "grad_norm": 0.250195175409317, + "learning_rate": 3.0798486216184753e-06, + "loss": 0.0229, + "step": 39281 + }, + { + "epoch": 0.9208319414898293, + "grad_norm": 0.4512886106967926, + "learning_rate": 3.07803482816329e-06, + "loss": 0.6419, + "step": 39282 + }, + { + "epoch": 0.9208553830646343, + "grad_norm": 0.4625447690486908, + "learning_rate": 3.076221560609416e-06, + "loss": 0.0871, + "step": 39283 + }, + { + "epoch": 0.9208788246394393, + "grad_norm": 0.19287274777889252, + "learning_rate": 3.0744088189666897e-06, + "loss": 0.0113, + "step": 39284 + }, + { + "epoch": 0.9209022662142443, + "grad_norm": 0.1749257594347, + "learning_rate": 3.072596603244937e-06, + "loss": 0.0232, + "step": 39285 + }, + { + "epoch": 0.9209257077890493, + "grad_norm": 0.23143894970417023, + "learning_rate": 3.0707849134539945e-06, + "loss": 0.0481, + "step": 39286 + }, + { + "epoch": 0.9209491493638543, + "grad_norm": 0.20938532054424286, + "learning_rate": 3.0689737496037208e-06, + "loss": 0.0358, + "step": 39287 + }, + { + "epoch": 0.9209725909386592, + "grad_norm": 0.5716761946678162, + "learning_rate": 3.067163111703897e-06, + "loss": 0.0973, + "step": 39288 + }, + { + "epoch": 0.9209960325134643, + "grad_norm": 0.17387256026268005, + "learning_rate": 3.0653529997643814e-06, + "loss": 0.035, + "step": 39289 + }, + { + "epoch": 0.9210194740882692, + "grad_norm": 0.48492521047592163, + "learning_rate": 3.0635434137949894e-06, + "loss": 0.0675, + "step": 39290 + }, + { + "epoch": 0.9210429156630743, + "grad_norm": 0.27301159501075745, + "learning_rate": 3.0617343538055455e-06, + "loss": 0.0492, + "step": 39291 + }, + { + "epoch": 0.9210663572378792, + "grad_norm": 0.5674002766609192, + "learning_rate": 3.0599258198058425e-06, + "loss": 0.5203, + "step": 39292 + }, + { + "epoch": 0.9210897988126843, + "grad_norm": 0.39115503430366516, + "learning_rate": 3.0581178118057053e-06, + "loss": 0.6681, + "step": 39293 + }, + { + "epoch": 0.9211132403874892, + "grad_norm": 0.24875688552856445, + "learning_rate": 3.05631032981496e-06, + "loss": 0.0484, + "step": 39294 + }, + { + "epoch": 0.9211366819622943, + "grad_norm": 0.5659935474395752, + "learning_rate": 3.054503373843387e-06, + "loss": 0.1067, + "step": 39295 + }, + { + "epoch": 0.9211601235370992, + "grad_norm": 0.1968740075826645, + "learning_rate": 3.0526969439008124e-06, + "loss": 0.036, + "step": 39296 + }, + { + "epoch": 0.9211835651119042, + "grad_norm": 0.7511557936668396, + "learning_rate": 3.050891039997039e-06, + "loss": 0.1049, + "step": 39297 + }, + { + "epoch": 0.9212070066867092, + "grad_norm": 0.5007676482200623, + "learning_rate": 3.0490856621418374e-06, + "loss": 0.0611, + "step": 39298 + }, + { + "epoch": 0.9212304482615142, + "grad_norm": 0.37903228402137756, + "learning_rate": 3.047280810345032e-06, + "loss": 0.0692, + "step": 39299 + }, + { + "epoch": 0.9212538898363192, + "grad_norm": 0.3350837528705597, + "learning_rate": 3.045476484616405e-06, + "loss": 0.0851, + "step": 39300 + }, + { + "epoch": 0.9212773314111242, + "grad_norm": 0.5668811202049255, + "learning_rate": 3.043672684965748e-06, + "loss": 0.0403, + "step": 39301 + }, + { + "epoch": 0.9213007729859292, + "grad_norm": 0.2752213776111603, + "learning_rate": 3.0418694114028533e-06, + "loss": 0.0324, + "step": 39302 + }, + { + "epoch": 0.9213242145607342, + "grad_norm": 0.29966971278190613, + "learning_rate": 3.0400666639375018e-06, + "loss": 0.0479, + "step": 39303 + }, + { + "epoch": 0.9213476561355392, + "grad_norm": 0.24354544281959534, + "learning_rate": 3.038264442579475e-06, + "loss": 0.0433, + "step": 39304 + }, + { + "epoch": 0.9213710977103442, + "grad_norm": 0.5419999361038208, + "learning_rate": 3.0364627473385533e-06, + "loss": 0.0869, + "step": 39305 + }, + { + "epoch": 0.9213945392851491, + "grad_norm": 0.5521354675292969, + "learning_rate": 3.034661578224507e-06, + "loss": 0.5468, + "step": 39306 + }, + { + "epoch": 0.9214179808599542, + "grad_norm": 0.5792531967163086, + "learning_rate": 3.032860935247106e-06, + "loss": 0.0809, + "step": 39307 + }, + { + "epoch": 0.9214414224347591, + "grad_norm": 0.20794616639614105, + "learning_rate": 3.0310608184161427e-06, + "loss": 0.0391, + "step": 39308 + }, + { + "epoch": 0.9214648640095642, + "grad_norm": 0.13631290197372437, + "learning_rate": 3.0292612277413648e-06, + "loss": 0.0209, + "step": 39309 + }, + { + "epoch": 0.9214883055843691, + "grad_norm": 0.155295729637146, + "learning_rate": 3.0274621632325308e-06, + "loss": 0.0198, + "step": 39310 + }, + { + "epoch": 0.9215117471591742, + "grad_norm": 0.20863088965415955, + "learning_rate": 3.025663624899433e-06, + "loss": 0.021, + "step": 39311 + }, + { + "epoch": 0.9215351887339791, + "grad_norm": 0.4859173595905304, + "learning_rate": 3.0238656127517974e-06, + "loss": 0.1188, + "step": 39312 + }, + { + "epoch": 0.9215586303087842, + "grad_norm": 0.3358803391456604, + "learning_rate": 3.0220681267993933e-06, + "loss": 0.0383, + "step": 39313 + }, + { + "epoch": 0.9215820718835891, + "grad_norm": 0.6482238173484802, + "learning_rate": 3.02027116705198e-06, + "loss": 0.056, + "step": 39314 + }, + { + "epoch": 0.9216055134583941, + "grad_norm": 0.6817457675933838, + "learning_rate": 3.0184747335193054e-06, + "loss": 0.6835, + "step": 39315 + }, + { + "epoch": 0.9216289550331991, + "grad_norm": 0.5848063230514526, + "learning_rate": 3.0166788262111055e-06, + "loss": 0.0939, + "step": 39316 + }, + { + "epoch": 0.9216523966080041, + "grad_norm": 0.5284278988838196, + "learning_rate": 3.01488344513714e-06, + "loss": 0.5419, + "step": 39317 + }, + { + "epoch": 0.9216758381828091, + "grad_norm": 0.4150436818599701, + "learning_rate": 3.013088590307145e-06, + "loss": 0.0466, + "step": 39318 + }, + { + "epoch": 0.9216992797576141, + "grad_norm": 0.06659258157014847, + "learning_rate": 3.011294261730846e-06, + "loss": 0.0121, + "step": 39319 + }, + { + "epoch": 0.9217227213324191, + "grad_norm": 0.24550485610961914, + "learning_rate": 3.009500459418002e-06, + "loss": 0.0271, + "step": 39320 + }, + { + "epoch": 0.9217461629072241, + "grad_norm": 0.36878105998039246, + "learning_rate": 3.0077071833783386e-06, + "loss": 0.0357, + "step": 39321 + }, + { + "epoch": 0.921769604482029, + "grad_norm": 0.2065865397453308, + "learning_rate": 3.0059144336215706e-06, + "loss": 0.0393, + "step": 39322 + }, + { + "epoch": 0.9217930460568341, + "grad_norm": 0.5684879422187805, + "learning_rate": 3.0041222101574453e-06, + "loss": 0.0392, + "step": 39323 + }, + { + "epoch": 0.9218164876316391, + "grad_norm": 0.42433983087539673, + "learning_rate": 3.002330512995688e-06, + "loss": 0.09, + "step": 39324 + }, + { + "epoch": 0.9218399292064441, + "grad_norm": 0.11279015243053436, + "learning_rate": 3.0005393421460027e-06, + "loss": 0.0201, + "step": 39325 + }, + { + "epoch": 0.9218633707812491, + "grad_norm": 0.32229501008987427, + "learning_rate": 2.9987486976181144e-06, + "loss": 0.0331, + "step": 39326 + }, + { + "epoch": 0.9218868123560541, + "grad_norm": 0.15191899240016937, + "learning_rate": 2.996958579421749e-06, + "loss": 0.0196, + "step": 39327 + }, + { + "epoch": 0.9219102539308591, + "grad_norm": 0.500586748123169, + "learning_rate": 2.995168987566621e-06, + "loss": 0.0885, + "step": 39328 + }, + { + "epoch": 0.9219336955056641, + "grad_norm": 0.5003471970558167, + "learning_rate": 2.9933799220624338e-06, + "loss": 0.0728, + "step": 39329 + }, + { + "epoch": 0.9219571370804691, + "grad_norm": 0.5635104179382324, + "learning_rate": 2.9915913829188903e-06, + "loss": 0.0595, + "step": 39330 + }, + { + "epoch": 0.921980578655274, + "grad_norm": 0.14102791249752045, + "learning_rate": 2.9898033701456828e-06, + "loss": 0.0171, + "step": 39331 + }, + { + "epoch": 0.9220040202300791, + "grad_norm": 0.03132752701640129, + "learning_rate": 2.9880158837525487e-06, + "loss": 0.0029, + "step": 39332 + }, + { + "epoch": 0.922027461804884, + "grad_norm": 0.3978639245033264, + "learning_rate": 2.986228923749168e-06, + "loss": 0.0905, + "step": 39333 + }, + { + "epoch": 0.9220509033796891, + "grad_norm": 0.46075645089149475, + "learning_rate": 2.984442490145223e-06, + "loss": 0.109, + "step": 39334 + }, + { + "epoch": 0.922074344954494, + "grad_norm": 0.20490045845508575, + "learning_rate": 2.982656582950427e-06, + "loss": 0.0224, + "step": 39335 + }, + { + "epoch": 0.9220977865292991, + "grad_norm": 0.14964133501052856, + "learning_rate": 2.980871202174473e-06, + "loss": 0.0442, + "step": 39336 + }, + { + "epoch": 0.922121228104104, + "grad_norm": 0.6611615419387817, + "learning_rate": 2.9790863478270427e-06, + "loss": 0.1138, + "step": 39337 + }, + { + "epoch": 0.9221446696789091, + "grad_norm": 0.2645050585269928, + "learning_rate": 2.9773020199177937e-06, + "loss": 0.0285, + "step": 39338 + }, + { + "epoch": 0.922168111253714, + "grad_norm": 0.26762932538986206, + "learning_rate": 2.9755182184564524e-06, + "loss": 0.0183, + "step": 39339 + }, + { + "epoch": 0.9221915528285191, + "grad_norm": 0.3631843626499176, + "learning_rate": 2.9737349434526775e-06, + "loss": 0.0412, + "step": 39340 + }, + { + "epoch": 0.922214994403324, + "grad_norm": 0.46688324213027954, + "learning_rate": 2.9719521949161387e-06, + "loss": 0.0332, + "step": 39341 + }, + { + "epoch": 0.922238435978129, + "grad_norm": 0.5144544839859009, + "learning_rate": 2.970169972856529e-06, + "loss": 0.0685, + "step": 39342 + }, + { + "epoch": 0.922261877552934, + "grad_norm": 0.28696009516716003, + "learning_rate": 2.968388277283496e-06, + "loss": 0.0677, + "step": 39343 + }, + { + "epoch": 0.922285319127739, + "grad_norm": 0.6485589742660522, + "learning_rate": 2.966607108206698e-06, + "loss": 0.0726, + "step": 39344 + }, + { + "epoch": 0.922308760702544, + "grad_norm": 0.6314916610717773, + "learning_rate": 2.9648264656358394e-06, + "loss": 0.1309, + "step": 39345 + }, + { + "epoch": 0.922332202277349, + "grad_norm": 0.11683528125286102, + "learning_rate": 2.963046349580556e-06, + "loss": 0.022, + "step": 39346 + }, + { + "epoch": 0.922355643852154, + "grad_norm": 0.22382792830467224, + "learning_rate": 2.9612667600505073e-06, + "loss": 0.0541, + "step": 39347 + }, + { + "epoch": 0.922379085426959, + "grad_norm": 0.5601843595504761, + "learning_rate": 2.959487697055352e-06, + "loss": 0.0619, + "step": 39348 + }, + { + "epoch": 0.922402527001764, + "grad_norm": 0.4081330895423889, + "learning_rate": 2.95770916060476e-06, + "loss": 0.091, + "step": 39349 + }, + { + "epoch": 0.922425968576569, + "grad_norm": 0.5225288271903992, + "learning_rate": 2.9559311507083576e-06, + "loss": 0.0768, + "step": 39350 + }, + { + "epoch": 0.9224494101513739, + "grad_norm": 0.08154336363077164, + "learning_rate": 2.954153667375792e-06, + "loss": 0.0095, + "step": 39351 + }, + { + "epoch": 0.922472851726179, + "grad_norm": 0.2767276167869568, + "learning_rate": 2.952376710616733e-06, + "loss": 0.0288, + "step": 39352 + }, + { + "epoch": 0.9224962933009839, + "grad_norm": 0.2973630428314209, + "learning_rate": 2.9506002804407963e-06, + "loss": 0.4204, + "step": 39353 + }, + { + "epoch": 0.922519734875789, + "grad_norm": 0.2837291955947876, + "learning_rate": 2.9488243768576394e-06, + "loss": 0.037, + "step": 39354 + }, + { + "epoch": 0.9225431764505939, + "grad_norm": 0.5005913972854614, + "learning_rate": 2.947048999876889e-06, + "loss": 0.574, + "step": 39355 + }, + { + "epoch": 0.922566618025399, + "grad_norm": 0.4846700429916382, + "learning_rate": 2.9452741495081814e-06, + "loss": 0.1004, + "step": 39356 + }, + { + "epoch": 0.9225900596002039, + "grad_norm": 0.5148449540138245, + "learning_rate": 2.943499825761131e-06, + "loss": 0.0767, + "step": 39357 + }, + { + "epoch": 0.922613501175009, + "grad_norm": 0.42098018527030945, + "learning_rate": 2.9417260286453863e-06, + "loss": 0.059, + "step": 39358 + }, + { + "epoch": 0.9226369427498139, + "grad_norm": 0.17378436028957367, + "learning_rate": 2.939952758170561e-06, + "loss": 0.0352, + "step": 39359 + }, + { + "epoch": 0.9226603843246189, + "grad_norm": 0.12788544595241547, + "learning_rate": 2.9381800143462923e-06, + "loss": 0.0209, + "step": 39360 + }, + { + "epoch": 0.9226838258994239, + "grad_norm": 0.1253773421049118, + "learning_rate": 2.9364077971821834e-06, + "loss": 0.0193, + "step": 39361 + }, + { + "epoch": 0.9227072674742289, + "grad_norm": 0.3461059033870697, + "learning_rate": 2.9346361066878604e-06, + "loss": 0.053, + "step": 39362 + }, + { + "epoch": 0.9227307090490339, + "grad_norm": 0.45584121346473694, + "learning_rate": 2.932864942872926e-06, + "loss": 0.0553, + "step": 39363 + }, + { + "epoch": 0.9227541506238389, + "grad_norm": 0.4484628438949585, + "learning_rate": 2.931094305746984e-06, + "loss": 0.0987, + "step": 39364 + }, + { + "epoch": 0.9227775921986439, + "grad_norm": 0.3553599417209625, + "learning_rate": 2.9293241953196713e-06, + "loss": 0.0629, + "step": 39365 + }, + { + "epoch": 0.9228010337734489, + "grad_norm": 0.4360114336013794, + "learning_rate": 2.927554611600569e-06, + "loss": 0.0432, + "step": 39366 + }, + { + "epoch": 0.9228244753482538, + "grad_norm": 0.37991994619369507, + "learning_rate": 2.925785554599292e-06, + "loss": 0.0345, + "step": 39367 + }, + { + "epoch": 0.9228479169230589, + "grad_norm": 0.17821846902370453, + "learning_rate": 2.9240170243254205e-06, + "loss": 0.0503, + "step": 39368 + }, + { + "epoch": 0.9228713584978638, + "grad_norm": 0.332600861787796, + "learning_rate": 2.922249020788559e-06, + "loss": 0.0482, + "step": 39369 + }, + { + "epoch": 0.9228948000726689, + "grad_norm": 0.43875405192375183, + "learning_rate": 2.9204815439983104e-06, + "loss": 0.044, + "step": 39370 + }, + { + "epoch": 0.9229182416474738, + "grad_norm": 0.44957205653190613, + "learning_rate": 2.9187145939642555e-06, + "loss": 0.0357, + "step": 39371 + }, + { + "epoch": 0.9229416832222789, + "grad_norm": 0.27451080083847046, + "learning_rate": 2.916948170695988e-06, + "loss": 0.0579, + "step": 39372 + }, + { + "epoch": 0.9229651247970838, + "grad_norm": 0.28044000267982483, + "learning_rate": 2.915182274203099e-06, + "loss": 0.0369, + "step": 39373 + }, + { + "epoch": 0.9229885663718889, + "grad_norm": 0.10484082251787186, + "learning_rate": 2.9134169044951478e-06, + "loss": 0.0166, + "step": 39374 + }, + { + "epoch": 0.9230120079466939, + "grad_norm": 0.454224556684494, + "learning_rate": 2.9116520615817376e-06, + "loss": 0.085, + "step": 39375 + }, + { + "epoch": 0.9230354495214989, + "grad_norm": 0.34046897292137146, + "learning_rate": 2.909887745472417e-06, + "loss": 0.0411, + "step": 39376 + }, + { + "epoch": 0.9230588910963039, + "grad_norm": 0.13125762343406677, + "learning_rate": 2.908123956176789e-06, + "loss": 0.0229, + "step": 39377 + }, + { + "epoch": 0.9230823326711088, + "grad_norm": 0.39920058846473694, + "learning_rate": 2.9063606937044018e-06, + "loss": 0.0673, + "step": 39378 + }, + { + "epoch": 0.9231057742459139, + "grad_norm": 0.3294682800769806, + "learning_rate": 2.9045979580648362e-06, + "loss": 0.0616, + "step": 39379 + }, + { + "epoch": 0.9231292158207188, + "grad_norm": 0.19221149384975433, + "learning_rate": 2.9028357492676628e-06, + "loss": 0.0331, + "step": 39380 + }, + { + "epoch": 0.9231526573955239, + "grad_norm": 0.2739986479282379, + "learning_rate": 2.901074067322418e-06, + "loss": 0.0445, + "step": 39381 + }, + { + "epoch": 0.9231760989703288, + "grad_norm": 0.37837469577789307, + "learning_rate": 2.899312912238672e-06, + "loss": 0.0386, + "step": 39382 + }, + { + "epoch": 0.9231995405451339, + "grad_norm": 0.09429330378770828, + "learning_rate": 2.8975522840259837e-06, + "loss": 0.0136, + "step": 39383 + }, + { + "epoch": 0.9232229821199388, + "grad_norm": 0.5201423764228821, + "learning_rate": 2.8957921826939128e-06, + "loss": 0.1268, + "step": 39384 + }, + { + "epoch": 0.9232464236947439, + "grad_norm": 0.3334914743900299, + "learning_rate": 2.8940326082520176e-06, + "loss": 0.0636, + "step": 39385 + }, + { + "epoch": 0.9232698652695488, + "grad_norm": 0.23822295665740967, + "learning_rate": 2.892273560709813e-06, + "loss": 0.1832, + "step": 39386 + }, + { + "epoch": 0.9232933068443538, + "grad_norm": 0.5780718326568604, + "learning_rate": 2.8905150400768797e-06, + "loss": 0.0633, + "step": 39387 + }, + { + "epoch": 0.9233167484191588, + "grad_norm": 0.6181434392929077, + "learning_rate": 2.8887570463627334e-06, + "loss": 0.0779, + "step": 39388 + }, + { + "epoch": 0.9233401899939638, + "grad_norm": 0.13474786281585693, + "learning_rate": 2.8869995795769102e-06, + "loss": 0.0141, + "step": 39389 + }, + { + "epoch": 0.9233636315687688, + "grad_norm": 0.29509851336479187, + "learning_rate": 2.885242639728969e-06, + "loss": 0.0687, + "step": 39390 + }, + { + "epoch": 0.9233870731435738, + "grad_norm": 0.3588877320289612, + "learning_rate": 2.8834862268284357e-06, + "loss": 0.0522, + "step": 39391 + }, + { + "epoch": 0.9234105147183788, + "grad_norm": 0.2626684010028839, + "learning_rate": 2.8817303408848363e-06, + "loss": 0.0648, + "step": 39392 + }, + { + "epoch": 0.9234339562931838, + "grad_norm": 0.29829660058021545, + "learning_rate": 2.879974981907696e-06, + "loss": 0.0436, + "step": 39393 + }, + { + "epoch": 0.9234573978679887, + "grad_norm": 0.13479746878147125, + "learning_rate": 2.8782201499065408e-06, + "loss": 0.0299, + "step": 39394 + }, + { + "epoch": 0.9234808394427938, + "grad_norm": 0.5866191983222961, + "learning_rate": 2.8764658448908856e-06, + "loss": 0.0791, + "step": 39395 + }, + { + "epoch": 0.9235042810175987, + "grad_norm": 0.5098704099655151, + "learning_rate": 2.8747120668702553e-06, + "loss": 0.0876, + "step": 39396 + }, + { + "epoch": 0.9235277225924038, + "grad_norm": 0.3921889364719391, + "learning_rate": 2.8729588158541875e-06, + "loss": 0.0907, + "step": 39397 + }, + { + "epoch": 0.9235511641672087, + "grad_norm": 0.2624662220478058, + "learning_rate": 2.8712060918521633e-06, + "loss": 0.0248, + "step": 39398 + }, + { + "epoch": 0.9235746057420138, + "grad_norm": 0.4605185091495514, + "learning_rate": 2.869453894873719e-06, + "loss": 0.0595, + "step": 39399 + }, + { + "epoch": 0.9235980473168187, + "grad_norm": 0.6623390913009644, + "learning_rate": 2.8677022249283368e-06, + "loss": 0.0866, + "step": 39400 + }, + { + "epoch": 0.9236214888916238, + "grad_norm": 0.13114677369594574, + "learning_rate": 2.865951082025542e-06, + "loss": 0.0255, + "step": 39401 + }, + { + "epoch": 0.9236449304664287, + "grad_norm": 0.10308866947889328, + "learning_rate": 2.8642004661748156e-06, + "loss": 0.02, + "step": 39402 + }, + { + "epoch": 0.9236683720412338, + "grad_norm": 0.2641230523586273, + "learning_rate": 2.8624503773856727e-06, + "loss": 0.0591, + "step": 39403 + }, + { + "epoch": 0.9236918136160387, + "grad_norm": 0.21865251660346985, + "learning_rate": 2.860700815667616e-06, + "loss": 0.0387, + "step": 39404 + }, + { + "epoch": 0.9237152551908437, + "grad_norm": 0.6043695211410522, + "learning_rate": 2.858951781030117e-06, + "loss": 0.1443, + "step": 39405 + }, + { + "epoch": 0.9237386967656487, + "grad_norm": 0.2737395465373993, + "learning_rate": 2.857203273482689e-06, + "loss": 0.0432, + "step": 39406 + }, + { + "epoch": 0.9237621383404537, + "grad_norm": 0.4088844954967499, + "learning_rate": 2.8554552930347812e-06, + "loss": 0.0571, + "step": 39407 + }, + { + "epoch": 0.9237855799152587, + "grad_norm": 0.26702889800071716, + "learning_rate": 2.853707839695918e-06, + "loss": 0.0654, + "step": 39408 + }, + { + "epoch": 0.9238090214900637, + "grad_norm": 0.27356186509132385, + "learning_rate": 2.851960913475571e-06, + "loss": 0.0683, + "step": 39409 + }, + { + "epoch": 0.9238324630648687, + "grad_norm": 0.09980619698762894, + "learning_rate": 2.8502145143832203e-06, + "loss": 0.0115, + "step": 39410 + }, + { + "epoch": 0.9238559046396737, + "grad_norm": 0.13102030754089355, + "learning_rate": 2.848468642428337e-06, + "loss": 0.0283, + "step": 39411 + }, + { + "epoch": 0.9238793462144786, + "grad_norm": 0.2512209713459015, + "learning_rate": 2.8467232976203906e-06, + "loss": 0.0276, + "step": 39412 + }, + { + "epoch": 0.9239027877892837, + "grad_norm": 0.4269753396511078, + "learning_rate": 2.844978479968863e-06, + "loss": 0.0709, + "step": 39413 + }, + { + "epoch": 0.9239262293640886, + "grad_norm": 0.1403827965259552, + "learning_rate": 2.8432341894831903e-06, + "loss": 0.0297, + "step": 39414 + }, + { + "epoch": 0.9239496709388937, + "grad_norm": 0.1878093183040619, + "learning_rate": 2.841490426172877e-06, + "loss": 0.0156, + "step": 39415 + }, + { + "epoch": 0.9239731125136986, + "grad_norm": 0.5083283185958862, + "learning_rate": 2.8397471900473703e-06, + "loss": 0.4973, + "step": 39416 + }, + { + "epoch": 0.9239965540885037, + "grad_norm": 0.14133991301059723, + "learning_rate": 2.8380044811161297e-06, + "loss": 0.0182, + "step": 39417 + }, + { + "epoch": 0.9240199956633086, + "grad_norm": 0.46600547432899475, + "learning_rate": 2.8362622993886145e-06, + "loss": 0.046, + "step": 39418 + }, + { + "epoch": 0.9240434372381137, + "grad_norm": 0.2553969919681549, + "learning_rate": 2.83452064487425e-06, + "loss": 0.0461, + "step": 39419 + }, + { + "epoch": 0.9240668788129186, + "grad_norm": 0.5398806929588318, + "learning_rate": 2.8327795175825176e-06, + "loss": 0.637, + "step": 39420 + }, + { + "epoch": 0.9240903203877237, + "grad_norm": 0.5700675249099731, + "learning_rate": 2.8310389175228545e-06, + "loss": 0.5776, + "step": 39421 + }, + { + "epoch": 0.9241137619625286, + "grad_norm": 0.42416369915008545, + "learning_rate": 2.8292988447047088e-06, + "loss": 0.0886, + "step": 39422 + }, + { + "epoch": 0.9241372035373336, + "grad_norm": 0.22940005362033844, + "learning_rate": 2.8275592991375278e-06, + "loss": 0.0407, + "step": 39423 + }, + { + "epoch": 0.9241606451121386, + "grad_norm": 0.30444473028182983, + "learning_rate": 2.8258202808307377e-06, + "loss": 0.0229, + "step": 39424 + }, + { + "epoch": 0.9241840866869436, + "grad_norm": 0.7687051892280579, + "learning_rate": 2.8240817897937754e-06, + "loss": 0.0859, + "step": 39425 + }, + { + "epoch": 0.9242075282617487, + "grad_norm": 0.21329402923583984, + "learning_rate": 2.822343826036078e-06, + "loss": 0.0288, + "step": 39426 + }, + { + "epoch": 0.9242309698365536, + "grad_norm": 0.28680312633514404, + "learning_rate": 2.8206063895670707e-06, + "loss": 0.0353, + "step": 39427 + }, + { + "epoch": 0.9242544114113587, + "grad_norm": 0.6284136772155762, + "learning_rate": 2.8188694803961913e-06, + "loss": 0.1399, + "step": 39428 + }, + { + "epoch": 0.9242778529861636, + "grad_norm": 0.4493997395038605, + "learning_rate": 2.8171330985328647e-06, + "loss": 0.1249, + "step": 39429 + }, + { + "epoch": 0.9243012945609687, + "grad_norm": 0.3191496729850769, + "learning_rate": 2.8153972439865063e-06, + "loss": 0.0399, + "step": 39430 + }, + { + "epoch": 0.9243247361357736, + "grad_norm": 0.3037704527378082, + "learning_rate": 2.8136619167665303e-06, + "loss": 0.0619, + "step": 39431 + }, + { + "epoch": 0.9243481777105786, + "grad_norm": 0.46176183223724365, + "learning_rate": 2.8119271168823515e-06, + "loss": 0.0882, + "step": 39432 + }, + { + "epoch": 0.9243716192853836, + "grad_norm": 0.4157375395298004, + "learning_rate": 2.8101928443433957e-06, + "loss": 0.0561, + "step": 39433 + }, + { + "epoch": 0.9243950608601886, + "grad_norm": 0.3355998992919922, + "learning_rate": 2.8084590991590553e-06, + "loss": 0.3491, + "step": 39434 + }, + { + "epoch": 0.9244185024349936, + "grad_norm": 0.5456990003585815, + "learning_rate": 2.8067258813387563e-06, + "loss": 0.0978, + "step": 39435 + }, + { + "epoch": 0.9244419440097986, + "grad_norm": 0.48900800943374634, + "learning_rate": 2.8049931908919023e-06, + "loss": 0.0777, + "step": 39436 + }, + { + "epoch": 0.9244653855846036, + "grad_norm": 0.4636813700199127, + "learning_rate": 2.803261027827886e-06, + "loss": 0.1073, + "step": 39437 + }, + { + "epoch": 0.9244888271594086, + "grad_norm": 0.5374918580055237, + "learning_rate": 2.8015293921560994e-06, + "loss": 0.1091, + "step": 39438 + }, + { + "epoch": 0.9245122687342135, + "grad_norm": 0.6375210285186768, + "learning_rate": 2.799798283885946e-06, + "loss": 0.0695, + "step": 39439 + }, + { + "epoch": 0.9245357103090186, + "grad_norm": 0.3151394724845886, + "learning_rate": 2.79806770302683e-06, + "loss": 0.0489, + "step": 39440 + }, + { + "epoch": 0.9245591518838235, + "grad_norm": 0.17937901616096497, + "learning_rate": 2.796337649588132e-06, + "loss": 0.0155, + "step": 39441 + }, + { + "epoch": 0.9245825934586286, + "grad_norm": 0.736049473285675, + "learning_rate": 2.7946081235792344e-06, + "loss": 0.1431, + "step": 39442 + }, + { + "epoch": 0.9246060350334335, + "grad_norm": 0.5424992442131042, + "learning_rate": 2.7928791250095287e-06, + "loss": 0.6492, + "step": 39443 + }, + { + "epoch": 0.9246294766082386, + "grad_norm": 0.23978720605373383, + "learning_rate": 2.7911506538883856e-06, + "loss": 0.0278, + "step": 39444 + }, + { + "epoch": 0.9246529181830435, + "grad_norm": 0.12789663672447205, + "learning_rate": 2.789422710225209e-06, + "loss": 0.0097, + "step": 39445 + }, + { + "epoch": 0.9246763597578486, + "grad_norm": 0.4873284697532654, + "learning_rate": 2.7876952940293354e-06, + "loss": 0.0448, + "step": 39446 + }, + { + "epoch": 0.9246998013326535, + "grad_norm": 0.501322865486145, + "learning_rate": 2.78596840531018e-06, + "loss": 0.0784, + "step": 39447 + }, + { + "epoch": 0.9247232429074586, + "grad_norm": 0.6609266996383667, + "learning_rate": 2.78424204407709e-06, + "loss": 0.1213, + "step": 39448 + }, + { + "epoch": 0.9247466844822635, + "grad_norm": 0.09517239034175873, + "learning_rate": 2.7825162103394364e-06, + "loss": 0.0202, + "step": 39449 + }, + { + "epoch": 0.9247701260570685, + "grad_norm": 0.2759005129337311, + "learning_rate": 2.780790904106578e-06, + "loss": 0.0349, + "step": 39450 + }, + { + "epoch": 0.9247935676318735, + "grad_norm": 0.4962424039840698, + "learning_rate": 2.7790661253878857e-06, + "loss": 0.0941, + "step": 39451 + }, + { + "epoch": 0.9248170092066785, + "grad_norm": 0.6652401685714722, + "learning_rate": 2.777341874192707e-06, + "loss": 0.1546, + "step": 39452 + }, + { + "epoch": 0.9248404507814835, + "grad_norm": 0.4124377369880676, + "learning_rate": 2.7756181505304126e-06, + "loss": 0.0179, + "step": 39453 + }, + { + "epoch": 0.9248638923562885, + "grad_norm": 0.2345476895570755, + "learning_rate": 2.7738949544103497e-06, + "loss": 0.0205, + "step": 39454 + }, + { + "epoch": 0.9248873339310935, + "grad_norm": 0.5885857343673706, + "learning_rate": 2.7721722858418674e-06, + "loss": 0.0926, + "step": 39455 + }, + { + "epoch": 0.9249107755058985, + "grad_norm": 0.6025717258453369, + "learning_rate": 2.770450144834291e-06, + "loss": 0.0781, + "step": 39456 + }, + { + "epoch": 0.9249342170807034, + "grad_norm": 0.67494797706604, + "learning_rate": 2.7687285313970134e-06, + "loss": 0.1515, + "step": 39457 + }, + { + "epoch": 0.9249576586555085, + "grad_norm": 0.3249850869178772, + "learning_rate": 2.767007445539338e-06, + "loss": 0.0503, + "step": 39458 + }, + { + "epoch": 0.9249811002303134, + "grad_norm": 0.24291767179965973, + "learning_rate": 2.7652868872706016e-06, + "loss": 0.0323, + "step": 39459 + }, + { + "epoch": 0.9250045418051185, + "grad_norm": 0.17827388644218445, + "learning_rate": 2.7635668566001637e-06, + "loss": 0.0213, + "step": 39460 + }, + { + "epoch": 0.9250279833799234, + "grad_norm": 0.13735733926296234, + "learning_rate": 2.76184735353735e-06, + "loss": 0.0259, + "step": 39461 + }, + { + "epoch": 0.9250514249547285, + "grad_norm": 0.5652856826782227, + "learning_rate": 2.7601283780914866e-06, + "loss": 0.5628, + "step": 39462 + }, + { + "epoch": 0.9250748665295334, + "grad_norm": 0.38902172446250916, + "learning_rate": 2.7584099302718992e-06, + "loss": 0.3776, + "step": 39463 + }, + { + "epoch": 0.9250983081043385, + "grad_norm": 0.24775677919387817, + "learning_rate": 2.7566920100879134e-06, + "loss": 0.0355, + "step": 39464 + }, + { + "epoch": 0.9251217496791434, + "grad_norm": 0.38110268115997314, + "learning_rate": 2.7549746175488443e-06, + "loss": 0.0523, + "step": 39465 + }, + { + "epoch": 0.9251451912539485, + "grad_norm": 0.37633800506591797, + "learning_rate": 2.753257752664029e-06, + "loss": 0.0457, + "step": 39466 + }, + { + "epoch": 0.9251686328287534, + "grad_norm": 0.24148984253406525, + "learning_rate": 2.751541415442771e-06, + "loss": 0.019, + "step": 39467 + }, + { + "epoch": 0.9251920744035584, + "grad_norm": 0.1536143571138382, + "learning_rate": 2.749825605894363e-06, + "loss": 0.0405, + "step": 39468 + }, + { + "epoch": 0.9252155159783634, + "grad_norm": 0.38028326630592346, + "learning_rate": 2.7481103240281526e-06, + "loss": 0.0928, + "step": 39469 + }, + { + "epoch": 0.9252389575531684, + "grad_norm": 0.5204362273216248, + "learning_rate": 2.7463955698534327e-06, + "loss": 0.0699, + "step": 39470 + }, + { + "epoch": 0.9252623991279734, + "grad_norm": 0.30751389265060425, + "learning_rate": 2.744681343379507e-06, + "loss": 0.0423, + "step": 39471 + }, + { + "epoch": 0.9252858407027784, + "grad_norm": 0.45921358466148376, + "learning_rate": 2.742967644615657e-06, + "loss": 0.1246, + "step": 39472 + }, + { + "epoch": 0.9253092822775834, + "grad_norm": 0.533784031867981, + "learning_rate": 2.7412544735712088e-06, + "loss": 0.6776, + "step": 39473 + }, + { + "epoch": 0.9253327238523884, + "grad_norm": 0.3968193531036377, + "learning_rate": 2.7395418302554542e-06, + "loss": 0.0498, + "step": 39474 + }, + { + "epoch": 0.9253561654271933, + "grad_norm": 0.5662557482719421, + "learning_rate": 2.7378297146776753e-06, + "loss": 0.0373, + "step": 39475 + }, + { + "epoch": 0.9253796070019984, + "grad_norm": 0.34198784828186035, + "learning_rate": 2.736118126847176e-06, + "loss": 0.0499, + "step": 39476 + }, + { + "epoch": 0.9254030485768034, + "grad_norm": 0.4545915126800537, + "learning_rate": 2.7344070667732145e-06, + "loss": 0.1112, + "step": 39477 + }, + { + "epoch": 0.9254264901516084, + "grad_norm": 0.3548617660999298, + "learning_rate": 2.7326965344651177e-06, + "loss": 0.5858, + "step": 39478 + }, + { + "epoch": 0.9254499317264134, + "grad_norm": 0.40886518359184265, + "learning_rate": 2.7309865299321334e-06, + "loss": 0.0901, + "step": 39479 + }, + { + "epoch": 0.9254733733012184, + "grad_norm": 0.3752625584602356, + "learning_rate": 2.729277053183543e-06, + "loss": 0.0483, + "step": 39480 + }, + { + "epoch": 0.9254968148760234, + "grad_norm": 0.4990631341934204, + "learning_rate": 2.7275681042286395e-06, + "loss": 0.4299, + "step": 39481 + }, + { + "epoch": 0.9255202564508284, + "grad_norm": 0.5771206021308899, + "learning_rate": 2.7258596830766925e-06, + "loss": 0.1163, + "step": 39482 + }, + { + "epoch": 0.9255436980256334, + "grad_norm": 0.463737815618515, + "learning_rate": 2.724151789736962e-06, + "loss": 0.1096, + "step": 39483 + }, + { + "epoch": 0.9255671396004383, + "grad_norm": 0.7638242244720459, + "learning_rate": 2.722444424218706e-06, + "loss": 0.138, + "step": 39484 + }, + { + "epoch": 0.9255905811752434, + "grad_norm": 0.4654725193977356, + "learning_rate": 2.720737586531219e-06, + "loss": 0.4516, + "step": 39485 + }, + { + "epoch": 0.9256140227500483, + "grad_norm": 0.27717140316963196, + "learning_rate": 2.7190312766837365e-06, + "loss": 0.0451, + "step": 39486 + }, + { + "epoch": 0.9256374643248534, + "grad_norm": 0.3168349266052246, + "learning_rate": 2.7173254946855298e-06, + "loss": 0.0297, + "step": 39487 + }, + { + "epoch": 0.9256609058996583, + "grad_norm": 0.14698992669582367, + "learning_rate": 2.7156202405458463e-06, + "loss": 0.0299, + "step": 39488 + }, + { + "epoch": 0.9256843474744634, + "grad_norm": 0.31665700674057007, + "learning_rate": 2.713915514273946e-06, + "loss": 0.0227, + "step": 39489 + }, + { + "epoch": 0.9257077890492683, + "grad_norm": 0.299380898475647, + "learning_rate": 2.7122113158790763e-06, + "loss": 0.0187, + "step": 39490 + }, + { + "epoch": 0.9257312306240734, + "grad_norm": 0.6021139621734619, + "learning_rate": 2.7105076453704747e-06, + "loss": 0.804, + "step": 39491 + }, + { + "epoch": 0.9257546721988783, + "grad_norm": 0.12066750973463058, + "learning_rate": 2.7088045027574006e-06, + "loss": 0.0155, + "step": 39492 + }, + { + "epoch": 0.9257781137736834, + "grad_norm": 0.30566614866256714, + "learning_rate": 2.7071018880490906e-06, + "loss": 0.0272, + "step": 39493 + }, + { + "epoch": 0.9258015553484883, + "grad_norm": 0.35903653502464294, + "learning_rate": 2.705399801254782e-06, + "loss": 0.0434, + "step": 39494 + }, + { + "epoch": 0.9258249969232933, + "grad_norm": 0.37926679849624634, + "learning_rate": 2.703698242383712e-06, + "loss": 0.0458, + "step": 39495 + }, + { + "epoch": 0.9258484384980983, + "grad_norm": 0.7374638319015503, + "learning_rate": 2.7019972114451064e-06, + "loss": 0.1523, + "step": 39496 + }, + { + "epoch": 0.9258718800729033, + "grad_norm": 0.5086217522621155, + "learning_rate": 2.700296708448191e-06, + "loss": 0.1049, + "step": 39497 + }, + { + "epoch": 0.9258953216477083, + "grad_norm": 0.0753820389509201, + "learning_rate": 2.6985967334022142e-06, + "loss": 0.0121, + "step": 39498 + }, + { + "epoch": 0.9259187632225133, + "grad_norm": 0.6300192475318909, + "learning_rate": 2.69689728631638e-06, + "loss": 0.0631, + "step": 39499 + }, + { + "epoch": 0.9259422047973183, + "grad_norm": 0.3264181315898895, + "learning_rate": 2.695198367199925e-06, + "loss": 0.0504, + "step": 39500 + }, + { + "epoch": 0.9259656463721233, + "grad_norm": 0.3196430802345276, + "learning_rate": 2.693499976062064e-06, + "loss": 0.0567, + "step": 39501 + }, + { + "epoch": 0.9259890879469282, + "grad_norm": 0.4939504861831665, + "learning_rate": 2.6918021129120006e-06, + "loss": 0.0796, + "step": 39502 + }, + { + "epoch": 0.9260125295217333, + "grad_norm": 0.12458141893148422, + "learning_rate": 2.6901047777589396e-06, + "loss": 0.0126, + "step": 39503 + }, + { + "epoch": 0.9260359710965382, + "grad_norm": 0.13858568668365479, + "learning_rate": 2.6884079706121277e-06, + "loss": 0.0162, + "step": 39504 + }, + { + "epoch": 0.9260594126713433, + "grad_norm": 0.21814361214637756, + "learning_rate": 2.6867116914807367e-06, + "loss": 0.0144, + "step": 39505 + }, + { + "epoch": 0.9260828542461482, + "grad_norm": 0.7827123403549194, + "learning_rate": 2.6850159403739917e-06, + "loss": 0.1481, + "step": 39506 + }, + { + "epoch": 0.9261062958209533, + "grad_norm": 0.17367233335971832, + "learning_rate": 2.6833207173010854e-06, + "loss": 0.0195, + "step": 39507 + }, + { + "epoch": 0.9261297373957582, + "grad_norm": 0.45739081501960754, + "learning_rate": 2.681626022271222e-06, + "loss": 0.0877, + "step": 39508 + }, + { + "epoch": 0.9261531789705633, + "grad_norm": 0.5488783121109009, + "learning_rate": 2.6799318552935825e-06, + "loss": 0.0541, + "step": 39509 + }, + { + "epoch": 0.9261766205453682, + "grad_norm": 0.14352591335773468, + "learning_rate": 2.678238216377371e-06, + "loss": 0.0192, + "step": 39510 + }, + { + "epoch": 0.9262000621201733, + "grad_norm": 0.5367730259895325, + "learning_rate": 2.67654510553178e-06, + "loss": 0.0559, + "step": 39511 + }, + { + "epoch": 0.9262235036949782, + "grad_norm": 0.5867800712585449, + "learning_rate": 2.6748525227659913e-06, + "loss": 0.1434, + "step": 39512 + }, + { + "epoch": 0.9262469452697832, + "grad_norm": 0.24592815339565277, + "learning_rate": 2.6731604680891865e-06, + "loss": 0.0469, + "step": 39513 + }, + { + "epoch": 0.9262703868445882, + "grad_norm": 0.28840717673301697, + "learning_rate": 2.6714689415105576e-06, + "loss": 0.0324, + "step": 39514 + }, + { + "epoch": 0.9262938284193932, + "grad_norm": 0.1595148742198944, + "learning_rate": 2.6697779430392646e-06, + "loss": 0.019, + "step": 39515 + }, + { + "epoch": 0.9263172699941982, + "grad_norm": 0.8970825672149658, + "learning_rate": 2.6680874726844994e-06, + "loss": 0.0813, + "step": 39516 + }, + { + "epoch": 0.9263407115690032, + "grad_norm": 0.48015743494033813, + "learning_rate": 2.6663975304554112e-06, + "loss": 0.0669, + "step": 39517 + }, + { + "epoch": 0.9263641531438082, + "grad_norm": 0.30610477924346924, + "learning_rate": 2.664708116361203e-06, + "loss": 0.0263, + "step": 39518 + }, + { + "epoch": 0.9263875947186132, + "grad_norm": 0.5817469358444214, + "learning_rate": 2.6630192304110344e-06, + "loss": 0.0506, + "step": 39519 + }, + { + "epoch": 0.9264110362934181, + "grad_norm": 0.29562273621559143, + "learning_rate": 2.661330872614054e-06, + "loss": 0.0399, + "step": 39520 + }, + { + "epoch": 0.9264344778682232, + "grad_norm": 0.1804133951663971, + "learning_rate": 2.6596430429794204e-06, + "loss": 0.0322, + "step": 39521 + }, + { + "epoch": 0.9264579194430281, + "grad_norm": 0.34250563383102417, + "learning_rate": 2.6579557415163048e-06, + "loss": 0.0777, + "step": 39522 + }, + { + "epoch": 0.9264813610178332, + "grad_norm": 0.4534280300140381, + "learning_rate": 2.6562689682338547e-06, + "loss": 0.1075, + "step": 39523 + }, + { + "epoch": 0.9265048025926381, + "grad_norm": 0.13351286947727203, + "learning_rate": 2.65458272314123e-06, + "loss": 0.0171, + "step": 39524 + }, + { + "epoch": 0.9265282441674432, + "grad_norm": 0.4853908121585846, + "learning_rate": 2.652897006247579e-06, + "loss": 0.4587, + "step": 39525 + }, + { + "epoch": 0.9265516857422481, + "grad_norm": 0.22491301596164703, + "learning_rate": 2.6512118175620492e-06, + "loss": 0.0176, + "step": 39526 + }, + { + "epoch": 0.9265751273170532, + "grad_norm": 0.4356488287448883, + "learning_rate": 2.6495271570937783e-06, + "loss": 0.0833, + "step": 39527 + }, + { + "epoch": 0.9265985688918582, + "grad_norm": 0.2748123109340668, + "learning_rate": 2.647843024851904e-06, + "loss": 0.0252, + "step": 39528 + }, + { + "epoch": 0.9266220104666631, + "grad_norm": 0.11901070177555084, + "learning_rate": 2.646159420845562e-06, + "loss": 0.007, + "step": 39529 + }, + { + "epoch": 0.9266454520414682, + "grad_norm": 0.18404845893383026, + "learning_rate": 2.644476345083913e-06, + "loss": 0.0328, + "step": 39530 + }, + { + "epoch": 0.9266688936162731, + "grad_norm": 0.06957726180553436, + "learning_rate": 2.6427937975760707e-06, + "loss": 0.0098, + "step": 39531 + }, + { + "epoch": 0.9266923351910782, + "grad_norm": 0.2814123332500458, + "learning_rate": 2.6411117783311735e-06, + "loss": 0.0317, + "step": 39532 + }, + { + "epoch": 0.9267157767658831, + "grad_norm": 0.2580436170101166, + "learning_rate": 2.6394302873583355e-06, + "loss": 0.0556, + "step": 39533 + }, + { + "epoch": 0.9267392183406882, + "grad_norm": 0.1625000536441803, + "learning_rate": 2.637749324666683e-06, + "loss": 0.0176, + "step": 39534 + }, + { + "epoch": 0.9267626599154931, + "grad_norm": 0.20406389236450195, + "learning_rate": 2.6360688902653306e-06, + "loss": 0.0126, + "step": 39535 + }, + { + "epoch": 0.9267861014902982, + "grad_norm": 0.1842021942138672, + "learning_rate": 2.6343889841634164e-06, + "loss": 0.0171, + "step": 39536 + }, + { + "epoch": 0.9268095430651031, + "grad_norm": 0.45411306619644165, + "learning_rate": 2.632709606370054e-06, + "loss": 0.1015, + "step": 39537 + }, + { + "epoch": 0.9268329846399082, + "grad_norm": 0.4840933680534363, + "learning_rate": 2.6310307568943373e-06, + "loss": 0.0554, + "step": 39538 + }, + { + "epoch": 0.9268564262147131, + "grad_norm": 0.2723136842250824, + "learning_rate": 2.6293524357453912e-06, + "loss": 0.0541, + "step": 39539 + }, + { + "epoch": 0.9268798677895181, + "grad_norm": 0.4036281704902649, + "learning_rate": 2.6276746429323097e-06, + "loss": 0.0958, + "step": 39540 + }, + { + "epoch": 0.9269033093643231, + "grad_norm": 0.6498264670372009, + "learning_rate": 2.6259973784641956e-06, + "loss": 0.0887, + "step": 39541 + }, + { + "epoch": 0.9269267509391281, + "grad_norm": 0.5152514576911926, + "learning_rate": 2.6243206423501644e-06, + "loss": 0.1671, + "step": 39542 + }, + { + "epoch": 0.9269501925139331, + "grad_norm": 0.31254658102989197, + "learning_rate": 2.6226444345993086e-06, + "loss": 0.0493, + "step": 39543 + }, + { + "epoch": 0.9269736340887381, + "grad_norm": 0.3214053213596344, + "learning_rate": 2.620968755220732e-06, + "loss": 0.0722, + "step": 39544 + }, + { + "epoch": 0.9269970756635431, + "grad_norm": 0.5647748112678528, + "learning_rate": 2.619293604223505e-06, + "loss": 0.0651, + "step": 39545 + }, + { + "epoch": 0.9270205172383481, + "grad_norm": 0.7476678490638733, + "learning_rate": 2.6176189816167317e-06, + "loss": 0.1457, + "step": 39546 + }, + { + "epoch": 0.927043958813153, + "grad_norm": 0.4782024919986725, + "learning_rate": 2.615944887409505e-06, + "loss": 0.0894, + "step": 39547 + }, + { + "epoch": 0.9270674003879581, + "grad_norm": 0.641453742980957, + "learning_rate": 2.6142713216108837e-06, + "loss": 0.0748, + "step": 39548 + }, + { + "epoch": 0.927090841962763, + "grad_norm": 0.20031726360321045, + "learning_rate": 2.6125982842299723e-06, + "loss": 0.0432, + "step": 39549 + }, + { + "epoch": 0.9271142835375681, + "grad_norm": 0.11751015484333038, + "learning_rate": 2.610925775275841e-06, + "loss": 0.0151, + "step": 39550 + }, + { + "epoch": 0.927137725112373, + "grad_norm": 0.06557200849056244, + "learning_rate": 2.609253794757571e-06, + "loss": 0.0051, + "step": 39551 + }, + { + "epoch": 0.9271611666871781, + "grad_norm": 0.49288809299468994, + "learning_rate": 2.6075823426842226e-06, + "loss": 0.0817, + "step": 39552 + }, + { + "epoch": 0.927184608261983, + "grad_norm": 0.33061063289642334, + "learning_rate": 2.6059114190648657e-06, + "loss": 0.0502, + "step": 39553 + }, + { + "epoch": 0.9272080498367881, + "grad_norm": 0.1578218638896942, + "learning_rate": 2.6042410239085824e-06, + "loss": 0.0199, + "step": 39554 + }, + { + "epoch": 0.927231491411593, + "grad_norm": 0.425625741481781, + "learning_rate": 2.6025711572244093e-06, + "loss": 0.0778, + "step": 39555 + }, + { + "epoch": 0.927254932986398, + "grad_norm": 0.44257864356040955, + "learning_rate": 2.6009018190214396e-06, + "loss": 0.0707, + "step": 39556 + }, + { + "epoch": 0.927278374561203, + "grad_norm": 0.3923092484474182, + "learning_rate": 2.5992330093087103e-06, + "loss": 0.0705, + "step": 39557 + }, + { + "epoch": 0.927301816136008, + "grad_norm": 0.6452748775482178, + "learning_rate": 2.597564728095281e-06, + "loss": 0.1895, + "step": 39558 + }, + { + "epoch": 0.927325257710813, + "grad_norm": 0.44378793239593506, + "learning_rate": 2.595896975390211e-06, + "loss": 0.0634, + "step": 39559 + }, + { + "epoch": 0.927348699285618, + "grad_norm": 0.33620545268058777, + "learning_rate": 2.594229751202526e-06, + "loss": 0.0645, + "step": 39560 + }, + { + "epoch": 0.927372140860423, + "grad_norm": 0.4846253991127014, + "learning_rate": 2.592563055541297e-06, + "loss": 0.0604, + "step": 39561 + }, + { + "epoch": 0.927395582435228, + "grad_norm": 0.4136323928833008, + "learning_rate": 2.5908968884155728e-06, + "loss": 0.0559, + "step": 39562 + }, + { + "epoch": 0.927419024010033, + "grad_norm": 0.49870485067367554, + "learning_rate": 2.5892312498343676e-06, + "loss": 0.0578, + "step": 39563 + }, + { + "epoch": 0.927442465584838, + "grad_norm": 0.49732571840286255, + "learning_rate": 2.587566139806741e-06, + "loss": 0.0922, + "step": 39564 + }, + { + "epoch": 0.9274659071596429, + "grad_norm": 0.5042515397071838, + "learning_rate": 2.585901558341697e-06, + "loss": 0.047, + "step": 39565 + }, + { + "epoch": 0.927489348734448, + "grad_norm": 0.3125109374523163, + "learning_rate": 2.584237505448317e-06, + "loss": 0.0511, + "step": 39566 + }, + { + "epoch": 0.9275127903092529, + "grad_norm": 0.2949400842189789, + "learning_rate": 2.5825739811355833e-06, + "loss": 0.0356, + "step": 39567 + }, + { + "epoch": 0.927536231884058, + "grad_norm": 0.1253243237733841, + "learning_rate": 2.5809109854125547e-06, + "loss": 0.017, + "step": 39568 + }, + { + "epoch": 0.9275596734588629, + "grad_norm": 0.3875078856945038, + "learning_rate": 2.5792485182882354e-06, + "loss": 0.0406, + "step": 39569 + }, + { + "epoch": 0.927583115033668, + "grad_norm": 0.2948315739631653, + "learning_rate": 2.5775865797716625e-06, + "loss": 0.0323, + "step": 39570 + }, + { + "epoch": 0.9276065566084729, + "grad_norm": 0.24605710804462433, + "learning_rate": 2.57592516987184e-06, + "loss": 0.0462, + "step": 39571 + }, + { + "epoch": 0.927629998183278, + "grad_norm": 0.5019896030426025, + "learning_rate": 2.574264288597794e-06, + "loss": 0.0622, + "step": 39572 + }, + { + "epoch": 0.9276534397580829, + "grad_norm": 0.38925930857658386, + "learning_rate": 2.5726039359585064e-06, + "loss": 0.4548, + "step": 39573 + }, + { + "epoch": 0.927676881332888, + "grad_norm": 0.4132390320301056, + "learning_rate": 2.570944111963036e-06, + "loss": 0.0704, + "step": 39574 + }, + { + "epoch": 0.9277003229076929, + "grad_norm": 0.21166417002677917, + "learning_rate": 2.5692848166203433e-06, + "loss": 0.0361, + "step": 39575 + }, + { + "epoch": 0.9277237644824979, + "grad_norm": 0.4575370252132416, + "learning_rate": 2.5676260499394643e-06, + "loss": 0.0783, + "step": 39576 + }, + { + "epoch": 0.9277472060573029, + "grad_norm": 0.5479967594146729, + "learning_rate": 2.565967811929382e-06, + "loss": 0.0789, + "step": 39577 + }, + { + "epoch": 0.9277706476321079, + "grad_norm": 0.5459845066070557, + "learning_rate": 2.564310102599088e-06, + "loss": 0.0817, + "step": 39578 + }, + { + "epoch": 0.927794089206913, + "grad_norm": 0.5242279171943665, + "learning_rate": 2.5626529219575977e-06, + "loss": 0.0803, + "step": 39579 + }, + { + "epoch": 0.9278175307817179, + "grad_norm": 0.13730204105377197, + "learning_rate": 2.5609962700138824e-06, + "loss": 0.0227, + "step": 39580 + }, + { + "epoch": 0.927840972356523, + "grad_norm": 0.6190561652183533, + "learning_rate": 2.559340146776945e-06, + "loss": 0.8074, + "step": 39581 + }, + { + "epoch": 0.9278644139313279, + "grad_norm": 0.26103973388671875, + "learning_rate": 2.557684552255768e-06, + "loss": 0.053, + "step": 39582 + }, + { + "epoch": 0.927887855506133, + "grad_norm": 0.7264842987060547, + "learning_rate": 2.5560294864593436e-06, + "loss": 0.1082, + "step": 39583 + }, + { + "epoch": 0.9279112970809379, + "grad_norm": 0.25480198860168457, + "learning_rate": 2.554374949396643e-06, + "loss": 0.0246, + "step": 39584 + }, + { + "epoch": 0.9279347386557429, + "grad_norm": 0.5952173471450806, + "learning_rate": 2.5527209410766363e-06, + "loss": 0.1235, + "step": 39585 + }, + { + "epoch": 0.9279581802305479, + "grad_norm": 0.33333465456962585, + "learning_rate": 2.551067461508294e-06, + "loss": 0.0668, + "step": 39586 + }, + { + "epoch": 0.9279816218053529, + "grad_norm": 0.5532070398330688, + "learning_rate": 2.5494145107006208e-06, + "loss": 0.0535, + "step": 39587 + }, + { + "epoch": 0.9280050633801579, + "grad_norm": 0.8589189052581787, + "learning_rate": 2.547762088662553e-06, + "loss": 0.1018, + "step": 39588 + }, + { + "epoch": 0.9280285049549629, + "grad_norm": 0.5093718767166138, + "learning_rate": 2.5461101954030732e-06, + "loss": 0.0806, + "step": 39589 + }, + { + "epoch": 0.9280519465297679, + "grad_norm": 0.3090704679489136, + "learning_rate": 2.544458830931129e-06, + "loss": 0.2783, + "step": 39590 + }, + { + "epoch": 0.9280753881045729, + "grad_norm": 0.19290493428707123, + "learning_rate": 2.5428079952557026e-06, + "loss": 0.0279, + "step": 39591 + }, + { + "epoch": 0.9280988296793778, + "grad_norm": 0.28254634141921997, + "learning_rate": 2.5411576883857424e-06, + "loss": 0.0312, + "step": 39592 + }, + { + "epoch": 0.9281222712541829, + "grad_norm": 0.39475128054618835, + "learning_rate": 2.5395079103301854e-06, + "loss": 0.0354, + "step": 39593 + }, + { + "epoch": 0.9281457128289878, + "grad_norm": 0.288308322429657, + "learning_rate": 2.5378586610980137e-06, + "loss": 0.037, + "step": 39594 + }, + { + "epoch": 0.9281691544037929, + "grad_norm": 0.3179360330104828, + "learning_rate": 2.5362099406981644e-06, + "loss": 0.3363, + "step": 39595 + }, + { + "epoch": 0.9281925959785978, + "grad_norm": 0.8625703454017639, + "learning_rate": 2.5345617491395744e-06, + "loss": 0.14, + "step": 39596 + }, + { + "epoch": 0.9282160375534029, + "grad_norm": 0.30271226167678833, + "learning_rate": 2.5329140864311927e-06, + "loss": 0.0524, + "step": 39597 + }, + { + "epoch": 0.9282394791282078, + "grad_norm": 0.11965698003768921, + "learning_rate": 2.5312669525819567e-06, + "loss": 0.0222, + "step": 39598 + }, + { + "epoch": 0.9282629207030129, + "grad_norm": 0.5563119649887085, + "learning_rate": 2.529620347600803e-06, + "loss": 0.6004, + "step": 39599 + }, + { + "epoch": 0.9282863622778178, + "grad_norm": 0.3085756301879883, + "learning_rate": 2.5279742714966804e-06, + "loss": 0.0275, + "step": 39600 + }, + { + "epoch": 0.9283098038526229, + "grad_norm": 0.3497069478034973, + "learning_rate": 2.526328724278504e-06, + "loss": 0.0517, + "step": 39601 + }, + { + "epoch": 0.9283332454274278, + "grad_norm": 0.18397413194179535, + "learning_rate": 2.5246837059552107e-06, + "loss": 0.0331, + "step": 39602 + }, + { + "epoch": 0.9283566870022328, + "grad_norm": 0.25876280665397644, + "learning_rate": 2.5230392165357163e-06, + "loss": 0.0459, + "step": 39603 + }, + { + "epoch": 0.9283801285770378, + "grad_norm": 0.24815930426120758, + "learning_rate": 2.521395256028969e-06, + "loss": 0.0398, + "step": 39604 + }, + { + "epoch": 0.9284035701518428, + "grad_norm": 0.4746960997581482, + "learning_rate": 2.51975182444385e-06, + "loss": 0.0793, + "step": 39605 + }, + { + "epoch": 0.9284270117266478, + "grad_norm": 0.6125414371490479, + "learning_rate": 2.5181089217893197e-06, + "loss": 0.1002, + "step": 39606 + }, + { + "epoch": 0.9284504533014528, + "grad_norm": 0.13692441582679749, + "learning_rate": 2.5164665480742587e-06, + "loss": 0.0363, + "step": 39607 + }, + { + "epoch": 0.9284738948762578, + "grad_norm": 0.33409494161605835, + "learning_rate": 2.5148247033075945e-06, + "loss": 0.0552, + "step": 39608 + }, + { + "epoch": 0.9284973364510628, + "grad_norm": 0.3754505515098572, + "learning_rate": 2.513183387498241e-06, + "loss": 0.039, + "step": 39609 + }, + { + "epoch": 0.9285207780258677, + "grad_norm": 0.21646031737327576, + "learning_rate": 2.5115426006550923e-06, + "loss": 0.0236, + "step": 39610 + }, + { + "epoch": 0.9285442196006728, + "grad_norm": 0.5020193457603455, + "learning_rate": 2.5099023427870407e-06, + "loss": 0.105, + "step": 39611 + }, + { + "epoch": 0.9285676611754777, + "grad_norm": 0.4189395606517792, + "learning_rate": 2.508262613903012e-06, + "loss": 0.0675, + "step": 39612 + }, + { + "epoch": 0.9285911027502828, + "grad_norm": 0.381556898355484, + "learning_rate": 2.5066234140118884e-06, + "loss": 0.0738, + "step": 39613 + }, + { + "epoch": 0.9286145443250877, + "grad_norm": 0.40534061193466187, + "learning_rate": 2.5049847431225627e-06, + "loss": 0.0534, + "step": 39614 + }, + { + "epoch": 0.9286379858998928, + "grad_norm": 0.12576813995838165, + "learning_rate": 2.503346601243939e-06, + "loss": 0.0119, + "step": 39615 + }, + { + "epoch": 0.9286614274746977, + "grad_norm": 0.38430097699165344, + "learning_rate": 2.5017089883848876e-06, + "loss": 0.064, + "step": 39616 + }, + { + "epoch": 0.9286848690495028, + "grad_norm": 0.442425012588501, + "learning_rate": 2.5000719045543243e-06, + "loss": 0.0453, + "step": 39617 + }, + { + "epoch": 0.9287083106243077, + "grad_norm": 0.25683459639549255, + "learning_rate": 2.4984353497610857e-06, + "loss": 0.0358, + "step": 39618 + }, + { + "epoch": 0.9287317521991127, + "grad_norm": 0.2954341173171997, + "learning_rate": 2.4967993240140985e-06, + "loss": 0.0341, + "step": 39619 + }, + { + "epoch": 0.9287551937739177, + "grad_norm": 0.3852499723434448, + "learning_rate": 2.495163827322211e-06, + "loss": 0.0842, + "step": 39620 + }, + { + "epoch": 0.9287786353487227, + "grad_norm": 0.5135354995727539, + "learning_rate": 2.4935288596943164e-06, + "loss": 0.0879, + "step": 39621 + }, + { + "epoch": 0.9288020769235277, + "grad_norm": 0.5345034599304199, + "learning_rate": 2.4918944211392736e-06, + "loss": 0.1212, + "step": 39622 + }, + { + "epoch": 0.9288255184983327, + "grad_norm": 0.49843111634254456, + "learning_rate": 2.490260511665943e-06, + "loss": 0.0428, + "step": 39623 + }, + { + "epoch": 0.9288489600731377, + "grad_norm": 0.2127937525510788, + "learning_rate": 2.4886271312832055e-06, + "loss": 0.0319, + "step": 39624 + }, + { + "epoch": 0.9288724016479427, + "grad_norm": 0.503402590751648, + "learning_rate": 2.4869942799999213e-06, + "loss": 0.0865, + "step": 39625 + }, + { + "epoch": 0.9288958432227477, + "grad_norm": 0.8866694569587708, + "learning_rate": 2.485361957824939e-06, + "loss": 0.0678, + "step": 39626 + }, + { + "epoch": 0.9289192847975527, + "grad_norm": 0.5844088196754456, + "learning_rate": 2.4837301647671287e-06, + "loss": 0.0928, + "step": 39627 + }, + { + "epoch": 0.9289427263723576, + "grad_norm": 0.4544989764690399, + "learning_rate": 2.48209890083535e-06, + "loss": 0.0889, + "step": 39628 + }, + { + "epoch": 0.9289661679471627, + "grad_norm": 0.3764660060405731, + "learning_rate": 2.480468166038441e-06, + "loss": 0.0293, + "step": 39629 + }, + { + "epoch": 0.9289896095219677, + "grad_norm": 0.6483609080314636, + "learning_rate": 2.4788379603852496e-06, + "loss": 0.0948, + "step": 39630 + }, + { + "epoch": 0.9290130510967727, + "grad_norm": 0.7021113038063049, + "learning_rate": 2.4772082838846245e-06, + "loss": 0.1336, + "step": 39631 + }, + { + "epoch": 0.9290364926715777, + "grad_norm": 0.3772090971469879, + "learning_rate": 2.475579136545414e-06, + "loss": 0.0614, + "step": 39632 + }, + { + "epoch": 0.9290599342463827, + "grad_norm": 0.40870824456214905, + "learning_rate": 2.4739505183764445e-06, + "loss": 0.0334, + "step": 39633 + }, + { + "epoch": 0.9290833758211877, + "grad_norm": 0.19299562275409698, + "learning_rate": 2.4723224293865756e-06, + "loss": 0.0405, + "step": 39634 + }, + { + "epoch": 0.9291068173959927, + "grad_norm": 0.35606658458709717, + "learning_rate": 2.470694869584622e-06, + "loss": 0.3761, + "step": 39635 + }, + { + "epoch": 0.9291302589707977, + "grad_norm": 0.33982011675834656, + "learning_rate": 2.469067838979411e-06, + "loss": 0.0565, + "step": 39636 + }, + { + "epoch": 0.9291537005456026, + "grad_norm": 0.10359852761030197, + "learning_rate": 2.46744133757979e-06, + "loss": 0.0195, + "step": 39637 + }, + { + "epoch": 0.9291771421204077, + "grad_norm": 0.4257062077522278, + "learning_rate": 2.4658153653945635e-06, + "loss": 0.0475, + "step": 39638 + }, + { + "epoch": 0.9292005836952126, + "grad_norm": 0.2124229222536087, + "learning_rate": 2.4641899224325693e-06, + "loss": 0.0521, + "step": 39639 + }, + { + "epoch": 0.9292240252700177, + "grad_norm": 0.6191243529319763, + "learning_rate": 2.462565008702633e-06, + "loss": 0.0714, + "step": 39640 + }, + { + "epoch": 0.9292474668448226, + "grad_norm": 0.6837379336357117, + "learning_rate": 2.460940624213559e-06, + "loss": 0.1065, + "step": 39641 + }, + { + "epoch": 0.9292709084196277, + "grad_norm": 0.33089736104011536, + "learning_rate": 2.4593167689741626e-06, + "loss": 0.0695, + "step": 39642 + }, + { + "epoch": 0.9292943499944326, + "grad_norm": 0.2450268268585205, + "learning_rate": 2.4576934429932473e-06, + "loss": 0.1418, + "step": 39643 + }, + { + "epoch": 0.9293177915692377, + "grad_norm": 0.4621208906173706, + "learning_rate": 2.45607064627964e-06, + "loss": 0.0861, + "step": 39644 + }, + { + "epoch": 0.9293412331440426, + "grad_norm": 0.4550333321094513, + "learning_rate": 2.4544483788421334e-06, + "loss": 0.078, + "step": 39645 + }, + { + "epoch": 0.9293646747188476, + "grad_norm": 0.38202589750289917, + "learning_rate": 2.4528266406895315e-06, + "loss": 0.0499, + "step": 39646 + }, + { + "epoch": 0.9293881162936526, + "grad_norm": 0.4035293757915497, + "learning_rate": 2.451205431830639e-06, + "loss": 0.0718, + "step": 39647 + }, + { + "epoch": 0.9294115578684576, + "grad_norm": 0.2601456046104431, + "learning_rate": 2.4495847522742476e-06, + "loss": 0.0182, + "step": 39648 + }, + { + "epoch": 0.9294349994432626, + "grad_norm": 0.30057093501091003, + "learning_rate": 2.4479646020291513e-06, + "loss": 0.0274, + "step": 39649 + }, + { + "epoch": 0.9294584410180676, + "grad_norm": 0.41116252541542053, + "learning_rate": 2.446344981104132e-06, + "loss": 0.0507, + "step": 39650 + }, + { + "epoch": 0.9294818825928726, + "grad_norm": 0.2576913833618164, + "learning_rate": 2.4447258895079926e-06, + "loss": 0.0302, + "step": 39651 + }, + { + "epoch": 0.9295053241676776, + "grad_norm": 0.5495731830596924, + "learning_rate": 2.4431073272495274e-06, + "loss": 0.0339, + "step": 39652 + }, + { + "epoch": 0.9295287657424826, + "grad_norm": 0.5073724389076233, + "learning_rate": 2.441489294337496e-06, + "loss": 0.0519, + "step": 39653 + }, + { + "epoch": 0.9295522073172876, + "grad_norm": 0.19278106093406677, + "learning_rate": 2.439871790780679e-06, + "loss": 0.0184, + "step": 39654 + }, + { + "epoch": 0.9295756488920925, + "grad_norm": 0.20410975813865662, + "learning_rate": 2.4382548165878705e-06, + "loss": 0.0366, + "step": 39655 + }, + { + "epoch": 0.9295990904668976, + "grad_norm": 0.4901798367500305, + "learning_rate": 2.43663837176783e-06, + "loss": 0.098, + "step": 39656 + }, + { + "epoch": 0.9296225320417025, + "grad_norm": 0.40282708406448364, + "learning_rate": 2.435022456329339e-06, + "loss": 0.0999, + "step": 39657 + }, + { + "epoch": 0.9296459736165076, + "grad_norm": 0.3470439910888672, + "learning_rate": 2.4334070702811572e-06, + "loss": 0.0608, + "step": 39658 + }, + { + "epoch": 0.9296694151913125, + "grad_norm": 0.5764090418815613, + "learning_rate": 2.4317922136320447e-06, + "loss": 0.0588, + "step": 39659 + }, + { + "epoch": 0.9296928567661176, + "grad_norm": 0.5667243599891663, + "learning_rate": 2.430177886390783e-06, + "loss": 0.0534, + "step": 39660 + }, + { + "epoch": 0.9297162983409225, + "grad_norm": 0.2014477550983429, + "learning_rate": 2.42856408856611e-06, + "loss": 0.0125, + "step": 39661 + }, + { + "epoch": 0.9297397399157276, + "grad_norm": 0.0609443336725235, + "learning_rate": 2.4269508201667957e-06, + "loss": 0.0043, + "step": 39662 + }, + { + "epoch": 0.9297631814905325, + "grad_norm": 0.5236748456954956, + "learning_rate": 2.425338081201578e-06, + "loss": 0.4153, + "step": 39663 + }, + { + "epoch": 0.9297866230653375, + "grad_norm": 0.6101469993591309, + "learning_rate": 2.423725871679228e-06, + "loss": 0.1361, + "step": 39664 + }, + { + "epoch": 0.9298100646401425, + "grad_norm": 0.4236448109149933, + "learning_rate": 2.422114191608493e-06, + "loss": 0.0464, + "step": 39665 + }, + { + "epoch": 0.9298335062149475, + "grad_norm": 0.1524774432182312, + "learning_rate": 2.4205030409981012e-06, + "loss": 0.0232, + "step": 39666 + }, + { + "epoch": 0.9298569477897525, + "grad_norm": 0.5850568413734436, + "learning_rate": 2.4188924198568106e-06, + "loss": 0.1211, + "step": 39667 + }, + { + "epoch": 0.9298803893645575, + "grad_norm": 0.3149416446685791, + "learning_rate": 2.417282328193349e-06, + "loss": 0.0391, + "step": 39668 + }, + { + "epoch": 0.9299038309393625, + "grad_norm": 0.43274983763694763, + "learning_rate": 2.4156727660164525e-06, + "loss": 0.0662, + "step": 39669 + }, + { + "epoch": 0.9299272725141675, + "grad_norm": 0.5112161636352539, + "learning_rate": 2.41406373333487e-06, + "loss": 0.5571, + "step": 39670 + }, + { + "epoch": 0.9299507140889725, + "grad_norm": 0.3293932378292084, + "learning_rate": 2.412455230157318e-06, + "loss": 0.0576, + "step": 39671 + }, + { + "epoch": 0.9299741556637775, + "grad_norm": 0.46803009510040283, + "learning_rate": 2.410847256492521e-06, + "loss": 0.0972, + "step": 39672 + }, + { + "epoch": 0.9299975972385824, + "grad_norm": 0.2943069040775299, + "learning_rate": 2.409239812349218e-06, + "loss": 0.0423, + "step": 39673 + }, + { + "epoch": 0.9300210388133875, + "grad_norm": 0.08370602875947952, + "learning_rate": 2.4076328977361117e-06, + "loss": 0.0068, + "step": 39674 + }, + { + "epoch": 0.9300444803881924, + "grad_norm": 0.06958069652318954, + "learning_rate": 2.406026512661941e-06, + "loss": 0.0068, + "step": 39675 + }, + { + "epoch": 0.9300679219629975, + "grad_norm": 0.6253231763839722, + "learning_rate": 2.40442065713542e-06, + "loss": 0.1233, + "step": 39676 + }, + { + "epoch": 0.9300913635378024, + "grad_norm": 0.157024547457695, + "learning_rate": 2.4028153311652535e-06, + "loss": 0.0284, + "step": 39677 + }, + { + "epoch": 0.9301148051126075, + "grad_norm": 0.4587274193763733, + "learning_rate": 2.4012105347601565e-06, + "loss": 0.0806, + "step": 39678 + }, + { + "epoch": 0.9301382466874124, + "grad_norm": 0.3013659715652466, + "learning_rate": 2.3996062679288333e-06, + "loss": 0.2082, + "step": 39679 + }, + { + "epoch": 0.9301616882622175, + "grad_norm": 0.05175082013010979, + "learning_rate": 2.398002530679999e-06, + "loss": 0.0069, + "step": 39680 + }, + { + "epoch": 0.9301851298370225, + "grad_norm": 0.3672591745853424, + "learning_rate": 2.3963993230223246e-06, + "loss": 0.0496, + "step": 39681 + }, + { + "epoch": 0.9302085714118274, + "grad_norm": 0.47998595237731934, + "learning_rate": 2.394796644964559e-06, + "loss": 0.0487, + "step": 39682 + }, + { + "epoch": 0.9302320129866325, + "grad_norm": 0.3486565351486206, + "learning_rate": 2.393194496515361e-06, + "loss": 0.0339, + "step": 39683 + }, + { + "epoch": 0.9302554545614374, + "grad_norm": 0.5417203903198242, + "learning_rate": 2.391592877683424e-06, + "loss": 0.1207, + "step": 39684 + }, + { + "epoch": 0.9302788961362425, + "grad_norm": 0.13283203542232513, + "learning_rate": 2.3899917884774634e-06, + "loss": 0.0242, + "step": 39685 + }, + { + "epoch": 0.9303023377110474, + "grad_norm": 0.36464622616767883, + "learning_rate": 2.3883912289061506e-06, + "loss": 0.037, + "step": 39686 + }, + { + "epoch": 0.9303257792858525, + "grad_norm": 0.5425735116004944, + "learning_rate": 2.3867911989781556e-06, + "loss": 0.1006, + "step": 39687 + }, + { + "epoch": 0.9303492208606574, + "grad_norm": 0.5205128788948059, + "learning_rate": 2.385191698702183e-06, + "loss": 0.7143, + "step": 39688 + }, + { + "epoch": 0.9303726624354625, + "grad_norm": 0.5312729477882385, + "learning_rate": 2.383592728086903e-06, + "loss": 0.0836, + "step": 39689 + }, + { + "epoch": 0.9303961040102674, + "grad_norm": 0.8553577065467834, + "learning_rate": 2.3819942871410093e-06, + "loss": 0.1895, + "step": 39690 + }, + { + "epoch": 0.9304195455850724, + "grad_norm": 0.15089468657970428, + "learning_rate": 2.3803963758731506e-06, + "loss": 0.0332, + "step": 39691 + }, + { + "epoch": 0.9304429871598774, + "grad_norm": 0.40773481130599976, + "learning_rate": 2.3787989942920086e-06, + "loss": 0.0495, + "step": 39692 + }, + { + "epoch": 0.9304664287346824, + "grad_norm": 0.5486031770706177, + "learning_rate": 2.3772021424062427e-06, + "loss": 0.0783, + "step": 39693 + }, + { + "epoch": 0.9304898703094874, + "grad_norm": 0.28533580899238586, + "learning_rate": 2.3756058202245134e-06, + "loss": 0.0473, + "step": 39694 + }, + { + "epoch": 0.9305133118842924, + "grad_norm": 0.7315354943275452, + "learning_rate": 2.3740100277555023e-06, + "loss": 0.072, + "step": 39695 + }, + { + "epoch": 0.9305367534590974, + "grad_norm": 0.14448197185993195, + "learning_rate": 2.3724147650078685e-06, + "loss": 0.029, + "step": 39696 + }, + { + "epoch": 0.9305601950339024, + "grad_norm": 0.09411785006523132, + "learning_rate": 2.3708200319902507e-06, + "loss": 0.0064, + "step": 39697 + }, + { + "epoch": 0.9305836366087074, + "grad_norm": 0.5448334813117981, + "learning_rate": 2.3692258287112966e-06, + "loss": 0.1045, + "step": 39698 + }, + { + "epoch": 0.9306070781835124, + "grad_norm": 0.1558474749326706, + "learning_rate": 2.3676321551796776e-06, + "loss": 0.019, + "step": 39699 + }, + { + "epoch": 0.9306305197583173, + "grad_norm": 0.6762322187423706, + "learning_rate": 2.366039011404031e-06, + "loss": 0.1507, + "step": 39700 + }, + { + "epoch": 0.9306539613331224, + "grad_norm": 0.15678104758262634, + "learning_rate": 2.3644463973929943e-06, + "loss": 0.0227, + "step": 39701 + }, + { + "epoch": 0.9306774029079273, + "grad_norm": 0.264438271522522, + "learning_rate": 2.3628543131552273e-06, + "loss": 0.0428, + "step": 39702 + }, + { + "epoch": 0.9307008444827324, + "grad_norm": 0.1620715707540512, + "learning_rate": 2.3612627586993565e-06, + "loss": 0.0114, + "step": 39703 + }, + { + "epoch": 0.9307242860575373, + "grad_norm": 0.3641142249107361, + "learning_rate": 2.359671734034019e-06, + "loss": 0.0774, + "step": 39704 + }, + { + "epoch": 0.9307477276323424, + "grad_norm": 0.3158729374408722, + "learning_rate": 2.358081239167853e-06, + "loss": 0.0516, + "step": 39705 + }, + { + "epoch": 0.9307711692071473, + "grad_norm": 0.2041471302509308, + "learning_rate": 2.3564912741094736e-06, + "loss": 0.0422, + "step": 39706 + }, + { + "epoch": 0.9307946107819524, + "grad_norm": 0.6059731245040894, + "learning_rate": 2.3549018388675182e-06, + "loss": 0.0891, + "step": 39707 + }, + { + "epoch": 0.9308180523567573, + "grad_norm": 0.10827692598104477, + "learning_rate": 2.3533129334506133e-06, + "loss": 0.0177, + "step": 39708 + }, + { + "epoch": 0.9308414939315623, + "grad_norm": 0.4201374053955078, + "learning_rate": 2.3517245578673854e-06, + "loss": 0.0853, + "step": 39709 + }, + { + "epoch": 0.9308649355063673, + "grad_norm": 0.4336424469947815, + "learning_rate": 2.3501367121264382e-06, + "loss": 0.037, + "step": 39710 + }, + { + "epoch": 0.9308883770811723, + "grad_norm": 0.4296145439147949, + "learning_rate": 2.348549396236388e-06, + "loss": 0.1027, + "step": 39711 + }, + { + "epoch": 0.9309118186559773, + "grad_norm": 0.7735172510147095, + "learning_rate": 2.3469626102058606e-06, + "loss": 0.1071, + "step": 39712 + }, + { + "epoch": 0.9309352602307823, + "grad_norm": 0.1464652121067047, + "learning_rate": 2.3453763540434604e-06, + "loss": 0.0237, + "step": 39713 + }, + { + "epoch": 0.9309587018055873, + "grad_norm": 0.5551452040672302, + "learning_rate": 2.3437906277577916e-06, + "loss": 0.5666, + "step": 39714 + }, + { + "epoch": 0.9309821433803923, + "grad_norm": 0.6208924055099487, + "learning_rate": 2.342205431357469e-06, + "loss": 0.0573, + "step": 39715 + }, + { + "epoch": 0.9310055849551973, + "grad_norm": 0.40773871541023254, + "learning_rate": 2.340620764851087e-06, + "loss": 0.6456, + "step": 39716 + }, + { + "epoch": 0.9310290265300023, + "grad_norm": 0.5279151201248169, + "learning_rate": 2.3390366282472378e-06, + "loss": 0.5044, + "step": 39717 + }, + { + "epoch": 0.9310524681048072, + "grad_norm": 0.3943181037902832, + "learning_rate": 2.3374530215545256e-06, + "loss": 0.0741, + "step": 39718 + }, + { + "epoch": 0.9310759096796123, + "grad_norm": 0.5792568922042847, + "learning_rate": 2.3358699447815215e-06, + "loss": 0.1063, + "step": 39719 + }, + { + "epoch": 0.9310993512544172, + "grad_norm": 0.5008188486099243, + "learning_rate": 2.3342873979368516e-06, + "loss": 0.0787, + "step": 39720 + }, + { + "epoch": 0.9311227928292223, + "grad_norm": 0.46409037709236145, + "learning_rate": 2.3327053810290766e-06, + "loss": 0.0444, + "step": 39721 + }, + { + "epoch": 0.9311462344040272, + "grad_norm": 0.38725048303604126, + "learning_rate": 2.3311238940668e-06, + "loss": 0.0487, + "step": 39722 + }, + { + "epoch": 0.9311696759788323, + "grad_norm": 0.5214921236038208, + "learning_rate": 2.3295429370585707e-06, + "loss": 0.0766, + "step": 39723 + }, + { + "epoch": 0.9311931175536372, + "grad_norm": 0.45798543095588684, + "learning_rate": 2.3279625100130042e-06, + "loss": 0.0403, + "step": 39724 + }, + { + "epoch": 0.9312165591284423, + "grad_norm": 0.342293918132782, + "learning_rate": 2.32638261293866e-06, + "loss": 0.0681, + "step": 39725 + }, + { + "epoch": 0.9312400007032472, + "grad_norm": 0.17484432458877563, + "learning_rate": 2.3248032458440984e-06, + "loss": 0.0188, + "step": 39726 + }, + { + "epoch": 0.9312634422780522, + "grad_norm": 0.4875777065753937, + "learning_rate": 2.323224408737901e-06, + "loss": 0.0169, + "step": 39727 + }, + { + "epoch": 0.9312868838528572, + "grad_norm": 0.3545413315296173, + "learning_rate": 2.3216461016286494e-06, + "loss": 0.0367, + "step": 39728 + }, + { + "epoch": 0.9313103254276622, + "grad_norm": 0.3195919692516327, + "learning_rate": 2.320068324524882e-06, + "loss": 0.049, + "step": 39729 + }, + { + "epoch": 0.9313337670024672, + "grad_norm": 0.2579483985900879, + "learning_rate": 2.31849107743517e-06, + "loss": 0.0226, + "step": 39730 + }, + { + "epoch": 0.9313572085772722, + "grad_norm": 0.34193745255470276, + "learning_rate": 2.316914360368083e-06, + "loss": 0.0195, + "step": 39731 + }, + { + "epoch": 0.9313806501520773, + "grad_norm": 0.6765389442443848, + "learning_rate": 2.3153381733321487e-06, + "loss": 0.6006, + "step": 39732 + }, + { + "epoch": 0.9314040917268822, + "grad_norm": 0.14535927772521973, + "learning_rate": 2.3137625163359377e-06, + "loss": 0.0277, + "step": 39733 + }, + { + "epoch": 0.9314275333016873, + "grad_norm": 0.2051989585161209, + "learning_rate": 2.3121873893880097e-06, + "loss": 0.0214, + "step": 39734 + }, + { + "epoch": 0.9314509748764922, + "grad_norm": 0.1469792276620865, + "learning_rate": 2.3106127924968913e-06, + "loss": 0.0206, + "step": 39735 + }, + { + "epoch": 0.9314744164512972, + "grad_norm": 0.5996032357215881, + "learning_rate": 2.309038725671131e-06, + "loss": 0.1011, + "step": 39736 + }, + { + "epoch": 0.9314978580261022, + "grad_norm": 0.4644567668437958, + "learning_rate": 2.307465188919289e-06, + "loss": 0.0427, + "step": 39737 + }, + { + "epoch": 0.9315212996009072, + "grad_norm": 0.17779521644115448, + "learning_rate": 2.3058921822498803e-06, + "loss": 0.0226, + "step": 39738 + }, + { + "epoch": 0.9315447411757122, + "grad_norm": 0.24625854194164276, + "learning_rate": 2.3043197056714315e-06, + "loss": 0.0493, + "step": 39739 + }, + { + "epoch": 0.9315681827505172, + "grad_norm": 0.11578468978404999, + "learning_rate": 2.3027477591925138e-06, + "loss": 0.0166, + "step": 39740 + }, + { + "epoch": 0.9315916243253222, + "grad_norm": 0.19454997777938843, + "learning_rate": 2.3011763428216203e-06, + "loss": 0.0347, + "step": 39741 + }, + { + "epoch": 0.9316150659001272, + "grad_norm": 0.5636680126190186, + "learning_rate": 2.2996054565672998e-06, + "loss": 0.5737, + "step": 39742 + }, + { + "epoch": 0.9316385074749322, + "grad_norm": 0.1539382040500641, + "learning_rate": 2.2980351004380674e-06, + "loss": 0.0226, + "step": 39743 + }, + { + "epoch": 0.9316619490497372, + "grad_norm": 0.3239201009273529, + "learning_rate": 2.2964652744424387e-06, + "loss": 0.032, + "step": 39744 + }, + { + "epoch": 0.9316853906245421, + "grad_norm": 0.15491130948066711, + "learning_rate": 2.294895978588929e-06, + "loss": 0.0311, + "step": 39745 + }, + { + "epoch": 0.9317088321993472, + "grad_norm": 0.4830906093120575, + "learning_rate": 2.2933272128860762e-06, + "loss": 0.0543, + "step": 39746 + }, + { + "epoch": 0.9317322737741521, + "grad_norm": 0.4383120536804199, + "learning_rate": 2.2917589773423732e-06, + "loss": 0.6152, + "step": 39747 + }, + { + "epoch": 0.9317557153489572, + "grad_norm": 0.23793596029281616, + "learning_rate": 2.2901912719663133e-06, + "loss": 0.0416, + "step": 39748 + }, + { + "epoch": 0.9317791569237621, + "grad_norm": 0.35992923378944397, + "learning_rate": 2.288624096766445e-06, + "loss": 0.064, + "step": 39749 + }, + { + "epoch": 0.9318025984985672, + "grad_norm": 0.36177369952201843, + "learning_rate": 2.287057451751251e-06, + "loss": 0.0515, + "step": 39750 + }, + { + "epoch": 0.9318260400733721, + "grad_norm": 0.345192015171051, + "learning_rate": 2.285491336929213e-06, + "loss": 0.0411, + "step": 39751 + }, + { + "epoch": 0.9318494816481772, + "grad_norm": 0.36640268564224243, + "learning_rate": 2.2839257523088575e-06, + "loss": 0.091, + "step": 39752 + }, + { + "epoch": 0.9318729232229821, + "grad_norm": 0.26784592866897583, + "learning_rate": 2.282360697898678e-06, + "loss": 0.0491, + "step": 39753 + }, + { + "epoch": 0.9318963647977871, + "grad_norm": 0.5511229038238525, + "learning_rate": 2.280796173707145e-06, + "loss": 0.0608, + "step": 39754 + }, + { + "epoch": 0.9319198063725921, + "grad_norm": 0.17108899354934692, + "learning_rate": 2.279232179742763e-06, + "loss": 0.0319, + "step": 39755 + }, + { + "epoch": 0.9319432479473971, + "grad_norm": 0.14084897935390472, + "learning_rate": 2.2776687160140144e-06, + "loss": 0.0249, + "step": 39756 + }, + { + "epoch": 0.9319666895222021, + "grad_norm": 0.9584654569625854, + "learning_rate": 2.276105782529381e-06, + "loss": 0.0311, + "step": 39757 + }, + { + "epoch": 0.9319901310970071, + "grad_norm": 0.4068286120891571, + "learning_rate": 2.274543379297345e-06, + "loss": 0.0572, + "step": 39758 + }, + { + "epoch": 0.9320135726718121, + "grad_norm": 0.2912074625492096, + "learning_rate": 2.272981506326388e-06, + "loss": 0.0378, + "step": 39759 + }, + { + "epoch": 0.9320370142466171, + "grad_norm": 0.29090064764022827, + "learning_rate": 2.271420163624971e-06, + "loss": 0.0469, + "step": 39760 + }, + { + "epoch": 0.932060455821422, + "grad_norm": 0.3822108209133148, + "learning_rate": 2.2698593512015863e-06, + "loss": 0.0653, + "step": 39761 + }, + { + "epoch": 0.9320838973962271, + "grad_norm": 0.3896195590496063, + "learning_rate": 2.2682990690646945e-06, + "loss": 0.3551, + "step": 39762 + }, + { + "epoch": 0.932107338971032, + "grad_norm": 0.5562012195587158, + "learning_rate": 2.266739317222766e-06, + "loss": 0.1343, + "step": 39763 + }, + { + "epoch": 0.9321307805458371, + "grad_norm": 0.4159952700138092, + "learning_rate": 2.2651800956842385e-06, + "loss": 0.4323, + "step": 39764 + }, + { + "epoch": 0.932154222120642, + "grad_norm": 0.40563103556632996, + "learning_rate": 2.2636214044576053e-06, + "loss": 0.0465, + "step": 39765 + }, + { + "epoch": 0.9321776636954471, + "grad_norm": 0.38877788186073303, + "learning_rate": 2.2620632435513046e-06, + "loss": 0.0751, + "step": 39766 + }, + { + "epoch": 0.932201105270252, + "grad_norm": 0.22983206808567047, + "learning_rate": 2.2605056129738065e-06, + "loss": 0.0428, + "step": 39767 + }, + { + "epoch": 0.9322245468450571, + "grad_norm": 0.5186128616333008, + "learning_rate": 2.2589485127335495e-06, + "loss": 0.0351, + "step": 39768 + }, + { + "epoch": 0.932247988419862, + "grad_norm": 0.7066411375999451, + "learning_rate": 2.2573919428389823e-06, + "loss": 0.1486, + "step": 39769 + }, + { + "epoch": 0.9322714299946671, + "grad_norm": 0.534591794013977, + "learning_rate": 2.2558359032985422e-06, + "loss": 0.0599, + "step": 39770 + }, + { + "epoch": 0.932294871569472, + "grad_norm": 0.17761240899562836, + "learning_rate": 2.2542803941207e-06, + "loss": 0.046, + "step": 39771 + }, + { + "epoch": 0.932318313144277, + "grad_norm": 0.2727435827255249, + "learning_rate": 2.2527254153138723e-06, + "loss": 0.0331, + "step": 39772 + }, + { + "epoch": 0.932341754719082, + "grad_norm": 0.507030725479126, + "learning_rate": 2.251170966886518e-06, + "loss": 0.5593, + "step": 39773 + }, + { + "epoch": 0.932365196293887, + "grad_norm": 0.43596524000167847, + "learning_rate": 2.2496170488470525e-06, + "loss": 0.0804, + "step": 39774 + }, + { + "epoch": 0.932388637868692, + "grad_norm": 0.16436007618904114, + "learning_rate": 2.248063661203914e-06, + "loss": 0.0287, + "step": 39775 + }, + { + "epoch": 0.932412079443497, + "grad_norm": 0.5894575715065002, + "learning_rate": 2.246510803965529e-06, + "loss": 0.1104, + "step": 39776 + }, + { + "epoch": 0.932435521018302, + "grad_norm": 0.8287805914878845, + "learning_rate": 2.2449584771403243e-06, + "loss": 0.1104, + "step": 39777 + }, + { + "epoch": 0.932458962593107, + "grad_norm": 0.457733690738678, + "learning_rate": 2.243406680736726e-06, + "loss": 0.0873, + "step": 39778 + }, + { + "epoch": 0.932482404167912, + "grad_norm": 0.18081286549568176, + "learning_rate": 2.24185541476315e-06, + "loss": 0.031, + "step": 39779 + }, + { + "epoch": 0.932505845742717, + "grad_norm": 0.28023475408554077, + "learning_rate": 2.240304679228011e-06, + "loss": 0.0566, + "step": 39780 + }, + { + "epoch": 0.9325292873175219, + "grad_norm": 0.5292462706565857, + "learning_rate": 2.238754474139737e-06, + "loss": 0.0956, + "step": 39781 + }, + { + "epoch": 0.932552728892327, + "grad_norm": 0.45074567198753357, + "learning_rate": 2.23720479950672e-06, + "loss": 0.0367, + "step": 39782 + }, + { + "epoch": 0.932576170467132, + "grad_norm": 0.13738663494586945, + "learning_rate": 2.2356556553373874e-06, + "loss": 0.0174, + "step": 39783 + }, + { + "epoch": 0.932599612041937, + "grad_norm": 0.43093836307525635, + "learning_rate": 2.234107041640121e-06, + "loss": 0.4662, + "step": 39784 + }, + { + "epoch": 0.932623053616742, + "grad_norm": 0.6419489979743958, + "learning_rate": 2.2325589584233475e-06, + "loss": 0.1282, + "step": 39785 + }, + { + "epoch": 0.932646495191547, + "grad_norm": 0.3366938531398773, + "learning_rate": 2.2310114056954713e-06, + "loss": 0.0715, + "step": 39786 + }, + { + "epoch": 0.932669936766352, + "grad_norm": 0.5749243497848511, + "learning_rate": 2.229464383464863e-06, + "loss": 0.0827, + "step": 39787 + }, + { + "epoch": 0.932693378341157, + "grad_norm": 0.4086943566799164, + "learning_rate": 2.227917891739939e-06, + "loss": 0.079, + "step": 39788 + }, + { + "epoch": 0.932716819915962, + "grad_norm": 0.34371018409729004, + "learning_rate": 2.2263719305290585e-06, + "loss": 0.0483, + "step": 39789 + }, + { + "epoch": 0.9327402614907669, + "grad_norm": 0.5443378686904907, + "learning_rate": 2.22482649984066e-06, + "loss": 0.6521, + "step": 39790 + }, + { + "epoch": 0.932763703065572, + "grad_norm": 0.35348886251449585, + "learning_rate": 2.2232815996830912e-06, + "loss": 0.041, + "step": 39791 + }, + { + "epoch": 0.9327871446403769, + "grad_norm": 0.662693202495575, + "learning_rate": 2.2217372300647463e-06, + "loss": 0.1285, + "step": 39792 + }, + { + "epoch": 0.932810586215182, + "grad_norm": 0.5192945003509521, + "learning_rate": 2.2201933909940074e-06, + "loss": 0.1173, + "step": 39793 + }, + { + "epoch": 0.9328340277899869, + "grad_norm": 0.6260468363761902, + "learning_rate": 2.2186500824792457e-06, + "loss": 0.079, + "step": 39794 + }, + { + "epoch": 0.932857469364792, + "grad_norm": 0.4656951427459717, + "learning_rate": 2.2171073045288316e-06, + "loss": 0.0597, + "step": 39795 + }, + { + "epoch": 0.9328809109395969, + "grad_norm": 0.5174257755279541, + "learning_rate": 2.2155650571511367e-06, + "loss": 0.1, + "step": 39796 + }, + { + "epoch": 0.932904352514402, + "grad_norm": 0.19002938270568848, + "learning_rate": 2.214023340354554e-06, + "loss": 0.0191, + "step": 39797 + }, + { + "epoch": 0.9329277940892069, + "grad_norm": 0.4015914797782898, + "learning_rate": 2.2124821541474217e-06, + "loss": 0.4234, + "step": 39798 + }, + { + "epoch": 0.932951235664012, + "grad_norm": 0.4219591021537781, + "learning_rate": 2.2109414985381216e-06, + "loss": 0.0589, + "step": 39799 + }, + { + "epoch": 0.9329746772388169, + "grad_norm": 0.3495115041732788, + "learning_rate": 2.2094013735349917e-06, + "loss": 0.0627, + "step": 39800 + }, + { + "epoch": 0.9329981188136219, + "grad_norm": 0.621571958065033, + "learning_rate": 2.207861779146403e-06, + "loss": 0.1387, + "step": 39801 + }, + { + "epoch": 0.9330215603884269, + "grad_norm": 0.17533688247203827, + "learning_rate": 2.2063227153806933e-06, + "loss": 0.0226, + "step": 39802 + }, + { + "epoch": 0.9330450019632319, + "grad_norm": 0.5817431807518005, + "learning_rate": 2.2047841822462444e-06, + "loss": 0.1309, + "step": 39803 + }, + { + "epoch": 0.9330684435380369, + "grad_norm": 0.7780975699424744, + "learning_rate": 2.2032461797513837e-06, + "loss": 0.1476, + "step": 39804 + }, + { + "epoch": 0.9330918851128419, + "grad_norm": 0.6225758194923401, + "learning_rate": 2.2017087079044486e-06, + "loss": 0.078, + "step": 39805 + }, + { + "epoch": 0.9331153266876469, + "grad_norm": 0.17850695550441742, + "learning_rate": 2.200171766713799e-06, + "loss": 0.0358, + "step": 39806 + }, + { + "epoch": 0.9331387682624519, + "grad_norm": 3.6148252487182617, + "learning_rate": 2.1986353561877727e-06, + "loss": 0.0847, + "step": 39807 + }, + { + "epoch": 0.9331622098372568, + "grad_norm": 0.3962526023387909, + "learning_rate": 2.197099476334685e-06, + "loss": 0.0779, + "step": 39808 + }, + { + "epoch": 0.9331856514120619, + "grad_norm": 0.4315115511417389, + "learning_rate": 2.195564127162886e-06, + "loss": 0.0673, + "step": 39809 + }, + { + "epoch": 0.9332090929868668, + "grad_norm": 0.2542467415332794, + "learning_rate": 2.1940293086807228e-06, + "loss": 0.0384, + "step": 39810 + }, + { + "epoch": 0.9332325345616719, + "grad_norm": 0.17885196208953857, + "learning_rate": 2.1924950208965013e-06, + "loss": 0.0255, + "step": 39811 + }, + { + "epoch": 0.9332559761364768, + "grad_norm": 0.5441973805427551, + "learning_rate": 2.1909612638185473e-06, + "loss": 0.6195, + "step": 39812 + }, + { + "epoch": 0.9332794177112819, + "grad_norm": 0.3633575439453125, + "learning_rate": 2.189428037455188e-06, + "loss": 0.0526, + "step": 39813 + }, + { + "epoch": 0.9333028592860868, + "grad_norm": 0.3374120891094208, + "learning_rate": 2.1878953418147498e-06, + "loss": 0.0456, + "step": 39814 + }, + { + "epoch": 0.9333263008608919, + "grad_norm": 0.337390661239624, + "learning_rate": 2.186363176905526e-06, + "loss": 0.0587, + "step": 39815 + }, + { + "epoch": 0.9333497424356968, + "grad_norm": 0.38943493366241455, + "learning_rate": 2.1848315427358655e-06, + "loss": 0.1116, + "step": 39816 + }, + { + "epoch": 0.9333731840105018, + "grad_norm": 0.1788061112165451, + "learning_rate": 2.183300439314051e-06, + "loss": 0.0381, + "step": 39817 + }, + { + "epoch": 0.9333966255853068, + "grad_norm": 0.7945706844329834, + "learning_rate": 2.1817698666483976e-06, + "loss": 0.1527, + "step": 39818 + }, + { + "epoch": 0.9334200671601118, + "grad_norm": 0.3879442512989044, + "learning_rate": 2.18023982474721e-06, + "loss": 0.058, + "step": 39819 + }, + { + "epoch": 0.9334435087349168, + "grad_norm": 0.891538679599762, + "learning_rate": 2.1787103136187923e-06, + "loss": 0.1613, + "step": 39820 + }, + { + "epoch": 0.9334669503097218, + "grad_norm": 0.15402323007583618, + "learning_rate": 2.1771813332714385e-06, + "loss": 0.0225, + "step": 39821 + }, + { + "epoch": 0.9334903918845268, + "grad_norm": 0.2114797681570053, + "learning_rate": 2.1756528837134525e-06, + "loss": 0.0156, + "step": 39822 + }, + { + "epoch": 0.9335138334593318, + "grad_norm": 0.1768581122159958, + "learning_rate": 2.174124964953128e-06, + "loss": 0.0316, + "step": 39823 + }, + { + "epoch": 0.9335372750341367, + "grad_norm": 0.23899199068546295, + "learning_rate": 2.1725975769987474e-06, + "loss": 0.2972, + "step": 39824 + }, + { + "epoch": 0.9335607166089418, + "grad_norm": 0.29180634021759033, + "learning_rate": 2.1710707198586035e-06, + "loss": 0.0715, + "step": 39825 + }, + { + "epoch": 0.9335841581837467, + "grad_norm": 0.5221964120864868, + "learning_rate": 2.16954439354099e-06, + "loss": 0.0594, + "step": 39826 + }, + { + "epoch": 0.9336075997585518, + "grad_norm": 0.12692026793956757, + "learning_rate": 2.1680185980541666e-06, + "loss": 0.0176, + "step": 39827 + }, + { + "epoch": 0.9336310413333567, + "grad_norm": 0.4879216253757477, + "learning_rate": 2.1664933334064274e-06, + "loss": 0.072, + "step": 39828 + }, + { + "epoch": 0.9336544829081618, + "grad_norm": 0.7960972189903259, + "learning_rate": 2.1649685996060543e-06, + "loss": 0.1387, + "step": 39829 + }, + { + "epoch": 0.9336779244829667, + "grad_norm": 0.16702106595039368, + "learning_rate": 2.163444396661307e-06, + "loss": 0.0215, + "step": 39830 + }, + { + "epoch": 0.9337013660577718, + "grad_norm": 0.23617783188819885, + "learning_rate": 2.1619207245804685e-06, + "loss": 0.0352, + "step": 39831 + }, + { + "epoch": 0.9337248076325767, + "grad_norm": 0.3057516813278198, + "learning_rate": 2.1603975833717982e-06, + "loss": 0.0605, + "step": 39832 + }, + { + "epoch": 0.9337482492073818, + "grad_norm": 0.47384732961654663, + "learning_rate": 2.1588749730435455e-06, + "loss": 0.0822, + "step": 39833 + }, + { + "epoch": 0.9337716907821867, + "grad_norm": 0.44182541966438293, + "learning_rate": 2.157352893603992e-06, + "loss": 0.0334, + "step": 39834 + }, + { + "epoch": 0.9337951323569917, + "grad_norm": 0.45902127027511597, + "learning_rate": 2.1558313450613987e-06, + "loss": 0.1012, + "step": 39835 + }, + { + "epoch": 0.9338185739317968, + "grad_norm": 0.8044886589050293, + "learning_rate": 2.154310327424014e-06, + "loss": 0.1019, + "step": 39836 + }, + { + "epoch": 0.9338420155066017, + "grad_norm": 0.3695215582847595, + "learning_rate": 2.152789840700098e-06, + "loss": 0.0683, + "step": 39837 + }, + { + "epoch": 0.9338654570814068, + "grad_norm": 0.7139784097671509, + "learning_rate": 2.1512698848978995e-06, + "loss": 0.0601, + "step": 39838 + }, + { + "epoch": 0.9338888986562117, + "grad_norm": 0.35796216130256653, + "learning_rate": 2.1497504600256566e-06, + "loss": 0.0499, + "step": 39839 + }, + { + "epoch": 0.9339123402310168, + "grad_norm": 0.5064833760261536, + "learning_rate": 2.1482315660916074e-06, + "loss": 0.0918, + "step": 39840 + }, + { + "epoch": 0.9339357818058217, + "grad_norm": 0.15341366827487946, + "learning_rate": 2.1467132031040227e-06, + "loss": 0.0134, + "step": 39841 + }, + { + "epoch": 0.9339592233806268, + "grad_norm": 0.0812474936246872, + "learning_rate": 2.1451953710711182e-06, + "loss": 0.012, + "step": 39842 + }, + { + "epoch": 0.9339826649554317, + "grad_norm": 0.8982177376747131, + "learning_rate": 2.143678070001132e-06, + "loss": 0.1321, + "step": 39843 + }, + { + "epoch": 0.9340061065302367, + "grad_norm": 0.20340125262737274, + "learning_rate": 2.142161299902301e-06, + "loss": 0.0231, + "step": 39844 + }, + { + "epoch": 0.9340295481050417, + "grad_norm": 0.4699096381664276, + "learning_rate": 2.140645060782853e-06, + "loss": 0.4786, + "step": 39845 + }, + { + "epoch": 0.9340529896798467, + "grad_norm": 0.31478703022003174, + "learning_rate": 2.139129352651015e-06, + "loss": 0.0474, + "step": 39846 + }, + { + "epoch": 0.9340764312546517, + "grad_norm": 0.3721247613430023, + "learning_rate": 2.1376141755150125e-06, + "loss": 0.076, + "step": 39847 + }, + { + "epoch": 0.9340998728294567, + "grad_norm": 0.46824711561203003, + "learning_rate": 2.1360995293830733e-06, + "loss": 0.0555, + "step": 39848 + }, + { + "epoch": 0.9341233144042617, + "grad_norm": 0.33606740832328796, + "learning_rate": 2.1345854142634015e-06, + "loss": 0.0573, + "step": 39849 + }, + { + "epoch": 0.9341467559790667, + "grad_norm": 0.5507623553276062, + "learning_rate": 2.133071830164235e-06, + "loss": 0.0886, + "step": 39850 + }, + { + "epoch": 0.9341701975538717, + "grad_norm": 0.45876920223236084, + "learning_rate": 2.1315587770937563e-06, + "loss": 0.0588, + "step": 39851 + }, + { + "epoch": 0.9341936391286767, + "grad_norm": 0.4421696066856384, + "learning_rate": 2.130046255060203e-06, + "loss": 0.5586, + "step": 39852 + }, + { + "epoch": 0.9342170807034816, + "grad_norm": 0.17115478217601776, + "learning_rate": 2.1285342640717575e-06, + "loss": 0.0183, + "step": 39853 + }, + { + "epoch": 0.9342405222782867, + "grad_norm": 0.3550136387348175, + "learning_rate": 2.127022804136647e-06, + "loss": 0.079, + "step": 39854 + }, + { + "epoch": 0.9342639638530916, + "grad_norm": 0.38201725482940674, + "learning_rate": 2.125511875263064e-06, + "loss": 0.079, + "step": 39855 + }, + { + "epoch": 0.9342874054278967, + "grad_norm": 0.08714677393436432, + "learning_rate": 2.1240014774592033e-06, + "loss": 0.0139, + "step": 39856 + }, + { + "epoch": 0.9343108470027016, + "grad_norm": 0.27337273955345154, + "learning_rate": 2.1224916107332572e-06, + "loss": 0.0289, + "step": 39857 + }, + { + "epoch": 0.9343342885775067, + "grad_norm": 0.5645495653152466, + "learning_rate": 2.1209822750934306e-06, + "loss": 0.5208, + "step": 39858 + }, + { + "epoch": 0.9343577301523116, + "grad_norm": 0.4672379195690155, + "learning_rate": 2.1194734705479057e-06, + "loss": 0.0449, + "step": 39859 + }, + { + "epoch": 0.9343811717271167, + "grad_norm": 0.40270155668258667, + "learning_rate": 2.1179651971048654e-06, + "loss": 0.0828, + "step": 39860 + }, + { + "epoch": 0.9344046133019216, + "grad_norm": 0.5558097958564758, + "learning_rate": 2.1164574547725025e-06, + "loss": 0.1284, + "step": 39861 + }, + { + "epoch": 0.9344280548767266, + "grad_norm": 0.14399218559265137, + "learning_rate": 2.1149502435589995e-06, + "loss": 0.0258, + "step": 39862 + }, + { + "epoch": 0.9344514964515316, + "grad_norm": 0.40845030546188354, + "learning_rate": 2.113443563472517e-06, + "loss": 0.0504, + "step": 39863 + }, + { + "epoch": 0.9344749380263366, + "grad_norm": 0.37319788336753845, + "learning_rate": 2.111937414521248e-06, + "loss": 0.4037, + "step": 39864 + }, + { + "epoch": 0.9344983796011416, + "grad_norm": 0.428621381521225, + "learning_rate": 2.1104317967133524e-06, + "loss": 0.0691, + "step": 39865 + }, + { + "epoch": 0.9345218211759466, + "grad_norm": 0.13919378817081451, + "learning_rate": 2.1089267100570133e-06, + "loss": 0.0081, + "step": 39866 + }, + { + "epoch": 0.9345452627507516, + "grad_norm": 0.5662994384765625, + "learning_rate": 2.1074221545603903e-06, + "loss": 0.0581, + "step": 39867 + }, + { + "epoch": 0.9345687043255566, + "grad_norm": 0.30460798740386963, + "learning_rate": 2.1059181302316546e-06, + "loss": 0.0476, + "step": 39868 + }, + { + "epoch": 0.9345921459003615, + "grad_norm": 0.3321567177772522, + "learning_rate": 2.104414637078944e-06, + "loss": 0.0374, + "step": 39869 + }, + { + "epoch": 0.9346155874751666, + "grad_norm": 0.11324702203273773, + "learning_rate": 2.1029116751104416e-06, + "loss": 0.0191, + "step": 39870 + }, + { + "epoch": 0.9346390290499715, + "grad_norm": 0.13179919123649597, + "learning_rate": 2.101409244334285e-06, + "loss": 0.0218, + "step": 39871 + }, + { + "epoch": 0.9346624706247766, + "grad_norm": 0.21291550993919373, + "learning_rate": 2.0999073447586336e-06, + "loss": 0.0482, + "step": 39872 + }, + { + "epoch": 0.9346859121995815, + "grad_norm": 0.3824774920940399, + "learning_rate": 2.0984059763916487e-06, + "loss": 0.114, + "step": 39873 + }, + { + "epoch": 0.9347093537743866, + "grad_norm": 0.5526514053344727, + "learning_rate": 2.0969051392414674e-06, + "loss": 0.0985, + "step": 39874 + }, + { + "epoch": 0.9347327953491915, + "grad_norm": 0.42289459705352783, + "learning_rate": 2.095404833316217e-06, + "loss": 0.0798, + "step": 39875 + }, + { + "epoch": 0.9347562369239966, + "grad_norm": 0.478893905878067, + "learning_rate": 2.0939050586240684e-06, + "loss": 0.0873, + "step": 39876 + }, + { + "epoch": 0.9347796784988015, + "grad_norm": 0.37110403180122375, + "learning_rate": 2.0924058151731374e-06, + "loss": 0.0676, + "step": 39877 + }, + { + "epoch": 0.9348031200736066, + "grad_norm": 0.41414162516593933, + "learning_rate": 2.090907102971562e-06, + "loss": 0.0643, + "step": 39878 + }, + { + "epoch": 0.9348265616484115, + "grad_norm": 0.11317402124404907, + "learning_rate": 2.0894089220274805e-06, + "loss": 0.0129, + "step": 39879 + }, + { + "epoch": 0.9348500032232165, + "grad_norm": 0.2839702069759369, + "learning_rate": 2.0879112723490193e-06, + "loss": 0.0399, + "step": 39880 + }, + { + "epoch": 0.9348734447980215, + "grad_norm": 0.32103490829467773, + "learning_rate": 2.086414153944294e-06, + "loss": 0.0428, + "step": 39881 + }, + { + "epoch": 0.9348968863728265, + "grad_norm": 0.3185843527317047, + "learning_rate": 2.084917566821454e-06, + "loss": 0.0396, + "step": 39882 + }, + { + "epoch": 0.9349203279476315, + "grad_norm": 0.31393635272979736, + "learning_rate": 2.0834215109886036e-06, + "loss": 0.0632, + "step": 39883 + }, + { + "epoch": 0.9349437695224365, + "grad_norm": 0.08483851701021194, + "learning_rate": 2.081925986453859e-06, + "loss": 0.0094, + "step": 39884 + }, + { + "epoch": 0.9349672110972415, + "grad_norm": 0.6364064812660217, + "learning_rate": 2.080430993225324e-06, + "loss": 0.161, + "step": 39885 + }, + { + "epoch": 0.9349906526720465, + "grad_norm": 0.44032520055770874, + "learning_rate": 2.0789365313111374e-06, + "loss": 0.0852, + "step": 39886 + }, + { + "epoch": 0.9350140942468516, + "grad_norm": 0.38289588689804077, + "learning_rate": 2.0774426007193924e-06, + "loss": 0.0736, + "step": 39887 + }, + { + "epoch": 0.9350375358216565, + "grad_norm": 0.19780461490154266, + "learning_rate": 2.075949201458194e-06, + "loss": 0.0317, + "step": 39888 + }, + { + "epoch": 0.9350609773964615, + "grad_norm": 0.2934314012527466, + "learning_rate": 2.074456333535657e-06, + "loss": 0.0373, + "step": 39889 + }, + { + "epoch": 0.9350844189712665, + "grad_norm": 0.5201332569122314, + "learning_rate": 2.0729639969598647e-06, + "loss": 0.7209, + "step": 39890 + }, + { + "epoch": 0.9351078605460715, + "grad_norm": 0.1287413090467453, + "learning_rate": 2.071472191738921e-06, + "loss": 0.0192, + "step": 39891 + }, + { + "epoch": 0.9351313021208765, + "grad_norm": 0.21106064319610596, + "learning_rate": 2.0699809178809314e-06, + "loss": 0.0497, + "step": 39892 + }, + { + "epoch": 0.9351547436956815, + "grad_norm": 0.5952029824256897, + "learning_rate": 2.0684901753939777e-06, + "loss": 0.1549, + "step": 39893 + }, + { + "epoch": 0.9351781852704865, + "grad_norm": 0.37468820810317993, + "learning_rate": 2.0669999642861427e-06, + "loss": 0.0676, + "step": 39894 + }, + { + "epoch": 0.9352016268452915, + "grad_norm": 0.2782238721847534, + "learning_rate": 2.0655102845655194e-06, + "loss": 0.0477, + "step": 39895 + }, + { + "epoch": 0.9352250684200965, + "grad_norm": 0.3180925250053406, + "learning_rate": 2.0640211362402017e-06, + "loss": 0.0448, + "step": 39896 + }, + { + "epoch": 0.9352485099949015, + "grad_norm": 0.6667543053627014, + "learning_rate": 2.06253251931825e-06, + "loss": 0.0581, + "step": 39897 + }, + { + "epoch": 0.9352719515697064, + "grad_norm": 0.12605898082256317, + "learning_rate": 2.0610444338077463e-06, + "loss": 0.0238, + "step": 39898 + }, + { + "epoch": 0.9352953931445115, + "grad_norm": 0.17485512793064117, + "learning_rate": 2.059556879716784e-06, + "loss": 0.0271, + "step": 39899 + }, + { + "epoch": 0.9353188347193164, + "grad_norm": 0.3810018002986908, + "learning_rate": 2.058069857053413e-06, + "loss": 0.0613, + "step": 39900 + }, + { + "epoch": 0.9353422762941215, + "grad_norm": 0.27018478512763977, + "learning_rate": 2.056583365825715e-06, + "loss": 0.0365, + "step": 39901 + }, + { + "epoch": 0.9353657178689264, + "grad_norm": 0.4021781384944916, + "learning_rate": 2.0550974060417394e-06, + "loss": 0.101, + "step": 39902 + }, + { + "epoch": 0.9353891594437315, + "grad_norm": 0.12645192444324493, + "learning_rate": 2.0536119777095576e-06, + "loss": 0.0102, + "step": 39903 + }, + { + "epoch": 0.9354126010185364, + "grad_norm": 0.9490944147109985, + "learning_rate": 2.0521270808372406e-06, + "loss": 0.8434, + "step": 39904 + }, + { + "epoch": 0.9354360425933415, + "grad_norm": 0.18316113948822021, + "learning_rate": 2.0506427154328266e-06, + "loss": 0.0175, + "step": 39905 + }, + { + "epoch": 0.9354594841681464, + "grad_norm": 0.3381277620792389, + "learning_rate": 2.049158881504387e-06, + "loss": 0.0447, + "step": 39906 + }, + { + "epoch": 0.9354829257429514, + "grad_norm": 0.36810341477394104, + "learning_rate": 2.04767557905996e-06, + "loss": 0.0537, + "step": 39907 + }, + { + "epoch": 0.9355063673177564, + "grad_norm": 0.37612417340278625, + "learning_rate": 2.0461928081076055e-06, + "loss": 0.0613, + "step": 39908 + }, + { + "epoch": 0.9355298088925614, + "grad_norm": 0.39481234550476074, + "learning_rate": 2.0447105686553616e-06, + "loss": 0.0643, + "step": 39909 + }, + { + "epoch": 0.9355532504673664, + "grad_norm": 0.18108835816383362, + "learning_rate": 2.043228860711266e-06, + "loss": 0.0114, + "step": 39910 + }, + { + "epoch": 0.9355766920421714, + "grad_norm": 0.15577900409698486, + "learning_rate": 2.041747684283368e-06, + "loss": 0.0148, + "step": 39911 + }, + { + "epoch": 0.9356001336169764, + "grad_norm": 0.2947985827922821, + "learning_rate": 2.0402670393797064e-06, + "loss": 0.034, + "step": 39912 + }, + { + "epoch": 0.9356235751917814, + "grad_norm": 0.48127222061157227, + "learning_rate": 2.038786926008307e-06, + "loss": 0.0676, + "step": 39913 + }, + { + "epoch": 0.9356470167665863, + "grad_norm": 0.3834303915500641, + "learning_rate": 2.0373073441772085e-06, + "loss": 0.6257, + "step": 39914 + }, + { + "epoch": 0.9356704583413914, + "grad_norm": 0.5297268629074097, + "learning_rate": 2.0358282938944372e-06, + "loss": 0.1027, + "step": 39915 + }, + { + "epoch": 0.9356938999161963, + "grad_norm": 0.36045584082603455, + "learning_rate": 2.0343497751679984e-06, + "loss": 0.0571, + "step": 39916 + }, + { + "epoch": 0.9357173414910014, + "grad_norm": 0.2647687792778015, + "learning_rate": 2.0328717880059413e-06, + "loss": 0.0316, + "step": 39917 + }, + { + "epoch": 0.9357407830658063, + "grad_norm": 0.5260137915611267, + "learning_rate": 2.03139433241627e-06, + "loss": 0.41, + "step": 39918 + }, + { + "epoch": 0.9357642246406114, + "grad_norm": 0.4670761525630951, + "learning_rate": 2.029917408407023e-06, + "loss": 0.0807, + "step": 39919 + }, + { + "epoch": 0.9357876662154163, + "grad_norm": 0.1244317889213562, + "learning_rate": 2.0284410159861934e-06, + "loss": 0.0179, + "step": 39920 + }, + { + "epoch": 0.9358111077902214, + "grad_norm": 0.4941141605377197, + "learning_rate": 2.0269651551617974e-06, + "loss": 0.5351, + "step": 39921 + }, + { + "epoch": 0.9358345493650263, + "grad_norm": 0.3392408788204193, + "learning_rate": 2.02548982594184e-06, + "loss": 0.0506, + "step": 39922 + }, + { + "epoch": 0.9358579909398314, + "grad_norm": 0.5743575692176819, + "learning_rate": 2.024015028334325e-06, + "loss": 0.0715, + "step": 39923 + }, + { + "epoch": 0.9358814325146363, + "grad_norm": 0.1625141054391861, + "learning_rate": 2.022540762347269e-06, + "loss": 0.0183, + "step": 39924 + }, + { + "epoch": 0.9359048740894413, + "grad_norm": 0.5664370656013489, + "learning_rate": 2.0210670279886657e-06, + "loss": 0.0226, + "step": 39925 + }, + { + "epoch": 0.9359283156642463, + "grad_norm": 0.4154132306575775, + "learning_rate": 2.019593825266497e-06, + "loss": 0.0724, + "step": 39926 + }, + { + "epoch": 0.9359517572390513, + "grad_norm": 0.4317761957645416, + "learning_rate": 2.018121154188779e-06, + "loss": 0.0588, + "step": 39927 + }, + { + "epoch": 0.9359751988138563, + "grad_norm": 0.1653164029121399, + "learning_rate": 2.016649014763483e-06, + "loss": 0.0186, + "step": 39928 + }, + { + "epoch": 0.9359986403886613, + "grad_norm": 0.45816564559936523, + "learning_rate": 2.015177406998592e-06, + "loss": 0.0606, + "step": 39929 + }, + { + "epoch": 0.9360220819634663, + "grad_norm": 0.2822161316871643, + "learning_rate": 2.0137063309021098e-06, + "loss": 0.0385, + "step": 39930 + }, + { + "epoch": 0.9360455235382713, + "grad_norm": 0.38961631059646606, + "learning_rate": 2.0122357864820198e-06, + "loss": 0.5416, + "step": 39931 + }, + { + "epoch": 0.9360689651130762, + "grad_norm": 0.3639504611492157, + "learning_rate": 2.0107657737462925e-06, + "loss": 0.0578, + "step": 39932 + }, + { + "epoch": 0.9360924066878813, + "grad_norm": 0.6332089304924011, + "learning_rate": 2.0092962927029112e-06, + "loss": 0.1309, + "step": 39933 + }, + { + "epoch": 0.9361158482626862, + "grad_norm": 0.15747234225273132, + "learning_rate": 2.007827343359836e-06, + "loss": 0.0435, + "step": 39934 + }, + { + "epoch": 0.9361392898374913, + "grad_norm": 0.23637071251869202, + "learning_rate": 2.006358925725049e-06, + "loss": 0.0142, + "step": 39935 + }, + { + "epoch": 0.9361627314122962, + "grad_norm": 0.41859450936317444, + "learning_rate": 2.0048910398065e-06, + "loss": 0.0683, + "step": 39936 + }, + { + "epoch": 0.9361861729871013, + "grad_norm": 0.2558039426803589, + "learning_rate": 2.0034236856121825e-06, + "loss": 0.0175, + "step": 39937 + }, + { + "epoch": 0.9362096145619063, + "grad_norm": 0.38312914967536926, + "learning_rate": 2.001956863150034e-06, + "loss": 0.0778, + "step": 39938 + }, + { + "epoch": 0.9362330561367113, + "grad_norm": 0.16140514612197876, + "learning_rate": 2.0004905724280378e-06, + "loss": 0.0129, + "step": 39939 + }, + { + "epoch": 0.9362564977115163, + "grad_norm": 0.3520386815071106, + "learning_rate": 1.99902481345412e-06, + "loss": 0.3825, + "step": 39940 + }, + { + "epoch": 0.9362799392863212, + "grad_norm": 0.19624769687652588, + "learning_rate": 1.997559586236253e-06, + "loss": 0.0334, + "step": 39941 + }, + { + "epoch": 0.9363033808611263, + "grad_norm": 0.5095928311347961, + "learning_rate": 1.996094890782385e-06, + "loss": 0.7033, + "step": 39942 + }, + { + "epoch": 0.9363268224359312, + "grad_norm": 0.38190263509750366, + "learning_rate": 1.994630727100455e-06, + "loss": 0.0622, + "step": 39943 + }, + { + "epoch": 0.9363502640107363, + "grad_norm": 0.26216867566108704, + "learning_rate": 1.9931670951984225e-06, + "loss": 0.0462, + "step": 39944 + }, + { + "epoch": 0.9363737055855412, + "grad_norm": 0.2994323670864105, + "learning_rate": 1.9917039950842265e-06, + "loss": 0.0246, + "step": 39945 + }, + { + "epoch": 0.9363971471603463, + "grad_norm": 0.7362661957740784, + "learning_rate": 1.9902414267657933e-06, + "loss": 0.1251, + "step": 39946 + }, + { + "epoch": 0.9364205887351512, + "grad_norm": 0.4192260801792145, + "learning_rate": 1.988779390251061e-06, + "loss": 0.0468, + "step": 39947 + }, + { + "epoch": 0.9364440303099563, + "grad_norm": 0.2789062261581421, + "learning_rate": 1.987317885547968e-06, + "loss": 0.063, + "step": 39948 + }, + { + "epoch": 0.9364674718847612, + "grad_norm": 0.45671775937080383, + "learning_rate": 1.9858569126644523e-06, + "loss": 0.0408, + "step": 39949 + }, + { + "epoch": 0.9364909134595663, + "grad_norm": 0.33015820384025574, + "learning_rate": 1.9843964716084297e-06, + "loss": 0.0476, + "step": 39950 + }, + { + "epoch": 0.9365143550343712, + "grad_norm": 0.5180457234382629, + "learning_rate": 1.982936562387827e-06, + "loss": 0.3497, + "step": 39951 + }, + { + "epoch": 0.9365377966091762, + "grad_norm": 0.20762914419174194, + "learning_rate": 1.981477185010572e-06, + "loss": 0.0294, + "step": 39952 + }, + { + "epoch": 0.9365612381839812, + "grad_norm": 0.20658549666404724, + "learning_rate": 1.9800183394845683e-06, + "loss": 0.0251, + "step": 39953 + }, + { + "epoch": 0.9365846797587862, + "grad_norm": 0.5235212445259094, + "learning_rate": 1.978560025817744e-06, + "loss": 0.639, + "step": 39954 + }, + { + "epoch": 0.9366081213335912, + "grad_norm": 0.31186550855636597, + "learning_rate": 1.977102244018003e-06, + "loss": 0.0576, + "step": 39955 + }, + { + "epoch": 0.9366315629083962, + "grad_norm": 0.3903229236602783, + "learning_rate": 1.9756449940932733e-06, + "loss": 0.0166, + "step": 39956 + }, + { + "epoch": 0.9366550044832012, + "grad_norm": 0.34047985076904297, + "learning_rate": 1.9741882760514474e-06, + "loss": 0.0559, + "step": 39957 + }, + { + "epoch": 0.9366784460580062, + "grad_norm": 0.3441188335418701, + "learning_rate": 1.9727320899004308e-06, + "loss": 0.0547, + "step": 39958 + }, + { + "epoch": 0.9367018876328111, + "grad_norm": 0.1465883105993271, + "learning_rate": 1.9712764356481283e-06, + "loss": 0.0247, + "step": 39959 + }, + { + "epoch": 0.9367253292076162, + "grad_norm": 0.4010471999645233, + "learning_rate": 1.9698213133024334e-06, + "loss": 0.0583, + "step": 39960 + }, + { + "epoch": 0.9367487707824211, + "grad_norm": 0.3359779417514801, + "learning_rate": 1.9683667228712397e-06, + "loss": 0.0348, + "step": 39961 + }, + { + "epoch": 0.9367722123572262, + "grad_norm": 0.45347508788108826, + "learning_rate": 1.9669126643624525e-06, + "loss": 0.0382, + "step": 39962 + }, + { + "epoch": 0.9367956539320311, + "grad_norm": 0.2758577764034271, + "learning_rate": 1.965459137783954e-06, + "loss": 0.026, + "step": 39963 + }, + { + "epoch": 0.9368190955068362, + "grad_norm": 0.3939272165298462, + "learning_rate": 1.9640061431436264e-06, + "loss": 0.0438, + "step": 39964 + }, + { + "epoch": 0.9368425370816411, + "grad_norm": 0.3285256028175354, + "learning_rate": 1.962553680449364e-06, + "loss": 0.0636, + "step": 39965 + }, + { + "epoch": 0.9368659786564462, + "grad_norm": 0.16651061177253723, + "learning_rate": 1.9611017497090378e-06, + "loss": 0.0294, + "step": 39966 + }, + { + "epoch": 0.9368894202312511, + "grad_norm": 0.5193815231323242, + "learning_rate": 1.9596503509305196e-06, + "loss": 0.1132, + "step": 39967 + }, + { + "epoch": 0.9369128618060562, + "grad_norm": 0.3237898647785187, + "learning_rate": 1.9581994841217034e-06, + "loss": 0.0329, + "step": 39968 + }, + { + "epoch": 0.9369363033808611, + "grad_norm": 0.19377966225147247, + "learning_rate": 1.956749149290449e-06, + "loss": 0.0287, + "step": 39969 + }, + { + "epoch": 0.9369597449556661, + "grad_norm": 0.5712501406669617, + "learning_rate": 1.955299346444639e-06, + "loss": 0.0753, + "step": 39970 + }, + { + "epoch": 0.9369831865304711, + "grad_norm": 0.35101109743118286, + "learning_rate": 1.953850075592134e-06, + "loss": 0.1326, + "step": 39971 + }, + { + "epoch": 0.9370066281052761, + "grad_norm": 0.4341702461242676, + "learning_rate": 1.9524013367407946e-06, + "loss": 0.0948, + "step": 39972 + }, + { + "epoch": 0.9370300696800811, + "grad_norm": 0.07468380779027939, + "learning_rate": 1.9509531298984806e-06, + "loss": 0.0107, + "step": 39973 + }, + { + "epoch": 0.9370535112548861, + "grad_norm": 0.19253262877464294, + "learning_rate": 1.949505455073042e-06, + "loss": 0.0232, + "step": 39974 + }, + { + "epoch": 0.9370769528296911, + "grad_norm": 0.3052769601345062, + "learning_rate": 1.9480583122723495e-06, + "loss": 0.0524, + "step": 39975 + }, + { + "epoch": 0.9371003944044961, + "grad_norm": 0.22341738641262054, + "learning_rate": 1.946611701504253e-06, + "loss": 0.0235, + "step": 39976 + }, + { + "epoch": 0.937123835979301, + "grad_norm": 0.5225411653518677, + "learning_rate": 1.9451656227766014e-06, + "loss": 0.0992, + "step": 39977 + }, + { + "epoch": 0.9371472775541061, + "grad_norm": 0.5572401881217957, + "learning_rate": 1.943720076097233e-06, + "loss": 0.0643, + "step": 39978 + }, + { + "epoch": 0.937170719128911, + "grad_norm": 0.33152642846107483, + "learning_rate": 1.942275061473986e-06, + "loss": 0.0639, + "step": 39979 + }, + { + "epoch": 0.9371941607037161, + "grad_norm": 0.06627831608057022, + "learning_rate": 1.940830578914732e-06, + "loss": 0.0062, + "step": 39980 + }, + { + "epoch": 0.937217602278521, + "grad_norm": 0.3575592041015625, + "learning_rate": 1.9393866284272644e-06, + "loss": 0.0636, + "step": 39981 + }, + { + "epoch": 0.9372410438533261, + "grad_norm": 0.35027965903282166, + "learning_rate": 1.937943210019466e-06, + "loss": 0.0587, + "step": 39982 + }, + { + "epoch": 0.937264485428131, + "grad_norm": 0.19002114236354828, + "learning_rate": 1.936500323699131e-06, + "loss": 0.0148, + "step": 39983 + }, + { + "epoch": 0.9372879270029361, + "grad_norm": 0.13489967584609985, + "learning_rate": 1.935057969474108e-06, + "loss": 0.0186, + "step": 39984 + }, + { + "epoch": 0.937311368577741, + "grad_norm": 0.647426426410675, + "learning_rate": 1.933616147352224e-06, + "loss": 0.1164, + "step": 39985 + }, + { + "epoch": 0.937334810152546, + "grad_norm": 0.19057194888591766, + "learning_rate": 1.9321748573412845e-06, + "loss": 0.0212, + "step": 39986 + }, + { + "epoch": 0.937358251727351, + "grad_norm": 0.33900538086891174, + "learning_rate": 1.9307340994491275e-06, + "loss": 0.4863, + "step": 39987 + }, + { + "epoch": 0.937381693302156, + "grad_norm": 0.06517338007688522, + "learning_rate": 1.9292938736835577e-06, + "loss": 0.0042, + "step": 39988 + }, + { + "epoch": 0.9374051348769611, + "grad_norm": 0.7544160485267639, + "learning_rate": 1.927854180052413e-06, + "loss": 0.1626, + "step": 39989 + }, + { + "epoch": 0.937428576451766, + "grad_norm": 0.45927757024765015, + "learning_rate": 1.9264150185634768e-06, + "loss": 0.0546, + "step": 39990 + }, + { + "epoch": 0.9374520180265711, + "grad_norm": 0.36496543884277344, + "learning_rate": 1.924976389224564e-06, + "loss": 0.073, + "step": 39991 + }, + { + "epoch": 0.937475459601376, + "grad_norm": 0.5727236270904541, + "learning_rate": 1.923538292043503e-06, + "loss": 0.0812, + "step": 39992 + }, + { + "epoch": 0.9374989011761811, + "grad_norm": 0.3974628448486328, + "learning_rate": 1.9221007270280643e-06, + "loss": 0.0589, + "step": 39993 + }, + { + "epoch": 0.937522342750986, + "grad_norm": 0.09205175936222076, + "learning_rate": 1.9206636941860757e-06, + "loss": 0.0188, + "step": 39994 + }, + { + "epoch": 0.937545784325791, + "grad_norm": 0.3579356372356415, + "learning_rate": 1.9192271935253193e-06, + "loss": 0.0425, + "step": 39995 + }, + { + "epoch": 0.937569225900596, + "grad_norm": 0.5497313141822815, + "learning_rate": 1.9177912250536e-06, + "loss": 0.0631, + "step": 39996 + }, + { + "epoch": 0.937592667475401, + "grad_norm": 0.6482866406440735, + "learning_rate": 1.916355788778701e-06, + "loss": 0.1476, + "step": 39997 + }, + { + "epoch": 0.937616109050206, + "grad_norm": 0.3536219000816345, + "learning_rate": 1.9149208847084156e-06, + "loss": 0.0607, + "step": 39998 + }, + { + "epoch": 0.937639550625011, + "grad_norm": 0.303012877702713, + "learning_rate": 1.9134865128505153e-06, + "loss": 0.0531, + "step": 39999 + }, + { + "epoch": 0.937662992199816, + "grad_norm": 0.5321277976036072, + "learning_rate": 1.912052673212805e-06, + "loss": 0.081, + "step": 40000 + }, + { + "epoch": 0.937686433774621, + "grad_norm": 0.07117856293916702, + "learning_rate": 1.910619365803057e-06, + "loss": 0.0146, + "step": 40001 + }, + { + "epoch": 0.937709875349426, + "grad_norm": 0.4396685063838959, + "learning_rate": 1.909186590629042e-06, + "loss": 0.0734, + "step": 40002 + }, + { + "epoch": 0.937733316924231, + "grad_norm": 0.6877076625823975, + "learning_rate": 1.9077543476985314e-06, + "loss": 0.1376, + "step": 40003 + }, + { + "epoch": 0.937756758499036, + "grad_norm": 0.3519687056541443, + "learning_rate": 1.9063226370193199e-06, + "loss": 0.0585, + "step": 40004 + }, + { + "epoch": 0.937780200073841, + "grad_norm": 0.17111079394817352, + "learning_rate": 1.904891458599145e-06, + "loss": 0.0272, + "step": 40005 + }, + { + "epoch": 0.9378036416486459, + "grad_norm": 0.40297019481658936, + "learning_rate": 1.9034608124457898e-06, + "loss": 0.4747, + "step": 40006 + }, + { + "epoch": 0.937827083223451, + "grad_norm": 0.28281447291374207, + "learning_rate": 1.9020306985670256e-06, + "loss": 0.052, + "step": 40007 + }, + { + "epoch": 0.9378505247982559, + "grad_norm": 0.5736522078514099, + "learning_rate": 1.9006011169705906e-06, + "loss": 0.0846, + "step": 40008 + }, + { + "epoch": 0.937873966373061, + "grad_norm": 0.436949223279953, + "learning_rate": 1.8991720676642565e-06, + "loss": 0.0689, + "step": 40009 + }, + { + "epoch": 0.9378974079478659, + "grad_norm": 0.3631491959095001, + "learning_rate": 1.8977435506557728e-06, + "loss": 0.0632, + "step": 40010 + }, + { + "epoch": 0.937920849522671, + "grad_norm": 0.38113945722579956, + "learning_rate": 1.8963155659528998e-06, + "loss": 0.3577, + "step": 40011 + }, + { + "epoch": 0.9379442910974759, + "grad_norm": 0.28498417139053345, + "learning_rate": 1.8948881135633534e-06, + "loss": 0.0232, + "step": 40012 + }, + { + "epoch": 0.937967732672281, + "grad_norm": 0.5295042991638184, + "learning_rate": 1.8934611934949276e-06, + "loss": 0.0443, + "step": 40013 + }, + { + "epoch": 0.9379911742470859, + "grad_norm": 0.6175867319107056, + "learning_rate": 1.8920348057553272e-06, + "loss": 0.0797, + "step": 40014 + }, + { + "epoch": 0.9380146158218909, + "grad_norm": 0.5618339776992798, + "learning_rate": 1.8906089503523018e-06, + "loss": 0.0931, + "step": 40015 + }, + { + "epoch": 0.9380380573966959, + "grad_norm": 0.5763516426086426, + "learning_rate": 1.8891836272936003e-06, + "loss": 0.5313, + "step": 40016 + }, + { + "epoch": 0.9380614989715009, + "grad_norm": 0.19383679330348969, + "learning_rate": 1.8877588365869391e-06, + "loss": 0.023, + "step": 40017 + }, + { + "epoch": 0.9380849405463059, + "grad_norm": 0.22827129065990448, + "learning_rate": 1.8863345782400565e-06, + "loss": 0.0261, + "step": 40018 + }, + { + "epoch": 0.9381083821211109, + "grad_norm": 0.37992432713508606, + "learning_rate": 1.8849108522606796e-06, + "loss": 0.0636, + "step": 40019 + }, + { + "epoch": 0.9381318236959159, + "grad_norm": 0.39151430130004883, + "learning_rate": 1.8834876586565464e-06, + "loss": 0.0577, + "step": 40020 + }, + { + "epoch": 0.9381552652707209, + "grad_norm": 0.3377501666545868, + "learning_rate": 1.8820649974353622e-06, + "loss": 0.0417, + "step": 40021 + }, + { + "epoch": 0.9381787068455258, + "grad_norm": 0.5136846303939819, + "learning_rate": 1.8806428686048539e-06, + "loss": 0.0432, + "step": 40022 + }, + { + "epoch": 0.9382021484203309, + "grad_norm": 0.11363967508077621, + "learning_rate": 1.8792212721727376e-06, + "loss": 0.0227, + "step": 40023 + }, + { + "epoch": 0.9382255899951358, + "grad_norm": 0.3636431097984314, + "learning_rate": 1.8778002081467183e-06, + "loss": 0.0333, + "step": 40024 + }, + { + "epoch": 0.9382490315699409, + "grad_norm": 0.4016365110874176, + "learning_rate": 1.8763796765345232e-06, + "loss": 0.0899, + "step": 40025 + }, + { + "epoch": 0.9382724731447458, + "grad_norm": 0.3687451183795929, + "learning_rate": 1.874959677343846e-06, + "loss": 0.0525, + "step": 40026 + }, + { + "epoch": 0.9382959147195509, + "grad_norm": 0.20808233320713043, + "learning_rate": 1.8735402105823918e-06, + "loss": 0.0249, + "step": 40027 + }, + { + "epoch": 0.9383193562943558, + "grad_norm": 0.5726333856582642, + "learning_rate": 1.8721212762578656e-06, + "loss": 0.1221, + "step": 40028 + }, + { + "epoch": 0.9383427978691609, + "grad_norm": 0.16268151998519897, + "learning_rate": 1.8707028743779831e-06, + "loss": 0.0146, + "step": 40029 + }, + { + "epoch": 0.9383662394439658, + "grad_norm": 0.2009616494178772, + "learning_rate": 1.8692850049504163e-06, + "loss": 0.035, + "step": 40030 + }, + { + "epoch": 0.9383896810187708, + "grad_norm": 0.28115662932395935, + "learning_rate": 1.86786766798287e-06, + "loss": 0.0322, + "step": 40031 + }, + { + "epoch": 0.9384131225935758, + "grad_norm": 0.5270015001296997, + "learning_rate": 1.8664508634830269e-06, + "loss": 0.1076, + "step": 40032 + }, + { + "epoch": 0.9384365641683808, + "grad_norm": 0.13480639457702637, + "learning_rate": 1.865034591458592e-06, + "loss": 0.0213, + "step": 40033 + }, + { + "epoch": 0.9384600057431858, + "grad_norm": 0.33168262243270874, + "learning_rate": 1.863618851917237e-06, + "loss": 0.0487, + "step": 40034 + }, + { + "epoch": 0.9384834473179908, + "grad_norm": 0.09840890765190125, + "learning_rate": 1.8622036448666447e-06, + "loss": 0.0147, + "step": 40035 + }, + { + "epoch": 0.9385068888927958, + "grad_norm": 0.39895060658454895, + "learning_rate": 1.8607889703144864e-06, + "loss": 0.3891, + "step": 40036 + }, + { + "epoch": 0.9385303304676008, + "grad_norm": 0.4958159029483795, + "learning_rate": 1.8593748282684454e-06, + "loss": 0.097, + "step": 40037 + }, + { + "epoch": 0.9385537720424058, + "grad_norm": 0.28339749574661255, + "learning_rate": 1.857961218736215e-06, + "loss": 0.0396, + "step": 40038 + }, + { + "epoch": 0.9385772136172108, + "grad_norm": 0.11111294478178024, + "learning_rate": 1.8565481417254337e-06, + "loss": 0.0081, + "step": 40039 + }, + { + "epoch": 0.9386006551920159, + "grad_norm": 0.13164624571800232, + "learning_rate": 1.8551355972437734e-06, + "loss": 0.0213, + "step": 40040 + }, + { + "epoch": 0.9386240967668208, + "grad_norm": 0.2417188435792923, + "learning_rate": 1.8537235852989165e-06, + "loss": 0.0416, + "step": 40041 + }, + { + "epoch": 0.9386475383416258, + "grad_norm": 0.11279703676700592, + "learning_rate": 1.8523121058985127e-06, + "loss": 0.0209, + "step": 40042 + }, + { + "epoch": 0.9386709799164308, + "grad_norm": 0.5374447107315063, + "learning_rate": 1.8509011590502335e-06, + "loss": 0.0841, + "step": 40043 + }, + { + "epoch": 0.9386944214912358, + "grad_norm": 0.38754338026046753, + "learning_rate": 1.849490744761706e-06, + "loss": 0.0659, + "step": 40044 + }, + { + "epoch": 0.9387178630660408, + "grad_norm": 0.2681936025619507, + "learning_rate": 1.8480808630406132e-06, + "loss": 0.0204, + "step": 40045 + }, + { + "epoch": 0.9387413046408458, + "grad_norm": 0.44736209511756897, + "learning_rate": 1.8466715138945934e-06, + "loss": 0.0631, + "step": 40046 + }, + { + "epoch": 0.9387647462156508, + "grad_norm": 0.6877533197402954, + "learning_rate": 1.8452626973312958e-06, + "loss": 0.1214, + "step": 40047 + }, + { + "epoch": 0.9387881877904558, + "grad_norm": 0.6246282458305359, + "learning_rate": 1.843854413358359e-06, + "loss": 0.1162, + "step": 40048 + }, + { + "epoch": 0.9388116293652607, + "grad_norm": 0.5256655216217041, + "learning_rate": 1.8424466619834325e-06, + "loss": 0.1306, + "step": 40049 + }, + { + "epoch": 0.9388350709400658, + "grad_norm": 0.401594340801239, + "learning_rate": 1.8410394432141321e-06, + "loss": 0.0797, + "step": 40050 + }, + { + "epoch": 0.9388585125148707, + "grad_norm": 0.2799285054206848, + "learning_rate": 1.8396327570581296e-06, + "loss": 0.0399, + "step": 40051 + }, + { + "epoch": 0.9388819540896758, + "grad_norm": 0.5560145378112793, + "learning_rate": 1.83822660352303e-06, + "loss": 0.1084, + "step": 40052 + }, + { + "epoch": 0.9389053956644807, + "grad_norm": 0.42428261041641235, + "learning_rate": 1.8368209826164829e-06, + "loss": 0.0857, + "step": 40053 + }, + { + "epoch": 0.9389288372392858, + "grad_norm": 0.12192137539386749, + "learning_rate": 1.8354158943461041e-06, + "loss": 0.0163, + "step": 40054 + }, + { + "epoch": 0.9389522788140907, + "grad_norm": 0.12956002354621887, + "learning_rate": 1.8340113387195102e-06, + "loss": 0.0161, + "step": 40055 + }, + { + "epoch": 0.9389757203888958, + "grad_norm": 0.28128767013549805, + "learning_rate": 1.832607315744328e-06, + "loss": 0.03, + "step": 40056 + }, + { + "epoch": 0.9389991619637007, + "grad_norm": 0.16520853340625763, + "learning_rate": 1.831203825428196e-06, + "loss": 0.0251, + "step": 40057 + }, + { + "epoch": 0.9390226035385058, + "grad_norm": 0.7034094929695129, + "learning_rate": 1.829800867778697e-06, + "loss": 0.1213, + "step": 40058 + }, + { + "epoch": 0.9390460451133107, + "grad_norm": 0.3884861171245575, + "learning_rate": 1.8283984428034694e-06, + "loss": 0.3833, + "step": 40059 + }, + { + "epoch": 0.9390694866881157, + "grad_norm": 0.5055989027023315, + "learning_rate": 1.8269965505101072e-06, + "loss": 0.062, + "step": 40060 + }, + { + "epoch": 0.9390929282629207, + "grad_norm": 0.19979704916477203, + "learning_rate": 1.8255951909062264e-06, + "loss": 0.0509, + "step": 40061 + }, + { + "epoch": 0.9391163698377257, + "grad_norm": 0.39644113183021545, + "learning_rate": 1.8241943639994208e-06, + "loss": 0.0402, + "step": 40062 + }, + { + "epoch": 0.9391398114125307, + "grad_norm": 0.4350472390651703, + "learning_rate": 1.8227940697972955e-06, + "loss": 0.0628, + "step": 40063 + }, + { + "epoch": 0.9391632529873357, + "grad_norm": 0.16254983842372894, + "learning_rate": 1.8213943083074447e-06, + "loss": 0.0072, + "step": 40064 + }, + { + "epoch": 0.9391866945621407, + "grad_norm": 0.46674883365631104, + "learning_rate": 1.8199950795374843e-06, + "loss": 0.3846, + "step": 40065 + }, + { + "epoch": 0.9392101361369457, + "grad_norm": 0.5155289769172668, + "learning_rate": 1.8185963834949859e-06, + "loss": 0.0911, + "step": 40066 + }, + { + "epoch": 0.9392335777117506, + "grad_norm": 0.3660295009613037, + "learning_rate": 1.8171982201875437e-06, + "loss": 0.0445, + "step": 40067 + }, + { + "epoch": 0.9392570192865557, + "grad_norm": 0.5437494516372681, + "learning_rate": 1.8158005896227515e-06, + "loss": 0.0667, + "step": 40068 + }, + { + "epoch": 0.9392804608613606, + "grad_norm": 0.7759196758270264, + "learning_rate": 1.8144034918081698e-06, + "loss": 0.5282, + "step": 40069 + }, + { + "epoch": 0.9393039024361657, + "grad_norm": 0.29662302136421204, + "learning_rate": 1.8130069267514037e-06, + "loss": 0.0516, + "step": 40070 + }, + { + "epoch": 0.9393273440109706, + "grad_norm": 0.4788452088832855, + "learning_rate": 1.8116108944600251e-06, + "loss": 0.0771, + "step": 40071 + }, + { + "epoch": 0.9393507855857757, + "grad_norm": 0.4516189992427826, + "learning_rate": 1.8102153949416056e-06, + "loss": 0.0892, + "step": 40072 + }, + { + "epoch": 0.9393742271605806, + "grad_norm": 0.22065772116184235, + "learning_rate": 1.8088204282037169e-06, + "loss": 0.028, + "step": 40073 + }, + { + "epoch": 0.9393976687353857, + "grad_norm": 0.4550366997718811, + "learning_rate": 1.8074259942539306e-06, + "loss": 0.0391, + "step": 40074 + }, + { + "epoch": 0.9394211103101906, + "grad_norm": 0.12597230076789856, + "learning_rate": 1.8060320930998076e-06, + "loss": 0.029, + "step": 40075 + }, + { + "epoch": 0.9394445518849956, + "grad_norm": 0.34111881256103516, + "learning_rate": 1.8046387247489083e-06, + "loss": 0.0202, + "step": 40076 + }, + { + "epoch": 0.9394679934598006, + "grad_norm": 0.08971527218818665, + "learning_rate": 1.8032458892088155e-06, + "loss": 0.0151, + "step": 40077 + }, + { + "epoch": 0.9394914350346056, + "grad_norm": 0.7986359000205994, + "learning_rate": 1.801853586487068e-06, + "loss": 0.0944, + "step": 40078 + }, + { + "epoch": 0.9395148766094106, + "grad_norm": 0.520699679851532, + "learning_rate": 1.8004618165912258e-06, + "loss": 0.0458, + "step": 40079 + }, + { + "epoch": 0.9395383181842156, + "grad_norm": 0.3287515938282013, + "learning_rate": 1.7990705795288386e-06, + "loss": 0.0604, + "step": 40080 + }, + { + "epoch": 0.9395617597590206, + "grad_norm": 0.3266325294971466, + "learning_rate": 1.7976798753074563e-06, + "loss": 0.0524, + "step": 40081 + }, + { + "epoch": 0.9395852013338256, + "grad_norm": 0.15864241123199463, + "learning_rate": 1.796289703934617e-06, + "loss": 0.0393, + "step": 40082 + }, + { + "epoch": 0.9396086429086306, + "grad_norm": 0.2932072579860687, + "learning_rate": 1.7949000654178705e-06, + "loss": 0.045, + "step": 40083 + }, + { + "epoch": 0.9396320844834356, + "grad_norm": 0.4206072986125946, + "learning_rate": 1.7935109597647658e-06, + "loss": 0.101, + "step": 40084 + }, + { + "epoch": 0.9396555260582405, + "grad_norm": 0.4464457631111145, + "learning_rate": 1.7921223869828308e-06, + "loss": 0.0595, + "step": 40085 + }, + { + "epoch": 0.9396789676330456, + "grad_norm": 0.1686873883008957, + "learning_rate": 1.7907343470796034e-06, + "loss": 0.033, + "step": 40086 + }, + { + "epoch": 0.9397024092078505, + "grad_norm": 0.2420562207698822, + "learning_rate": 1.7893468400626e-06, + "loss": 0.0226, + "step": 40087 + }, + { + "epoch": 0.9397258507826556, + "grad_norm": 0.1793738156557083, + "learning_rate": 1.7879598659393704e-06, + "loss": 0.0181, + "step": 40088 + }, + { + "epoch": 0.9397492923574605, + "grad_norm": 0.12002759426832199, + "learning_rate": 1.7865734247174192e-06, + "loss": 0.019, + "step": 40089 + }, + { + "epoch": 0.9397727339322656, + "grad_norm": 0.5997506976127625, + "learning_rate": 1.7851875164042963e-06, + "loss": 0.0537, + "step": 40090 + }, + { + "epoch": 0.9397961755070706, + "grad_norm": 0.394919753074646, + "learning_rate": 1.7838021410075068e-06, + "loss": 0.0543, + "step": 40091 + }, + { + "epoch": 0.9398196170818756, + "grad_norm": 0.0962577760219574, + "learning_rate": 1.7824172985345777e-06, + "loss": 0.0151, + "step": 40092 + }, + { + "epoch": 0.9398430586566806, + "grad_norm": 0.1797555685043335, + "learning_rate": 1.7810329889930034e-06, + "loss": 0.0155, + "step": 40093 + }, + { + "epoch": 0.9398665002314855, + "grad_norm": 0.6691985130310059, + "learning_rate": 1.7796492123902997e-06, + "loss": 0.0437, + "step": 40094 + }, + { + "epoch": 0.9398899418062906, + "grad_norm": 0.631018340587616, + "learning_rate": 1.778265968733983e-06, + "loss": 0.1046, + "step": 40095 + }, + { + "epoch": 0.9399133833810955, + "grad_norm": 0.3760426938533783, + "learning_rate": 1.7768832580315697e-06, + "loss": 0.0962, + "step": 40096 + }, + { + "epoch": 0.9399368249559006, + "grad_norm": 0.2939806580543518, + "learning_rate": 1.7755010802905426e-06, + "loss": 0.0557, + "step": 40097 + }, + { + "epoch": 0.9399602665307055, + "grad_norm": 0.6008968353271484, + "learning_rate": 1.7741194355184066e-06, + "loss": 0.0432, + "step": 40098 + }, + { + "epoch": 0.9399837081055106, + "grad_norm": 0.36157849431037903, + "learning_rate": 1.7727383237226669e-06, + "loss": 0.0333, + "step": 40099 + }, + { + "epoch": 0.9400071496803155, + "grad_norm": 0.517616331577301, + "learning_rate": 1.7713577449107954e-06, + "loss": 0.0866, + "step": 40100 + }, + { + "epoch": 0.9400305912551206, + "grad_norm": 0.2625315189361572, + "learning_rate": 1.7699776990902972e-06, + "loss": 0.0351, + "step": 40101 + }, + { + "epoch": 0.9400540328299255, + "grad_norm": 0.7086888551712036, + "learning_rate": 1.7685981862686775e-06, + "loss": 0.1258, + "step": 40102 + }, + { + "epoch": 0.9400774744047306, + "grad_norm": 0.8454569578170776, + "learning_rate": 1.7672192064534076e-06, + "loss": 0.1223, + "step": 40103 + }, + { + "epoch": 0.9401009159795355, + "grad_norm": 0.33949992060661316, + "learning_rate": 1.76584075965196e-06, + "loss": 0.5169, + "step": 40104 + }, + { + "epoch": 0.9401243575543405, + "grad_norm": 0.28846457600593567, + "learning_rate": 1.7644628458718282e-06, + "loss": 0.0383, + "step": 40105 + }, + { + "epoch": 0.9401477991291455, + "grad_norm": 0.20313379168510437, + "learning_rate": 1.7630854651204843e-06, + "loss": 0.0261, + "step": 40106 + }, + { + "epoch": 0.9401712407039505, + "grad_norm": 0.1445382684469223, + "learning_rate": 1.7617086174053887e-06, + "loss": 0.0281, + "step": 40107 + }, + { + "epoch": 0.9401946822787555, + "grad_norm": 0.20163841545581818, + "learning_rate": 1.7603323027340358e-06, + "loss": 0.0221, + "step": 40108 + }, + { + "epoch": 0.9402181238535605, + "grad_norm": 0.09123773127794266, + "learning_rate": 1.7589565211138748e-06, + "loss": 0.0146, + "step": 40109 + }, + { + "epoch": 0.9402415654283655, + "grad_norm": 0.21101628243923187, + "learning_rate": 1.7575812725523889e-06, + "loss": 0.0185, + "step": 40110 + }, + { + "epoch": 0.9402650070031705, + "grad_norm": 0.35889461636543274, + "learning_rate": 1.7562065570570163e-06, + "loss": 0.0415, + "step": 40111 + }, + { + "epoch": 0.9402884485779754, + "grad_norm": 0.10968522727489471, + "learning_rate": 1.754832374635229e-06, + "loss": 0.0115, + "step": 40112 + }, + { + "epoch": 0.9403118901527805, + "grad_norm": 0.46005505323410034, + "learning_rate": 1.7534587252944989e-06, + "loss": 0.6021, + "step": 40113 + }, + { + "epoch": 0.9403353317275854, + "grad_norm": 0.17766794562339783, + "learning_rate": 1.752085609042242e-06, + "loss": 0.0286, + "step": 40114 + }, + { + "epoch": 0.9403587733023905, + "grad_norm": 0.11372445523738861, + "learning_rate": 1.7507130258859416e-06, + "loss": 0.0138, + "step": 40115 + }, + { + "epoch": 0.9403822148771954, + "grad_norm": 0.49756568670272827, + "learning_rate": 1.749340975833047e-06, + "loss": 0.0966, + "step": 40116 + }, + { + "epoch": 0.9404056564520005, + "grad_norm": 0.4815429449081421, + "learning_rate": 1.7479694588909857e-06, + "loss": 0.6508, + "step": 40117 + }, + { + "epoch": 0.9404290980268054, + "grad_norm": 0.301620215177536, + "learning_rate": 1.746598475067196e-06, + "loss": 0.0804, + "step": 40118 + }, + { + "epoch": 0.9404525396016105, + "grad_norm": 0.6175934672355652, + "learning_rate": 1.7452280243691277e-06, + "loss": 0.0926, + "step": 40119 + }, + { + "epoch": 0.9404759811764154, + "grad_norm": 0.17127960920333862, + "learning_rate": 1.7438581068042193e-06, + "loss": 0.0382, + "step": 40120 + }, + { + "epoch": 0.9404994227512204, + "grad_norm": 0.38132745027542114, + "learning_rate": 1.742488722379898e-06, + "loss": 0.0701, + "step": 40121 + }, + { + "epoch": 0.9405228643260254, + "grad_norm": 0.17689551413059235, + "learning_rate": 1.7411198711035914e-06, + "loss": 0.0312, + "step": 40122 + }, + { + "epoch": 0.9405463059008304, + "grad_norm": 0.24080407619476318, + "learning_rate": 1.7397515529827379e-06, + "loss": 0.044, + "step": 40123 + }, + { + "epoch": 0.9405697474756354, + "grad_norm": 0.12108306586742401, + "learning_rate": 1.7383837680247538e-06, + "loss": 0.0211, + "step": 40124 + }, + { + "epoch": 0.9405931890504404, + "grad_norm": 0.3704826831817627, + "learning_rate": 1.7370165162370556e-06, + "loss": 0.0331, + "step": 40125 + }, + { + "epoch": 0.9406166306252454, + "grad_norm": 0.5131414532661438, + "learning_rate": 1.7356497976270702e-06, + "loss": 0.0569, + "step": 40126 + }, + { + "epoch": 0.9406400722000504, + "grad_norm": 0.6191065311431885, + "learning_rate": 1.7342836122022144e-06, + "loss": 0.1113, + "step": 40127 + }, + { + "epoch": 0.9406635137748554, + "grad_norm": 0.3855001628398895, + "learning_rate": 1.732917959969893e-06, + "loss": 0.0887, + "step": 40128 + }, + { + "epoch": 0.9406869553496604, + "grad_norm": 0.11221090704202652, + "learning_rate": 1.7315528409375336e-06, + "loss": 0.0214, + "step": 40129 + }, + { + "epoch": 0.9407103969244653, + "grad_norm": 0.21973051130771637, + "learning_rate": 1.7301882551125303e-06, + "loss": 0.0203, + "step": 40130 + }, + { + "epoch": 0.9407338384992704, + "grad_norm": 0.37174850702285767, + "learning_rate": 1.7288242025022882e-06, + "loss": 0.0792, + "step": 40131 + }, + { + "epoch": 0.9407572800740753, + "grad_norm": 0.7424126863479614, + "learning_rate": 1.7274606831142014e-06, + "loss": 0.591, + "step": 40132 + }, + { + "epoch": 0.9407807216488804, + "grad_norm": 0.2792416214942932, + "learning_rate": 1.7260976969556753e-06, + "loss": 0.0628, + "step": 40133 + }, + { + "epoch": 0.9408041632236853, + "grad_norm": 0.6374897360801697, + "learning_rate": 1.7247352440341146e-06, + "loss": 0.1511, + "step": 40134 + }, + { + "epoch": 0.9408276047984904, + "grad_norm": 0.15307092666625977, + "learning_rate": 1.7233733243569027e-06, + "loss": 0.0365, + "step": 40135 + }, + { + "epoch": 0.9408510463732953, + "grad_norm": 0.37243351340293884, + "learning_rate": 1.7220119379314336e-06, + "loss": 0.0533, + "step": 40136 + }, + { + "epoch": 0.9408744879481004, + "grad_norm": 0.5120794177055359, + "learning_rate": 1.7206510847650792e-06, + "loss": 0.064, + "step": 40137 + }, + { + "epoch": 0.9408979295229053, + "grad_norm": 0.0866178646683693, + "learning_rate": 1.7192907648652335e-06, + "loss": 0.0089, + "step": 40138 + }, + { + "epoch": 0.9409213710977103, + "grad_norm": 0.2769559919834137, + "learning_rate": 1.7179309782392904e-06, + "loss": 0.0213, + "step": 40139 + }, + { + "epoch": 0.9409448126725153, + "grad_norm": 0.12491583824157715, + "learning_rate": 1.7165717248946111e-06, + "loss": 0.0151, + "step": 40140 + }, + { + "epoch": 0.9409682542473203, + "grad_norm": 0.48236769437789917, + "learning_rate": 1.7152130048385783e-06, + "loss": 0.0566, + "step": 40141 + }, + { + "epoch": 0.9409916958221254, + "grad_norm": 0.36357346177101135, + "learning_rate": 1.7138548180785752e-06, + "loss": 0.051, + "step": 40142 + }, + { + "epoch": 0.9410151373969303, + "grad_norm": 0.460312157869339, + "learning_rate": 1.7124971646219513e-06, + "loss": 0.0841, + "step": 40143 + }, + { + "epoch": 0.9410385789717354, + "grad_norm": 0.5723007321357727, + "learning_rate": 1.7111400444760783e-06, + "loss": 0.5008, + "step": 40144 + }, + { + "epoch": 0.9410620205465403, + "grad_norm": 0.4897805452346802, + "learning_rate": 1.7097834576483175e-06, + "loss": 0.0738, + "step": 40145 + }, + { + "epoch": 0.9410854621213454, + "grad_norm": 0.4055715501308441, + "learning_rate": 1.7084274041460402e-06, + "loss": 0.0618, + "step": 40146 + }, + { + "epoch": 0.9411089036961503, + "grad_norm": 0.673647403717041, + "learning_rate": 1.7070718839766075e-06, + "loss": 0.0953, + "step": 40147 + }, + { + "epoch": 0.9411323452709554, + "grad_norm": 0.45782604813575745, + "learning_rate": 1.7057168971473691e-06, + "loss": 0.0385, + "step": 40148 + }, + { + "epoch": 0.9411557868457603, + "grad_norm": 0.29627782106399536, + "learning_rate": 1.7043624436656525e-06, + "loss": 0.0378, + "step": 40149 + }, + { + "epoch": 0.9411792284205653, + "grad_norm": 0.3911879360675812, + "learning_rate": 1.7030085235388516e-06, + "loss": 0.0491, + "step": 40150 + }, + { + "epoch": 0.9412026699953703, + "grad_norm": 0.20860707759857178, + "learning_rate": 1.7016551367742828e-06, + "loss": 0.0255, + "step": 40151 + }, + { + "epoch": 0.9412261115701753, + "grad_norm": 0.393868625164032, + "learning_rate": 1.7003022833792848e-06, + "loss": 0.0813, + "step": 40152 + }, + { + "epoch": 0.9412495531449803, + "grad_norm": 0.4305109977722168, + "learning_rate": 1.6989499633612293e-06, + "loss": 0.6054, + "step": 40153 + }, + { + "epoch": 0.9412729947197853, + "grad_norm": 0.6970010995864868, + "learning_rate": 1.6975981767274217e-06, + "loss": 0.5382, + "step": 40154 + }, + { + "epoch": 0.9412964362945903, + "grad_norm": 1.340933918952942, + "learning_rate": 1.696246923485223e-06, + "loss": 0.3069, + "step": 40155 + }, + { + "epoch": 0.9413198778693953, + "grad_norm": 0.3745129406452179, + "learning_rate": 1.694896203641949e-06, + "loss": 0.4338, + "step": 40156 + }, + { + "epoch": 0.9413433194442002, + "grad_norm": 0.33715537190437317, + "learning_rate": 1.6935460172049277e-06, + "loss": 0.0613, + "step": 40157 + }, + { + "epoch": 0.9413667610190053, + "grad_norm": 0.38653337955474854, + "learning_rate": 1.6921963641814864e-06, + "loss": 0.4439, + "step": 40158 + }, + { + "epoch": 0.9413902025938102, + "grad_norm": 0.6445977091789246, + "learning_rate": 1.6908472445789526e-06, + "loss": 0.0727, + "step": 40159 + }, + { + "epoch": 0.9414136441686153, + "grad_norm": 0.48747357726097107, + "learning_rate": 1.6894986584046536e-06, + "loss": 0.078, + "step": 40160 + }, + { + "epoch": 0.9414370857434202, + "grad_norm": 0.22246254980564117, + "learning_rate": 1.6881506056658946e-06, + "loss": 0.0244, + "step": 40161 + }, + { + "epoch": 0.9414605273182253, + "grad_norm": 0.08606420457363129, + "learning_rate": 1.6868030863699923e-06, + "loss": 0.008, + "step": 40162 + }, + { + "epoch": 0.9414839688930302, + "grad_norm": 0.415446013212204, + "learning_rate": 1.685456100524263e-06, + "loss": 0.0528, + "step": 40163 + }, + { + "epoch": 0.9415074104678353, + "grad_norm": 0.3247677981853485, + "learning_rate": 1.684109648136023e-06, + "loss": 0.0659, + "step": 40164 + }, + { + "epoch": 0.9415308520426402, + "grad_norm": 0.5900903940200806, + "learning_rate": 1.6827637292125553e-06, + "loss": 0.0679, + "step": 40165 + }, + { + "epoch": 0.9415542936174452, + "grad_norm": 0.49803370237350464, + "learning_rate": 1.6814183437611763e-06, + "loss": 0.0439, + "step": 40166 + }, + { + "epoch": 0.9415777351922502, + "grad_norm": 0.44655483961105347, + "learning_rate": 1.6800734917891913e-06, + "loss": 0.1099, + "step": 40167 + }, + { + "epoch": 0.9416011767670552, + "grad_norm": 0.0883779525756836, + "learning_rate": 1.6787291733038946e-06, + "loss": 0.0162, + "step": 40168 + }, + { + "epoch": 0.9416246183418602, + "grad_norm": 0.5336536765098572, + "learning_rate": 1.67738538831258e-06, + "loss": 0.5497, + "step": 40169 + }, + { + "epoch": 0.9416480599166652, + "grad_norm": 0.3421061635017395, + "learning_rate": 1.67604213682252e-06, + "loss": 0.0365, + "step": 40170 + }, + { + "epoch": 0.9416715014914702, + "grad_norm": 0.2350286990404129, + "learning_rate": 1.6746994188410303e-06, + "loss": 0.0468, + "step": 40171 + }, + { + "epoch": 0.9416949430662752, + "grad_norm": 0.6602011919021606, + "learning_rate": 1.6733572343753944e-06, + "loss": 0.042, + "step": 40172 + }, + { + "epoch": 0.9417183846410802, + "grad_norm": 0.47521859407424927, + "learning_rate": 1.6720155834328731e-06, + "loss": 0.0838, + "step": 40173 + }, + { + "epoch": 0.9417418262158852, + "grad_norm": 0.49473491311073303, + "learning_rate": 1.6706744660207606e-06, + "loss": 0.0521, + "step": 40174 + }, + { + "epoch": 0.9417652677906901, + "grad_norm": 0.3390595018863678, + "learning_rate": 1.6693338821463511e-06, + "loss": 0.0337, + "step": 40175 + }, + { + "epoch": 0.9417887093654952, + "grad_norm": 0.4158101975917816, + "learning_rate": 1.6679938318168831e-06, + "loss": 0.0316, + "step": 40176 + }, + { + "epoch": 0.9418121509403001, + "grad_norm": 0.6360387802124023, + "learning_rate": 1.6666543150396507e-06, + "loss": 0.161, + "step": 40177 + }, + { + "epoch": 0.9418355925151052, + "grad_norm": 0.5154298543930054, + "learning_rate": 1.665315331821915e-06, + "loss": 0.5985, + "step": 40178 + }, + { + "epoch": 0.9418590340899101, + "grad_norm": 0.22282253205776215, + "learning_rate": 1.6639768821709478e-06, + "loss": 0.0322, + "step": 40179 + }, + { + "epoch": 0.9418824756647152, + "grad_norm": 0.24992454051971436, + "learning_rate": 1.6626389660940101e-06, + "loss": 0.0524, + "step": 40180 + }, + { + "epoch": 0.9419059172395201, + "grad_norm": 0.14809787273406982, + "learning_rate": 1.6613015835983515e-06, + "loss": 0.0245, + "step": 40181 + }, + { + "epoch": 0.9419293588143252, + "grad_norm": 0.18986278772354126, + "learning_rate": 1.6599647346912328e-06, + "loss": 0.0325, + "step": 40182 + }, + { + "epoch": 0.9419528003891301, + "grad_norm": 0.4588620364665985, + "learning_rate": 1.6586284193799152e-06, + "loss": 0.0639, + "step": 40183 + }, + { + "epoch": 0.9419762419639351, + "grad_norm": 0.34573325514793396, + "learning_rate": 1.657292637671648e-06, + "loss": 0.067, + "step": 40184 + }, + { + "epoch": 0.9419996835387401, + "grad_norm": 0.2642410695552826, + "learning_rate": 1.655957389573659e-06, + "loss": 0.0491, + "step": 40185 + }, + { + "epoch": 0.9420231251135451, + "grad_norm": 0.926558256149292, + "learning_rate": 1.6546226750932314e-06, + "loss": 0.0377, + "step": 40186 + }, + { + "epoch": 0.9420465666883501, + "grad_norm": 0.36257192492485046, + "learning_rate": 1.6532884942375814e-06, + "loss": 0.0855, + "step": 40187 + }, + { + "epoch": 0.9420700082631551, + "grad_norm": 0.27971896529197693, + "learning_rate": 1.651954847013948e-06, + "loss": 0.0481, + "step": 40188 + }, + { + "epoch": 0.9420934498379601, + "grad_norm": 0.4567437767982483, + "learning_rate": 1.6506217334295805e-06, + "loss": 0.0828, + "step": 40189 + }, + { + "epoch": 0.9421168914127651, + "grad_norm": 0.30572059750556946, + "learning_rate": 1.6492891534916954e-06, + "loss": 0.0335, + "step": 40190 + }, + { + "epoch": 0.94214033298757, + "grad_norm": 0.48628923296928406, + "learning_rate": 1.6479571072075316e-06, + "loss": 0.0661, + "step": 40191 + }, + { + "epoch": 0.9421637745623751, + "grad_norm": 0.5367494821548462, + "learning_rate": 1.646625594584328e-06, + "loss": 0.0901, + "step": 40192 + }, + { + "epoch": 0.9421872161371802, + "grad_norm": 0.43123406171798706, + "learning_rate": 1.6452946156293004e-06, + "loss": 0.0585, + "step": 40193 + }, + { + "epoch": 0.9422106577119851, + "grad_norm": 0.47113001346588135, + "learning_rate": 1.6439641703496655e-06, + "loss": 0.0832, + "step": 40194 + }, + { + "epoch": 0.9422340992867901, + "grad_norm": 0.3806498348712921, + "learning_rate": 1.64263425875264e-06, + "loss": 0.0335, + "step": 40195 + }, + { + "epoch": 0.9422575408615951, + "grad_norm": 0.1403292566537857, + "learning_rate": 1.641304880845451e-06, + "loss": 0.0265, + "step": 40196 + }, + { + "epoch": 0.9422809824364001, + "grad_norm": 0.5018053650856018, + "learning_rate": 1.6399760366353044e-06, + "loss": 0.4525, + "step": 40197 + }, + { + "epoch": 0.9423044240112051, + "grad_norm": 0.4341435134410858, + "learning_rate": 1.6386477261294165e-06, + "loss": 0.0753, + "step": 40198 + }, + { + "epoch": 0.9423278655860101, + "grad_norm": 0.2860329747200012, + "learning_rate": 1.6373199493349922e-06, + "loss": 0.0408, + "step": 40199 + }, + { + "epoch": 0.942351307160815, + "grad_norm": 0.5123937129974365, + "learning_rate": 1.6359927062592373e-06, + "loss": 0.0647, + "step": 40200 + }, + { + "epoch": 0.9423747487356201, + "grad_norm": 0.40443623065948486, + "learning_rate": 1.6346659969093569e-06, + "loss": 0.5651, + "step": 40201 + }, + { + "epoch": 0.942398190310425, + "grad_norm": 0.1486452966928482, + "learning_rate": 1.6333398212925343e-06, + "loss": 0.0259, + "step": 40202 + }, + { + "epoch": 0.9424216318852301, + "grad_norm": 0.2064020037651062, + "learning_rate": 1.632014179415975e-06, + "loss": 0.0238, + "step": 40203 + }, + { + "epoch": 0.942445073460035, + "grad_norm": 0.32426032423973083, + "learning_rate": 1.6306890712868727e-06, + "loss": 0.0547, + "step": 40204 + }, + { + "epoch": 0.9424685150348401, + "grad_norm": 0.2313740849494934, + "learning_rate": 1.6293644969124334e-06, + "loss": 0.0374, + "step": 40205 + }, + { + "epoch": 0.942491956609645, + "grad_norm": 0.5135072469711304, + "learning_rate": 1.6280404562998174e-06, + "loss": 0.6042, + "step": 40206 + }, + { + "epoch": 0.9425153981844501, + "grad_norm": 0.4676656126976013, + "learning_rate": 1.6267169494562196e-06, + "loss": 0.0771, + "step": 40207 + }, + { + "epoch": 0.942538839759255, + "grad_norm": 0.363467812538147, + "learning_rate": 1.6253939763888115e-06, + "loss": 0.7544, + "step": 40208 + }, + { + "epoch": 0.9425622813340601, + "grad_norm": 0.36163902282714844, + "learning_rate": 1.6240715371047988e-06, + "loss": 0.0671, + "step": 40209 + }, + { + "epoch": 0.942585722908865, + "grad_norm": 0.17138439416885376, + "learning_rate": 1.62274963161132e-06, + "loss": 0.0229, + "step": 40210 + }, + { + "epoch": 0.94260916448367, + "grad_norm": 0.34775421023368835, + "learning_rate": 1.6214282599155805e-06, + "loss": 0.0278, + "step": 40211 + }, + { + "epoch": 0.942632606058475, + "grad_norm": 0.8371204137802124, + "learning_rate": 1.6201074220247414e-06, + "loss": 0.196, + "step": 40212 + }, + { + "epoch": 0.94265604763328, + "grad_norm": 0.15569816529750824, + "learning_rate": 1.6187871179459635e-06, + "loss": 0.0229, + "step": 40213 + }, + { + "epoch": 0.942679489208085, + "grad_norm": 0.1628042757511139, + "learning_rate": 1.6174673476864078e-06, + "loss": 0.0229, + "step": 40214 + }, + { + "epoch": 0.94270293078289, + "grad_norm": 0.5511121153831482, + "learning_rate": 1.6161481112532461e-06, + "loss": 0.0982, + "step": 40215 + }, + { + "epoch": 0.942726372357695, + "grad_norm": 0.5163761377334595, + "learning_rate": 1.6148294086536286e-06, + "loss": 0.0959, + "step": 40216 + }, + { + "epoch": 0.9427498139325, + "grad_norm": 0.4593321979045868, + "learning_rate": 1.6135112398947161e-06, + "loss": 0.0392, + "step": 40217 + }, + { + "epoch": 0.942773255507305, + "grad_norm": 0.36393097043037415, + "learning_rate": 1.6121936049836584e-06, + "loss": 0.0567, + "step": 40218 + }, + { + "epoch": 0.94279669708211, + "grad_norm": 0.17605307698249817, + "learning_rate": 1.6108765039275941e-06, + "loss": 0.0276, + "step": 40219 + }, + { + "epoch": 0.9428201386569149, + "grad_norm": 0.618500292301178, + "learning_rate": 1.6095599367336954e-06, + "loss": 0.1055, + "step": 40220 + }, + { + "epoch": 0.94284358023172, + "grad_norm": 0.26998379826545715, + "learning_rate": 1.608243903409079e-06, + "loss": 0.3449, + "step": 40221 + }, + { + "epoch": 0.9428670218065249, + "grad_norm": 0.4205257296562195, + "learning_rate": 1.6069284039608945e-06, + "loss": 0.0427, + "step": 40222 + }, + { + "epoch": 0.94289046338133, + "grad_norm": 0.32782888412475586, + "learning_rate": 1.6056134383962917e-06, + "loss": 0.0706, + "step": 40223 + }, + { + "epoch": 0.9429139049561349, + "grad_norm": 0.2664417326450348, + "learning_rate": 1.6042990067223984e-06, + "loss": 0.1593, + "step": 40224 + }, + { + "epoch": 0.94293734653094, + "grad_norm": 0.5155622363090515, + "learning_rate": 1.6029851089463422e-06, + "loss": 0.1049, + "step": 40225 + }, + { + "epoch": 0.9429607881057449, + "grad_norm": 0.184006005525589, + "learning_rate": 1.6016717450752506e-06, + "loss": 0.0229, + "step": 40226 + }, + { + "epoch": 0.94298422968055, + "grad_norm": 0.9138336777687073, + "learning_rate": 1.6003589151162624e-06, + "loss": 0.176, + "step": 40227 + }, + { + "epoch": 0.9430076712553549, + "grad_norm": 0.15948547422885895, + "learning_rate": 1.599046619076483e-06, + "loss": 0.0216, + "step": 40228 + }, + { + "epoch": 0.94303111283016, + "grad_norm": 0.16413335502147675, + "learning_rate": 1.5977348569630623e-06, + "loss": 0.0281, + "step": 40229 + }, + { + "epoch": 0.9430545544049649, + "grad_norm": 0.4006338119506836, + "learning_rate": 1.5964236287830837e-06, + "loss": 0.0442, + "step": 40230 + }, + { + "epoch": 0.9430779959797699, + "grad_norm": 0.3308058977127075, + "learning_rate": 1.5951129345436856e-06, + "loss": 0.0406, + "step": 40231 + }, + { + "epoch": 0.9431014375545749, + "grad_norm": 0.4226281940937042, + "learning_rate": 1.5938027742519735e-06, + "loss": 0.0552, + "step": 40232 + }, + { + "epoch": 0.9431248791293799, + "grad_norm": 0.4907762110233307, + "learning_rate": 1.5924931479150529e-06, + "loss": 0.0781, + "step": 40233 + }, + { + "epoch": 0.9431483207041849, + "grad_norm": 0.5899336338043213, + "learning_rate": 1.5911840555400182e-06, + "loss": 0.0903, + "step": 40234 + }, + { + "epoch": 0.9431717622789899, + "grad_norm": 0.46638891100883484, + "learning_rate": 1.5898754971339858e-06, + "loss": 0.3704, + "step": 40235 + }, + { + "epoch": 0.9431952038537949, + "grad_norm": 0.5780438780784607, + "learning_rate": 1.5885674727040723e-06, + "loss": 0.0342, + "step": 40236 + }, + { + "epoch": 0.9432186454285999, + "grad_norm": 0.5198798179626465, + "learning_rate": 1.5872599822573497e-06, + "loss": 0.5123, + "step": 40237 + }, + { + "epoch": 0.9432420870034048, + "grad_norm": 0.5671568512916565, + "learning_rate": 1.5859530258009235e-06, + "loss": 0.0656, + "step": 40238 + }, + { + "epoch": 0.9432655285782099, + "grad_norm": 0.4450943171977997, + "learning_rate": 1.5846466033418883e-06, + "loss": 0.0517, + "step": 40239 + }, + { + "epoch": 0.9432889701530148, + "grad_norm": 0.5160343647003174, + "learning_rate": 1.583340714887327e-06, + "loss": 0.0674, + "step": 40240 + }, + { + "epoch": 0.9433124117278199, + "grad_norm": 0.4497208595275879, + "learning_rate": 1.5820353604443118e-06, + "loss": 0.0943, + "step": 40241 + }, + { + "epoch": 0.9433358533026248, + "grad_norm": 0.32375457882881165, + "learning_rate": 1.5807305400199478e-06, + "loss": 0.0502, + "step": 40242 + }, + { + "epoch": 0.9433592948774299, + "grad_norm": 0.2550301253795624, + "learning_rate": 1.579426253621319e-06, + "loss": 0.0218, + "step": 40243 + }, + { + "epoch": 0.9433827364522349, + "grad_norm": 0.5147461295127869, + "learning_rate": 1.5781225012554856e-06, + "loss": 0.072, + "step": 40244 + }, + { + "epoch": 0.9434061780270399, + "grad_norm": 0.3470575213432312, + "learning_rate": 1.5768192829295203e-06, + "loss": 0.0516, + "step": 40245 + }, + { + "epoch": 0.9434296196018449, + "grad_norm": 0.41683676838874817, + "learning_rate": 1.575516598650495e-06, + "loss": 0.0878, + "step": 40246 + }, + { + "epoch": 0.9434530611766498, + "grad_norm": 0.5971061587333679, + "learning_rate": 1.5742144484254928e-06, + "loss": 0.0989, + "step": 40247 + }, + { + "epoch": 0.9434765027514549, + "grad_norm": 0.5094653367996216, + "learning_rate": 1.5729128322615637e-06, + "loss": 0.109, + "step": 40248 + }, + { + "epoch": 0.9434999443262598, + "grad_norm": 0.36972203850746155, + "learning_rate": 1.5716117501657912e-06, + "loss": 0.102, + "step": 40249 + }, + { + "epoch": 0.9435233859010649, + "grad_norm": 0.168370321393013, + "learning_rate": 1.5703112021452138e-06, + "loss": 0.032, + "step": 40250 + }, + { + "epoch": 0.9435468274758698, + "grad_norm": 1.1717885732650757, + "learning_rate": 1.5690111882068925e-06, + "loss": 0.1932, + "step": 40251 + }, + { + "epoch": 0.9435702690506749, + "grad_norm": 0.5750923752784729, + "learning_rate": 1.5677117083578885e-06, + "loss": 0.1288, + "step": 40252 + }, + { + "epoch": 0.9435937106254798, + "grad_norm": 0.7339345812797546, + "learning_rate": 1.5664127626052405e-06, + "loss": 0.1178, + "step": 40253 + }, + { + "epoch": 0.9436171522002849, + "grad_norm": 0.1897279918193817, + "learning_rate": 1.5651143509560207e-06, + "loss": 0.0283, + "step": 40254 + }, + { + "epoch": 0.9436405937750898, + "grad_norm": 0.40014657378196716, + "learning_rate": 1.5638164734172456e-06, + "loss": 0.0818, + "step": 40255 + }, + { + "epoch": 0.9436640353498948, + "grad_norm": 0.511025607585907, + "learning_rate": 1.5625191299959762e-06, + "loss": 0.1268, + "step": 40256 + }, + { + "epoch": 0.9436874769246998, + "grad_norm": 0.44877898693084717, + "learning_rate": 1.5612223206992404e-06, + "loss": 0.0785, + "step": 40257 + }, + { + "epoch": 0.9437109184995048, + "grad_norm": 0.30995383858680725, + "learning_rate": 1.559926045534077e-06, + "loss": 0.0412, + "step": 40258 + }, + { + "epoch": 0.9437343600743098, + "grad_norm": 0.3655376434326172, + "learning_rate": 1.5586303045075246e-06, + "loss": 0.0778, + "step": 40259 + }, + { + "epoch": 0.9437578016491148, + "grad_norm": 0.46519726514816284, + "learning_rate": 1.557335097626611e-06, + "loss": 0.1189, + "step": 40260 + }, + { + "epoch": 0.9437812432239198, + "grad_norm": 0.36631062626838684, + "learning_rate": 1.5560404248983751e-06, + "loss": 0.0357, + "step": 40261 + }, + { + "epoch": 0.9438046847987248, + "grad_norm": 0.5729988813400269, + "learning_rate": 1.5547462863298222e-06, + "loss": 0.1207, + "step": 40262 + }, + { + "epoch": 0.9438281263735298, + "grad_norm": 0.12379718571901321, + "learning_rate": 1.5534526819279915e-06, + "loss": 0.0246, + "step": 40263 + }, + { + "epoch": 0.9438515679483348, + "grad_norm": 0.186481311917305, + "learning_rate": 1.5521596116998994e-06, + "loss": 0.0182, + "step": 40264 + }, + { + "epoch": 0.9438750095231397, + "grad_norm": 0.6204285621643066, + "learning_rate": 1.5508670756525512e-06, + "loss": 0.5494, + "step": 40265 + }, + { + "epoch": 0.9438984510979448, + "grad_norm": 0.15391512215137482, + "learning_rate": 1.5495750737929527e-06, + "loss": 0.0282, + "step": 40266 + }, + { + "epoch": 0.9439218926727497, + "grad_norm": 0.4233510196208954, + "learning_rate": 1.5482836061281424e-06, + "loss": 0.0543, + "step": 40267 + }, + { + "epoch": 0.9439453342475548, + "grad_norm": 0.2976509630680084, + "learning_rate": 1.5469926726651151e-06, + "loss": 0.0332, + "step": 40268 + }, + { + "epoch": 0.9439687758223597, + "grad_norm": 0.1001119539141655, + "learning_rate": 1.545702273410865e-06, + "loss": 0.0135, + "step": 40269 + }, + { + "epoch": 0.9439922173971648, + "grad_norm": 0.5179186463356018, + "learning_rate": 1.5444124083724087e-06, + "loss": 0.1407, + "step": 40270 + }, + { + "epoch": 0.9440156589719697, + "grad_norm": 0.13168612122535706, + "learning_rate": 1.5431230775567296e-06, + "loss": 0.0195, + "step": 40271 + }, + { + "epoch": 0.9440391005467748, + "grad_norm": 0.22205394506454468, + "learning_rate": 1.5418342809708442e-06, + "loss": 0.0283, + "step": 40272 + }, + { + "epoch": 0.9440625421215797, + "grad_norm": 0.6799941062927246, + "learning_rate": 1.5405460186217247e-06, + "loss": 0.0951, + "step": 40273 + }, + { + "epoch": 0.9440859836963847, + "grad_norm": 0.34322667121887207, + "learning_rate": 1.5392582905163766e-06, + "loss": 0.0208, + "step": 40274 + }, + { + "epoch": 0.9441094252711897, + "grad_norm": 0.3418826758861542, + "learning_rate": 1.5379710966617834e-06, + "loss": 0.0578, + "step": 40275 + }, + { + "epoch": 0.9441328668459947, + "grad_norm": 0.2927594482898712, + "learning_rate": 1.536684437064928e-06, + "loss": 0.0935, + "step": 40276 + }, + { + "epoch": 0.9441563084207997, + "grad_norm": 0.3687887489795685, + "learning_rate": 1.535398311732794e-06, + "loss": 0.0993, + "step": 40277 + }, + { + "epoch": 0.9441797499956047, + "grad_norm": 0.38212060928344727, + "learning_rate": 1.5341127206723538e-06, + "loss": 0.0444, + "step": 40278 + }, + { + "epoch": 0.9442031915704097, + "grad_norm": 0.41411808133125305, + "learning_rate": 1.5328276638905792e-06, + "loss": 0.0749, + "step": 40279 + }, + { + "epoch": 0.9442266331452147, + "grad_norm": 0.3782363831996918, + "learning_rate": 1.5315431413944647e-06, + "loss": 0.0413, + "step": 40280 + }, + { + "epoch": 0.9442500747200196, + "grad_norm": 0.5853935480117798, + "learning_rate": 1.5302591531909606e-06, + "loss": 0.0907, + "step": 40281 + }, + { + "epoch": 0.9442735162948247, + "grad_norm": 0.47757667303085327, + "learning_rate": 1.5289756992870386e-06, + "loss": 0.0464, + "step": 40282 + }, + { + "epoch": 0.9442969578696296, + "grad_norm": 0.3965863585472107, + "learning_rate": 1.5276927796896489e-06, + "loss": 0.0663, + "step": 40283 + }, + { + "epoch": 0.9443203994444347, + "grad_norm": 0.21060410141944885, + "learning_rate": 1.526410394405786e-06, + "loss": 0.01, + "step": 40284 + }, + { + "epoch": 0.9443438410192396, + "grad_norm": 0.18644890189170837, + "learning_rate": 1.5251285434423779e-06, + "loss": 0.0204, + "step": 40285 + }, + { + "epoch": 0.9443672825940447, + "grad_norm": 0.21255740523338318, + "learning_rate": 1.5238472268063962e-06, + "loss": 0.0376, + "step": 40286 + }, + { + "epoch": 0.9443907241688496, + "grad_norm": 0.5753472447395325, + "learning_rate": 1.5225664445047805e-06, + "loss": 0.156, + "step": 40287 + }, + { + "epoch": 0.9444141657436547, + "grad_norm": 0.5678192377090454, + "learning_rate": 1.5212861965445025e-06, + "loss": 0.1392, + "step": 40288 + }, + { + "epoch": 0.9444376073184596, + "grad_norm": 0.38195839524269104, + "learning_rate": 1.52000648293249e-06, + "loss": 0.0759, + "step": 40289 + }, + { + "epoch": 0.9444610488932647, + "grad_norm": 0.6550577282905579, + "learning_rate": 1.5187273036756932e-06, + "loss": 0.63, + "step": 40290 + }, + { + "epoch": 0.9444844904680696, + "grad_norm": 0.6512896418571472, + "learning_rate": 1.51744865878104e-06, + "loss": 0.1343, + "step": 40291 + }, + { + "epoch": 0.9445079320428746, + "grad_norm": 0.06173331290483475, + "learning_rate": 1.51617054825548e-06, + "loss": 0.0158, + "step": 40292 + }, + { + "epoch": 0.9445313736176796, + "grad_norm": 0.6256596446037292, + "learning_rate": 1.5148929721059634e-06, + "loss": 0.1255, + "step": 40293 + }, + { + "epoch": 0.9445548151924846, + "grad_norm": 0.2754829227924347, + "learning_rate": 1.513615930339396e-06, + "loss": 0.045, + "step": 40294 + }, + { + "epoch": 0.9445782567672897, + "grad_norm": 0.4262964725494385, + "learning_rate": 1.5123394229627164e-06, + "loss": 0.1125, + "step": 40295 + }, + { + "epoch": 0.9446016983420946, + "grad_norm": 0.22375188767910004, + "learning_rate": 1.5110634499828525e-06, + "loss": 0.0386, + "step": 40296 + }, + { + "epoch": 0.9446251399168997, + "grad_norm": 0.5655654072761536, + "learning_rate": 1.509788011406732e-06, + "loss": 0.1654, + "step": 40297 + }, + { + "epoch": 0.9446485814917046, + "grad_norm": 0.19895261526107788, + "learning_rate": 1.5085131072412607e-06, + "loss": 0.0418, + "step": 40298 + }, + { + "epoch": 0.9446720230665097, + "grad_norm": 0.2636268436908722, + "learning_rate": 1.5072387374933772e-06, + "loss": 0.0651, + "step": 40299 + }, + { + "epoch": 0.9446954646413146, + "grad_norm": 0.5420771837234497, + "learning_rate": 1.5059649021699872e-06, + "loss": 0.7557, + "step": 40300 + }, + { + "epoch": 0.9447189062161196, + "grad_norm": 0.40539512038230896, + "learning_rate": 1.5046916012779965e-06, + "loss": 0.1098, + "step": 40301 + }, + { + "epoch": 0.9447423477909246, + "grad_norm": 0.6759646534919739, + "learning_rate": 1.5034188348243217e-06, + "loss": 0.1167, + "step": 40302 + }, + { + "epoch": 0.9447657893657296, + "grad_norm": 0.211061492562294, + "learning_rate": 1.5021466028158681e-06, + "loss": 0.0225, + "step": 40303 + }, + { + "epoch": 0.9447892309405346, + "grad_norm": 0.41961708664894104, + "learning_rate": 1.5008749052595195e-06, + "loss": 0.5836, + "step": 40304 + }, + { + "epoch": 0.9448126725153396, + "grad_norm": 0.5421909689903259, + "learning_rate": 1.4996037421622145e-06, + "loss": 0.0985, + "step": 40305 + }, + { + "epoch": 0.9448361140901446, + "grad_norm": 0.5727459788322449, + "learning_rate": 1.4983331135308142e-06, + "loss": 0.096, + "step": 40306 + }, + { + "epoch": 0.9448595556649496, + "grad_norm": 0.2696989178657532, + "learning_rate": 1.4970630193722356e-06, + "loss": 0.051, + "step": 40307 + }, + { + "epoch": 0.9448829972397546, + "grad_norm": 0.38316357135772705, + "learning_rate": 1.4957934596933621e-06, + "loss": 0.0304, + "step": 40308 + }, + { + "epoch": 0.9449064388145596, + "grad_norm": 0.45261338353157043, + "learning_rate": 1.4945244345010767e-06, + "loss": 0.0501, + "step": 40309 + }, + { + "epoch": 0.9449298803893645, + "grad_norm": 0.504557192325592, + "learning_rate": 1.4932559438022854e-06, + "loss": 0.0615, + "step": 40310 + }, + { + "epoch": 0.9449533219641696, + "grad_norm": 0.2806970775127411, + "learning_rate": 1.491987987603838e-06, + "loss": 0.0537, + "step": 40311 + }, + { + "epoch": 0.9449767635389745, + "grad_norm": 0.22825320065021515, + "learning_rate": 1.490720565912651e-06, + "loss": 0.0352, + "step": 40312 + }, + { + "epoch": 0.9450002051137796, + "grad_norm": 0.15033769607543945, + "learning_rate": 1.489453678735575e-06, + "loss": 0.0306, + "step": 40313 + }, + { + "epoch": 0.9450236466885845, + "grad_norm": 0.17234642803668976, + "learning_rate": 1.4881873260794932e-06, + "loss": 0.0249, + "step": 40314 + }, + { + "epoch": 0.9450470882633896, + "grad_norm": 0.3596208989620209, + "learning_rate": 1.4869215079512888e-06, + "loss": 0.0738, + "step": 40315 + }, + { + "epoch": 0.9450705298381945, + "grad_norm": 0.3032201826572418, + "learning_rate": 1.485656224357801e-06, + "loss": 0.0502, + "step": 40316 + }, + { + "epoch": 0.9450939714129996, + "grad_norm": 0.4636690616607666, + "learning_rate": 1.484391475305924e-06, + "loss": 0.0646, + "step": 40317 + }, + { + "epoch": 0.9451174129878045, + "grad_norm": 0.7449455857276917, + "learning_rate": 1.483127260802497e-06, + "loss": 0.141, + "step": 40318 + }, + { + "epoch": 0.9451408545626095, + "grad_norm": 0.3395855724811554, + "learning_rate": 1.4818635808544034e-06, + "loss": 0.0857, + "step": 40319 + }, + { + "epoch": 0.9451642961374145, + "grad_norm": 0.4836272895336151, + "learning_rate": 1.4806004354684821e-06, + "loss": 0.6205, + "step": 40320 + }, + { + "epoch": 0.9451877377122195, + "grad_norm": 0.23334065079689026, + "learning_rate": 1.4793378246515942e-06, + "loss": 0.037, + "step": 40321 + }, + { + "epoch": 0.9452111792870245, + "grad_norm": 0.4707893133163452, + "learning_rate": 1.4780757484105901e-06, + "loss": 0.0503, + "step": 40322 + }, + { + "epoch": 0.9452346208618295, + "grad_norm": 0.1695885956287384, + "learning_rate": 1.4768142067523195e-06, + "loss": 0.0164, + "step": 40323 + }, + { + "epoch": 0.9452580624366345, + "grad_norm": 0.052380818873643875, + "learning_rate": 1.4755531996836213e-06, + "loss": 0.0061, + "step": 40324 + }, + { + "epoch": 0.9452815040114395, + "grad_norm": 0.19165220856666565, + "learning_rate": 1.474292727211346e-06, + "loss": 0.0507, + "step": 40325 + }, + { + "epoch": 0.9453049455862444, + "grad_norm": 0.23063668608665466, + "learning_rate": 1.4730327893423213e-06, + "loss": 0.0281, + "step": 40326 + }, + { + "epoch": 0.9453283871610495, + "grad_norm": 1.083604097366333, + "learning_rate": 1.471773386083397e-06, + "loss": 0.113, + "step": 40327 + }, + { + "epoch": 0.9453518287358544, + "grad_norm": 0.13951171934604645, + "learning_rate": 1.4705145174414016e-06, + "loss": 0.0123, + "step": 40328 + }, + { + "epoch": 0.9453752703106595, + "grad_norm": 0.4206148087978363, + "learning_rate": 1.4692561834231622e-06, + "loss": 0.0431, + "step": 40329 + }, + { + "epoch": 0.9453987118854644, + "grad_norm": 0.6367829442024231, + "learning_rate": 1.4679983840355072e-06, + "loss": 0.1063, + "step": 40330 + }, + { + "epoch": 0.9454221534602695, + "grad_norm": 0.38957375288009644, + "learning_rate": 1.4667411192852642e-06, + "loss": 0.0846, + "step": 40331 + }, + { + "epoch": 0.9454455950350744, + "grad_norm": 0.4336162507534027, + "learning_rate": 1.4654843891792502e-06, + "loss": 0.0491, + "step": 40332 + }, + { + "epoch": 0.9454690366098795, + "grad_norm": 0.5249316096305847, + "learning_rate": 1.464228193724304e-06, + "loss": 0.1275, + "step": 40333 + }, + { + "epoch": 0.9454924781846844, + "grad_norm": 0.29221484065055847, + "learning_rate": 1.4629725329272204e-06, + "loss": 0.0451, + "step": 40334 + }, + { + "epoch": 0.9455159197594895, + "grad_norm": 0.21822406351566315, + "learning_rate": 1.4617174067948159e-06, + "loss": 0.0291, + "step": 40335 + }, + { + "epoch": 0.9455393613342944, + "grad_norm": 0.23837910592556, + "learning_rate": 1.460462815333896e-06, + "loss": 0.0379, + "step": 40336 + }, + { + "epoch": 0.9455628029090994, + "grad_norm": 0.16237585246562958, + "learning_rate": 1.4592087585512782e-06, + "loss": 0.0247, + "step": 40337 + }, + { + "epoch": 0.9455862444839044, + "grad_norm": 0.1812955141067505, + "learning_rate": 1.4579552364537675e-06, + "loss": 0.0379, + "step": 40338 + }, + { + "epoch": 0.9456096860587094, + "grad_norm": 0.19150036573410034, + "learning_rate": 1.4567022490481697e-06, + "loss": 0.0211, + "step": 40339 + }, + { + "epoch": 0.9456331276335144, + "grad_norm": 0.37171101570129395, + "learning_rate": 1.4554497963412684e-06, + "loss": 0.4761, + "step": 40340 + }, + { + "epoch": 0.9456565692083194, + "grad_norm": 0.07208847999572754, + "learning_rate": 1.4541978783398691e-06, + "loss": 0.0137, + "step": 40341 + }, + { + "epoch": 0.9456800107831244, + "grad_norm": 0.49539947509765625, + "learning_rate": 1.4529464950507554e-06, + "loss": 0.3749, + "step": 40342 + }, + { + "epoch": 0.9457034523579294, + "grad_norm": 0.24420461058616638, + "learning_rate": 1.451695646480733e-06, + "loss": 0.0296, + "step": 40343 + }, + { + "epoch": 0.9457268939327343, + "grad_norm": 0.20033830404281616, + "learning_rate": 1.450445332636574e-06, + "loss": 0.0264, + "step": 40344 + }, + { + "epoch": 0.9457503355075394, + "grad_norm": 0.45454075932502747, + "learning_rate": 1.4491955535250845e-06, + "loss": 0.1208, + "step": 40345 + }, + { + "epoch": 0.9457737770823444, + "grad_norm": 0.385185569524765, + "learning_rate": 1.4479463091530143e-06, + "loss": 0.032, + "step": 40346 + }, + { + "epoch": 0.9457972186571494, + "grad_norm": 0.5492005348205566, + "learning_rate": 1.4466975995271693e-06, + "loss": 0.0607, + "step": 40347 + }, + { + "epoch": 0.9458206602319544, + "grad_norm": 0.5517034530639648, + "learning_rate": 1.4454494246543104e-06, + "loss": 0.5966, + "step": 40348 + }, + { + "epoch": 0.9458441018067594, + "grad_norm": 0.5998849272727966, + "learning_rate": 1.4442017845412104e-06, + "loss": 0.1411, + "step": 40349 + }, + { + "epoch": 0.9458675433815644, + "grad_norm": 0.4210022985935211, + "learning_rate": 1.4429546791946524e-06, + "loss": 0.0609, + "step": 40350 + }, + { + "epoch": 0.9458909849563694, + "grad_norm": 0.27382487058639526, + "learning_rate": 1.4417081086213868e-06, + "loss": 0.0521, + "step": 40351 + }, + { + "epoch": 0.9459144265311744, + "grad_norm": 0.32334813475608826, + "learning_rate": 1.440462072828186e-06, + "loss": 0.0339, + "step": 40352 + }, + { + "epoch": 0.9459378681059794, + "grad_norm": 0.35749804973602295, + "learning_rate": 1.4392165718218108e-06, + "loss": 0.0634, + "step": 40353 + }, + { + "epoch": 0.9459613096807844, + "grad_norm": 0.2530304789543152, + "learning_rate": 1.437971605609023e-06, + "loss": 0.0443, + "step": 40354 + }, + { + "epoch": 0.9459847512555893, + "grad_norm": 0.7172262668609619, + "learning_rate": 1.4367271741965504e-06, + "loss": 0.0985, + "step": 40355 + }, + { + "epoch": 0.9460081928303944, + "grad_norm": 0.7223042845726013, + "learning_rate": 1.4354832775911763e-06, + "loss": 0.6107, + "step": 40356 + }, + { + "epoch": 0.9460316344051993, + "grad_norm": 0.3117673099040985, + "learning_rate": 1.4342399157996512e-06, + "loss": 0.0511, + "step": 40357 + }, + { + "epoch": 0.9460550759800044, + "grad_norm": 0.46974825859069824, + "learning_rate": 1.4329970888287136e-06, + "loss": 0.0435, + "step": 40358 + }, + { + "epoch": 0.9460785175548093, + "grad_norm": 0.2158087193965912, + "learning_rate": 1.431754796685103e-06, + "loss": 0.0362, + "step": 40359 + }, + { + "epoch": 0.9461019591296144, + "grad_norm": 0.34725281596183777, + "learning_rate": 1.4305130393755584e-06, + "loss": 0.064, + "step": 40360 + }, + { + "epoch": 0.9461254007044193, + "grad_norm": 0.6977618932723999, + "learning_rate": 1.4292718169068075e-06, + "loss": 0.1226, + "step": 40361 + }, + { + "epoch": 0.9461488422792244, + "grad_norm": 0.5724250674247742, + "learning_rate": 1.428031129285623e-06, + "loss": 0.0723, + "step": 40362 + }, + { + "epoch": 0.9461722838540293, + "grad_norm": 0.08673932403326035, + "learning_rate": 1.4267909765186994e-06, + "loss": 0.0153, + "step": 40363 + }, + { + "epoch": 0.9461957254288343, + "grad_norm": 0.6688936948776245, + "learning_rate": 1.4255513586127867e-06, + "loss": 0.0484, + "step": 40364 + }, + { + "epoch": 0.9462191670036393, + "grad_norm": 0.11136951297521591, + "learning_rate": 1.424312275574602e-06, + "loss": 0.025, + "step": 40365 + }, + { + "epoch": 0.9462426085784443, + "grad_norm": 0.5378316640853882, + "learning_rate": 1.4230737274108729e-06, + "loss": 0.1202, + "step": 40366 + }, + { + "epoch": 0.9462660501532493, + "grad_norm": 0.3085378408432007, + "learning_rate": 1.4218357141283166e-06, + "loss": 0.0372, + "step": 40367 + }, + { + "epoch": 0.9462894917280543, + "grad_norm": 0.14943067729473114, + "learning_rate": 1.420598235733639e-06, + "loss": 0.0188, + "step": 40368 + }, + { + "epoch": 0.9463129333028593, + "grad_norm": 0.36396586894989014, + "learning_rate": 1.4193612922335785e-06, + "loss": 0.0328, + "step": 40369 + }, + { + "epoch": 0.9463363748776643, + "grad_norm": 0.5582082867622375, + "learning_rate": 1.4181248836348414e-06, + "loss": 0.1269, + "step": 40370 + }, + { + "epoch": 0.9463598164524692, + "grad_norm": 0.3442985713481903, + "learning_rate": 1.4168890099441224e-06, + "loss": 0.0526, + "step": 40371 + }, + { + "epoch": 0.9463832580272743, + "grad_norm": 0.302427738904953, + "learning_rate": 1.4156536711681378e-06, + "loss": 0.0274, + "step": 40372 + }, + { + "epoch": 0.9464066996020792, + "grad_norm": 0.2936333417892456, + "learning_rate": 1.4144188673135827e-06, + "loss": 0.0378, + "step": 40373 + }, + { + "epoch": 0.9464301411768843, + "grad_norm": 0.38197168707847595, + "learning_rate": 1.4131845983871628e-06, + "loss": 0.0626, + "step": 40374 + }, + { + "epoch": 0.9464535827516892, + "grad_norm": 0.10400184988975525, + "learning_rate": 1.4119508643955837e-06, + "loss": 0.0102, + "step": 40375 + }, + { + "epoch": 0.9464770243264943, + "grad_norm": 0.28637903928756714, + "learning_rate": 1.410717665345529e-06, + "loss": 0.0292, + "step": 40376 + }, + { + "epoch": 0.9465004659012992, + "grad_norm": 0.723562479019165, + "learning_rate": 1.4094850012436934e-06, + "loss": 0.1394, + "step": 40377 + }, + { + "epoch": 0.9465239074761043, + "grad_norm": 0.5861868858337402, + "learning_rate": 1.4082528720967602e-06, + "loss": 0.1168, + "step": 40378 + }, + { + "epoch": 0.9465473490509092, + "grad_norm": 0.3220553994178772, + "learning_rate": 1.4070212779114244e-06, + "loss": 0.0544, + "step": 40379 + }, + { + "epoch": 0.9465707906257143, + "grad_norm": 0.11945798248052597, + "learning_rate": 1.4057902186943583e-06, + "loss": 0.0202, + "step": 40380 + }, + { + "epoch": 0.9465942322005192, + "grad_norm": 0.4153221547603607, + "learning_rate": 1.4045596944522343e-06, + "loss": 0.0531, + "step": 40381 + }, + { + "epoch": 0.9466176737753242, + "grad_norm": 0.47267624735832214, + "learning_rate": 1.4033297051917581e-06, + "loss": 0.0942, + "step": 40382 + }, + { + "epoch": 0.9466411153501292, + "grad_norm": 0.2703016996383667, + "learning_rate": 1.4021002509195912e-06, + "loss": 0.0632, + "step": 40383 + }, + { + "epoch": 0.9466645569249342, + "grad_norm": 0.5767351984977722, + "learning_rate": 1.400871331642395e-06, + "loss": 0.0898, + "step": 40384 + }, + { + "epoch": 0.9466879984997392, + "grad_norm": 0.39726772904396057, + "learning_rate": 1.3996429473668414e-06, + "loss": 0.05, + "step": 40385 + }, + { + "epoch": 0.9467114400745442, + "grad_norm": 0.3879777789115906, + "learning_rate": 1.3984150980996036e-06, + "loss": 0.0837, + "step": 40386 + }, + { + "epoch": 0.9467348816493492, + "grad_norm": 0.3280538022518158, + "learning_rate": 1.3971877838473203e-06, + "loss": 0.3962, + "step": 40387 + }, + { + "epoch": 0.9467583232241542, + "grad_norm": 0.8826581835746765, + "learning_rate": 1.395961004616686e-06, + "loss": 0.1054, + "step": 40388 + }, + { + "epoch": 0.9467817647989591, + "grad_norm": 0.1475784033536911, + "learning_rate": 1.3947347604143402e-06, + "loss": 0.0275, + "step": 40389 + }, + { + "epoch": 0.9468052063737642, + "grad_norm": 0.42021000385284424, + "learning_rate": 1.393509051246933e-06, + "loss": 0.0594, + "step": 40390 + }, + { + "epoch": 0.9468286479485691, + "grad_norm": 0.12350817024707794, + "learning_rate": 1.3922838771211143e-06, + "loss": 0.0122, + "step": 40391 + }, + { + "epoch": 0.9468520895233742, + "grad_norm": 0.45035162568092346, + "learning_rate": 1.391059238043535e-06, + "loss": 0.0599, + "step": 40392 + }, + { + "epoch": 0.9468755310981791, + "grad_norm": 0.3031960427761078, + "learning_rate": 1.3898351340208448e-06, + "loss": 0.0402, + "step": 40393 + }, + { + "epoch": 0.9468989726729842, + "grad_norm": 0.35900986194610596, + "learning_rate": 1.388611565059672e-06, + "loss": 0.0519, + "step": 40394 + }, + { + "epoch": 0.9469224142477891, + "grad_norm": 0.48157164454460144, + "learning_rate": 1.387388531166678e-06, + "loss": 0.0583, + "step": 40395 + }, + { + "epoch": 0.9469458558225942, + "grad_norm": 0.3866709768772125, + "learning_rate": 1.3861660323484793e-06, + "loss": 0.4162, + "step": 40396 + }, + { + "epoch": 0.9469692973973992, + "grad_norm": 0.5741996765136719, + "learning_rate": 1.3849440686117266e-06, + "loss": 0.1069, + "step": 40397 + }, + { + "epoch": 0.9469927389722042, + "grad_norm": 0.38351157307624817, + "learning_rate": 1.3837226399630365e-06, + "loss": 0.0687, + "step": 40398 + }, + { + "epoch": 0.9470161805470092, + "grad_norm": 0.6278079152107239, + "learning_rate": 1.3825017464090262e-06, + "loss": 0.1091, + "step": 40399 + }, + { + "epoch": 0.9470396221218141, + "grad_norm": 0.20146781206130981, + "learning_rate": 1.3812813879563457e-06, + "loss": 0.0467, + "step": 40400 + }, + { + "epoch": 0.9470630636966192, + "grad_norm": 0.35839712619781494, + "learning_rate": 1.3800615646116123e-06, + "loss": 0.0693, + "step": 40401 + }, + { + "epoch": 0.9470865052714241, + "grad_norm": 0.6128615140914917, + "learning_rate": 1.3788422763814313e-06, + "loss": 0.3046, + "step": 40402 + }, + { + "epoch": 0.9471099468462292, + "grad_norm": 0.2602400779724121, + "learning_rate": 1.37762352327242e-06, + "loss": 0.0444, + "step": 40403 + }, + { + "epoch": 0.9471333884210341, + "grad_norm": 0.26681846380233765, + "learning_rate": 1.3764053052911953e-06, + "loss": 0.0367, + "step": 40404 + }, + { + "epoch": 0.9471568299958392, + "grad_norm": 0.35298484563827515, + "learning_rate": 1.3751876224443738e-06, + "loss": 0.0543, + "step": 40405 + }, + { + "epoch": 0.9471802715706441, + "grad_norm": 0.30411362648010254, + "learning_rate": 1.3739704747385506e-06, + "loss": 0.037, + "step": 40406 + }, + { + "epoch": 0.9472037131454492, + "grad_norm": 0.5322631001472473, + "learning_rate": 1.3727538621803538e-06, + "loss": 0.1031, + "step": 40407 + }, + { + "epoch": 0.9472271547202541, + "grad_norm": 0.3467201888561249, + "learning_rate": 1.3715377847763554e-06, + "loss": 0.0728, + "step": 40408 + }, + { + "epoch": 0.9472505962950591, + "grad_norm": 0.509427011013031, + "learning_rate": 1.370322242533173e-06, + "loss": 0.1015, + "step": 40409 + }, + { + "epoch": 0.9472740378698641, + "grad_norm": 0.15831732749938965, + "learning_rate": 1.3691072354573898e-06, + "loss": 0.0253, + "step": 40410 + }, + { + "epoch": 0.9472974794446691, + "grad_norm": 0.19862253963947296, + "learning_rate": 1.3678927635556004e-06, + "loss": 0.017, + "step": 40411 + }, + { + "epoch": 0.9473209210194741, + "grad_norm": 0.24847565591335297, + "learning_rate": 1.3666788268344e-06, + "loss": 0.0269, + "step": 40412 + }, + { + "epoch": 0.9473443625942791, + "grad_norm": 0.27792415022850037, + "learning_rate": 1.3654654253003828e-06, + "loss": 0.0257, + "step": 40413 + }, + { + "epoch": 0.9473678041690841, + "grad_norm": 0.2008262574672699, + "learning_rate": 1.3642525589601106e-06, + "loss": 0.0382, + "step": 40414 + }, + { + "epoch": 0.9473912457438891, + "grad_norm": 0.6339659690856934, + "learning_rate": 1.3630402278201892e-06, + "loss": 0.1052, + "step": 40415 + }, + { + "epoch": 0.947414687318694, + "grad_norm": 0.2550514340400696, + "learning_rate": 1.361828431887169e-06, + "loss": 0.0459, + "step": 40416 + }, + { + "epoch": 0.9474381288934991, + "grad_norm": 0.3889862895011902, + "learning_rate": 1.3606171711676552e-06, + "loss": 0.0595, + "step": 40417 + }, + { + "epoch": 0.947461570468304, + "grad_norm": 0.24126745760440826, + "learning_rate": 1.35940644566821e-06, + "loss": 0.0242, + "step": 40418 + }, + { + "epoch": 0.9474850120431091, + "grad_norm": 0.030614130198955536, + "learning_rate": 1.3581962553953832e-06, + "loss": 0.0015, + "step": 40419 + }, + { + "epoch": 0.947508453617914, + "grad_norm": 0.41540682315826416, + "learning_rate": 1.3569866003557697e-06, + "loss": 0.0478, + "step": 40420 + }, + { + "epoch": 0.9475318951927191, + "grad_norm": 0.6054580807685852, + "learning_rate": 1.3557774805559086e-06, + "loss": 0.0871, + "step": 40421 + }, + { + "epoch": 0.947555336767524, + "grad_norm": 0.6709165573120117, + "learning_rate": 1.354568896002384e-06, + "loss": 0.1072, + "step": 40422 + }, + { + "epoch": 0.9475787783423291, + "grad_norm": 0.34368202090263367, + "learning_rate": 1.3533608467017455e-06, + "loss": 0.0636, + "step": 40423 + }, + { + "epoch": 0.947602219917134, + "grad_norm": 0.08863506466150284, + "learning_rate": 1.3521533326605329e-06, + "loss": 0.0126, + "step": 40424 + }, + { + "epoch": 0.947625661491939, + "grad_norm": 0.31639334559440613, + "learning_rate": 1.3509463538853074e-06, + "loss": 0.0721, + "step": 40425 + }, + { + "epoch": 0.947649103066744, + "grad_norm": 0.1663818657398224, + "learning_rate": 1.3497399103826303e-06, + "loss": 0.0222, + "step": 40426 + }, + { + "epoch": 0.947672544641549, + "grad_norm": 0.12318269908428192, + "learning_rate": 1.348534002159041e-06, + "loss": 0.018, + "step": 40427 + }, + { + "epoch": 0.947695986216354, + "grad_norm": 0.34877678751945496, + "learning_rate": 1.3473286292210785e-06, + "loss": 0.036, + "step": 40428 + }, + { + "epoch": 0.947719427791159, + "grad_norm": 0.31697821617126465, + "learning_rate": 1.346123791575271e-06, + "loss": 0.0397, + "step": 40429 + }, + { + "epoch": 0.947742869365964, + "grad_norm": 0.15256138145923615, + "learning_rate": 1.3449194892281802e-06, + "loss": 0.0218, + "step": 40430 + }, + { + "epoch": 0.947766310940769, + "grad_norm": 0.10594267398118973, + "learning_rate": 1.343715722186334e-06, + "loss": 0.0083, + "step": 40431 + }, + { + "epoch": 0.947789752515574, + "grad_norm": 0.721725583076477, + "learning_rate": 1.342512490456249e-06, + "loss": 0.6053, + "step": 40432 + }, + { + "epoch": 0.947813194090379, + "grad_norm": 0.13684573769569397, + "learning_rate": 1.3413097940444764e-06, + "loss": 0.0109, + "step": 40433 + }, + { + "epoch": 0.947836635665184, + "grad_norm": 0.12991808354854584, + "learning_rate": 1.3401076329575325e-06, + "loss": 0.0162, + "step": 40434 + }, + { + "epoch": 0.947860077239989, + "grad_norm": 0.5707672834396362, + "learning_rate": 1.3389060072019344e-06, + "loss": 0.0592, + "step": 40435 + }, + { + "epoch": 0.9478835188147939, + "grad_norm": 0.5203763246536255, + "learning_rate": 1.3377049167841994e-06, + "loss": 0.0617, + "step": 40436 + }, + { + "epoch": 0.947906960389599, + "grad_norm": 0.1566026359796524, + "learning_rate": 1.3365043617108553e-06, + "loss": 0.018, + "step": 40437 + }, + { + "epoch": 0.9479304019644039, + "grad_norm": 0.16277413070201874, + "learning_rate": 1.3353043419884192e-06, + "loss": 0.0137, + "step": 40438 + }, + { + "epoch": 0.947953843539209, + "grad_norm": 0.7572587132453918, + "learning_rate": 1.3341048576233862e-06, + "loss": 0.1057, + "step": 40439 + }, + { + "epoch": 0.9479772851140139, + "grad_norm": 0.4826236963272095, + "learning_rate": 1.332905908622284e-06, + "loss": 0.0868, + "step": 40440 + }, + { + "epoch": 0.948000726688819, + "grad_norm": 0.3300603926181793, + "learning_rate": 1.3317074949915965e-06, + "loss": 0.0642, + "step": 40441 + }, + { + "epoch": 0.9480241682636239, + "grad_norm": 0.3592052161693573, + "learning_rate": 1.3305096167378516e-06, + "loss": 0.0516, + "step": 40442 + }, + { + "epoch": 0.948047609838429, + "grad_norm": 0.4362410604953766, + "learning_rate": 1.3293122738675333e-06, + "loss": 0.0482, + "step": 40443 + }, + { + "epoch": 0.9480710514132339, + "grad_norm": 0.12915009260177612, + "learning_rate": 1.3281154663871254e-06, + "loss": 0.0166, + "step": 40444 + }, + { + "epoch": 0.9480944929880389, + "grad_norm": 0.562139093875885, + "learning_rate": 1.3269191943031555e-06, + "loss": 0.1029, + "step": 40445 + }, + { + "epoch": 0.9481179345628439, + "grad_norm": 0.7109981179237366, + "learning_rate": 1.3257234576220967e-06, + "loss": 0.113, + "step": 40446 + }, + { + "epoch": 0.9481413761376489, + "grad_norm": 0.19232673943042755, + "learning_rate": 1.3245282563504323e-06, + "loss": 0.0263, + "step": 40447 + }, + { + "epoch": 0.948164817712454, + "grad_norm": 0.15257789194583893, + "learning_rate": 1.3233335904946465e-06, + "loss": 0.0275, + "step": 40448 + }, + { + "epoch": 0.9481882592872589, + "grad_norm": 0.1624457985162735, + "learning_rate": 1.3221394600612335e-06, + "loss": 0.013, + "step": 40449 + }, + { + "epoch": 0.948211700862064, + "grad_norm": 0.22929860651493073, + "learning_rate": 1.3209458650566665e-06, + "loss": 0.0147, + "step": 40450 + }, + { + "epoch": 0.9482351424368689, + "grad_norm": 0.3244331479072571, + "learning_rate": 1.3197528054874175e-06, + "loss": 0.0373, + "step": 40451 + }, + { + "epoch": 0.948258584011674, + "grad_norm": 0.3813541531562805, + "learning_rate": 1.3185602813599708e-06, + "loss": 0.0776, + "step": 40452 + }, + { + "epoch": 0.9482820255864789, + "grad_norm": 0.36110904812812805, + "learning_rate": 1.3173682926807763e-06, + "loss": 0.0629, + "step": 40453 + }, + { + "epoch": 0.9483054671612839, + "grad_norm": 0.4822019934654236, + "learning_rate": 1.316176839456329e-06, + "loss": 0.4067, + "step": 40454 + }, + { + "epoch": 0.9483289087360889, + "grad_norm": 0.6035541892051697, + "learning_rate": 1.3149859216930794e-06, + "loss": 0.4766, + "step": 40455 + }, + { + "epoch": 0.9483523503108939, + "grad_norm": 0.4469802975654602, + "learning_rate": 1.3137955393974889e-06, + "loss": 0.0934, + "step": 40456 + }, + { + "epoch": 0.9483757918856989, + "grad_norm": 0.20998382568359375, + "learning_rate": 1.3126056925760189e-06, + "loss": 0.0361, + "step": 40457 + }, + { + "epoch": 0.9483992334605039, + "grad_norm": 0.6527416110038757, + "learning_rate": 1.311416381235131e-06, + "loss": 0.1165, + "step": 40458 + }, + { + "epoch": 0.9484226750353089, + "grad_norm": 0.5433974862098694, + "learning_rate": 1.3102276053812646e-06, + "loss": 0.0903, + "step": 40459 + }, + { + "epoch": 0.9484461166101139, + "grad_norm": 0.516170084476471, + "learning_rate": 1.309039365020881e-06, + "loss": 0.117, + "step": 40460 + }, + { + "epoch": 0.9484695581849188, + "grad_norm": 0.7097107768058777, + "learning_rate": 1.3078516601604308e-06, + "loss": 0.0659, + "step": 40461 + }, + { + "epoch": 0.9484929997597239, + "grad_norm": 0.7664266228675842, + "learning_rate": 1.306664490806353e-06, + "loss": 0.0984, + "step": 40462 + }, + { + "epoch": 0.9485164413345288, + "grad_norm": 0.5628859996795654, + "learning_rate": 1.305477856965076e-06, + "loss": 0.4953, + "step": 40463 + }, + { + "epoch": 0.9485398829093339, + "grad_norm": 0.3913798928260803, + "learning_rate": 1.3042917586430504e-06, + "loss": 0.0391, + "step": 40464 + }, + { + "epoch": 0.9485633244841388, + "grad_norm": 0.3391769528388977, + "learning_rate": 1.303106195846715e-06, + "loss": 0.0662, + "step": 40465 + }, + { + "epoch": 0.9485867660589439, + "grad_norm": 0.3454459011554718, + "learning_rate": 1.3019211685825095e-06, + "loss": 0.0676, + "step": 40466 + }, + { + "epoch": 0.9486102076337488, + "grad_norm": 0.326874703168869, + "learning_rate": 1.3007366768568506e-06, + "loss": 0.0608, + "step": 40467 + }, + { + "epoch": 0.9486336492085539, + "grad_norm": 0.504621148109436, + "learning_rate": 1.2995527206761782e-06, + "loss": 0.083, + "step": 40468 + }, + { + "epoch": 0.9486570907833588, + "grad_norm": 0.3926531672477722, + "learning_rate": 1.2983693000468976e-06, + "loss": 0.0556, + "step": 40469 + }, + { + "epoch": 0.9486805323581639, + "grad_norm": 0.10197929292917252, + "learning_rate": 1.2971864149754375e-06, + "loss": 0.017, + "step": 40470 + }, + { + "epoch": 0.9487039739329688, + "grad_norm": 0.3604060709476471, + "learning_rate": 1.296004065468226e-06, + "loss": 0.0394, + "step": 40471 + }, + { + "epoch": 0.9487274155077738, + "grad_norm": 0.4831751883029938, + "learning_rate": 1.2948222515316688e-06, + "loss": 0.077, + "step": 40472 + }, + { + "epoch": 0.9487508570825788, + "grad_norm": 0.29406583309173584, + "learning_rate": 1.2936409731721832e-06, + "loss": 0.0415, + "step": 40473 + }, + { + "epoch": 0.9487742986573838, + "grad_norm": 0.4885011613368988, + "learning_rate": 1.2924602303961864e-06, + "loss": 0.0511, + "step": 40474 + }, + { + "epoch": 0.9487977402321888, + "grad_norm": 0.6578994393348694, + "learning_rate": 1.291280023210062e-06, + "loss": 0.0747, + "step": 40475 + }, + { + "epoch": 0.9488211818069938, + "grad_norm": 0.12707042694091797, + "learning_rate": 1.290100351620227e-06, + "loss": 0.0313, + "step": 40476 + }, + { + "epoch": 0.9488446233817988, + "grad_norm": 0.35408005118370056, + "learning_rate": 1.2889212156330877e-06, + "loss": 0.0744, + "step": 40477 + }, + { + "epoch": 0.9488680649566038, + "grad_norm": 0.3715212047100067, + "learning_rate": 1.2877426152550387e-06, + "loss": 0.0734, + "step": 40478 + }, + { + "epoch": 0.9488915065314087, + "grad_norm": 0.2626695930957794, + "learning_rate": 1.286564550492475e-06, + "loss": 0.0245, + "step": 40479 + }, + { + "epoch": 0.9489149481062138, + "grad_norm": 0.22547176480293274, + "learning_rate": 1.2853870213517805e-06, + "loss": 0.0311, + "step": 40480 + }, + { + "epoch": 0.9489383896810187, + "grad_norm": 0.2760680913925171, + "learning_rate": 1.284210027839361e-06, + "loss": 0.0204, + "step": 40481 + }, + { + "epoch": 0.9489618312558238, + "grad_norm": 0.44527822732925415, + "learning_rate": 1.283033569961578e-06, + "loss": 0.0497, + "step": 40482 + }, + { + "epoch": 0.9489852728306287, + "grad_norm": 0.4250393509864807, + "learning_rate": 1.281857647724849e-06, + "loss": 0.0718, + "step": 40483 + }, + { + "epoch": 0.9490087144054338, + "grad_norm": 0.393232524394989, + "learning_rate": 1.280682261135524e-06, + "loss": 0.665, + "step": 40484 + }, + { + "epoch": 0.9490321559802387, + "grad_norm": 0.24036827683448792, + "learning_rate": 1.2795074101999982e-06, + "loss": 0.0224, + "step": 40485 + }, + { + "epoch": 0.9490555975550438, + "grad_norm": 0.32464680075645447, + "learning_rate": 1.2783330949246441e-06, + "loss": 0.0516, + "step": 40486 + }, + { + "epoch": 0.9490790391298487, + "grad_norm": 0.5077084302902222, + "learning_rate": 1.2771593153158234e-06, + "loss": 0.0868, + "step": 40487 + }, + { + "epoch": 0.9491024807046538, + "grad_norm": 0.5177802443504333, + "learning_rate": 1.2759860713799088e-06, + "loss": 0.0939, + "step": 40488 + }, + { + "epoch": 0.9491259222794587, + "grad_norm": 0.3984009325504303, + "learning_rate": 1.2748133631232729e-06, + "loss": 0.0529, + "step": 40489 + }, + { + "epoch": 0.9491493638542637, + "grad_norm": 0.16090041399002075, + "learning_rate": 1.2736411905522882e-06, + "loss": 0.0124, + "step": 40490 + }, + { + "epoch": 0.9491728054290687, + "grad_norm": 0.12223412841558456, + "learning_rate": 1.2724695536732833e-06, + "loss": 0.0133, + "step": 40491 + }, + { + "epoch": 0.9491962470038737, + "grad_norm": 0.38863450288772583, + "learning_rate": 1.2712984524926531e-06, + "loss": 0.0746, + "step": 40492 + }, + { + "epoch": 0.9492196885786787, + "grad_norm": 0.6433807611465454, + "learning_rate": 1.2701278870167255e-06, + "loss": 0.1123, + "step": 40493 + }, + { + "epoch": 0.9492431301534837, + "grad_norm": 0.41515135765075684, + "learning_rate": 1.2689578572518622e-06, + "loss": 0.0475, + "step": 40494 + }, + { + "epoch": 0.9492665717282887, + "grad_norm": 0.14333108067512512, + "learning_rate": 1.2677883632043918e-06, + "loss": 0.0251, + "step": 40495 + }, + { + "epoch": 0.9492900133030937, + "grad_norm": 0.518088161945343, + "learning_rate": 1.2666194048806978e-06, + "loss": 0.0541, + "step": 40496 + }, + { + "epoch": 0.9493134548778986, + "grad_norm": 0.14845961332321167, + "learning_rate": 1.2654509822870974e-06, + "loss": 0.0242, + "step": 40497 + }, + { + "epoch": 0.9493368964527037, + "grad_norm": 0.16197721660137177, + "learning_rate": 1.264283095429941e-06, + "loss": 0.0273, + "step": 40498 + }, + { + "epoch": 0.9493603380275087, + "grad_norm": 0.6531252861022949, + "learning_rate": 1.2631157443155572e-06, + "loss": 0.0653, + "step": 40499 + }, + { + "epoch": 0.9493837796023137, + "grad_norm": 0.21751068532466888, + "learning_rate": 1.261948928950285e-06, + "loss": 0.0297, + "step": 40500 + }, + { + "epoch": 0.9494072211771187, + "grad_norm": 0.357159286737442, + "learning_rate": 1.2607826493404417e-06, + "loss": 0.0891, + "step": 40501 + }, + { + "epoch": 0.9494306627519237, + "grad_norm": 0.37523728609085083, + "learning_rate": 1.259616905492378e-06, + "loss": 0.0838, + "step": 40502 + }, + { + "epoch": 0.9494541043267287, + "grad_norm": 0.312838077545166, + "learning_rate": 1.2584516974124105e-06, + "loss": 0.0323, + "step": 40503 + }, + { + "epoch": 0.9494775459015337, + "grad_norm": 0.3319358229637146, + "learning_rate": 1.2572870251068569e-06, + "loss": 0.0487, + "step": 40504 + }, + { + "epoch": 0.9495009874763387, + "grad_norm": 0.1351892650127411, + "learning_rate": 1.2561228885820452e-06, + "loss": 0.0121, + "step": 40505 + }, + { + "epoch": 0.9495244290511436, + "grad_norm": 0.82402104139328, + "learning_rate": 1.2549592878442816e-06, + "loss": 0.1249, + "step": 40506 + }, + { + "epoch": 0.9495478706259487, + "grad_norm": 0.34452468156814575, + "learning_rate": 1.2537962228998946e-06, + "loss": 0.0529, + "step": 40507 + }, + { + "epoch": 0.9495713122007536, + "grad_norm": 0.29337114095687866, + "learning_rate": 1.2526336937551786e-06, + "loss": 0.0288, + "step": 40508 + }, + { + "epoch": 0.9495947537755587, + "grad_norm": 0.647982656955719, + "learning_rate": 1.2514717004164512e-06, + "loss": 0.1024, + "step": 40509 + }, + { + "epoch": 0.9496181953503636, + "grad_norm": 0.4550682008266449, + "learning_rate": 1.2503102428900072e-06, + "loss": 0.5773, + "step": 40510 + }, + { + "epoch": 0.9496416369251687, + "grad_norm": 0.517572820186615, + "learning_rate": 1.2491493211821637e-06, + "loss": 0.0802, + "step": 40511 + }, + { + "epoch": 0.9496650784999736, + "grad_norm": 0.08022312074899673, + "learning_rate": 1.2479889352992158e-06, + "loss": 0.0159, + "step": 40512 + }, + { + "epoch": 0.9496885200747787, + "grad_norm": 0.5198554396629333, + "learning_rate": 1.2468290852474474e-06, + "loss": 0.0679, + "step": 40513 + }, + { + "epoch": 0.9497119616495836, + "grad_norm": 0.5986656546592712, + "learning_rate": 1.2456697710331644e-06, + "loss": 0.0781, + "step": 40514 + }, + { + "epoch": 0.9497354032243887, + "grad_norm": 0.376979261636734, + "learning_rate": 1.2445109926626508e-06, + "loss": 0.0806, + "step": 40515 + }, + { + "epoch": 0.9497588447991936, + "grad_norm": 0.43887436389923096, + "learning_rate": 1.2433527501422015e-06, + "loss": 0.0701, + "step": 40516 + }, + { + "epoch": 0.9497822863739986, + "grad_norm": 0.15001174807548523, + "learning_rate": 1.2421950434781004e-06, + "loss": 0.0253, + "step": 40517 + }, + { + "epoch": 0.9498057279488036, + "grad_norm": 0.46888864040374756, + "learning_rate": 1.2410378726766313e-06, + "loss": 0.069, + "step": 40518 + }, + { + "epoch": 0.9498291695236086, + "grad_norm": 0.29527872800827026, + "learning_rate": 1.239881237744056e-06, + "loss": 0.0423, + "step": 40519 + }, + { + "epoch": 0.9498526110984136, + "grad_norm": 0.47904908657073975, + "learning_rate": 1.2387251386866583e-06, + "loss": 0.0755, + "step": 40520 + }, + { + "epoch": 0.9498760526732186, + "grad_norm": 0.09544934332370758, + "learning_rate": 1.237569575510722e-06, + "loss": 0.0106, + "step": 40521 + }, + { + "epoch": 0.9498994942480236, + "grad_norm": 0.4289622902870178, + "learning_rate": 1.2364145482225198e-06, + "loss": 0.0771, + "step": 40522 + }, + { + "epoch": 0.9499229358228286, + "grad_norm": 0.18522138893604279, + "learning_rate": 1.2352600568282913e-06, + "loss": 0.0289, + "step": 40523 + }, + { + "epoch": 0.9499463773976335, + "grad_norm": 0.7479978203773499, + "learning_rate": 1.2341061013343313e-06, + "loss": 0.0973, + "step": 40524 + }, + { + "epoch": 0.9499698189724386, + "grad_norm": 0.6015759706497192, + "learning_rate": 1.2329526817468906e-06, + "loss": 0.0675, + "step": 40525 + }, + { + "epoch": 0.9499932605472435, + "grad_norm": 0.41735368967056274, + "learning_rate": 1.2317997980722084e-06, + "loss": 0.053, + "step": 40526 + }, + { + "epoch": 0.9500167021220486, + "grad_norm": 0.15497061610221863, + "learning_rate": 1.2306474503165687e-06, + "loss": 0.0194, + "step": 40527 + }, + { + "epoch": 0.9500401436968535, + "grad_norm": 0.5546902418136597, + "learning_rate": 1.229495638486211e-06, + "loss": 0.1096, + "step": 40528 + }, + { + "epoch": 0.9500635852716586, + "grad_norm": 0.29996219277381897, + "learning_rate": 1.2283443625873857e-06, + "loss": 0.0559, + "step": 40529 + }, + { + "epoch": 0.9500870268464635, + "grad_norm": 0.5563765168190002, + "learning_rate": 1.2271936226263546e-06, + "loss": 0.16, + "step": 40530 + }, + { + "epoch": 0.9501104684212686, + "grad_norm": 0.5065228343009949, + "learning_rate": 1.2260434186093351e-06, + "loss": 0.0832, + "step": 40531 + }, + { + "epoch": 0.9501339099960735, + "grad_norm": 0.334728479385376, + "learning_rate": 1.2248937505425885e-06, + "loss": 0.052, + "step": 40532 + }, + { + "epoch": 0.9501573515708786, + "grad_norm": 0.07519848644733429, + "learning_rate": 1.223744618432332e-06, + "loss": 0.0052, + "step": 40533 + }, + { + "epoch": 0.9501807931456835, + "grad_norm": 0.0973915383219719, + "learning_rate": 1.2225960222848277e-06, + "loss": 0.017, + "step": 40534 + }, + { + "epoch": 0.9502042347204885, + "grad_norm": 0.42515265941619873, + "learning_rate": 1.2214479621062925e-06, + "loss": 0.0897, + "step": 40535 + }, + { + "epoch": 0.9502276762952935, + "grad_norm": 0.08079776912927628, + "learning_rate": 1.2203004379029547e-06, + "loss": 0.0241, + "step": 40536 + }, + { + "epoch": 0.9502511178700985, + "grad_norm": 0.3723233640193939, + "learning_rate": 1.2191534496810542e-06, + "loss": 0.0532, + "step": 40537 + }, + { + "epoch": 0.9502745594449035, + "grad_norm": 0.4644182026386261, + "learning_rate": 1.2180069974467856e-06, + "loss": 0.1157, + "step": 40538 + }, + { + "epoch": 0.9502980010197085, + "grad_norm": 0.5186147689819336, + "learning_rate": 1.2168610812063996e-06, + "loss": 0.0776, + "step": 40539 + }, + { + "epoch": 0.9503214425945135, + "grad_norm": 0.43057408928871155, + "learning_rate": 1.2157157009661023e-06, + "loss": 0.0642, + "step": 40540 + }, + { + "epoch": 0.9503448841693185, + "grad_norm": 0.6317954659461975, + "learning_rate": 1.2145708567321113e-06, + "loss": 0.1013, + "step": 40541 + }, + { + "epoch": 0.9503683257441234, + "grad_norm": 0.4397159814834595, + "learning_rate": 1.2134265485106322e-06, + "loss": 0.0602, + "step": 40542 + }, + { + "epoch": 0.9503917673189285, + "grad_norm": 0.23698335886001587, + "learning_rate": 1.2122827763078825e-06, + "loss": 0.0315, + "step": 40543 + }, + { + "epoch": 0.9504152088937334, + "grad_norm": 0.17900361120700836, + "learning_rate": 1.2111395401300685e-06, + "loss": 0.0298, + "step": 40544 + }, + { + "epoch": 0.9504386504685385, + "grad_norm": 0.5193642377853394, + "learning_rate": 1.209996839983385e-06, + "loss": 0.7669, + "step": 40545 + }, + { + "epoch": 0.9504620920433434, + "grad_norm": 0.27862054109573364, + "learning_rate": 1.208854675874027e-06, + "loss": 0.0438, + "step": 40546 + }, + { + "epoch": 0.9504855336181485, + "grad_norm": 0.2423161268234253, + "learning_rate": 1.207713047808212e-06, + "loss": 0.0264, + "step": 40547 + }, + { + "epoch": 0.9505089751929534, + "grad_norm": 0.22817282378673553, + "learning_rate": 1.2065719557921128e-06, + "loss": 0.0239, + "step": 40548 + }, + { + "epoch": 0.9505324167677585, + "grad_norm": 0.6142289638519287, + "learning_rate": 1.2054313998319467e-06, + "loss": 0.1284, + "step": 40549 + }, + { + "epoch": 0.9505558583425635, + "grad_norm": 0.16431035101413727, + "learning_rate": 1.204291379933875e-06, + "loss": 0.0236, + "step": 40550 + }, + { + "epoch": 0.9505792999173684, + "grad_norm": 0.48559457063674927, + "learning_rate": 1.2031518961041043e-06, + "loss": 0.106, + "step": 40551 + }, + { + "epoch": 0.9506027414921735, + "grad_norm": 0.49905598163604736, + "learning_rate": 1.2020129483488074e-06, + "loss": 0.0799, + "step": 40552 + }, + { + "epoch": 0.9506261830669784, + "grad_norm": 0.5781129598617554, + "learning_rate": 1.2008745366741569e-06, + "loss": 0.0564, + "step": 40553 + }, + { + "epoch": 0.9506496246417835, + "grad_norm": 0.1199892982840538, + "learning_rate": 1.199736661086348e-06, + "loss": 0.0093, + "step": 40554 + }, + { + "epoch": 0.9506730662165884, + "grad_norm": 0.07408218085765839, + "learning_rate": 1.1985993215915536e-06, + "loss": 0.0074, + "step": 40555 + }, + { + "epoch": 0.9506965077913935, + "grad_norm": 0.45736148953437805, + "learning_rate": 1.1974625181959242e-06, + "loss": 0.0477, + "step": 40556 + }, + { + "epoch": 0.9507199493661984, + "grad_norm": 0.26003944873809814, + "learning_rate": 1.196326250905655e-06, + "loss": 0.0284, + "step": 40557 + }, + { + "epoch": 0.9507433909410035, + "grad_norm": 0.0777403861284256, + "learning_rate": 1.1951905197268854e-06, + "loss": 0.0161, + "step": 40558 + }, + { + "epoch": 0.9507668325158084, + "grad_norm": 0.4897334575653076, + "learning_rate": 1.194055324665788e-06, + "loss": 0.0536, + "step": 40559 + }, + { + "epoch": 0.9507902740906135, + "grad_norm": 0.7553247809410095, + "learning_rate": 1.1929206657285364e-06, + "loss": 0.143, + "step": 40560 + }, + { + "epoch": 0.9508137156654184, + "grad_norm": 0.12657774984836578, + "learning_rate": 1.1917865429212693e-06, + "loss": 0.0076, + "step": 40561 + }, + { + "epoch": 0.9508371572402234, + "grad_norm": 0.4399222731590271, + "learning_rate": 1.1906529562501378e-06, + "loss": 0.0746, + "step": 40562 + }, + { + "epoch": 0.9508605988150284, + "grad_norm": 0.2777739465236664, + "learning_rate": 1.1895199057213147e-06, + "loss": 0.0336, + "step": 40563 + }, + { + "epoch": 0.9508840403898334, + "grad_norm": 0.7411735653877258, + "learning_rate": 1.1883873913409393e-06, + "loss": 0.0921, + "step": 40564 + }, + { + "epoch": 0.9509074819646384, + "grad_norm": 0.47182586789131165, + "learning_rate": 1.1872554131151291e-06, + "loss": 0.0848, + "step": 40565 + }, + { + "epoch": 0.9509309235394434, + "grad_norm": 0.1235254630446434, + "learning_rate": 1.1861239710500683e-06, + "loss": 0.0126, + "step": 40566 + }, + { + "epoch": 0.9509543651142484, + "grad_norm": 0.3868570625782013, + "learning_rate": 1.1849930651518737e-06, + "loss": 0.0461, + "step": 40567 + }, + { + "epoch": 0.9509778066890534, + "grad_norm": 0.51207435131073, + "learning_rate": 1.1838626954266851e-06, + "loss": 0.1272, + "step": 40568 + }, + { + "epoch": 0.9510012482638583, + "grad_norm": 0.6240926384925842, + "learning_rate": 1.1827328618806422e-06, + "loss": 0.0782, + "step": 40569 + }, + { + "epoch": 0.9510246898386634, + "grad_norm": 0.1165241152048111, + "learning_rate": 1.181603564519862e-06, + "loss": 0.0196, + "step": 40570 + }, + { + "epoch": 0.9510481314134683, + "grad_norm": 0.11606574058532715, + "learning_rate": 1.1804748033504732e-06, + "loss": 0.0171, + "step": 40571 + }, + { + "epoch": 0.9510715729882734, + "grad_norm": 0.2249659150838852, + "learning_rate": 1.1793465783786151e-06, + "loss": 0.012, + "step": 40572 + }, + { + "epoch": 0.9510950145630783, + "grad_norm": 0.1068229228258133, + "learning_rate": 1.1782188896104052e-06, + "loss": 0.0165, + "step": 40573 + }, + { + "epoch": 0.9511184561378834, + "grad_norm": 0.3175791800022125, + "learning_rate": 1.1770917370519497e-06, + "loss": 0.0629, + "step": 40574 + }, + { + "epoch": 0.9511418977126883, + "grad_norm": 0.13587592542171478, + "learning_rate": 1.175965120709377e-06, + "loss": 0.0133, + "step": 40575 + }, + { + "epoch": 0.9511653392874934, + "grad_norm": 0.1468309760093689, + "learning_rate": 1.1748390405887933e-06, + "loss": 0.0084, + "step": 40576 + }, + { + "epoch": 0.9511887808622983, + "grad_norm": 0.776995062828064, + "learning_rate": 1.1737134966963158e-06, + "loss": 0.1485, + "step": 40577 + }, + { + "epoch": 0.9512122224371033, + "grad_norm": 0.1106356605887413, + "learning_rate": 1.1725884890380401e-06, + "loss": 0.0096, + "step": 40578 + }, + { + "epoch": 0.9512356640119083, + "grad_norm": 0.4183795750141144, + "learning_rate": 1.1714640176200831e-06, + "loss": 0.0814, + "step": 40579 + }, + { + "epoch": 0.9512591055867133, + "grad_norm": 0.2837531864643097, + "learning_rate": 1.17034008244854e-06, + "loss": 0.0389, + "step": 40580 + }, + { + "epoch": 0.9512825471615183, + "grad_norm": 0.2546326220035553, + "learning_rate": 1.1692166835295172e-06, + "loss": 0.021, + "step": 40581 + }, + { + "epoch": 0.9513059887363233, + "grad_norm": 0.2928239405155182, + "learning_rate": 1.1680938208690983e-06, + "loss": 0.0543, + "step": 40582 + }, + { + "epoch": 0.9513294303111283, + "grad_norm": 0.13677459955215454, + "learning_rate": 1.1669714944733901e-06, + "loss": 0.0128, + "step": 40583 + }, + { + "epoch": 0.9513528718859333, + "grad_norm": 0.11779443174600601, + "learning_rate": 1.165849704348454e-06, + "loss": 0.019, + "step": 40584 + }, + { + "epoch": 0.9513763134607383, + "grad_norm": 0.18221895396709442, + "learning_rate": 1.1647284505004075e-06, + "loss": 0.0199, + "step": 40585 + }, + { + "epoch": 0.9513997550355433, + "grad_norm": 0.23822325468063354, + "learning_rate": 1.1636077329353235e-06, + "loss": 0.0262, + "step": 40586 + }, + { + "epoch": 0.9514231966103482, + "grad_norm": 0.1369512528181076, + "learning_rate": 1.162487551659275e-06, + "loss": 0.0135, + "step": 40587 + }, + { + "epoch": 0.9514466381851533, + "grad_norm": 0.5899613499641418, + "learning_rate": 1.1613679066783568e-06, + "loss": 0.1074, + "step": 40588 + }, + { + "epoch": 0.9514700797599582, + "grad_norm": 0.5008262991905212, + "learning_rate": 1.1602487979986421e-06, + "loss": 0.0956, + "step": 40589 + }, + { + "epoch": 0.9514935213347633, + "grad_norm": 0.3732127845287323, + "learning_rate": 1.1591302256261815e-06, + "loss": 0.0628, + "step": 40590 + }, + { + "epoch": 0.9515169629095682, + "grad_norm": 0.7472918629646301, + "learning_rate": 1.1580121895670704e-06, + "loss": 0.1039, + "step": 40591 + }, + { + "epoch": 0.9515404044843733, + "grad_norm": 0.1350259780883789, + "learning_rate": 1.1568946898273591e-06, + "loss": 0.0203, + "step": 40592 + }, + { + "epoch": 0.9515638460591782, + "grad_norm": 0.46222543716430664, + "learning_rate": 1.155777726413121e-06, + "loss": 0.0258, + "step": 40593 + }, + { + "epoch": 0.9515872876339833, + "grad_norm": 0.2755447030067444, + "learning_rate": 1.1546612993304062e-06, + "loss": 0.0434, + "step": 40594 + }, + { + "epoch": 0.9516107292087882, + "grad_norm": 0.23781031370162964, + "learning_rate": 1.1535454085852881e-06, + "loss": 0.0451, + "step": 40595 + }, + { + "epoch": 0.9516341707835932, + "grad_norm": 0.2050762176513672, + "learning_rate": 1.152430054183795e-06, + "loss": 0.0162, + "step": 40596 + }, + { + "epoch": 0.9516576123583982, + "grad_norm": 0.6526134014129639, + "learning_rate": 1.1513152361320112e-06, + "loss": 0.1311, + "step": 40597 + }, + { + "epoch": 0.9516810539332032, + "grad_norm": 0.6435054540634155, + "learning_rate": 1.1502009544359648e-06, + "loss": 0.6179, + "step": 40598 + }, + { + "epoch": 0.9517044955080082, + "grad_norm": 0.48434698581695557, + "learning_rate": 1.1490872091016958e-06, + "loss": 0.0686, + "step": 40599 + }, + { + "epoch": 0.9517279370828132, + "grad_norm": 0.627418041229248, + "learning_rate": 1.1479740001352769e-06, + "loss": 0.0741, + "step": 40600 + }, + { + "epoch": 0.9517513786576182, + "grad_norm": 0.6753853559494019, + "learning_rate": 1.1468613275427253e-06, + "loss": 0.5852, + "step": 40601 + }, + { + "epoch": 0.9517748202324232, + "grad_norm": 0.40815243124961853, + "learning_rate": 1.145749191330081e-06, + "loss": 0.0457, + "step": 40602 + }, + { + "epoch": 0.9517982618072283, + "grad_norm": 0.5129702687263489, + "learning_rate": 1.1446375915033835e-06, + "loss": 0.0871, + "step": 40603 + }, + { + "epoch": 0.9518217033820332, + "grad_norm": 0.5815621018409729, + "learning_rate": 1.1435265280686614e-06, + "loss": 0.0856, + "step": 40604 + }, + { + "epoch": 0.9518451449568383, + "grad_norm": 0.15230831503868103, + "learning_rate": 1.1424160010319429e-06, + "loss": 0.0157, + "step": 40605 + }, + { + "epoch": 0.9518685865316432, + "grad_norm": 0.5459635257720947, + "learning_rate": 1.1413060103992568e-06, + "loss": 0.0858, + "step": 40606 + }, + { + "epoch": 0.9518920281064482, + "grad_norm": 0.4343830645084381, + "learning_rate": 1.1401965561766315e-06, + "loss": 0.1086, + "step": 40607 + }, + { + "epoch": 0.9519154696812532, + "grad_norm": 0.381734699010849, + "learning_rate": 1.1390876383700732e-06, + "loss": 0.0489, + "step": 40608 + }, + { + "epoch": 0.9519389112560582, + "grad_norm": 0.4337579011917114, + "learning_rate": 1.1379792569855886e-06, + "loss": 0.0599, + "step": 40609 + }, + { + "epoch": 0.9519623528308632, + "grad_norm": 0.45846983790397644, + "learning_rate": 1.1368714120292279e-06, + "loss": 0.1039, + "step": 40610 + }, + { + "epoch": 0.9519857944056682, + "grad_norm": 0.4108163118362427, + "learning_rate": 1.1357641035069756e-06, + "loss": 0.0873, + "step": 40611 + }, + { + "epoch": 0.9520092359804732, + "grad_norm": 0.1804448664188385, + "learning_rate": 1.1346573314248487e-06, + "loss": 0.0355, + "step": 40612 + }, + { + "epoch": 0.9520326775552782, + "grad_norm": 0.33327558636665344, + "learning_rate": 1.133551095788854e-06, + "loss": 0.0366, + "step": 40613 + }, + { + "epoch": 0.9520561191300831, + "grad_norm": 0.2053690105676651, + "learning_rate": 1.1324453966049864e-06, + "loss": 0.0341, + "step": 40614 + }, + { + "epoch": 0.9520795607048882, + "grad_norm": 0.10153365880250931, + "learning_rate": 1.1313402338792522e-06, + "loss": 0.011, + "step": 40615 + }, + { + "epoch": 0.9521030022796931, + "grad_norm": 0.14816558361053467, + "learning_rate": 1.1302356076176358e-06, + "loss": 0.0257, + "step": 40616 + }, + { + "epoch": 0.9521264438544982, + "grad_norm": 0.5519100427627563, + "learning_rate": 1.1291315178261542e-06, + "loss": 0.0593, + "step": 40617 + }, + { + "epoch": 0.9521498854293031, + "grad_norm": 0.8509051203727722, + "learning_rate": 1.1280279645107806e-06, + "loss": 0.1002, + "step": 40618 + }, + { + "epoch": 0.9521733270041082, + "grad_norm": 0.5134045481681824, + "learning_rate": 1.12692494767751e-06, + "loss": 0.597, + "step": 40619 + }, + { + "epoch": 0.9521967685789131, + "grad_norm": 0.5068297386169434, + "learning_rate": 1.125822467332327e-06, + "loss": 0.0777, + "step": 40620 + }, + { + "epoch": 0.9522202101537182, + "grad_norm": 0.4707913100719452, + "learning_rate": 1.124720523481204e-06, + "loss": 0.5417, + "step": 40621 + }, + { + "epoch": 0.9522436517285231, + "grad_norm": 0.3222946226596832, + "learning_rate": 1.1236191161301257e-06, + "loss": 0.0583, + "step": 40622 + }, + { + "epoch": 0.9522670933033281, + "grad_norm": 0.48507457971572876, + "learning_rate": 1.1225182452850758e-06, + "loss": 0.0905, + "step": 40623 + }, + { + "epoch": 0.9522905348781331, + "grad_norm": 0.14647367596626282, + "learning_rate": 1.1214179109520163e-06, + "loss": 0.0208, + "step": 40624 + }, + { + "epoch": 0.9523139764529381, + "grad_norm": 0.2559357285499573, + "learning_rate": 1.1203181131369312e-06, + "loss": 0.0269, + "step": 40625 + }, + { + "epoch": 0.9523374180277431, + "grad_norm": 0.3384733498096466, + "learning_rate": 1.1192188518457824e-06, + "loss": 0.3852, + "step": 40626 + }, + { + "epoch": 0.9523608596025481, + "grad_norm": 0.16572055220603943, + "learning_rate": 1.118120127084532e-06, + "loss": 0.0207, + "step": 40627 + }, + { + "epoch": 0.9523843011773531, + "grad_norm": 0.6887216567993164, + "learning_rate": 1.117021938859142e-06, + "loss": 0.3455, + "step": 40628 + }, + { + "epoch": 0.9524077427521581, + "grad_norm": 0.1757853627204895, + "learning_rate": 1.1159242871755515e-06, + "loss": 0.0112, + "step": 40629 + }, + { + "epoch": 0.952431184326963, + "grad_norm": 0.44505125284194946, + "learning_rate": 1.1148271720397563e-06, + "loss": 0.0565, + "step": 40630 + }, + { + "epoch": 0.9524546259017681, + "grad_norm": 0.39419499039649963, + "learning_rate": 1.1137305934576847e-06, + "loss": 0.0756, + "step": 40631 + }, + { + "epoch": 0.952478067476573, + "grad_norm": 0.2320052683353424, + "learning_rate": 1.1126345514352987e-06, + "loss": 0.0293, + "step": 40632 + }, + { + "epoch": 0.9525015090513781, + "grad_norm": 0.15697313845157623, + "learning_rate": 1.111539045978527e-06, + "loss": 0.0327, + "step": 40633 + }, + { + "epoch": 0.952524950626183, + "grad_norm": 0.22477617859840393, + "learning_rate": 1.110444077093331e-06, + "loss": 0.0233, + "step": 40634 + }, + { + "epoch": 0.9525483922009881, + "grad_norm": 0.6737518906593323, + "learning_rate": 1.1093496447856288e-06, + "loss": 0.1298, + "step": 40635 + }, + { + "epoch": 0.952571833775793, + "grad_norm": 0.32686522603034973, + "learning_rate": 1.108255749061393e-06, + "loss": 0.0575, + "step": 40636 + }, + { + "epoch": 0.9525952753505981, + "grad_norm": 0.8067222237586975, + "learning_rate": 1.1071623899265416e-06, + "loss": 0.17, + "step": 40637 + }, + { + "epoch": 0.952618716925403, + "grad_norm": 0.5714096426963806, + "learning_rate": 1.1060695673870137e-06, + "loss": 0.0903, + "step": 40638 + }, + { + "epoch": 0.9526421585002081, + "grad_norm": 0.35324251651763916, + "learning_rate": 1.104977281448727e-06, + "loss": 0.0469, + "step": 40639 + }, + { + "epoch": 0.952665600075013, + "grad_norm": 0.49609097838401794, + "learning_rate": 1.10388553211761e-06, + "loss": 0.0932, + "step": 40640 + }, + { + "epoch": 0.952689041649818, + "grad_norm": 0.46286505460739136, + "learning_rate": 1.1027943193995915e-06, + "loss": 0.0741, + "step": 40641 + }, + { + "epoch": 0.952712483224623, + "grad_norm": 0.11655271798372269, + "learning_rate": 1.1017036433005889e-06, + "loss": 0.0178, + "step": 40642 + }, + { + "epoch": 0.952735924799428, + "grad_norm": 0.40961122512817383, + "learning_rate": 1.1006135038265308e-06, + "loss": 0.1003, + "step": 40643 + }, + { + "epoch": 0.952759366374233, + "grad_norm": 0.4998547434806824, + "learning_rate": 1.0995239009833236e-06, + "loss": 0.0797, + "step": 40644 + }, + { + "epoch": 0.952782807949038, + "grad_norm": 0.4408716857433319, + "learning_rate": 1.0984348347768847e-06, + "loss": 0.0762, + "step": 40645 + }, + { + "epoch": 0.952806249523843, + "grad_norm": 0.4155134856700897, + "learning_rate": 1.0973463052131095e-06, + "loss": 0.0708, + "step": 40646 + }, + { + "epoch": 0.952829691098648, + "grad_norm": 0.5013408660888672, + "learning_rate": 1.0962583122979152e-06, + "loss": 0.0539, + "step": 40647 + }, + { + "epoch": 0.952853132673453, + "grad_norm": 0.363899290561676, + "learning_rate": 1.0951708560371975e-06, + "loss": 0.0847, + "step": 40648 + }, + { + "epoch": 0.952876574248258, + "grad_norm": 0.31284573674201965, + "learning_rate": 1.0940839364368738e-06, + "loss": 0.0504, + "step": 40649 + }, + { + "epoch": 0.9529000158230629, + "grad_norm": 0.46424227952957153, + "learning_rate": 1.092997553502839e-06, + "loss": 0.0413, + "step": 40650 + }, + { + "epoch": 0.952923457397868, + "grad_norm": 0.40357300639152527, + "learning_rate": 1.0919117072409668e-06, + "loss": 0.0771, + "step": 40651 + }, + { + "epoch": 0.9529468989726729, + "grad_norm": 0.3036089241504669, + "learning_rate": 1.0908263976571743e-06, + "loss": 0.0516, + "step": 40652 + }, + { + "epoch": 0.952970340547478, + "grad_norm": 0.42619970440864563, + "learning_rate": 1.0897416247573344e-06, + "loss": 0.0778, + "step": 40653 + }, + { + "epoch": 0.952993782122283, + "grad_norm": 0.6792253255844116, + "learning_rate": 1.0886573885473316e-06, + "loss": 0.0705, + "step": 40654 + }, + { + "epoch": 0.953017223697088, + "grad_norm": 0.7277395129203796, + "learning_rate": 1.0875736890330613e-06, + "loss": 0.0719, + "step": 40655 + }, + { + "epoch": 0.953040665271893, + "grad_norm": 0.0770622119307518, + "learning_rate": 1.086490526220396e-06, + "loss": 0.0162, + "step": 40656 + }, + { + "epoch": 0.953064106846698, + "grad_norm": 0.3650130331516266, + "learning_rate": 1.0854079001152095e-06, + "loss": 0.0626, + "step": 40657 + }, + { + "epoch": 0.953087548421503, + "grad_norm": 0.13025441765785217, + "learning_rate": 1.0843258107233856e-06, + "loss": 0.0162, + "step": 40658 + }, + { + "epoch": 0.953110989996308, + "grad_norm": 0.8171401023864746, + "learning_rate": 1.0832442580507973e-06, + "loss": 0.1243, + "step": 40659 + }, + { + "epoch": 0.953134431571113, + "grad_norm": 0.2987273931503296, + "learning_rate": 1.0821632421032957e-06, + "loss": 0.0697, + "step": 40660 + }, + { + "epoch": 0.9531578731459179, + "grad_norm": 0.36786627769470215, + "learning_rate": 1.0810827628867538e-06, + "loss": 0.0301, + "step": 40661 + }, + { + "epoch": 0.953181314720723, + "grad_norm": 0.2890348732471466, + "learning_rate": 1.0800028204070555e-06, + "loss": 0.0272, + "step": 40662 + }, + { + "epoch": 0.9532047562955279, + "grad_norm": 0.17214536666870117, + "learning_rate": 1.07892341467003e-06, + "loss": 0.0376, + "step": 40663 + }, + { + "epoch": 0.953228197870333, + "grad_norm": 0.5685164928436279, + "learning_rate": 1.07784454568155e-06, + "loss": 0.0559, + "step": 40664 + }, + { + "epoch": 0.9532516394451379, + "grad_norm": 0.9157774448394775, + "learning_rate": 1.0767662134474777e-06, + "loss": 0.4346, + "step": 40665 + }, + { + "epoch": 0.953275081019943, + "grad_norm": 0.33649682998657227, + "learning_rate": 1.0756884179736304e-06, + "loss": 0.0735, + "step": 40666 + }, + { + "epoch": 0.9532985225947479, + "grad_norm": 0.0991382896900177, + "learning_rate": 1.0746111592659035e-06, + "loss": 0.0158, + "step": 40667 + }, + { + "epoch": 0.953321964169553, + "grad_norm": 0.5970698595046997, + "learning_rate": 1.0735344373301036e-06, + "loss": 0.4501, + "step": 40668 + }, + { + "epoch": 0.9533454057443579, + "grad_norm": 0.5933292508125305, + "learning_rate": 1.0724582521720928e-06, + "loss": 0.0796, + "step": 40669 + }, + { + "epoch": 0.9533688473191629, + "grad_norm": 0.17890594899654388, + "learning_rate": 1.0713826037977103e-06, + "loss": 0.023, + "step": 40670 + }, + { + "epoch": 0.9533922888939679, + "grad_norm": 0.5673342943191528, + "learning_rate": 1.0703074922127853e-06, + "loss": 0.0978, + "step": 40671 + }, + { + "epoch": 0.9534157304687729, + "grad_norm": 0.13247142732143402, + "learning_rate": 1.0692329174231352e-06, + "loss": 0.0248, + "step": 40672 + }, + { + "epoch": 0.9534391720435779, + "grad_norm": 0.49696359038352966, + "learning_rate": 1.068158879434622e-06, + "loss": 0.0731, + "step": 40673 + }, + { + "epoch": 0.9534626136183829, + "grad_norm": 0.5132684707641602, + "learning_rate": 1.0670853782530632e-06, + "loss": 0.0834, + "step": 40674 + }, + { + "epoch": 0.9534860551931879, + "grad_norm": 0.728572428226471, + "learning_rate": 1.0660124138842764e-06, + "loss": 0.3834, + "step": 40675 + }, + { + "epoch": 0.9535094967679929, + "grad_norm": 0.5582441091537476, + "learning_rate": 1.0649399863340903e-06, + "loss": 0.1237, + "step": 40676 + }, + { + "epoch": 0.9535329383427978, + "grad_norm": 0.4892941415309906, + "learning_rate": 1.0638680956083336e-06, + "loss": 0.0908, + "step": 40677 + }, + { + "epoch": 0.9535563799176029, + "grad_norm": 0.21690231561660767, + "learning_rate": 1.0627967417127905e-06, + "loss": 0.0354, + "step": 40678 + }, + { + "epoch": 0.9535798214924078, + "grad_norm": 0.808734655380249, + "learning_rate": 1.0617259246533006e-06, + "loss": 0.1014, + "step": 40679 + }, + { + "epoch": 0.9536032630672129, + "grad_norm": 0.3185117840766907, + "learning_rate": 1.0606556444356709e-06, + "loss": 0.0375, + "step": 40680 + }, + { + "epoch": 0.9536267046420178, + "grad_norm": 0.0967397689819336, + "learning_rate": 1.0595859010657073e-06, + "loss": 0.0133, + "step": 40681 + }, + { + "epoch": 0.9536501462168229, + "grad_norm": 0.9091845154762268, + "learning_rate": 1.0585166945492165e-06, + "loss": 0.1302, + "step": 40682 + }, + { + "epoch": 0.9536735877916278, + "grad_norm": 0.08177853375673294, + "learning_rate": 1.0574480248919826e-06, + "loss": 0.0072, + "step": 40683 + }, + { + "epoch": 0.9536970293664329, + "grad_norm": 0.16447897255420685, + "learning_rate": 1.0563798920998235e-06, + "loss": 0.0149, + "step": 40684 + }, + { + "epoch": 0.9537204709412378, + "grad_norm": 0.18905119597911835, + "learning_rate": 1.0553122961785344e-06, + "loss": 0.0181, + "step": 40685 + }, + { + "epoch": 0.9537439125160428, + "grad_norm": 0.3246152400970459, + "learning_rate": 1.0542452371338885e-06, + "loss": 0.0668, + "step": 40686 + }, + { + "epoch": 0.9537673540908478, + "grad_norm": 0.38012322783470154, + "learning_rate": 1.053178714971703e-06, + "loss": 0.0664, + "step": 40687 + }, + { + "epoch": 0.9537907956656528, + "grad_norm": 0.36868399381637573, + "learning_rate": 1.0521127296977518e-06, + "loss": 0.0641, + "step": 40688 + }, + { + "epoch": 0.9538142372404578, + "grad_norm": 0.35658010840415955, + "learning_rate": 1.0510472813178185e-06, + "loss": 0.0358, + "step": 40689 + }, + { + "epoch": 0.9538376788152628, + "grad_norm": 0.5368506908416748, + "learning_rate": 1.0499823698376876e-06, + "loss": 0.5225, + "step": 40690 + }, + { + "epoch": 0.9538611203900678, + "grad_norm": 0.26989394426345825, + "learning_rate": 1.0489179952631212e-06, + "loss": 0.0443, + "step": 40691 + }, + { + "epoch": 0.9538845619648728, + "grad_norm": 0.4676671624183655, + "learning_rate": 1.0478541575999145e-06, + "loss": 0.1338, + "step": 40692 + }, + { + "epoch": 0.9539080035396778, + "grad_norm": 0.32858771085739136, + "learning_rate": 1.046790856853841e-06, + "loss": 0.3956, + "step": 40693 + }, + { + "epoch": 0.9539314451144828, + "grad_norm": 0.6049622297286987, + "learning_rate": 1.0457280930306513e-06, + "loss": 0.1148, + "step": 40694 + }, + { + "epoch": 0.9539548866892877, + "grad_norm": 0.6470912098884583, + "learning_rate": 1.0446658661361408e-06, + "loss": 0.1171, + "step": 40695 + }, + { + "epoch": 0.9539783282640928, + "grad_norm": 0.5241056084632874, + "learning_rate": 1.0436041761760384e-06, + "loss": 0.1025, + "step": 40696 + }, + { + "epoch": 0.9540017698388977, + "grad_norm": 0.48743537068367004, + "learning_rate": 1.0425430231561284e-06, + "loss": 0.0964, + "step": 40697 + }, + { + "epoch": 0.9540252114137028, + "grad_norm": 0.11465509235858917, + "learning_rate": 1.0414824070821617e-06, + "loss": 0.0215, + "step": 40698 + }, + { + "epoch": 0.9540486529885077, + "grad_norm": 0.2209225445985794, + "learning_rate": 1.040422327959889e-06, + "loss": 0.0394, + "step": 40699 + }, + { + "epoch": 0.9540720945633128, + "grad_norm": 0.637069821357727, + "learning_rate": 1.0393627857950727e-06, + "loss": 0.0743, + "step": 40700 + }, + { + "epoch": 0.9540955361381177, + "grad_norm": 0.7556689381599426, + "learning_rate": 1.0383037805934525e-06, + "loss": 0.1347, + "step": 40701 + }, + { + "epoch": 0.9541189777129228, + "grad_norm": 0.2218518853187561, + "learning_rate": 1.0372453123607905e-06, + "loss": 0.061, + "step": 40702 + }, + { + "epoch": 0.9541424192877277, + "grad_norm": 0.3934742510318756, + "learning_rate": 1.0361873811028044e-06, + "loss": 0.0458, + "step": 40703 + }, + { + "epoch": 0.9541658608625327, + "grad_norm": 0.567861020565033, + "learning_rate": 1.0351299868252561e-06, + "loss": 0.0886, + "step": 40704 + }, + { + "epoch": 0.9541893024373378, + "grad_norm": 0.1946106255054474, + "learning_rate": 1.0340731295338635e-06, + "loss": 0.0296, + "step": 40705 + }, + { + "epoch": 0.9542127440121427, + "grad_norm": 0.39017245173454285, + "learning_rate": 1.0330168092343884e-06, + "loss": 0.0559, + "step": 40706 + }, + { + "epoch": 0.9542361855869478, + "grad_norm": 0.2777080833911896, + "learning_rate": 1.0319610259325373e-06, + "loss": 0.0485, + "step": 40707 + }, + { + "epoch": 0.9542596271617527, + "grad_norm": 1.073752760887146, + "learning_rate": 1.0309057796340504e-06, + "loss": 0.0538, + "step": 40708 + }, + { + "epoch": 0.9542830687365578, + "grad_norm": 0.370564341545105, + "learning_rate": 1.0298510703446563e-06, + "loss": 0.102, + "step": 40709 + }, + { + "epoch": 0.9543065103113627, + "grad_norm": 0.446886271238327, + "learning_rate": 1.0287968980700725e-06, + "loss": 0.0661, + "step": 40710 + }, + { + "epoch": 0.9543299518861678, + "grad_norm": 0.35190412402153015, + "learning_rate": 1.0277432628160054e-06, + "loss": 0.0491, + "step": 40711 + }, + { + "epoch": 0.9543533934609727, + "grad_norm": 0.12889862060546875, + "learning_rate": 1.0266901645881955e-06, + "loss": 0.0166, + "step": 40712 + }, + { + "epoch": 0.9543768350357777, + "grad_norm": 0.14228878915309906, + "learning_rate": 1.0256376033923487e-06, + "loss": 0.0192, + "step": 40713 + }, + { + "epoch": 0.9544002766105827, + "grad_norm": 0.2658807039260864, + "learning_rate": 1.024585579234183e-06, + "loss": 0.0183, + "step": 40714 + }, + { + "epoch": 0.9544237181853877, + "grad_norm": 0.33874592185020447, + "learning_rate": 1.0235340921193936e-06, + "loss": 0.0652, + "step": 40715 + }, + { + "epoch": 0.9544471597601927, + "grad_norm": 0.26321396231651306, + "learning_rate": 1.0224831420536873e-06, + "loss": 0.02, + "step": 40716 + }, + { + "epoch": 0.9544706013349977, + "grad_norm": 0.28424569964408875, + "learning_rate": 1.0214327290427706e-06, + "loss": 0.0299, + "step": 40717 + }, + { + "epoch": 0.9544940429098027, + "grad_norm": 0.6019930839538574, + "learning_rate": 1.0203828530923386e-06, + "loss": 0.0779, + "step": 40718 + }, + { + "epoch": 0.9545174844846077, + "grad_norm": 0.4613334834575653, + "learning_rate": 1.0193335142081096e-06, + "loss": 0.0786, + "step": 40719 + }, + { + "epoch": 0.9545409260594127, + "grad_norm": 0.15142391622066498, + "learning_rate": 1.0182847123957339e-06, + "loss": 0.0157, + "step": 40720 + }, + { + "epoch": 0.9545643676342177, + "grad_norm": 0.5485819578170776, + "learning_rate": 1.017236447660952e-06, + "loss": 0.1301, + "step": 40721 + }, + { + "epoch": 0.9545878092090226, + "grad_norm": 0.4442622661590576, + "learning_rate": 1.0161887200094144e-06, + "loss": 0.0569, + "step": 40722 + }, + { + "epoch": 0.9546112507838277, + "grad_norm": 0.40324297547340393, + "learning_rate": 1.015141529446828e-06, + "loss": 0.0376, + "step": 40723 + }, + { + "epoch": 0.9546346923586326, + "grad_norm": 0.39680951833724976, + "learning_rate": 1.014094875978866e-06, + "loss": 0.0473, + "step": 40724 + }, + { + "epoch": 0.9546581339334377, + "grad_norm": 0.13698415458202362, + "learning_rate": 1.0130487596112014e-06, + "loss": 0.0121, + "step": 40725 + }, + { + "epoch": 0.9546815755082426, + "grad_norm": 0.3027750253677368, + "learning_rate": 1.01200318034953e-06, + "loss": 0.0726, + "step": 40726 + }, + { + "epoch": 0.9547050170830477, + "grad_norm": 0.26347076892852783, + "learning_rate": 1.0109581381995026e-06, + "loss": 0.0578, + "step": 40727 + }, + { + "epoch": 0.9547284586578526, + "grad_norm": 0.6270579099655151, + "learning_rate": 1.0099136331668035e-06, + "loss": 0.122, + "step": 40728 + }, + { + "epoch": 0.9547519002326577, + "grad_norm": 0.10181014984846115, + "learning_rate": 1.008869665257095e-06, + "loss": 0.013, + "step": 40729 + }, + { + "epoch": 0.9547753418074626, + "grad_norm": 0.4173474609851837, + "learning_rate": 1.0078262344760393e-06, + "loss": 0.083, + "step": 40730 + }, + { + "epoch": 0.9547987833822676, + "grad_norm": 0.5202323794364929, + "learning_rate": 1.0067833408293092e-06, + "loss": 0.1176, + "step": 40731 + }, + { + "epoch": 0.9548222249570726, + "grad_norm": 0.11150200664997101, + "learning_rate": 1.0057409843225563e-06, + "loss": 0.0129, + "step": 40732 + }, + { + "epoch": 0.9548456665318776, + "grad_norm": 0.18703119456768036, + "learning_rate": 1.0046991649614201e-06, + "loss": 0.029, + "step": 40733 + }, + { + "epoch": 0.9548691081066826, + "grad_norm": 0.2776292860507965, + "learning_rate": 1.0036578827515852e-06, + "loss": 0.0342, + "step": 40734 + }, + { + "epoch": 0.9548925496814876, + "grad_norm": 0.24727728962898254, + "learning_rate": 1.0026171376986914e-06, + "loss": 0.0343, + "step": 40735 + }, + { + "epoch": 0.9549159912562926, + "grad_norm": 0.25773870944976807, + "learning_rate": 1.0015769298083677e-06, + "loss": 0.0375, + "step": 40736 + }, + { + "epoch": 0.9549394328310976, + "grad_norm": 0.20482632517814636, + "learning_rate": 1.000537259086276e-06, + "loss": 0.0128, + "step": 40737 + }, + { + "epoch": 0.9549628744059026, + "grad_norm": 0.4674157202243805, + "learning_rate": 9.994981255380565e-07, + "loss": 0.0664, + "step": 40738 + }, + { + "epoch": 0.9549863159807076, + "grad_norm": 1.1677325963974, + "learning_rate": 9.984595291693377e-07, + "loss": 0.1281, + "step": 40739 + }, + { + "epoch": 0.9550097575555125, + "grad_norm": 0.42541685700416565, + "learning_rate": 9.974214699857598e-07, + "loss": 0.0456, + "step": 40740 + }, + { + "epoch": 0.9550331991303176, + "grad_norm": 0.5886870622634888, + "learning_rate": 9.963839479929625e-07, + "loss": 0.0862, + "step": 40741 + }, + { + "epoch": 0.9550566407051225, + "grad_norm": 0.14390213787555695, + "learning_rate": 9.95346963196564e-07, + "loss": 0.0303, + "step": 40742 + }, + { + "epoch": 0.9550800822799276, + "grad_norm": 0.4612956941127777, + "learning_rate": 9.943105156022037e-07, + "loss": 0.0708, + "step": 40743 + }, + { + "epoch": 0.9551035238547325, + "grad_norm": 0.3655621111392975, + "learning_rate": 9.932746052154996e-07, + "loss": 0.0743, + "step": 40744 + }, + { + "epoch": 0.9551269654295376, + "grad_norm": 0.2457963526248932, + "learning_rate": 9.922392320420582e-07, + "loss": 0.0147, + "step": 40745 + }, + { + "epoch": 0.9551504070043425, + "grad_norm": 0.5814184546470642, + "learning_rate": 9.912043960875195e-07, + "loss": 0.1212, + "step": 40746 + }, + { + "epoch": 0.9551738485791476, + "grad_norm": 0.17876912653446198, + "learning_rate": 9.901700973574902e-07, + "loss": 0.032, + "step": 40747 + }, + { + "epoch": 0.9551972901539525, + "grad_norm": 0.184523805975914, + "learning_rate": 9.891363358575768e-07, + "loss": 0.0171, + "step": 40748 + }, + { + "epoch": 0.9552207317287575, + "grad_norm": 0.024009443819522858, + "learning_rate": 9.88103111593397e-07, + "loss": 0.0018, + "step": 40749 + }, + { + "epoch": 0.9552441733035625, + "grad_norm": 0.16590386629104614, + "learning_rate": 9.870704245705575e-07, + "loss": 0.0204, + "step": 40750 + }, + { + "epoch": 0.9552676148783675, + "grad_norm": 0.12050742655992508, + "learning_rate": 9.86038274794654e-07, + "loss": 0.0206, + "step": 40751 + }, + { + "epoch": 0.9552910564531725, + "grad_norm": 0.23992259800434113, + "learning_rate": 9.850066622712927e-07, + "loss": 0.0463, + "step": 40752 + }, + { + "epoch": 0.9553144980279775, + "grad_norm": 0.275849848985672, + "learning_rate": 9.839755870060697e-07, + "loss": 0.04, + "step": 40753 + }, + { + "epoch": 0.9553379396027825, + "grad_norm": 0.4962916970252991, + "learning_rate": 9.829450490045799e-07, + "loss": 0.5647, + "step": 40754 + }, + { + "epoch": 0.9553613811775875, + "grad_norm": 0.2783516049385071, + "learning_rate": 9.819150482724082e-07, + "loss": 0.0467, + "step": 40755 + }, + { + "epoch": 0.9553848227523926, + "grad_norm": 0.5607463121414185, + "learning_rate": 9.8088558481515e-07, + "loss": 0.0744, + "step": 40756 + }, + { + "epoch": 0.9554082643271975, + "grad_norm": 0.4291212558746338, + "learning_rate": 9.798566586383895e-07, + "loss": 0.0721, + "step": 40757 + }, + { + "epoch": 0.9554317059020025, + "grad_norm": 0.4623945653438568, + "learning_rate": 9.788282697477224e-07, + "loss": 0.0596, + "step": 40758 + }, + { + "epoch": 0.9554551474768075, + "grad_norm": 0.28019067645072937, + "learning_rate": 9.77800418148711e-07, + "loss": 0.0608, + "step": 40759 + }, + { + "epoch": 0.9554785890516125, + "grad_norm": 0.32462963461875916, + "learning_rate": 9.767731038469396e-07, + "loss": 0.0498, + "step": 40760 + }, + { + "epoch": 0.9555020306264175, + "grad_norm": 0.4944612681865692, + "learning_rate": 9.757463268479816e-07, + "loss": 0.4102, + "step": 40761 + }, + { + "epoch": 0.9555254722012225, + "grad_norm": 0.4002329111099243, + "learning_rate": 9.747200871573991e-07, + "loss": 0.0728, + "step": 40762 + }, + { + "epoch": 0.9555489137760275, + "grad_norm": 0.4215301275253296, + "learning_rate": 9.736943847807766e-07, + "loss": 0.0522, + "step": 40763 + }, + { + "epoch": 0.9555723553508325, + "grad_norm": 0.13163639605045319, + "learning_rate": 9.726692197236764e-07, + "loss": 0.0331, + "step": 40764 + }, + { + "epoch": 0.9555957969256375, + "grad_norm": 0.0885029211640358, + "learning_rate": 9.716445919916607e-07, + "loss": 0.0108, + "step": 40765 + }, + { + "epoch": 0.9556192385004425, + "grad_norm": 0.5211764574050903, + "learning_rate": 9.706205015902803e-07, + "loss": 0.1024, + "step": 40766 + }, + { + "epoch": 0.9556426800752474, + "grad_norm": 0.6457346081733704, + "learning_rate": 9.695969485250977e-07, + "loss": 0.6082, + "step": 40767 + }, + { + "epoch": 0.9556661216500525, + "grad_norm": 0.30666282773017883, + "learning_rate": 9.685739328016642e-07, + "loss": 0.0554, + "step": 40768 + }, + { + "epoch": 0.9556895632248574, + "grad_norm": 0.16902251541614532, + "learning_rate": 9.675514544255304e-07, + "loss": 0.0088, + "step": 40769 + }, + { + "epoch": 0.9557130047996625, + "grad_norm": 0.47946441173553467, + "learning_rate": 9.665295134022479e-07, + "loss": 0.0637, + "step": 40770 + }, + { + "epoch": 0.9557364463744674, + "grad_norm": 0.43119969964027405, + "learning_rate": 9.655081097373676e-07, + "loss": 0.0557, + "step": 40771 + }, + { + "epoch": 0.9557598879492725, + "grad_norm": 0.6882429718971252, + "learning_rate": 9.644872434364293e-07, + "loss": 0.1629, + "step": 40772 + }, + { + "epoch": 0.9557833295240774, + "grad_norm": 0.5386879444122314, + "learning_rate": 9.634669145049513e-07, + "loss": 0.0763, + "step": 40773 + }, + { + "epoch": 0.9558067710988825, + "grad_norm": 0.3835824429988861, + "learning_rate": 9.624471229485066e-07, + "loss": 0.0402, + "step": 40774 + }, + { + "epoch": 0.9558302126736874, + "grad_norm": 0.13395318388938904, + "learning_rate": 9.61427868772591e-07, + "loss": 0.0134, + "step": 40775 + }, + { + "epoch": 0.9558536542484924, + "grad_norm": 0.2645132839679718, + "learning_rate": 9.604091519827552e-07, + "loss": 0.0463, + "step": 40776 + }, + { + "epoch": 0.9558770958232974, + "grad_norm": 0.4603755474090576, + "learning_rate": 9.593909725845285e-07, + "loss": 0.0403, + "step": 40777 + }, + { + "epoch": 0.9559005373981024, + "grad_norm": 0.32691490650177, + "learning_rate": 9.583733305834397e-07, + "loss": 0.0502, + "step": 40778 + }, + { + "epoch": 0.9559239789729074, + "grad_norm": 0.11792958527803421, + "learning_rate": 9.573562259849844e-07, + "loss": 0.0153, + "step": 40779 + }, + { + "epoch": 0.9559474205477124, + "grad_norm": 0.08277004957199097, + "learning_rate": 9.563396587947028e-07, + "loss": 0.0185, + "step": 40780 + }, + { + "epoch": 0.9559708621225174, + "grad_norm": 0.5401207804679871, + "learning_rate": 9.553236290181234e-07, + "loss": 0.0819, + "step": 40781 + }, + { + "epoch": 0.9559943036973224, + "grad_norm": 0.488586962223053, + "learning_rate": 9.543081366607198e-07, + "loss": 0.0843, + "step": 40782 + }, + { + "epoch": 0.9560177452721274, + "grad_norm": 0.11696416884660721, + "learning_rate": 9.532931817280432e-07, + "loss": 0.0281, + "step": 40783 + }, + { + "epoch": 0.9560411868469324, + "grad_norm": 0.5686468482017517, + "learning_rate": 9.522787642255781e-07, + "loss": 0.6284, + "step": 40784 + }, + { + "epoch": 0.9560646284217373, + "grad_norm": 0.3162049353122711, + "learning_rate": 9.512648841588312e-07, + "loss": 0.0491, + "step": 40785 + }, + { + "epoch": 0.9560880699965424, + "grad_norm": 0.13922707736492157, + "learning_rate": 9.502515415333091e-07, + "loss": 0.0068, + "step": 40786 + }, + { + "epoch": 0.9561115115713473, + "grad_norm": 0.2967971861362457, + "learning_rate": 9.492387363544964e-07, + "loss": 0.0432, + "step": 40787 + }, + { + "epoch": 0.9561349531461524, + "grad_norm": 0.3184710443019867, + "learning_rate": 9.482264686279108e-07, + "loss": 0.032, + "step": 40788 + }, + { + "epoch": 0.9561583947209573, + "grad_norm": 0.6465634703636169, + "learning_rate": 9.472147383590257e-07, + "loss": 0.4658, + "step": 40789 + }, + { + "epoch": 0.9561818362957624, + "grad_norm": 0.30349618196487427, + "learning_rate": 9.46203545553348e-07, + "loss": 0.0412, + "step": 40790 + }, + { + "epoch": 0.9562052778705673, + "grad_norm": 0.5411026477813721, + "learning_rate": 9.45192890216351e-07, + "loss": 0.1272, + "step": 40791 + }, + { + "epoch": 0.9562287194453724, + "grad_norm": 0.2734907567501068, + "learning_rate": 9.44182772353519e-07, + "loss": 0.0355, + "step": 40792 + }, + { + "epoch": 0.9562521610201773, + "grad_norm": 0.7961986064910889, + "learning_rate": 9.431731919703368e-07, + "loss": 0.1867, + "step": 40793 + }, + { + "epoch": 0.9562756025949823, + "grad_norm": 0.8549504280090332, + "learning_rate": 9.421641490722777e-07, + "loss": 0.0821, + "step": 40794 + }, + { + "epoch": 0.9562990441697873, + "grad_norm": 0.17629259824752808, + "learning_rate": 9.411556436648261e-07, + "loss": 0.0324, + "step": 40795 + }, + { + "epoch": 0.9563224857445923, + "grad_norm": 0.31962475180625916, + "learning_rate": 9.401476757534555e-07, + "loss": 0.0768, + "step": 40796 + }, + { + "epoch": 0.9563459273193973, + "grad_norm": 0.45698556303977966, + "learning_rate": 9.391402453436283e-07, + "loss": 0.108, + "step": 40797 + }, + { + "epoch": 0.9563693688942023, + "grad_norm": 0.675128161907196, + "learning_rate": 9.381333524408065e-07, + "loss": 0.09, + "step": 40798 + }, + { + "epoch": 0.9563928104690073, + "grad_norm": 0.5717183351516724, + "learning_rate": 9.371269970504637e-07, + "loss": 0.0795, + "step": 40799 + }, + { + "epoch": 0.9564162520438123, + "grad_norm": 0.4381336569786072, + "learning_rate": 9.3612117917804e-07, + "loss": 0.0659, + "step": 40800 + }, + { + "epoch": 0.9564396936186172, + "grad_norm": 0.34633904695510864, + "learning_rate": 9.351158988290199e-07, + "loss": 0.0333, + "step": 40801 + }, + { + "epoch": 0.9564631351934223, + "grad_norm": 0.09250561892986298, + "learning_rate": 9.341111560088545e-07, + "loss": 0.0123, + "step": 40802 + }, + { + "epoch": 0.9564865767682272, + "grad_norm": 0.5267870426177979, + "learning_rate": 9.331069507229728e-07, + "loss": 0.3557, + "step": 40803 + }, + { + "epoch": 0.9565100183430323, + "grad_norm": 0.3221501410007477, + "learning_rate": 9.321032829768484e-07, + "loss": 0.0533, + "step": 40804 + }, + { + "epoch": 0.9565334599178372, + "grad_norm": 0.39452195167541504, + "learning_rate": 9.3110015277591e-07, + "loss": 0.0558, + "step": 40805 + }, + { + "epoch": 0.9565569014926423, + "grad_norm": 0.21605293452739716, + "learning_rate": 9.30097560125609e-07, + "loss": 0.0288, + "step": 40806 + }, + { + "epoch": 0.9565803430674473, + "grad_norm": 0.34632179141044617, + "learning_rate": 9.290955050313855e-07, + "loss": 0.0315, + "step": 40807 + }, + { + "epoch": 0.9566037846422523, + "grad_norm": 0.25842776894569397, + "learning_rate": 9.280939874986794e-07, + "loss": 0.0354, + "step": 40808 + }, + { + "epoch": 0.9566272262170573, + "grad_norm": 0.625463604927063, + "learning_rate": 9.270930075329198e-07, + "loss": 0.0481, + "step": 40809 + }, + { + "epoch": 0.9566506677918623, + "grad_norm": 0.12641748785972595, + "learning_rate": 9.260925651395358e-07, + "loss": 0.0278, + "step": 40810 + }, + { + "epoch": 0.9566741093666673, + "grad_norm": 0.07841984182596207, + "learning_rate": 9.250926603239562e-07, + "loss": 0.0061, + "step": 40811 + }, + { + "epoch": 0.9566975509414722, + "grad_norm": 0.2853684723377228, + "learning_rate": 9.240932930916213e-07, + "loss": 0.0325, + "step": 40812 + }, + { + "epoch": 0.9567209925162773, + "grad_norm": 0.3105901777744293, + "learning_rate": 9.230944634479266e-07, + "loss": 0.045, + "step": 40813 + }, + { + "epoch": 0.9567444340910822, + "grad_norm": 0.25350478291511536, + "learning_rate": 9.220961713983123e-07, + "loss": 0.0202, + "step": 40814 + }, + { + "epoch": 0.9567678756658873, + "grad_norm": 0.49376818537712097, + "learning_rate": 9.210984169481962e-07, + "loss": 0.0839, + "step": 40815 + }, + { + "epoch": 0.9567913172406922, + "grad_norm": 0.5668584704399109, + "learning_rate": 9.201012001029851e-07, + "loss": 0.1001, + "step": 40816 + }, + { + "epoch": 0.9568147588154973, + "grad_norm": 0.4967673420906067, + "learning_rate": 9.19104520868086e-07, + "loss": 0.0673, + "step": 40817 + }, + { + "epoch": 0.9568382003903022, + "grad_norm": 0.5870100259780884, + "learning_rate": 9.181083792489053e-07, + "loss": 0.0616, + "step": 40818 + }, + { + "epoch": 0.9568616419651073, + "grad_norm": 0.17886705696582794, + "learning_rate": 9.171127752508724e-07, + "loss": 0.0116, + "step": 40819 + }, + { + "epoch": 0.9568850835399122, + "grad_norm": 0.40557488799095154, + "learning_rate": 9.161177088793493e-07, + "loss": 0.08, + "step": 40820 + }, + { + "epoch": 0.9569085251147172, + "grad_norm": 0.3017851710319519, + "learning_rate": 9.151231801397764e-07, + "loss": 0.0215, + "step": 40821 + }, + { + "epoch": 0.9569319666895222, + "grad_norm": 0.1271951049566269, + "learning_rate": 9.14129189037527e-07, + "loss": 0.0182, + "step": 40822 + }, + { + "epoch": 0.9569554082643272, + "grad_norm": 0.17886479198932648, + "learning_rate": 9.131357355779968e-07, + "loss": 0.0251, + "step": 40823 + }, + { + "epoch": 0.9569788498391322, + "grad_norm": 0.3413037955760956, + "learning_rate": 9.121428197665816e-07, + "loss": 0.0579, + "step": 40824 + }, + { + "epoch": 0.9570022914139372, + "grad_norm": 0.20474839210510254, + "learning_rate": 9.111504416086548e-07, + "loss": 0.043, + "step": 40825 + }, + { + "epoch": 0.9570257329887422, + "grad_norm": 0.36153027415275574, + "learning_rate": 9.101586011096231e-07, + "loss": 0.0622, + "step": 40826 + }, + { + "epoch": 0.9570491745635472, + "grad_norm": 0.107472263276577, + "learning_rate": 9.091672982748489e-07, + "loss": 0.0169, + "step": 40827 + }, + { + "epoch": 0.9570726161383522, + "grad_norm": 0.5027325749397278, + "learning_rate": 9.081765331097281e-07, + "loss": 0.0986, + "step": 40828 + }, + { + "epoch": 0.9570960577131572, + "grad_norm": 0.1904589831829071, + "learning_rate": 9.071863056196228e-07, + "loss": 0.0226, + "step": 40829 + }, + { + "epoch": 0.9571194992879621, + "grad_norm": 0.5638597011566162, + "learning_rate": 9.061966158099067e-07, + "loss": 0.078, + "step": 40830 + }, + { + "epoch": 0.9571429408627672, + "grad_norm": 0.4141785800457001, + "learning_rate": 9.052074636859531e-07, + "loss": 0.0401, + "step": 40831 + }, + { + "epoch": 0.9571663824375721, + "grad_norm": 0.2600850462913513, + "learning_rate": 9.042188492531356e-07, + "loss": 0.0555, + "step": 40832 + }, + { + "epoch": 0.9571898240123772, + "grad_norm": 0.1205984428524971, + "learning_rate": 9.032307725168165e-07, + "loss": 0.0177, + "step": 40833 + }, + { + "epoch": 0.9572132655871821, + "grad_norm": 0.5238701701164246, + "learning_rate": 9.022432334823472e-07, + "loss": 0.124, + "step": 40834 + }, + { + "epoch": 0.9572367071619872, + "grad_norm": 0.5491751432418823, + "learning_rate": 9.01256232155101e-07, + "loss": 0.5107, + "step": 40835 + }, + { + "epoch": 0.9572601487367921, + "grad_norm": 0.07071728259325027, + "learning_rate": 9.002697685404182e-07, + "loss": 0.0103, + "step": 40836 + }, + { + "epoch": 0.9572835903115972, + "grad_norm": 0.45470112562179565, + "learning_rate": 8.992838426436611e-07, + "loss": 0.0781, + "step": 40837 + }, + { + "epoch": 0.9573070318864021, + "grad_norm": 0.48863837122917175, + "learning_rate": 8.9829845447017e-07, + "loss": 0.0517, + "step": 40838 + }, + { + "epoch": 0.9573304734612071, + "grad_norm": 0.2678319215774536, + "learning_rate": 8.973136040252961e-07, + "loss": 0.0437, + "step": 40839 + }, + { + "epoch": 0.9573539150360121, + "grad_norm": 0.05122207850217819, + "learning_rate": 8.963292913143906e-07, + "loss": 0.0079, + "step": 40840 + }, + { + "epoch": 0.9573773566108171, + "grad_norm": 0.4630504846572876, + "learning_rate": 8.953455163427938e-07, + "loss": 0.0881, + "step": 40841 + }, + { + "epoch": 0.9574007981856221, + "grad_norm": 0.10883484780788422, + "learning_rate": 8.943622791158235e-07, + "loss": 0.0138, + "step": 40842 + }, + { + "epoch": 0.9574242397604271, + "grad_norm": 0.09692442417144775, + "learning_rate": 8.933795796388422e-07, + "loss": 0.0082, + "step": 40843 + }, + { + "epoch": 0.9574476813352321, + "grad_norm": 0.6359015703201294, + "learning_rate": 8.92397417917179e-07, + "loss": 0.0752, + "step": 40844 + }, + { + "epoch": 0.9574711229100371, + "grad_norm": 0.8278089761734009, + "learning_rate": 8.914157939561296e-07, + "loss": 0.0985, + "step": 40845 + }, + { + "epoch": 0.957494564484842, + "grad_norm": 0.3950800597667694, + "learning_rate": 8.904347077610676e-07, + "loss": 0.0832, + "step": 40846 + }, + { + "epoch": 0.9575180060596471, + "grad_norm": 0.236693874001503, + "learning_rate": 8.894541593372774e-07, + "loss": 0.036, + "step": 40847 + }, + { + "epoch": 0.957541447634452, + "grad_norm": 0.34389728307724, + "learning_rate": 8.884741486901104e-07, + "loss": 0.0656, + "step": 40848 + }, + { + "epoch": 0.9575648892092571, + "grad_norm": 0.1047632023692131, + "learning_rate": 8.874946758248625e-07, + "loss": 0.0041, + "step": 40849 + }, + { + "epoch": 0.957588330784062, + "grad_norm": 0.44792479276657104, + "learning_rate": 8.865157407468516e-07, + "loss": 0.0659, + "step": 40850 + }, + { + "epoch": 0.9576117723588671, + "grad_norm": 0.10409451276063919, + "learning_rate": 8.855373434613957e-07, + "loss": 0.0145, + "step": 40851 + }, + { + "epoch": 0.957635213933672, + "grad_norm": 0.44454285502433777, + "learning_rate": 8.845594839738014e-07, + "loss": 0.0695, + "step": 40852 + }, + { + "epoch": 0.9576586555084771, + "grad_norm": 0.06232753396034241, + "learning_rate": 8.835821622893759e-07, + "loss": 0.0112, + "step": 40853 + }, + { + "epoch": 0.957682097083282, + "grad_norm": 0.4618862569332123, + "learning_rate": 8.826053784134147e-07, + "loss": 0.0578, + "step": 40854 + }, + { + "epoch": 0.957705538658087, + "grad_norm": 0.4298951029777527, + "learning_rate": 8.816291323512249e-07, + "loss": 0.0185, + "step": 40855 + }, + { + "epoch": 0.957728980232892, + "grad_norm": 0.20405107736587524, + "learning_rate": 8.806534241081133e-07, + "loss": 0.0458, + "step": 40856 + }, + { + "epoch": 0.957752421807697, + "grad_norm": 0.1263856589794159, + "learning_rate": 8.796782536893533e-07, + "loss": 0.0201, + "step": 40857 + }, + { + "epoch": 0.9577758633825021, + "grad_norm": 0.27510711550712585, + "learning_rate": 8.787036211002408e-07, + "loss": 0.0502, + "step": 40858 + }, + { + "epoch": 0.957799304957307, + "grad_norm": 0.2072277069091797, + "learning_rate": 8.777295263460828e-07, + "loss": 0.0469, + "step": 40859 + }, + { + "epoch": 0.9578227465321121, + "grad_norm": 0.4967327117919922, + "learning_rate": 8.767559694321415e-07, + "loss": 0.083, + "step": 40860 + }, + { + "epoch": 0.957846188106917, + "grad_norm": 0.3730184733867645, + "learning_rate": 8.757829503637238e-07, + "loss": 0.0503, + "step": 40861 + }, + { + "epoch": 0.9578696296817221, + "grad_norm": 0.574434220790863, + "learning_rate": 8.748104691460813e-07, + "loss": 0.0938, + "step": 40862 + }, + { + "epoch": 0.957893071256527, + "grad_norm": 0.4888726770877838, + "learning_rate": 8.738385257844983e-07, + "loss": 0.085, + "step": 40863 + }, + { + "epoch": 0.9579165128313321, + "grad_norm": 0.5759554505348206, + "learning_rate": 8.728671202842709e-07, + "loss": 0.0489, + "step": 40864 + }, + { + "epoch": 0.957939954406137, + "grad_norm": 0.09012379497289658, + "learning_rate": 8.718962526506502e-07, + "loss": 0.0071, + "step": 40865 + }, + { + "epoch": 0.957963395980942, + "grad_norm": 0.4919450283050537, + "learning_rate": 8.709259228888988e-07, + "loss": 0.0887, + "step": 40866 + }, + { + "epoch": 0.957986837555747, + "grad_norm": 0.5586044192314148, + "learning_rate": 8.699561310042903e-07, + "loss": 0.0578, + "step": 40867 + }, + { + "epoch": 0.958010279130552, + "grad_norm": 0.3500494062900543, + "learning_rate": 8.689868770020981e-07, + "loss": 0.0714, + "step": 40868 + }, + { + "epoch": 0.958033720705357, + "grad_norm": 0.34971192479133606, + "learning_rate": 8.680181608875626e-07, + "loss": 0.0531, + "step": 40869 + }, + { + "epoch": 0.958057162280162, + "grad_norm": 0.7830959558486938, + "learning_rate": 8.670499826659462e-07, + "loss": 0.0969, + "step": 40870 + }, + { + "epoch": 0.958080603854967, + "grad_norm": 0.2537459135055542, + "learning_rate": 8.660823423425113e-07, + "loss": 0.0273, + "step": 40871 + }, + { + "epoch": 0.958104045429772, + "grad_norm": 0.14921671152114868, + "learning_rate": 8.651152399224982e-07, + "loss": 0.0101, + "step": 40872 + }, + { + "epoch": 0.958127487004577, + "grad_norm": 0.06879954040050507, + "learning_rate": 8.641486754111582e-07, + "loss": 0.0087, + "step": 40873 + }, + { + "epoch": 0.958150928579382, + "grad_norm": 0.09824836254119873, + "learning_rate": 8.631826488137429e-07, + "loss": 0.0138, + "step": 40874 + }, + { + "epoch": 0.9581743701541869, + "grad_norm": 0.4674322307109833, + "learning_rate": 8.622171601354701e-07, + "loss": 0.0498, + "step": 40875 + }, + { + "epoch": 0.958197811728992, + "grad_norm": 0.550317108631134, + "learning_rate": 8.612522093816022e-07, + "loss": 0.083, + "step": 40876 + }, + { + "epoch": 0.9582212533037969, + "grad_norm": 0.3650210201740265, + "learning_rate": 8.602877965573686e-07, + "loss": 0.4097, + "step": 40877 + }, + { + "epoch": 0.958244694878602, + "grad_norm": 0.2603977620601654, + "learning_rate": 8.593239216679872e-07, + "loss": 0.016, + "step": 40878 + }, + { + "epoch": 0.9582681364534069, + "grad_norm": 0.5021613240242004, + "learning_rate": 8.583605847187203e-07, + "loss": 0.1129, + "step": 40879 + }, + { + "epoch": 0.958291578028212, + "grad_norm": 0.10008829832077026, + "learning_rate": 8.573977857147641e-07, + "loss": 0.0067, + "step": 40880 + }, + { + "epoch": 0.9583150196030169, + "grad_norm": 0.3822406828403473, + "learning_rate": 8.564355246613698e-07, + "loss": 0.0574, + "step": 40881 + }, + { + "epoch": 0.958338461177822, + "grad_norm": 0.3603895902633667, + "learning_rate": 8.554738015637331e-07, + "loss": 0.0464, + "step": 40882 + }, + { + "epoch": 0.9583619027526269, + "grad_norm": 0.08689062297344208, + "learning_rate": 8.545126164270834e-07, + "loss": 0.0105, + "step": 40883 + }, + { + "epoch": 0.9583853443274319, + "grad_norm": 0.50091952085495, + "learning_rate": 8.535519692566385e-07, + "loss": 0.0658, + "step": 40884 + }, + { + "epoch": 0.9584087859022369, + "grad_norm": 0.5068066716194153, + "learning_rate": 8.525918600576055e-07, + "loss": 0.095, + "step": 40885 + }, + { + "epoch": 0.9584322274770419, + "grad_norm": 0.36189004778862, + "learning_rate": 8.516322888352024e-07, + "loss": 0.061, + "step": 40886 + }, + { + "epoch": 0.9584556690518469, + "grad_norm": 0.27783721685409546, + "learning_rate": 8.506732555946251e-07, + "loss": 0.0261, + "step": 40887 + }, + { + "epoch": 0.9584791106266519, + "grad_norm": 0.48428183794021606, + "learning_rate": 8.497147603410804e-07, + "loss": 0.0932, + "step": 40888 + }, + { + "epoch": 0.9585025522014569, + "grad_norm": 0.6817962527275085, + "learning_rate": 8.487568030797755e-07, + "loss": 0.0517, + "step": 40889 + }, + { + "epoch": 0.9585259937762619, + "grad_norm": 0.47079789638519287, + "learning_rate": 8.477993838159059e-07, + "loss": 0.0848, + "step": 40890 + }, + { + "epoch": 0.9585494353510668, + "grad_norm": 0.5465855598449707, + "learning_rate": 8.468425025546567e-07, + "loss": 0.0703, + "step": 40891 + }, + { + "epoch": 0.9585728769258719, + "grad_norm": 0.4785575568675995, + "learning_rate": 8.458861593012346e-07, + "loss": 0.0433, + "step": 40892 + }, + { + "epoch": 0.9585963185006768, + "grad_norm": 0.17902730405330658, + "learning_rate": 8.449303540608244e-07, + "loss": 0.0196, + "step": 40893 + }, + { + "epoch": 0.9586197600754819, + "grad_norm": 0.63165283203125, + "learning_rate": 8.439750868386109e-07, + "loss": 0.0576, + "step": 40894 + }, + { + "epoch": 0.9586432016502868, + "grad_norm": 0.19322890043258667, + "learning_rate": 8.430203576397678e-07, + "loss": 0.0293, + "step": 40895 + }, + { + "epoch": 0.9586666432250919, + "grad_norm": 0.14908425509929657, + "learning_rate": 8.420661664694907e-07, + "loss": 0.0163, + "step": 40896 + }, + { + "epoch": 0.9586900847998968, + "grad_norm": 0.25858068466186523, + "learning_rate": 8.411125133329422e-07, + "loss": 0.0334, + "step": 40897 + }, + { + "epoch": 0.9587135263747019, + "grad_norm": 0.4359133243560791, + "learning_rate": 8.401593982353184e-07, + "loss": 0.0904, + "step": 40898 + }, + { + "epoch": 0.9587369679495068, + "grad_norm": 0.3409287929534912, + "learning_rate": 8.392068211817705e-07, + "loss": 0.0464, + "step": 40899 + }, + { + "epoch": 0.9587604095243119, + "grad_norm": 0.29658105969429016, + "learning_rate": 8.382547821774722e-07, + "loss": 0.033, + "step": 40900 + }, + { + "epoch": 0.9587838510991168, + "grad_norm": 0.36855456233024597, + "learning_rate": 8.373032812275972e-07, + "loss": 0.035, + "step": 40901 + }, + { + "epoch": 0.9588072926739218, + "grad_norm": 0.3526892364025116, + "learning_rate": 8.363523183372968e-07, + "loss": 0.06, + "step": 40902 + }, + { + "epoch": 0.9588307342487268, + "grad_norm": 0.5023066401481628, + "learning_rate": 8.354018935117447e-07, + "loss": 0.6467, + "step": 40903 + }, + { + "epoch": 0.9588541758235318, + "grad_norm": 0.7001486420631409, + "learning_rate": 8.344520067560813e-07, + "loss": 0.1131, + "step": 40904 + }, + { + "epoch": 0.9588776173983368, + "grad_norm": 0.5689641833305359, + "learning_rate": 8.335026580754801e-07, + "loss": 0.0751, + "step": 40905 + }, + { + "epoch": 0.9589010589731418, + "grad_norm": 0.1870666891336441, + "learning_rate": 8.325538474750705e-07, + "loss": 0.0282, + "step": 40906 + }, + { + "epoch": 0.9589245005479468, + "grad_norm": 0.11639492213726044, + "learning_rate": 8.31605574960026e-07, + "loss": 0.022, + "step": 40907 + }, + { + "epoch": 0.9589479421227518, + "grad_norm": 0.22697414457798004, + "learning_rate": 8.306578405354648e-07, + "loss": 0.0498, + "step": 40908 + }, + { + "epoch": 0.9589713836975569, + "grad_norm": 0.5709290504455566, + "learning_rate": 8.297106442065495e-07, + "loss": 0.1448, + "step": 40909 + }, + { + "epoch": 0.9589948252723618, + "grad_norm": 0.42082732915878296, + "learning_rate": 8.28763985978409e-07, + "loss": 0.0597, + "step": 40910 + }, + { + "epoch": 0.9590182668471668, + "grad_norm": 0.7214022278785706, + "learning_rate": 8.278178658561841e-07, + "loss": 0.1168, + "step": 40911 + }, + { + "epoch": 0.9590417084219718, + "grad_norm": 0.3963438868522644, + "learning_rate": 8.268722838450149e-07, + "loss": 0.0545, + "step": 40912 + }, + { + "epoch": 0.9590651499967768, + "grad_norm": 0.13312727212905884, + "learning_rate": 8.259272399500195e-07, + "loss": 0.0184, + "step": 40913 + }, + { + "epoch": 0.9590885915715818, + "grad_norm": 0.33599144220352173, + "learning_rate": 8.249827341763273e-07, + "loss": 0.0301, + "step": 40914 + }, + { + "epoch": 0.9591120331463868, + "grad_norm": 0.2977815568447113, + "learning_rate": 8.240387665290672e-07, + "loss": 0.0541, + "step": 40915 + }, + { + "epoch": 0.9591354747211918, + "grad_norm": 0.41924768686294556, + "learning_rate": 8.230953370133688e-07, + "loss": 0.0501, + "step": 40916 + }, + { + "epoch": 0.9591589162959968, + "grad_norm": 0.5251157283782959, + "learning_rate": 8.221524456343388e-07, + "loss": 0.0337, + "step": 40917 + }, + { + "epoch": 0.9591823578708017, + "grad_norm": 0.06345956027507782, + "learning_rate": 8.212100923971067e-07, + "loss": 0.0107, + "step": 40918 + }, + { + "epoch": 0.9592057994456068, + "grad_norm": 0.17996098101139069, + "learning_rate": 8.202682773067683e-07, + "loss": 0.025, + "step": 40919 + }, + { + "epoch": 0.9592292410204117, + "grad_norm": 0.40431681275367737, + "learning_rate": 8.193270003684417e-07, + "loss": 0.0779, + "step": 40920 + }, + { + "epoch": 0.9592526825952168, + "grad_norm": 0.21183204650878906, + "learning_rate": 8.18386261587234e-07, + "loss": 0.1456, + "step": 40921 + }, + { + "epoch": 0.9592761241700217, + "grad_norm": 0.7259838581085205, + "learning_rate": 8.174460609682633e-07, + "loss": 0.0942, + "step": 40922 + }, + { + "epoch": 0.9592995657448268, + "grad_norm": 0.19682636857032776, + "learning_rate": 8.165063985166143e-07, + "loss": 0.0137, + "step": 40923 + }, + { + "epoch": 0.9593230073196317, + "grad_norm": 0.26030853390693665, + "learning_rate": 8.155672742373833e-07, + "loss": 0.0471, + "step": 40924 + }, + { + "epoch": 0.9593464488944368, + "grad_norm": 0.27008986473083496, + "learning_rate": 8.146286881356769e-07, + "loss": 0.0574, + "step": 40925 + }, + { + "epoch": 0.9593698904692417, + "grad_norm": 0.43136534094810486, + "learning_rate": 8.136906402165912e-07, + "loss": 0.0636, + "step": 40926 + }, + { + "epoch": 0.9593933320440468, + "grad_norm": 0.14180442690849304, + "learning_rate": 8.127531304851998e-07, + "loss": 0.0246, + "step": 40927 + }, + { + "epoch": 0.9594167736188517, + "grad_norm": 0.36328649520874023, + "learning_rate": 8.118161589465989e-07, + "loss": 0.0487, + "step": 40928 + }, + { + "epoch": 0.9594402151936567, + "grad_norm": 0.5223789811134338, + "learning_rate": 8.108797256058731e-07, + "loss": 0.0507, + "step": 40929 + }, + { + "epoch": 0.9594636567684617, + "grad_norm": 0.24868813157081604, + "learning_rate": 8.099438304681073e-07, + "loss": 0.0252, + "step": 40930 + }, + { + "epoch": 0.9594870983432667, + "grad_norm": 0.2096508890390396, + "learning_rate": 8.090084735383751e-07, + "loss": 0.0316, + "step": 40931 + }, + { + "epoch": 0.9595105399180717, + "grad_norm": 0.49852025508880615, + "learning_rate": 8.080736548217505e-07, + "loss": 0.0987, + "step": 40932 + }, + { + "epoch": 0.9595339814928767, + "grad_norm": 0.6057229042053223, + "learning_rate": 8.071393743233069e-07, + "loss": 0.1185, + "step": 40933 + }, + { + "epoch": 0.9595574230676817, + "grad_norm": 0.26959288120269775, + "learning_rate": 8.06205632048107e-07, + "loss": 0.0407, + "step": 40934 + }, + { + "epoch": 0.9595808646424867, + "grad_norm": 0.3291759490966797, + "learning_rate": 8.052724280012358e-07, + "loss": 0.0669, + "step": 40935 + }, + { + "epoch": 0.9596043062172916, + "grad_norm": 0.6319139003753662, + "learning_rate": 8.043397621877446e-07, + "loss": 0.0714, + "step": 40936 + }, + { + "epoch": 0.9596277477920967, + "grad_norm": 0.7893906831741333, + "learning_rate": 8.03407634612685e-07, + "loss": 0.1586, + "step": 40937 + }, + { + "epoch": 0.9596511893669016, + "grad_norm": 0.3238494396209717, + "learning_rate": 8.02476045281142e-07, + "loss": 0.0729, + "step": 40938 + }, + { + "epoch": 0.9596746309417067, + "grad_norm": 0.2667678892612457, + "learning_rate": 8.015449941981334e-07, + "loss": 0.0251, + "step": 40939 + }, + { + "epoch": 0.9596980725165116, + "grad_norm": 0.12461531162261963, + "learning_rate": 8.006144813687333e-07, + "loss": 0.0361, + "step": 40940 + }, + { + "epoch": 0.9597215140913167, + "grad_norm": 0.2112937569618225, + "learning_rate": 7.99684506797993e-07, + "loss": 0.0185, + "step": 40941 + }, + { + "epoch": 0.9597449556661216, + "grad_norm": 0.11741974949836731, + "learning_rate": 7.987550704909641e-07, + "loss": 0.0126, + "step": 40942 + }, + { + "epoch": 0.9597683972409267, + "grad_norm": 0.35631492733955383, + "learning_rate": 7.978261724526648e-07, + "loss": 0.0586, + "step": 40943 + }, + { + "epoch": 0.9597918388157316, + "grad_norm": 0.41172468662261963, + "learning_rate": 7.968978126881577e-07, + "loss": 0.0621, + "step": 40944 + }, + { + "epoch": 0.9598152803905367, + "grad_norm": 0.6287060976028442, + "learning_rate": 7.959699912024721e-07, + "loss": 0.1339, + "step": 40945 + }, + { + "epoch": 0.9598387219653416, + "grad_norm": 0.27712371945381165, + "learning_rate": 7.950427080006373e-07, + "loss": 0.0285, + "step": 40946 + }, + { + "epoch": 0.9598621635401466, + "grad_norm": 0.7308387160301208, + "learning_rate": 7.941159630877049e-07, + "loss": 0.0828, + "step": 40947 + }, + { + "epoch": 0.9598856051149516, + "grad_norm": 0.4049719572067261, + "learning_rate": 7.931897564686707e-07, + "loss": 0.0652, + "step": 40948 + }, + { + "epoch": 0.9599090466897566, + "grad_norm": 0.31507885456085205, + "learning_rate": 7.922640881485865e-07, + "loss": 0.0317, + "step": 40949 + }, + { + "epoch": 0.9599324882645616, + "grad_norm": 0.2734004259109497, + "learning_rate": 7.913389581324704e-07, + "loss": 0.0461, + "step": 40950 + }, + { + "epoch": 0.9599559298393666, + "grad_norm": 0.46616682410240173, + "learning_rate": 7.904143664253294e-07, + "loss": 0.0625, + "step": 40951 + }, + { + "epoch": 0.9599793714141716, + "grad_norm": 0.519534707069397, + "learning_rate": 7.894903130321929e-07, + "loss": 0.0455, + "step": 40952 + }, + { + "epoch": 0.9600028129889766, + "grad_norm": 0.2405700981616974, + "learning_rate": 7.88566797958068e-07, + "loss": 0.0209, + "step": 40953 + }, + { + "epoch": 0.9600262545637815, + "grad_norm": 0.4064336121082306, + "learning_rate": 7.87643821207984e-07, + "loss": 0.077, + "step": 40954 + }, + { + "epoch": 0.9600496961385866, + "grad_norm": 0.16731485724449158, + "learning_rate": 7.867213827869258e-07, + "loss": 0.0296, + "step": 40955 + }, + { + "epoch": 0.9600731377133915, + "grad_norm": 0.6338719129562378, + "learning_rate": 7.857994826999004e-07, + "loss": 0.1005, + "step": 40956 + }, + { + "epoch": 0.9600965792881966, + "grad_norm": 0.6092003583908081, + "learning_rate": 7.848781209519263e-07, + "loss": 0.0227, + "step": 40957 + }, + { + "epoch": 0.9601200208630015, + "grad_norm": 0.23644664883613586, + "learning_rate": 7.839572975479881e-07, + "loss": 0.0371, + "step": 40958 + }, + { + "epoch": 0.9601434624378066, + "grad_norm": 0.2951560318470001, + "learning_rate": 7.830370124930819e-07, + "loss": 0.0508, + "step": 40959 + }, + { + "epoch": 0.9601669040126116, + "grad_norm": 0.12020846456289291, + "learning_rate": 7.821172657922149e-07, + "loss": 0.0089, + "step": 40960 + }, + { + "epoch": 0.9601903455874166, + "grad_norm": 0.28726711869239807, + "learning_rate": 7.811980574503718e-07, + "loss": 0.0539, + "step": 40961 + }, + { + "epoch": 0.9602137871622216, + "grad_norm": 0.11026296019554138, + "learning_rate": 7.802793874725378e-07, + "loss": 0.0302, + "step": 40962 + }, + { + "epoch": 0.9602372287370265, + "grad_norm": 0.3274676203727722, + "learning_rate": 7.793612558636864e-07, + "loss": 0.0239, + "step": 40963 + }, + { + "epoch": 0.9602606703118316, + "grad_norm": 0.48551425337791443, + "learning_rate": 7.784436626288139e-07, + "loss": 0.0886, + "step": 40964 + }, + { + "epoch": 0.9602841118866365, + "grad_norm": 0.19792470335960388, + "learning_rate": 7.775266077729049e-07, + "loss": 0.0327, + "step": 40965 + }, + { + "epoch": 0.9603075534614416, + "grad_norm": 0.2645089030265808, + "learning_rate": 7.766100913009111e-07, + "loss": 0.0439, + "step": 40966 + }, + { + "epoch": 0.9603309950362465, + "grad_norm": 0.4366612136363983, + "learning_rate": 7.756941132178397e-07, + "loss": 0.4837, + "step": 40967 + }, + { + "epoch": 0.9603544366110516, + "grad_norm": 0.8765183687210083, + "learning_rate": 7.74778673528631e-07, + "loss": 0.1304, + "step": 40968 + }, + { + "epoch": 0.9603778781858565, + "grad_norm": 0.28900325298309326, + "learning_rate": 7.738637722382703e-07, + "loss": 0.0373, + "step": 40969 + }, + { + "epoch": 0.9604013197606616, + "grad_norm": 0.2214410901069641, + "learning_rate": 7.729494093517086e-07, + "loss": 0.0191, + "step": 40970 + }, + { + "epoch": 0.9604247613354665, + "grad_norm": 0.5985098481178284, + "learning_rate": 7.720355848739202e-07, + "loss": 0.1179, + "step": 40971 + }, + { + "epoch": 0.9604482029102716, + "grad_norm": 0.27142664790153503, + "learning_rate": 7.711222988098455e-07, + "loss": 0.0518, + "step": 40972 + }, + { + "epoch": 0.9604716444850765, + "grad_norm": 0.7137444615364075, + "learning_rate": 7.70209551164458e-07, + "loss": 0.1288, + "step": 40973 + }, + { + "epoch": 0.9604950860598815, + "grad_norm": 0.2023545801639557, + "learning_rate": 7.692973419426985e-07, + "loss": 0.0427, + "step": 40974 + }, + { + "epoch": 0.9605185276346865, + "grad_norm": 0.4118281900882721, + "learning_rate": 7.683856711495296e-07, + "loss": 0.0514, + "step": 40975 + }, + { + "epoch": 0.9605419692094915, + "grad_norm": 0.5322848558425903, + "learning_rate": 7.674745387898807e-07, + "loss": 0.0982, + "step": 40976 + }, + { + "epoch": 0.9605654107842965, + "grad_norm": 0.3255932927131653, + "learning_rate": 7.665639448687145e-07, + "loss": 0.0521, + "step": 40977 + }, + { + "epoch": 0.9605888523591015, + "grad_norm": 0.25550705194473267, + "learning_rate": 7.656538893909493e-07, + "loss": 0.0222, + "step": 40978 + }, + { + "epoch": 0.9606122939339065, + "grad_norm": 0.2947569489479065, + "learning_rate": 7.647443723615477e-07, + "loss": 0.0457, + "step": 40979 + }, + { + "epoch": 0.9606357355087115, + "grad_norm": 0.3132898509502411, + "learning_rate": 7.638353937854392e-07, + "loss": 0.033, + "step": 40980 + }, + { + "epoch": 0.9606591770835164, + "grad_norm": 0.13406382501125336, + "learning_rate": 7.629269536675421e-07, + "loss": 0.0211, + "step": 40981 + }, + { + "epoch": 0.9606826186583215, + "grad_norm": 0.6533810496330261, + "learning_rate": 7.620190520127968e-07, + "loss": 0.122, + "step": 40982 + }, + { + "epoch": 0.9607060602331264, + "grad_norm": 0.2086481899023056, + "learning_rate": 7.611116888261216e-07, + "loss": 0.0204, + "step": 40983 + }, + { + "epoch": 0.9607295018079315, + "grad_norm": 0.6155955791473389, + "learning_rate": 7.60204864112446e-07, + "loss": 0.0858, + "step": 40984 + }, + { + "epoch": 0.9607529433827364, + "grad_norm": 0.2764158844947815, + "learning_rate": 7.592985778766881e-07, + "loss": 0.0535, + "step": 40985 + }, + { + "epoch": 0.9607763849575415, + "grad_norm": 0.27854353189468384, + "learning_rate": 7.583928301237664e-07, + "loss": 0.0327, + "step": 40986 + }, + { + "epoch": 0.9607998265323464, + "grad_norm": 0.14248675107955933, + "learning_rate": 7.57487620858599e-07, + "loss": 0.03, + "step": 40987 + }, + { + "epoch": 0.9608232681071515, + "grad_norm": 0.4823664426803589, + "learning_rate": 7.565829500860932e-07, + "loss": 0.065, + "step": 40988 + }, + { + "epoch": 0.9608467096819564, + "grad_norm": 0.70245361328125, + "learning_rate": 7.556788178111563e-07, + "loss": 0.0863, + "step": 40989 + }, + { + "epoch": 0.9608701512567615, + "grad_norm": 0.2558106482028961, + "learning_rate": 7.547752240386952e-07, + "loss": 0.0262, + "step": 40990 + }, + { + "epoch": 0.9608935928315664, + "grad_norm": 0.4484921097755432, + "learning_rate": 7.538721687736173e-07, + "loss": 0.0566, + "step": 40991 + }, + { + "epoch": 0.9609170344063714, + "grad_norm": 0.4822624921798706, + "learning_rate": 7.529696520208184e-07, + "loss": 0.0663, + "step": 40992 + }, + { + "epoch": 0.9609404759811764, + "grad_norm": 0.4795970916748047, + "learning_rate": 7.520676737852061e-07, + "loss": 0.1143, + "step": 40993 + }, + { + "epoch": 0.9609639175559814, + "grad_norm": 0.5184338092803955, + "learning_rate": 7.511662340716541e-07, + "loss": 0.0936, + "step": 40994 + }, + { + "epoch": 0.9609873591307864, + "grad_norm": 0.12917816638946533, + "learning_rate": 7.502653328850695e-07, + "loss": 0.016, + "step": 40995 + }, + { + "epoch": 0.9610108007055914, + "grad_norm": 0.3724953830242157, + "learning_rate": 7.493649702303373e-07, + "loss": 0.0586, + "step": 40996 + }, + { + "epoch": 0.9610342422803964, + "grad_norm": 0.2303185909986496, + "learning_rate": 7.484651461123315e-07, + "loss": 0.0293, + "step": 40997 + }, + { + "epoch": 0.9610576838552014, + "grad_norm": 0.28867292404174805, + "learning_rate": 7.475658605359592e-07, + "loss": 0.038, + "step": 40998 + }, + { + "epoch": 0.9610811254300063, + "grad_norm": 0.4497532546520233, + "learning_rate": 7.466671135060721e-07, + "loss": 0.032, + "step": 40999 + }, + { + "epoch": 0.9611045670048114, + "grad_norm": 0.3516960144042969, + "learning_rate": 7.457689050275662e-07, + "loss": 0.0823, + "step": 41000 + }, + { + "epoch": 0.9611280085796163, + "grad_norm": 0.5148528814315796, + "learning_rate": 7.448712351053044e-07, + "loss": 0.1595, + "step": 41001 + }, + { + "epoch": 0.9611514501544214, + "grad_norm": 0.1464974731206894, + "learning_rate": 7.439741037441717e-07, + "loss": 0.0154, + "step": 41002 + }, + { + "epoch": 0.9611748917292263, + "grad_norm": 0.5264211893081665, + "learning_rate": 7.430775109490196e-07, + "loss": 0.0926, + "step": 41003 + }, + { + "epoch": 0.9611983333040314, + "grad_norm": 0.38105136156082153, + "learning_rate": 7.421814567247221e-07, + "loss": 0.0331, + "step": 41004 + }, + { + "epoch": 0.9612217748788363, + "grad_norm": 0.46981632709503174, + "learning_rate": 7.412859410761308e-07, + "loss": 0.0629, + "step": 41005 + }, + { + "epoch": 0.9612452164536414, + "grad_norm": 0.9000663161277771, + "learning_rate": 7.403909640081197e-07, + "loss": 0.0991, + "step": 41006 + }, + { + "epoch": 0.9612686580284463, + "grad_norm": 0.5481259226799011, + "learning_rate": 7.394965255255404e-07, + "loss": 0.0605, + "step": 41007 + }, + { + "epoch": 0.9612920996032513, + "grad_norm": 0.3200632631778717, + "learning_rate": 7.386026256332446e-07, + "loss": 0.462, + "step": 41008 + }, + { + "epoch": 0.9613155411780563, + "grad_norm": 0.5440254211425781, + "learning_rate": 7.377092643360728e-07, + "loss": 0.0651, + "step": 41009 + }, + { + "epoch": 0.9613389827528613, + "grad_norm": 0.6996987462043762, + "learning_rate": 7.36816441638899e-07, + "loss": 0.1227, + "step": 41010 + }, + { + "epoch": 0.9613624243276664, + "grad_norm": 0.38008660078048706, + "learning_rate": 7.359241575465303e-07, + "loss": 0.0633, + "step": 41011 + }, + { + "epoch": 0.9613858659024713, + "grad_norm": 0.4781181514263153, + "learning_rate": 7.350324120638408e-07, + "loss": 0.1324, + "step": 41012 + }, + { + "epoch": 0.9614093074772764, + "grad_norm": 0.27387043833732605, + "learning_rate": 7.341412051956487e-07, + "loss": 0.0414, + "step": 41013 + }, + { + "epoch": 0.9614327490520813, + "grad_norm": 0.46625006198883057, + "learning_rate": 7.332505369468057e-07, + "loss": 0.0445, + "step": 41014 + }, + { + "epoch": 0.9614561906268864, + "grad_norm": 0.18769386410713196, + "learning_rate": 7.323604073221301e-07, + "loss": 0.0062, + "step": 41015 + }, + { + "epoch": 0.9614796322016913, + "grad_norm": 1.1238123178482056, + "learning_rate": 7.314708163264627e-07, + "loss": 0.0496, + "step": 41016 + }, + { + "epoch": 0.9615030737764964, + "grad_norm": 0.13189934194087982, + "learning_rate": 7.305817639646217e-07, + "loss": 0.0132, + "step": 41017 + }, + { + "epoch": 0.9615265153513013, + "grad_norm": 0.12116637825965881, + "learning_rate": 7.296932502414366e-07, + "loss": 0.0171, + "step": 41018 + }, + { + "epoch": 0.9615499569261063, + "grad_norm": 0.48468849062919617, + "learning_rate": 7.288052751617258e-07, + "loss": 0.0873, + "step": 41019 + }, + { + "epoch": 0.9615733985009113, + "grad_norm": 0.537036120891571, + "learning_rate": 7.279178387303188e-07, + "loss": 0.0881, + "step": 41020 + }, + { + "epoch": 0.9615968400757163, + "grad_norm": 0.3731094300746918, + "learning_rate": 7.270309409520116e-07, + "loss": 0.07, + "step": 41021 + }, + { + "epoch": 0.9616202816505213, + "grad_norm": 0.3469616174697876, + "learning_rate": 7.261445818316226e-07, + "loss": 0.0591, + "step": 41022 + }, + { + "epoch": 0.9616437232253263, + "grad_norm": 0.5937955379486084, + "learning_rate": 7.252587613739703e-07, + "loss": 0.0793, + "step": 41023 + }, + { + "epoch": 0.9616671648001313, + "grad_norm": 0.3315173387527466, + "learning_rate": 7.243734795838509e-07, + "loss": 0.0483, + "step": 41024 + }, + { + "epoch": 0.9616906063749363, + "grad_norm": 0.41909459233283997, + "learning_rate": 7.234887364660714e-07, + "loss": 0.07, + "step": 41025 + }, + { + "epoch": 0.9617140479497412, + "grad_norm": 0.3004055321216583, + "learning_rate": 7.226045320254394e-07, + "loss": 0.046, + "step": 41026 + }, + { + "epoch": 0.9617374895245463, + "grad_norm": 0.32006970047950745, + "learning_rate": 7.217208662667507e-07, + "loss": 0.0806, + "step": 41027 + }, + { + "epoch": 0.9617609310993512, + "grad_norm": 0.3462013602256775, + "learning_rate": 7.208377391947907e-07, + "loss": 0.0463, + "step": 41028 + }, + { + "epoch": 0.9617843726741563, + "grad_norm": 0.1954786479473114, + "learning_rate": 7.199551508143443e-07, + "loss": 0.0115, + "step": 41029 + }, + { + "epoch": 0.9618078142489612, + "grad_norm": 0.25136232376098633, + "learning_rate": 7.190731011302299e-07, + "loss": 0.0193, + "step": 41030 + }, + { + "epoch": 0.9618312558237663, + "grad_norm": 0.12385747581720352, + "learning_rate": 7.181915901472102e-07, + "loss": 0.0199, + "step": 41031 + }, + { + "epoch": 0.9618546973985712, + "grad_norm": 0.34390321373939514, + "learning_rate": 7.173106178700706e-07, + "loss": 0.3891, + "step": 41032 + }, + { + "epoch": 0.9618781389733763, + "grad_norm": 0.5903275609016418, + "learning_rate": 7.164301843036069e-07, + "loss": 0.0729, + "step": 41033 + }, + { + "epoch": 0.9619015805481812, + "grad_norm": 0.5504831671714783, + "learning_rate": 7.155502894525823e-07, + "loss": 0.099, + "step": 41034 + }, + { + "epoch": 0.9619250221229863, + "grad_norm": 0.38015973567962646, + "learning_rate": 7.146709333217594e-07, + "loss": 0.0408, + "step": 41035 + }, + { + "epoch": 0.9619484636977912, + "grad_norm": 0.3731142580509186, + "learning_rate": 7.137921159159233e-07, + "loss": 0.0321, + "step": 41036 + }, + { + "epoch": 0.9619719052725962, + "grad_norm": 0.21024027466773987, + "learning_rate": 7.129138372398592e-07, + "loss": 0.0197, + "step": 41037 + }, + { + "epoch": 0.9619953468474012, + "grad_norm": 0.2734372317790985, + "learning_rate": 7.120360972983075e-07, + "loss": 0.0251, + "step": 41038 + }, + { + "epoch": 0.9620187884222062, + "grad_norm": 0.2013244479894638, + "learning_rate": 7.111588960960425e-07, + "loss": 0.0289, + "step": 41039 + }, + { + "epoch": 0.9620422299970112, + "grad_norm": 0.08946307748556137, + "learning_rate": 7.102822336378267e-07, + "loss": 0.0166, + "step": 41040 + }, + { + "epoch": 0.9620656715718162, + "grad_norm": 0.22377926111221313, + "learning_rate": 7.09406109928401e-07, + "loss": 0.0341, + "step": 41041 + }, + { + "epoch": 0.9620891131466212, + "grad_norm": 0.2893756330013275, + "learning_rate": 7.085305249725282e-07, + "loss": 0.0559, + "step": 41042 + }, + { + "epoch": 0.9621125547214262, + "grad_norm": 0.4833306670188904, + "learning_rate": 7.076554787749712e-07, + "loss": 0.0839, + "step": 41043 + }, + { + "epoch": 0.9621359962962311, + "grad_norm": 0.336556613445282, + "learning_rate": 7.067809713404594e-07, + "loss": 0.0281, + "step": 41044 + }, + { + "epoch": 0.9621594378710362, + "grad_norm": 0.1323613077402115, + "learning_rate": 7.059070026737557e-07, + "loss": 0.0258, + "step": 41045 + }, + { + "epoch": 0.9621828794458411, + "grad_norm": 0.47610363364219666, + "learning_rate": 7.050335727795787e-07, + "loss": 0.1102, + "step": 41046 + }, + { + "epoch": 0.9622063210206462, + "grad_norm": 0.3047488331794739, + "learning_rate": 7.041606816626911e-07, + "loss": 0.024, + "step": 41047 + }, + { + "epoch": 0.9622297625954511, + "grad_norm": 0.4704292416572571, + "learning_rate": 7.032883293278114e-07, + "loss": 0.4775, + "step": 41048 + }, + { + "epoch": 0.9622532041702562, + "grad_norm": 0.47118788957595825, + "learning_rate": 7.024165157796803e-07, + "loss": 0.6787, + "step": 41049 + }, + { + "epoch": 0.9622766457450611, + "grad_norm": 0.1824205219745636, + "learning_rate": 7.015452410230383e-07, + "loss": 0.0159, + "step": 41050 + }, + { + "epoch": 0.9623000873198662, + "grad_norm": 0.25696441531181335, + "learning_rate": 7.00674505062604e-07, + "loss": 0.039, + "step": 41051 + }, + { + "epoch": 0.9623235288946711, + "grad_norm": 0.44681715965270996, + "learning_rate": 6.998043079030958e-07, + "loss": 0.0644, + "step": 41052 + }, + { + "epoch": 0.9623469704694761, + "grad_norm": 0.4644842743873596, + "learning_rate": 6.989346495492544e-07, + "loss": 0.0695, + "step": 41053 + }, + { + "epoch": 0.9623704120442811, + "grad_norm": 0.4072171449661255, + "learning_rate": 6.980655300057648e-07, + "loss": 0.0554, + "step": 41054 + }, + { + "epoch": 0.9623938536190861, + "grad_norm": 0.10957791656255722, + "learning_rate": 6.971969492773789e-07, + "loss": 0.007, + "step": 41055 + }, + { + "epoch": 0.9624172951938911, + "grad_norm": 0.1402096003293991, + "learning_rate": 6.963289073687929e-07, + "loss": 0.0135, + "step": 41056 + }, + { + "epoch": 0.9624407367686961, + "grad_norm": 0.05550704151391983, + "learning_rate": 6.95461404284714e-07, + "loss": 0.0057, + "step": 41057 + }, + { + "epoch": 0.9624641783435011, + "grad_norm": 0.37345215678215027, + "learning_rate": 6.945944400298609e-07, + "loss": 0.029, + "step": 41058 + }, + { + "epoch": 0.9624876199183061, + "grad_norm": 0.14816604554653168, + "learning_rate": 6.937280146089186e-07, + "loss": 0.0177, + "step": 41059 + }, + { + "epoch": 0.962511061493111, + "grad_norm": 0.735567033290863, + "learning_rate": 6.928621280266056e-07, + "loss": 0.1582, + "step": 41060 + }, + { + "epoch": 0.9625345030679161, + "grad_norm": 0.49038299918174744, + "learning_rate": 6.919967802876182e-07, + "loss": 0.0629, + "step": 41061 + }, + { + "epoch": 0.9625579446427212, + "grad_norm": 0.17169152200222015, + "learning_rate": 6.911319713966414e-07, + "loss": 0.0291, + "step": 41062 + }, + { + "epoch": 0.9625813862175261, + "grad_norm": 0.614987313747406, + "learning_rate": 6.902677013583825e-07, + "loss": 0.1304, + "step": 41063 + }, + { + "epoch": 0.9626048277923311, + "grad_norm": 0.19944779574871063, + "learning_rate": 6.894039701775268e-07, + "loss": 0.0283, + "step": 41064 + }, + { + "epoch": 0.9626282693671361, + "grad_norm": 0.7248570322990417, + "learning_rate": 6.885407778587482e-07, + "loss": 0.1256, + "step": 41065 + }, + { + "epoch": 0.9626517109419411, + "grad_norm": 0.4330292344093323, + "learning_rate": 6.876781244067432e-07, + "loss": 0.0964, + "step": 41066 + }, + { + "epoch": 0.9626751525167461, + "grad_norm": 0.38206878304481506, + "learning_rate": 6.868160098261855e-07, + "loss": 0.0646, + "step": 41067 + }, + { + "epoch": 0.9626985940915511, + "grad_norm": 0.3940691351890564, + "learning_rate": 6.859544341217605e-07, + "loss": 0.0556, + "step": 41068 + }, + { + "epoch": 0.9627220356663561, + "grad_norm": 0.32629042863845825, + "learning_rate": 6.850933972981421e-07, + "loss": 0.0529, + "step": 41069 + }, + { + "epoch": 0.9627454772411611, + "grad_norm": 0.5778192281723022, + "learning_rate": 6.842328993599933e-07, + "loss": 0.1451, + "step": 41070 + }, + { + "epoch": 0.962768918815966, + "grad_norm": 0.31080687046051025, + "learning_rate": 6.833729403119993e-07, + "loss": 0.0378, + "step": 41071 + }, + { + "epoch": 0.9627923603907711, + "grad_norm": 0.3368004560470581, + "learning_rate": 6.825135201588117e-07, + "loss": 0.079, + "step": 41072 + }, + { + "epoch": 0.962815801965576, + "grad_norm": 0.3993781507015228, + "learning_rate": 6.816546389050826e-07, + "loss": 0.0562, + "step": 41073 + }, + { + "epoch": 0.9628392435403811, + "grad_norm": 0.4796711206436157, + "learning_rate": 6.80796296555497e-07, + "loss": 0.0816, + "step": 41074 + }, + { + "epoch": 0.962862685115186, + "grad_norm": 0.38486248254776, + "learning_rate": 6.799384931147179e-07, + "loss": 0.0454, + "step": 41075 + }, + { + "epoch": 0.9628861266899911, + "grad_norm": 0.41785621643066406, + "learning_rate": 6.790812285873749e-07, + "loss": 0.0885, + "step": 41076 + }, + { + "epoch": 0.962909568264796, + "grad_norm": 0.2989330291748047, + "learning_rate": 6.782245029781309e-07, + "loss": 0.0422, + "step": 41077 + }, + { + "epoch": 0.9629330098396011, + "grad_norm": 0.1481042504310608, + "learning_rate": 6.773683162916267e-07, + "loss": 0.0323, + "step": 41078 + }, + { + "epoch": 0.962956451414406, + "grad_norm": 0.40993934869766235, + "learning_rate": 6.765126685325252e-07, + "loss": 0.0297, + "step": 41079 + }, + { + "epoch": 0.962979892989211, + "grad_norm": 0.1487792432308197, + "learning_rate": 6.756575597054449e-07, + "loss": 0.0196, + "step": 41080 + }, + { + "epoch": 0.963003334564016, + "grad_norm": 0.11655662208795547, + "learning_rate": 6.748029898150488e-07, + "loss": 0.013, + "step": 41081 + }, + { + "epoch": 0.963026776138821, + "grad_norm": 0.3079022169113159, + "learning_rate": 6.739489588659664e-07, + "loss": 0.0362, + "step": 41082 + }, + { + "epoch": 0.963050217713626, + "grad_norm": 0.3115655183792114, + "learning_rate": 6.730954668628275e-07, + "loss": 0.0362, + "step": 41083 + }, + { + "epoch": 0.963073659288431, + "grad_norm": 0.4736233055591583, + "learning_rate": 6.722425138102728e-07, + "loss": 0.052, + "step": 41084 + }, + { + "epoch": 0.963097100863236, + "grad_norm": 0.3707461357116699, + "learning_rate": 6.713900997129096e-07, + "loss": 0.0933, + "step": 41085 + }, + { + "epoch": 0.963120542438041, + "grad_norm": 0.24560976028442383, + "learning_rate": 6.705382245753788e-07, + "loss": 0.0415, + "step": 41086 + }, + { + "epoch": 0.963143984012846, + "grad_norm": 0.49037042260169983, + "learning_rate": 6.696868884022988e-07, + "loss": 0.0845, + "step": 41087 + }, + { + "epoch": 0.963167425587651, + "grad_norm": 0.6603348851203918, + "learning_rate": 6.688360911982994e-07, + "loss": 0.1441, + "step": 41088 + }, + { + "epoch": 0.9631908671624559, + "grad_norm": 0.28770679235458374, + "learning_rate": 6.679858329679877e-07, + "loss": 0.0421, + "step": 41089 + }, + { + "epoch": 0.963214308737261, + "grad_norm": 0.22698476910591125, + "learning_rate": 6.671361137159715e-07, + "loss": 0.0276, + "step": 41090 + }, + { + "epoch": 0.9632377503120659, + "grad_norm": 0.47027334570884705, + "learning_rate": 6.662869334468802e-07, + "loss": 0.0567, + "step": 41091 + }, + { + "epoch": 0.963261191886871, + "grad_norm": 0.1821049600839615, + "learning_rate": 6.65438292165288e-07, + "loss": 0.0247, + "step": 41092 + }, + { + "epoch": 0.9632846334616759, + "grad_norm": 0.4241989552974701, + "learning_rate": 6.645901898758355e-07, + "loss": 0.0424, + "step": 41093 + }, + { + "epoch": 0.963308075036481, + "grad_norm": 0.5251614451408386, + "learning_rate": 6.637426265831081e-07, + "loss": 0.5959, + "step": 41094 + }, + { + "epoch": 0.9633315166112859, + "grad_norm": 0.3419330418109894, + "learning_rate": 6.628956022917132e-07, + "loss": 0.0517, + "step": 41095 + }, + { + "epoch": 0.963354958186091, + "grad_norm": 0.5926201343536377, + "learning_rate": 6.620491170062249e-07, + "loss": 0.1163, + "step": 41096 + }, + { + "epoch": 0.9633783997608959, + "grad_norm": 0.5348316431045532, + "learning_rate": 6.612031707312616e-07, + "loss": 0.0512, + "step": 41097 + }, + { + "epoch": 0.963401841335701, + "grad_norm": 0.10648605227470398, + "learning_rate": 6.603577634713975e-07, + "loss": 0.0189, + "step": 41098 + }, + { + "epoch": 0.9634252829105059, + "grad_norm": 0.26791587471961975, + "learning_rate": 6.595128952312291e-07, + "loss": 0.0315, + "step": 41099 + }, + { + "epoch": 0.9634487244853109, + "grad_norm": 0.20570893585681915, + "learning_rate": 6.586685660153302e-07, + "loss": 0.0174, + "step": 41100 + }, + { + "epoch": 0.9634721660601159, + "grad_norm": 0.7108896374702454, + "learning_rate": 6.578247758282974e-07, + "loss": 0.1429, + "step": 41101 + }, + { + "epoch": 0.9634956076349209, + "grad_norm": 0.258710652589798, + "learning_rate": 6.569815246746935e-07, + "loss": 0.0272, + "step": 41102 + }, + { + "epoch": 0.9635190492097259, + "grad_norm": 0.6197416186332703, + "learning_rate": 6.561388125591039e-07, + "loss": 0.1414, + "step": 41103 + }, + { + "epoch": 0.9635424907845309, + "grad_norm": 0.28377777338027954, + "learning_rate": 6.552966394861027e-07, + "loss": 0.0483, + "step": 41104 + }, + { + "epoch": 0.9635659323593359, + "grad_norm": 0.48044079542160034, + "learning_rate": 6.544550054602527e-07, + "loss": 0.054, + "step": 41105 + }, + { + "epoch": 0.9635893739341409, + "grad_norm": 0.5562681555747986, + "learning_rate": 6.536139104861172e-07, + "loss": 0.0829, + "step": 41106 + }, + { + "epoch": 0.9636128155089458, + "grad_norm": 0.3834061026573181, + "learning_rate": 6.527733545682812e-07, + "loss": 0.0698, + "step": 41107 + }, + { + "epoch": 0.9636362570837509, + "grad_norm": 0.3305455148220062, + "learning_rate": 6.519333377112746e-07, + "loss": 0.0429, + "step": 41108 + }, + { + "epoch": 0.9636596986585558, + "grad_norm": 0.5234754681587219, + "learning_rate": 6.510938599196714e-07, + "loss": 0.4581, + "step": 41109 + }, + { + "epoch": 0.9636831402333609, + "grad_norm": 0.3853833079338074, + "learning_rate": 6.502549211980346e-07, + "loss": 0.0857, + "step": 41110 + }, + { + "epoch": 0.9637065818081658, + "grad_norm": 0.10474301129579544, + "learning_rate": 6.494165215509051e-07, + "loss": 0.012, + "step": 41111 + }, + { + "epoch": 0.9637300233829709, + "grad_norm": 0.48074549436569214, + "learning_rate": 6.485786609828348e-07, + "loss": 0.051, + "step": 41112 + }, + { + "epoch": 0.9637534649577759, + "grad_norm": 0.12105169892311096, + "learning_rate": 6.477413394983755e-07, + "loss": 0.0192, + "step": 41113 + }, + { + "epoch": 0.9637769065325809, + "grad_norm": 0.3584354519844055, + "learning_rate": 6.469045571020571e-07, + "loss": 0.0667, + "step": 41114 + }, + { + "epoch": 0.9638003481073859, + "grad_norm": 0.6223691701889038, + "learning_rate": 6.460683137984313e-07, + "loss": 0.0595, + "step": 41115 + }, + { + "epoch": 0.9638237896821908, + "grad_norm": 0.4703941345214844, + "learning_rate": 6.452326095920391e-07, + "loss": 0.0897, + "step": 41116 + }, + { + "epoch": 0.9638472312569959, + "grad_norm": 0.1342398226261139, + "learning_rate": 6.44397444487399e-07, + "loss": 0.0096, + "step": 41117 + }, + { + "epoch": 0.9638706728318008, + "grad_norm": 0.4690491259098053, + "learning_rate": 6.435628184890518e-07, + "loss": 0.0537, + "step": 41118 + }, + { + "epoch": 0.9638941144066059, + "grad_norm": 0.5881820321083069, + "learning_rate": 6.427287316015274e-07, + "loss": 0.0699, + "step": 41119 + }, + { + "epoch": 0.9639175559814108, + "grad_norm": 0.2966492474079132, + "learning_rate": 6.418951838293552e-07, + "loss": 0.0376, + "step": 41120 + }, + { + "epoch": 0.9639409975562159, + "grad_norm": 0.5833913683891296, + "learning_rate": 6.41062175177054e-07, + "loss": 0.0954, + "step": 41121 + }, + { + "epoch": 0.9639644391310208, + "grad_norm": 0.1366400420665741, + "learning_rate": 6.402297056491313e-07, + "loss": 0.0228, + "step": 41122 + }, + { + "epoch": 0.9639878807058259, + "grad_norm": 0.19365929067134857, + "learning_rate": 6.393977752501279e-07, + "loss": 0.0234, + "step": 41123 + }, + { + "epoch": 0.9640113222806308, + "grad_norm": 0.36351990699768066, + "learning_rate": 6.385663839845402e-07, + "loss": 0.0605, + "step": 41124 + }, + { + "epoch": 0.9640347638554359, + "grad_norm": 0.18950404226779938, + "learning_rate": 6.377355318568867e-07, + "loss": 0.0196, + "step": 41125 + }, + { + "epoch": 0.9640582054302408, + "grad_norm": 0.30811721086502075, + "learning_rate": 6.369052188716751e-07, + "loss": 0.0398, + "step": 41126 + }, + { + "epoch": 0.9640816470050458, + "grad_norm": 0.26081305742263794, + "learning_rate": 6.360754450334128e-07, + "loss": 0.0437, + "step": 41127 + }, + { + "epoch": 0.9641050885798508, + "grad_norm": 0.7776021361351013, + "learning_rate": 6.352462103465961e-07, + "loss": 0.1762, + "step": 41128 + }, + { + "epoch": 0.9641285301546558, + "grad_norm": 0.08582865446805954, + "learning_rate": 6.344175148157328e-07, + "loss": 0.0103, + "step": 41129 + }, + { + "epoch": 0.9641519717294608, + "grad_norm": 0.7434211373329163, + "learning_rate": 6.335893584452968e-07, + "loss": 0.1891, + "step": 41130 + }, + { + "epoch": 0.9641754133042658, + "grad_norm": 0.09807911515235901, + "learning_rate": 6.327617412398179e-07, + "loss": 0.0139, + "step": 41131 + }, + { + "epoch": 0.9641988548790708, + "grad_norm": 0.7193595767021179, + "learning_rate": 6.319346632037593e-07, + "loss": 0.1256, + "step": 41132 + }, + { + "epoch": 0.9642222964538758, + "grad_norm": 0.3077755570411682, + "learning_rate": 6.311081243416173e-07, + "loss": 0.0465, + "step": 41133 + }, + { + "epoch": 0.9642457380286807, + "grad_norm": 0.8419116139411926, + "learning_rate": 6.302821246578771e-07, + "loss": 0.1221, + "step": 41134 + }, + { + "epoch": 0.9642691796034858, + "grad_norm": 0.44319844245910645, + "learning_rate": 6.294566641570132e-07, + "loss": 0.0614, + "step": 41135 + }, + { + "epoch": 0.9642926211782907, + "grad_norm": 0.24758552014827728, + "learning_rate": 6.286317428435217e-07, + "loss": 0.0344, + "step": 41136 + }, + { + "epoch": 0.9643160627530958, + "grad_norm": 0.3003290295600891, + "learning_rate": 6.27807360721866e-07, + "loss": 0.0315, + "step": 41137 + }, + { + "epoch": 0.9643395043279007, + "grad_norm": 0.4325200915336609, + "learning_rate": 6.269835177965199e-07, + "loss": 0.0753, + "step": 41138 + }, + { + "epoch": 0.9643629459027058, + "grad_norm": 0.4371977150440216, + "learning_rate": 6.26160214071958e-07, + "loss": 0.0508, + "step": 41139 + }, + { + "epoch": 0.9643863874775107, + "grad_norm": 0.17182476818561554, + "learning_rate": 6.253374495526431e-07, + "loss": 0.0258, + "step": 41140 + }, + { + "epoch": 0.9644098290523158, + "grad_norm": 0.8045862317085266, + "learning_rate": 6.245152242430385e-07, + "loss": 0.1508, + "step": 41141 + }, + { + "epoch": 0.9644332706271207, + "grad_norm": 0.37467893958091736, + "learning_rate": 6.236935381476184e-07, + "loss": 0.0409, + "step": 41142 + }, + { + "epoch": 0.9644567122019257, + "grad_norm": 0.08732177317142487, + "learning_rate": 6.228723912708234e-07, + "loss": 0.0089, + "step": 41143 + }, + { + "epoch": 0.9644801537767307, + "grad_norm": 0.5227794051170349, + "learning_rate": 6.220517836171169e-07, + "loss": 0.0657, + "step": 41144 + }, + { + "epoch": 0.9645035953515357, + "grad_norm": 0.35410335659980774, + "learning_rate": 6.212317151909508e-07, + "loss": 0.083, + "step": 41145 + }, + { + "epoch": 0.9645270369263407, + "grad_norm": 0.29042544960975647, + "learning_rate": 6.204121859967771e-07, + "loss": 0.034, + "step": 41146 + }, + { + "epoch": 0.9645504785011457, + "grad_norm": 0.5841838121414185, + "learning_rate": 6.195931960390477e-07, + "loss": 0.0852, + "step": 41147 + }, + { + "epoch": 0.9645739200759507, + "grad_norm": 0.41083070635795593, + "learning_rate": 6.187747453221926e-07, + "loss": 0.085, + "step": 41148 + }, + { + "epoch": 0.9645973616507557, + "grad_norm": 0.41009587049484253, + "learning_rate": 6.179568338506636e-07, + "loss": 0.1015, + "step": 41149 + }, + { + "epoch": 0.9646208032255607, + "grad_norm": 0.5859571695327759, + "learning_rate": 6.171394616288906e-07, + "loss": 0.5727, + "step": 41150 + }, + { + "epoch": 0.9646442448003657, + "grad_norm": 0.46487799286842346, + "learning_rate": 6.163226286613144e-07, + "loss": 0.1111, + "step": 41151 + }, + { + "epoch": 0.9646676863751706, + "grad_norm": 0.7435576319694519, + "learning_rate": 6.155063349523649e-07, + "loss": 0.1437, + "step": 41152 + }, + { + "epoch": 0.9646911279499757, + "grad_norm": 0.43962910771369934, + "learning_rate": 6.146905805064829e-07, + "loss": 0.1203, + "step": 41153 + }, + { + "epoch": 0.9647145695247806, + "grad_norm": 0.29307475686073303, + "learning_rate": 6.13875365328076e-07, + "loss": 0.0515, + "step": 41154 + }, + { + "epoch": 0.9647380110995857, + "grad_norm": 0.310674250125885, + "learning_rate": 6.13060689421574e-07, + "loss": 0.0501, + "step": 41155 + }, + { + "epoch": 0.9647614526743906, + "grad_norm": 0.31852632761001587, + "learning_rate": 6.122465527913956e-07, + "loss": 0.027, + "step": 41156 + }, + { + "epoch": 0.9647848942491957, + "grad_norm": 0.4197334051132202, + "learning_rate": 6.114329554419707e-07, + "loss": 0.0707, + "step": 41157 + }, + { + "epoch": 0.9648083358240006, + "grad_norm": 0.20773327350616455, + "learning_rate": 6.106198973776956e-07, + "loss": 0.0326, + "step": 41158 + }, + { + "epoch": 0.9648317773988057, + "grad_norm": 0.30478689074516296, + "learning_rate": 6.09807378603e-07, + "loss": 0.058, + "step": 41159 + }, + { + "epoch": 0.9648552189736106, + "grad_norm": 0.3706792891025543, + "learning_rate": 6.089953991222919e-07, + "loss": 0.0656, + "step": 41160 + }, + { + "epoch": 0.9648786605484156, + "grad_norm": 0.13867716491222382, + "learning_rate": 6.081839589399563e-07, + "loss": 0.0289, + "step": 41161 + }, + { + "epoch": 0.9649021021232206, + "grad_norm": 0.16743813455104828, + "learning_rate": 6.07373058060412e-07, + "loss": 0.0247, + "step": 41162 + }, + { + "epoch": 0.9649255436980256, + "grad_norm": 0.367486834526062, + "learning_rate": 6.065626964880556e-07, + "loss": 0.0827, + "step": 41163 + }, + { + "epoch": 0.9649489852728307, + "grad_norm": 0.08342266827821732, + "learning_rate": 6.057528742272833e-07, + "loss": 0.0152, + "step": 41164 + }, + { + "epoch": 0.9649724268476356, + "grad_norm": 0.5517120361328125, + "learning_rate": 6.04943591282503e-07, + "loss": 0.0826, + "step": 41165 + }, + { + "epoch": 0.9649958684224407, + "grad_norm": 0.4153989851474762, + "learning_rate": 6.041348476580777e-07, + "loss": 0.1038, + "step": 41166 + }, + { + "epoch": 0.9650193099972456, + "grad_norm": 0.37086930871009827, + "learning_rate": 6.033266433584261e-07, + "loss": 0.0428, + "step": 41167 + }, + { + "epoch": 0.9650427515720507, + "grad_norm": 0.29912644624710083, + "learning_rate": 6.025189783879004e-07, + "loss": 0.0524, + "step": 41168 + }, + { + "epoch": 0.9650661931468556, + "grad_norm": 0.11306340247392654, + "learning_rate": 6.017118527509192e-07, + "loss": 0.0129, + "step": 41169 + }, + { + "epoch": 0.9650896347216606, + "grad_norm": 0.21553169190883636, + "learning_rate": 6.009052664518233e-07, + "loss": 0.0179, + "step": 41170 + }, + { + "epoch": 0.9651130762964656, + "grad_norm": 0.5003157258033752, + "learning_rate": 6.000992194950317e-07, + "loss": 0.1201, + "step": 41171 + }, + { + "epoch": 0.9651365178712706, + "grad_norm": 0.13453969359397888, + "learning_rate": 5.992937118848851e-07, + "loss": 0.02, + "step": 41172 + }, + { + "epoch": 0.9651599594460756, + "grad_norm": 0.47691771388053894, + "learning_rate": 5.984887436257691e-07, + "loss": 0.083, + "step": 41173 + }, + { + "epoch": 0.9651834010208806, + "grad_norm": 0.09860862046480179, + "learning_rate": 5.976843147220579e-07, + "loss": 0.0089, + "step": 41174 + }, + { + "epoch": 0.9652068425956856, + "grad_norm": 0.16991035640239716, + "learning_rate": 5.968804251780924e-07, + "loss": 0.0183, + "step": 41175 + }, + { + "epoch": 0.9652302841704906, + "grad_norm": 0.2568972408771515, + "learning_rate": 5.960770749982469e-07, + "loss": 0.0214, + "step": 41176 + }, + { + "epoch": 0.9652537257452956, + "grad_norm": 0.5225445032119751, + "learning_rate": 5.952742641868958e-07, + "loss": 0.0779, + "step": 41177 + }, + { + "epoch": 0.9652771673201006, + "grad_norm": 0.39929234981536865, + "learning_rate": 5.944719927483689e-07, + "loss": 0.0633, + "step": 41178 + }, + { + "epoch": 0.9653006088949055, + "grad_norm": 0.47070279717445374, + "learning_rate": 5.936702606870403e-07, + "loss": 0.4608, + "step": 41179 + }, + { + "epoch": 0.9653240504697106, + "grad_norm": 0.5426129698753357, + "learning_rate": 5.928690680072512e-07, + "loss": 0.1154, + "step": 41180 + }, + { + "epoch": 0.9653474920445155, + "grad_norm": 0.22442808747291565, + "learning_rate": 5.920684147133427e-07, + "loss": 0.035, + "step": 41181 + }, + { + "epoch": 0.9653709336193206, + "grad_norm": 0.5517731308937073, + "learning_rate": 5.912683008096665e-07, + "loss": 0.146, + "step": 41182 + }, + { + "epoch": 0.9653943751941255, + "grad_norm": 0.39818060398101807, + "learning_rate": 5.904687263005748e-07, + "loss": 0.0454, + "step": 41183 + }, + { + "epoch": 0.9654178167689306, + "grad_norm": 0.7936476469039917, + "learning_rate": 5.896696911903865e-07, + "loss": 0.1214, + "step": 41184 + }, + { + "epoch": 0.9654412583437355, + "grad_norm": 0.43059417605400085, + "learning_rate": 5.888711954834536e-07, + "loss": 0.0573, + "step": 41185 + }, + { + "epoch": 0.9654646999185406, + "grad_norm": 0.2528657615184784, + "learning_rate": 5.880732391840948e-07, + "loss": 0.0429, + "step": 41186 + }, + { + "epoch": 0.9654881414933455, + "grad_norm": 0.09197243303060532, + "learning_rate": 5.872758222966401e-07, + "loss": 0.0106, + "step": 41187 + }, + { + "epoch": 0.9655115830681505, + "grad_norm": 0.48941770195961, + "learning_rate": 5.864789448254304e-07, + "loss": 0.0477, + "step": 41188 + }, + { + "epoch": 0.9655350246429555, + "grad_norm": 0.16351093351840973, + "learning_rate": 5.856826067747734e-07, + "loss": 0.0211, + "step": 41189 + }, + { + "epoch": 0.9655584662177605, + "grad_norm": 0.18247397243976593, + "learning_rate": 5.848868081490099e-07, + "loss": 0.0394, + "step": 41190 + }, + { + "epoch": 0.9655819077925655, + "grad_norm": 0.2634810209274292, + "learning_rate": 5.840915489524368e-07, + "loss": 0.0428, + "step": 41191 + }, + { + "epoch": 0.9656053493673705, + "grad_norm": 0.42236098647117615, + "learning_rate": 5.832968291893837e-07, + "loss": 0.0746, + "step": 41192 + }, + { + "epoch": 0.9656287909421755, + "grad_norm": 0.34784120321273804, + "learning_rate": 5.825026488641471e-07, + "loss": 0.0738, + "step": 41193 + }, + { + "epoch": 0.9656522325169805, + "grad_norm": 0.22497296333312988, + "learning_rate": 5.817090079810572e-07, + "loss": 0.0405, + "step": 41194 + }, + { + "epoch": 0.9656756740917855, + "grad_norm": 0.23738688230514526, + "learning_rate": 5.809159065443992e-07, + "loss": 0.049, + "step": 41195 + }, + { + "epoch": 0.9656991156665905, + "grad_norm": 0.2991088330745697, + "learning_rate": 5.80123344558492e-07, + "loss": 0.0646, + "step": 41196 + }, + { + "epoch": 0.9657225572413954, + "grad_norm": 0.468295156955719, + "learning_rate": 5.793313220276319e-07, + "loss": 0.4944, + "step": 41197 + }, + { + "epoch": 0.9657459988162005, + "grad_norm": 0.3240731656551361, + "learning_rate": 5.785398389561159e-07, + "loss": 0.0411, + "step": 41198 + }, + { + "epoch": 0.9657694403910054, + "grad_norm": 0.2044500857591629, + "learning_rate": 5.777488953482401e-07, + "loss": 0.037, + "step": 41199 + }, + { + "epoch": 0.9657928819658105, + "grad_norm": 0.502211332321167, + "learning_rate": 5.769584912082904e-07, + "loss": 0.0476, + "step": 41200 + }, + { + "epoch": 0.9658163235406154, + "grad_norm": 0.2138679325580597, + "learning_rate": 5.76168626540563e-07, + "loss": 0.0286, + "step": 41201 + }, + { + "epoch": 0.9658397651154205, + "grad_norm": 0.2553774118423462, + "learning_rate": 5.753793013493436e-07, + "loss": 0.0353, + "step": 41202 + }, + { + "epoch": 0.9658632066902254, + "grad_norm": 0.4025675654411316, + "learning_rate": 5.745905156389064e-07, + "loss": 0.0443, + "step": 41203 + }, + { + "epoch": 0.9658866482650305, + "grad_norm": 0.32372573018074036, + "learning_rate": 5.73802269413537e-07, + "loss": 0.0489, + "step": 41204 + }, + { + "epoch": 0.9659100898398354, + "grad_norm": 0.19383291900157928, + "learning_rate": 5.730145626775096e-07, + "loss": 0.0464, + "step": 41205 + }, + { + "epoch": 0.9659335314146404, + "grad_norm": 0.6244617700576782, + "learning_rate": 5.722273954350988e-07, + "loss": 0.1107, + "step": 41206 + }, + { + "epoch": 0.9659569729894454, + "grad_norm": 0.2212146669626236, + "learning_rate": 5.714407676905898e-07, + "loss": 0.048, + "step": 41207 + }, + { + "epoch": 0.9659804145642504, + "grad_norm": 0.6249246001243591, + "learning_rate": 5.706546794482237e-07, + "loss": 0.1377, + "step": 41208 + }, + { + "epoch": 0.9660038561390554, + "grad_norm": 0.29800570011138916, + "learning_rate": 5.698691307122861e-07, + "loss": 0.0534, + "step": 41209 + }, + { + "epoch": 0.9660272977138604, + "grad_norm": 0.3854096531867981, + "learning_rate": 5.690841214870401e-07, + "loss": 0.0856, + "step": 41210 + }, + { + "epoch": 0.9660507392886654, + "grad_norm": 0.22249366343021393, + "learning_rate": 5.68299651776738e-07, + "loss": 0.0276, + "step": 41211 + }, + { + "epoch": 0.9660741808634704, + "grad_norm": 0.2606607973575592, + "learning_rate": 5.675157215856431e-07, + "loss": 0.0787, + "step": 41212 + }, + { + "epoch": 0.9660976224382753, + "grad_norm": 0.3887190818786621, + "learning_rate": 5.667323309179851e-07, + "loss": 0.0422, + "step": 41213 + }, + { + "epoch": 0.9661210640130804, + "grad_norm": 0.2455066293478012, + "learning_rate": 5.659494797780496e-07, + "loss": 0.0319, + "step": 41214 + }, + { + "epoch": 0.9661445055878854, + "grad_norm": 0.5350968837738037, + "learning_rate": 5.651671681700666e-07, + "loss": 0.1232, + "step": 41215 + }, + { + "epoch": 0.9661679471626904, + "grad_norm": 0.3931936025619507, + "learning_rate": 5.64385396098277e-07, + "loss": 0.0236, + "step": 41216 + }, + { + "epoch": 0.9661913887374954, + "grad_norm": 0.45232000946998596, + "learning_rate": 5.636041635669331e-07, + "loss": 0.0997, + "step": 41217 + }, + { + "epoch": 0.9662148303123004, + "grad_norm": 0.324357807636261, + "learning_rate": 5.628234705802648e-07, + "loss": 0.0555, + "step": 41218 + }, + { + "epoch": 0.9662382718871054, + "grad_norm": 0.3396627902984619, + "learning_rate": 5.620433171425133e-07, + "loss": 0.0741, + "step": 41219 + }, + { + "epoch": 0.9662617134619104, + "grad_norm": 0.20497240126132965, + "learning_rate": 5.612637032579083e-07, + "loss": 0.035, + "step": 41220 + }, + { + "epoch": 0.9662851550367154, + "grad_norm": 0.14258311688899994, + "learning_rate": 5.604846289306798e-07, + "loss": 0.0198, + "step": 41221 + }, + { + "epoch": 0.9663085966115204, + "grad_norm": 0.5419883728027344, + "learning_rate": 5.597060941650689e-07, + "loss": 0.0927, + "step": 41222 + }, + { + "epoch": 0.9663320381863254, + "grad_norm": 0.49703487753868103, + "learning_rate": 5.589280989652834e-07, + "loss": 0.0583, + "step": 41223 + }, + { + "epoch": 0.9663554797611303, + "grad_norm": 0.22549642622470856, + "learning_rate": 5.581506433355533e-07, + "loss": 0.0368, + "step": 41224 + }, + { + "epoch": 0.9663789213359354, + "grad_norm": 0.4450666904449463, + "learning_rate": 5.573737272800861e-07, + "loss": 0.0653, + "step": 41225 + }, + { + "epoch": 0.9664023629107403, + "grad_norm": 0.4485720992088318, + "learning_rate": 5.565973508031009e-07, + "loss": 0.084, + "step": 41226 + }, + { + "epoch": 0.9664258044855454, + "grad_norm": 0.44289860129356384, + "learning_rate": 5.558215139088274e-07, + "loss": 0.0316, + "step": 41227 + }, + { + "epoch": 0.9664492460603503, + "grad_norm": 0.5102439522743225, + "learning_rate": 5.550462166014624e-07, + "loss": 0.0736, + "step": 41228 + }, + { + "epoch": 0.9664726876351554, + "grad_norm": 0.2583935558795929, + "learning_rate": 5.542714588852027e-07, + "loss": 0.0157, + "step": 41229 + }, + { + "epoch": 0.9664961292099603, + "grad_norm": 0.5268718600273132, + "learning_rate": 5.534972407642669e-07, + "loss": 0.0972, + "step": 41230 + }, + { + "epoch": 0.9665195707847654, + "grad_norm": 0.3271876871585846, + "learning_rate": 5.527235622428517e-07, + "loss": 0.0239, + "step": 41231 + }, + { + "epoch": 0.9665430123595703, + "grad_norm": 0.16802392899990082, + "learning_rate": 5.519504233251538e-07, + "loss": 0.0245, + "step": 41232 + }, + { + "epoch": 0.9665664539343753, + "grad_norm": 0.3596118986606598, + "learning_rate": 5.511778240153698e-07, + "loss": 0.3406, + "step": 41233 + }, + { + "epoch": 0.9665898955091803, + "grad_norm": 0.17810474336147308, + "learning_rate": 5.504057643176963e-07, + "loss": 0.0321, + "step": 41234 + }, + { + "epoch": 0.9666133370839853, + "grad_norm": 0.36512061953544617, + "learning_rate": 5.496342442363078e-07, + "loss": 0.0473, + "step": 41235 + }, + { + "epoch": 0.9666367786587903, + "grad_norm": 0.38174968957901, + "learning_rate": 5.488632637754121e-07, + "loss": 0.0768, + "step": 41236 + }, + { + "epoch": 0.9666602202335953, + "grad_norm": 0.4260251820087433, + "learning_rate": 5.480928229391724e-07, + "loss": 0.0792, + "step": 41237 + }, + { + "epoch": 0.9666836618084003, + "grad_norm": 0.27225109934806824, + "learning_rate": 5.473229217317855e-07, + "loss": 0.0338, + "step": 41238 + }, + { + "epoch": 0.9667071033832053, + "grad_norm": 0.3928528130054474, + "learning_rate": 5.465535601574035e-07, + "loss": 0.1035, + "step": 41239 + }, + { + "epoch": 0.9667305449580103, + "grad_norm": 0.17043901979923248, + "learning_rate": 5.457847382202341e-07, + "loss": 0.0355, + "step": 41240 + }, + { + "epoch": 0.9667539865328153, + "grad_norm": 0.3162928819656372, + "learning_rate": 5.450164559244298e-07, + "loss": 0.0657, + "step": 41241 + }, + { + "epoch": 0.9667774281076202, + "grad_norm": 0.1821894347667694, + "learning_rate": 5.442487132741537e-07, + "loss": 0.0427, + "step": 41242 + }, + { + "epoch": 0.9668008696824253, + "grad_norm": 0.13350090384483337, + "learning_rate": 5.434815102735912e-07, + "loss": 0.0149, + "step": 41243 + }, + { + "epoch": 0.9668243112572302, + "grad_norm": 0.4859611988067627, + "learning_rate": 5.427148469268839e-07, + "loss": 0.0639, + "step": 41244 + }, + { + "epoch": 0.9668477528320353, + "grad_norm": 0.1663658320903778, + "learning_rate": 5.419487232382059e-07, + "loss": 0.019, + "step": 41245 + }, + { + "epoch": 0.9668711944068402, + "grad_norm": 0.2015194296836853, + "learning_rate": 5.411831392117206e-07, + "loss": 0.0301, + "step": 41246 + }, + { + "epoch": 0.9668946359816453, + "grad_norm": 0.10099203139543533, + "learning_rate": 5.404180948515692e-07, + "loss": 0.0112, + "step": 41247 + }, + { + "epoch": 0.9669180775564502, + "grad_norm": 1.0179215669631958, + "learning_rate": 5.396535901619038e-07, + "loss": 0.1455, + "step": 41248 + }, + { + "epoch": 0.9669415191312553, + "grad_norm": 0.408545583486557, + "learning_rate": 5.388896251468656e-07, + "loss": 0.0355, + "step": 41249 + }, + { + "epoch": 0.9669649607060602, + "grad_norm": 0.2837847173213959, + "learning_rate": 5.38126199810618e-07, + "loss": 0.0609, + "step": 41250 + }, + { + "epoch": 0.9669884022808652, + "grad_norm": 0.300567090511322, + "learning_rate": 5.373633141572909e-07, + "loss": 0.0402, + "step": 41251 + }, + { + "epoch": 0.9670118438556702, + "grad_norm": 0.6429770588874817, + "learning_rate": 5.366009681910256e-07, + "loss": 0.0627, + "step": 41252 + }, + { + "epoch": 0.9670352854304752, + "grad_norm": 0.28750747442245483, + "learning_rate": 5.358391619159742e-07, + "loss": 0.0429, + "step": 41253 + }, + { + "epoch": 0.9670587270052802, + "grad_norm": 0.09573154896497726, + "learning_rate": 5.350778953362445e-07, + "loss": 0.0131, + "step": 41254 + }, + { + "epoch": 0.9670821685800852, + "grad_norm": 0.5138252973556519, + "learning_rate": 5.343171684559778e-07, + "loss": 0.0444, + "step": 41255 + }, + { + "epoch": 0.9671056101548902, + "grad_norm": 0.2751423120498657, + "learning_rate": 5.33556981279304e-07, + "loss": 0.0552, + "step": 41256 + }, + { + "epoch": 0.9671290517296952, + "grad_norm": 0.21714474260807037, + "learning_rate": 5.32797333810342e-07, + "loss": 0.0286, + "step": 41257 + }, + { + "epoch": 0.9671524933045001, + "grad_norm": 0.2478310763835907, + "learning_rate": 5.32038226053222e-07, + "loss": 0.0542, + "step": 41258 + }, + { + "epoch": 0.9671759348793052, + "grad_norm": 0.2976873815059662, + "learning_rate": 5.312796580120627e-07, + "loss": 0.0435, + "step": 41259 + }, + { + "epoch": 0.9671993764541101, + "grad_norm": 0.5245452523231506, + "learning_rate": 5.30521629690972e-07, + "loss": 0.0276, + "step": 41260 + }, + { + "epoch": 0.9672228180289152, + "grad_norm": 0.403109073638916, + "learning_rate": 5.297641410940801e-07, + "loss": 0.0713, + "step": 41261 + }, + { + "epoch": 0.9672462596037201, + "grad_norm": 0.04975743964314461, + "learning_rate": 5.290071922254724e-07, + "loss": 0.0093, + "step": 41262 + }, + { + "epoch": 0.9672697011785252, + "grad_norm": 0.4014016389846802, + "learning_rate": 5.28250783089268e-07, + "loss": 0.0364, + "step": 41263 + }, + { + "epoch": 0.9672931427533301, + "grad_norm": 0.7781665921211243, + "learning_rate": 5.274949136895746e-07, + "loss": 0.5572, + "step": 41264 + }, + { + "epoch": 0.9673165843281352, + "grad_norm": 0.3946179449558258, + "learning_rate": 5.267395840305001e-07, + "loss": 0.0346, + "step": 41265 + }, + { + "epoch": 0.9673400259029402, + "grad_norm": 0.5001387000083923, + "learning_rate": 5.259847941161189e-07, + "loss": 0.7562, + "step": 41266 + }, + { + "epoch": 0.9673634674777452, + "grad_norm": 0.071086585521698, + "learning_rate": 5.2523054395055e-07, + "loss": 0.0107, + "step": 41267 + }, + { + "epoch": 0.9673869090525502, + "grad_norm": 0.23364657163619995, + "learning_rate": 5.24476833537868e-07, + "loss": 0.041, + "step": 41268 + }, + { + "epoch": 0.9674103506273551, + "grad_norm": 0.183534637093544, + "learning_rate": 5.237236628821807e-07, + "loss": 0.0418, + "step": 41269 + }, + { + "epoch": 0.9674337922021602, + "grad_norm": 0.6093235015869141, + "learning_rate": 5.229710319875624e-07, + "loss": 0.0558, + "step": 41270 + }, + { + "epoch": 0.9674572337769651, + "grad_norm": 0.1713898926973343, + "learning_rate": 5.222189408580991e-07, + "loss": 0.0392, + "step": 41271 + }, + { + "epoch": 0.9674806753517702, + "grad_norm": 0.7659431099891663, + "learning_rate": 5.214673894978761e-07, + "loss": 0.1426, + "step": 41272 + }, + { + "epoch": 0.9675041169265751, + "grad_norm": 0.2207145392894745, + "learning_rate": 5.20716377910968e-07, + "loss": 0.043, + "step": 41273 + }, + { + "epoch": 0.9675275585013802, + "grad_norm": 0.46573010087013245, + "learning_rate": 5.199659061014494e-07, + "loss": 0.087, + "step": 41274 + }, + { + "epoch": 0.9675510000761851, + "grad_norm": 0.641559362411499, + "learning_rate": 5.192159740733948e-07, + "loss": 0.1004, + "step": 41275 + }, + { + "epoch": 0.9675744416509902, + "grad_norm": 0.3199508488178253, + "learning_rate": 5.184665818308788e-07, + "loss": 0.0233, + "step": 41276 + }, + { + "epoch": 0.9675978832257951, + "grad_norm": 0.3719087839126587, + "learning_rate": 5.177177293779423e-07, + "loss": 0.0354, + "step": 41277 + }, + { + "epoch": 0.9676213248006001, + "grad_norm": 0.7348541021347046, + "learning_rate": 5.169694167186711e-07, + "loss": 0.0886, + "step": 41278 + }, + { + "epoch": 0.9676447663754051, + "grad_norm": 0.24123091995716095, + "learning_rate": 5.162216438571288e-07, + "loss": 0.0327, + "step": 41279 + }, + { + "epoch": 0.9676682079502101, + "grad_norm": 0.20961758494377136, + "learning_rate": 5.154744107973564e-07, + "loss": 0.0228, + "step": 41280 + }, + { + "epoch": 0.9676916495250151, + "grad_norm": 0.11701122671365738, + "learning_rate": 5.147277175434284e-07, + "loss": 0.0148, + "step": 41281 + }, + { + "epoch": 0.9677150910998201, + "grad_norm": 0.4051367938518524, + "learning_rate": 5.139815640993751e-07, + "loss": 0.0487, + "step": 41282 + }, + { + "epoch": 0.9677385326746251, + "grad_norm": 0.4351285994052887, + "learning_rate": 5.132359504692596e-07, + "loss": 0.5698, + "step": 41283 + }, + { + "epoch": 0.9677619742494301, + "grad_norm": 0.2644651234149933, + "learning_rate": 5.124908766571235e-07, + "loss": 0.0581, + "step": 41284 + }, + { + "epoch": 0.967785415824235, + "grad_norm": 0.44457677006721497, + "learning_rate": 5.117463426670077e-07, + "loss": 0.0659, + "step": 41285 + }, + { + "epoch": 0.9678088573990401, + "grad_norm": 0.5541166067123413, + "learning_rate": 5.110023485029536e-07, + "loss": 0.3118, + "step": 41286 + }, + { + "epoch": 0.967832298973845, + "grad_norm": 0.4590087831020355, + "learning_rate": 5.102588941690023e-07, + "loss": 0.0525, + "step": 41287 + }, + { + "epoch": 0.9678557405486501, + "grad_norm": 0.34144774079322815, + "learning_rate": 5.09515979669184e-07, + "loss": 0.0554, + "step": 41288 + }, + { + "epoch": 0.967879182123455, + "grad_norm": 0.4317806661128998, + "learning_rate": 5.087736050075176e-07, + "loss": 0.0159, + "step": 41289 + }, + { + "epoch": 0.9679026236982601, + "grad_norm": 0.20833706855773926, + "learning_rate": 5.080317701880555e-07, + "loss": 0.0347, + "step": 41290 + }, + { + "epoch": 0.967926065273065, + "grad_norm": 0.47841188311576843, + "learning_rate": 5.072904752148055e-07, + "loss": 0.0553, + "step": 41291 + }, + { + "epoch": 0.9679495068478701, + "grad_norm": 0.3293730318546295, + "learning_rate": 5.065497200917868e-07, + "loss": 0.0395, + "step": 41292 + }, + { + "epoch": 0.967972948422675, + "grad_norm": 0.5237672924995422, + "learning_rate": 5.058095048230405e-07, + "loss": 0.0554, + "step": 41293 + }, + { + "epoch": 0.9679963899974801, + "grad_norm": 0.5430865287780762, + "learning_rate": 5.050698294125634e-07, + "loss": 0.0679, + "step": 41294 + }, + { + "epoch": 0.968019831572285, + "grad_norm": 0.4662126302719116, + "learning_rate": 5.043306938643744e-07, + "loss": 0.02, + "step": 41295 + }, + { + "epoch": 0.96804327314709, + "grad_norm": 0.4286860525608063, + "learning_rate": 5.035920981824816e-07, + "loss": 0.1049, + "step": 41296 + }, + { + "epoch": 0.968066714721895, + "grad_norm": 0.5886991024017334, + "learning_rate": 5.028540423709038e-07, + "loss": 0.0589, + "step": 41297 + }, + { + "epoch": 0.9680901562967, + "grad_norm": 0.5786355137825012, + "learning_rate": 5.02116526433638e-07, + "loss": 0.0876, + "step": 41298 + }, + { + "epoch": 0.968113597871505, + "grad_norm": 0.3277825713157654, + "learning_rate": 5.01379550374681e-07, + "loss": 0.0306, + "step": 41299 + }, + { + "epoch": 0.96813703944631, + "grad_norm": 0.6092146635055542, + "learning_rate": 5.006431141980406e-07, + "loss": 0.1028, + "step": 41300 + }, + { + "epoch": 0.968160481021115, + "grad_norm": 0.463150292634964, + "learning_rate": 4.999072179077024e-07, + "loss": 0.0716, + "step": 41301 + }, + { + "epoch": 0.96818392259592, + "grad_norm": 0.549359142780304, + "learning_rate": 4.991718615076635e-07, + "loss": 0.6489, + "step": 41302 + }, + { + "epoch": 0.968207364170725, + "grad_norm": 0.10215679556131363, + "learning_rate": 4.984370450019204e-07, + "loss": 0.0166, + "step": 41303 + }, + { + "epoch": 0.96823080574553, + "grad_norm": 0.4072202742099762, + "learning_rate": 4.97702768394448e-07, + "loss": 0.0369, + "step": 41304 + }, + { + "epoch": 0.9682542473203349, + "grad_norm": 0.10305880010128021, + "learning_rate": 4.969690316892428e-07, + "loss": 0.0178, + "step": 41305 + }, + { + "epoch": 0.96827768889514, + "grad_norm": 0.13655781745910645, + "learning_rate": 4.962358348902907e-07, + "loss": 0.0202, + "step": 41306 + }, + { + "epoch": 0.9683011304699449, + "grad_norm": 0.5706242918968201, + "learning_rate": 4.955031780015551e-07, + "loss": 0.0884, + "step": 41307 + }, + { + "epoch": 0.96832457204475, + "grad_norm": 0.9415941834449768, + "learning_rate": 4.947710610270107e-07, + "loss": 0.1777, + "step": 41308 + }, + { + "epoch": 0.9683480136195549, + "grad_norm": 0.4161263406276703, + "learning_rate": 4.94039483970643e-07, + "loss": 0.0954, + "step": 41309 + }, + { + "epoch": 0.96837145519436, + "grad_norm": 0.7279808521270752, + "learning_rate": 4.933084468364157e-07, + "loss": 0.099, + "step": 41310 + }, + { + "epoch": 0.9683948967691649, + "grad_norm": 0.29611119627952576, + "learning_rate": 4.925779496282923e-07, + "loss": 0.0302, + "step": 41311 + }, + { + "epoch": 0.96841833834397, + "grad_norm": 0.1255689263343811, + "learning_rate": 4.918479923502362e-07, + "loss": 0.0077, + "step": 41312 + }, + { + "epoch": 0.9684417799187749, + "grad_norm": 0.6951481699943542, + "learning_rate": 4.911185750062109e-07, + "loss": 0.0935, + "step": 41313 + }, + { + "epoch": 0.9684652214935799, + "grad_norm": 0.3997479975223541, + "learning_rate": 4.903896976001798e-07, + "loss": 0.0384, + "step": 41314 + }, + { + "epoch": 0.9684886630683849, + "grad_norm": 0.45481276512145996, + "learning_rate": 4.896613601360845e-07, + "loss": 0.0909, + "step": 41315 + }, + { + "epoch": 0.9685121046431899, + "grad_norm": 0.6094151735305786, + "learning_rate": 4.889335626178882e-07, + "loss": 0.0679, + "step": 41316 + }, + { + "epoch": 0.968535546217995, + "grad_norm": 0.1459766924381256, + "learning_rate": 4.882063050495322e-07, + "loss": 0.0228, + "step": 41317 + }, + { + "epoch": 0.9685589877927999, + "grad_norm": 0.27324342727661133, + "learning_rate": 4.874795874349803e-07, + "loss": 0.0416, + "step": 41318 + }, + { + "epoch": 0.968582429367605, + "grad_norm": 0.2554875612258911, + "learning_rate": 4.867534097781512e-07, + "loss": 0.0335, + "step": 41319 + }, + { + "epoch": 0.9686058709424099, + "grad_norm": 0.6134104132652283, + "learning_rate": 4.860277720829975e-07, + "loss": 0.1532, + "step": 41320 + }, + { + "epoch": 0.968629312517215, + "grad_norm": 0.4067537784576416, + "learning_rate": 4.853026743534495e-07, + "loss": 0.0665, + "step": 41321 + }, + { + "epoch": 0.9686527540920199, + "grad_norm": 0.4077427387237549, + "learning_rate": 4.845781165934594e-07, + "loss": 0.0589, + "step": 41322 + }, + { + "epoch": 0.968676195666825, + "grad_norm": 0.6047583222389221, + "learning_rate": 4.838540988069462e-07, + "loss": 0.0569, + "step": 41323 + }, + { + "epoch": 0.9686996372416299, + "grad_norm": 0.2741324007511139, + "learning_rate": 4.831306209978293e-07, + "loss": 0.0451, + "step": 41324 + }, + { + "epoch": 0.9687230788164349, + "grad_norm": 0.2564946115016937, + "learning_rate": 4.824076831700496e-07, + "loss": 0.0353, + "step": 41325 + }, + { + "epoch": 0.9687465203912399, + "grad_norm": 0.5382590293884277, + "learning_rate": 4.816852853275266e-07, + "loss": 0.0956, + "step": 41326 + }, + { + "epoch": 0.9687699619660449, + "grad_norm": 0.3977481722831726, + "learning_rate": 4.809634274741681e-07, + "loss": 0.0696, + "step": 41327 + }, + { + "epoch": 0.9687934035408499, + "grad_norm": 0.13620907068252563, + "learning_rate": 4.802421096139042e-07, + "loss": 0.0247, + "step": 41328 + }, + { + "epoch": 0.9688168451156549, + "grad_norm": 0.5988713502883911, + "learning_rate": 4.79521331750643e-07, + "loss": 0.1022, + "step": 41329 + }, + { + "epoch": 0.9688402866904599, + "grad_norm": 0.3596166670322418, + "learning_rate": 4.788010938883036e-07, + "loss": 0.0279, + "step": 41330 + }, + { + "epoch": 0.9688637282652649, + "grad_norm": 0.15908204019069672, + "learning_rate": 4.780813960307829e-07, + "loss": 0.0208, + "step": 41331 + }, + { + "epoch": 0.9688871698400698, + "grad_norm": 0.45163360238075256, + "learning_rate": 4.773622381819997e-07, + "loss": 0.0662, + "step": 41332 + }, + { + "epoch": 0.9689106114148749, + "grad_norm": 0.5840075612068176, + "learning_rate": 4.766436203458402e-07, + "loss": 0.0661, + "step": 41333 + }, + { + "epoch": 0.9689340529896798, + "grad_norm": 0.09197814017534256, + "learning_rate": 4.7592554252621215e-07, + "loss": 0.01, + "step": 41334 + }, + { + "epoch": 0.9689574945644849, + "grad_norm": 0.22093789279460907, + "learning_rate": 4.7520800472700133e-07, + "loss": 0.0348, + "step": 41335 + }, + { + "epoch": 0.9689809361392898, + "grad_norm": 0.6612725853919983, + "learning_rate": 4.7449100695212687e-07, + "loss": 0.0594, + "step": 41336 + }, + { + "epoch": 0.9690043777140949, + "grad_norm": 0.1864078789949417, + "learning_rate": 4.7377454920545237e-07, + "loss": 0.0289, + "step": 41337 + }, + { + "epoch": 0.9690278192888998, + "grad_norm": 0.33839377760887146, + "learning_rate": 4.730586314908747e-07, + "loss": 0.0536, + "step": 41338 + }, + { + "epoch": 0.9690512608637049, + "grad_norm": 0.2017533928155899, + "learning_rate": 4.7234325381229073e-07, + "loss": 0.0327, + "step": 41339 + }, + { + "epoch": 0.9690747024385098, + "grad_norm": 0.568943977355957, + "learning_rate": 4.716284161735529e-07, + "loss": 0.1015, + "step": 41340 + }, + { + "epoch": 0.9690981440133148, + "grad_norm": 0.383348286151886, + "learning_rate": 4.709141185785693e-07, + "loss": 0.0702, + "step": 41341 + }, + { + "epoch": 0.9691215855881198, + "grad_norm": 0.798082172870636, + "learning_rate": 4.7020036103119225e-07, + "loss": 0.185, + "step": 41342 + }, + { + "epoch": 0.9691450271629248, + "grad_norm": 0.5070557594299316, + "learning_rate": 4.6948714353531875e-07, + "loss": 0.0879, + "step": 41343 + }, + { + "epoch": 0.9691684687377298, + "grad_norm": 0.11303811520338058, + "learning_rate": 4.6877446609481235e-07, + "loss": 0.0098, + "step": 41344 + }, + { + "epoch": 0.9691919103125348, + "grad_norm": 0.4483814239501953, + "learning_rate": 4.680623287135255e-07, + "loss": 0.0566, + "step": 41345 + }, + { + "epoch": 0.9692153518873398, + "grad_norm": 0.5634050965309143, + "learning_rate": 4.673507313953329e-07, + "loss": 0.0733, + "step": 41346 + }, + { + "epoch": 0.9692387934621448, + "grad_norm": 0.40448617935180664, + "learning_rate": 4.6663967414408706e-07, + "loss": 0.0441, + "step": 41347 + }, + { + "epoch": 0.9692622350369497, + "grad_norm": 0.6195502281188965, + "learning_rate": 4.6592915696366255e-07, + "loss": 0.0528, + "step": 41348 + }, + { + "epoch": 0.9692856766117548, + "grad_norm": 0.32257401943206787, + "learning_rate": 4.652191798579009e-07, + "loss": 0.0382, + "step": 41349 + }, + { + "epoch": 0.9693091181865597, + "grad_norm": 0.14308999478816986, + "learning_rate": 4.645097428306544e-07, + "loss": 0.0143, + "step": 41350 + }, + { + "epoch": 0.9693325597613648, + "grad_norm": 0.47544020414352417, + "learning_rate": 4.638008458857868e-07, + "loss": 0.0991, + "step": 41351 + }, + { + "epoch": 0.9693560013361697, + "grad_norm": 0.253190815448761, + "learning_rate": 4.630924890271171e-07, + "loss": 0.0216, + "step": 41352 + }, + { + "epoch": 0.9693794429109748, + "grad_norm": 0.5463657379150391, + "learning_rate": 4.623846722585201e-07, + "loss": 0.0894, + "step": 41353 + }, + { + "epoch": 0.9694028844857797, + "grad_norm": 0.4127645492553711, + "learning_rate": 4.6167739558381493e-07, + "loss": 0.0961, + "step": 41354 + }, + { + "epoch": 0.9694263260605848, + "grad_norm": 0.3445611596107483, + "learning_rate": 4.609706590068541e-07, + "loss": 0.0394, + "step": 41355 + }, + { + "epoch": 0.9694497676353897, + "grad_norm": 0.5053282976150513, + "learning_rate": 4.602644625314678e-07, + "loss": 0.7876, + "step": 41356 + }, + { + "epoch": 0.9694732092101948, + "grad_norm": 0.6626826524734497, + "learning_rate": 4.595588061614864e-07, + "loss": 0.0711, + "step": 41357 + }, + { + "epoch": 0.9694966507849997, + "grad_norm": 0.07366698980331421, + "learning_rate": 4.5885368990072896e-07, + "loss": 0.0041, + "step": 41358 + }, + { + "epoch": 0.9695200923598047, + "grad_norm": 0.4042973816394806, + "learning_rate": 4.581491137530258e-07, + "loss": 0.0548, + "step": 41359 + }, + { + "epoch": 0.9695435339346097, + "grad_norm": 0.24217036366462708, + "learning_rate": 4.5744507772221835e-07, + "loss": 0.0174, + "step": 41360 + }, + { + "epoch": 0.9695669755094147, + "grad_norm": 0.5251394510269165, + "learning_rate": 4.567415818121035e-07, + "loss": 0.5416, + "step": 41361 + }, + { + "epoch": 0.9695904170842197, + "grad_norm": 0.08806294947862625, + "learning_rate": 4.560386260265115e-07, + "loss": 0.0095, + "step": 41362 + }, + { + "epoch": 0.9696138586590247, + "grad_norm": 0.4596317410469055, + "learning_rate": 4.5533621036926154e-07, + "loss": 0.0591, + "step": 41363 + }, + { + "epoch": 0.9696373002338297, + "grad_norm": 0.21207347512245178, + "learning_rate": 4.546343348441395e-07, + "loss": 0.0418, + "step": 41364 + }, + { + "epoch": 0.9696607418086347, + "grad_norm": 0.3330165147781372, + "learning_rate": 4.539329994549757e-07, + "loss": 0.3549, + "step": 41365 + }, + { + "epoch": 0.9696841833834396, + "grad_norm": 0.21143899857997894, + "learning_rate": 4.53232204205567e-07, + "loss": 0.0258, + "step": 41366 + }, + { + "epoch": 0.9697076249582447, + "grad_norm": 0.3619228005409241, + "learning_rate": 4.5253194909973264e-07, + "loss": 0.0873, + "step": 41367 + }, + { + "epoch": 0.9697310665330496, + "grad_norm": 0.5329346060752869, + "learning_rate": 4.518322341412473e-07, + "loss": 0.0664, + "step": 41368 + }, + { + "epoch": 0.9697545081078547, + "grad_norm": 0.20877094566822052, + "learning_rate": 4.511330593339302e-07, + "loss": 0.0315, + "step": 41369 + }, + { + "epoch": 0.9697779496826597, + "grad_norm": 0.35806745290756226, + "learning_rate": 4.5043442468155617e-07, + "loss": 0.0638, + "step": 41370 + }, + { + "epoch": 0.9698013912574647, + "grad_norm": 0.30597180128097534, + "learning_rate": 4.4973633018792207e-07, + "loss": 0.0485, + "step": 41371 + }, + { + "epoch": 0.9698248328322697, + "grad_norm": 0.07875994592905045, + "learning_rate": 4.4903877585682487e-07, + "loss": 0.0136, + "step": 41372 + }, + { + "epoch": 0.9698482744070747, + "grad_norm": 1.136902928352356, + "learning_rate": 4.4834176169203935e-07, + "loss": 0.5802, + "step": 41373 + }, + { + "epoch": 0.9698717159818797, + "grad_norm": 0.9761365056037903, + "learning_rate": 4.476452876973514e-07, + "loss": 0.2029, + "step": 41374 + }, + { + "epoch": 0.9698951575566847, + "grad_norm": 0.09041246771812439, + "learning_rate": 4.469493538765357e-07, + "loss": 0.0082, + "step": 41375 + }, + { + "epoch": 0.9699185991314897, + "grad_norm": 0.24681280553340912, + "learning_rate": 4.4625396023337815e-07, + "loss": 0.0319, + "step": 41376 + }, + { + "epoch": 0.9699420407062946, + "grad_norm": 0.35770437121391296, + "learning_rate": 4.4555910677163136e-07, + "loss": 0.042, + "step": 41377 + }, + { + "epoch": 0.9699654822810997, + "grad_norm": 0.10212886333465576, + "learning_rate": 4.4486479349509224e-07, + "loss": 0.0134, + "step": 41378 + }, + { + "epoch": 0.9699889238559046, + "grad_norm": 0.3668799102306366, + "learning_rate": 4.4417102040751335e-07, + "loss": 0.0312, + "step": 41379 + }, + { + "epoch": 0.9700123654307097, + "grad_norm": 0.5023020505905151, + "learning_rate": 4.4347778751265834e-07, + "loss": 0.4801, + "step": 41380 + }, + { + "epoch": 0.9700358070055146, + "grad_norm": 0.5038196444511414, + "learning_rate": 4.4278509481430195e-07, + "loss": 0.0466, + "step": 41381 + }, + { + "epoch": 0.9700592485803197, + "grad_norm": 0.20853249728679657, + "learning_rate": 4.420929423161857e-07, + "loss": 0.028, + "step": 41382 + }, + { + "epoch": 0.9700826901551246, + "grad_norm": 0.09481845051050186, + "learning_rate": 4.414013300220732e-07, + "loss": 0.0229, + "step": 41383 + }, + { + "epoch": 0.9701061317299297, + "grad_norm": 0.3422118127346039, + "learning_rate": 4.4071025793571694e-07, + "loss": 0.0602, + "step": 41384 + }, + { + "epoch": 0.9701295733047346, + "grad_norm": 0.5718521475791931, + "learning_rate": 4.400197260608696e-07, + "loss": 0.0798, + "step": 41385 + }, + { + "epoch": 0.9701530148795396, + "grad_norm": 0.36417946219444275, + "learning_rate": 4.393297344012726e-07, + "loss": 0.0704, + "step": 41386 + }, + { + "epoch": 0.9701764564543446, + "grad_norm": 0.5577048659324646, + "learning_rate": 4.3864028296066727e-07, + "loss": 0.5874, + "step": 41387 + }, + { + "epoch": 0.9701998980291496, + "grad_norm": 0.4574790596961975, + "learning_rate": 4.379513717428063e-07, + "loss": 0.099, + "step": 41388 + }, + { + "epoch": 0.9702233396039546, + "grad_norm": 0.5973219871520996, + "learning_rate": 4.372630007514089e-07, + "loss": 0.1237, + "step": 41389 + }, + { + "epoch": 0.9702467811787596, + "grad_norm": 0.17765338718891144, + "learning_rate": 4.3657516999023875e-07, + "loss": 0.0261, + "step": 41390 + }, + { + "epoch": 0.9702702227535646, + "grad_norm": 0.7221333980560303, + "learning_rate": 4.3588787946300394e-07, + "loss": 0.0922, + "step": 41391 + }, + { + "epoch": 0.9702936643283696, + "grad_norm": 0.39724045991897583, + "learning_rate": 4.352011291734348e-07, + "loss": 0.0528, + "step": 41392 + }, + { + "epoch": 0.9703171059031745, + "grad_norm": 0.29328998923301697, + "learning_rate": 4.3451491912526177e-07, + "loss": 0.0414, + "step": 41393 + }, + { + "epoch": 0.9703405474779796, + "grad_norm": 0.2660689353942871, + "learning_rate": 4.3382924932222625e-07, + "loss": 0.061, + "step": 41394 + }, + { + "epoch": 0.9703639890527845, + "grad_norm": 0.38273367285728455, + "learning_rate": 4.3314411976802525e-07, + "loss": 0.0436, + "step": 41395 + }, + { + "epoch": 0.9703874306275896, + "grad_norm": 0.516768217086792, + "learning_rate": 4.324595304663781e-07, + "loss": 0.0863, + "step": 41396 + }, + { + "epoch": 0.9704108722023945, + "grad_norm": 0.3254733085632324, + "learning_rate": 4.31775481421004e-07, + "loss": 0.2348, + "step": 41397 + }, + { + "epoch": 0.9704343137771996, + "grad_norm": 0.19273675978183746, + "learning_rate": 4.310919726356222e-07, + "loss": 0.0268, + "step": 41398 + }, + { + "epoch": 0.9704577553520045, + "grad_norm": 0.11032325774431229, + "learning_rate": 4.304090041139408e-07, + "loss": 0.0155, + "step": 41399 + }, + { + "epoch": 0.9704811969268096, + "grad_norm": 0.2996954023838043, + "learning_rate": 4.2972657585964584e-07, + "loss": 0.0283, + "step": 41400 + }, + { + "epoch": 0.9705046385016145, + "grad_norm": 0.6152167916297913, + "learning_rate": 4.290446878764565e-07, + "loss": 0.0857, + "step": 41401 + }, + { + "epoch": 0.9705280800764196, + "grad_norm": 0.5204153656959534, + "learning_rate": 4.283633401680809e-07, + "loss": 0.0506, + "step": 41402 + }, + { + "epoch": 0.9705515216512245, + "grad_norm": 0.5796751976013184, + "learning_rate": 4.2768253273820504e-07, + "loss": 0.0854, + "step": 41403 + }, + { + "epoch": 0.9705749632260295, + "grad_norm": 0.2883206009864807, + "learning_rate": 4.270022655905148e-07, + "loss": 0.0237, + "step": 41404 + }, + { + "epoch": 0.9705984048008345, + "grad_norm": 0.371218204498291, + "learning_rate": 4.2632253872870733e-07, + "loss": 0.0891, + "step": 41405 + }, + { + "epoch": 0.9706218463756395, + "grad_norm": 0.04499758407473564, + "learning_rate": 4.2564335215647957e-07, + "loss": 0.0056, + "step": 41406 + }, + { + "epoch": 0.9706452879504445, + "grad_norm": 0.5870208144187927, + "learning_rate": 4.249647058775175e-07, + "loss": 0.0757, + "step": 41407 + }, + { + "epoch": 0.9706687295252495, + "grad_norm": 0.4497622847557068, + "learning_rate": 4.242865998954848e-07, + "loss": 0.0935, + "step": 41408 + }, + { + "epoch": 0.9706921711000545, + "grad_norm": 0.37785691022872925, + "learning_rate": 4.236090342140786e-07, + "loss": 0.0442, + "step": 41409 + }, + { + "epoch": 0.9707156126748595, + "grad_norm": 0.28388696908950806, + "learning_rate": 4.229320088369626e-07, + "loss": 0.0356, + "step": 41410 + }, + { + "epoch": 0.9707390542496644, + "grad_norm": 0.40873685479164124, + "learning_rate": 4.222555237678338e-07, + "loss": 0.5492, + "step": 41411 + }, + { + "epoch": 0.9707624958244695, + "grad_norm": 0.5474952459335327, + "learning_rate": 4.2157957901033386e-07, + "loss": 0.1133, + "step": 41412 + }, + { + "epoch": 0.9707859373992744, + "grad_norm": 0.10602469742298126, + "learning_rate": 4.2090417456813747e-07, + "loss": 0.0155, + "step": 41413 + }, + { + "epoch": 0.9708093789740795, + "grad_norm": 0.5335320234298706, + "learning_rate": 4.202293104449306e-07, + "loss": 0.0892, + "step": 41414 + }, + { + "epoch": 0.9708328205488844, + "grad_norm": 0.4791156053543091, + "learning_rate": 4.195549866443549e-07, + "loss": 0.0634, + "step": 41415 + }, + { + "epoch": 0.9708562621236895, + "grad_norm": 0.3404882252216339, + "learning_rate": 4.1888120317006285e-07, + "loss": 0.0313, + "step": 41416 + }, + { + "epoch": 0.9708797036984944, + "grad_norm": 0.2882966101169586, + "learning_rate": 4.1820796002572936e-07, + "loss": 0.0407, + "step": 41417 + }, + { + "epoch": 0.9709031452732995, + "grad_norm": 0.41531699895858765, + "learning_rate": 4.17535257214996e-07, + "loss": 0.0803, + "step": 41418 + }, + { + "epoch": 0.9709265868481044, + "grad_norm": 0.38404762744903564, + "learning_rate": 4.168630947415264e-07, + "loss": 0.0561, + "step": 41419 + }, + { + "epoch": 0.9709500284229095, + "grad_norm": 0.224908709526062, + "learning_rate": 4.1619147260895106e-07, + "loss": 0.0513, + "step": 41420 + }, + { + "epoch": 0.9709734699977145, + "grad_norm": 0.37819117307662964, + "learning_rate": 4.1552039082092267e-07, + "loss": 0.057, + "step": 41421 + }, + { + "epoch": 0.9709969115725194, + "grad_norm": 0.359444260597229, + "learning_rate": 4.1484984938108263e-07, + "loss": 0.0381, + "step": 41422 + }, + { + "epoch": 0.9710203531473245, + "grad_norm": 0.156735360622406, + "learning_rate": 4.1417984829305037e-07, + "loss": 0.0078, + "step": 41423 + }, + { + "epoch": 0.9710437947221294, + "grad_norm": 0.6436558365821838, + "learning_rate": 4.1351038756050063e-07, + "loss": 0.0713, + "step": 41424 + }, + { + "epoch": 0.9710672362969345, + "grad_norm": 0.3825942277908325, + "learning_rate": 4.1284146718703066e-07, + "loss": 0.4431, + "step": 41425 + }, + { + "epoch": 0.9710906778717394, + "grad_norm": 0.22980448603630066, + "learning_rate": 4.121730871762819e-07, + "loss": 0.0295, + "step": 41426 + }, + { + "epoch": 0.9711141194465445, + "grad_norm": 0.08611294627189636, + "learning_rate": 4.115052475318959e-07, + "loss": 0.0114, + "step": 41427 + }, + { + "epoch": 0.9711375610213494, + "grad_norm": 0.028982073068618774, + "learning_rate": 4.108379482574698e-07, + "loss": 0.0017, + "step": 41428 + }, + { + "epoch": 0.9711610025961545, + "grad_norm": 0.1751657873392105, + "learning_rate": 4.101711893566451e-07, + "loss": 0.0207, + "step": 41429 + }, + { + "epoch": 0.9711844441709594, + "grad_norm": 0.34544068574905396, + "learning_rate": 4.09504970833019e-07, + "loss": 0.0533, + "step": 41430 + }, + { + "epoch": 0.9712078857457644, + "grad_norm": 0.444583535194397, + "learning_rate": 4.0883929269023293e-07, + "loss": 0.049, + "step": 41431 + }, + { + "epoch": 0.9712313273205694, + "grad_norm": 0.40595290064811707, + "learning_rate": 4.0817415493188406e-07, + "loss": 0.0587, + "step": 41432 + }, + { + "epoch": 0.9712547688953744, + "grad_norm": 0.37695637345314026, + "learning_rate": 4.075095575615695e-07, + "loss": 0.0523, + "step": 41433 + }, + { + "epoch": 0.9712782104701794, + "grad_norm": 0.1778693050146103, + "learning_rate": 4.068455005829197e-07, + "loss": 0.0283, + "step": 41434 + }, + { + "epoch": 0.9713016520449844, + "grad_norm": 0.5427206754684448, + "learning_rate": 4.061819839995207e-07, + "loss": 0.4467, + "step": 41435 + }, + { + "epoch": 0.9713250936197894, + "grad_norm": 0.4071846008300781, + "learning_rate": 4.0551900781499175e-07, + "loss": 0.0422, + "step": 41436 + }, + { + "epoch": 0.9713485351945944, + "grad_norm": 0.1061653271317482, + "learning_rate": 4.0485657203290786e-07, + "loss": 0.0101, + "step": 41437 + }, + { + "epoch": 0.9713719767693993, + "grad_norm": 0.12373761832714081, + "learning_rate": 4.0419467665686605e-07, + "loss": 0.0142, + "step": 41438 + }, + { + "epoch": 0.9713954183442044, + "grad_norm": 0.49134427309036255, + "learning_rate": 4.035333216904746e-07, + "loss": 0.4362, + "step": 41439 + }, + { + "epoch": 0.9714188599190093, + "grad_norm": 0.46058881282806396, + "learning_rate": 4.028725071373085e-07, + "loss": 0.0586, + "step": 41440 + }, + { + "epoch": 0.9714423014938144, + "grad_norm": 0.33779552578926086, + "learning_rate": 4.0221223300095365e-07, + "loss": 0.4474, + "step": 41441 + }, + { + "epoch": 0.9714657430686193, + "grad_norm": 0.49973028898239136, + "learning_rate": 4.0155249928499615e-07, + "loss": 0.0422, + "step": 41442 + }, + { + "epoch": 0.9714891846434244, + "grad_norm": 0.38534975051879883, + "learning_rate": 4.008933059930109e-07, + "loss": 0.0692, + "step": 41443 + }, + { + "epoch": 0.9715126262182293, + "grad_norm": 0.5406096577644348, + "learning_rate": 4.00234653128595e-07, + "loss": 0.1112, + "step": 41444 + }, + { + "epoch": 0.9715360677930344, + "grad_norm": 0.45499223470687866, + "learning_rate": 3.9957654069529004e-07, + "loss": 0.0976, + "step": 41445 + }, + { + "epoch": 0.9715595093678393, + "grad_norm": 0.46492013335227966, + "learning_rate": 3.989189686966932e-07, + "loss": 0.0641, + "step": 41446 + }, + { + "epoch": 0.9715829509426444, + "grad_norm": 0.3787330687046051, + "learning_rate": 3.9826193713635717e-07, + "loss": 0.0419, + "step": 41447 + }, + { + "epoch": 0.9716063925174493, + "grad_norm": 0.5352205038070679, + "learning_rate": 3.9760544601785685e-07, + "loss": 0.0586, + "step": 41448 + }, + { + "epoch": 0.9716298340922543, + "grad_norm": 0.2313642054796219, + "learning_rate": 3.96949495344745e-07, + "loss": 0.0569, + "step": 41449 + }, + { + "epoch": 0.9716532756670593, + "grad_norm": 0.13757050037384033, + "learning_rate": 3.962940851205854e-07, + "loss": 0.0167, + "step": 41450 + }, + { + "epoch": 0.9716767172418643, + "grad_norm": 0.6151402592658997, + "learning_rate": 3.9563921534894187e-07, + "loss": 0.5465, + "step": 41451 + }, + { + "epoch": 0.9717001588166693, + "grad_norm": 0.6098654866218567, + "learning_rate": 3.9498488603335605e-07, + "loss": 0.1081, + "step": 41452 + }, + { + "epoch": 0.9717236003914743, + "grad_norm": 0.14033329486846924, + "learning_rate": 3.943310971773806e-07, + "loss": 0.0152, + "step": 41453 + }, + { + "epoch": 0.9717470419662793, + "grad_norm": 0.08865802735090256, + "learning_rate": 3.936778487845683e-07, + "loss": 0.0187, + "step": 41454 + }, + { + "epoch": 0.9717704835410843, + "grad_norm": 0.21511410176753998, + "learning_rate": 3.9302514085846066e-07, + "loss": 0.0285, + "step": 41455 + }, + { + "epoch": 0.9717939251158892, + "grad_norm": 0.21856342256069183, + "learning_rate": 3.9237297340259936e-07, + "loss": 0.0206, + "step": 41456 + }, + { + "epoch": 0.9718173666906943, + "grad_norm": 0.4451315701007843, + "learning_rate": 3.9172134642051493e-07, + "loss": 0.1294, + "step": 41457 + }, + { + "epoch": 0.9718408082654992, + "grad_norm": 0.27139464020729065, + "learning_rate": 3.910702599157601e-07, + "loss": 0.0391, + "step": 41458 + }, + { + "epoch": 0.9718642498403043, + "grad_norm": 0.6773297190666199, + "learning_rate": 3.904197138918653e-07, + "loss": 0.0822, + "step": 41459 + }, + { + "epoch": 0.9718876914151092, + "grad_norm": 0.23143406212329865, + "learning_rate": 3.8976970835233884e-07, + "loss": 0.0178, + "step": 41460 + }, + { + "epoch": 0.9719111329899143, + "grad_norm": 0.6827218532562256, + "learning_rate": 3.891202433007335e-07, + "loss": 0.1114, + "step": 41461 + }, + { + "epoch": 0.9719345745647192, + "grad_norm": 0.668215811252594, + "learning_rate": 3.884713187405575e-07, + "loss": 0.1779, + "step": 41462 + }, + { + "epoch": 0.9719580161395243, + "grad_norm": 0.1993371546268463, + "learning_rate": 3.8782293467533036e-07, + "loss": 0.0384, + "step": 41463 + }, + { + "epoch": 0.9719814577143292, + "grad_norm": 0.3324033319950104, + "learning_rate": 3.8717509110859363e-07, + "loss": 0.0428, + "step": 41464 + }, + { + "epoch": 0.9720048992891342, + "grad_norm": 0.3798294961452484, + "learning_rate": 3.865277880438334e-07, + "loss": 0.0614, + "step": 41465 + }, + { + "epoch": 0.9720283408639392, + "grad_norm": 0.18093335628509521, + "learning_rate": 3.8588102548458016e-07, + "loss": 0.036, + "step": 41466 + }, + { + "epoch": 0.9720517824387442, + "grad_norm": 0.8559172749519348, + "learning_rate": 3.8523480343433115e-07, + "loss": 0.2793, + "step": 41467 + }, + { + "epoch": 0.9720752240135492, + "grad_norm": 0.057015661150217056, + "learning_rate": 3.845891218966058e-07, + "loss": 0.0072, + "step": 41468 + }, + { + "epoch": 0.9720986655883542, + "grad_norm": 0.11612646281719208, + "learning_rate": 3.8394398087490125e-07, + "loss": 0.021, + "step": 41469 + }, + { + "epoch": 0.9721221071631592, + "grad_norm": 0.4839322865009308, + "learning_rate": 3.8329938037271474e-07, + "loss": 0.0743, + "step": 41470 + }, + { + "epoch": 0.9721455487379642, + "grad_norm": 0.2876754701137543, + "learning_rate": 3.8265532039354345e-07, + "loss": 0.0336, + "step": 41471 + }, + { + "epoch": 0.9721689903127693, + "grad_norm": 0.7651578187942505, + "learning_rate": 3.820118009408846e-07, + "loss": 0.4418, + "step": 41472 + }, + { + "epoch": 0.9721924318875742, + "grad_norm": 0.2897208631038666, + "learning_rate": 3.8136882201823545e-07, + "loss": 0.0302, + "step": 41473 + }, + { + "epoch": 0.9722158734623793, + "grad_norm": 0.2077774703502655, + "learning_rate": 3.8072638362908195e-07, + "loss": 0.0422, + "step": 41474 + }, + { + "epoch": 0.9722393150371842, + "grad_norm": 0.4236350357532501, + "learning_rate": 3.8008448577689927e-07, + "loss": 0.0813, + "step": 41475 + }, + { + "epoch": 0.9722627566119892, + "grad_norm": 0.1485920250415802, + "learning_rate": 3.794431284651845e-07, + "loss": 0.0264, + "step": 41476 + }, + { + "epoch": 0.9722861981867942, + "grad_norm": 0.26976513862609863, + "learning_rate": 3.7880231169741267e-07, + "loss": 0.0454, + "step": 41477 + }, + { + "epoch": 0.9723096397615992, + "grad_norm": 0.30789488554000854, + "learning_rate": 3.7816203547705877e-07, + "loss": 0.0286, + "step": 41478 + }, + { + "epoch": 0.9723330813364042, + "grad_norm": 0.3155902624130249, + "learning_rate": 3.775222998076089e-07, + "loss": 0.0617, + "step": 41479 + }, + { + "epoch": 0.9723565229112092, + "grad_norm": 0.3985252380371094, + "learning_rate": 3.768831046925159e-07, + "loss": 0.0998, + "step": 41480 + }, + { + "epoch": 0.9723799644860142, + "grad_norm": 0.37585046887397766, + "learning_rate": 3.762444501352547e-07, + "loss": 0.0858, + "step": 41481 + }, + { + "epoch": 0.9724034060608192, + "grad_norm": 0.5872209072113037, + "learning_rate": 3.756063361392892e-07, + "loss": 0.0698, + "step": 41482 + }, + { + "epoch": 0.9724268476356241, + "grad_norm": 0.28434473276138306, + "learning_rate": 3.7496876270809444e-07, + "loss": 0.031, + "step": 41483 + }, + { + "epoch": 0.9724502892104292, + "grad_norm": 0.4178433120250702, + "learning_rate": 3.743317298451121e-07, + "loss": 0.0665, + "step": 41484 + }, + { + "epoch": 0.9724737307852341, + "grad_norm": 0.5045269131660461, + "learning_rate": 3.736952375538172e-07, + "loss": 0.1043, + "step": 41485 + }, + { + "epoch": 0.9724971723600392, + "grad_norm": 0.11408555507659912, + "learning_rate": 3.730592858376403e-07, + "loss": 0.0236, + "step": 41486 + }, + { + "epoch": 0.9725206139348441, + "grad_norm": 1.0693528652191162, + "learning_rate": 3.7242387470005634e-07, + "loss": 0.1496, + "step": 41487 + }, + { + "epoch": 0.9725440555096492, + "grad_norm": 0.5057504773139954, + "learning_rate": 3.7178900414449605e-07, + "loss": 0.4909, + "step": 41488 + }, + { + "epoch": 0.9725674970844541, + "grad_norm": 0.316112220287323, + "learning_rate": 3.711546741744121e-07, + "loss": 0.0592, + "step": 41489 + }, + { + "epoch": 0.9725909386592592, + "grad_norm": 0.29723039269447327, + "learning_rate": 3.705208847932351e-07, + "loss": 0.0675, + "step": 41490 + }, + { + "epoch": 0.9726143802340641, + "grad_norm": 0.193046435713768, + "learning_rate": 3.69887636004429e-07, + "loss": 0.0245, + "step": 41491 + }, + { + "epoch": 0.9726378218088692, + "grad_norm": 0.6061840653419495, + "learning_rate": 3.692549278114021e-07, + "loss": 0.1015, + "step": 41492 + }, + { + "epoch": 0.9726612633836741, + "grad_norm": 0.7250769138336182, + "learning_rate": 3.6862276021759623e-07, + "loss": 0.0927, + "step": 41493 + }, + { + "epoch": 0.9726847049584791, + "grad_norm": 0.216175839304924, + "learning_rate": 3.6799113322645297e-07, + "loss": 0.0183, + "step": 41494 + }, + { + "epoch": 0.9727081465332841, + "grad_norm": 0.2302519530057907, + "learning_rate": 3.673600468413918e-07, + "loss": 0.0304, + "step": 41495 + }, + { + "epoch": 0.9727315881080891, + "grad_norm": 0.06445769220590591, + "learning_rate": 3.6672950106583224e-07, + "loss": 0.0044, + "step": 41496 + }, + { + "epoch": 0.9727550296828941, + "grad_norm": 0.40380337834358215, + "learning_rate": 3.6609949590320494e-07, + "loss": 0.0872, + "step": 41497 + }, + { + "epoch": 0.9727784712576991, + "grad_norm": 0.26759007573127747, + "learning_rate": 3.6547003135691815e-07, + "loss": 0.0356, + "step": 41498 + }, + { + "epoch": 0.9728019128325041, + "grad_norm": 0.19480611383914948, + "learning_rate": 3.6484110743040256e-07, + "loss": 0.0477, + "step": 41499 + }, + { + "epoch": 0.9728253544073091, + "grad_norm": 0.46729370951652527, + "learning_rate": 3.6421272412704434e-07, + "loss": 0.0547, + "step": 41500 + }, + { + "epoch": 0.972848795982114, + "grad_norm": 0.20427633821964264, + "learning_rate": 3.6358488145027403e-07, + "loss": 0.02, + "step": 41501 + }, + { + "epoch": 0.9728722375569191, + "grad_norm": 0.38382041454315186, + "learning_rate": 3.6295757940350005e-07, + "loss": 0.0749, + "step": 41502 + }, + { + "epoch": 0.972895679131724, + "grad_norm": 0.5291053652763367, + "learning_rate": 3.623308179901197e-07, + "loss": 0.0807, + "step": 41503 + }, + { + "epoch": 0.9729191207065291, + "grad_norm": 0.41980692744255066, + "learning_rate": 3.6170459721353025e-07, + "loss": 0.0493, + "step": 41504 + }, + { + "epoch": 0.972942562281334, + "grad_norm": 0.6334311962127686, + "learning_rate": 3.6107891707714e-07, + "loss": 0.105, + "step": 41505 + }, + { + "epoch": 0.9729660038561391, + "grad_norm": 0.09300345927476883, + "learning_rate": 3.604537775843242e-07, + "loss": 0.0145, + "step": 41506 + }, + { + "epoch": 0.972989445430944, + "grad_norm": 0.22372883558273315, + "learning_rate": 3.598291787385022e-07, + "loss": 0.0384, + "step": 41507 + }, + { + "epoch": 0.9730128870057491, + "grad_norm": 0.3717555105686188, + "learning_rate": 3.59205120543038e-07, + "loss": 0.0751, + "step": 41508 + }, + { + "epoch": 0.973036328580554, + "grad_norm": 0.22743934392929077, + "learning_rate": 3.585816030013289e-07, + "loss": 0.04, + "step": 41509 + }, + { + "epoch": 0.973059770155359, + "grad_norm": 0.23660804331302643, + "learning_rate": 3.579586261167611e-07, + "loss": 0.0546, + "step": 41510 + }, + { + "epoch": 0.973083211730164, + "grad_norm": 0.09460056573152542, + "learning_rate": 3.5733618989270966e-07, + "loss": 0.007, + "step": 41511 + }, + { + "epoch": 0.973106653304969, + "grad_norm": 0.46775755286216736, + "learning_rate": 3.567142943325608e-07, + "loss": 0.0881, + "step": 41512 + }, + { + "epoch": 0.973130094879774, + "grad_norm": 0.5337628126144409, + "learning_rate": 3.5609293943966725e-07, + "loss": 0.054, + "step": 41513 + }, + { + "epoch": 0.973153536454579, + "grad_norm": 0.5248114466667175, + "learning_rate": 3.5547212521742645e-07, + "loss": 0.0962, + "step": 41514 + }, + { + "epoch": 0.973176978029384, + "grad_norm": 0.3460076153278351, + "learning_rate": 3.548518516691912e-07, + "loss": 0.0498, + "step": 41515 + }, + { + "epoch": 0.973200419604189, + "grad_norm": 0.5429167747497559, + "learning_rate": 3.542321187983366e-07, + "loss": 0.5005, + "step": 41516 + }, + { + "epoch": 0.973223861178994, + "grad_norm": 0.18778486549854279, + "learning_rate": 3.5361292660821553e-07, + "loss": 0.0214, + "step": 41517 + }, + { + "epoch": 0.973247302753799, + "grad_norm": 0.148214191198349, + "learning_rate": 3.5299427510219197e-07, + "loss": 0.0295, + "step": 41518 + }, + { + "epoch": 0.9732707443286039, + "grad_norm": 0.4266478717327118, + "learning_rate": 3.5237616428362987e-07, + "loss": 0.0639, + "step": 41519 + }, + { + "epoch": 0.973294185903409, + "grad_norm": 0.5227568745613098, + "learning_rate": 3.51758594155871e-07, + "loss": 0.0929, + "step": 41520 + }, + { + "epoch": 0.9733176274782139, + "grad_norm": 0.09683050215244293, + "learning_rate": 3.511415647222793e-07, + "loss": 0.0195, + "step": 41521 + }, + { + "epoch": 0.973341069053019, + "grad_norm": 0.3299354612827301, + "learning_rate": 3.5052507598618556e-07, + "loss": 0.0476, + "step": 41522 + }, + { + "epoch": 0.973364510627824, + "grad_norm": 0.25648799538612366, + "learning_rate": 3.4990912795095367e-07, + "loss": 0.0526, + "step": 41523 + }, + { + "epoch": 0.973387952202629, + "grad_norm": 0.1378723680973053, + "learning_rate": 3.492937206199032e-07, + "loss": 0.03, + "step": 41524 + }, + { + "epoch": 0.973411393777434, + "grad_norm": 0.15109068155288696, + "learning_rate": 3.486788539963981e-07, + "loss": 0.0275, + "step": 41525 + }, + { + "epoch": 0.973434835352239, + "grad_norm": 0.8272125124931335, + "learning_rate": 3.4806452808375803e-07, + "loss": 0.5281, + "step": 41526 + }, + { + "epoch": 0.973458276927044, + "grad_norm": 0.3166492283344269, + "learning_rate": 3.4745074288532466e-07, + "loss": 0.027, + "step": 41527 + }, + { + "epoch": 0.973481718501849, + "grad_norm": 0.30222463607788086, + "learning_rate": 3.4683749840442867e-07, + "loss": 0.0182, + "step": 41528 + }, + { + "epoch": 0.973505160076654, + "grad_norm": 0.17529582977294922, + "learning_rate": 3.4622479464438975e-07, + "loss": 0.0184, + "step": 41529 + }, + { + "epoch": 0.9735286016514589, + "grad_norm": 0.09956447780132294, + "learning_rate": 3.4561263160853843e-07, + "loss": 0.0098, + "step": 41530 + }, + { + "epoch": 0.973552043226264, + "grad_norm": 0.5276921391487122, + "learning_rate": 3.450010093001943e-07, + "loss": 0.1156, + "step": 41531 + }, + { + "epoch": 0.9735754848010689, + "grad_norm": 0.31561511754989624, + "learning_rate": 3.44389927722677e-07, + "loss": 0.3463, + "step": 41532 + }, + { + "epoch": 0.973598926375874, + "grad_norm": 0.5775660872459412, + "learning_rate": 3.43779386879306e-07, + "loss": 0.0941, + "step": 41533 + }, + { + "epoch": 0.9736223679506789, + "grad_norm": 0.40505701303482056, + "learning_rate": 3.4316938677338986e-07, + "loss": 0.0405, + "step": 41534 + }, + { + "epoch": 0.973645809525484, + "grad_norm": 0.5923963785171509, + "learning_rate": 3.4255992740822585e-07, + "loss": 0.6594, + "step": 41535 + }, + { + "epoch": 0.9736692511002889, + "grad_norm": 1.0656791925430298, + "learning_rate": 3.419510087871447e-07, + "loss": 0.0772, + "step": 41536 + }, + { + "epoch": 0.973692692675094, + "grad_norm": 0.09542442113161087, + "learning_rate": 3.4134263091344374e-07, + "loss": 0.0062, + "step": 41537 + }, + { + "epoch": 0.9737161342498989, + "grad_norm": 0.0694449171423912, + "learning_rate": 3.407347937904093e-07, + "loss": 0.0089, + "step": 41538 + }, + { + "epoch": 0.9737395758247039, + "grad_norm": 0.29745444655418396, + "learning_rate": 3.4012749742136087e-07, + "loss": 0.0562, + "step": 41539 + }, + { + "epoch": 0.9737630173995089, + "grad_norm": 0.30205637216567993, + "learning_rate": 3.395207418095736e-07, + "loss": 0.0513, + "step": 41540 + }, + { + "epoch": 0.9737864589743139, + "grad_norm": 0.596688449382782, + "learning_rate": 3.38914526958356e-07, + "loss": 0.0926, + "step": 41541 + }, + { + "epoch": 0.9738099005491189, + "grad_norm": 0.2647489607334137, + "learning_rate": 3.3830885287098326e-07, + "loss": 0.2701, + "step": 41542 + }, + { + "epoch": 0.9738333421239239, + "grad_norm": 0.13571399450302124, + "learning_rate": 3.3770371955075263e-07, + "loss": 0.0282, + "step": 41543 + }, + { + "epoch": 0.9738567836987289, + "grad_norm": 0.3158809244632721, + "learning_rate": 3.370991270009394e-07, + "loss": 0.0322, + "step": 41544 + }, + { + "epoch": 0.9738802252735339, + "grad_norm": 0.4545130133628845, + "learning_rate": 3.364950752248408e-07, + "loss": 0.0799, + "step": 41545 + }, + { + "epoch": 0.9739036668483388, + "grad_norm": 0.19668082892894745, + "learning_rate": 3.358915642257099e-07, + "loss": 0.0241, + "step": 41546 + }, + { + "epoch": 0.9739271084231439, + "grad_norm": 0.14522361755371094, + "learning_rate": 3.3528859400683287e-07, + "loss": 0.0189, + "step": 41547 + }, + { + "epoch": 0.9739505499979488, + "grad_norm": 0.5577630996704102, + "learning_rate": 3.3468616457149604e-07, + "loss": 0.2369, + "step": 41548 + }, + { + "epoch": 0.9739739915727539, + "grad_norm": 0.6500875949859619, + "learning_rate": 3.340842759229412e-07, + "loss": 0.0575, + "step": 41549 + }, + { + "epoch": 0.9739974331475588, + "grad_norm": 0.3944137692451477, + "learning_rate": 3.334829280644547e-07, + "loss": 0.5082, + "step": 41550 + }, + { + "epoch": 0.9740208747223639, + "grad_norm": 0.5757200121879578, + "learning_rate": 3.328821209992894e-07, + "loss": 0.0896, + "step": 41551 + }, + { + "epoch": 0.9740443162971688, + "grad_norm": 0.22118085622787476, + "learning_rate": 3.322818547307094e-07, + "loss": 0.0215, + "step": 41552 + }, + { + "epoch": 0.9740677578719739, + "grad_norm": 0.3186684846878052, + "learning_rate": 3.316821292619676e-07, + "loss": 0.0643, + "step": 41553 + }, + { + "epoch": 0.9740911994467788, + "grad_norm": 0.2796969711780548, + "learning_rate": 3.310829445963171e-07, + "loss": 0.0323, + "step": 41554 + }, + { + "epoch": 0.9741146410215838, + "grad_norm": 0.10100767761468887, + "learning_rate": 3.3048430073702175e-07, + "loss": 0.0194, + "step": 41555 + }, + { + "epoch": 0.9741380825963888, + "grad_norm": 0.3673040270805359, + "learning_rate": 3.2988619768732353e-07, + "loss": 0.1079, + "step": 41556 + }, + { + "epoch": 0.9741615241711938, + "grad_norm": 0.14704503118991852, + "learning_rate": 3.292886354504532e-07, + "loss": 0.0179, + "step": 41557 + }, + { + "epoch": 0.9741849657459988, + "grad_norm": 0.10201355069875717, + "learning_rate": 3.2869161402967473e-07, + "loss": 0.0151, + "step": 41558 + }, + { + "epoch": 0.9742084073208038, + "grad_norm": 0.2657921314239502, + "learning_rate": 3.280951334282079e-07, + "loss": 0.0479, + "step": 41559 + }, + { + "epoch": 0.9742318488956088, + "grad_norm": 0.6291139125823975, + "learning_rate": 3.274991936493166e-07, + "loss": 0.1455, + "step": 41560 + }, + { + "epoch": 0.9742552904704138, + "grad_norm": 0.24986831843852997, + "learning_rate": 3.2690379469619834e-07, + "loss": 0.0409, + "step": 41561 + }, + { + "epoch": 0.9742787320452188, + "grad_norm": 0.23343071341514587, + "learning_rate": 3.2630893657211725e-07, + "loss": 0.0224, + "step": 41562 + }, + { + "epoch": 0.9743021736200238, + "grad_norm": 0.2836979925632477, + "learning_rate": 3.257146192802818e-07, + "loss": 0.0347, + "step": 41563 + }, + { + "epoch": 0.9743256151948287, + "grad_norm": 0.5372315645217896, + "learning_rate": 3.251208428239116e-07, + "loss": 0.0916, + "step": 41564 + }, + { + "epoch": 0.9743490567696338, + "grad_norm": 0.6663637757301331, + "learning_rate": 3.245276072062486e-07, + "loss": 0.6269, + "step": 41565 + }, + { + "epoch": 0.9743724983444387, + "grad_norm": 0.25978323817253113, + "learning_rate": 3.239349124305013e-07, + "loss": 0.0243, + "step": 41566 + }, + { + "epoch": 0.9743959399192438, + "grad_norm": 0.4306632876396179, + "learning_rate": 3.233427584998894e-07, + "loss": 0.0781, + "step": 41567 + }, + { + "epoch": 0.9744193814940487, + "grad_norm": 0.1699114739894867, + "learning_rate": 3.227511454176213e-07, + "loss": 0.0414, + "step": 41568 + }, + { + "epoch": 0.9744428230688538, + "grad_norm": 0.6085443496704102, + "learning_rate": 3.2216007318689456e-07, + "loss": 0.0552, + "step": 41569 + }, + { + "epoch": 0.9744662646436587, + "grad_norm": 0.07474693655967712, + "learning_rate": 3.215695418109399e-07, + "loss": 0.0123, + "step": 41570 + }, + { + "epoch": 0.9744897062184638, + "grad_norm": 0.12149310857057571, + "learning_rate": 3.209795512929548e-07, + "loss": 0.0205, + "step": 41571 + }, + { + "epoch": 0.9745131477932687, + "grad_norm": 0.5630302429199219, + "learning_rate": 3.203901016361366e-07, + "loss": 0.1003, + "step": 41572 + }, + { + "epoch": 0.9745365893680737, + "grad_norm": 0.6142531633377075, + "learning_rate": 3.1980119284367174e-07, + "loss": 0.11, + "step": 41573 + }, + { + "epoch": 0.9745600309428788, + "grad_norm": 0.18983004987239838, + "learning_rate": 3.192128249187798e-07, + "loss": 0.0267, + "step": 41574 + }, + { + "epoch": 0.9745834725176837, + "grad_norm": 0.38622337579727173, + "learning_rate": 3.1862499786464715e-07, + "loss": 0.0507, + "step": 41575 + }, + { + "epoch": 0.9746069140924888, + "grad_norm": 0.4921562969684601, + "learning_rate": 3.1803771168444907e-07, + "loss": 0.0835, + "step": 41576 + }, + { + "epoch": 0.9746303556672937, + "grad_norm": 0.22593046724796295, + "learning_rate": 3.174509663813829e-07, + "loss": 0.0269, + "step": 41577 + }, + { + "epoch": 0.9746537972420988, + "grad_norm": 0.5779138803482056, + "learning_rate": 3.168647619586351e-07, + "loss": 0.4277, + "step": 41578 + }, + { + "epoch": 0.9746772388169037, + "grad_norm": 0.5187584161758423, + "learning_rate": 3.1627909841936975e-07, + "loss": 0.1245, + "step": 41579 + }, + { + "epoch": 0.9747006803917088, + "grad_norm": 0.2380494326353073, + "learning_rate": 3.156939757667954e-07, + "loss": 0.0275, + "step": 41580 + }, + { + "epoch": 0.9747241219665137, + "grad_norm": 0.31170767545700073, + "learning_rate": 3.1510939400405395e-07, + "loss": 0.0442, + "step": 41581 + }, + { + "epoch": 0.9747475635413188, + "grad_norm": 0.15760792791843414, + "learning_rate": 3.145253531343317e-07, + "loss": 0.0376, + "step": 41582 + }, + { + "epoch": 0.9747710051161237, + "grad_norm": 0.5663465857505798, + "learning_rate": 3.1394185316080403e-07, + "loss": 0.1287, + "step": 41583 + }, + { + "epoch": 0.9747944466909287, + "grad_norm": 0.45720648765563965, + "learning_rate": 3.133588940866239e-07, + "loss": 0.081, + "step": 41584 + }, + { + "epoch": 0.9748178882657337, + "grad_norm": 0.5831313133239746, + "learning_rate": 3.127764759149665e-07, + "loss": 0.0799, + "step": 41585 + }, + { + "epoch": 0.9748413298405387, + "grad_norm": 0.3544454276561737, + "learning_rate": 3.12194598648996e-07, + "loss": 0.0542, + "step": 41586 + }, + { + "epoch": 0.9748647714153437, + "grad_norm": 0.15916728973388672, + "learning_rate": 3.116132622918433e-07, + "loss": 0.0295, + "step": 41587 + }, + { + "epoch": 0.9748882129901487, + "grad_norm": 0.1787969470024109, + "learning_rate": 3.1103246684668354e-07, + "loss": 0.0362, + "step": 41588 + }, + { + "epoch": 0.9749116545649537, + "grad_norm": 0.13135959208011627, + "learning_rate": 3.104522123166698e-07, + "loss": 0.0067, + "step": 41589 + }, + { + "epoch": 0.9749350961397587, + "grad_norm": 0.5413962602615356, + "learning_rate": 3.0987249870495505e-07, + "loss": 0.5512, + "step": 41590 + }, + { + "epoch": 0.9749585377145636, + "grad_norm": 0.44951367378234863, + "learning_rate": 3.092933260146591e-07, + "loss": 0.0243, + "step": 41591 + }, + { + "epoch": 0.9749819792893687, + "grad_norm": 0.8172479867935181, + "learning_rate": 3.0871469424894605e-07, + "loss": 0.1067, + "step": 41592 + }, + { + "epoch": 0.9750054208641736, + "grad_norm": 0.10092612355947495, + "learning_rate": 3.0813660341095785e-07, + "loss": 0.0084, + "step": 41593 + }, + { + "epoch": 0.9750288624389787, + "grad_norm": 0.3413601219654083, + "learning_rate": 3.075590535038142e-07, + "loss": 0.0399, + "step": 41594 + }, + { + "epoch": 0.9750523040137836, + "grad_norm": 0.2149399369955063, + "learning_rate": 3.0698204453065705e-07, + "loss": 0.0201, + "step": 41595 + }, + { + "epoch": 0.9750757455885887, + "grad_norm": 0.5538527965545654, + "learning_rate": 3.064055764946172e-07, + "loss": 0.0611, + "step": 41596 + }, + { + "epoch": 0.9750991871633936, + "grad_norm": 0.40571653842926025, + "learning_rate": 3.0582964939882555e-07, + "loss": 0.0742, + "step": 41597 + }, + { + "epoch": 0.9751226287381987, + "grad_norm": 0.5045491456985474, + "learning_rate": 3.052542632464128e-07, + "loss": 0.0438, + "step": 41598 + }, + { + "epoch": 0.9751460703130036, + "grad_norm": 0.4681400656700134, + "learning_rate": 3.046794180404877e-07, + "loss": 0.0808, + "step": 41599 + }, + { + "epoch": 0.9751695118878086, + "grad_norm": 0.08569910377264023, + "learning_rate": 3.0410511378418103e-07, + "loss": 0.0136, + "step": 41600 + }, + { + "epoch": 0.9751929534626136, + "grad_norm": 0.14539946615695953, + "learning_rate": 3.0353135048059034e-07, + "loss": 0.0317, + "step": 41601 + }, + { + "epoch": 0.9752163950374186, + "grad_norm": 0.15827012062072754, + "learning_rate": 3.0295812813285753e-07, + "loss": 0.0148, + "step": 41602 + }, + { + "epoch": 0.9752398366122236, + "grad_norm": 0.8843204975128174, + "learning_rate": 3.02385446744069e-07, + "loss": 0.1365, + "step": 41603 + }, + { + "epoch": 0.9752632781870286, + "grad_norm": 0.5717313289642334, + "learning_rate": 3.018133063173334e-07, + "loss": 0.0606, + "step": 41604 + }, + { + "epoch": 0.9752867197618336, + "grad_norm": 0.27989307045936584, + "learning_rate": 3.0124170685577047e-07, + "loss": 0.0409, + "step": 41605 + }, + { + "epoch": 0.9753101613366386, + "grad_norm": 0.6028111577033997, + "learning_rate": 3.006706483624777e-07, + "loss": 0.1027, + "step": 41606 + }, + { + "epoch": 0.9753336029114436, + "grad_norm": 0.4830862581729889, + "learning_rate": 3.0010013084054154e-07, + "loss": 0.0661, + "step": 41607 + }, + { + "epoch": 0.9753570444862486, + "grad_norm": 0.2021019458770752, + "learning_rate": 2.9953015429305954e-07, + "loss": 0.0158, + "step": 41608 + }, + { + "epoch": 0.9753804860610535, + "grad_norm": 0.12180930376052856, + "learning_rate": 2.989607187231402e-07, + "loss": 0.0112, + "step": 41609 + }, + { + "epoch": 0.9754039276358586, + "grad_norm": 0.4893020987510681, + "learning_rate": 2.9839182413387013e-07, + "loss": 0.0898, + "step": 41610 + }, + { + "epoch": 0.9754273692106635, + "grad_norm": 0.3822706937789917, + "learning_rate": 2.978234705283134e-07, + "loss": 0.4306, + "step": 41611 + }, + { + "epoch": 0.9754508107854686, + "grad_norm": 0.16061189770698547, + "learning_rate": 2.972556579095787e-07, + "loss": 0.0236, + "step": 41612 + }, + { + "epoch": 0.9754742523602735, + "grad_norm": 0.43101605772972107, + "learning_rate": 2.9668838628074125e-07, + "loss": 0.0735, + "step": 41613 + }, + { + "epoch": 0.9754976939350786, + "grad_norm": 0.2942920923233032, + "learning_rate": 2.961216556448654e-07, + "loss": 0.0389, + "step": 41614 + }, + { + "epoch": 0.9755211355098835, + "grad_norm": 0.683502733707428, + "learning_rate": 2.955554660050486e-07, + "loss": 0.2691, + "step": 41615 + }, + { + "epoch": 0.9755445770846886, + "grad_norm": 0.601709246635437, + "learning_rate": 2.9498981736434396e-07, + "loss": 0.0898, + "step": 41616 + }, + { + "epoch": 0.9755680186594935, + "grad_norm": 0.2293391078710556, + "learning_rate": 2.944247097258268e-07, + "loss": 0.0448, + "step": 41617 + }, + { + "epoch": 0.9755914602342985, + "grad_norm": 0.20077760517597198, + "learning_rate": 2.938601430925725e-07, + "loss": 0.0259, + "step": 41618 + }, + { + "epoch": 0.9756149018091035, + "grad_norm": 0.5367979407310486, + "learning_rate": 2.932961174676341e-07, + "loss": 0.0767, + "step": 41619 + }, + { + "epoch": 0.9756383433839085, + "grad_norm": 1.199867606163025, + "learning_rate": 2.927326328540647e-07, + "loss": 0.112, + "step": 41620 + }, + { + "epoch": 0.9756617849587135, + "grad_norm": 0.33695876598358154, + "learning_rate": 2.921696892549397e-07, + "loss": 0.0739, + "step": 41621 + }, + { + "epoch": 0.9756852265335185, + "grad_norm": 0.2801748514175415, + "learning_rate": 2.9160728667330104e-07, + "loss": 0.0316, + "step": 41622 + }, + { + "epoch": 0.9757086681083235, + "grad_norm": 0.3872241973876953, + "learning_rate": 2.91045425112213e-07, + "loss": 0.0695, + "step": 41623 + }, + { + "epoch": 0.9757321096831285, + "grad_norm": 0.15402022004127502, + "learning_rate": 2.904841045747175e-07, + "loss": 0.0181, + "step": 41624 + }, + { + "epoch": 0.9757555512579336, + "grad_norm": 0.3019478917121887, + "learning_rate": 2.8992332506384557e-07, + "loss": 0.0517, + "step": 41625 + }, + { + "epoch": 0.9757789928327385, + "grad_norm": 0.0984470471739769, + "learning_rate": 2.8936308658266133e-07, + "loss": 0.0154, + "step": 41626 + }, + { + "epoch": 0.9758024344075436, + "grad_norm": 0.3287818431854248, + "learning_rate": 2.888033891341957e-07, + "loss": 0.0452, + "step": 41627 + }, + { + "epoch": 0.9758258759823485, + "grad_norm": 0.14037811756134033, + "learning_rate": 2.882442327214907e-07, + "loss": 0.016, + "step": 41628 + }, + { + "epoch": 0.9758493175571535, + "grad_norm": 0.3356958031654358, + "learning_rate": 2.8768561734756614e-07, + "loss": 0.3696, + "step": 41629 + }, + { + "epoch": 0.9758727591319585, + "grad_norm": 0.082548126578331, + "learning_rate": 2.8712754301547517e-07, + "loss": 0.007, + "step": 41630 + }, + { + "epoch": 0.9758962007067635, + "grad_norm": 0.45885005593299866, + "learning_rate": 2.8657000972822645e-07, + "loss": 0.6477, + "step": 41631 + }, + { + "epoch": 0.9759196422815685, + "grad_norm": 0.6235707998275757, + "learning_rate": 2.860130174888509e-07, + "loss": 0.0967, + "step": 41632 + }, + { + "epoch": 0.9759430838563735, + "grad_norm": 0.34459295868873596, + "learning_rate": 2.8545656630037943e-07, + "loss": 0.0631, + "step": 41633 + }, + { + "epoch": 0.9759665254311785, + "grad_norm": 0.1917751580476761, + "learning_rate": 2.8490065616582074e-07, + "loss": 0.0359, + "step": 41634 + }, + { + "epoch": 0.9759899670059835, + "grad_norm": 0.42218267917633057, + "learning_rate": 2.8434528708819464e-07, + "loss": 0.018, + "step": 41635 + }, + { + "epoch": 0.9760134085807884, + "grad_norm": 0.607062041759491, + "learning_rate": 2.8379045907050983e-07, + "loss": 0.155, + "step": 41636 + }, + { + "epoch": 0.9760368501555935, + "grad_norm": 0.2199750393629074, + "learning_rate": 2.8323617211578615e-07, + "loss": 0.0206, + "step": 41637 + }, + { + "epoch": 0.9760602917303984, + "grad_norm": 0.2291431427001953, + "learning_rate": 2.8268242622703225e-07, + "loss": 0.0179, + "step": 41638 + }, + { + "epoch": 0.9760837333052035, + "grad_norm": 0.28974074125289917, + "learning_rate": 2.8212922140723463e-07, + "loss": 0.0416, + "step": 41639 + }, + { + "epoch": 0.9761071748800084, + "grad_norm": 0.35232895612716675, + "learning_rate": 2.8157655765941315e-07, + "loss": 0.0425, + "step": 41640 + }, + { + "epoch": 0.9761306164548135, + "grad_norm": 0.1525421291589737, + "learning_rate": 2.810244349865654e-07, + "loss": 0.0197, + "step": 41641 + }, + { + "epoch": 0.9761540580296184, + "grad_norm": 0.36333826184272766, + "learning_rate": 2.8047285339168894e-07, + "loss": 0.0499, + "step": 41642 + }, + { + "epoch": 0.9761774996044235, + "grad_norm": 0.3214302062988281, + "learning_rate": 2.799218128777592e-07, + "loss": 0.0579, + "step": 41643 + }, + { + "epoch": 0.9762009411792284, + "grad_norm": 0.6451200842857361, + "learning_rate": 2.793713134477738e-07, + "loss": 0.5831, + "step": 41644 + }, + { + "epoch": 0.9762243827540334, + "grad_norm": 0.08691227436065674, + "learning_rate": 2.7882135510473027e-07, + "loss": 0.0188, + "step": 41645 + }, + { + "epoch": 0.9762478243288384, + "grad_norm": 0.5319817662239075, + "learning_rate": 2.782719378516041e-07, + "loss": 0.1155, + "step": 41646 + }, + { + "epoch": 0.9762712659036434, + "grad_norm": 0.14921218156814575, + "learning_rate": 2.777230616913817e-07, + "loss": 0.0237, + "step": 41647 + }, + { + "epoch": 0.9762947074784484, + "grad_norm": 0.2420395165681839, + "learning_rate": 2.7717472662703856e-07, + "loss": 0.0503, + "step": 41648 + }, + { + "epoch": 0.9763181490532534, + "grad_norm": 0.4799952507019043, + "learning_rate": 2.7662693266155004e-07, + "loss": 0.0262, + "step": 41649 + }, + { + "epoch": 0.9763415906280584, + "grad_norm": 0.5713198184967041, + "learning_rate": 2.760796797978804e-07, + "loss": 0.0639, + "step": 41650 + }, + { + "epoch": 0.9763650322028634, + "grad_norm": 0.7096765041351318, + "learning_rate": 2.755329680390162e-07, + "loss": 0.064, + "step": 41651 + }, + { + "epoch": 0.9763884737776684, + "grad_norm": 0.4618309736251831, + "learning_rate": 2.749867973879106e-07, + "loss": 0.4773, + "step": 41652 + }, + { + "epoch": 0.9764119153524734, + "grad_norm": 0.22300712764263153, + "learning_rate": 2.7444116784752784e-07, + "loss": 0.0397, + "step": 41653 + }, + { + "epoch": 0.9764353569272783, + "grad_norm": 0.6337341666221619, + "learning_rate": 2.738960794208323e-07, + "loss": 0.5137, + "step": 41654 + }, + { + "epoch": 0.9764587985020834, + "grad_norm": 0.4879240095615387, + "learning_rate": 2.7335153211077713e-07, + "loss": 0.078, + "step": 41655 + }, + { + "epoch": 0.9764822400768883, + "grad_norm": 0.37400689721107483, + "learning_rate": 2.7280752592032664e-07, + "loss": 0.0577, + "step": 41656 + }, + { + "epoch": 0.9765056816516934, + "grad_norm": 0.3898325264453888, + "learning_rate": 2.72264060852423e-07, + "loss": 0.0653, + "step": 41657 + }, + { + "epoch": 0.9765291232264983, + "grad_norm": 0.6666608452796936, + "learning_rate": 2.717211369100192e-07, + "loss": 0.1469, + "step": 41658 + }, + { + "epoch": 0.9765525648013034, + "grad_norm": 0.4081328213214874, + "learning_rate": 2.711787540960575e-07, + "loss": 0.0551, + "step": 41659 + }, + { + "epoch": 0.9765760063761083, + "grad_norm": 0.18686863780021667, + "learning_rate": 2.706369124134911e-07, + "loss": 0.0416, + "step": 41660 + }, + { + "epoch": 0.9765994479509134, + "grad_norm": 0.7140713930130005, + "learning_rate": 2.700956118652509e-07, + "loss": 0.1278, + "step": 41661 + }, + { + "epoch": 0.9766228895257183, + "grad_norm": 0.30068737268447876, + "learning_rate": 2.69554852454279e-07, + "loss": 0.0366, + "step": 41662 + }, + { + "epoch": 0.9766463311005233, + "grad_norm": 0.7308823466300964, + "learning_rate": 2.690146341835065e-07, + "loss": 0.1153, + "step": 41663 + }, + { + "epoch": 0.9766697726753283, + "grad_norm": 0.3709397614002228, + "learning_rate": 2.6847495705585314e-07, + "loss": 0.042, + "step": 41664 + }, + { + "epoch": 0.9766932142501333, + "grad_norm": 0.5033468008041382, + "learning_rate": 2.679358210742722e-07, + "loss": 0.1131, + "step": 41665 + }, + { + "epoch": 0.9767166558249383, + "grad_norm": 0.3912692368030548, + "learning_rate": 2.6739722624166133e-07, + "loss": 0.0256, + "step": 41666 + }, + { + "epoch": 0.9767400973997433, + "grad_norm": 0.6530262231826782, + "learning_rate": 2.668591725609737e-07, + "loss": 0.0952, + "step": 41667 + }, + { + "epoch": 0.9767635389745483, + "grad_norm": 0.40383586287498474, + "learning_rate": 2.6632166003510703e-07, + "loss": 0.0723, + "step": 41668 + }, + { + "epoch": 0.9767869805493533, + "grad_norm": 0.5045298337936401, + "learning_rate": 2.6578468866697015e-07, + "loss": 0.1105, + "step": 41669 + }, + { + "epoch": 0.9768104221241583, + "grad_norm": 0.3194981515407562, + "learning_rate": 2.6524825845950505e-07, + "loss": 0.0669, + "step": 41670 + }, + { + "epoch": 0.9768338636989633, + "grad_norm": 0.2683860957622528, + "learning_rate": 2.647123694155984e-07, + "loss": 0.0313, + "step": 41671 + }, + { + "epoch": 0.9768573052737682, + "grad_norm": 0.2770397961139679, + "learning_rate": 2.6417702153818114e-07, + "loss": 0.0293, + "step": 41672 + }, + { + "epoch": 0.9768807468485733, + "grad_norm": 0.42623940110206604, + "learning_rate": 2.6364221483012875e-07, + "loss": 0.0562, + "step": 41673 + }, + { + "epoch": 0.9769041884233782, + "grad_norm": 0.1418055146932602, + "learning_rate": 2.631079492943611e-07, + "loss": 0.0214, + "step": 41674 + }, + { + "epoch": 0.9769276299981833, + "grad_norm": 0.28074878454208374, + "learning_rate": 2.625742249337759e-07, + "loss": 0.0493, + "step": 41675 + }, + { + "epoch": 0.9769510715729883, + "grad_norm": 0.4127236306667328, + "learning_rate": 2.6204104175127084e-07, + "loss": 0.0407, + "step": 41676 + }, + { + "epoch": 0.9769745131477933, + "grad_norm": 0.12179816514253616, + "learning_rate": 2.615083997497325e-07, + "loss": 0.0181, + "step": 41677 + }, + { + "epoch": 0.9769979547225983, + "grad_norm": 0.27831587195396423, + "learning_rate": 2.6097629893204744e-07, + "loss": 0.042, + "step": 41678 + }, + { + "epoch": 0.9770213962974033, + "grad_norm": 0.8729441165924072, + "learning_rate": 2.6044473930112445e-07, + "loss": 0.134, + "step": 41679 + }, + { + "epoch": 0.9770448378722083, + "grad_norm": 0.2893163561820984, + "learning_rate": 2.599137208598168e-07, + "loss": 0.0309, + "step": 41680 + }, + { + "epoch": 0.9770682794470132, + "grad_norm": 0.5178006887435913, + "learning_rate": 2.5938324361103326e-07, + "loss": 0.5877, + "step": 41681 + }, + { + "epoch": 0.9770917210218183, + "grad_norm": 0.1207503229379654, + "learning_rate": 2.5885330755763826e-07, + "loss": 0.0154, + "step": 41682 + }, + { + "epoch": 0.9771151625966232, + "grad_norm": 0.24075086414813995, + "learning_rate": 2.583239127025072e-07, + "loss": 0.0234, + "step": 41683 + }, + { + "epoch": 0.9771386041714283, + "grad_norm": 0.3204386830329895, + "learning_rate": 2.5779505904851563e-07, + "loss": 0.0643, + "step": 41684 + }, + { + "epoch": 0.9771620457462332, + "grad_norm": 0.4196198880672455, + "learning_rate": 2.572667465985279e-07, + "loss": 0.0674, + "step": 41685 + }, + { + "epoch": 0.9771854873210383, + "grad_norm": 0.4539051651954651, + "learning_rate": 2.5673897535541947e-07, + "loss": 0.0402, + "step": 41686 + }, + { + "epoch": 0.9772089288958432, + "grad_norm": 0.07033985108137131, + "learning_rate": 2.562117453220547e-07, + "loss": 0.01, + "step": 41687 + }, + { + "epoch": 0.9772323704706483, + "grad_norm": 0.17921559512615204, + "learning_rate": 2.5568505650128696e-07, + "loss": 0.034, + "step": 41688 + }, + { + "epoch": 0.9772558120454532, + "grad_norm": 0.26471737027168274, + "learning_rate": 2.551589088959694e-07, + "loss": 0.0292, + "step": 41689 + }, + { + "epoch": 0.9772792536202582, + "grad_norm": 0.4766101837158203, + "learning_rate": 2.5463330250896644e-07, + "loss": 0.1044, + "step": 41690 + }, + { + "epoch": 0.9773026951950632, + "grad_norm": 0.2978250980377197, + "learning_rate": 2.541082373431314e-07, + "loss": 0.0176, + "step": 41691 + }, + { + "epoch": 0.9773261367698682, + "grad_norm": 0.2810971736907959, + "learning_rate": 2.535837134013175e-07, + "loss": 0.0201, + "step": 41692 + }, + { + "epoch": 0.9773495783446732, + "grad_norm": 0.14783339202404022, + "learning_rate": 2.530597306863447e-07, + "loss": 0.0301, + "step": 41693 + }, + { + "epoch": 0.9773730199194782, + "grad_norm": 0.6864621043205261, + "learning_rate": 2.5253628920108853e-07, + "loss": 0.1084, + "step": 41694 + }, + { + "epoch": 0.9773964614942832, + "grad_norm": 0.5217574834823608, + "learning_rate": 2.520133889483689e-07, + "loss": 0.0797, + "step": 41695 + }, + { + "epoch": 0.9774199030690882, + "grad_norm": 0.5060504078865051, + "learning_rate": 2.5149102993103914e-07, + "loss": 0.0751, + "step": 41696 + }, + { + "epoch": 0.9774433446438932, + "grad_norm": 0.2519475221633911, + "learning_rate": 2.5096921215190803e-07, + "loss": 0.0414, + "step": 41697 + }, + { + "epoch": 0.9774667862186982, + "grad_norm": 0.16265366971492767, + "learning_rate": 2.5044793561384006e-07, + "loss": 0.0246, + "step": 41698 + }, + { + "epoch": 0.9774902277935031, + "grad_norm": 0.5123361349105835, + "learning_rate": 2.499272003196329e-07, + "loss": 0.078, + "step": 41699 + }, + { + "epoch": 0.9775136693683082, + "grad_norm": 0.4811262786388397, + "learning_rate": 2.4940700627212876e-07, + "loss": 0.3063, + "step": 41700 + }, + { + "epoch": 0.9775371109431131, + "grad_norm": 0.4228725731372833, + "learning_rate": 2.488873534741476e-07, + "loss": 0.0423, + "step": 41701 + }, + { + "epoch": 0.9775605525179182, + "grad_norm": 0.41002780199050903, + "learning_rate": 2.483682419285094e-07, + "loss": 0.059, + "step": 41702 + }, + { + "epoch": 0.9775839940927231, + "grad_norm": 0.5181585550308228, + "learning_rate": 2.4784967163802297e-07, + "loss": 0.0811, + "step": 41703 + }, + { + "epoch": 0.9776074356675282, + "grad_norm": 0.18140453100204468, + "learning_rate": 2.4733164260551945e-07, + "loss": 0.0223, + "step": 41704 + }, + { + "epoch": 0.9776308772423331, + "grad_norm": 0.10853242874145508, + "learning_rate": 2.4681415483379653e-07, + "loss": 0.0111, + "step": 41705 + }, + { + "epoch": 0.9776543188171382, + "grad_norm": 0.3209123909473419, + "learning_rate": 2.4629720832566316e-07, + "loss": 0.0667, + "step": 41706 + }, + { + "epoch": 0.9776777603919431, + "grad_norm": 0.18112386763095856, + "learning_rate": 2.457808030839281e-07, + "loss": 0.0173, + "step": 41707 + }, + { + "epoch": 0.9777012019667481, + "grad_norm": 0.2854728400707245, + "learning_rate": 2.4526493911138926e-07, + "loss": 0.0232, + "step": 41708 + }, + { + "epoch": 0.9777246435415531, + "grad_norm": 0.36559784412384033, + "learning_rate": 2.447496164108443e-07, + "loss": 0.0546, + "step": 41709 + }, + { + "epoch": 0.9777480851163581, + "grad_norm": 0.48719289898872375, + "learning_rate": 2.442348349850909e-07, + "loss": 0.1042, + "step": 41710 + }, + { + "epoch": 0.9777715266911631, + "grad_norm": 0.4668625593185425, + "learning_rate": 2.437205948369381e-07, + "loss": 0.0693, + "step": 41711 + }, + { + "epoch": 0.9777949682659681, + "grad_norm": 0.5491011738777161, + "learning_rate": 2.4320689596915025e-07, + "loss": 0.1041, + "step": 41712 + }, + { + "epoch": 0.9778184098407731, + "grad_norm": 0.2433517873287201, + "learning_rate": 2.4269373838452513e-07, + "loss": 0.0535, + "step": 41713 + }, + { + "epoch": 0.9778418514155781, + "grad_norm": 0.45258629322052, + "learning_rate": 2.421811220858605e-07, + "loss": 0.0525, + "step": 41714 + }, + { + "epoch": 0.977865292990383, + "grad_norm": 0.5823667049407959, + "learning_rate": 2.4166904707592084e-07, + "loss": 0.0497, + "step": 41715 + }, + { + "epoch": 0.9778887345651881, + "grad_norm": 0.49456027150154114, + "learning_rate": 2.4115751335748173e-07, + "loss": 0.5085, + "step": 41716 + }, + { + "epoch": 0.977912176139993, + "grad_norm": 0.12393028289079666, + "learning_rate": 2.406465209333297e-07, + "loss": 0.0218, + "step": 41717 + }, + { + "epoch": 0.9779356177147981, + "grad_norm": 0.5907318592071533, + "learning_rate": 2.401360698062405e-07, + "loss": 0.5387, + "step": 41718 + }, + { + "epoch": 0.977959059289603, + "grad_norm": 0.4389694035053253, + "learning_rate": 2.396261599789784e-07, + "loss": 0.0875, + "step": 41719 + }, + { + "epoch": 0.9779825008644081, + "grad_norm": 0.5021017789840698, + "learning_rate": 2.3911679145430797e-07, + "loss": 0.0876, + "step": 41720 + }, + { + "epoch": 0.978005942439213, + "grad_norm": 0.3611912429332733, + "learning_rate": 2.3860796423499364e-07, + "loss": 0.0756, + "step": 41721 + }, + { + "epoch": 0.9780293840140181, + "grad_norm": 0.11542624235153198, + "learning_rate": 2.3809967832379986e-07, + "loss": 0.0131, + "step": 41722 + }, + { + "epoch": 0.978052825588823, + "grad_norm": 0.4766249656677246, + "learning_rate": 2.3759193372347998e-07, + "loss": 0.0849, + "step": 41723 + }, + { + "epoch": 0.978076267163628, + "grad_norm": 0.33693480491638184, + "learning_rate": 2.3708473043678737e-07, + "loss": 0.0283, + "step": 41724 + }, + { + "epoch": 0.978099708738433, + "grad_norm": 0.4373353123664856, + "learning_rate": 2.3657806846648645e-07, + "loss": 0.0895, + "step": 41725 + }, + { + "epoch": 0.978123150313238, + "grad_norm": 0.47459739446640015, + "learning_rate": 2.360719478153084e-07, + "loss": 0.0863, + "step": 41726 + }, + { + "epoch": 0.9781465918880431, + "grad_norm": 0.3588717579841614, + "learning_rate": 2.3556636848601765e-07, + "loss": 0.0502, + "step": 41727 + }, + { + "epoch": 0.978170033462848, + "grad_norm": 0.23051586747169495, + "learning_rate": 2.3506133048134537e-07, + "loss": 0.0451, + "step": 41728 + }, + { + "epoch": 0.9781934750376531, + "grad_norm": 0.49598944187164307, + "learning_rate": 2.3455683380403383e-07, + "loss": 0.1346, + "step": 41729 + }, + { + "epoch": 0.978216916612458, + "grad_norm": 0.4369044303894043, + "learning_rate": 2.3405287845682522e-07, + "loss": 0.0592, + "step": 41730 + }, + { + "epoch": 0.9782403581872631, + "grad_norm": 0.45961815118789673, + "learning_rate": 2.3354946444245074e-07, + "loss": 0.0776, + "step": 41731 + }, + { + "epoch": 0.978263799762068, + "grad_norm": 0.5683275461196899, + "learning_rate": 2.3304659176364153e-07, + "loss": 0.0814, + "step": 41732 + }, + { + "epoch": 0.9782872413368731, + "grad_norm": 0.47863173484802246, + "learning_rate": 2.325442604231287e-07, + "loss": 0.1051, + "step": 41733 + }, + { + "epoch": 0.978310682911678, + "grad_norm": 0.4075719118118286, + "learning_rate": 2.3204247042363235e-07, + "loss": 0.0426, + "step": 41734 + }, + { + "epoch": 0.978334124486483, + "grad_norm": 0.05670875683426857, + "learning_rate": 2.315412217678725e-07, + "loss": 0.007, + "step": 41735 + }, + { + "epoch": 0.978357566061288, + "grad_norm": 0.09786340594291687, + "learning_rate": 2.3104051445859144e-07, + "loss": 0.0084, + "step": 41736 + }, + { + "epoch": 0.978381007636093, + "grad_norm": 0.9284936785697937, + "learning_rate": 2.3054034849847584e-07, + "loss": 0.0896, + "step": 41737 + }, + { + "epoch": 0.978404449210898, + "grad_norm": 0.5255183577537537, + "learning_rate": 2.3004072389025687e-07, + "loss": 0.5593, + "step": 41738 + }, + { + "epoch": 0.978427890785703, + "grad_norm": 0.3728402554988861, + "learning_rate": 2.2954164063665463e-07, + "loss": 0.1029, + "step": 41739 + }, + { + "epoch": 0.978451332360508, + "grad_norm": 0.39257070422172546, + "learning_rate": 2.2904309874034468e-07, + "loss": 0.0553, + "step": 41740 + }, + { + "epoch": 0.978474773935313, + "grad_norm": 0.3896694481372833, + "learning_rate": 2.2854509820406932e-07, + "loss": 0.0439, + "step": 41741 + }, + { + "epoch": 0.978498215510118, + "grad_norm": 0.2795068025588989, + "learning_rate": 2.280476390305042e-07, + "loss": 0.0461, + "step": 41742 + }, + { + "epoch": 0.978521657084923, + "grad_norm": 0.5055195689201355, + "learning_rate": 2.275507212223582e-07, + "loss": 0.0985, + "step": 41743 + }, + { + "epoch": 0.9785450986597279, + "grad_norm": 0.6130907535552979, + "learning_rate": 2.2705434478232923e-07, + "loss": 0.1115, + "step": 41744 + }, + { + "epoch": 0.978568540234533, + "grad_norm": 0.4535430371761322, + "learning_rate": 2.26558509713104e-07, + "loss": 0.0981, + "step": 41745 + }, + { + "epoch": 0.9785919818093379, + "grad_norm": 0.0956551656126976, + "learning_rate": 2.2606321601738034e-07, + "loss": 0.0214, + "step": 41746 + }, + { + "epoch": 0.978615423384143, + "grad_norm": 0.378464937210083, + "learning_rate": 2.2556846369784502e-07, + "loss": 0.0245, + "step": 41747 + }, + { + "epoch": 0.9786388649589479, + "grad_norm": 0.26001086831092834, + "learning_rate": 2.2507425275718476e-07, + "loss": 0.026, + "step": 41748 + }, + { + "epoch": 0.978662306533753, + "grad_norm": 0.634005606174469, + "learning_rate": 2.2458058319807517e-07, + "loss": 0.0792, + "step": 41749 + }, + { + "epoch": 0.9786857481085579, + "grad_norm": 0.39006760716438293, + "learning_rate": 2.2408745502319194e-07, + "loss": 0.0504, + "step": 41750 + }, + { + "epoch": 0.978709189683363, + "grad_norm": 0.2083665132522583, + "learning_rate": 2.235948682352218e-07, + "loss": 0.0394, + "step": 41751 + }, + { + "epoch": 0.9787326312581679, + "grad_norm": 0.1549803912639618, + "learning_rate": 2.2310282283682927e-07, + "loss": 0.0211, + "step": 41752 + }, + { + "epoch": 0.978756072832973, + "grad_norm": 0.24458302557468414, + "learning_rate": 2.2261131883067888e-07, + "loss": 0.0363, + "step": 41753 + }, + { + "epoch": 0.9787795144077779, + "grad_norm": 0.4700598120689392, + "learning_rate": 2.2212035621944628e-07, + "loss": 0.4452, + "step": 41754 + }, + { + "epoch": 0.9788029559825829, + "grad_norm": 0.4467019736766815, + "learning_rate": 2.21629935005796e-07, + "loss": 0.0591, + "step": 41755 + }, + { + "epoch": 0.9788263975573879, + "grad_norm": 0.5539354681968689, + "learning_rate": 2.2114005519239257e-07, + "loss": 0.0819, + "step": 41756 + }, + { + "epoch": 0.9788498391321929, + "grad_norm": 0.45893043279647827, + "learning_rate": 2.2065071678188943e-07, + "loss": 0.0773, + "step": 41757 + }, + { + "epoch": 0.9788732807069979, + "grad_norm": 0.12377826869487762, + "learning_rate": 2.2016191977694e-07, + "loss": 0.0204, + "step": 41758 + }, + { + "epoch": 0.9788967222818029, + "grad_norm": 0.2810186445713043, + "learning_rate": 2.1967366418019775e-07, + "loss": 0.0546, + "step": 41759 + }, + { + "epoch": 0.9789201638566079, + "grad_norm": 0.6470649242401123, + "learning_rate": 2.1918594999430497e-07, + "loss": 0.1232, + "step": 41760 + }, + { + "epoch": 0.9789436054314129, + "grad_norm": 0.03418358042836189, + "learning_rate": 2.1869877722191512e-07, + "loss": 0.0025, + "step": 41761 + }, + { + "epoch": 0.9789670470062178, + "grad_norm": 0.6805321574211121, + "learning_rate": 2.1821214586568161e-07, + "loss": 0.118, + "step": 41762 + }, + { + "epoch": 0.9789904885810229, + "grad_norm": 0.35828280448913574, + "learning_rate": 2.177260559282246e-07, + "loss": 0.0248, + "step": 41763 + }, + { + "epoch": 0.9790139301558278, + "grad_norm": 0.42551133036613464, + "learning_rate": 2.1724050741219748e-07, + "loss": 0.0728, + "step": 41764 + }, + { + "epoch": 0.9790373717306329, + "grad_norm": 0.6267035007476807, + "learning_rate": 2.167555003202315e-07, + "loss": 0.0451, + "step": 41765 + }, + { + "epoch": 0.9790608133054378, + "grad_norm": 0.5521897077560425, + "learning_rate": 2.162710346549468e-07, + "loss": 0.0964, + "step": 41766 + }, + { + "epoch": 0.9790842548802429, + "grad_norm": 0.15908940136432648, + "learning_rate": 2.157871104189857e-07, + "loss": 0.0166, + "step": 41767 + }, + { + "epoch": 0.9791076964550478, + "grad_norm": 0.6795697212219238, + "learning_rate": 2.153037276149683e-07, + "loss": 0.1065, + "step": 41768 + }, + { + "epoch": 0.9791311380298529, + "grad_norm": 0.6145167946815491, + "learning_rate": 2.1482088624551477e-07, + "loss": 0.6702, + "step": 41769 + }, + { + "epoch": 0.9791545796046578, + "grad_norm": 0.7316187024116516, + "learning_rate": 2.1433858631325632e-07, + "loss": 0.0588, + "step": 41770 + }, + { + "epoch": 0.9791780211794628, + "grad_norm": 0.6388740539550781, + "learning_rate": 2.1385682782080197e-07, + "loss": 0.6105, + "step": 41771 + }, + { + "epoch": 0.9792014627542678, + "grad_norm": 0.12137464433908463, + "learning_rate": 2.1337561077076073e-07, + "loss": 0.023, + "step": 41772 + }, + { + "epoch": 0.9792249043290728, + "grad_norm": 0.49717026948928833, + "learning_rate": 2.1289493516575276e-07, + "loss": 0.1035, + "step": 41773 + }, + { + "epoch": 0.9792483459038778, + "grad_norm": 0.7232780456542969, + "learning_rate": 2.1241480100837597e-07, + "loss": 0.0567, + "step": 41774 + }, + { + "epoch": 0.9792717874786828, + "grad_norm": 0.3369750678539276, + "learning_rate": 2.1193520830125046e-07, + "loss": 0.0431, + "step": 41775 + }, + { + "epoch": 0.9792952290534878, + "grad_norm": 0.5846775770187378, + "learning_rate": 2.114561570469742e-07, + "loss": 0.5881, + "step": 41776 + }, + { + "epoch": 0.9793186706282928, + "grad_norm": 0.1682887077331543, + "learning_rate": 2.1097764724813395e-07, + "loss": 0.0188, + "step": 41777 + }, + { + "epoch": 0.9793421122030979, + "grad_norm": 0.25380200147628784, + "learning_rate": 2.104996789073388e-07, + "loss": 0.2362, + "step": 41778 + }, + { + "epoch": 0.9793655537779028, + "grad_norm": 0.36655551195144653, + "learning_rate": 2.100222520271755e-07, + "loss": 0.0468, + "step": 41779 + }, + { + "epoch": 0.9793889953527078, + "grad_norm": 0.6159796118736267, + "learning_rate": 2.0954536661024205e-07, + "loss": 0.0309, + "step": 41780 + }, + { + "epoch": 0.9794124369275128, + "grad_norm": 0.24582011997699738, + "learning_rate": 2.0906902265911409e-07, + "loss": 0.0626, + "step": 41781 + }, + { + "epoch": 0.9794358785023178, + "grad_norm": 0.11944077163934708, + "learning_rate": 2.085932201763896e-07, + "loss": 0.016, + "step": 41782 + }, + { + "epoch": 0.9794593200771228, + "grad_norm": 0.21626420319080353, + "learning_rate": 2.0811795916464427e-07, + "loss": 0.0357, + "step": 41783 + }, + { + "epoch": 0.9794827616519278, + "grad_norm": 0.3044677674770355, + "learning_rate": 2.076432396264538e-07, + "loss": 0.0296, + "step": 41784 + }, + { + "epoch": 0.9795062032267328, + "grad_norm": 0.7373274564743042, + "learning_rate": 2.0716906156439396e-07, + "loss": 0.0637, + "step": 41785 + }, + { + "epoch": 0.9795296448015378, + "grad_norm": 0.33993831276893616, + "learning_rate": 2.0669542498104043e-07, + "loss": 0.0613, + "step": 41786 + }, + { + "epoch": 0.9795530863763428, + "grad_norm": 0.07886166125535965, + "learning_rate": 2.0622232987896896e-07, + "loss": 0.0167, + "step": 41787 + }, + { + "epoch": 0.9795765279511478, + "grad_norm": 0.26023662090301514, + "learning_rate": 2.057497762607441e-07, + "loss": 0.0335, + "step": 41788 + }, + { + "epoch": 0.9795999695259527, + "grad_norm": 1.0747162103652954, + "learning_rate": 2.052777641289194e-07, + "loss": 0.0158, + "step": 41789 + }, + { + "epoch": 0.9796234111007578, + "grad_norm": 0.3037055432796478, + "learning_rate": 2.0480629348607062e-07, + "loss": 0.0377, + "step": 41790 + }, + { + "epoch": 0.9796468526755627, + "grad_norm": 0.4607141315937042, + "learning_rate": 2.043353643347401e-07, + "loss": 0.1152, + "step": 41791 + }, + { + "epoch": 0.9796702942503678, + "grad_norm": 0.5820479393005371, + "learning_rate": 2.038649766774925e-07, + "loss": 0.069, + "step": 41792 + }, + { + "epoch": 0.9796937358251727, + "grad_norm": 0.44832223653793335, + "learning_rate": 2.033951305168924e-07, + "loss": 0.0725, + "step": 41793 + }, + { + "epoch": 0.9797171773999778, + "grad_norm": 0.5330086350440979, + "learning_rate": 2.0292582585546005e-07, + "loss": 0.3793, + "step": 41794 + }, + { + "epoch": 0.9797406189747827, + "grad_norm": 0.2287791222333908, + "learning_rate": 2.0245706269577114e-07, + "loss": 0.02, + "step": 41795 + }, + { + "epoch": 0.9797640605495878, + "grad_norm": 0.16849206387996674, + "learning_rate": 2.0198884104034588e-07, + "loss": 0.0293, + "step": 41796 + }, + { + "epoch": 0.9797875021243927, + "grad_norm": 0.36833497881889343, + "learning_rate": 2.015211608917489e-07, + "loss": 0.0522, + "step": 41797 + }, + { + "epoch": 0.9798109436991977, + "grad_norm": 0.353887677192688, + "learning_rate": 2.0105402225248927e-07, + "loss": 0.0553, + "step": 41798 + }, + { + "epoch": 0.9798343852740027, + "grad_norm": 0.5799263119697571, + "learning_rate": 2.0058742512512053e-07, + "loss": 0.1024, + "step": 41799 + }, + { + "epoch": 0.9798578268488077, + "grad_norm": 0.4864993989467621, + "learning_rate": 2.0012136951217397e-07, + "loss": 0.0487, + "step": 41800 + }, + { + "epoch": 0.9798812684236127, + "grad_norm": 0.24931862950325012, + "learning_rate": 1.9965585541618093e-07, + "loss": 0.0378, + "step": 41801 + }, + { + "epoch": 0.9799047099984177, + "grad_norm": 0.5145075917243958, + "learning_rate": 1.991908828396505e-07, + "loss": 0.0252, + "step": 41802 + }, + { + "epoch": 0.9799281515732227, + "grad_norm": 0.397504061460495, + "learning_rate": 1.9872645178512505e-07, + "loss": 0.0849, + "step": 41803 + }, + { + "epoch": 0.9799515931480277, + "grad_norm": 0.5211818218231201, + "learning_rate": 1.9826256225510264e-07, + "loss": 0.0665, + "step": 41804 + }, + { + "epoch": 0.9799750347228326, + "grad_norm": 0.4288274943828583, + "learning_rate": 1.9779921425212568e-07, + "loss": 0.0535, + "step": 41805 + }, + { + "epoch": 0.9799984762976377, + "grad_norm": 0.43211594223976135, + "learning_rate": 1.9733640777869211e-07, + "loss": 0.0631, + "step": 41806 + }, + { + "epoch": 0.9800219178724426, + "grad_norm": 0.6754220128059387, + "learning_rate": 1.968741428373222e-07, + "loss": 0.0843, + "step": 41807 + }, + { + "epoch": 0.9800453594472477, + "grad_norm": 0.18418747186660767, + "learning_rate": 1.9641241943051393e-07, + "loss": 0.025, + "step": 41808 + }, + { + "epoch": 0.9800688010220526, + "grad_norm": 0.1457941085100174, + "learning_rate": 1.9595123756077638e-07, + "loss": 0.0167, + "step": 41809 + }, + { + "epoch": 0.9800922425968577, + "grad_norm": 0.2700909972190857, + "learning_rate": 1.9549059723061868e-07, + "loss": 0.0324, + "step": 41810 + }, + { + "epoch": 0.9801156841716626, + "grad_norm": 0.13889533281326294, + "learning_rate": 1.9503049844253885e-07, + "loss": 0.0179, + "step": 41811 + }, + { + "epoch": 0.9801391257464677, + "grad_norm": 0.11679656058549881, + "learning_rate": 1.9457094119902376e-07, + "loss": 0.0181, + "step": 41812 + }, + { + "epoch": 0.9801625673212726, + "grad_norm": 0.5925211310386658, + "learning_rate": 1.941119255025714e-07, + "loss": 0.1091, + "step": 41813 + }, + { + "epoch": 0.9801860088960777, + "grad_norm": 0.38912510871887207, + "learning_rate": 1.9365345135567982e-07, + "loss": 0.0843, + "step": 41814 + }, + { + "epoch": 0.9802094504708826, + "grad_norm": 0.47338351607322693, + "learning_rate": 1.9319551876082474e-07, + "loss": 0.1044, + "step": 41815 + }, + { + "epoch": 0.9802328920456876, + "grad_norm": 0.42892518639564514, + "learning_rate": 1.9273812772050425e-07, + "loss": 0.0362, + "step": 41816 + }, + { + "epoch": 0.9802563336204926, + "grad_norm": 0.33487287163734436, + "learning_rate": 1.92281278237183e-07, + "loss": 0.0383, + "step": 41817 + }, + { + "epoch": 0.9802797751952976, + "grad_norm": 0.20897117257118225, + "learning_rate": 1.91824970313359e-07, + "loss": 0.0341, + "step": 41818 + }, + { + "epoch": 0.9803032167701026, + "grad_norm": 0.4336984157562256, + "learning_rate": 1.9136920395149693e-07, + "loss": 0.3918, + "step": 41819 + }, + { + "epoch": 0.9803266583449076, + "grad_norm": 0.14764226973056793, + "learning_rate": 1.909139791540726e-07, + "loss": 0.0311, + "step": 41820 + }, + { + "epoch": 0.9803500999197126, + "grad_norm": 0.14184929430484772, + "learning_rate": 1.904592959235507e-07, + "loss": 0.0159, + "step": 41821 + }, + { + "epoch": 0.9803735414945176, + "grad_norm": 0.9894676804542542, + "learning_rate": 1.9000515426240706e-07, + "loss": 0.1947, + "step": 41822 + }, + { + "epoch": 0.9803969830693225, + "grad_norm": 0.4496782124042511, + "learning_rate": 1.8955155417309523e-07, + "loss": 0.0724, + "step": 41823 + }, + { + "epoch": 0.9804204246441276, + "grad_norm": 0.5223049521446228, + "learning_rate": 1.89098495658091e-07, + "loss": 0.0951, + "step": 41824 + }, + { + "epoch": 0.9804438662189325, + "grad_norm": 0.5802381038665771, + "learning_rate": 1.886459787198369e-07, + "loss": 0.1192, + "step": 41825 + }, + { + "epoch": 0.9804673077937376, + "grad_norm": 0.2681005299091339, + "learning_rate": 1.8819400336080873e-07, + "loss": 0.023, + "step": 41826 + }, + { + "epoch": 0.9804907493685425, + "grad_norm": 0.32889029383659363, + "learning_rate": 1.8774256958343784e-07, + "loss": 0.0453, + "step": 41827 + }, + { + "epoch": 0.9805141909433476, + "grad_norm": 0.3149087727069855, + "learning_rate": 1.8729167739017784e-07, + "loss": 0.0585, + "step": 41828 + }, + { + "epoch": 0.9805376325181526, + "grad_norm": 0.4638943672180176, + "learning_rate": 1.8684132678349342e-07, + "loss": 0.0696, + "step": 41829 + }, + { + "epoch": 0.9805610740929576, + "grad_norm": 0.2167852818965912, + "learning_rate": 1.8639151776580488e-07, + "loss": 0.0209, + "step": 41830 + }, + { + "epoch": 0.9805845156677626, + "grad_norm": 0.45815378427505493, + "learning_rate": 1.859422503395658e-07, + "loss": 0.1123, + "step": 41831 + }, + { + "epoch": 0.9806079572425676, + "grad_norm": 0.28688332438468933, + "learning_rate": 1.854935245072076e-07, + "loss": 0.068, + "step": 41832 + }, + { + "epoch": 0.9806313988173726, + "grad_norm": 0.18438710272312164, + "learning_rate": 1.850453402711616e-07, + "loss": 0.0233, + "step": 41833 + }, + { + "epoch": 0.9806548403921775, + "grad_norm": 0.21060670912265778, + "learning_rate": 1.8459769763388146e-07, + "loss": 0.031, + "step": 41834 + }, + { + "epoch": 0.9806782819669826, + "grad_norm": 0.2160222977399826, + "learning_rate": 1.8415059659777633e-07, + "loss": 0.0314, + "step": 41835 + }, + { + "epoch": 0.9807017235417875, + "grad_norm": 0.28851816058158875, + "learning_rate": 1.8370403716526652e-07, + "loss": 0.0369, + "step": 41836 + }, + { + "epoch": 0.9807251651165926, + "grad_norm": 0.3121459186077118, + "learning_rate": 1.832580193387945e-07, + "loss": 0.0519, + "step": 41837 + }, + { + "epoch": 0.9807486066913975, + "grad_norm": 0.3665255010128021, + "learning_rate": 1.8281254312076945e-07, + "loss": 0.0532, + "step": 41838 + }, + { + "epoch": 0.9807720482662026, + "grad_norm": 0.3361470401287079, + "learning_rate": 1.8236760851360056e-07, + "loss": 0.0604, + "step": 41839 + }, + { + "epoch": 0.9807954898410075, + "grad_norm": 0.20938746631145477, + "learning_rate": 1.8192321551973036e-07, + "loss": 0.0219, + "step": 41840 + }, + { + "epoch": 0.9808189314158126, + "grad_norm": 0.18128404021263123, + "learning_rate": 1.8147936414153466e-07, + "loss": 0.0198, + "step": 41841 + }, + { + "epoch": 0.9808423729906175, + "grad_norm": 0.3326958119869232, + "learning_rate": 1.810360543814449e-07, + "loss": 0.0647, + "step": 41842 + }, + { + "epoch": 0.9808658145654225, + "grad_norm": 0.3587381839752197, + "learning_rate": 1.8059328624185913e-07, + "loss": 0.0276, + "step": 41843 + }, + { + "epoch": 0.9808892561402275, + "grad_norm": 0.32830122113227844, + "learning_rate": 1.8015105972518653e-07, + "loss": 0.0517, + "step": 41844 + }, + { + "epoch": 0.9809126977150325, + "grad_norm": 0.423358291387558, + "learning_rate": 1.7970937483381412e-07, + "loss": 0.5213, + "step": 41845 + }, + { + "epoch": 0.9809361392898375, + "grad_norm": 0.2976880371570587, + "learning_rate": 1.7926823157015104e-07, + "loss": 0.0348, + "step": 41846 + }, + { + "epoch": 0.9809595808646425, + "grad_norm": 0.5657429099082947, + "learning_rate": 1.788276299365843e-07, + "loss": 0.1308, + "step": 41847 + }, + { + "epoch": 0.9809830224394475, + "grad_norm": 0.09596746414899826, + "learning_rate": 1.7838756993550087e-07, + "loss": 0.0132, + "step": 41848 + }, + { + "epoch": 0.9810064640142525, + "grad_norm": 0.14743973314762115, + "learning_rate": 1.7794805156929884e-07, + "loss": 0.0215, + "step": 41849 + }, + { + "epoch": 0.9810299055890574, + "grad_norm": 0.3688238263130188, + "learning_rate": 1.7750907484035407e-07, + "loss": 0.0611, + "step": 41850 + }, + { + "epoch": 0.9810533471638625, + "grad_norm": 0.06908730417490005, + "learning_rate": 1.7707063975105353e-07, + "loss": 0.0049, + "step": 41851 + }, + { + "epoch": 0.9810767887386674, + "grad_norm": 0.5897018313407898, + "learning_rate": 1.7663274630377314e-07, + "loss": 0.131, + "step": 41852 + }, + { + "epoch": 0.9811002303134725, + "grad_norm": 0.8064554929733276, + "learning_rate": 1.7619539450088872e-07, + "loss": 0.0222, + "step": 41853 + }, + { + "epoch": 0.9811236718882774, + "grad_norm": 0.5655439496040344, + "learning_rate": 1.7575858434477622e-07, + "loss": 0.1099, + "step": 41854 + }, + { + "epoch": 0.9811471134630825, + "grad_norm": 0.41689175367355347, + "learning_rate": 1.7532231583781145e-07, + "loss": 0.0664, + "step": 41855 + }, + { + "epoch": 0.9811705550378874, + "grad_norm": 0.2170831710100174, + "learning_rate": 1.74886588982337e-07, + "loss": 0.0397, + "step": 41856 + }, + { + "epoch": 0.9811939966126925, + "grad_norm": 0.6460263133049011, + "learning_rate": 1.74451403780751e-07, + "loss": 0.1488, + "step": 41857 + }, + { + "epoch": 0.9812174381874974, + "grad_norm": 0.40698501467704773, + "learning_rate": 1.7401676023539593e-07, + "loss": 0.0525, + "step": 41858 + }, + { + "epoch": 0.9812408797623025, + "grad_norm": 0.46351832151412964, + "learning_rate": 1.7358265834862553e-07, + "loss": 0.6138, + "step": 41859 + }, + { + "epoch": 0.9812643213371074, + "grad_norm": 0.26257652044296265, + "learning_rate": 1.7314909812280455e-07, + "loss": 0.0371, + "step": 41860 + }, + { + "epoch": 0.9812877629119124, + "grad_norm": 0.427310049533844, + "learning_rate": 1.727160795602867e-07, + "loss": 0.0722, + "step": 41861 + }, + { + "epoch": 0.9813112044867174, + "grad_norm": 0.5187690854072571, + "learning_rate": 1.7228360266342558e-07, + "loss": 0.0976, + "step": 41862 + }, + { + "epoch": 0.9813346460615224, + "grad_norm": 0.35107120871543884, + "learning_rate": 1.7185166743454162e-07, + "loss": 0.0412, + "step": 41863 + }, + { + "epoch": 0.9813580876363274, + "grad_norm": 0.4464735984802246, + "learning_rate": 1.7142027387601068e-07, + "loss": 0.0924, + "step": 41864 + }, + { + "epoch": 0.9813815292111324, + "grad_norm": 0.4025188982486725, + "learning_rate": 1.7098942199016422e-07, + "loss": 0.0628, + "step": 41865 + }, + { + "epoch": 0.9814049707859374, + "grad_norm": 0.5106678605079651, + "learning_rate": 1.705591117793226e-07, + "loss": 0.1142, + "step": 41866 + }, + { + "epoch": 0.9814284123607424, + "grad_norm": 0.6455875039100647, + "learning_rate": 1.7012934324585063e-07, + "loss": 0.1488, + "step": 41867 + }, + { + "epoch": 0.9814518539355473, + "grad_norm": 0.3762390911579132, + "learning_rate": 1.697001163920464e-07, + "loss": 0.3808, + "step": 41868 + }, + { + "epoch": 0.9814752955103524, + "grad_norm": 0.4842541217803955, + "learning_rate": 1.6927143122026367e-07, + "loss": 0.2783, + "step": 41869 + }, + { + "epoch": 0.9814987370851573, + "grad_norm": 0.42163053154945374, + "learning_rate": 1.6884328773282276e-07, + "loss": 0.2388, + "step": 41870 + }, + { + "epoch": 0.9815221786599624, + "grad_norm": 0.41661643981933594, + "learning_rate": 1.6841568593204405e-07, + "loss": 0.0638, + "step": 41871 + }, + { + "epoch": 0.9815456202347673, + "grad_norm": 0.30764511227607727, + "learning_rate": 1.6798862582025898e-07, + "loss": 0.0338, + "step": 41872 + }, + { + "epoch": 0.9815690618095724, + "grad_norm": 0.4225918650627136, + "learning_rate": 1.6756210739975465e-07, + "loss": 0.116, + "step": 41873 + }, + { + "epoch": 0.9815925033843773, + "grad_norm": 0.45505601167678833, + "learning_rate": 1.671361306728736e-07, + "loss": 0.1068, + "step": 41874 + }, + { + "epoch": 0.9816159449591824, + "grad_norm": 0.20018638670444489, + "learning_rate": 1.667106956419251e-07, + "loss": 0.0424, + "step": 41875 + }, + { + "epoch": 0.9816393865339873, + "grad_norm": 0.6080533862113953, + "learning_rate": 1.6628580230920732e-07, + "loss": 0.1823, + "step": 41876 + }, + { + "epoch": 0.9816628281087924, + "grad_norm": 0.49293622374534607, + "learning_rate": 1.6586145067702951e-07, + "loss": 0.0664, + "step": 41877 + }, + { + "epoch": 0.9816862696835973, + "grad_norm": 0.283052533864975, + "learning_rate": 1.6543764074770096e-07, + "loss": 0.0535, + "step": 41878 + }, + { + "epoch": 0.9817097112584023, + "grad_norm": 0.3093879520893097, + "learning_rate": 1.650143725235087e-07, + "loss": 0.0484, + "step": 41879 + }, + { + "epoch": 0.9817331528332074, + "grad_norm": 0.5963696837425232, + "learning_rate": 1.6459164600675091e-07, + "loss": 0.1103, + "step": 41880 + }, + { + "epoch": 0.9817565944080123, + "grad_norm": 0.0908840224146843, + "learning_rate": 1.6416946119972576e-07, + "loss": 0.012, + "step": 41881 + }, + { + "epoch": 0.9817800359828174, + "grad_norm": 0.2164912223815918, + "learning_rate": 1.6374781810473138e-07, + "loss": 0.0184, + "step": 41882 + }, + { + "epoch": 0.9818034775576223, + "grad_norm": 0.1325928419828415, + "learning_rate": 1.6332671672404376e-07, + "loss": 0.015, + "step": 41883 + }, + { + "epoch": 0.9818269191324274, + "grad_norm": 0.3183380365371704, + "learning_rate": 1.6290615705996104e-07, + "loss": 0.0425, + "step": 41884 + }, + { + "epoch": 0.9818503607072323, + "grad_norm": 0.3188893795013428, + "learning_rate": 1.6248613911474807e-07, + "loss": 0.0284, + "step": 41885 + }, + { + "epoch": 0.9818738022820374, + "grad_norm": 0.3561600148677826, + "learning_rate": 1.6206666289070304e-07, + "loss": 0.0513, + "step": 41886 + }, + { + "epoch": 0.9818972438568423, + "grad_norm": 0.17438000440597534, + "learning_rate": 1.616477283900797e-07, + "loss": 0.0177, + "step": 41887 + }, + { + "epoch": 0.9819206854316473, + "grad_norm": 0.4373629689216614, + "learning_rate": 1.6122933561516505e-07, + "loss": 0.0669, + "step": 41888 + }, + { + "epoch": 0.9819441270064523, + "grad_norm": 0.4049380421638489, + "learning_rate": 1.6081148456822404e-07, + "loss": 0.04, + "step": 41889 + }, + { + "epoch": 0.9819675685812573, + "grad_norm": 0.47218605875968933, + "learning_rate": 1.6039417525153255e-07, + "loss": 0.0868, + "step": 41890 + }, + { + "epoch": 0.9819910101560623, + "grad_norm": 0.21333301067352295, + "learning_rate": 1.5997740766735547e-07, + "loss": 0.0382, + "step": 41891 + }, + { + "epoch": 0.9820144517308673, + "grad_norm": 0.4210343658924103, + "learning_rate": 1.5956118181793544e-07, + "loss": 0.0642, + "step": 41892 + }, + { + "epoch": 0.9820378933056723, + "grad_norm": 0.3302364647388458, + "learning_rate": 1.5914549770554842e-07, + "loss": 0.1999, + "step": 41893 + }, + { + "epoch": 0.9820613348804773, + "grad_norm": 0.4535204768180847, + "learning_rate": 1.5873035533244819e-07, + "loss": 0.1076, + "step": 41894 + }, + { + "epoch": 0.9820847764552822, + "grad_norm": 0.29726317524909973, + "learning_rate": 1.5831575470088843e-07, + "loss": 0.0747, + "step": 41895 + }, + { + "epoch": 0.9821082180300873, + "grad_norm": 0.29123786091804504, + "learning_rate": 1.5790169581311188e-07, + "loss": 0.0456, + "step": 41896 + }, + { + "epoch": 0.9821316596048922, + "grad_norm": 0.18023471534252167, + "learning_rate": 1.5748817867136112e-07, + "loss": 0.0254, + "step": 41897 + }, + { + "epoch": 0.9821551011796973, + "grad_norm": 0.1395614743232727, + "learning_rate": 1.5707520327788993e-07, + "loss": 0.0116, + "step": 41898 + }, + { + "epoch": 0.9821785427545022, + "grad_norm": 0.7769050598144531, + "learning_rate": 1.5666276963494097e-07, + "loss": 0.2062, + "step": 41899 + }, + { + "epoch": 0.9822019843293073, + "grad_norm": 0.4429590106010437, + "learning_rate": 1.5625087774473468e-07, + "loss": 0.0751, + "step": 41900 + }, + { + "epoch": 0.9822254259041122, + "grad_norm": 0.536106526851654, + "learning_rate": 1.558395276095359e-07, + "loss": 0.0968, + "step": 41901 + }, + { + "epoch": 0.9822488674789173, + "grad_norm": 0.4624122083187103, + "learning_rate": 1.554287192315429e-07, + "loss": 0.0841, + "step": 41902 + }, + { + "epoch": 0.9822723090537222, + "grad_norm": 0.604324460029602, + "learning_rate": 1.550184526130094e-07, + "loss": 0.0799, + "step": 41903 + }, + { + "epoch": 0.9822957506285273, + "grad_norm": 0.20886069536209106, + "learning_rate": 1.5460872775615586e-07, + "loss": 0.0432, + "step": 41904 + }, + { + "epoch": 0.9823191922033322, + "grad_norm": 0.30205410718917847, + "learning_rate": 1.5419954466320274e-07, + "loss": 0.0268, + "step": 41905 + }, + { + "epoch": 0.9823426337781372, + "grad_norm": 0.5517333149909973, + "learning_rate": 1.5379090333635937e-07, + "loss": 0.0989, + "step": 41906 + }, + { + "epoch": 0.9823660753529422, + "grad_norm": 0.7017940282821655, + "learning_rate": 1.533828037778684e-07, + "loss": 0.053, + "step": 41907 + }, + { + "epoch": 0.9823895169277472, + "grad_norm": 0.3170270025730133, + "learning_rate": 1.5297524598991697e-07, + "loss": 0.0435, + "step": 41908 + }, + { + "epoch": 0.9824129585025522, + "grad_norm": 0.08042744547128677, + "learning_rate": 1.5256822997473664e-07, + "loss": 0.0104, + "step": 41909 + }, + { + "epoch": 0.9824364000773572, + "grad_norm": 0.36841756105422974, + "learning_rate": 1.5216175573453672e-07, + "loss": 0.0558, + "step": 41910 + }, + { + "epoch": 0.9824598416521622, + "grad_norm": 0.2432529777288437, + "learning_rate": 1.5175582327150439e-07, + "loss": 0.0474, + "step": 41911 + }, + { + "epoch": 0.9824832832269672, + "grad_norm": 0.2205418348312378, + "learning_rate": 1.5135043258784899e-07, + "loss": 0.0365, + "step": 41912 + }, + { + "epoch": 0.9825067248017721, + "grad_norm": 0.5623947978019714, + "learning_rate": 1.5094558368577982e-07, + "loss": 0.0845, + "step": 41913 + }, + { + "epoch": 0.9825301663765772, + "grad_norm": 0.5618204474449158, + "learning_rate": 1.5054127656748406e-07, + "loss": 0.0716, + "step": 41914 + }, + { + "epoch": 0.9825536079513821, + "grad_norm": 0.49950358271598816, + "learning_rate": 1.5013751123515995e-07, + "loss": 0.0928, + "step": 41915 + }, + { + "epoch": 0.9825770495261872, + "grad_norm": 0.47938814759254456, + "learning_rate": 1.4973428769099462e-07, + "loss": 0.0984, + "step": 41916 + }, + { + "epoch": 0.9826004911009921, + "grad_norm": 0.7390487790107727, + "learning_rate": 1.4933160593718632e-07, + "loss": 0.0842, + "step": 41917 + }, + { + "epoch": 0.9826239326757972, + "grad_norm": 0.3496438264846802, + "learning_rate": 1.4892946597589997e-07, + "loss": 0.0533, + "step": 41918 + }, + { + "epoch": 0.9826473742506021, + "grad_norm": 0.3836252689361572, + "learning_rate": 1.4852786780934492e-07, + "loss": 0.0595, + "step": 41919 + }, + { + "epoch": 0.9826708158254072, + "grad_norm": 0.39196112751960754, + "learning_rate": 1.48126811439675e-07, + "loss": 0.0328, + "step": 41920 + }, + { + "epoch": 0.9826942574002121, + "grad_norm": 0.19154863059520721, + "learning_rate": 1.4772629686907735e-07, + "loss": 0.0222, + "step": 41921 + }, + { + "epoch": 0.9827176989750172, + "grad_norm": 0.20846490561962128, + "learning_rate": 1.47326324099728e-07, + "loss": 0.0203, + "step": 41922 + }, + { + "epoch": 0.9827411405498221, + "grad_norm": 0.15789826214313507, + "learning_rate": 1.4692689313379194e-07, + "loss": 0.0194, + "step": 41923 + }, + { + "epoch": 0.9827645821246271, + "grad_norm": 0.5634744763374329, + "learning_rate": 1.4652800397343402e-07, + "loss": 0.0437, + "step": 41924 + }, + { + "epoch": 0.9827880236994321, + "grad_norm": 0.5659254193305969, + "learning_rate": 1.4612965662083033e-07, + "loss": 0.1073, + "step": 41925 + }, + { + "epoch": 0.9828114652742371, + "grad_norm": 0.19852954149246216, + "learning_rate": 1.4573185107812358e-07, + "loss": 0.0388, + "step": 41926 + }, + { + "epoch": 0.9828349068490421, + "grad_norm": 0.16303354501724243, + "learning_rate": 1.4533458734750094e-07, + "loss": 0.0167, + "step": 41927 + }, + { + "epoch": 0.9828583484238471, + "grad_norm": 0.6687421798706055, + "learning_rate": 1.449378654310829e-07, + "loss": 0.5736, + "step": 41928 + }, + { + "epoch": 0.9828817899986521, + "grad_norm": 0.09073147922754288, + "learning_rate": 1.445416853310566e-07, + "loss": 0.009, + "step": 41929 + }, + { + "epoch": 0.9829052315734571, + "grad_norm": 0.6455991268157959, + "learning_rate": 1.441460470495426e-07, + "loss": 0.0585, + "step": 41930 + }, + { + "epoch": 0.9829286731482622, + "grad_norm": 0.4037434160709381, + "learning_rate": 1.437509505887058e-07, + "loss": 0.0972, + "step": 41931 + }, + { + "epoch": 0.9829521147230671, + "grad_norm": 1.0272655487060547, + "learning_rate": 1.4335639595067784e-07, + "loss": 0.1572, + "step": 41932 + }, + { + "epoch": 0.9829755562978721, + "grad_norm": 0.14025349915027618, + "learning_rate": 1.4296238313761257e-07, + "loss": 0.0117, + "step": 41933 + }, + { + "epoch": 0.9829989978726771, + "grad_norm": 0.5343084931373596, + "learning_rate": 1.4256891215163047e-07, + "loss": 0.0978, + "step": 41934 + }, + { + "epoch": 0.9830224394474821, + "grad_norm": 0.54084312915802, + "learning_rate": 1.4217598299488544e-07, + "loss": 0.0387, + "step": 41935 + }, + { + "epoch": 0.9830458810222871, + "grad_norm": 0.1598769575357437, + "learning_rate": 1.4178359566949795e-07, + "loss": 0.023, + "step": 41936 + }, + { + "epoch": 0.9830693225970921, + "grad_norm": 0.6376850605010986, + "learning_rate": 1.4139175017761076e-07, + "loss": 0.0935, + "step": 41937 + }, + { + "epoch": 0.9830927641718971, + "grad_norm": 0.33483585715293884, + "learning_rate": 1.4100044652132216e-07, + "loss": 0.0369, + "step": 41938 + }, + { + "epoch": 0.9831162057467021, + "grad_norm": 0.3972729742527008, + "learning_rate": 1.4060968470278602e-07, + "loss": 0.5783, + "step": 41939 + }, + { + "epoch": 0.983139647321507, + "grad_norm": 0.7551599740982056, + "learning_rate": 1.4021946472411175e-07, + "loss": 0.158, + "step": 41940 + }, + { + "epoch": 0.9831630888963121, + "grad_norm": 0.565270721912384, + "learning_rate": 1.3982978658740874e-07, + "loss": 0.1241, + "step": 41941 + }, + { + "epoch": 0.983186530471117, + "grad_norm": 0.3689015507698059, + "learning_rate": 1.3944065029479757e-07, + "loss": 0.0504, + "step": 41942 + }, + { + "epoch": 0.9832099720459221, + "grad_norm": 0.22592787444591522, + "learning_rate": 1.3905205584838765e-07, + "loss": 0.0148, + "step": 41943 + }, + { + "epoch": 0.983233413620727, + "grad_norm": 0.4554089605808258, + "learning_rate": 1.3866400325028838e-07, + "loss": 0.0898, + "step": 41944 + }, + { + "epoch": 0.9832568551955321, + "grad_norm": 0.22069691121578217, + "learning_rate": 1.382764925026092e-07, + "loss": 0.0384, + "step": 41945 + }, + { + "epoch": 0.983280296770337, + "grad_norm": 0.6363845467567444, + "learning_rate": 1.3788952360744845e-07, + "loss": 0.1056, + "step": 41946 + }, + { + "epoch": 0.9833037383451421, + "grad_norm": 0.4843929708003998, + "learning_rate": 1.3750309656690442e-07, + "loss": 0.5839, + "step": 41947 + }, + { + "epoch": 0.983327179919947, + "grad_norm": 0.25675663352012634, + "learning_rate": 1.3711721138307544e-07, + "loss": 0.0386, + "step": 41948 + }, + { + "epoch": 0.983350621494752, + "grad_norm": 0.19510096311569214, + "learning_rate": 1.3673186805805983e-07, + "loss": 0.035, + "step": 41949 + }, + { + "epoch": 0.983374063069557, + "grad_norm": 0.3444376289844513, + "learning_rate": 1.363470665939559e-07, + "loss": 0.0818, + "step": 41950 + }, + { + "epoch": 0.983397504644362, + "grad_norm": 0.9362179636955261, + "learning_rate": 1.359628069928176e-07, + "loss": 0.1245, + "step": 41951 + }, + { + "epoch": 0.983420946219167, + "grad_norm": 0.34919384121894836, + "learning_rate": 1.3557908925677653e-07, + "loss": 0.0279, + "step": 41952 + }, + { + "epoch": 0.983444387793972, + "grad_norm": 0.18957439064979553, + "learning_rate": 1.3519591338787551e-07, + "loss": 0.0246, + "step": 41953 + }, + { + "epoch": 0.983467829368777, + "grad_norm": 0.4285665452480316, + "learning_rate": 1.3481327938821287e-07, + "loss": 0.0669, + "step": 41954 + }, + { + "epoch": 0.983491270943582, + "grad_norm": 0.4969714879989624, + "learning_rate": 1.344311872598647e-07, + "loss": 0.0605, + "step": 41955 + }, + { + "epoch": 0.983514712518387, + "grad_norm": 0.34540894627571106, + "learning_rate": 1.3404963700489603e-07, + "loss": 0.0475, + "step": 41956 + }, + { + "epoch": 0.983538154093192, + "grad_norm": 0.21778464317321777, + "learning_rate": 1.3366862862538298e-07, + "loss": 0.0321, + "step": 41957 + }, + { + "epoch": 0.983561595667997, + "grad_norm": 0.9590562582015991, + "learning_rate": 1.3328816212340166e-07, + "loss": 0.1904, + "step": 41958 + }, + { + "epoch": 0.983585037242802, + "grad_norm": 0.18502140045166016, + "learning_rate": 1.329082375009949e-07, + "loss": 0.0439, + "step": 41959 + }, + { + "epoch": 0.9836084788176069, + "grad_norm": 0.09260252118110657, + "learning_rate": 1.3252885476023879e-07, + "loss": 0.0239, + "step": 41960 + }, + { + "epoch": 0.983631920392412, + "grad_norm": 0.3128923177719116, + "learning_rate": 1.3215001390318726e-07, + "loss": 0.0461, + "step": 41961 + }, + { + "epoch": 0.9836553619672169, + "grad_norm": 0.634684145450592, + "learning_rate": 1.3177171493190532e-07, + "loss": 0.0908, + "step": 41962 + }, + { + "epoch": 0.983678803542022, + "grad_norm": 0.1839522272348404, + "learning_rate": 1.313939578484358e-07, + "loss": 0.0171, + "step": 41963 + }, + { + "epoch": 0.9837022451168269, + "grad_norm": 1.096036672592163, + "learning_rate": 1.3101674265482146e-07, + "loss": 0.065, + "step": 41964 + }, + { + "epoch": 0.983725686691632, + "grad_norm": 0.34686118364334106, + "learning_rate": 1.3064006935312734e-07, + "loss": 0.0595, + "step": 41965 + }, + { + "epoch": 0.9837491282664369, + "grad_norm": 0.2846546173095703, + "learning_rate": 1.3026393794538516e-07, + "loss": 0.0399, + "step": 41966 + }, + { + "epoch": 0.983772569841242, + "grad_norm": 0.4160541594028473, + "learning_rate": 1.2988834843364884e-07, + "loss": 0.0431, + "step": 41967 + }, + { + "epoch": 0.9837960114160469, + "grad_norm": 0.395751416683197, + "learning_rate": 1.2951330081993896e-07, + "loss": 0.0478, + "step": 41968 + }, + { + "epoch": 0.9838194529908519, + "grad_norm": 0.15125323832035065, + "learning_rate": 1.2913879510628723e-07, + "loss": 0.0259, + "step": 41969 + }, + { + "epoch": 0.9838428945656569, + "grad_norm": 0.5220348238945007, + "learning_rate": 1.2876483129474758e-07, + "loss": 0.0793, + "step": 41970 + }, + { + "epoch": 0.9838663361404619, + "grad_norm": 0.08183189481496811, + "learning_rate": 1.2839140938734063e-07, + "loss": 0.0108, + "step": 41971 + }, + { + "epoch": 0.9838897777152669, + "grad_norm": 0.538541853427887, + "learning_rate": 1.2801852938608695e-07, + "loss": 0.1103, + "step": 41972 + }, + { + "epoch": 0.9839132192900719, + "grad_norm": 0.7847616076469421, + "learning_rate": 1.2764619129300715e-07, + "loss": 0.1104, + "step": 41973 + }, + { + "epoch": 0.9839366608648769, + "grad_norm": 0.3992239534854889, + "learning_rate": 1.2727439511013295e-07, + "loss": 0.0976, + "step": 41974 + }, + { + "epoch": 0.9839601024396819, + "grad_norm": 0.4601772129535675, + "learning_rate": 1.2690314083947385e-07, + "loss": 0.0694, + "step": 41975 + }, + { + "epoch": 0.9839835440144868, + "grad_norm": 0.9172854423522949, + "learning_rate": 1.2653242848305048e-07, + "loss": 0.1233, + "step": 41976 + }, + { + "epoch": 0.9840069855892919, + "grad_norm": 0.45191675424575806, + "learning_rate": 1.261622580428612e-07, + "loss": 0.358, + "step": 41977 + }, + { + "epoch": 0.9840304271640968, + "grad_norm": 0.08925717324018478, + "learning_rate": 1.2579262952093773e-07, + "loss": 0.011, + "step": 41978 + }, + { + "epoch": 0.9840538687389019, + "grad_norm": 0.26910191774368286, + "learning_rate": 1.254235429192674e-07, + "loss": 0.028, + "step": 41979 + }, + { + "epoch": 0.9840773103137068, + "grad_norm": 0.5816749334335327, + "learning_rate": 1.2505499823984857e-07, + "loss": 0.5216, + "step": 41980 + }, + { + "epoch": 0.9841007518885119, + "grad_norm": 0.11085669696331024, + "learning_rate": 1.2468699548469077e-07, + "loss": 0.0189, + "step": 41981 + }, + { + "epoch": 0.9841241934633169, + "grad_norm": 0.36220139265060425, + "learning_rate": 1.2431953465579238e-07, + "loss": 0.0512, + "step": 41982 + }, + { + "epoch": 0.9841476350381219, + "grad_norm": 0.39745357632637024, + "learning_rate": 1.2395261575515182e-07, + "loss": 0.0563, + "step": 41983 + }, + { + "epoch": 0.9841710766129269, + "grad_norm": 0.3520825207233429, + "learning_rate": 1.235862387847453e-07, + "loss": 0.0491, + "step": 41984 + }, + { + "epoch": 0.9841945181877318, + "grad_norm": 0.3785061538219452, + "learning_rate": 1.2322040374657118e-07, + "loss": 0.0202, + "step": 41985 + }, + { + "epoch": 0.9842179597625369, + "grad_norm": 0.3033590018749237, + "learning_rate": 1.2285511064260567e-07, + "loss": 0.0517, + "step": 41986 + }, + { + "epoch": 0.9842414013373418, + "grad_norm": 0.42127954959869385, + "learning_rate": 1.224903594748472e-07, + "loss": 0.057, + "step": 41987 + }, + { + "epoch": 0.9842648429121469, + "grad_norm": 0.03428720682859421, + "learning_rate": 1.2212615024524977e-07, + "loss": 0.0083, + "step": 41988 + }, + { + "epoch": 0.9842882844869518, + "grad_norm": 0.14360782504081726, + "learning_rate": 1.2176248295581173e-07, + "loss": 0.0282, + "step": 41989 + }, + { + "epoch": 0.9843117260617569, + "grad_norm": 0.7164157032966614, + "learning_rate": 1.2139935760849818e-07, + "loss": 0.1377, + "step": 41990 + }, + { + "epoch": 0.9843351676365618, + "grad_norm": 0.5119879841804504, + "learning_rate": 1.2103677420528537e-07, + "loss": 0.103, + "step": 41991 + }, + { + "epoch": 0.9843586092113669, + "grad_norm": 0.384700745344162, + "learning_rate": 1.2067473274812723e-07, + "loss": 0.0808, + "step": 41992 + }, + { + "epoch": 0.9843820507861718, + "grad_norm": 0.06292804330587387, + "learning_rate": 1.2031323323901112e-07, + "loss": 0.0051, + "step": 41993 + }, + { + "epoch": 0.9844054923609769, + "grad_norm": 0.3253318667411804, + "learning_rate": 1.1995227567986878e-07, + "loss": 0.0544, + "step": 41994 + }, + { + "epoch": 0.9844289339357818, + "grad_norm": 0.4447341859340668, + "learning_rate": 1.1959186007267643e-07, + "loss": 0.0891, + "step": 41995 + }, + { + "epoch": 0.9844523755105868, + "grad_norm": 0.2808839976787567, + "learning_rate": 1.1923198641938803e-07, + "loss": 0.0608, + "step": 41996 + }, + { + "epoch": 0.9844758170853918, + "grad_norm": 0.35074254870414734, + "learning_rate": 1.1887265472195764e-07, + "loss": 0.0405, + "step": 41997 + }, + { + "epoch": 0.9844992586601968, + "grad_norm": 0.30053427815437317, + "learning_rate": 1.1851386498232809e-07, + "loss": 0.0347, + "step": 41998 + }, + { + "epoch": 0.9845227002350018, + "grad_norm": 0.5001954436302185, + "learning_rate": 1.1815561720245338e-07, + "loss": 0.0712, + "step": 41999 + }, + { + "epoch": 0.9845461418098068, + "grad_norm": 0.5503443479537964, + "learning_rate": 1.1779791138427643e-07, + "loss": 0.5989, + "step": 42000 + }, + { + "epoch": 0.9845695833846118, + "grad_norm": 0.4237864017486572, + "learning_rate": 1.17440747529729e-07, + "loss": 0.0685, + "step": 42001 + }, + { + "epoch": 0.9845930249594168, + "grad_norm": 0.3733888566493988, + "learning_rate": 1.170841256407651e-07, + "loss": 0.0714, + "step": 42002 + }, + { + "epoch": 0.9846164665342217, + "grad_norm": 0.6364831328392029, + "learning_rate": 1.167280457193054e-07, + "loss": 0.0859, + "step": 42003 + }, + { + "epoch": 0.9846399081090268, + "grad_norm": 0.1850546896457672, + "learning_rate": 1.163725077672928e-07, + "loss": 0.087, + "step": 42004 + }, + { + "epoch": 0.9846633496838317, + "grad_norm": 0.6933931708335876, + "learning_rate": 1.16017511786648e-07, + "loss": 0.6146, + "step": 42005 + }, + { + "epoch": 0.9846867912586368, + "grad_norm": 0.3579496145248413, + "learning_rate": 1.1566305777930275e-07, + "loss": 0.0342, + "step": 42006 + }, + { + "epoch": 0.9847102328334417, + "grad_norm": 0.45280909538269043, + "learning_rate": 1.1530914574717777e-07, + "loss": 0.0523, + "step": 42007 + }, + { + "epoch": 0.9847336744082468, + "grad_norm": 0.6225023865699768, + "learning_rate": 1.1495577569219373e-07, + "loss": 0.1177, + "step": 42008 + }, + { + "epoch": 0.9847571159830517, + "grad_norm": 0.05638313293457031, + "learning_rate": 1.1460294761627133e-07, + "loss": 0.0035, + "step": 42009 + }, + { + "epoch": 0.9847805575578568, + "grad_norm": 0.6007921695709229, + "learning_rate": 1.1425066152132014e-07, + "loss": 0.1246, + "step": 42010 + }, + { + "epoch": 0.9848039991326617, + "grad_norm": 0.12262248247861862, + "learning_rate": 1.1389891740924974e-07, + "loss": 0.0159, + "step": 42011 + }, + { + "epoch": 0.9848274407074668, + "grad_norm": 1.096692442893982, + "learning_rate": 1.1354771528196973e-07, + "loss": 0.1978, + "step": 42012 + }, + { + "epoch": 0.9848508822822717, + "grad_norm": 0.7899664044380188, + "learning_rate": 1.1319705514140077e-07, + "loss": 0.1202, + "step": 42013 + }, + { + "epoch": 0.9848743238570767, + "grad_norm": 0.6076713800430298, + "learning_rate": 1.1284693698941917e-07, + "loss": 0.607, + "step": 42014 + }, + { + "epoch": 0.9848977654318817, + "grad_norm": 0.5982524156570435, + "learning_rate": 1.124973608279456e-07, + "loss": 0.0938, + "step": 42015 + }, + { + "epoch": 0.9849212070066867, + "grad_norm": 0.08314923197031021, + "learning_rate": 1.1214832665886743e-07, + "loss": 0.0101, + "step": 42016 + }, + { + "epoch": 0.9849446485814917, + "grad_norm": 0.8309585452079773, + "learning_rate": 1.1179983448408316e-07, + "loss": 0.2227, + "step": 42017 + }, + { + "epoch": 0.9849680901562967, + "grad_norm": 0.44983765482902527, + "learning_rate": 1.1145188430549125e-07, + "loss": 0.1179, + "step": 42018 + }, + { + "epoch": 0.9849915317311017, + "grad_norm": 0.30229923129081726, + "learning_rate": 1.1110447612495689e-07, + "loss": 0.0394, + "step": 42019 + }, + { + "epoch": 0.9850149733059067, + "grad_norm": 0.9164735078811646, + "learning_rate": 1.1075760994437856e-07, + "loss": 0.1987, + "step": 42020 + }, + { + "epoch": 0.9850384148807116, + "grad_norm": 0.529484748840332, + "learning_rate": 1.1041128576564364e-07, + "loss": 0.6096, + "step": 42021 + }, + { + "epoch": 0.9850618564555167, + "grad_norm": 0.1239880621433258, + "learning_rate": 1.1006550359061729e-07, + "loss": 0.0126, + "step": 42022 + }, + { + "epoch": 0.9850852980303216, + "grad_norm": 0.42678335309028625, + "learning_rate": 1.0972026342118691e-07, + "loss": 0.0585, + "step": 42023 + }, + { + "epoch": 0.9851087396051267, + "grad_norm": 0.5782219171524048, + "learning_rate": 1.0937556525922877e-07, + "loss": 0.0979, + "step": 42024 + }, + { + "epoch": 0.9851321811799316, + "grad_norm": 0.3951333165168762, + "learning_rate": 1.0903140910660802e-07, + "loss": 0.0858, + "step": 42025 + }, + { + "epoch": 0.9851556227547367, + "grad_norm": 0.10901570320129395, + "learning_rate": 1.0868779496518988e-07, + "loss": 0.014, + "step": 42026 + }, + { + "epoch": 0.9851790643295416, + "grad_norm": 0.40668007731437683, + "learning_rate": 1.083447228368395e-07, + "loss": 0.061, + "step": 42027 + }, + { + "epoch": 0.9852025059043467, + "grad_norm": 0.4507385194301605, + "learning_rate": 1.0800219272342205e-07, + "loss": 0.0806, + "step": 42028 + }, + { + "epoch": 0.9852259474791516, + "grad_norm": 0.7031485438346863, + "learning_rate": 1.076602046267805e-07, + "loss": 0.1595, + "step": 42029 + }, + { + "epoch": 0.9852493890539566, + "grad_norm": 0.28267741203308105, + "learning_rate": 1.0731875854880225e-07, + "loss": 0.028, + "step": 42030 + }, + { + "epoch": 0.9852728306287616, + "grad_norm": 0.2196260392665863, + "learning_rate": 1.0697785449130803e-07, + "loss": 0.043, + "step": 42031 + }, + { + "epoch": 0.9852962722035666, + "grad_norm": 0.5862098932266235, + "learning_rate": 1.0663749245616306e-07, + "loss": 0.081, + "step": 42032 + }, + { + "epoch": 0.9853197137783717, + "grad_norm": 0.5576050877571106, + "learning_rate": 1.0629767244521028e-07, + "loss": 0.0997, + "step": 42033 + }, + { + "epoch": 0.9853431553531766, + "grad_norm": 0.3085091710090637, + "learning_rate": 1.0595839446030375e-07, + "loss": 0.0539, + "step": 42034 + }, + { + "epoch": 0.9853665969279817, + "grad_norm": 0.2879692316055298, + "learning_rate": 1.0561965850326427e-07, + "loss": 0.0475, + "step": 42035 + }, + { + "epoch": 0.9853900385027866, + "grad_norm": 0.39311835169792175, + "learning_rate": 1.05281464575957e-07, + "loss": 0.021, + "step": 42036 + }, + { + "epoch": 0.9854134800775917, + "grad_norm": 0.10885690152645111, + "learning_rate": 1.049438126801916e-07, + "loss": 0.011, + "step": 42037 + }, + { + "epoch": 0.9854369216523966, + "grad_norm": 0.27655717730522156, + "learning_rate": 1.0460670281781105e-07, + "loss": 0.0346, + "step": 42038 + }, + { + "epoch": 0.9854603632272017, + "grad_norm": 0.17593461275100708, + "learning_rate": 1.042701349906361e-07, + "loss": 0.0371, + "step": 42039 + }, + { + "epoch": 0.9854838048020066, + "grad_norm": 0.1842488944530487, + "learning_rate": 1.0393410920050973e-07, + "loss": 0.035, + "step": 42040 + }, + { + "epoch": 0.9855072463768116, + "grad_norm": 0.5517042875289917, + "learning_rate": 1.035986254492416e-07, + "loss": 0.1109, + "step": 42041 + }, + { + "epoch": 0.9855306879516166, + "grad_norm": 0.40663546323776245, + "learning_rate": 1.0326368373865247e-07, + "loss": 0.0836, + "step": 42042 + }, + { + "epoch": 0.9855541295264216, + "grad_norm": 0.2704940736293793, + "learning_rate": 1.0292928407057422e-07, + "loss": 0.0223, + "step": 42043 + }, + { + "epoch": 0.9855775711012266, + "grad_norm": 0.3781537115573883, + "learning_rate": 1.0259542644680542e-07, + "loss": 0.0449, + "step": 42044 + }, + { + "epoch": 0.9856010126760316, + "grad_norm": 0.5984703302383423, + "learning_rate": 1.0226211086915571e-07, + "loss": 0.1027, + "step": 42045 + }, + { + "epoch": 0.9856244542508366, + "grad_norm": 0.507300078868866, + "learning_rate": 1.0192933733944588e-07, + "loss": 0.0874, + "step": 42046 + }, + { + "epoch": 0.9856478958256416, + "grad_norm": 0.17014357447624207, + "learning_rate": 1.0159710585947446e-07, + "loss": 0.0221, + "step": 42047 + }, + { + "epoch": 0.9856713374004465, + "grad_norm": 0.12210649251937866, + "learning_rate": 1.0126541643105114e-07, + "loss": 0.0102, + "step": 42048 + }, + { + "epoch": 0.9856947789752516, + "grad_norm": 0.6584795713424683, + "learning_rate": 1.0093426905596337e-07, + "loss": 0.0857, + "step": 42049 + }, + { + "epoch": 0.9857182205500565, + "grad_norm": 0.6648910045623779, + "learning_rate": 1.0060366373602081e-07, + "loss": 0.0858, + "step": 42050 + }, + { + "epoch": 0.9857416621248616, + "grad_norm": 0.27956005930900574, + "learning_rate": 1.0027360047301093e-07, + "loss": 0.0357, + "step": 42051 + }, + { + "epoch": 0.9857651036996665, + "grad_norm": 0.15696382522583008, + "learning_rate": 9.994407926872118e-08, + "loss": 0.0335, + "step": 42052 + }, + { + "epoch": 0.9857885452744716, + "grad_norm": 0.41130709648132324, + "learning_rate": 9.961510012495012e-08, + "loss": 0.0646, + "step": 42053 + }, + { + "epoch": 0.9858119868492765, + "grad_norm": 0.40101414918899536, + "learning_rate": 9.928666304347411e-08, + "loss": 0.0796, + "step": 42054 + }, + { + "epoch": 0.9858354284240816, + "grad_norm": 0.541459321975708, + "learning_rate": 9.895876802608062e-08, + "loss": 0.1437, + "step": 42055 + }, + { + "epoch": 0.9858588699988865, + "grad_norm": 0.4302560091018677, + "learning_rate": 9.863141507454598e-08, + "loss": 0.0788, + "step": 42056 + }, + { + "epoch": 0.9858823115736915, + "grad_norm": 0.7416347861289978, + "learning_rate": 9.830460419064657e-08, + "loss": 0.0827, + "step": 42057 + }, + { + "epoch": 0.9859057531484965, + "grad_norm": 0.1722879707813263, + "learning_rate": 9.797833537614764e-08, + "loss": 0.0159, + "step": 42058 + }, + { + "epoch": 0.9859291947233015, + "grad_norm": 0.29921793937683105, + "learning_rate": 9.765260863282554e-08, + "loss": 0.0384, + "step": 42059 + }, + { + "epoch": 0.9859526362981065, + "grad_norm": 0.20972363650798798, + "learning_rate": 9.732742396245664e-08, + "loss": 0.0145, + "step": 42060 + }, + { + "epoch": 0.9859760778729115, + "grad_norm": 0.3839825391769409, + "learning_rate": 9.700278136679508e-08, + "loss": 0.0283, + "step": 42061 + }, + { + "epoch": 0.9859995194477165, + "grad_norm": 0.3943808674812317, + "learning_rate": 9.667868084760612e-08, + "loss": 0.0592, + "step": 42062 + }, + { + "epoch": 0.9860229610225215, + "grad_norm": 0.3085247576236725, + "learning_rate": 9.635512240665501e-08, + "loss": 0.0421, + "step": 42063 + }, + { + "epoch": 0.9860464025973265, + "grad_norm": 0.37584084272384644, + "learning_rate": 9.60321060456848e-08, + "loss": 0.0583, + "step": 42064 + }, + { + "epoch": 0.9860698441721315, + "grad_norm": 0.23956583440303802, + "learning_rate": 9.570963176644965e-08, + "loss": 0.0263, + "step": 42065 + }, + { + "epoch": 0.9860932857469364, + "grad_norm": 0.1866261065006256, + "learning_rate": 9.538769957070371e-08, + "loss": 0.0413, + "step": 42066 + }, + { + "epoch": 0.9861167273217415, + "grad_norm": 0.3738050162792206, + "learning_rate": 9.506630946020112e-08, + "loss": 0.0745, + "step": 42067 + }, + { + "epoch": 0.9861401688965464, + "grad_norm": 0.37527355551719666, + "learning_rate": 9.474546143667384e-08, + "loss": 0.0674, + "step": 42068 + }, + { + "epoch": 0.9861636104713515, + "grad_norm": 0.14489516615867615, + "learning_rate": 9.442515550186492e-08, + "loss": 0.0206, + "step": 42069 + }, + { + "epoch": 0.9861870520461564, + "grad_norm": 0.3827863037586212, + "learning_rate": 9.410539165750632e-08, + "loss": 0.0359, + "step": 42070 + }, + { + "epoch": 0.9862104936209615, + "grad_norm": 0.3803991675376892, + "learning_rate": 9.378616990535217e-08, + "loss": 0.0666, + "step": 42071 + }, + { + "epoch": 0.9862339351957664, + "grad_norm": 0.3307223320007324, + "learning_rate": 9.346749024712331e-08, + "loss": 0.0742, + "step": 42072 + }, + { + "epoch": 0.9862573767705715, + "grad_norm": 0.3358094394207001, + "learning_rate": 9.314935268454061e-08, + "loss": 0.0427, + "step": 42073 + }, + { + "epoch": 0.9862808183453764, + "grad_norm": 0.28304004669189453, + "learning_rate": 9.283175721934712e-08, + "loss": 0.0507, + "step": 42074 + }, + { + "epoch": 0.9863042599201814, + "grad_norm": 0.42791086435317993, + "learning_rate": 9.251470385325256e-08, + "loss": 0.0392, + "step": 42075 + }, + { + "epoch": 0.9863277014949864, + "grad_norm": 0.4647276997566223, + "learning_rate": 9.21981925879889e-08, + "loss": 0.5501, + "step": 42076 + }, + { + "epoch": 0.9863511430697914, + "grad_norm": 0.33851805329322815, + "learning_rate": 9.188222342525476e-08, + "loss": 0.0439, + "step": 42077 + }, + { + "epoch": 0.9863745846445964, + "grad_norm": 0.39674803614616394, + "learning_rate": 9.156679636678212e-08, + "loss": 0.0432, + "step": 42078 + }, + { + "epoch": 0.9863980262194014, + "grad_norm": 0.5519326329231262, + "learning_rate": 9.125191141428069e-08, + "loss": 0.0966, + "step": 42079 + }, + { + "epoch": 0.9864214677942064, + "grad_norm": 0.29552093148231506, + "learning_rate": 9.093756856946024e-08, + "loss": 0.0295, + "step": 42080 + }, + { + "epoch": 0.9864449093690114, + "grad_norm": 0.13264411687850952, + "learning_rate": 9.062376783400827e-08, + "loss": 0.0213, + "step": 42081 + }, + { + "epoch": 0.9864683509438164, + "grad_norm": 0.33762654662132263, + "learning_rate": 9.031050920965677e-08, + "loss": 0.0418, + "step": 42082 + }, + { + "epoch": 0.9864917925186214, + "grad_norm": 0.4047330319881439, + "learning_rate": 8.999779269808217e-08, + "loss": 0.0572, + "step": 42083 + }, + { + "epoch": 0.9865152340934265, + "grad_norm": 0.4284387528896332, + "learning_rate": 8.96856183009831e-08, + "loss": 0.6457, + "step": 42084 + }, + { + "epoch": 0.9865386756682314, + "grad_norm": 0.46562322974205017, + "learning_rate": 8.93739860200582e-08, + "loss": 0.0235, + "step": 42085 + }, + { + "epoch": 0.9865621172430364, + "grad_norm": 0.12628372013568878, + "learning_rate": 8.906289585700611e-08, + "loss": 0.0191, + "step": 42086 + }, + { + "epoch": 0.9865855588178414, + "grad_norm": 0.7161909937858582, + "learning_rate": 8.875234781351439e-08, + "loss": 0.05, + "step": 42087 + }, + { + "epoch": 0.9866090003926464, + "grad_norm": 0.42668211460113525, + "learning_rate": 8.844234189124834e-08, + "loss": 0.0688, + "step": 42088 + }, + { + "epoch": 0.9866324419674514, + "grad_norm": 0.4798277020454407, + "learning_rate": 8.813287809191772e-08, + "loss": 0.5845, + "step": 42089 + }, + { + "epoch": 0.9866558835422564, + "grad_norm": 0.30780333280563354, + "learning_rate": 8.782395641717678e-08, + "loss": 0.0316, + "step": 42090 + }, + { + "epoch": 0.9866793251170614, + "grad_norm": 0.6005271673202515, + "learning_rate": 8.751557686871303e-08, + "loss": 0.097, + "step": 42091 + }, + { + "epoch": 0.9867027666918664, + "grad_norm": 0.7450222969055176, + "learning_rate": 8.720773944820293e-08, + "loss": 0.0276, + "step": 42092 + }, + { + "epoch": 0.9867262082666713, + "grad_norm": 0.19981957972049713, + "learning_rate": 8.69004441573118e-08, + "loss": 0.0133, + "step": 42093 + }, + { + "epoch": 0.9867496498414764, + "grad_norm": 0.8647661805152893, + "learning_rate": 8.659369099770498e-08, + "loss": 0.0853, + "step": 42094 + }, + { + "epoch": 0.9867730914162813, + "grad_norm": 0.5923416018486023, + "learning_rate": 8.62874799710478e-08, + "loss": 0.1165, + "step": 42095 + }, + { + "epoch": 0.9867965329910864, + "grad_norm": 0.2419341653585434, + "learning_rate": 8.59818110790056e-08, + "loss": 0.0183, + "step": 42096 + }, + { + "epoch": 0.9868199745658913, + "grad_norm": 0.49332863092422485, + "learning_rate": 8.567668432324372e-08, + "loss": 0.1106, + "step": 42097 + }, + { + "epoch": 0.9868434161406964, + "grad_norm": 0.4812883138656616, + "learning_rate": 8.537209970540527e-08, + "loss": 0.0442, + "step": 42098 + }, + { + "epoch": 0.9868668577155013, + "grad_norm": 0.17711718380451202, + "learning_rate": 8.506805722714451e-08, + "loss": 0.0234, + "step": 42099 + }, + { + "epoch": 0.9868902992903064, + "grad_norm": 0.7672727108001709, + "learning_rate": 8.476455689011565e-08, + "loss": 0.0611, + "step": 42100 + }, + { + "epoch": 0.9869137408651113, + "grad_norm": 0.6638063788414001, + "learning_rate": 8.446159869596182e-08, + "loss": 0.0581, + "step": 42101 + }, + { + "epoch": 0.9869371824399163, + "grad_norm": 0.28094032406806946, + "learning_rate": 8.415918264632617e-08, + "loss": 0.059, + "step": 42102 + }, + { + "epoch": 0.9869606240147213, + "grad_norm": 0.25489744544029236, + "learning_rate": 8.38573087428518e-08, + "loss": 0.0205, + "step": 42103 + }, + { + "epoch": 0.9869840655895263, + "grad_norm": 0.37499332427978516, + "learning_rate": 8.355597698718187e-08, + "loss": 0.0768, + "step": 42104 + }, + { + "epoch": 0.9870075071643313, + "grad_norm": 0.5644780993461609, + "learning_rate": 8.32551873809484e-08, + "loss": 0.0803, + "step": 42105 + }, + { + "epoch": 0.9870309487391363, + "grad_norm": 0.5493227243423462, + "learning_rate": 8.295493992578341e-08, + "loss": 0.1071, + "step": 42106 + }, + { + "epoch": 0.9870543903139413, + "grad_norm": 0.3109005093574524, + "learning_rate": 8.265523462330783e-08, + "loss": 0.0325, + "step": 42107 + }, + { + "epoch": 0.9870778318887463, + "grad_norm": 0.1993645876646042, + "learning_rate": 8.235607147515367e-08, + "loss": 0.0487, + "step": 42108 + }, + { + "epoch": 0.9871012734635513, + "grad_norm": 0.37545084953308105, + "learning_rate": 8.2057450482953e-08, + "loss": 0.1219, + "step": 42109 + }, + { + "epoch": 0.9871247150383563, + "grad_norm": 0.17862673103809357, + "learning_rate": 8.17593716483156e-08, + "loss": 0.0369, + "step": 42110 + }, + { + "epoch": 0.9871481566131612, + "grad_norm": 0.11561991274356842, + "learning_rate": 8.146183497285131e-08, + "loss": 0.0256, + "step": 42111 + }, + { + "epoch": 0.9871715981879663, + "grad_norm": 0.4924525022506714, + "learning_rate": 8.116484045819218e-08, + "loss": 0.0561, + "step": 42112 + }, + { + "epoch": 0.9871950397627712, + "grad_norm": 0.2203286737203598, + "learning_rate": 8.086838810594799e-08, + "loss": 0.0312, + "step": 42113 + }, + { + "epoch": 0.9872184813375763, + "grad_norm": 0.4222933351993561, + "learning_rate": 8.057247791771749e-08, + "loss": 0.0318, + "step": 42114 + }, + { + "epoch": 0.9872419229123812, + "grad_norm": 0.1388387829065323, + "learning_rate": 8.027710989511051e-08, + "loss": 0.0303, + "step": 42115 + }, + { + "epoch": 0.9872653644871863, + "grad_norm": 0.4290582835674286, + "learning_rate": 7.998228403972574e-08, + "loss": 0.0704, + "step": 42116 + }, + { + "epoch": 0.9872888060619912, + "grad_norm": 0.28043332695961, + "learning_rate": 7.968800035316194e-08, + "loss": 0.0431, + "step": 42117 + }, + { + "epoch": 0.9873122476367963, + "grad_norm": 0.515012800693512, + "learning_rate": 7.939425883702889e-08, + "loss": 0.0626, + "step": 42118 + }, + { + "epoch": 0.9873356892116012, + "grad_norm": 0.47404569387435913, + "learning_rate": 7.910105949290314e-08, + "loss": 0.0782, + "step": 42119 + }, + { + "epoch": 0.9873591307864062, + "grad_norm": 0.5571755766868591, + "learning_rate": 7.88084023223945e-08, + "loss": 0.1082, + "step": 42120 + }, + { + "epoch": 0.9873825723612112, + "grad_norm": 0.3495112955570221, + "learning_rate": 7.851628732707949e-08, + "loss": 0.0454, + "step": 42121 + }, + { + "epoch": 0.9874060139360162, + "grad_norm": 0.13991472125053406, + "learning_rate": 7.822471450854574e-08, + "loss": 0.0218, + "step": 42122 + }, + { + "epoch": 0.9874294555108212, + "grad_norm": 0.559307336807251, + "learning_rate": 7.793368386836975e-08, + "loss": 1.1149, + "step": 42123 + }, + { + "epoch": 0.9874528970856262, + "grad_norm": 0.3367871642112732, + "learning_rate": 7.764319540813914e-08, + "loss": 0.0484, + "step": 42124 + }, + { + "epoch": 0.9874763386604312, + "grad_norm": 0.5232043266296387, + "learning_rate": 7.735324912941932e-08, + "loss": 0.0727, + "step": 42125 + }, + { + "epoch": 0.9874997802352362, + "grad_norm": 0.13495083153247833, + "learning_rate": 7.706384503379793e-08, + "loss": 0.0203, + "step": 42126 + }, + { + "epoch": 0.9875232218100412, + "grad_norm": 0.09606447070837021, + "learning_rate": 7.677498312282927e-08, + "loss": 0.0209, + "step": 42127 + }, + { + "epoch": 0.9875466633848462, + "grad_norm": 0.1554432362318039, + "learning_rate": 7.648666339810096e-08, + "loss": 0.0208, + "step": 42128 + }, + { + "epoch": 0.9875701049596511, + "grad_norm": 0.7337034940719604, + "learning_rate": 7.61988858611562e-08, + "loss": 0.1125, + "step": 42129 + }, + { + "epoch": 0.9875935465344562, + "grad_norm": 0.1831580400466919, + "learning_rate": 7.591165051357152e-08, + "loss": 0.0215, + "step": 42130 + }, + { + "epoch": 0.9876169881092611, + "grad_norm": 0.124233677983284, + "learning_rate": 7.562495735689013e-08, + "loss": 0.0301, + "step": 42131 + }, + { + "epoch": 0.9876404296840662, + "grad_norm": 0.10811980068683624, + "learning_rate": 7.533880639268853e-08, + "loss": 0.0171, + "step": 42132 + }, + { + "epoch": 0.9876638712588711, + "grad_norm": 0.5524460077285767, + "learning_rate": 7.505319762249885e-08, + "loss": 0.0348, + "step": 42133 + }, + { + "epoch": 0.9876873128336762, + "grad_norm": 0.26638418436050415, + "learning_rate": 7.476813104788649e-08, + "loss": 0.0311, + "step": 42134 + }, + { + "epoch": 0.9877107544084811, + "grad_norm": 0.639371931552887, + "learning_rate": 7.448360667039466e-08, + "loss": 0.1035, + "step": 42135 + }, + { + "epoch": 0.9877341959832862, + "grad_norm": 0.4077465534210205, + "learning_rate": 7.419962449155548e-08, + "loss": 0.1036, + "step": 42136 + }, + { + "epoch": 0.9877576375580912, + "grad_norm": 0.4124591052532196, + "learning_rate": 7.391618451292325e-08, + "loss": 0.0988, + "step": 42137 + }, + { + "epoch": 0.9877810791328961, + "grad_norm": 0.24183334410190582, + "learning_rate": 7.363328673604119e-08, + "loss": 0.0377, + "step": 42138 + }, + { + "epoch": 0.9878045207077012, + "grad_norm": 0.2510017454624176, + "learning_rate": 7.33509311624192e-08, + "loss": 0.0366, + "step": 42139 + }, + { + "epoch": 0.9878279622825061, + "grad_norm": 0.3921220004558563, + "learning_rate": 7.306911779361159e-08, + "loss": 0.0544, + "step": 42140 + }, + { + "epoch": 0.9878514038573112, + "grad_norm": 0.9475036263465881, + "learning_rate": 7.278784663113935e-08, + "loss": 0.1017, + "step": 42141 + }, + { + "epoch": 0.9878748454321161, + "grad_norm": 0.9971645474433899, + "learning_rate": 7.250711767653461e-08, + "loss": 0.1072, + "step": 42142 + }, + { + "epoch": 0.9878982870069212, + "grad_norm": 0.5482659935951233, + "learning_rate": 7.222693093130729e-08, + "loss": 0.1009, + "step": 42143 + }, + { + "epoch": 0.9879217285817261, + "grad_norm": 0.4681205153465271, + "learning_rate": 7.194728639698945e-08, + "loss": 0.083, + "step": 42144 + }, + { + "epoch": 0.9879451701565312, + "grad_norm": 0.30076533555984497, + "learning_rate": 7.166818407510212e-08, + "loss": 0.0519, + "step": 42145 + }, + { + "epoch": 0.9879686117313361, + "grad_norm": 0.21902981400489807, + "learning_rate": 7.13896239671441e-08, + "loss": 0.0339, + "step": 42146 + }, + { + "epoch": 0.9879920533061411, + "grad_norm": 0.042400166392326355, + "learning_rate": 7.111160607463641e-08, + "loss": 0.0023, + "step": 42147 + }, + { + "epoch": 0.9880154948809461, + "grad_norm": 0.16653099656105042, + "learning_rate": 7.083413039908892e-08, + "loss": 0.0342, + "step": 42148 + }, + { + "epoch": 0.9880389364557511, + "grad_norm": 0.34633368253707886, + "learning_rate": 7.055719694201157e-08, + "loss": 0.0502, + "step": 42149 + }, + { + "epoch": 0.9880623780305561, + "grad_norm": 0.4719533622264862, + "learning_rate": 7.028080570489204e-08, + "loss": 0.1087, + "step": 42150 + }, + { + "epoch": 0.9880858196053611, + "grad_norm": 0.721126139163971, + "learning_rate": 7.000495668924023e-08, + "loss": 0.2217, + "step": 42151 + }, + { + "epoch": 0.9881092611801661, + "grad_norm": 0.34543371200561523, + "learning_rate": 6.972964989655495e-08, + "loss": 0.0462, + "step": 42152 + }, + { + "epoch": 0.9881327027549711, + "grad_norm": 0.1714506298303604, + "learning_rate": 6.94548853283239e-08, + "loss": 0.0225, + "step": 42153 + }, + { + "epoch": 0.988156144329776, + "grad_norm": 0.09344326704740524, + "learning_rate": 6.918066298604586e-08, + "loss": 0.0133, + "step": 42154 + }, + { + "epoch": 0.9881795859045811, + "grad_norm": 0.2974047064781189, + "learning_rate": 6.890698287119746e-08, + "loss": 0.0384, + "step": 42155 + }, + { + "epoch": 0.988203027479386, + "grad_norm": 0.6089444756507874, + "learning_rate": 6.863384498526638e-08, + "loss": 0.1356, + "step": 42156 + }, + { + "epoch": 0.9882264690541911, + "grad_norm": 0.3843810558319092, + "learning_rate": 6.836124932974031e-08, + "loss": 0.0499, + "step": 42157 + }, + { + "epoch": 0.988249910628996, + "grad_norm": 0.6437132954597473, + "learning_rate": 6.808919590610696e-08, + "loss": 0.1656, + "step": 42158 + }, + { + "epoch": 0.9882733522038011, + "grad_norm": 0.3051312565803528, + "learning_rate": 6.781768471582073e-08, + "loss": 0.0668, + "step": 42159 + }, + { + "epoch": 0.988296793778606, + "grad_norm": 0.4766090512275696, + "learning_rate": 6.75467157603693e-08, + "loss": 0.07, + "step": 42160 + }, + { + "epoch": 0.9883202353534111, + "grad_norm": 0.05817878618836403, + "learning_rate": 6.727628904120708e-08, + "loss": 0.0082, + "step": 42161 + }, + { + "epoch": 0.988343676928216, + "grad_norm": 0.6510971784591675, + "learning_rate": 6.700640455982176e-08, + "loss": 0.1189, + "step": 42162 + }, + { + "epoch": 0.9883671185030211, + "grad_norm": 0.7650601863861084, + "learning_rate": 6.673706231767884e-08, + "loss": 0.06, + "step": 42163 + }, + { + "epoch": 0.988390560077826, + "grad_norm": 0.5148385167121887, + "learning_rate": 6.646826231622161e-08, + "loss": 0.5534, + "step": 42164 + }, + { + "epoch": 0.988414001652631, + "grad_norm": 0.36612874269485474, + "learning_rate": 6.620000455691555e-08, + "loss": 0.3668, + "step": 42165 + }, + { + "epoch": 0.988437443227436, + "grad_norm": 0.47683024406433105, + "learning_rate": 6.593228904122617e-08, + "loss": 0.0569, + "step": 42166 + }, + { + "epoch": 0.988460884802241, + "grad_norm": 0.1959168165922165, + "learning_rate": 6.566511577058565e-08, + "loss": 0.0177, + "step": 42167 + }, + { + "epoch": 0.988484326377046, + "grad_norm": 0.3905276358127594, + "learning_rate": 6.539848474647058e-08, + "loss": 0.0757, + "step": 42168 + }, + { + "epoch": 0.988507767951851, + "grad_norm": 0.4674215018749237, + "learning_rate": 6.513239597030207e-08, + "loss": 0.0869, + "step": 42169 + }, + { + "epoch": 0.988531209526656, + "grad_norm": 0.46117842197418213, + "learning_rate": 6.48668494435456e-08, + "loss": 0.4823, + "step": 42170 + }, + { + "epoch": 0.988554651101461, + "grad_norm": 0.5718197822570801, + "learning_rate": 6.460184516762224e-08, + "loss": 0.1208, + "step": 42171 + }, + { + "epoch": 0.988578092676266, + "grad_norm": 0.5487556457519531, + "learning_rate": 6.43373831439864e-08, + "loss": 0.0719, + "step": 42172 + }, + { + "epoch": 0.988601534251071, + "grad_norm": 0.3067089021205902, + "learning_rate": 6.407346337405918e-08, + "loss": 0.0709, + "step": 42173 + }, + { + "epoch": 0.9886249758258759, + "grad_norm": 0.2626161575317383, + "learning_rate": 6.381008585928383e-08, + "loss": 0.0437, + "step": 42174 + }, + { + "epoch": 0.988648417400681, + "grad_norm": 0.1355496197938919, + "learning_rate": 6.354725060108146e-08, + "loss": 0.0143, + "step": 42175 + }, + { + "epoch": 0.9886718589754859, + "grad_norm": 0.6464430093765259, + "learning_rate": 6.328495760088426e-08, + "loss": 0.1269, + "step": 42176 + }, + { + "epoch": 0.988695300550291, + "grad_norm": 0.11030406504869461, + "learning_rate": 6.302320686011332e-08, + "loss": 0.0116, + "step": 42177 + }, + { + "epoch": 0.9887187421250959, + "grad_norm": 0.269894003868103, + "learning_rate": 6.27619983801786e-08, + "loss": 0.0519, + "step": 42178 + }, + { + "epoch": 0.988742183699901, + "grad_norm": 0.11369600147008896, + "learning_rate": 6.250133216252341e-08, + "loss": 0.016, + "step": 42179 + }, + { + "epoch": 0.9887656252747059, + "grad_norm": 0.27231481671333313, + "learning_rate": 6.224120820853552e-08, + "loss": 0.0635, + "step": 42180 + }, + { + "epoch": 0.988789066849511, + "grad_norm": 0.23858408629894257, + "learning_rate": 6.198162651963601e-08, + "loss": 0.0186, + "step": 42181 + }, + { + "epoch": 0.9888125084243159, + "grad_norm": 0.09542933106422424, + "learning_rate": 6.172258709723488e-08, + "loss": 0.0197, + "step": 42182 + }, + { + "epoch": 0.9888359499991209, + "grad_norm": 0.37423720955848694, + "learning_rate": 6.14640899427421e-08, + "loss": 0.0592, + "step": 42183 + }, + { + "epoch": 0.9888593915739259, + "grad_norm": 0.4706951379776001, + "learning_rate": 6.120613505755656e-08, + "loss": 0.0747, + "step": 42184 + }, + { + "epoch": 0.9888828331487309, + "grad_norm": 0.11936637759208679, + "learning_rate": 6.094872244307715e-08, + "loss": 0.0127, + "step": 42185 + }, + { + "epoch": 0.9889062747235359, + "grad_norm": 0.20630556344985962, + "learning_rate": 6.069185210069161e-08, + "loss": 0.0265, + "step": 42186 + }, + { + "epoch": 0.9889297162983409, + "grad_norm": 0.3844689428806305, + "learning_rate": 6.043552403179887e-08, + "loss": 0.0892, + "step": 42187 + }, + { + "epoch": 0.988953157873146, + "grad_norm": 0.13394135236740112, + "learning_rate": 6.017973823780887e-08, + "loss": 0.0226, + "step": 42188 + }, + { + "epoch": 0.9889765994479509, + "grad_norm": 0.43472883105278015, + "learning_rate": 5.992449472008721e-08, + "loss": 0.0852, + "step": 42189 + }, + { + "epoch": 0.989000041022756, + "grad_norm": 0.10445386171340942, + "learning_rate": 5.966979348002167e-08, + "loss": 0.014, + "step": 42190 + }, + { + "epoch": 0.9890234825975609, + "grad_norm": 0.8830959796905518, + "learning_rate": 5.9415634519000006e-08, + "loss": 0.1275, + "step": 42191 + }, + { + "epoch": 0.989046924172366, + "grad_norm": 0.4217556416988373, + "learning_rate": 5.916201783839892e-08, + "loss": 0.048, + "step": 42192 + }, + { + "epoch": 0.9890703657471709, + "grad_norm": 0.7146165370941162, + "learning_rate": 5.8908943439583974e-08, + "loss": 0.0465, + "step": 42193 + }, + { + "epoch": 0.9890938073219759, + "grad_norm": 0.7778220772743225, + "learning_rate": 5.865641132395405e-08, + "loss": 0.0984, + "step": 42194 + }, + { + "epoch": 0.9891172488967809, + "grad_norm": 0.458696186542511, + "learning_rate": 5.840442149286363e-08, + "loss": 0.0821, + "step": 42195 + }, + { + "epoch": 0.9891406904715859, + "grad_norm": 0.2955787479877472, + "learning_rate": 5.815297394767827e-08, + "loss": 0.0195, + "step": 42196 + }, + { + "epoch": 0.9891641320463909, + "grad_norm": 0.295759379863739, + "learning_rate": 5.790206868976356e-08, + "loss": 0.0235, + "step": 42197 + }, + { + "epoch": 0.9891875736211959, + "grad_norm": 0.13500291109085083, + "learning_rate": 5.765170572048506e-08, + "loss": 0.0127, + "step": 42198 + }, + { + "epoch": 0.9892110151960009, + "grad_norm": 0.0937768891453743, + "learning_rate": 5.7401885041197254e-08, + "loss": 0.0176, + "step": 42199 + }, + { + "epoch": 0.9892344567708059, + "grad_norm": 0.44422447681427, + "learning_rate": 5.7152606653254615e-08, + "loss": 0.0432, + "step": 42200 + }, + { + "epoch": 0.9892578983456108, + "grad_norm": 0.4816979467868805, + "learning_rate": 5.690387055801161e-08, + "loss": 0.0768, + "step": 42201 + }, + { + "epoch": 0.9892813399204159, + "grad_norm": 0.16228336095809937, + "learning_rate": 5.6655676756822705e-08, + "loss": 0.0279, + "step": 42202 + }, + { + "epoch": 0.9893047814952208, + "grad_norm": 0.37541353702545166, + "learning_rate": 5.640802525103128e-08, + "loss": 0.0756, + "step": 42203 + }, + { + "epoch": 0.9893282230700259, + "grad_norm": 0.32011932134628296, + "learning_rate": 5.6160916041969605e-08, + "loss": 0.2977, + "step": 42204 + }, + { + "epoch": 0.9893516646448308, + "grad_norm": 0.40219613909721375, + "learning_rate": 5.591434913099214e-08, + "loss": 0.087, + "step": 42205 + }, + { + "epoch": 0.9893751062196359, + "grad_norm": 0.3842833638191223, + "learning_rate": 5.5668324519442263e-08, + "loss": 0.4745, + "step": 42206 + }, + { + "epoch": 0.9893985477944408, + "grad_norm": 0.3219163715839386, + "learning_rate": 5.542284220864114e-08, + "loss": 0.0621, + "step": 42207 + }, + { + "epoch": 0.9894219893692459, + "grad_norm": 0.5157294273376465, + "learning_rate": 5.517790219993213e-08, + "loss": 0.108, + "step": 42208 + }, + { + "epoch": 0.9894454309440508, + "grad_norm": 0.6434780955314636, + "learning_rate": 5.493350449462531e-08, + "loss": 0.1557, + "step": 42209 + }, + { + "epoch": 0.9894688725188558, + "grad_norm": 0.09923660010099411, + "learning_rate": 5.4689649094075145e-08, + "loss": 0.0101, + "step": 42210 + }, + { + "epoch": 0.9894923140936608, + "grad_norm": 0.48750388622283936, + "learning_rate": 5.44463359995695e-08, + "loss": 0.1142, + "step": 42211 + }, + { + "epoch": 0.9895157556684658, + "grad_norm": 0.5194850564002991, + "learning_rate": 5.420356521246284e-08, + "loss": 0.0936, + "step": 42212 + }, + { + "epoch": 0.9895391972432708, + "grad_norm": 0.12982887029647827, + "learning_rate": 5.396133673405412e-08, + "loss": 0.0196, + "step": 42213 + }, + { + "epoch": 0.9895626388180758, + "grad_norm": 0.41373977065086365, + "learning_rate": 5.3719650565664525e-08, + "loss": 0.0429, + "step": 42214 + }, + { + "epoch": 0.9895860803928808, + "grad_norm": 0.18530716001987457, + "learning_rate": 5.3478506708604106e-08, + "loss": 0.0161, + "step": 42215 + }, + { + "epoch": 0.9896095219676858, + "grad_norm": 0.35796085000038147, + "learning_rate": 5.323790516417182e-08, + "loss": 0.0379, + "step": 42216 + }, + { + "epoch": 0.9896329635424908, + "grad_norm": 0.4799957573413849, + "learning_rate": 5.299784593367774e-08, + "loss": 0.1028, + "step": 42217 + }, + { + "epoch": 0.9896564051172958, + "grad_norm": 0.38964536786079407, + "learning_rate": 5.275832901843192e-08, + "loss": 0.0642, + "step": 42218 + }, + { + "epoch": 0.9896798466921007, + "grad_norm": 0.5873771905899048, + "learning_rate": 5.251935441973332e-08, + "loss": 0.0989, + "step": 42219 + }, + { + "epoch": 0.9897032882669058, + "grad_norm": 0.5085461735725403, + "learning_rate": 5.228092213888092e-08, + "loss": 0.0422, + "step": 42220 + }, + { + "epoch": 0.9897267298417107, + "grad_norm": 0.22813411056995392, + "learning_rate": 5.204303217715145e-08, + "loss": 0.0456, + "step": 42221 + }, + { + "epoch": 0.9897501714165158, + "grad_norm": 0.04712836444377899, + "learning_rate": 5.180568453585499e-08, + "loss": 0.0051, + "step": 42222 + }, + { + "epoch": 0.9897736129913207, + "grad_norm": 0.39796268939971924, + "learning_rate": 5.1568879216257194e-08, + "loss": 0.102, + "step": 42223 + }, + { + "epoch": 0.9897970545661258, + "grad_norm": 0.4627770185470581, + "learning_rate": 5.133261621966812e-08, + "loss": 0.0788, + "step": 42224 + }, + { + "epoch": 0.9898204961409307, + "grad_norm": 0.1845645010471344, + "learning_rate": 5.109689554735342e-08, + "loss": 0.0364, + "step": 42225 + }, + { + "epoch": 0.9898439377157358, + "grad_norm": 0.5094046592712402, + "learning_rate": 5.0861717200600955e-08, + "loss": 0.0634, + "step": 42226 + }, + { + "epoch": 0.9898673792905407, + "grad_norm": 0.2415044754743576, + "learning_rate": 5.0627081180687483e-08, + "loss": 0.0483, + "step": 42227 + }, + { + "epoch": 0.9898908208653457, + "grad_norm": 0.43710067868232727, + "learning_rate": 5.0392987488878664e-08, + "loss": 0.0644, + "step": 42228 + }, + { + "epoch": 0.9899142624401507, + "grad_norm": 0.19495002925395966, + "learning_rate": 5.015943612644014e-08, + "loss": 0.0137, + "step": 42229 + }, + { + "epoch": 0.9899377040149557, + "grad_norm": 0.4431056082248688, + "learning_rate": 4.992642709465978e-08, + "loss": 0.0581, + "step": 42230 + }, + { + "epoch": 0.9899611455897607, + "grad_norm": 0.41058772802352905, + "learning_rate": 4.969396039478102e-08, + "loss": 0.4241, + "step": 42231 + }, + { + "epoch": 0.9899845871645657, + "grad_norm": 0.43697208166122437, + "learning_rate": 4.9462036028069535e-08, + "loss": 0.0391, + "step": 42232 + }, + { + "epoch": 0.9900080287393707, + "grad_norm": 0.3539876639842987, + "learning_rate": 4.923065399579096e-08, + "loss": 0.4342, + "step": 42233 + }, + { + "epoch": 0.9900314703141757, + "grad_norm": 0.5349600315093994, + "learning_rate": 4.8999814299199866e-08, + "loss": 0.1377, + "step": 42234 + }, + { + "epoch": 0.9900549118889806, + "grad_norm": 0.3837912082672119, + "learning_rate": 4.8769516939550786e-08, + "loss": 0.0374, + "step": 42235 + }, + { + "epoch": 0.9900783534637857, + "grad_norm": 0.5026434659957886, + "learning_rate": 4.853976191807608e-08, + "loss": 0.5594, + "step": 42236 + }, + { + "epoch": 0.9901017950385906, + "grad_norm": 0.24890893697738647, + "learning_rate": 4.8310549236041394e-08, + "loss": 0.0242, + "step": 42237 + }, + { + "epoch": 0.9901252366133957, + "grad_norm": 0.5370760560035706, + "learning_rate": 4.808187889469018e-08, + "loss": 0.1358, + "step": 42238 + }, + { + "epoch": 0.9901486781882007, + "grad_norm": 0.11909305304288864, + "learning_rate": 4.7853750895243686e-08, + "loss": 0.0222, + "step": 42239 + }, + { + "epoch": 0.9901721197630057, + "grad_norm": 0.06152347847819328, + "learning_rate": 4.762616523896757e-08, + "loss": 0.0072, + "step": 42240 + }, + { + "epoch": 0.9901955613378107, + "grad_norm": 0.571974515914917, + "learning_rate": 4.7399121927071964e-08, + "loss": 0.0942, + "step": 42241 + }, + { + "epoch": 0.9902190029126157, + "grad_norm": 0.14422082901000977, + "learning_rate": 4.717262096080033e-08, + "loss": 0.0263, + "step": 42242 + }, + { + "epoch": 0.9902424444874207, + "grad_norm": 0.19359953701496124, + "learning_rate": 4.6946662341385004e-08, + "loss": 0.0257, + "step": 42243 + }, + { + "epoch": 0.9902658860622257, + "grad_norm": 0.24476473033428192, + "learning_rate": 4.672124607004724e-08, + "loss": 0.0404, + "step": 42244 + }, + { + "epoch": 0.9902893276370307, + "grad_norm": 0.10636338591575623, + "learning_rate": 4.649637214800828e-08, + "loss": 0.0199, + "step": 42245 + }, + { + "epoch": 0.9903127692118356, + "grad_norm": 0.17284509539604187, + "learning_rate": 4.627204057648937e-08, + "loss": 0.0172, + "step": 42246 + }, + { + "epoch": 0.9903362107866407, + "grad_norm": 0.2780320942401886, + "learning_rate": 4.604825135671176e-08, + "loss": 0.0436, + "step": 42247 + }, + { + "epoch": 0.9903596523614456, + "grad_norm": 0.2831045091152191, + "learning_rate": 4.5825004489885584e-08, + "loss": 0.0328, + "step": 42248 + }, + { + "epoch": 0.9903830939362507, + "grad_norm": 0.35057342052459717, + "learning_rate": 4.5602299977220984e-08, + "loss": 0.0485, + "step": 42249 + }, + { + "epoch": 0.9904065355110556, + "grad_norm": 0.2382776290178299, + "learning_rate": 4.538013781992811e-08, + "loss": 0.0287, + "step": 42250 + }, + { + "epoch": 0.9904299770858607, + "grad_norm": 0.19489912688732147, + "learning_rate": 4.515851801921711e-08, + "loss": 0.0298, + "step": 42251 + }, + { + "epoch": 0.9904534186606656, + "grad_norm": 0.127293199300766, + "learning_rate": 4.4937440576287014e-08, + "loss": 0.0102, + "step": 42252 + }, + { + "epoch": 0.9904768602354707, + "grad_norm": 0.22451719641685486, + "learning_rate": 4.4716905492325766e-08, + "loss": 0.0409, + "step": 42253 + }, + { + "epoch": 0.9905003018102756, + "grad_norm": 0.43032893538475037, + "learning_rate": 4.449691276855461e-08, + "loss": 0.0587, + "step": 42254 + }, + { + "epoch": 0.9905237433850806, + "grad_norm": 0.33948731422424316, + "learning_rate": 4.427746240615038e-08, + "loss": 0.0604, + "step": 42255 + }, + { + "epoch": 0.9905471849598856, + "grad_norm": 0.3249223828315735, + "learning_rate": 4.405855440630102e-08, + "loss": 0.0453, + "step": 42256 + }, + { + "epoch": 0.9905706265346906, + "grad_norm": 0.23152369260787964, + "learning_rate": 4.384018877020557e-08, + "loss": 0.0164, + "step": 42257 + }, + { + "epoch": 0.9905940681094956, + "grad_norm": 0.38937053084373474, + "learning_rate": 4.3622365499051966e-08, + "loss": 0.0728, + "step": 42258 + }, + { + "epoch": 0.9906175096843006, + "grad_norm": 0.13900837302207947, + "learning_rate": 4.3405084593994835e-08, + "loss": 0.0253, + "step": 42259 + }, + { + "epoch": 0.9906409512591056, + "grad_norm": 0.4356619715690613, + "learning_rate": 4.318834605624433e-08, + "loss": 0.0642, + "step": 42260 + }, + { + "epoch": 0.9906643928339106, + "grad_norm": 0.6241639256477356, + "learning_rate": 4.2972149886966184e-08, + "loss": 0.0768, + "step": 42261 + }, + { + "epoch": 0.9906878344087156, + "grad_norm": 0.2903369069099426, + "learning_rate": 4.275649608732613e-08, + "loss": 0.0475, + "step": 42262 + }, + { + "epoch": 0.9907112759835206, + "grad_norm": 0.398651659488678, + "learning_rate": 4.2541384658500996e-08, + "loss": 0.0505, + "step": 42263 + }, + { + "epoch": 0.9907347175583255, + "grad_norm": 0.35694652795791626, + "learning_rate": 4.2326815601656525e-08, + "loss": 0.0663, + "step": 42264 + }, + { + "epoch": 0.9907581591331306, + "grad_norm": 0.5137602686882019, + "learning_rate": 4.2112788917958444e-08, + "loss": 0.522, + "step": 42265 + }, + { + "epoch": 0.9907816007079355, + "grad_norm": 0.5822026133537292, + "learning_rate": 4.18993046085725e-08, + "loss": 0.4922, + "step": 42266 + }, + { + "epoch": 0.9908050422827406, + "grad_norm": 0.2645319104194641, + "learning_rate": 4.168636267464221e-08, + "loss": 0.0335, + "step": 42267 + }, + { + "epoch": 0.9908284838575455, + "grad_norm": 0.3686918020248413, + "learning_rate": 4.1473963117344414e-08, + "loss": 0.0665, + "step": 42268 + }, + { + "epoch": 0.9908519254323506, + "grad_norm": 0.24845203757286072, + "learning_rate": 4.1262105937811546e-08, + "loss": 0.0325, + "step": 42269 + }, + { + "epoch": 0.9908753670071555, + "grad_norm": 0.6631859540939331, + "learning_rate": 4.105079113719823e-08, + "loss": 0.1223, + "step": 42270 + }, + { + "epoch": 0.9908988085819606, + "grad_norm": 0.5290471315383911, + "learning_rate": 4.08400187166591e-08, + "loss": 0.0927, + "step": 42271 + }, + { + "epoch": 0.9909222501567655, + "grad_norm": 0.6180135011672974, + "learning_rate": 4.0629788677326587e-08, + "loss": 0.5736, + "step": 42272 + }, + { + "epoch": 0.9909456917315705, + "grad_norm": 0.5306348204612732, + "learning_rate": 4.0420101020355316e-08, + "loss": 0.0369, + "step": 42273 + }, + { + "epoch": 0.9909691333063755, + "grad_norm": 0.30318683385849, + "learning_rate": 4.021095574686662e-08, + "loss": 0.0466, + "step": 42274 + }, + { + "epoch": 0.9909925748811805, + "grad_norm": 0.46342983841896057, + "learning_rate": 4.000235285800402e-08, + "loss": 0.0789, + "step": 42275 + }, + { + "epoch": 0.9910160164559855, + "grad_norm": 0.11790353059768677, + "learning_rate": 3.979429235491105e-08, + "loss": 0.0235, + "step": 42276 + }, + { + "epoch": 0.9910394580307905, + "grad_norm": 0.7403796911239624, + "learning_rate": 3.9586774238697944e-08, + "loss": 0.0987, + "step": 42277 + }, + { + "epoch": 0.9910628996055955, + "grad_norm": 0.4102618396282196, + "learning_rate": 3.937979851049711e-08, + "loss": 0.0568, + "step": 42278 + }, + { + "epoch": 0.9910863411804005, + "grad_norm": 1.1264599561691284, + "learning_rate": 3.917336517142989e-08, + "loss": 0.0785, + "step": 42279 + }, + { + "epoch": 0.9911097827552054, + "grad_norm": 0.1041637659072876, + "learning_rate": 3.8967474222617596e-08, + "loss": 0.0068, + "step": 42280 + }, + { + "epoch": 0.9911332243300105, + "grad_norm": 0.4890068769454956, + "learning_rate": 3.876212566518156e-08, + "loss": 0.0606, + "step": 42281 + }, + { + "epoch": 0.9911566659048154, + "grad_norm": 0.257579505443573, + "learning_rate": 3.8557319500232e-08, + "loss": 0.0308, + "step": 42282 + }, + { + "epoch": 0.9911801074796205, + "grad_norm": 0.5156827569007874, + "learning_rate": 3.8353055728890255e-08, + "loss": 0.0315, + "step": 42283 + }, + { + "epoch": 0.9912035490544254, + "grad_norm": 0.36452195048332214, + "learning_rate": 3.814933435224432e-08, + "loss": 0.0631, + "step": 42284 + }, + { + "epoch": 0.9912269906292305, + "grad_norm": 0.4437994956970215, + "learning_rate": 3.794615537141555e-08, + "loss": 0.062, + "step": 42285 + }, + { + "epoch": 0.9912504322040354, + "grad_norm": 0.33262234926223755, + "learning_rate": 3.774351878749194e-08, + "loss": 0.0639, + "step": 42286 + }, + { + "epoch": 0.9912738737788405, + "grad_norm": 0.09557893872261047, + "learning_rate": 3.754142460159482e-08, + "loss": 0.01, + "step": 42287 + }, + { + "epoch": 0.9912973153536454, + "grad_norm": 0.4381329417228699, + "learning_rate": 3.7339872814801115e-08, + "loss": 0.0738, + "step": 42288 + }, + { + "epoch": 0.9913207569284505, + "grad_norm": 0.298166960477829, + "learning_rate": 3.713886342822104e-08, + "loss": 0.4156, + "step": 42289 + }, + { + "epoch": 0.9913441985032555, + "grad_norm": 0.3543127477169037, + "learning_rate": 3.693839644293151e-08, + "loss": 0.0389, + "step": 42290 + }, + { + "epoch": 0.9913676400780604, + "grad_norm": 0.3349531888961792, + "learning_rate": 3.673847186002055e-08, + "loss": 0.0725, + "step": 42291 + }, + { + "epoch": 0.9913910816528655, + "grad_norm": 0.41519007086753845, + "learning_rate": 3.6539089680576176e-08, + "loss": 0.0894, + "step": 42292 + }, + { + "epoch": 0.9914145232276704, + "grad_norm": 0.13277989625930786, + "learning_rate": 3.634024990568641e-08, + "loss": 0.0187, + "step": 42293 + }, + { + "epoch": 0.9914379648024755, + "grad_norm": 0.3926146626472473, + "learning_rate": 3.614195253642816e-08, + "loss": 0.0606, + "step": 42294 + }, + { + "epoch": 0.9914614063772804, + "grad_norm": 0.3778384029865265, + "learning_rate": 3.5944197573867244e-08, + "loss": 0.0678, + "step": 42295 + }, + { + "epoch": 0.9914848479520855, + "grad_norm": 0.4210412800312042, + "learning_rate": 3.574698501909168e-08, + "loss": 0.0566, + "step": 42296 + }, + { + "epoch": 0.9915082895268904, + "grad_norm": 0.600345253944397, + "learning_rate": 3.555031487315619e-08, + "loss": 0.3823, + "step": 42297 + }, + { + "epoch": 0.9915317311016955, + "grad_norm": 0.602356493473053, + "learning_rate": 3.535418713714877e-08, + "loss": 0.0838, + "step": 42298 + }, + { + "epoch": 0.9915551726765004, + "grad_norm": 0.6142903566360474, + "learning_rate": 3.5158601812101956e-08, + "loss": 0.1112, + "step": 42299 + }, + { + "epoch": 0.9915786142513054, + "grad_norm": 0.1918637752532959, + "learning_rate": 3.496355889910374e-08, + "loss": 0.0265, + "step": 42300 + }, + { + "epoch": 0.9916020558261104, + "grad_norm": 0.21403658390045166, + "learning_rate": 3.4769058399208855e-08, + "loss": 0.0334, + "step": 42301 + }, + { + "epoch": 0.9916254974009154, + "grad_norm": 0.1332116276025772, + "learning_rate": 3.4575100313460896e-08, + "loss": 0.0156, + "step": 42302 + }, + { + "epoch": 0.9916489389757204, + "grad_norm": 0.36637580394744873, + "learning_rate": 3.438168464292568e-08, + "loss": 0.0493, + "step": 42303 + }, + { + "epoch": 0.9916723805505254, + "grad_norm": 0.11446575075387955, + "learning_rate": 3.418881138863572e-08, + "loss": 0.0108, + "step": 42304 + }, + { + "epoch": 0.9916958221253304, + "grad_norm": 0.27595511078834534, + "learning_rate": 3.399648055165683e-08, + "loss": 0.0572, + "step": 42305 + }, + { + "epoch": 0.9917192637001354, + "grad_norm": 0.23065884411334991, + "learning_rate": 3.380469213302151e-08, + "loss": 0.0331, + "step": 42306 + }, + { + "epoch": 0.9917427052749404, + "grad_norm": 0.14663758873939514, + "learning_rate": 3.361344613377338e-08, + "loss": 0.0271, + "step": 42307 + }, + { + "epoch": 0.9917661468497454, + "grad_norm": 0.14427882432937622, + "learning_rate": 3.342274255494493e-08, + "loss": 0.0207, + "step": 42308 + }, + { + "epoch": 0.9917895884245503, + "grad_norm": 0.1528874933719635, + "learning_rate": 3.323258139757979e-08, + "loss": 0.0195, + "step": 42309 + }, + { + "epoch": 0.9918130299993554, + "grad_norm": 0.5633341670036316, + "learning_rate": 3.304296266271045e-08, + "loss": 0.1142, + "step": 42310 + }, + { + "epoch": 0.9918364715741603, + "grad_norm": 0.4805759787559509, + "learning_rate": 3.285388635134723e-08, + "loss": 0.0553, + "step": 42311 + }, + { + "epoch": 0.9918599131489654, + "grad_norm": 0.31641215085983276, + "learning_rate": 3.266535246454483e-08, + "loss": 0.0225, + "step": 42312 + }, + { + "epoch": 0.9918833547237703, + "grad_norm": 0.36643245816230774, + "learning_rate": 3.2477361003302455e-08, + "loss": 0.0861, + "step": 42313 + }, + { + "epoch": 0.9919067962985754, + "grad_norm": 0.12812046706676483, + "learning_rate": 3.228991196864151e-08, + "loss": 0.0185, + "step": 42314 + }, + { + "epoch": 0.9919302378733803, + "grad_norm": 0.16430626809597015, + "learning_rate": 3.2103005361605596e-08, + "loss": 0.0214, + "step": 42315 + }, + { + "epoch": 0.9919536794481854, + "grad_norm": 0.3299798369407654, + "learning_rate": 3.1916641183171724e-08, + "loss": 0.0403, + "step": 42316 + }, + { + "epoch": 0.9919771210229903, + "grad_norm": 0.5662962794303894, + "learning_rate": 3.1730819434383496e-08, + "loss": 0.4037, + "step": 42317 + }, + { + "epoch": 0.9920005625977953, + "grad_norm": 0.12338119000196457, + "learning_rate": 3.1545540116217906e-08, + "loss": 0.0117, + "step": 42318 + }, + { + "epoch": 0.9920240041726003, + "grad_norm": 0.41906705498695374, + "learning_rate": 3.136080322970747e-08, + "loss": 0.0779, + "step": 42319 + }, + { + "epoch": 0.9920474457474053, + "grad_norm": 0.6471841335296631, + "learning_rate": 3.117660877585138e-08, + "loss": 0.6398, + "step": 42320 + }, + { + "epoch": 0.9920708873222103, + "grad_norm": 0.460880845785141, + "learning_rate": 3.0992956755626635e-08, + "loss": 0.0728, + "step": 42321 + }, + { + "epoch": 0.9920943288970153, + "grad_norm": 0.5935813784599304, + "learning_rate": 3.080984717005464e-08, + "loss": 0.0629, + "step": 42322 + }, + { + "epoch": 0.9921177704718203, + "grad_norm": 0.36605405807495117, + "learning_rate": 3.0627280020123495e-08, + "loss": 0.0387, + "step": 42323 + }, + { + "epoch": 0.9921412120466253, + "grad_norm": 0.8102303147315979, + "learning_rate": 3.0445255306810194e-08, + "loss": 0.0858, + "step": 42324 + }, + { + "epoch": 0.9921646536214302, + "grad_norm": 0.053202901035547256, + "learning_rate": 3.026377303111394e-08, + "loss": 0.0059, + "step": 42325 + }, + { + "epoch": 0.9921880951962353, + "grad_norm": 0.08517326414585114, + "learning_rate": 3.008283319401173e-08, + "loss": 0.0098, + "step": 42326 + }, + { + "epoch": 0.9922115367710402, + "grad_norm": 0.49918821454048157, + "learning_rate": 2.990243579650276e-08, + "loss": 0.1076, + "step": 42327 + }, + { + "epoch": 0.9922349783458453, + "grad_norm": 0.40184083580970764, + "learning_rate": 2.9722580839552928e-08, + "loss": 0.0843, + "step": 42328 + }, + { + "epoch": 0.9922584199206502, + "grad_norm": 0.5655950307846069, + "learning_rate": 2.9543268324128127e-08, + "loss": 0.0557, + "step": 42329 + }, + { + "epoch": 0.9922818614954553, + "grad_norm": 0.2769758999347687, + "learning_rate": 2.9364498251216453e-08, + "loss": 0.0479, + "step": 42330 + }, + { + "epoch": 0.9923053030702602, + "grad_norm": 0.19743680953979492, + "learning_rate": 2.918627062179491e-08, + "loss": 0.034, + "step": 42331 + }, + { + "epoch": 0.9923287446450653, + "grad_norm": 0.33427852392196655, + "learning_rate": 2.900858543680718e-08, + "loss": 0.0378, + "step": 42332 + }, + { + "epoch": 0.9923521862198702, + "grad_norm": 0.17414793372154236, + "learning_rate": 2.883144269723026e-08, + "loss": 0.0099, + "step": 42333 + }, + { + "epoch": 0.9923756277946753, + "grad_norm": 0.3174278736114502, + "learning_rate": 2.8654842404030047e-08, + "loss": 0.0615, + "step": 42334 + }, + { + "epoch": 0.9923990693694802, + "grad_norm": 0.5632694959640503, + "learning_rate": 2.847878455816133e-08, + "loss": 0.6134, + "step": 42335 + }, + { + "epoch": 0.9924225109442852, + "grad_norm": 0.0869276225566864, + "learning_rate": 2.8303269160578903e-08, + "loss": 0.0126, + "step": 42336 + }, + { + "epoch": 0.9924459525190902, + "grad_norm": 0.4864979386329651, + "learning_rate": 2.8128296212226458e-08, + "loss": 0.0317, + "step": 42337 + }, + { + "epoch": 0.9924693940938952, + "grad_norm": 0.3667182922363281, + "learning_rate": 2.795386571405878e-08, + "loss": 0.3771, + "step": 42338 + }, + { + "epoch": 0.9924928356687002, + "grad_norm": 0.27570590376853943, + "learning_rate": 2.7779977667030664e-08, + "loss": 0.0362, + "step": 42339 + }, + { + "epoch": 0.9925162772435052, + "grad_norm": 0.7056038975715637, + "learning_rate": 2.76066320720858e-08, + "loss": 0.1232, + "step": 42340 + }, + { + "epoch": 0.9925397188183103, + "grad_norm": 0.4030088484287262, + "learning_rate": 2.7433828930145677e-08, + "loss": 0.0762, + "step": 42341 + }, + { + "epoch": 0.9925631603931152, + "grad_norm": 0.41505807638168335, + "learning_rate": 2.726156824216508e-08, + "loss": 0.0623, + "step": 42342 + }, + { + "epoch": 0.9925866019679203, + "grad_norm": 0.3591727316379547, + "learning_rate": 2.70898500090766e-08, + "loss": 0.0684, + "step": 42343 + }, + { + "epoch": 0.9926100435427252, + "grad_norm": 0.5035836696624756, + "learning_rate": 2.6918674231812823e-08, + "loss": 0.0555, + "step": 42344 + }, + { + "epoch": 0.9926334851175302, + "grad_norm": 0.0781799778342247, + "learning_rate": 2.6748040911295234e-08, + "loss": 0.0108, + "step": 42345 + }, + { + "epoch": 0.9926569266923352, + "grad_norm": 0.14365530014038086, + "learning_rate": 2.657795004844532e-08, + "loss": 0.0214, + "step": 42346 + }, + { + "epoch": 0.9926803682671402, + "grad_norm": 0.6909621357917786, + "learning_rate": 2.640840164420677e-08, + "loss": 0.138, + "step": 42347 + }, + { + "epoch": 0.9927038098419452, + "grad_norm": 0.4320240616798401, + "learning_rate": 2.6239395699478863e-08, + "loss": 0.0561, + "step": 42348 + }, + { + "epoch": 0.9927272514167502, + "grad_norm": 0.2799643576145172, + "learning_rate": 2.6070932215194187e-08, + "loss": 0.051, + "step": 42349 + }, + { + "epoch": 0.9927506929915552, + "grad_norm": 0.12413470447063446, + "learning_rate": 2.5903011192252025e-08, + "loss": 0.0231, + "step": 42350 + }, + { + "epoch": 0.9927741345663602, + "grad_norm": 0.4215034246444702, + "learning_rate": 2.573563263157386e-08, + "loss": 0.0741, + "step": 42351 + }, + { + "epoch": 0.9927975761411652, + "grad_norm": 0.4350281357765198, + "learning_rate": 2.5568796534070072e-08, + "loss": 0.0967, + "step": 42352 + }, + { + "epoch": 0.9928210177159702, + "grad_norm": 0.20972268283367157, + "learning_rate": 2.5402502900639945e-08, + "loss": 0.0379, + "step": 42353 + }, + { + "epoch": 0.9928444592907751, + "grad_norm": 0.523361325263977, + "learning_rate": 2.523675173218276e-08, + "loss": 0.6542, + "step": 42354 + }, + { + "epoch": 0.9928679008655802, + "grad_norm": 0.24079297482967377, + "learning_rate": 2.50715430295978e-08, + "loss": 0.037, + "step": 42355 + }, + { + "epoch": 0.9928913424403851, + "grad_norm": 0.5130137801170349, + "learning_rate": 2.490687679379544e-08, + "loss": 0.0815, + "step": 42356 + }, + { + "epoch": 0.9929147840151902, + "grad_norm": 0.09639463573694229, + "learning_rate": 2.4742753025652766e-08, + "loss": 0.0105, + "step": 42357 + }, + { + "epoch": 0.9929382255899951, + "grad_norm": 0.3467722237110138, + "learning_rate": 2.4579171726069052e-08, + "loss": 0.0715, + "step": 42358 + }, + { + "epoch": 0.9929616671648002, + "grad_norm": 0.12942729890346527, + "learning_rate": 2.4416132895932474e-08, + "loss": 0.0141, + "step": 42359 + }, + { + "epoch": 0.9929851087396051, + "grad_norm": 0.14825430512428284, + "learning_rate": 2.4253636536131218e-08, + "loss": 0.0237, + "step": 42360 + }, + { + "epoch": 0.9930085503144102, + "grad_norm": 0.5931531190872192, + "learning_rate": 2.4091682647531254e-08, + "loss": 0.0577, + "step": 42361 + }, + { + "epoch": 0.9930319918892151, + "grad_norm": 0.25984635949134827, + "learning_rate": 2.3930271231020762e-08, + "loss": 0.0607, + "step": 42362 + }, + { + "epoch": 0.9930554334640201, + "grad_norm": 0.40285593271255493, + "learning_rate": 2.3769402287476817e-08, + "loss": 0.0424, + "step": 42363 + }, + { + "epoch": 0.9930788750388251, + "grad_norm": 0.5545793175697327, + "learning_rate": 2.3609075817765392e-08, + "loss": 0.0785, + "step": 42364 + }, + { + "epoch": 0.9931023166136301, + "grad_norm": 0.26710763573646545, + "learning_rate": 2.3449291822774665e-08, + "loss": 0.0307, + "step": 42365 + }, + { + "epoch": 0.9931257581884351, + "grad_norm": 0.4452107846736908, + "learning_rate": 2.329005030334841e-08, + "loss": 0.0904, + "step": 42366 + }, + { + "epoch": 0.9931491997632401, + "grad_norm": 0.13160394132137299, + "learning_rate": 2.3131351260363697e-08, + "loss": 0.0142, + "step": 42367 + }, + { + "epoch": 0.9931726413380451, + "grad_norm": 0.25707298517227173, + "learning_rate": 2.2973194694686507e-08, + "loss": 0.0422, + "step": 42368 + }, + { + "epoch": 0.9931960829128501, + "grad_norm": 0.41077154874801636, + "learning_rate": 2.2815580607160604e-08, + "loss": 0.0807, + "step": 42369 + }, + { + "epoch": 0.993219524487655, + "grad_norm": 0.46690982580184937, + "learning_rate": 2.265850899865196e-08, + "loss": 0.4297, + "step": 42370 + }, + { + "epoch": 0.9932429660624601, + "grad_norm": 0.09710434079170227, + "learning_rate": 2.2501979870004354e-08, + "loss": 0.009, + "step": 42371 + }, + { + "epoch": 0.993266407637265, + "grad_norm": 0.08899674564599991, + "learning_rate": 2.234599322208375e-08, + "loss": 0.0087, + "step": 42372 + }, + { + "epoch": 0.9932898492120701, + "grad_norm": 0.22998641431331635, + "learning_rate": 2.2190549055711718e-08, + "loss": 0.0465, + "step": 42373 + }, + { + "epoch": 0.993313290786875, + "grad_norm": 0.1652119755744934, + "learning_rate": 2.203564737175423e-08, + "loss": 0.0231, + "step": 42374 + }, + { + "epoch": 0.9933367323616801, + "grad_norm": 0.2644537091255188, + "learning_rate": 2.1881288171032855e-08, + "loss": 0.0347, + "step": 42375 + }, + { + "epoch": 0.993360173936485, + "grad_norm": 0.2654842436313629, + "learning_rate": 2.172747145440246e-08, + "loss": 0.0462, + "step": 42376 + }, + { + "epoch": 0.9933836155112901, + "grad_norm": 0.2342842072248459, + "learning_rate": 2.1574197222695714e-08, + "loss": 0.0466, + "step": 42377 + }, + { + "epoch": 0.993407057086095, + "grad_norm": 0.4192575514316559, + "learning_rate": 2.142146547672308e-08, + "loss": 0.0971, + "step": 42378 + }, + { + "epoch": 0.9934304986609, + "grad_norm": 0.2908177077770233, + "learning_rate": 2.1269276217339428e-08, + "loss": 0.0276, + "step": 42379 + }, + { + "epoch": 0.993453940235705, + "grad_norm": 0.2747500240802765, + "learning_rate": 2.1117629445355225e-08, + "loss": 0.0577, + "step": 42380 + }, + { + "epoch": 0.99347738181051, + "grad_norm": 0.42088982462882996, + "learning_rate": 2.0966525161603136e-08, + "loss": 0.053, + "step": 42381 + }, + { + "epoch": 0.993500823385315, + "grad_norm": 0.3936236798763275, + "learning_rate": 2.0815963366893622e-08, + "loss": 0.05, + "step": 42382 + }, + { + "epoch": 0.99352426496012, + "grad_norm": 0.39554038643836975, + "learning_rate": 2.0665944062048247e-08, + "loss": 0.5982, + "step": 42383 + }, + { + "epoch": 0.993547706534925, + "grad_norm": 0.06678128242492676, + "learning_rate": 2.051646724787748e-08, + "loss": 0.0107, + "step": 42384 + }, + { + "epoch": 0.99357114810973, + "grad_norm": 0.15544751286506653, + "learning_rate": 2.0367532925202882e-08, + "loss": 0.0196, + "step": 42385 + }, + { + "epoch": 0.993594589684535, + "grad_norm": 0.64194655418396, + "learning_rate": 2.0219141094823812e-08, + "loss": 0.0921, + "step": 42386 + }, + { + "epoch": 0.99361803125934, + "grad_norm": 0.1199822798371315, + "learning_rate": 2.0071291757539633e-08, + "loss": 0.0081, + "step": 42387 + }, + { + "epoch": 0.993641472834145, + "grad_norm": 0.6829016804695129, + "learning_rate": 1.992398491416081e-08, + "loss": 0.1098, + "step": 42388 + }, + { + "epoch": 0.99366491440895, + "grad_norm": 0.32802847027778625, + "learning_rate": 1.97772205654978e-08, + "loss": 0.0508, + "step": 42389 + }, + { + "epoch": 0.9936883559837549, + "grad_norm": 0.5186978578567505, + "learning_rate": 1.9630998712316663e-08, + "loss": 0.2369, + "step": 42390 + }, + { + "epoch": 0.99371179755856, + "grad_norm": 0.4631960988044739, + "learning_rate": 1.9485319355438958e-08, + "loss": 0.0956, + "step": 42391 + }, + { + "epoch": 0.993735239133365, + "grad_norm": 0.37735581398010254, + "learning_rate": 1.9340182495652948e-08, + "loss": 0.0439, + "step": 42392 + }, + { + "epoch": 0.99375868070817, + "grad_norm": 0.4531356692314148, + "learning_rate": 1.9195588133724684e-08, + "loss": 0.0936, + "step": 42393 + }, + { + "epoch": 0.993782122282975, + "grad_norm": 0.415035218000412, + "learning_rate": 1.9051536270464633e-08, + "loss": 0.0528, + "step": 42394 + }, + { + "epoch": 0.99380556385778, + "grad_norm": 0.37600430846214294, + "learning_rate": 1.890802690662774e-08, + "loss": 0.088, + "step": 42395 + }, + { + "epoch": 0.993829005432585, + "grad_norm": 0.4864996075630188, + "learning_rate": 1.876506004301337e-08, + "loss": 0.051, + "step": 42396 + }, + { + "epoch": 0.99385244700739, + "grad_norm": 0.362334668636322, + "learning_rate": 1.862263568039868e-08, + "loss": 0.0547, + "step": 42397 + }, + { + "epoch": 0.993875888582195, + "grad_norm": 0.2974774241447449, + "learning_rate": 1.848075381953862e-08, + "loss": 0.0249, + "step": 42398 + }, + { + "epoch": 0.9938993301569999, + "grad_norm": 0.29578638076782227, + "learning_rate": 1.833941446122145e-08, + "loss": 0.0557, + "step": 42399 + }, + { + "epoch": 0.993922771731805, + "grad_norm": 0.2350553423166275, + "learning_rate": 1.8198617606202118e-08, + "loss": 0.0476, + "step": 42400 + }, + { + "epoch": 0.9939462133066099, + "grad_norm": 0.4558153450489044, + "learning_rate": 1.805836325524668e-08, + "loss": 0.0637, + "step": 42401 + }, + { + "epoch": 0.993969654881415, + "grad_norm": 0.3904416263103485, + "learning_rate": 1.791865140912119e-08, + "loss": 0.0292, + "step": 42402 + }, + { + "epoch": 0.9939930964562199, + "grad_norm": 0.5912901759147644, + "learning_rate": 1.77794820685695e-08, + "loss": 0.0758, + "step": 42403 + }, + { + "epoch": 0.994016538031025, + "grad_norm": 0.437928706407547, + "learning_rate": 1.7640855234368758e-08, + "loss": 0.4878, + "step": 42404 + }, + { + "epoch": 0.9940399796058299, + "grad_norm": 0.8743762969970703, + "learning_rate": 1.750277090726282e-08, + "loss": 0.2428, + "step": 42405 + }, + { + "epoch": 0.994063421180635, + "grad_norm": 0.40529558062553406, + "learning_rate": 1.7365229087984436e-08, + "loss": 0.0573, + "step": 42406 + }, + { + "epoch": 0.9940868627554399, + "grad_norm": 0.15086540579795837, + "learning_rate": 1.7228229777310754e-08, + "loss": 0.0159, + "step": 42407 + }, + { + "epoch": 0.9941103043302449, + "grad_norm": 0.533710241317749, + "learning_rate": 1.7091772975952323e-08, + "loss": 0.1259, + "step": 42408 + }, + { + "epoch": 0.9941337459050499, + "grad_norm": 0.16321291029453278, + "learning_rate": 1.6955858684675198e-08, + "loss": 0.0157, + "step": 42409 + }, + { + "epoch": 0.9941571874798549, + "grad_norm": 0.6487709879875183, + "learning_rate": 1.6820486904201015e-08, + "loss": 0.5784, + "step": 42410 + }, + { + "epoch": 0.9941806290546599, + "grad_norm": 0.8818499445915222, + "learning_rate": 1.6685657635273633e-08, + "loss": 0.1109, + "step": 42411 + }, + { + "epoch": 0.9942040706294649, + "grad_norm": 0.5958880186080933, + "learning_rate": 1.6551370878614692e-08, + "loss": 0.1459, + "step": 42412 + }, + { + "epoch": 0.9942275122042699, + "grad_norm": 0.37964093685150146, + "learning_rate": 1.6417626634968043e-08, + "loss": 0.0465, + "step": 42413 + }, + { + "epoch": 0.9942509537790749, + "grad_norm": 0.5037074685096741, + "learning_rate": 1.6284424905044227e-08, + "loss": 0.0655, + "step": 42414 + }, + { + "epoch": 0.9942743953538798, + "grad_norm": 0.18860465288162231, + "learning_rate": 1.6151765689575993e-08, + "loss": 0.0268, + "step": 42415 + }, + { + "epoch": 0.9942978369286849, + "grad_norm": 0.46160244941711426, + "learning_rate": 1.6019648989284985e-08, + "loss": 0.0301, + "step": 42416 + }, + { + "epoch": 0.9943212785034898, + "grad_norm": 0.24404610693454742, + "learning_rate": 1.5888074804870645e-08, + "loss": 0.0252, + "step": 42417 + }, + { + "epoch": 0.9943447200782949, + "grad_norm": 0.24605219066143036, + "learning_rate": 1.575704313706572e-08, + "loss": 0.0425, + "step": 42418 + }, + { + "epoch": 0.9943681616530998, + "grad_norm": 0.5663122534751892, + "learning_rate": 1.562655398658075e-08, + "loss": 0.1062, + "step": 42419 + }, + { + "epoch": 0.9943916032279049, + "grad_norm": 0.5208838582038879, + "learning_rate": 1.5496607354104076e-08, + "loss": 0.0376, + "step": 42420 + }, + { + "epoch": 0.9944150448027098, + "grad_norm": 0.36843210458755493, + "learning_rate": 1.5367203240368445e-08, + "loss": 0.0465, + "step": 42421 + }, + { + "epoch": 0.9944384863775149, + "grad_norm": 0.7454407811164856, + "learning_rate": 1.523834164605109e-08, + "loss": 0.1648, + "step": 42422 + }, + { + "epoch": 0.9944619279523198, + "grad_norm": 0.45666369795799255, + "learning_rate": 1.511002257186256e-08, + "loss": 0.0487, + "step": 42423 + }, + { + "epoch": 0.9944853695271249, + "grad_norm": 0.2556038200855255, + "learning_rate": 1.4982246018502287e-08, + "loss": 0.0505, + "step": 42424 + }, + { + "epoch": 0.9945088111019298, + "grad_norm": 0.1859961450099945, + "learning_rate": 1.4855011986658618e-08, + "loss": 0.0145, + "step": 42425 + }, + { + "epoch": 0.9945322526767348, + "grad_norm": 0.04510728642344475, + "learning_rate": 1.4728320477019886e-08, + "loss": 0.0032, + "step": 42426 + }, + { + "epoch": 0.9945556942515398, + "grad_norm": 0.32505255937576294, + "learning_rate": 1.460217149027443e-08, + "loss": 0.0441, + "step": 42427 + }, + { + "epoch": 0.9945791358263448, + "grad_norm": 0.20446136593818665, + "learning_rate": 1.447656502711059e-08, + "loss": 0.0444, + "step": 42428 + }, + { + "epoch": 0.9946025774011498, + "grad_norm": 0.7585521340370178, + "learning_rate": 1.4351501088216701e-08, + "loss": 0.1671, + "step": 42429 + }, + { + "epoch": 0.9946260189759548, + "grad_norm": 0.4832674264907837, + "learning_rate": 1.42269796742589e-08, + "loss": 0.0674, + "step": 42430 + }, + { + "epoch": 0.9946494605507598, + "grad_norm": 0.08145549148321152, + "learning_rate": 1.4103000785914422e-08, + "loss": 0.0107, + "step": 42431 + }, + { + "epoch": 0.9946729021255648, + "grad_norm": 0.3551490604877472, + "learning_rate": 1.3979564423860503e-08, + "loss": 0.0867, + "step": 42432 + }, + { + "epoch": 0.9946963437003697, + "grad_norm": 0.5468000769615173, + "learning_rate": 1.3856670588763276e-08, + "loss": 0.7166, + "step": 42433 + }, + { + "epoch": 0.9947197852751748, + "grad_norm": 0.5573776960372925, + "learning_rate": 1.3734319281288877e-08, + "loss": 0.6441, + "step": 42434 + }, + { + "epoch": 0.9947432268499797, + "grad_norm": 0.40229612588882446, + "learning_rate": 1.3612510502103438e-08, + "loss": 0.0294, + "step": 42435 + }, + { + "epoch": 0.9947666684247848, + "grad_norm": 0.34529685974121094, + "learning_rate": 1.3491244251873092e-08, + "loss": 0.0459, + "step": 42436 + }, + { + "epoch": 0.9947901099995897, + "grad_norm": 0.42864206433296204, + "learning_rate": 1.337052053124177e-08, + "loss": 0.0872, + "step": 42437 + }, + { + "epoch": 0.9948135515743948, + "grad_norm": 0.08717826753854752, + "learning_rate": 1.3250339340886709e-08, + "loss": 0.0119, + "step": 42438 + }, + { + "epoch": 0.9948369931491997, + "grad_norm": 0.10250870138406754, + "learning_rate": 1.3130700681440733e-08, + "loss": 0.0162, + "step": 42439 + }, + { + "epoch": 0.9948604347240048, + "grad_norm": 0.29322096705436707, + "learning_rate": 1.3011604553547774e-08, + "loss": 0.0354, + "step": 42440 + }, + { + "epoch": 0.9948838762988097, + "grad_norm": 0.10131536424160004, + "learning_rate": 1.2893050957885067e-08, + "loss": 0.0132, + "step": 42441 + }, + { + "epoch": 0.9949073178736147, + "grad_norm": 0.49406829476356506, + "learning_rate": 1.2775039895063234e-08, + "loss": 0.0689, + "step": 42442 + }, + { + "epoch": 0.9949307594484198, + "grad_norm": 0.4286225736141205, + "learning_rate": 1.2657571365737308e-08, + "loss": 0.0608, + "step": 42443 + }, + { + "epoch": 0.9949542010232247, + "grad_norm": 0.4598623812198639, + "learning_rate": 1.2540645370540115e-08, + "loss": 0.0716, + "step": 42444 + }, + { + "epoch": 0.9949776425980298, + "grad_norm": 0.13159765303134918, + "learning_rate": 1.2424261910115587e-08, + "loss": 0.0114, + "step": 42445 + }, + { + "epoch": 0.9950010841728347, + "grad_norm": 0.1324038803577423, + "learning_rate": 1.2308420985085444e-08, + "loss": 0.0387, + "step": 42446 + }, + { + "epoch": 0.9950245257476398, + "grad_norm": 0.0657496377825737, + "learning_rate": 1.2193122596082519e-08, + "loss": 0.0073, + "step": 42447 + }, + { + "epoch": 0.9950479673224447, + "grad_norm": 0.5291213989257812, + "learning_rate": 1.2078366743728531e-08, + "loss": 0.0749, + "step": 42448 + }, + { + "epoch": 0.9950714088972498, + "grad_norm": 0.2669876217842102, + "learning_rate": 1.1964153428645208e-08, + "loss": 0.0442, + "step": 42449 + }, + { + "epoch": 0.9950948504720547, + "grad_norm": 0.7952284812927246, + "learning_rate": 1.1850482651465378e-08, + "loss": 0.1253, + "step": 42450 + }, + { + "epoch": 0.9951182920468598, + "grad_norm": 0.2853546142578125, + "learning_rate": 1.173735441278856e-08, + "loss": 0.0498, + "step": 42451 + }, + { + "epoch": 0.9951417336216647, + "grad_norm": 0.4616400897502899, + "learning_rate": 1.1624768713236479e-08, + "loss": 0.0599, + "step": 42452 + }, + { + "epoch": 0.9951651751964697, + "grad_norm": 0.1364843100309372, + "learning_rate": 1.1512725553419756e-08, + "loss": 0.0086, + "step": 42453 + }, + { + "epoch": 0.9951886167712747, + "grad_norm": 0.44887664914131165, + "learning_rate": 1.1401224933949017e-08, + "loss": 0.0685, + "step": 42454 + }, + { + "epoch": 0.9952120583460797, + "grad_norm": 0.15996257960796356, + "learning_rate": 1.1290266855434883e-08, + "loss": 0.029, + "step": 42455 + }, + { + "epoch": 0.9952354999208847, + "grad_norm": 0.21408894658088684, + "learning_rate": 1.1179851318465772e-08, + "loss": 0.0342, + "step": 42456 + }, + { + "epoch": 0.9952589414956897, + "grad_norm": 0.5648854374885559, + "learning_rate": 1.1069978323641206e-08, + "loss": 0.0633, + "step": 42457 + }, + { + "epoch": 0.9952823830704947, + "grad_norm": 0.2453932762145996, + "learning_rate": 1.0960647871560703e-08, + "loss": 0.0489, + "step": 42458 + }, + { + "epoch": 0.9953058246452997, + "grad_norm": 0.260806143283844, + "learning_rate": 1.0851859962823785e-08, + "loss": 0.0683, + "step": 42459 + }, + { + "epoch": 0.9953292662201046, + "grad_norm": 0.16761930286884308, + "learning_rate": 1.0743614598018869e-08, + "loss": 0.016, + "step": 42460 + }, + { + "epoch": 0.9953527077949097, + "grad_norm": 0.44687795639038086, + "learning_rate": 1.0635911777723273e-08, + "loss": 0.0836, + "step": 42461 + }, + { + "epoch": 0.9953761493697146, + "grad_norm": 0.33832889795303345, + "learning_rate": 1.0528751502536516e-08, + "loss": 0.0268, + "step": 42462 + }, + { + "epoch": 0.9953995909445197, + "grad_norm": 0.5651414394378662, + "learning_rate": 1.0422133773024812e-08, + "loss": 0.0947, + "step": 42463 + }, + { + "epoch": 0.9954230325193246, + "grad_norm": 0.35879501700401306, + "learning_rate": 1.0316058589776579e-08, + "loss": 0.0413, + "step": 42464 + }, + { + "epoch": 0.9954464740941297, + "grad_norm": 0.2883790135383606, + "learning_rate": 1.0210525953369133e-08, + "loss": 0.0641, + "step": 42465 + }, + { + "epoch": 0.9954699156689346, + "grad_norm": 0.3582037687301636, + "learning_rate": 1.0105535864368688e-08, + "loss": 0.0358, + "step": 42466 + }, + { + "epoch": 0.9954933572437397, + "grad_norm": 0.38982686400413513, + "learning_rate": 1.0001088323341457e-08, + "loss": 0.0421, + "step": 42467 + }, + { + "epoch": 0.9955167988185446, + "grad_norm": 0.3586445450782776, + "learning_rate": 9.897183330864757e-09, + "loss": 0.0662, + "step": 42468 + }, + { + "epoch": 0.9955402403933497, + "grad_norm": 0.5329959988594055, + "learning_rate": 9.793820887493698e-09, + "loss": 0.0666, + "step": 42469 + }, + { + "epoch": 0.9955636819681546, + "grad_norm": 0.13490253686904907, + "learning_rate": 9.691000993794497e-09, + "loss": 0.0209, + "step": 42470 + }, + { + "epoch": 0.9955871235429596, + "grad_norm": 0.19508831202983856, + "learning_rate": 9.588723650322262e-09, + "loss": 0.0377, + "step": 42471 + }, + { + "epoch": 0.9956105651177646, + "grad_norm": 0.45077428221702576, + "learning_rate": 9.486988857643208e-09, + "loss": 0.0697, + "step": 42472 + }, + { + "epoch": 0.9956340066925696, + "grad_norm": 0.35965844988822937, + "learning_rate": 9.385796616290244e-09, + "loss": 0.0662, + "step": 42473 + }, + { + "epoch": 0.9956574482673746, + "grad_norm": 0.10555705428123474, + "learning_rate": 9.285146926818478e-09, + "loss": 0.012, + "step": 42474 + }, + { + "epoch": 0.9956808898421796, + "grad_norm": 0.267216295003891, + "learning_rate": 9.185039789783023e-09, + "loss": 0.0559, + "step": 42475 + }, + { + "epoch": 0.9957043314169846, + "grad_norm": 0.448566198348999, + "learning_rate": 9.085475205716787e-09, + "loss": 0.0882, + "step": 42476 + }, + { + "epoch": 0.9957277729917896, + "grad_norm": 0.2904037833213806, + "learning_rate": 8.986453175174881e-09, + "loss": 0.0191, + "step": 42477 + }, + { + "epoch": 0.9957512145665945, + "grad_norm": 0.44151580333709717, + "learning_rate": 8.887973698668006e-09, + "loss": 0.0639, + "step": 42478 + }, + { + "epoch": 0.9957746561413996, + "grad_norm": 0.2763591408729553, + "learning_rate": 8.790036776762379e-09, + "loss": 0.0394, + "step": 42479 + }, + { + "epoch": 0.9957980977162045, + "grad_norm": 0.0910792425274849, + "learning_rate": 8.692642409957596e-09, + "loss": 0.0133, + "step": 42480 + }, + { + "epoch": 0.9958215392910096, + "grad_norm": 0.2317035049200058, + "learning_rate": 8.595790598808773e-09, + "loss": 0.0265, + "step": 42481 + }, + { + "epoch": 0.9958449808658145, + "grad_norm": 0.2686232328414917, + "learning_rate": 8.49948134382661e-09, + "loss": 0.0191, + "step": 42482 + }, + { + "epoch": 0.9958684224406196, + "grad_norm": 0.7309125661849976, + "learning_rate": 8.403714645544014e-09, + "loss": 0.104, + "step": 42483 + }, + { + "epoch": 0.9958918640154245, + "grad_norm": 0.16505369544029236, + "learning_rate": 8.308490504471689e-09, + "loss": 0.0191, + "step": 42484 + }, + { + "epoch": 0.9959153055902296, + "grad_norm": 0.2669267952442169, + "learning_rate": 8.213808921120337e-09, + "loss": 0.0347, + "step": 42485 + }, + { + "epoch": 0.9959387471650345, + "grad_norm": 0.31851688027381897, + "learning_rate": 8.119669896022864e-09, + "loss": 0.4287, + "step": 42486 + }, + { + "epoch": 0.9959621887398395, + "grad_norm": 0.4123251736164093, + "learning_rate": 8.02607342967887e-09, + "loss": 0.5694, + "step": 42487 + }, + { + "epoch": 0.9959856303146445, + "grad_norm": 0.13233451545238495, + "learning_rate": 7.933019522587959e-09, + "loss": 0.0092, + "step": 42488 + }, + { + "epoch": 0.9960090718894495, + "grad_norm": 0.37912020087242126, + "learning_rate": 7.840508175271933e-09, + "loss": 0.0612, + "step": 42489 + }, + { + "epoch": 0.9960325134642545, + "grad_norm": 0.34884342551231384, + "learning_rate": 7.74853938821929e-09, + "loss": 0.0509, + "step": 42490 + }, + { + "epoch": 0.9960559550390595, + "grad_norm": 0.42614489793777466, + "learning_rate": 7.657113161940732e-09, + "loss": 0.0654, + "step": 42491 + }, + { + "epoch": 0.9960793966138645, + "grad_norm": 0.5439199209213257, + "learning_rate": 7.56622949692476e-09, + "loss": 0.6653, + "step": 42492 + }, + { + "epoch": 0.9961028381886695, + "grad_norm": 0.6702876091003418, + "learning_rate": 7.475888393659868e-09, + "loss": 0.0784, + "step": 42493 + }, + { + "epoch": 0.9961262797634746, + "grad_norm": 0.2684859335422516, + "learning_rate": 7.3860898526456615e-09, + "loss": 0.0151, + "step": 42494 + }, + { + "epoch": 0.9961497213382795, + "grad_norm": 0.15680623054504395, + "learning_rate": 7.296833874370634e-09, + "loss": 0.0168, + "step": 42495 + }, + { + "epoch": 0.9961731629130846, + "grad_norm": 0.11899803578853607, + "learning_rate": 7.2081204593121845e-09, + "loss": 0.0264, + "step": 42496 + }, + { + "epoch": 0.9961966044878895, + "grad_norm": 0.4948109984397888, + "learning_rate": 7.119949607947707e-09, + "loss": 0.0626, + "step": 42497 + }, + { + "epoch": 0.9962200460626945, + "grad_norm": 0.1402018666267395, + "learning_rate": 7.032321320776802e-09, + "loss": 0.0183, + "step": 42498 + }, + { + "epoch": 0.9962434876374995, + "grad_norm": 0.4914238154888153, + "learning_rate": 6.94523559824356e-09, + "loss": 0.0678, + "step": 42499 + }, + { + "epoch": 0.9962669292123045, + "grad_norm": 0.656125009059906, + "learning_rate": 6.858692440847581e-09, + "loss": 0.1045, + "step": 42500 + }, + { + "epoch": 0.9962903707871095, + "grad_norm": 0.07688167691230774, + "learning_rate": 6.772691849044055e-09, + "loss": 0.0108, + "step": 42501 + }, + { + "epoch": 0.9963138123619145, + "grad_norm": 0.47394832968711853, + "learning_rate": 6.687233823299277e-09, + "loss": 0.0694, + "step": 42502 + }, + { + "epoch": 0.9963372539367195, + "grad_norm": 0.35621827840805054, + "learning_rate": 6.602318364090643e-09, + "loss": 0.0405, + "step": 42503 + }, + { + "epoch": 0.9963606955115245, + "grad_norm": 0.3811296224594116, + "learning_rate": 6.517945471862241e-09, + "loss": 0.0554, + "step": 42504 + }, + { + "epoch": 0.9963841370863294, + "grad_norm": 0.20396316051483154, + "learning_rate": 6.434115147080366e-09, + "loss": 0.0337, + "step": 42505 + }, + { + "epoch": 0.9964075786611345, + "grad_norm": 0.2380816638469696, + "learning_rate": 6.350827390200209e-09, + "loss": 0.0252, + "step": 42506 + }, + { + "epoch": 0.9964310202359394, + "grad_norm": 0.5195209980010986, + "learning_rate": 6.2680822016769615e-09, + "loss": 0.0857, + "step": 42507 + }, + { + "epoch": 0.9964544618107445, + "grad_norm": 0.3944440186023712, + "learning_rate": 6.18587958194361e-09, + "loss": 0.0709, + "step": 42508 + }, + { + "epoch": 0.9964779033855494, + "grad_norm": 0.2439185082912445, + "learning_rate": 6.1042195314664486e-09, + "loss": 0.0187, + "step": 42509 + }, + { + "epoch": 0.9965013449603545, + "grad_norm": 0.4504718482494354, + "learning_rate": 6.023102050678464e-09, + "loss": 0.0752, + "step": 42510 + }, + { + "epoch": 0.9965247865351594, + "grad_norm": 0.3403731882572174, + "learning_rate": 5.942527140012643e-09, + "loss": 0.0694, + "step": 42511 + }, + { + "epoch": 0.9965482281099645, + "grad_norm": 0.24879314005374908, + "learning_rate": 5.8624947999241785e-09, + "loss": 0.026, + "step": 42512 + }, + { + "epoch": 0.9965716696847694, + "grad_norm": 0.20625512301921844, + "learning_rate": 5.783005030823852e-09, + "loss": 0.0317, + "step": 42513 + }, + { + "epoch": 0.9965951112595745, + "grad_norm": 0.25702327489852905, + "learning_rate": 5.7040578331668534e-09, + "loss": 0.0383, + "step": 42514 + }, + { + "epoch": 0.9966185528343794, + "grad_norm": 0.09272702783346176, + "learning_rate": 5.62565320737507e-09, + "loss": 0.0119, + "step": 42515 + }, + { + "epoch": 0.9966419944091844, + "grad_norm": 0.1952817291021347, + "learning_rate": 5.547791153859283e-09, + "loss": 0.0292, + "step": 42516 + }, + { + "epoch": 0.9966654359839894, + "grad_norm": 0.472404807806015, + "learning_rate": 5.4704716730635815e-09, + "loss": 0.0747, + "step": 42517 + }, + { + "epoch": 0.9966888775587944, + "grad_norm": 0.5323529839515686, + "learning_rate": 5.393694765387647e-09, + "loss": 0.0658, + "step": 42518 + }, + { + "epoch": 0.9967123191335994, + "grad_norm": 0.3567611277103424, + "learning_rate": 5.317460431264465e-09, + "loss": 0.0554, + "step": 42519 + }, + { + "epoch": 0.9967357607084044, + "grad_norm": 0.19129930436611176, + "learning_rate": 5.241768671093717e-09, + "loss": 0.0185, + "step": 42520 + }, + { + "epoch": 0.9967592022832094, + "grad_norm": 0.5768582820892334, + "learning_rate": 5.166619485297286e-09, + "loss": 0.069, + "step": 42521 + }, + { + "epoch": 0.9967826438580144, + "grad_norm": 0.3189857602119446, + "learning_rate": 5.092012874274854e-09, + "loss": 0.0316, + "step": 42522 + }, + { + "epoch": 0.9968060854328193, + "grad_norm": 0.5407537817955017, + "learning_rate": 5.017948838437203e-09, + "loss": 0.0696, + "step": 42523 + }, + { + "epoch": 0.9968295270076244, + "grad_norm": 0.3553767204284668, + "learning_rate": 4.944427378184013e-09, + "loss": 0.0508, + "step": 42524 + }, + { + "epoch": 0.9968529685824293, + "grad_norm": 0.10037284344434738, + "learning_rate": 4.871448493914965e-09, + "loss": 0.0153, + "step": 42525 + }, + { + "epoch": 0.9968764101572344, + "grad_norm": 0.40558701753616333, + "learning_rate": 4.799012186018636e-09, + "loss": 0.0827, + "step": 42526 + }, + { + "epoch": 0.9968998517320393, + "grad_norm": 0.24424687027931213, + "learning_rate": 4.72711845490581e-09, + "loss": 0.0179, + "step": 42527 + }, + { + "epoch": 0.9969232933068444, + "grad_norm": 0.18844956159591675, + "learning_rate": 4.655767300942859e-09, + "loss": 0.0372, + "step": 42528 + }, + { + "epoch": 0.9969467348816493, + "grad_norm": 0.5828653573989868, + "learning_rate": 4.584958724540566e-09, + "loss": 0.1176, + "step": 42529 + }, + { + "epoch": 0.9969701764564544, + "grad_norm": 0.42591169476509094, + "learning_rate": 4.514692726065306e-09, + "loss": 0.0557, + "step": 42530 + }, + { + "epoch": 0.9969936180312593, + "grad_norm": 0.3432867228984833, + "learning_rate": 4.4449693059056556e-09, + "loss": 0.0523, + "step": 42531 + }, + { + "epoch": 0.9970170596060643, + "grad_norm": 0.19387564063072205, + "learning_rate": 4.37578846443909e-09, + "loss": 0.0255, + "step": 42532 + }, + { + "epoch": 0.9970405011808693, + "grad_norm": 0.06978865712881088, + "learning_rate": 4.3071502020430865e-09, + "loss": 0.0055, + "step": 42533 + }, + { + "epoch": 0.9970639427556743, + "grad_norm": 0.5687222480773926, + "learning_rate": 4.239054519084018e-09, + "loss": 0.0692, + "step": 42534 + }, + { + "epoch": 0.9970873843304793, + "grad_norm": 0.809515655040741, + "learning_rate": 4.171501415939361e-09, + "loss": 0.1472, + "step": 42535 + }, + { + "epoch": 0.9971108259052843, + "grad_norm": 0.17893075942993164, + "learning_rate": 4.1044908929643856e-09, + "loss": 0.0168, + "step": 42536 + }, + { + "epoch": 0.9971342674800893, + "grad_norm": 0.369650661945343, + "learning_rate": 4.038022950536568e-09, + "loss": 0.2974, + "step": 42537 + }, + { + "epoch": 0.9971577090548943, + "grad_norm": 0.35162779688835144, + "learning_rate": 3.9720975890111815e-09, + "loss": 0.0411, + "step": 42538 + }, + { + "epoch": 0.9971811506296993, + "grad_norm": 0.6467986106872559, + "learning_rate": 3.906714808743494e-09, + "loss": 0.1024, + "step": 42539 + }, + { + "epoch": 0.9972045922045043, + "grad_norm": 0.3569040298461914, + "learning_rate": 3.841874610088781e-09, + "loss": 0.0528, + "step": 42540 + }, + { + "epoch": 0.9972280337793092, + "grad_norm": 0.387321799993515, + "learning_rate": 3.77757699340231e-09, + "loss": 0.0475, + "step": 42541 + }, + { + "epoch": 0.9972514753541143, + "grad_norm": 0.18786492943763733, + "learning_rate": 3.7138219590282517e-09, + "loss": 0.0182, + "step": 42542 + }, + { + "epoch": 0.9972749169289192, + "grad_norm": 0.23830834031105042, + "learning_rate": 3.650609507321878e-09, + "loss": 0.03, + "step": 42543 + }, + { + "epoch": 0.9972983585037243, + "grad_norm": 0.41022226214408875, + "learning_rate": 3.5879396386051534e-09, + "loss": 0.0492, + "step": 42544 + }, + { + "epoch": 0.9973218000785293, + "grad_norm": 0.28189098834991455, + "learning_rate": 3.525812353244451e-09, + "loss": 0.027, + "step": 42545 + }, + { + "epoch": 0.9973452416533343, + "grad_norm": 0.16841824352741241, + "learning_rate": 3.4642276515506334e-09, + "loss": 0.0341, + "step": 42546 + }, + { + "epoch": 0.9973686832281393, + "grad_norm": 0.38481253385543823, + "learning_rate": 3.4031855338900743e-09, + "loss": 0.0462, + "step": 42547 + }, + { + "epoch": 0.9973921248029443, + "grad_norm": 0.31901565194129944, + "learning_rate": 3.342686000562534e-09, + "loss": 0.0305, + "step": 42548 + }, + { + "epoch": 0.9974155663777493, + "grad_norm": 0.5775046944618225, + "learning_rate": 3.282729051912181e-09, + "loss": 0.0879, + "step": 42549 + }, + { + "epoch": 0.9974390079525542, + "grad_norm": 0.21756388247013092, + "learning_rate": 3.223314688260981e-09, + "loss": 0.02, + "step": 42550 + }, + { + "epoch": 0.9974624495273593, + "grad_norm": 0.6157242655754089, + "learning_rate": 3.164442909930898e-09, + "loss": 0.1244, + "step": 42551 + }, + { + "epoch": 0.9974858911021642, + "grad_norm": 0.5400121212005615, + "learning_rate": 3.106113717243897e-09, + "loss": 0.1744, + "step": 42552 + }, + { + "epoch": 0.9975093326769693, + "grad_norm": 0.3918223977088928, + "learning_rate": 3.0483271105108404e-09, + "loss": 0.042, + "step": 42553 + }, + { + "epoch": 0.9975327742517742, + "grad_norm": 0.1451845020055771, + "learning_rate": 2.991083090053692e-09, + "loss": 0.0116, + "step": 42554 + }, + { + "epoch": 0.9975562158265793, + "grad_norm": 0.5300381183624268, + "learning_rate": 2.9343816561722138e-09, + "loss": 0.0979, + "step": 42555 + }, + { + "epoch": 0.9975796574013842, + "grad_norm": 0.42802587151527405, + "learning_rate": 2.878222809188369e-09, + "loss": 0.068, + "step": 42556 + }, + { + "epoch": 0.9976030989761893, + "grad_norm": 0.5603247284889221, + "learning_rate": 2.8226065494019184e-09, + "loss": 0.1536, + "step": 42557 + }, + { + "epoch": 0.9976265405509942, + "grad_norm": 0.22990985214710236, + "learning_rate": 2.7675328771015197e-09, + "loss": 0.0374, + "step": 42558 + }, + { + "epoch": 0.9976499821257993, + "grad_norm": 0.43425193428993225, + "learning_rate": 2.7130017925980357e-09, + "loss": 0.09, + "step": 42559 + }, + { + "epoch": 0.9976734237006042, + "grad_norm": 0.36460354924201965, + "learning_rate": 2.6590132961912263e-09, + "loss": 0.0314, + "step": 42560 + }, + { + "epoch": 0.9976968652754092, + "grad_norm": 0.39607861638069153, + "learning_rate": 2.6055673881586474e-09, + "loss": 0.586, + "step": 42561 + }, + { + "epoch": 0.9977203068502142, + "grad_norm": 0.36671850085258484, + "learning_rate": 2.5526640688111614e-09, + "loss": 0.4116, + "step": 42562 + }, + { + "epoch": 0.9977437484250192, + "grad_norm": 0.5147928595542908, + "learning_rate": 2.5003033384152218e-09, + "loss": 0.0948, + "step": 42563 + }, + { + "epoch": 0.9977671899998242, + "grad_norm": 0.5612813830375671, + "learning_rate": 2.448485197270589e-09, + "loss": 0.0992, + "step": 42564 + }, + { + "epoch": 0.9977906315746292, + "grad_norm": 0.347146213054657, + "learning_rate": 2.3972096456548187e-09, + "loss": 0.0632, + "step": 42565 + }, + { + "epoch": 0.9978140731494342, + "grad_norm": 0.1831720769405365, + "learning_rate": 2.3464766838343643e-09, + "loss": 0.0292, + "step": 42566 + }, + { + "epoch": 0.9978375147242392, + "grad_norm": 0.14680629968643188, + "learning_rate": 2.296286312097884e-09, + "loss": 0.0215, + "step": 42567 + }, + { + "epoch": 0.9978609562990441, + "grad_norm": 0.5672134757041931, + "learning_rate": 2.246638530711831e-09, + "loss": 0.0815, + "step": 42568 + }, + { + "epoch": 0.9978843978738492, + "grad_norm": 0.4714570939540863, + "learning_rate": 2.1975333399537614e-09, + "loss": 0.5015, + "step": 42569 + }, + { + "epoch": 0.9979078394486541, + "grad_norm": 0.4478397071361542, + "learning_rate": 2.148970740079026e-09, + "loss": 0.0824, + "step": 42570 + }, + { + "epoch": 0.9979312810234592, + "grad_norm": 0.7237768769264221, + "learning_rate": 2.100950731354079e-09, + "loss": 0.1471, + "step": 42571 + }, + { + "epoch": 0.9979547225982641, + "grad_norm": 0.5344544053077698, + "learning_rate": 2.053473314045373e-09, + "loss": 0.5912, + "step": 42572 + }, + { + "epoch": 0.9979781641730692, + "grad_norm": 0.37858936190605164, + "learning_rate": 2.006538488397158e-09, + "loss": 0.0854, + "step": 42573 + }, + { + "epoch": 0.9980016057478741, + "grad_norm": 0.34818771481513977, + "learning_rate": 1.96014625468699e-09, + "loss": 0.0555, + "step": 42574 + }, + { + "epoch": 0.9980250473226792, + "grad_norm": 0.16201572120189667, + "learning_rate": 1.9142966131480144e-09, + "loss": 0.0149, + "step": 42575 + }, + { + "epoch": 0.9980484888974841, + "grad_norm": 0.39903879165649414, + "learning_rate": 1.8689895640244813e-09, + "loss": 0.0541, + "step": 42576 + }, + { + "epoch": 0.9980719304722891, + "grad_norm": 0.6099494099617004, + "learning_rate": 1.8242251075828442e-09, + "loss": 0.1304, + "step": 42577 + }, + { + "epoch": 0.9980953720470941, + "grad_norm": 0.722006618976593, + "learning_rate": 1.7800032440451475e-09, + "loss": 0.0587, + "step": 42578 + }, + { + "epoch": 0.9981188136218991, + "grad_norm": 0.2986277639865875, + "learning_rate": 1.7363239736667425e-09, + "loss": 0.0235, + "step": 42579 + }, + { + "epoch": 0.9981422551967041, + "grad_norm": 0.2810395658016205, + "learning_rate": 1.693187296680776e-09, + "loss": 0.0298, + "step": 42580 + }, + { + "epoch": 0.9981656967715091, + "grad_norm": 0.39467036724090576, + "learning_rate": 1.6505932133092928e-09, + "loss": 0.0491, + "step": 42581 + }, + { + "epoch": 0.9981891383463141, + "grad_norm": 0.3992513418197632, + "learning_rate": 1.6085417238076438e-09, + "loss": 0.033, + "step": 42582 + }, + { + "epoch": 0.9982125799211191, + "grad_norm": 0.4027562439441681, + "learning_rate": 1.5670328283756697e-09, + "loss": 0.3985, + "step": 42583 + }, + { + "epoch": 0.998236021495924, + "grad_norm": 0.4434415400028229, + "learning_rate": 1.526066527257619e-09, + "loss": 0.0584, + "step": 42584 + }, + { + "epoch": 0.9982594630707291, + "grad_norm": 0.39082181453704834, + "learning_rate": 1.485642820675537e-09, + "loss": 0.0865, + "step": 42585 + }, + { + "epoch": 0.998282904645534, + "grad_norm": 0.1111859381198883, + "learning_rate": 1.4457617088403653e-09, + "loss": 0.0274, + "step": 42586 + }, + { + "epoch": 0.9983063462203391, + "grad_norm": 0.22020696103572845, + "learning_rate": 1.4064231919741488e-09, + "loss": 0.0412, + "step": 42587 + }, + { + "epoch": 0.998329787795144, + "grad_norm": 0.048209819942712784, + "learning_rate": 1.3676272702878301e-09, + "loss": 0.0068, + "step": 42588 + }, + { + "epoch": 0.9983532293699491, + "grad_norm": 0.6140576004981995, + "learning_rate": 1.3293739439923514e-09, + "loss": 0.0919, + "step": 42589 + }, + { + "epoch": 0.998376670944754, + "grad_norm": 0.3726675510406494, + "learning_rate": 1.291663213298655e-09, + "loss": 0.0703, + "step": 42590 + }, + { + "epoch": 0.9984001125195591, + "grad_norm": 0.4014558792114258, + "learning_rate": 1.254495078395479e-09, + "loss": 0.0426, + "step": 42591 + }, + { + "epoch": 0.998423554094364, + "grad_norm": 0.5104010701179504, + "learning_rate": 1.2178695395048678e-09, + "loss": 0.0717, + "step": 42592 + }, + { + "epoch": 0.9984469956691691, + "grad_norm": 0.5288566946983337, + "learning_rate": 1.1817865968266616e-09, + "loss": 0.0985, + "step": 42593 + }, + { + "epoch": 0.998470437243974, + "grad_norm": 0.503233790397644, + "learning_rate": 1.1462462505384963e-09, + "loss": 0.1366, + "step": 42594 + }, + { + "epoch": 0.998493878818779, + "grad_norm": 0.5993900299072266, + "learning_rate": 1.1112485008402118e-09, + "loss": 0.093, + "step": 42595 + }, + { + "epoch": 0.9985173203935841, + "grad_norm": 0.7485541105270386, + "learning_rate": 1.0767933479316483e-09, + "loss": 0.1915, + "step": 42596 + }, + { + "epoch": 0.998540761968389, + "grad_norm": 0.5405333042144775, + "learning_rate": 1.0428807919904416e-09, + "loss": 0.454, + "step": 42597 + }, + { + "epoch": 0.9985642035431941, + "grad_norm": 0.49144184589385986, + "learning_rate": 1.0095108331942271e-09, + "loss": 0.082, + "step": 42598 + }, + { + "epoch": 0.998587645117999, + "grad_norm": 0.4338247776031494, + "learning_rate": 9.766834717428453e-10, + "loss": 0.0315, + "step": 42599 + }, + { + "epoch": 0.9986110866928041, + "grad_norm": 0.22782734036445618, + "learning_rate": 9.443987077917272e-10, + "loss": 0.05, + "step": 42600 + }, + { + "epoch": 0.998634528267609, + "grad_norm": 0.7476584315299988, + "learning_rate": 9.126565415407129e-10, + "loss": 0.1644, + "step": 42601 + }, + { + "epoch": 0.9986579698424141, + "grad_norm": 0.15397167205810547, + "learning_rate": 8.814569731452337e-10, + "loss": 0.0313, + "step": 42602 + }, + { + "epoch": 0.998681411417219, + "grad_norm": 0.2607061564922333, + "learning_rate": 8.508000027718233e-10, + "loss": 0.0315, + "step": 42603 + }, + { + "epoch": 0.998704852992024, + "grad_norm": 0.5141978859901428, + "learning_rate": 8.20685630598117e-10, + "loss": 0.6287, + "step": 42604 + }, + { + "epoch": 0.998728294566829, + "grad_norm": 0.1831005960702896, + "learning_rate": 7.911138567906484e-10, + "loss": 0.0375, + "step": 42605 + }, + { + "epoch": 0.998751736141634, + "grad_norm": 0.31689268350601196, + "learning_rate": 7.620846814937466e-10, + "loss": 0.0282, + "step": 42606 + }, + { + "epoch": 0.998775177716439, + "grad_norm": 0.5289425253868103, + "learning_rate": 7.335981048739449e-10, + "loss": 0.129, + "step": 42607 + }, + { + "epoch": 0.998798619291244, + "grad_norm": 0.3172128200531006, + "learning_rate": 7.056541270866746e-10, + "loss": 0.0433, + "step": 42608 + }, + { + "epoch": 0.998822060866049, + "grad_norm": 0.3604276478290558, + "learning_rate": 6.782527482873669e-10, + "loss": 0.0424, + "step": 42609 + }, + { + "epoch": 0.998845502440854, + "grad_norm": 0.375817209482193, + "learning_rate": 6.513939686092485e-10, + "loss": 0.0673, + "step": 42610 + }, + { + "epoch": 0.998868944015659, + "grad_norm": 0.4642806947231293, + "learning_rate": 6.25077788218853e-10, + "loss": 0.0622, + "step": 42611 + }, + { + "epoch": 0.998892385590464, + "grad_norm": 0.31393784284591675, + "learning_rate": 5.993042072383048e-10, + "loss": 0.033, + "step": 42612 + }, + { + "epoch": 0.9989158271652689, + "grad_norm": 0.4363534450531006, + "learning_rate": 5.740732258230353e-10, + "loss": 0.0819, + "step": 42613 + }, + { + "epoch": 0.998939268740074, + "grad_norm": 0.5292100310325623, + "learning_rate": 5.49384844106271e-10, + "loss": 0.07, + "step": 42614 + }, + { + "epoch": 0.9989627103148789, + "grad_norm": 0.4064843952655792, + "learning_rate": 5.252390622212388e-10, + "loss": 0.0237, + "step": 42615 + }, + { + "epoch": 0.998986151889684, + "grad_norm": 0.1633068472146988, + "learning_rate": 5.016358802900634e-10, + "loss": 0.015, + "step": 42616 + }, + { + "epoch": 0.9990095934644889, + "grad_norm": 0.495906800031662, + "learning_rate": 4.785752984570735e-10, + "loss": 0.0489, + "step": 42617 + }, + { + "epoch": 0.999033035039294, + "grad_norm": 0.09490009397268295, + "learning_rate": 4.5605731683329157e-10, + "loss": 0.0236, + "step": 42618 + }, + { + "epoch": 0.9990564766140989, + "grad_norm": 0.10305051505565643, + "learning_rate": 4.3408193555194433e-10, + "loss": 0.015, + "step": 42619 + }, + { + "epoch": 0.999079918188904, + "grad_norm": 0.4332115650177002, + "learning_rate": 4.1264915472405407e-10, + "loss": 0.3757, + "step": 42620 + }, + { + "epoch": 0.9991033597637089, + "grad_norm": 0.12587349116802216, + "learning_rate": 3.917589744717454e-10, + "loss": 0.0209, + "step": 42621 + }, + { + "epoch": 0.999126801338514, + "grad_norm": 0.4477699398994446, + "learning_rate": 3.714113949060405e-10, + "loss": 0.1036, + "step": 42622 + }, + { + "epoch": 0.9991502429133189, + "grad_norm": 0.8500177264213562, + "learning_rate": 3.516064161268595e-10, + "loss": 0.1985, + "step": 42623 + }, + { + "epoch": 0.9991736844881239, + "grad_norm": 0.641930103302002, + "learning_rate": 3.3234403825632696e-10, + "loss": 0.1303, + "step": 42624 + }, + { + "epoch": 0.9991971260629289, + "grad_norm": 0.300460547208786, + "learning_rate": 3.136242614054652e-10, + "loss": 0.0371, + "step": 42625 + }, + { + "epoch": 0.9992205676377339, + "grad_norm": 0.44313329458236694, + "learning_rate": 2.9544708565198975e-10, + "loss": 0.3195, + "step": 42626 + }, + { + "epoch": 0.9992440092125389, + "grad_norm": 0.3754320740699768, + "learning_rate": 2.7781251110692297e-10, + "loss": 0.0663, + "step": 42627 + }, + { + "epoch": 0.9992674507873439, + "grad_norm": 0.26968735456466675, + "learning_rate": 2.607205378701849e-10, + "loss": 0.0543, + "step": 42628 + }, + { + "epoch": 0.9992908923621489, + "grad_norm": 0.0893096774816513, + "learning_rate": 2.4417116603059344e-10, + "loss": 0.0099, + "step": 42629 + }, + { + "epoch": 0.9993143339369539, + "grad_norm": 0.12848104536533356, + "learning_rate": 2.2816439567696635e-10, + "loss": 0.0185, + "step": 42630 + }, + { + "epoch": 0.9993377755117588, + "grad_norm": 0.5980463027954102, + "learning_rate": 2.1270022689812152e-10, + "loss": 0.0662, + "step": 42631 + }, + { + "epoch": 0.9993612170865639, + "grad_norm": 0.33522093296051025, + "learning_rate": 1.977786597828768e-10, + "loss": 0.0405, + "step": 42632 + }, + { + "epoch": 0.9993846586613688, + "grad_norm": 0.17957447469234467, + "learning_rate": 1.8339969439784554e-10, + "loss": 0.0303, + "step": 42633 + }, + { + "epoch": 0.9994081002361739, + "grad_norm": 0.26919516921043396, + "learning_rate": 1.695633308318456e-10, + "loss": 0.0476, + "step": 42634 + }, + { + "epoch": 0.9994315418109788, + "grad_norm": 0.700171947479248, + "learning_rate": 1.5626956915149037e-10, + "loss": 0.1817, + "step": 42635 + }, + { + "epoch": 0.9994549833857839, + "grad_norm": 0.10965391993522644, + "learning_rate": 1.435184094455977e-10, + "loss": 0.0218, + "step": 42636 + }, + { + "epoch": 0.9994784249605888, + "grad_norm": 0.12758459150791168, + "learning_rate": 1.313098517696787e-10, + "loss": 0.0142, + "step": 42637 + }, + { + "epoch": 0.9995018665353939, + "grad_norm": 0.39541488885879517, + "learning_rate": 1.1964389619034678e-10, + "loss": 0.0255, + "step": 42638 + }, + { + "epoch": 0.9995253081101988, + "grad_norm": 0.5011173486709595, + "learning_rate": 1.085205427742153e-10, + "loss": 0.0657, + "step": 42639 + }, + { + "epoch": 0.9995487496850038, + "grad_norm": 0.4893856644630432, + "learning_rate": 9.793979157679545e-11, + "loss": 0.0651, + "step": 42640 + }, + { + "epoch": 0.9995721912598088, + "grad_norm": 0.3475714921951294, + "learning_rate": 8.790164266470058e-11, + "loss": 0.0497, + "step": 42641 + }, + { + "epoch": 0.9995956328346138, + "grad_norm": 0.6080512404441833, + "learning_rate": 7.840609608233963e-11, + "loss": 0.1509, + "step": 42642 + }, + { + "epoch": 0.9996190744094188, + "grad_norm": 0.7854772210121155, + "learning_rate": 6.945315189632595e-11, + "loss": 0.0855, + "step": 42643 + }, + { + "epoch": 0.9996425159842238, + "grad_norm": 0.3737565279006958, + "learning_rate": 6.104281013996627e-11, + "loss": 0.042, + "step": 42644 + }, + { + "epoch": 0.9996659575590288, + "grad_norm": 0.7229587435722351, + "learning_rate": 5.3175070857669483e-11, + "loss": 0.1216, + "step": 42645 + }, + { + "epoch": 0.9996893991338338, + "grad_norm": 0.21503369510173798, + "learning_rate": 4.584993410494676e-11, + "loss": 0.0496, + "step": 42646 + }, + { + "epoch": 0.9997128407086389, + "grad_norm": 0.6082687377929688, + "learning_rate": 3.906739991510477e-11, + "loss": 0.6055, + "step": 42647 + }, + { + "epoch": 0.9997362822834438, + "grad_norm": 0.4781608283519745, + "learning_rate": 3.282746833255246e-11, + "loss": 0.0369, + "step": 42648 + }, + { + "epoch": 0.9997597238582488, + "grad_norm": 0.32074275612831116, + "learning_rate": 2.7130139368392037e-11, + "loss": 0.0566, + "step": 42649 + }, + { + "epoch": 0.9997831654330538, + "grad_norm": 0.2633090615272522, + "learning_rate": 2.1975413078134666e-11, + "loss": 0.0328, + "step": 42650 + }, + { + "epoch": 0.9998066070078588, + "grad_norm": 0.3968164622783661, + "learning_rate": 1.7363289483984802e-11, + "loss": 0.0327, + "step": 42651 + }, + { + "epoch": 0.9998300485826638, + "grad_norm": 0.3332281708717346, + "learning_rate": 1.3293768597044675e-11, + "loss": 0.0261, + "step": 42652 + }, + { + "epoch": 0.9998534901574688, + "grad_norm": 0.2560446858406067, + "learning_rate": 9.766850461723209e-12, + "loss": 0.0331, + "step": 42653 + }, + { + "epoch": 0.9998769317322738, + "grad_norm": 0.512607991695404, + "learning_rate": 6.782535078020402e-12, + "loss": 0.0396, + "step": 42654 + }, + { + "epoch": 0.9999003733070788, + "grad_norm": 0.17004559934139252, + "learning_rate": 4.340822468140715e-12, + "loss": 0.0301, + "step": 42655 + }, + { + "epoch": 0.9999238148818838, + "grad_norm": 0.30125752091407776, + "learning_rate": 2.441712643186378e-12, + "loss": 0.0615, + "step": 42656 + }, + { + "epoch": 0.9999472564566888, + "grad_norm": 0.7820804119110107, + "learning_rate": 1.0852056253618514e-12, + "loss": 0.1606, + "step": 42657 + }, + { + "epoch": 0.9999706980314937, + "grad_norm": 0.37642017006874084, + "learning_rate": 2.713014035649053e-13, + "loss": 0.0427, + "step": 42658 + }, + { + "epoch": 0.9999941396062988, + "grad_norm": 0.2449454963207245, + "learning_rate": 0.0, + "loss": 0.074, + "step": 42659 + } + ], + "logging_steps": 1, + "max_steps": 42659, + "num_input_tokens_seen": 0, + "num_train_epochs": 1, + "save_steps": 500, + "stateful_callbacks": { + "TrainerControl": { + "args": { + "should_epoch_stop": false, + "should_evaluate": false, + "should_log": false, + "should_save": true, + "should_training_stop": true + }, + "attributes": {} + } + }, + "total_flos": 3.6836648442459587e+18, + "train_batch_size": 2, + "trial_name": null, + "trial_params": null +}